Home  | Publications | LAW+24a

CromSS: Cross-Modal Pre-Training With Noisy Labels for Remote Sensing Image Segmentation

MCML Authors

Abstract

We study the potential of noisy labels y to pretrain semantic segmentation models in a multi-modal learning framework for geospatial applications. Specifically, we propose a novel Cross-modal Sample Selection method (CromSS) that utilizes the class distributions P^{(d)}(x,c) over pixels x and classes c modelled by multiple sensors/modalities d of a given geospatial scene. Consistency of predictions across sensors d is jointly informed by the entropy of P^{(d)}(x,c). Noisy label sampling we determine by the confidence of each sensor d in the noisy class label, P^{(d)}(x,c=y(x)). To verify the performance of our approach, we conduct experiments with Sentinel-1 (radar) and Sentinel-2 (optical) satellite imagery from the globally-sampled SSL4EO-S12 dataset. We pair those scenes with 9-class noisy labels sourced from the Google Dynamic World project for pretraining. Transfer learning evaluations (downstream task) on the DFC2020 dataset confirm the effectiveness of the proposed method for remote sensing image segmentation.

inproceedings


ML4RS @ICLR 2024

2nd Workshop Machine Learning for Remote Sensing at the 12th International Conference on Learning Representations. Vienna, Austria, May 07-11, 2024.

Authors

C. Liu • C. M. Albrecht • Y. Wang • X. Zhu

Links

PDF

Research Area

 C3 | Physics and Geo Sciences

BibTeXKey: LAW+24a

Back to Top