April 15, 2024, 4:45 a.m. | Guillaume Astruc, Nicolas Gonthier, Clement Mallet, Loic Landrieu

cs.CV updates on arXiv.org arxiv.org

arXiv:2404.08351v1 Announce Type: new
Abstract: The field of Earth Observations (EO) offers a wealth of data from diverse sensors, presenting a great opportunity for advancing self-supervised multimodal learning. However, current multimodal EO datasets and models focus on a single data type, either mono-date images or time series, which limits their expressivity. We introduce OmniSat, a novel architecture that exploits the spatial alignment between multiple EO modalities to learn expressive multimodal representations without labels. To demonstrate the advantages of combining modalities …

abstract arxiv cs.cv current data datasets diverse earth earth observation focus fusion however images multimodal multimodal learning observation presenting sensors series time series type wealth

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Data Analyst (Digital Business Analyst)

@ Activate Interactive Pte Ltd | Singapore, Central Singapore, Singapore