Nov. 15, 2022, 2:16 a.m. | Di Wang, Qiming Zhang, Yufei Xu, Jing Zhang, Bo Du, Dacheng Tao, Liangpei Zhang

cs.CV updates on arXiv.org arxiv.org

Large-scale vision foundation models have made significant progress in visual
tasks on natural images, with vision transformers being the primary choice due
to their good scalability and representation ability. However, large-scale
models in remote sensing (RS) have not yet been sufficiently explored. In this
paper, we resort to plain vision transformers with about 100 million parameters
and make the first attempt to propose large vision models tailored to RS tasks
and investigate how such large models perform. To handle the …

arxiv foundation model remote sensing transformer vision

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne