April 15, 2024, 4:42 a.m. | Josie Harrison, Alexander Hollberg, Yinan Yu

cs.LG updates on arXiv.org arxiv.org

arXiv:2404.08557v1 Announce Type: cross
Abstract: Computer vision models trained on Google Street View images can create material cadastres. However, current approaches need manually annotated datasets that are difficult to obtain and often have class imbalance. To address these challenges, this paper fine-tuned a Swin Transformer model on a synthetic dataset generated with DALL-E and compared the performance to a similar manually annotated dataset. Although manual annotation remains the gold standard, the synthetic dataset performance demonstrates a reasonable alternative. The findings …

abstract annotation arxiv building challenges class classification computer computer vision create cs.cv cs.lg current data data annotation datasets google however images material paper scalability street swin swin transformer synthetic synthetic data transformer transformer model type view vision vision models

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Senior Data Scientist

@ ITE Management | New York City, United States