March 15, 2024, 4:45 a.m. | Hyung-Il Kim, Kimin Yun, Jun-Seok Yun, Yuseok Bae

cs.CV updates on arXiv.org arxiv.org

arXiv:2403.09199v1 Announce Type: new
Abstract: Recently, foundation models trained on massive datasets to adapt to a wide range of domains have attracted considerable attention and are actively being explored within the computer vision community. Among these, the Segment Anything Model (SAM) stands out for its remarkable progress in generalizability and flexibility for image segmentation tasks, achieved through prompt-based object mask generation. However, despite its strength, SAM faces two key limitations when applied to customized instance segmentation that segments specific objects …

abstract adapt arxiv attention community computer computer vision cs.ai cs.cv datasets domains foundation foundation model instance massive progress prompt prompt learning sam segment segment anything segment anything model segmentation type via vision

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne