March 12, 2024, 4:47 a.m. | Thang Doan, Sima Behpour, Xin Li, Wenbin He, Liang Gou, Liu Ren

cs.CV updates on arXiv.org arxiv.org

arXiv:2403.06295v1 Announce Type: new
Abstract: Few-shot Class-Incremental Learning (FSCIL) poses the challenge of retaining prior knowledge while learning from limited new data streams, all without overfitting. The rise of Vision-Language models (VLMs) has unlocked numerous applications, leveraging their existing knowledge to fine-tune on custom data. However, training the whole model is computationally prohibitive, and VLMs while being versatile in general domains still struggle with fine-grained datasets crucial for many applications. We tackle these challenges with two proposed simple modules. The …

abstract applications arxiv challenge class cs.cv data datasets data streams few-shot fine-grained however incremental knowledge language language models multimodal overfitting prior training type unlocked vision vision-language models vlms

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US