all AI news
Cross-Level Multi-Instance Distillation for Self-Supervised Fine-Grained Visual Categorization
Feb. 28, 2024, 5:47 a.m. | Qi Bi, Wei Ji, Jingjun Yi, Haolan Zhan, Gui-Song Xia
cs.CV updates on arXiv.org arxiv.org
Abstract: High-quality annotation of fine-grained visual categories demands great expert knowledge, which is taxing and time consuming. Alternatively, learning fine-grained visual representation from enormous unlabeled images (e.g., species, brands) by self-supervised learning becomes a feasible solution. However, recent researches find that existing self-supervised learning methods are less qualified to represent fine-grained categories. The bottleneck lies in that the pre-text representation is built from every patch-wise embedding, while fine-grained categories are only determined by several key patches …
abstract annotation arxiv brands cs.cv distillation expert fine-grained images instance knowledge quality representation self-supervised learning solution supervised learning type visual
More from arxiv.org / cs.CV updates on arXiv.org
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
#13721 - Data Engineer - AI Model Testing
@ Qualitest | Miami, Florida, United States
Elasticsearch Administrator
@ ManTech | 201BF - Customer Site, Chantilly, VA