April 24, 2024, 4:43 a.m. | Chaoyue Ding, Kunchi Li, Jun Wan, Shan Yu

cs.LG updates on arXiv.org arxiv.org

arXiv:2301.05180v2 Announce Type: replace
Abstract: Rehearsal approaches in class incremental learning (CIL) suffer from decision boundary overfitting to new classes, which is mainly caused by two factors: insufficiency of old classes data for knowledge distillation and imbalanced data learning between the learned and new classes because of the limited storage memory. In this work, we present a simple but effective approach to tackle these two factors. First, we employ a re-sampling strategy and Mixup K}nowledge D}istillation (Re-MKD) to improve the …

abstract arxiv class cs.cv cs.lg data decision distillation incremental knowledge memory overfitting storage type

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote