Jan. 31, 2024, 4:41 p.m. | Thuy-Trang Vu, Shahram Khadivi, Mahsa Ghorbanali, Dinh Phung, Gholamreza Haffari

cs.CL updates on arXiv.org arxiv.org

Acquiring new knowledge without forgetting what has been learned in a
sequence of tasks is the central focus of continual learning (CL). While tasks
arrive sequentially, the training data are often prepared and annotated
independently, leading to the CL of incoming supervised learning tasks. This
paper considers the under-explored problem of active continual learning (ACL)
for a sequence of active learning (AL) tasks, where each incoming task includes
a pool of unlabelled data and an annotation budget. We investigate the …

arxiv continual cs.lg data focus knowledge paper retention supervised learning tasks training training data

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Robotics Technician - 3rd Shift

@ GXO Logistics | Perris, CA, US, 92571