all AI news
ViT-MUL: A Baseline Study on Recent Machine Unlearning Methods Applied to Vision Transformers
March 18, 2024, 4:41 a.m. | Ikhyun Cho, Changyeon Park, Julia Hockenmaier
cs.LG updates on arXiv.org arxiv.org
Abstract: Machine unlearning (MUL) is an arising field in machine learning that seeks to erase the learned information of specific training data points from a trained model. Despite the recent active research in MUL within computer vision, the majority of work has focused on ResNet-based models. Given that Vision Transformers (ViT) have become the predominant model architecture, a detailed study of MUL specifically tailored to ViT is essential. In this paper, we present comprehensive experiments on …
abstract arxiv computer computer vision cs.cv cs.lg data information machine machine learning research study training training data transformers type unlearning vision vision transformers vit work
More from arxiv.org / cs.LG updates on arXiv.org
Testing the Segment Anything Model on radiology data
1 day, 4 hours ago |
arxiv.org
Calorimeter shower superresolution
1 day, 4 hours ago |
arxiv.org
Jobs in AI, ML, Big Data
Software Engineer for AI Training Data (School Specific)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Python)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Tier 2)
@ G2i Inc | Remote
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US