March 7, 2024, 5:43 a.m. | Wonjoon Chang, Dahee Kwon, Jaesik Choi

cs.LG updates on arXiv.org arxiv.org

arXiv:2312.17285v2 Announce Type: replace-cross
Abstract: Understanding intermediate representations of the concepts learned by deep learning classifiers is indispensable for interpreting general model behaviors. Existing approaches to reveal learned concepts often rely on human supervision, such as pre-defined concept sets or segmentation processes. In this paper, we propose a novel unsupervised method for discovering distributed representations of concepts by selecting a principal subset of neurons. Our empirical findings demonstrate that instances with similar neuron activation states tend to share coherent concepts. …

abstract arxiv classifiers concept concepts cs.ai cs.cv cs.lg deep learning distributed general human intermediate networks neural networks novel paper processes segmentation supervision type understanding

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne