Feb. 1, 2024, 12:41 p.m. | Peili Chen Linyang He Li Fu Lu Fan Edward F. Chang Yuanning Li

cs.CL updates on arXiv.org arxiv.org

Speech and language models trained through self-supervised learning (SSL) demonstrate strong alignment with brain activity during speech and language perception. However, given their distinct training modalities, it remains unclear whether they correlate with the same neural aspects. We directly address this question by evaluating the brain prediction performance of two representative SSL models, Wav2Vec2.0 and GPT-2, designed for speech and language tasks. Our findings reveal that both models accurately predict speech responses in the auditory cortex, with a significant correlation …

alignment brain brain activity cs.ai cs.cl eess.as extract human language language models perception performance prediction q-bio.nc question self-supervised learning speech ssl supervised learning through training

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne