all AI news
What Do Language Models Hear? Probing for Auditory Representations in Language Models
Feb. 28, 2024, 5:42 a.m. | Jerry Ngo, Yoon Kim
cs.LG updates on arXiv.org arxiv.org
Abstract: This work explores whether language models encode meaningfully grounded representations of sounds of objects. We learn a linear probe that retrieves the correct text representation of an object given a snippet of audio related to that object, where the sound representation is given by a pretrained audio model. This probe is trained via a contrastive loss that pushes the language representations and sound representations of an object to be close to one another. After training, …
abstract arxiv audio cs.ai cs.cl cs.lg cs.sd eess.as encode language language models learn linear objects probe representation sound text type work
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US
Research Engineer
@ Allora Labs | Remote
Ecosystem Manager
@ Allora Labs | Remote
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US