Oct. 12, 2022, 1:17 a.m. | Gaëlle Laperrière, Valentin Pelloin, Mickaël Rouvier, Themos Stafylakis, Yannick Estève

cs.CL updates on arXiv.org arxiv.org

In this paper we examine the use of semantically-aligned speech
representations for end-to-end spoken language understanding (SLU). We employ
the recently-introduced SAMU-XLSR model, which is designed to generate a single
embedding that captures the semantics at the utterance level, semantically
aligned across different languages. This model combines the acoustic
frame-level speech representation learning model (XLS-R) with the Language
Agnostic BERT Sentence Embedding (LaBSE) model. We show that the use of the
SAMU-XLSR model instead of the initial XLS-R model improves …

arxiv language speech spoken language understanding understanding

Software Engineer for AI Training Data (School Specific)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Python)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Tier 2)

@ G2i Inc | Remote

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US