April 20, 2022, 1:12 a.m. | Yingzhi Wang, Abdelmoumene Boumadane, Abdelwahab Heba

cs.CL updates on arXiv.org arxiv.org

Speech self-supervised models such as wav2vec 2.0 and HuBERT are making
revolutionary progress in Automatic Speech Recognition (ASR). However, they
have not been totally proved to produce better performance on tasks other than
ASR. In this work, we explore partial fine-tuning and entire fine-tuning on
wav2vec 2.0 and HuBERT pre-trained models for three non-ASR speech tasks :
Speech Emotion Recognition, Speaker Verification and Spoken Language
Understanding. With simple proposed down-stream frameworks, the best scores
reach 79.58% weighted accuracy for Speech …

arxiv benchmark emotion language speech spoken language understanding understanding verification

Software Engineer for AI Training Data (School Specific)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Python)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Tier 2)

@ G2i Inc | Remote

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US