July 25, 2022, 1:12 a.m. | Yu Zhang, Daniel S. Park, Wei Han, James Qin, Anmol Gulati, Joel Shor, Aren Jansen, Yuanzhong Xu, Yanping Huang, Shibo Wang, Zongwei Zhou, Bo Li, Min

cs.CL updates on arXiv.org arxiv.org

We summarize the results of a host of efforts using giant automatic speech
recognition (ASR) models pre-trained using large, diverse unlabeled datasets
containing approximately a million hours of audio. We find that the combination
of pre-training, self-training and scaling up model size greatly increases data
efficiency, even for extremely large tasks with tens of thousands of hours of
labeled data. In particular, on an ASR task with 34k hours of labeled data, by
fine-tuning an 8 billion parameter pre-trained Conformer …

arxiv automatic speech recognition learning scale semi-supervised semi-supervised learning speech speech recognition supervised learning

Data Scientist (m/f/x/d)

@ Symanto Research GmbH & Co. KG | Spain, Germany

Research Analyst

@ Cypris | Los Angeles, California, United States

Data Manager H/F

@ ASSYSTEM | Courbevoie, France

Software Engineer III - Java Scala BigData AWS

@ JPMorgan Chase & Co. | Bengaluru, Karnataka, India

Reference Data Specialist

@ JPMorgan Chase & Co. | Bengaluru, Karnataka, India

Data Visualization Manager

@ PatientPoint | Cincinnati, Ohio, United States