all AI news
Online Continual Learning of End-to-End Speech Recognition Models. (arXiv:2207.05071v1 [cs.LG])
July 13, 2022, 1:10 a.m. | Muqiao Yang, Ian Lane, Shinji Watanabe
cs.LG updates on arXiv.org arxiv.org
Continual Learning, also known as Lifelong Learning, aims to continually
learn from new data as it becomes available. While prior research on continual
learning in automatic speech recognition has focused on the adaptation of
models across multiple different speech recognition tasks, in this paper we
propose an experimental setting for \textit{online continual learning} for
automatic speech recognition of a single task. Specifically focusing on the
case where additional training data for the same task becomes available
incrementally over time, we …
arxiv continual learning lg speech speech recognition speech recognition models
More from arxiv.org / cs.LG updates on arXiv.org
Efficient Data-Driven MPC for Demand Response of Commercial Buildings
2 days, 11 hours ago |
arxiv.org
Testing the Segment Anything Model on radiology data
2 days, 11 hours ago |
arxiv.org
Calorimeter shower superresolution
2 days, 11 hours ago |
arxiv.org
Jobs in AI, ML, Big Data
Software Engineer for AI Training Data (School Specific)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Python)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Tier 2)
@ G2i Inc | Remote
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US