Web: http://arxiv.org/abs/2206.07860

June 17, 2022, 1:10 a.m. | Li-Chin Chen, Po-Hsun Chen, Richard Tzong-Han Tsai, Yu Tsao

cs.LG updates on arXiv.org arxiv.org

Speech generation and enhancement based on articulatory movements facilitate
communication when the scope of verbal communication is absent, e.g., in
patients who have lost the ability to speak. Although various techniques have
been proposed to this end, electropalatography (EPG), which is a monitoring
technique that records contact between the tongue and hard palate during
speech, has not been adequately explored. Herein, we propose a novel multimodal
EPG-to-speech (EPG2S) system that utilizes EPG and speech signals for speech
generation and enhancement. …

arxiv audio generation learning multimodal multimodal learning on speech

More from arxiv.org / cs.LG updates on arXiv.org

Machine Learning Researcher - Saalfeld Lab

@ Howard Hughes Medical Institute - Chevy Chase, MD | Ashburn, Virginia

Project Director, Machine Learning in US Health

@ ideas42.org | Remote, US

Data Science Intern

@ NannyML | Remote

Machine Learning Engineer NLP/Speech

@ Play.ht | Remote

Research Scientist, 3D Reconstruction

@ Yembo | Remote, US

Clinical Assistant or Associate Professor of Management Science and Systems

@ University at Buffalo | Buffalo, NY