Web: http://arxiv.org/abs/2110.04946

Jan. 27, 2022, 2:11 a.m. | Hieu-Thi Luong, Junichi Yamagishi

cs.LG updates on arXiv.org arxiv.org

Emotional and controllable speech synthesis is a topic that has received much
attention. However, most studies focused on improving the expressiveness and
controllability in the context of linguistic content, even though natural
verbal human communication is inseparable from spontaneous non-speech
expressions such as laughter, crying, or grunting. We propose a model called
LaughNet for synthesizing laughter by using waveform silhouettes as inputs. The
motivation is not simply synthesizing new laughter utterances, but testing a
novel synthesis-control paradigm that uses an …


More from arxiv.org / cs.LG updates on arXiv.org

Data Analytics and Technical support Lead

@ Coupa Software, Inc. | Bogota, Colombia

Data Science Manager

@ Vectra | San Jose, CA

Data Analyst Sr

@ Capco | Brazil - Sao Paulo

Data Scientist (NLP)

@ Builder.ai | London, England, United Kingdom - Remote

Senior Data Analyst

@ BuildZoom | Scottsdale, AZ/ San Francisco, CA/ Remote

Senior Research Scientist, Speech Recognition

@ SoundHound Inc. | Toronto, Canada