Nov. 11, 2023, 5:58 a.m. | Essam Wisam

Towards Data Science - Medium towardsdatascience.com

Learn about text-to-speech and how it’s realized by transformers

In 2019, FastSpeech has pushed the frontier of neural text-to-speech by offering significant improvement in inference speed while maintaining robustness to prevent word repetition or omission. It also allowed for controllability of the output speech in terms of speech and prosody.

In this story, we aim to familiarize you with how transformers are employed for text-to-speech, provide you with a concise overview of the FastSpeech paper and point you to how …

aim deep learning implementation improvement inference learn nlp overview paper robustness speech speed story terms text text-to-speech thoughts-and-theory transformers word

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne