Seamless: Multilingual Expressive and Streaming Speech Translation

Seamless Communication,Loïc Barrault,Yu-An Chung, Mariano Coria Meglioli,David Dale,Ning Dong, Mark Duppenthaler,Paul-Ambroise Duquenne,Brian Ellis, Hady Elsahar, Justin Haaheim, John Hoffman, Min-Jae Hwang,Hirofumi Inaguma, Christopher Klaiber,Ilia Kulikov,Pengwei Li,Daniel Licht,Jean Maillard, Ruslan Mavlyutov, Alice Rakotoarison,Kaushik Ram Sadagopan, Abinesh Ramakrishnan, Tuan Tran,Guillaume Wenzek,Yilin Yang, Ethan Ye,Ivan Evtimov, Pierre Fernandez,Cynthia Gao, Prangthip Hansanti, Elahe Kalbassi, Amanda Kallet, Artyom Kozhevnikov, Gabriel Mejia Gonzalez, Robin San Roman, Christophe Touret, Corinne Wong, Carleigh Wood, Bokai Yu,Pierre Andrews, Can Balioglu,Peng-Jen Chen,Marta R. Costa-jussà,Maha Elbayad,Hongyu Gong, Francisco Guzmán, Kevin Heffernan,Somya Jain,Justine Kao, Ann Lee,Xutai Ma, Alex Mourachko, Benjamin Peloquin, Juan Pino, Sravya Popuri, Christophe Ropers, Safiyyah Saleem, Holger Schwenk, Anna Sun, Paden Tomasello, Changhan Wang, Jeff Wang, Skyler Wang, Mary Williamson

CoRR(2023)

引用 0|浏览22
暂无评分
摘要
Large-scale automatic speech translation systems today lack key features that help machine-mediated communication feel seamless when compared to human-to-human dialogue. In this work, we introduce a family of models that enable end-to-end expressive and multilingual translations in a streaming fashion. First, we contribute an improved version of the massively multilingual and multimodal SeamlessM4T model-SeamlessM4T v2. This newer model, incorporating an updated UnitY2 framework, was trained on more low-resource language data. SeamlessM4T v2 provides the foundation on which our next two models are initiated. SeamlessExpressive enables translation that preserves vocal styles and prosody. Compared to previous efforts in expressive speech research, our work addresses certain underexplored aspects of prosody, such as speech rate and pauses, while also preserving the style of one's voice. As for SeamlessStreaming, our model leverages the Efficient Monotonic Multihead Attention mechanism to generate low-latency target translations without waiting for complete source utterances. As the first of its kind, SeamlessStreaming enables simultaneous speech-to-speech/text translation for multiple source and target languages. To ensure that our models can be used safely and responsibly, we implemented the first known red-teaming effort for multimodal machine translation, a system for the detection and mitigation of added toxicity, a systematic evaluation of gender bias, and an inaudible localized watermarking mechanism designed to dampen the impact of deepfakes. Consequently, we bring major components from SeamlessExpressive and SeamlessStreaming together to form Seamless, the first publicly available system that unlocks expressive cross-lingual communication in real-time. The contributions to this work are publicly released and accessible at https://github.com/facebookresearch/seamless_communication
更多
查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要