We've just announced IQ AI.
Yinghao Aaron Li is an artificial intelligence research scientist known for his contributions to speech synthesis, voice conversion, and multimodal large language models. He is currently a research scientist at Meta Superintelligence Labs and has developed several notable text-to-speech (TTS) models, including StyleTTS, StyleTTS 2, and DMOSpeech 2. [1]
Li completed his doctoral studies at Columbia University, earning a Ph.D. from the Department of Electrical Engineering. His research at Columbia focused on generative speech modeling, including text-to-speech synthesis and voice conversion, under the supervision of Professor Nima Mesgarani. Throughout his academic career, his work has been published in various IEEE journals and presented at prominent AI and computational linguistics conferences. [2] [3] [10]
During his Ph.D. studies, Li undertook a research internship at Adobe, where his work contributed to the development of the DMOSpeech project. Upon completing his doctorate, Li announced in 2025 that he would be joining Meta Superintelligence Labs as an AI Research Scientist. His work there focuses on multimodal large language models. His projects are often made publicly available through platforms like GitHub and Hugging Face, where models such as StyleTTS 2 have gained significant traction within the open-source community.
Li's research primarily addresses challenges in generating natural, diverse, and efficient human-like speech. His work spans style-based generative models, zero-shot synthesis, metric-optimized TTS, and integrated spoken dialogue systems. [1] [2] [3] [10] [11]
StyleTTS is a style-based generative model for text-to-speech synthesis designed to produce speech with naturalistic prosodic variations and emotional tones. The model synthesizes speech directly from a style vector, which is a latent variable sampled from a reference speech signal, without requiring explicit prosody modeling. This approach allows for the generation of diverse speech outputs from the same text input. [5]
Building on this foundation, StyleTTS 2 was developed to advance toward human-level TTS quality. It integrates style diffusion and adversarial training with large speech language models. This model improves upon its predecessor by enhancing the naturalness and speaker similarity of the synthesized speech. The project gained considerable attention in the open-source community, accumulating over 5,900 stars on GitHub and forming the basis for other popular TTS projects. [6] [1]
Presented at the 2025 Conference of the North American Chapter of the Association for Computational Linguistics (NAACL), StyleTTS-ZS is an efficient, high-quality zero-shot TTS model. The model addresses common issues in large-scale TTS, such as slow inference speeds and reliance on complex neural codec representations. It introduces a method that uses distilled time-varying style diffusion to capture diverse speaker identities and prosodies from a short reference audio clip.
Key features of StyleTTS-ZS include:
The model was demonstrated to be 10 to 20 times faster than other state-of-the-art large-scale zero-shot TTS systems at the time of its publication. [7]
DMOSpeech 2 represents an advancement in metric-optimized speech synthesis and was Li's final project during his Ph.D. It extends the work of the original DMOSpeech by incorporating reinforcement learning (RL) to optimize the duration predictor, a component previously not optimized for perceptual metrics. The system aims to create a more complete metric-optimized pipeline for zero-shot TTS.
The core innovations of DMOSpeech 2 are:
Evaluations showed that DMOSpeech 2 achieved superior performance across all metrics compared to previous systems and could perform inference in just 4-8 steps, reducing sampling steps by half without degrading quality. [8] [1]
Style-Talker is a framework designed for fast and natural spoken dialogue generation, presented at CoLM 2024. It addresses the latency and prosody limitations of traditional cascaded systems that chain together automatic speech recognition (ASR), a large language model (LLM), and a TTS model. Style-Talker fine-tunes an audio LLM and a style-based TTS model to work in concert.
The system operates by taking user input audio and using the transcribed chat history and speech style to generate both the text and the speaking style for the response. While the response is being synthesized and played, the system processes the next turn's input audio in parallel to extract its transcription and style. This pipeline design significantly reduces latency and allows the model to incorporate paralinguistic information from the input speech into the output, resulting in more natural and coherent dialogue. Experiments showed that Style-Talker was over 50% faster than conventional cascaded and speech-to-speech baseline models. [9]
Li has also contributed to several other projects in speech processing. His work includes PL-BERT, a phoneme-level BERT model for enhancing TTS prosody; SLMGAN, which uses speech language model representations for unsupervised zero-shot voice conversion; and an examination of the Mamba architecture for speech-related tasks titled Speech Slytherin. These projects further demonstrate his focus on improving the efficiency, naturalness, and controllability of generative speech models. [2]