Huck Yang

Sr. Research Scientist, NVIDIA Research

Ph.D., Georgia Institute of Technology

About

I am interested text-space alignment of speech, voice-agents, and multimodality. I worked at NVIDIA Research and Amazon ASR-LM, working with Andreas Stolcke, and as a research intern at Google (now DeepMind Audio), co-hosted by Bo Li and Yu Zhang in Tara N. Sainath's team.

🎓 My Ph.D. topic is on noise-robust voice model adaptation (now post-training), advised by Prof. Chin-Hui Lee.

🧬 I visited Prof. Jesper Tegnér's group on self-evolutionary algorithms and interned at TSMC in mixed-signal IC design before starting my PhD.





⚛️ Fun fact: I also work on Quantum ML part-time for fun, where I created the first variational circuit based speech [ICASSP 21] and language understanding [ICASSP 22] and received the Xanadu AI Quantum ML Award in 2019; recently, on quantum parameter adaptation for LLMs in [ICLR 25].

Recent Tutorials

EMNLP 2025

Spoken Conversational Agents with Large Language Models

voice agent and fun at test-time scaling.

Interspeech 2025

Efficient Adaptation in Speech Language Modeling

post-training for speech models.

Interspeech 2023

Cross-Modal Alignment for Voice Foundational Models

multimodal speech and speech-lms.