Ph.D., Georgia Institute of Technology
I focus on π£οΈ speech-language alignment and scaling laws. Prior to joining NVIDIA, I worked full-time at Amazon AGI, working with Andreas Stolcke in Ivan Bulyko's team, and as a Research Scientist intern at Google Speech & Brain teams (now DeepMind), co-hosted by Bo Li and Yu Zhang in Tara N. Sainath's team.
π My Ph.D. topic is on noise-robust voice model adaptation (now post-training), advised by Prof. Chin-Hui Lee.
𧬠I visited Prof. Jesper Tegnér's group on self-evolutionary algorithms and interned at TSMC in mixed-signal IC design before starting my PhD.
Exploring semantic and non-semantic alignment for LLMs.
Developing sample-efficient and cross-modal inference.
Building robust evaluation frameworks and intervention-resilient architectures.
A comprehensive tutorial on integrating LLMs with speech recognition systems, covering task-activating prompting and cross-modal alignment techniques.
Introduction to parameter-efficient adaptation methods for speech models, including prompt-tuning and in-context learning approaches.