About Guan-Ting (Daniel) Lin

Guan-Ting is currently a final-year Ph.D. student (expected to graduate in December 2025, looking for full-time research scientist/engineer roles starting from 2026) at the Speech Processing and Machine Learning Lab, National Taiwan University (NTU), under the guidance of Prof. Hung-yi Lee. His research interests include Speech LLMs, Full-Duplex Interaction, Spoken Language Understanding / Generation, and Test-Time Adaptation for Automatic Speech Recognition.
Guan-Ting has published more than 10 first/co-first author top-tier Speech & NLP conferences papers (ACL, EMNLP, ICASSP, INTERSPEECH, ASRU, SLT). Notably, he received the Best Paper Award at IEEE SLT 2022 in Doha, Qatar. He also regularly serves as an official reviewer for prestigious conferences, including ICLR, NeurIPS, ACL, EMNLP, NAACL, and ICASSP. He is recognized as ICLR 2025 Notable Reviewer.
He has been fortunate to gain extensive research experience through valuable opportunities:
Google DeepMind (2025 Spring): Student Researcher at Gemini Speech team (New York City), collaborating with Kartik Audhkhasi, Soheil Khorram, and Bhuvana Ramabhadran to develop methods enhancing Gemini speech capabilities in low-resource languages.
Amazon AGI (2024 Summer): Applied Scientist Intern at Speech team in Seattle, USA (under Ivan Bulyko’s team), working with Prashanth Gurunath Shivakumar, Yile Gu, and Ankur Gandhe on Align-SLM, the first end-to-end spoken language model with reinforcement learning.
Amazon Alexa AI (2023 Summer): Applied Scientist Intern at Speech Recognition and LM team in Seattle, USA (under Ivan Bulyko’s team), working with Prashanth Gurunath Shivakumar and Andreas Stolcke on a paralinguistics-enhanced LLM.
Amazon Alexa AI (2022 Summer): Applied Scientist Intern at in Cambridge, USA (under Chao Wang’s team), working with Chieh-Chi Kao and Qingming Tang on acoustic event classification using neural architecture search.
🔎 For more details, please see the [CV] and the Google Scholar for the most updated profile.
Open to discussing or collaborating on speech research—feel free to reach out at daniel094144[at]gmail[dot]com
.
Beyond academia, he enjoys singing 🎤, photography 📷, and watching MLB games ⚾️.
(2021-Present)
(2022/2023 Summer)
(2024 Summer)
(2025 Spring)
(2025 Fall)
Update 🚨
2025/08 Three papers accepted by ASRU 2025 — see you in Hawaii 🏝
2025/05 Align-SLM accepted by ACL 2025 — see you in Vienna!
2025/03 Released Full-Duplex-Bench — the first benchmark for full-duplex spoken dialogue models.
2024/11 Preprint of Align-SLM released — first RLAIF framework for end-to-end textless SLMs.
2024/09 Continual TTA & Emphasized-Talk accepted by EMNLP 2024 (main & findings).
2024/05 Advancing LLMs to Capture Speaking Styles accepted by ACL 2024.
2024/01 Received IEEE SPS Travel Grant for ICASSP 2024!
2023/12 Three papers accepted by ICASSP 2024 — see you in Seoul!
2023/02 Internship work with Amazon Alexa accepted by ICASSP 2023.
2023/01 Paper with Prof. Nigel Ward won Best Paper Award at IEEE SLT 2022!
2022/07 Received ISCA Travel Grant for Interspeech 2022.
2022/06 Two first-author papers accepted at Interspeech 2022.
Education
- Ph.D. in Communication Engineering, EECS, National Taiwan University [2021/9 - 2025/12]
- Advisor: Prof. Hung-yi Lee
- Transferred from M.S. program in Feb. 2023.
Selected Publications & Preprints
(For full publication list, please see the Google Scholar).
[Speech/Text Large Language Models]
Speech understanding and generation toward human-like spoken dialogue
- Full-Duplex-Bench v1.5: Evaluating Overlap Handling for Full-Duplex Speech Models
Guan-Ting Lin, Shih-Yun Shan Kuan, Qirui Wang, Jiachen Lian, Tingle Li, Hung-yi Lee
Arxiv 2025
paper / code - Full-Duplex-Bench: A Benchmark to Evaluate Full-duplex Spoken Dialogue Models on Turn-taking Capabilities
Guan-Ting Lin, Jiachen Lian(co-second), Tingle Li(co-second), Qirui Wang(co-second), Gopala Anumanchipalli, Alexander H. Liu, Hung-yi Lee
ASRU 2025
paper / code - Align-SLM: Textless Spoken Language Models with Reinforcement Learning from AI Feedback
Guan-Ting Lin, Prashanth Gurunath Shivakumar, Aditya Gourav, Yile Gu, Ankur Gandhe, Hung-yi Lee, Ivan Bulyko
ACL 2025
paper - Advancing Large Language Models to Capture Varied Speaking Styles and Respond Properly in Spoken Conversations
Guan-Ting Lin, Cheng-Han Chiang, Hung-yi Lee
ACL 2024
paper / data - Paralinguistics-Enhanced Large Language Modeling of Spoken Dialogue
Guan-Ting Lin, Prashanth Gurunath Shivakumar, Ankur Gandhe, Chao-Han Huck Yang, Yile Gu, Shalini Ghosh, Andreas Stolcke, Hung-yi Lee, Ivan Bulyko
ICASSP 2024
paper - Can LLMs Understand the Implication of Emphasized Sentences in Dialogue?
Guan-Ting Lin, Hung-yi Lee
EMNLP 2024 Findings
paper / data
[Self-supervised Speech Models]
Explore the utilities of self-supervised speech representations models
- On the Utility of Self-supervised Models for Prosody-related Task
Guan-Ting Lin(co-first), Chi-Luen Feng(co-first), Wei-Ping Huang, Yuan Tseng, Tzu-Han Lin, Chen-An Li, Hung-yi Lee, Nigel G. Ward
SLT 2022 (Best Paper Award)
paper / code - Analyzing the Robustness of Unsupervised Speech Recognition
Guan-Ting Lin(co-first), Chan-Jan Hsu(co-first), Da-Rong Liu, Hung-Yi Lee, Yu Tsao
ICASSP 2022
paper / code
[Spoken Language Understanding and Spoken Question Answering]
End-to-end approaches to understand high-level semantic information in speech signals
- DUAL: Discrete Spoken Unit Adaptive Learning for Textless Spoken Question Answering
Guan-Ting Lin, Yung-Sung Chuang, Ho-Lam Chung, Shu-wen Yang, Hsuan-Jui Chen, Shuyan Dong, Shang-Wen Li, Abdelrahman Mohamed, Hung-yi Lee, Lin-shan Lee
Interspeech 2022
paper / code - Improving Textless Spoken Language Understanding with Discrete Units as Intermediate Target
Guan-Wei Wu(co-first), Guan-Ting Lin(co-first), Shang-Wen Li, Hung-yi Lee
Interspeech 2023
paper - SpeechDPR: End-to-End Spoken Passage Retrieval for Open-Domain Spoken Question Answering
Chyi-Jiunn Lin, Guan-Ting Lin, Yung-Sung Chuang, Wei-Lun Wu, Shang-Wen Li, Abdelrahman Mohamed, Hung-yi Lee, Lin-shan Lee
ICASSP 2024
paper
[End-to-end ASR Test-time Adaptation]
Sample-dependent test-time adaptation to improve ASR on out-of-domain speech
- SUTA-LM: Bridging Test-Time Adaptation and Language Model Rescoring for Robust ASR
Wei-Ping Huang(co-first), Guan-Ting Lin(co-first), Hung-yi Lee
ASRU 2025
paper / code - Continual Test-time Adaptation for End-to-end Speech Recognition on Noisy Speech
Guan-Ting Lin(co-first), Wei-Ping Huang(co-first), Hung-yi Lee
EMNLP 2024
paper / code - Listen, Adapt, Better WER: Source-free Single-utterance Test-time Adaptation for Automatic Speech Recognition
Guan-Ting Lin, Shang-Wen Li, Hung-Yi Lee
Interspeech 2022 (Oral)
paper / code
Patents
- Inventor on a pending U.S. patent application in speech and language processing, filed by Google DeepMind (details confidential until publication)
Award
- IEEE Signal Processing Society Travel Grant @ ICASSP 2024
- Best paper award @ IEEE SLT 2022
- NTU Elite Doctoral Scholarship
- GICE Elite Doctoral Scholarship with NVIDIA
- ISCA travel grant @ Interspeech 2022
- Appier top-tier conference scholarship
- Dean’s list * 3 @ NTHU
- Phi Tau Phi Award @ NTHU
- The Zhu Shun Yi He Qin Scholarship @ NTHU
Academic Services
- Official Reviewer: ICLR’24’25, NeurIPS’24’25, ACL’24’25, EMNLP’24, NAACL’23’24, ICASSP’23’24, ISCSLP’22’23’24, COLING’25