About Guan-Ting (Daniel) Lin
Guan-Ting is currently a final-year Ph.D. student (expected to graduate in December 2025) at the Speech Processing and Machine Learning Lab, National Taiwan University (NTU), under the guidance of Prof. Hung-yi Lee. His research interests include Speech LLMs, Full-Duplex Interaction, Spoken Language Understanding / Generation, and Test-Time Adaptation for Automatic Speech Recognition.
Guan-Ting has authored more than 10 papers at top-tier Speech and NLP conferences (ACL, EMNLP, ICASSP, INTERSPEECH, IEEE SLT) as a first or co-first author. Notably, he received the Best Paper Award at IEEE SLT 2022 in Doha, Qatar. He also regularly serves as an official reviewer for prestigious conferences, including ICLR, NeurIPS, ACL, EMNLP, NAACL, and ICASSP.
He has been fortunate to gain extensive research experience through valuable opportunities:
Google DeepMind (2025 Spring): Student Researcher at Gemini Speech team (New York City), collaborating with Kartik Audhkhasi, Soheil Khorram, and Bhuvana Ramabhadran.
Amazon AGI (2024 Summer): Applied Scientist Intern at Speech team in Seattle, USA (under Ivan Bulyko’s team), working with Prashanth Gurunath Shivakumar, Yile Gu, and Ankur Gandhe on Align-SLM, the first end-to-end spoken language model with reinforcement learning.
Amazon Alexa AI (2023 Summer): Applied Scientist Intern at Speech Recognition and LM team in Seattle, USA (under Ivan Bulyko’s team), working with Prashanth Gurunath Shivakumar and Andreas Stolcke on a paralinguistics-enhanced LLM.
Amazon Alexa AI (2022 Summer): Applied Scientist Intern at in Cambridge, USA (under Chao Wang’s team), working with Chieh-Chi Kao and Qingming Tang on acoustic event classification using neural architecture search.
For more details, please see the [CV].
Beyond academia, he enjoys singing 🎤, photography 📷, and watching MLB games ⚾️.
(2021-Present)
(2022/2023 Summer)
(2024 Summer)
(2025 Spring)
(2025 Fall)
Update 🚨
- (05/2025) Align-SLM is accepted by ACL 2025. See you in Vienna!
- (03/2025) We release the Full-Duplex-Bench, which is the first benchmark that evaluates the full-duplex spoken dialogue models on turn-taking interaction.
- (11/2024) The preprint of Align-SLM is released, which is the first RLAIF framework for end-to-end textless spoken language models with state-of-the-art performance on SLM benchmarks!
- (09/2024) Continual TTA and Emphasized-Talk are accepted by EMNLP 2024 (one as main and one as findings).
- (05/2024) Advancing Large Language Models to Capture Varied Speaking Styles and Respond Properly in Spoken Conversations is accepted by ACL 2024 as the main conference paper!
- (01/2024) Received IEEE Signal Processing Society Travel Grant for participating ICASSP 2024!
- (12/2023) Three papers are accepted by ICASSP 2024 (one first-author and two co-author). See you in Seoul!
- (02/2023) My internship work with Amazon Alexa is accepted by ICASSP 2023!
- (01/2023) Our paper “On the Utility of Self-supervised Models for Prosody-related Tasks”, collaborating with Prof. Nigel Ward of UTEP, won the Best Paper Award of IEEE SLT 2022!
- (07/2022) Received ISCA Travel Grants for Interspeech 2022.
- (06/2022) Two first-author papers are accepted at Interspeech 2022.
Education
- Ph.D. in Communication Engineering, EECS, National Taiwan University [2021/9 - 2025/12]
- Advisor: Prof. Hung-yi Lee
- Transferred from M.S. program in Feb. 2023.
Selected Publications & Preprints
(For full publication list, please see the Google Scholar).
[Speech/Text Large Language Models]
Speech understanding and generation toward human-like spoken dialogue
- Full-Duplex-Bench: A Benchmark to Evaluate Full-duplex Spoken Dialogue Models on Turn-taking Capabilities
Guan-Ting Lin, Jiachen Lian(co-second), Tingle Li(co-second), Qirui Wang(co-second), Gopala Anumanchipalli, Alexander H. Liu, Hung-yi Lee
arXiv preprint
paper / code - Align-SLM: Textless Spoken Language Models with Reinforcement Learning from AI Feedback
Guan-Ting Lin, Prashanth Gurunath Shivakumar, Aditya Gourav, Yile Gu, Ankur Gandhe, Hung-yi Lee, Ivan Bulyko
ACL 2025
paper - Advancing Large Language Models to Capture Varied Speaking Styles and Respond Properly in Spoken Conversations
Guan-Ting Lin, Cheng-Han Chiang, Hung-yi Lee
ACL 2024
paper / data - Paralinguistics-Enhanced Large Language Modeling of Spoken Dialogue
Guan-Ting Lin, Prashanth Gurunath Shivakumar, Ankur Gandhe, Chao-Han Huck Yang, Yile Gu, Shalini Ghosh, Andreas Stolcke, Hung-yi Lee, Ivan Bulyko
ICASSP 2024
paper - Can LLMs Understand the Implication of Emphasized Sentences in Dialogue?
Guan-Ting Lin, Hung-yi Lee
EMNLP 2024 Findings
paper / data
[Self-supervised Speech Models]
Explore the utilities of self-supervised speech representations models
- On the Utility of Self-supervised Models for Prosody-related Task
Guan-Ting Lin(co-first), Chi-Luen Feng(co-first), Wei-Ping Huang, Yuan Tseng, Tzu-Han Lin, Chen-An Li, Hung-yi Lee, Nigel G. Ward
SLT 2022 (Best Paper Award)
paper / code - SUPERB: Speech processing Universal PERformance Benchmark
Shu-wen Yang, Po-Han Chi, Yung-Sung Chuang, Cheng-I Lai, Kushal Lakhotia, Yist Y. Lin, Andy T. Liu, Jiatong Shi, Xuankai Chang, Guan-Ting Lin, Tzu-Hsien Huang, Wei-Cheng Tseng, Ko-tik Lee, Da-Rong Liu, Zili Huang, Shuyan Dong, Shang-Wen Li, Shinji Watanabe, Abdelrahman Mohamed, Hung-yi Lee
Interspeech 2021
paper / code - Analyzing the Robustness of Unsupervised Speech Recognition
Guan-Ting Lin(co-first), Chan-Jan Hsu(co-first), Da-Rong Liu, Hung-Yi Lee, Yu Tsao
ICASSP 2022
paper / code
[Spoken Language Understanding and Spoken Question Answering]
End-to-end approaches to understand high-level semantic information in speech signals
- Improving Textless Spoken Language Understanding with Discrete Units as Intermediate Target
Guan-Wei Wu(co-first), Guan-Ting Lin(co-first), Shang-Wen Li, Hung-yi Lee
Interspeech 2023
paper - DUAL: Discrete Spoken Unit Adaptive Learning for Textless Spoken Question Answering
Guan-Ting Lin, Yung-Sung Chuang, Ho-Lam Chung, Shu-wen Yang, Hsuan-Jui Chen, Shuyan Dong, Shang-Wen Li, Abdelrahman Mohamed, Hung-yi Lee, Lin-shan Lee
Interspeech 2022 (Poster)
paper / code - SpeechDPR: End-to-End Spoken Passage Retrieval for Open-Domain Spoken Question Answering
Chyi-Jiunn Lin, Guan-Ting Lin, Yung-Sung Chuang, Wei-Lun Wu, Shang-Wen Li, Abdelrahman Mohamed, Hung-yi Lee, Lin-shan Lee
ICASSP 2024
paper
[End-to-end ASR Test-time Adaptation]
Sample-dependent test-time adaptation to improve ASR on out-of-domain speech
- Continual Test-time Adaptation for End-to-end Speech Recognition on Noisy Speech
Guan-Ting Lin(co-first), Wei-Ping Huang(co-first), Hung-yi Lee
EMNLP 2024
paper - Listen, Adapt, Better WER: Source-free Single-utterance Test-time Adaptation for Automatic Speech Recognition
Guan-Ting Lin, Shang-Wen Li, Hung-Yi Lee
Interspeech 2022 (Oral)
paper / code
Award
- IEEE Signal Processing Society Travel Grant @ ICASSP 2024
- Best paper award @ IEEE SLT 2022
- NTU Elite Doctoral Scholarship
- GICE Elite Doctoral Scholarship with NVIDIA
- ISCA travel grant @ Interspeech 2022
- Appier top-tier conference scholarship
- Dean’s list * 3 @ NTHU
- Phi Tau Phi Award @ NTHU
- The Zhu Shun Yi He Qin Scholarship @ NTHU
Academic Services
- Official Reviewer: ICLR’24’25, NeurIPS’24’25, ACL’24’25, EMNLP’24, NAACL’23’24, ICASSP’23’24, ISCSLP’22’23’24, COLING’25