Guan-Ting (Daniel) Lin

Taipei, Taiwan

Guan-Ting received his Ph.D. from the Speech Processing and Machine Learning Lab at National Taiwan University (NTU), advised by Prof. Hung-yi Lee. His research focuses on Speech LLMs, Full-Duplex Interaction, Spoken Language Understanding/Generation, and Test-Time Adaptation for ASR.

He has published 15+ first/co-first author papers at top-tier conferences (ACL, EMNLP, ICASSP, Interspeech, ASRU, SLT) and received the Best Paper Award at IEEE SLT 2022. He serves as a reviewer for ICLR, NeurIPS, ACL, EMNLP, and is recognized as ICLR 2025 Notable Reviewer.

Research Experience:

Guan-Ting (Daniel) Lin
Dumbo, USA (Feb. 2025)
NTU
Alexa AI
Amazon AGI
Google DeepMind
Meta

Open to discussing or collaborating on speech research—feel free to reach out at daniel094144[at]gmail[dot]com. Beyond academia, he enjoys singing 🎤, photography 📷, and watching MLB games ⚾️.

Update

  • 2026/01 Started writing blog posts! Shared learnings on my Ph.D. journey in the post.
  • 2026/01 Full-Duplex-Bench v1.5 accepted by ICASSP 2026 🇪🇸!
  • 2025/12 Successfully defended my Ph.D. dissertation! 🎓
  • 2025/08 Three papers accepted by ASRU 2025. See you in Hawaii 🏝
  • 2025/05 Align-SLM accepted by ACL 2025. See you in Vienna!
  • 2025/03 Released Full-Duplex-Bench — the first benchmark for full-duplex spoken dialogue models.
  • 2024/09 Continual TTA & Emphasized-Talk accepted by EMNLP 2024 (main & findings).
  • 2024/05 Spoken-LLM and StyleTalk accepted by ACL 2024.
  • 2024/01 Received IEEE SPS Travel Grant for ICASSP 2024!
  • 2023/12 Three papers accepted by ICASSP 2024. See you in Seoul!
  • 2023/02 Internship work with Amazon Alexa accepted by ICASSP 2023.
  • 2023/01 Paper with Prof. Nigel Ward won Best Paper Award at IEEE SLT 2022!
  • 2022/07 Received ISCA Travel Grant for Interspeech 2022.
  • 2022/06 Two first-author papers accepted at Interspeech 2022.

Selected Publications & Preprints

(For full publication list, please see the Google Scholar).

Speech/Text Large Language Models

Speech understanding and generation toward human-like spoken dialogue

  • Full-Duplex-Bench: A Benchmark to Evaluate Full-duplex Spoken Dialogue Models on Turn-taking Capabilities
    Guan-Ting Lin, Jiachen Lian(co-second), Tingle Li(co-second), Qirui Wang(co-second), Gopala Anumanchipalli, Alexander H. Liu, Hung-yi Lee
    ASRU 2025
    PAPER CODE
  • Align-SLM: Textless Spoken Language Models with Reinforcement Learning from AI Feedback
    Guan-Ting Lin, Prashanth Gurunath Shivakumar, Aditya Gourav, Yile Gu, Ankur Gandhe, Hung-yi Lee, Ivan Bulyko
    ACL 2025
    PAPER
  • Advancing Large Language Models to Capture Varied Speaking Styles and Respond Properly in Spoken Conversations
    Guan-Ting Lin, Cheng-Han Chiang, Hung-yi Lee
    ACL 2024
    PAPER DATA

  • Full-Duplex-Bench-v2: A Multi-Turn Evaluation Framework for Duplex Dialogue Systems with an Automated Examiner
    Guan-Ting Lin(co-first), Shih-Yun Shan Kuan(co-first), Jiatong Shi, Kai-Wei Chang, Siddhant Arora, Shinji Watanabe, Hung-yi Lee
    Arxiv 2025
    PAPER CODE

  • Full-Duplex-Bench v1.5: Evaluating Overlap Handling for Full-Duplex Speech Models
    Guan-Ting Lin, Shih-Yun Shan Kuan, Qirui Wang, Jiachen Lian, Tingle Li, Shinji Watanabe,Hung-yi Lee
    ICASSP 2026
    PAPER CODE

  • Paralinguistics-Enhanced Large Language Modeling of Spoken Dialogue
    Guan-Ting Lin, Prashanth Gurunath Shivakumar, Ankur Gandhe, Chao-Han Huck Yang, Yile Gu, Shalini Ghosh, Andreas Stolcke, Hung-yi Lee, Ivan Bulyko
    ICASSP 2024
    PAPER

  • On the Utility of Self-supervised Models for Prosody-related Task
    Guan-Ting Lin(co-first), Chi-Luen Feng(co-first), Wei-Ping Huang, Yuan Tseng, Tzu-Han Lin, Chen-An Li, Hung-yi Lee, Nigel G. Ward
    SLT 2022 (Best Paper Award)
    PAPER CODE

  • Can LLMs Understand the Implication of Emphasized Sentences in Dialogue?
    Guan-Ting Lin, Hung-yi Lee
    EMNLP 2024 Findings
    PAPER DATA

Spoken Language Understanding and Spoken Question Answering

End-to-end approaches to understand high-level semantic information in speech signals

  • DUAL: Discrete Spoken Unit Adaptive Learning for Textless Spoken Question Answering
    Guan-Ting Lin, Yung-Sung Chuang, Ho-Lam Chung, Shu-wen Yang, Hsuan-Jui Chen, Shuyan Dong, Shang-Wen Li, Abdelrahman Mohamed, Hung-yi Lee, Lin-shan Lee
    Interspeech 2022
    PAPER CODE
  • Improving Textless Spoken Language Understanding with Discrete Units as Intermediate Target
    Guan-Wei Wu(co-first), Guan-Ting Lin(co-first), Shang-Wen Li, Hung-yi Lee
    Interspeech 2023
    PAPER

End-to-end ASR Test-time Adaptation

Sample-dependent test-time adaptation to improve ASR on out-of-domain speech

  • Listen, Adapt, Better WER: Source-free Single-utterance Test-time Adaptation for Automatic Speech Recognition
    Guan-Ting Lin, Shang-Wen Li, Hung-Yi Lee
    Interspeech 2022
    PAPER CODE

  • Continual Test-time Adaptation for End-to-end Speech Recognition on Noisy Speech
    Guan-Ting Lin(co-first), Wei-Ping Huang(co-first), Hung-yi Lee
    EMNLP 2024
    PAPER CODE

  • SUTA-LM: Bridging Test-Time Adaptation and Language Model Rescoring for Robust ASR
    Wei-Ping Huang(co-first), Guan-Ting Lin(co-first), Hung-yi Lee
    ASRU 2025
    PAPER CODE

Patents

  • Inventor on a pending U.S. patent application in speech and language processing, filed by Google DeepMind (details confidential until publication)

Award

  • IEEE Signal Processing Society Travel Grant @ ICASSP 2024
  • Best paper award @ IEEE SLT 2022
  • NTU Elite Doctoral Scholarship
  • GICE Elite Doctoral Scholarship with NVIDIA
  • ISCA travel grant @ Interspeech 2022
  • Appier top-tier conference scholarship
  • Dean’s list * 3 @ NTHU
  • Phi Tau Phi Award @ NTHU
  • The Zhu Shun Yi He Qin Scholarship @ NTHU

Academic Services

  • Official Reviewer: ICLR’24’25, NeurIPS’24’25, ACL’24’25, EMNLP’24, NAACL’23’24, ICASSP’23’24, ISCSLP’22’23’24, COLING’25