About Guan-Ting (Daniel) Lin

Guan-Ting is currently a final-year Ph.D. student (expected to graduate in December 2025) at the Speech Processing and Machine Learning Lab, National Taiwan University (NTU), under the guidance of Prof. Hung-yi Lee. His research interests include Speech LLMs, Full-Duplex Interaction, Spoken Language Understanding / Generation, and Test-Time Adaptation for Automatic Speech Recognition.

Guan-Ting has authored more than 10 papers at top-tier Speech and NLP conferences (ACL, EMNLP, ICASSP, INTERSPEECH, IEEE SLT) as a first or co-first author. Notably, he received the Best Paper Award at IEEE SLT 2022 in Doha, Qatar. He also regularly serves as an official reviewer for prestigious conferences, including ICLR, NeurIPS, ACL, EMNLP, NAACL, and ICASSP.

He has been fortunate to gain extensive research experience through valuable opportunities:

For more details, please see the [CV].

Beyond academia, he enjoys singing 🎤, photography 📷, and watching MLB games ⚾️.

Image 1
NTU
(2021-Present)
Image 2
Alexa AI
(2022/2023 Summer)
Image 3
Amazon AGI
(2024 Summer)
Image 4
Google DeepMind
(2025 Spring)
Image 5
Meta GenAI
(2025 Fall)

Update 🚨

  • (05/2025) Align-SLM is accepted by ACL 2025. See you in Vienna!
  • (03/2025) We release the Full-Duplex-Bench, which is the first benchmark that evaluates the full-duplex spoken dialogue models on turn-taking interaction.
  • (11/2024) The preprint of Align-SLM is released, which is the first RLAIF framework for end-to-end textless spoken language models with state-of-the-art performance on SLM benchmarks!
  • (09/2024) Continual TTA and Emphasized-Talk are accepted by EMNLP 2024 (one as main and one as findings).
  • (05/2024) Advancing Large Language Models to Capture Varied Speaking Styles and Respond Properly in Spoken Conversations is accepted by ACL 2024 as the main conference paper!
  • (01/2024) Received IEEE Signal Processing Society Travel Grant for participating ICASSP 2024!
  • (12/2023) Three papers are accepted by ICASSP 2024 (one first-author and two co-author). See you in Seoul!
  • (02/2023) My internship work with Amazon Alexa is accepted by ICASSP 2023!
  • (01/2023) Our paper “On the Utility of Self-supervised Models for Prosody-related Tasks”, collaborating with Prof. Nigel Ward of UTEP, won the Best Paper Award of IEEE SLT 2022!
  • (07/2022) Received ISCA Travel Grants for Interspeech 2022.
  • (06/2022) Two first-author papers are accepted at Interspeech 2022.

Education

  • Ph.D. in Communication Engineering, EECS, National Taiwan University [2021/9 - 2025/12]
    • Advisor: Prof. Hung-yi Lee
    • Transferred from M.S. program in Feb. 2023.

Selected Publications & Preprints

(For full publication list, please see the Google Scholar).

[Speech/Text Large Language Models]
Speech understanding and generation toward human-like spoken dialogue

  • Full-Duplex-Bench: A Benchmark to Evaluate Full-duplex Spoken Dialogue Models on Turn-taking Capabilities
    Guan-Ting Lin, Jiachen Lian(co-second), Tingle Li(co-second), Qirui Wang(co-second), Gopala Anumanchipalli, Alexander H. Liu, Hung-yi Lee
    arXiv preprint
    paper / code
  • Align-SLM: Textless Spoken Language Models with Reinforcement Learning from AI Feedback
    Guan-Ting Lin, Prashanth Gurunath Shivakumar, Aditya Gourav, Yile Gu, Ankur Gandhe, Hung-yi Lee, Ivan Bulyko
    ACL 2025
    paper
  • Advancing Large Language Models to Capture Varied Speaking Styles and Respond Properly in Spoken Conversations
    Guan-Ting Lin, Cheng-Han Chiang, Hung-yi Lee
    ACL 2024
    paper / data
  • Paralinguistics-Enhanced Large Language Modeling of Spoken Dialogue
    Guan-Ting Lin, Prashanth Gurunath Shivakumar, Ankur Gandhe, Chao-Han Huck Yang, Yile Gu, Shalini Ghosh, Andreas Stolcke, Hung-yi Lee, Ivan Bulyko
    ICASSP 2024
    paper
  • Can LLMs Understand the Implication of Emphasized Sentences in Dialogue?
    Guan-Ting Lin, Hung-yi Lee
    EMNLP 2024 Findings
    paper / data

[Self-supervised Speech Models]
Explore the utilities of self-supervised speech representations models

  • On the Utility of Self-supervised Models for Prosody-related Task
    Guan-Ting Lin(co-first), Chi-Luen Feng(co-first), Wei-Ping Huang, Yuan Tseng, Tzu-Han Lin, Chen-An Li, Hung-yi Lee, Nigel G. Ward
    SLT 2022 (Best Paper Award)
    paper / code
  • SUPERB: Speech processing Universal PERformance Benchmark
    Shu-wen Yang, Po-Han Chi, Yung-Sung Chuang, Cheng-I Lai, Kushal Lakhotia, Yist Y. Lin, Andy T. Liu, Jiatong Shi, Xuankai Chang, Guan-Ting Lin, Tzu-Hsien Huang, Wei-Cheng Tseng, Ko-tik Lee, Da-Rong Liu, Zili Huang, Shuyan Dong, Shang-Wen Li, Shinji Watanabe, Abdelrahman Mohamed, Hung-yi Lee
    Interspeech 2021
    paper / code
  • Analyzing the Robustness of Unsupervised Speech Recognition
    Guan-Ting Lin(co-first), Chan-Jan Hsu(co-first), Da-Rong Liu, Hung-Yi Lee, Yu Tsao
    ICASSP 2022
    paper / code

[Spoken Language Understanding and Spoken Question Answering]
End-to-end approaches to understand high-level semantic information in speech signals

  • Improving Textless Spoken Language Understanding with Discrete Units as Intermediate Target
    Guan-Wei Wu(co-first), Guan-Ting Lin(co-first), Shang-Wen Li, Hung-yi Lee
    Interspeech 2023
    paper
  • DUAL: Discrete Spoken Unit Adaptive Learning for Textless Spoken Question Answering
    Guan-Ting Lin, Yung-Sung Chuang, Ho-Lam Chung, Shu-wen Yang, Hsuan-Jui Chen, Shuyan Dong, Shang-Wen Li, Abdelrahman Mohamed, Hung-yi Lee, Lin-shan Lee
    Interspeech 2022 (Poster)
    paper / code
  • SpeechDPR: End-to-End Spoken Passage Retrieval for Open-Domain Spoken Question Answering
    Chyi-Jiunn Lin, Guan-Ting Lin, Yung-Sung Chuang, Wei-Lun Wu, Shang-Wen Li, Abdelrahman Mohamed, Hung-yi Lee, Lin-shan Lee
    ICASSP 2024
    paper

[End-to-end ASR Test-time Adaptation]
Sample-dependent test-time adaptation to improve ASR on out-of-domain speech

  • Continual Test-time Adaptation for End-to-end Speech Recognition on Noisy Speech
    Guan-Ting Lin(co-first), Wei-Ping Huang(co-first), Hung-yi Lee
    EMNLP 2024
    paper
  • Listen, Adapt, Better WER: Source-free Single-utterance Test-time Adaptation for Automatic Speech Recognition
    Guan-Ting Lin, Shang-Wen Li, Hung-Yi Lee
    Interspeech 2022 (Oral)
    paper / code

Award

  • IEEE Signal Processing Society Travel Grant @ ICASSP 2024
  • Best paper award @ IEEE SLT 2022
  • NTU Elite Doctoral Scholarship
  • GICE Elite Doctoral Scholarship with NVIDIA
  • ISCA travel grant @ Interspeech 2022
  • Appier top-tier conference scholarship
  • Dean’s list * 3 @ NTHU
  • Phi Tau Phi Award @ NTHU
  • The Zhu Shun Yi He Qin Scholarship @ NTHU

Academic Services

  • Official Reviewer: ICLR’24’25, NeurIPS’24’25, ACL’24’25, EMNLP’24, NAACL’23’24, ICASSP’23’24, ISCSLP’22’23’24, COLING’25