๐Ÿ‘จโ€๐ŸŽ“ Biography

Dr. Zhe Li is a Postdoctoral Fellow at The University of Hong Kong (HKU). His research focuses on speech large language models (Speech LLMs) and robust speaker representation learning, with broader interests in multimodal AI for healthcare applications. He received his Ph.D. degree from the Department of Electrical and Electronic Engineering at The Hong Kong Polytechnic University (PolyU). He was a research intern at Microsoft Research Asia (MSRA) and previously conducted international collaborative research as a visiting student scholar with the Department of Electrical Engineering, Stanford University. As a key contributor, he received the 2020 Excellent Science and Technology Achievement Award from the Chinese Association for Artificial Intelligence, and his co-authored paper received the Best Student Paper Runner-Up Award at PRICAI 2024.

โ€œYou are more than what you have become!โ€


๐Ÿ“ฐ News

๐Ÿ† 2026

  • Apr. 2026 ๐ŸŽ‰ Our paper โ€œDB-SMGA: Dual-Branch Sequential Multi-Granularity Attention for Speech Depression Detectionโ€ has been accepted for publication in IEEE Signal Processing Letters (SPL). Congratulations to Dr. Meirong Song for her excellent work!
  • Apr. 2026 ๐ŸŽ‰ Our paper โ€œUncertainty-Aware Multi-Head Multi-Mode Knowledge Distillation for Self-Supervised Speaker Verificationโ€ has been accepted by IEEE Transactions on Audio, Speech, and Language Processing (T-ASLP)! Thanks to Dr. Jin!
  • Apr. 2026 ๐ŸŽ‰ Our tutorial Speech Large Language Models for Under-Resourced Languages has been accepted by InterSpeech 2026 โ€” see you in September 27โ€“October 1, Sydney, Australia ๐Ÿ‡ฆ๐Ÿ‡บ!
  • Mar. 2026 ๐ŸŽ‰ Our paper โ€œTowards A Unified Perspective on Parameter-Efficient Fine Tuning for Speaker Verificationโ€ has been accepted by IEEE Transactions on Audio, Speech, and Language Processing (T-ASLP)! Thanks to Prof. Mak!
  • Jan. 2026 ๐ŸŽ‰ Two papers accepted to ICASSP 2026 โ€” see you on May 4โ€“8, 2026, in Barcelona, Spain! ๐Ÿ‡ช๐Ÿ‡ธ

๐Ÿ† 2025

  • Dec. 2025 ๐ŸŽ‰ My First Tutorial! Our tutorial Speech Large Language Models: Architectures, Efficient Adaptation, and Applications has been accepted by IEEE ICME 2026 โ€” see you in Bangkok, Thailand ๐Ÿ‡น๐Ÿ‡ญ, July 5โ€“9, 2026!
  • Sep. 29, 2025 ๐ŸŽ‰ Our paper โ€œWhisMultiNet: Advancing End-to-End Speech Topic Classification with Whisper and MultiGateGNNโ€ has been accepted by IEEE Transactions on Audio, Speech, and Language Processing (T-ASLP)! Thanks to Xiaozhe Qi!
  • Sep. 4, 2025 ๐ŸŽ‰ Our paper โ€œDisentangling Speech Representations Learning with Latent Diffusion for Speaker Verificationโ€ has been accepted by IEEE Transactions on Audio, Speech, and Language Processing (T-ASLP)! Thanks to Prof. Mak!
  • Aug. 20, 2025 ๐ŸŽ‰ One paper accepted to EMNLP 2025 โ€” see you in Suzhou, China ๐Ÿ‡จ๐Ÿ‡ณ!
  • Jun. 18, 2025 ๐ŸŽ‰ One paper accepted to MICCAI 2025 โ€” see you in Daejeon, South Korea ๐Ÿ‡ฐ๐Ÿ‡ท!
  • Jun. 14, 2025 ๐ŸŽ‰ Our paper โ€œMutual Information-Enhanced Contrastive Learning with Margin for Maximal Speaker Separabilityโ€ has been accepted by IEEE Transactions on Audio, Speech, and Language Processing (T-ASLP). Thanks to Prof. Mak!
  • May 19, 2025 ๐ŸŽ‰ Two papers accepted to Interspeech 2025 โ€” see you in Rotterdam, Netherland ๐Ÿ‡ณ๐Ÿ‡ฑ!
  • Mar. 4, 2025 ๐Ÿง‘๐Ÿปโ€๐Ÿซ Paper Sharing Session: I gave a talk on Spectral-Aware Low-Rank Adaptation for Speaker Verification (ICASSP 2025).
  • Feb. 11, 2025 ๐Ÿง‘๐Ÿปโ€๐Ÿ’ป Joined Microsoft Research Asia (MSRA) as a Research Intern, focusing on multimodal large models for healthcare.

๐Ÿ† 2024

  • Dec. 21, 2024 ๐ŸŽ‰ Four papers accepted to ICASSP 2025 โ€” see you in Hyderabad, India ๐Ÿ‡ฎ๐Ÿ‡ณ!
  • Dec. 4, 2024 ๐Ÿ… Enhancing Multimodal Rumor Detection with Statistical Image Features and Modal Alignment via Contrastive Learning received the Best Student Paper Runner-Up Award ๐Ÿฅˆ at PRICAI 2024.
  • Jun. 17, 2024 ๐Ÿง‘๐Ÿปโ€๐Ÿซ Paper Sharing Session: Parameter-efficient Fine-tuning of Speaker-Aware Dynamic Prompts for Speaker Verification (Interspeech 2024).
  • Apr. 3, 2024 ๐Ÿง‘๐Ÿปโ€๐Ÿซ Paper Sharing Session: Dual Parameter-Efficient Fine-Tuning for Speaker Representation via Speaker Prompt Tuning and Adapters (ICASSP 2024).

๐ŸŽค 2023

  • Dec. 8, 2023 Presented Maximal Speaker Separability via Robust Speaker Representation Learning at NCMMSC 2023, Soochow, China ๐Ÿ‡จ๐Ÿ‡ณ.
  • Dec. 3, 2023 Presented Maximal Speaker Separability via Contrastive Learning with Angular Margin and Class-Aware Attention for Hard Samples at International Doctoral Forum 2023, Hong Kong SAR ๐Ÿ‡ญ๐Ÿ‡ฐ.

๐Ÿ“š 2022โ€“2020

  • May 15, 2023 Paper Sharing Session: Discriminative Speaker Representation via Contrastive Learning with Class-Aware Attention in Angular Space (ICASSP 2023).
  • Jul. 1, 2022 Participant Talk: Shared on speaker verification at Odyssey-CNSRC Workshop 2022.
  • May 29, 2021 ๐ŸŽ“ Completed Masterโ€™s oral examination.
  • Nov. 14, 2020 ๐Ÿ… CAAI Award: Received the Excellent Scientific and Technological Achievements Award of the Chinese Association for Artificial Intelligence.
  • Oct. 29, 2020 Video: Uploaded CCL 2020 oral presentation.
  • Oct. 11, 2020 Video: Uploaded CCMT 2020 oral presentation.

๐Ÿ”ฌ Research Interests

  • ๐Ÿง  Speech Large Language Models (Speech LLMs) โ€“ efficient fine-tuning, post-training alignment, and speech-based healthcare applications
  • ๐Ÿ—ฃ๏ธ Speech Signal Processing โ€“ speaker representation learning, accent recognition, and robust speech modeling
  • ๐Ÿฉบ Multimodal and Deep Learning โ€“ multimodal representation learning, cross-modal fusion

๐Ÿ’ผ Research Experience


Hidden Visit Tracker