Jing-Tong (Roger) Tzeng

prof_pic.jpg

I am a Master’s student at National Tsing Hua University (College of Semiconductor Research), advised by Prof. Chi-Chun (Jeremy) Lee. During my master’s studies, I also visited Prof. Carlos Busso’s Multimodal Speech Processing (MSP) Lab at The University of Texas at Dallas.

My work focuses on emotion recognition and biosignal analysis, with an emphasis on reducing aliasing introduced by front-end deep-learning preprocessing to improve downstream performance. In parallel, I develop human-in-the-loop generative components that enable targeted intervention in model outputs, increasing transparency and user trust. Broadly, I’m interested in multi-task learning and multimodal interaction.

selected publications

  1. arXiv
    SparseMOE.png
    Joint Learning using Mixture-of-Expert-Based Representation for Enhanced Speech Generation and Robust Emotion Recognition
    Jing-Tong Tzeng, Carlos Busso, and Chi-Chun Lee
    arXiv preprint arXiv:2509.08470, 2025
  2. Interspeech
    Lesson.png
    Lessons Learnt: Revisit Key Training Strategies for Effective Speech Emotion Recognition in the Wild
    Jing-Tong Tzeng, Bo-Hao Su, Ya-Tse Wu, and 2 more authors
    In Proceedings of Interspeech, 2025
  3. ICASSP
    NoiseSER.png
    Noise-Robust Speech Emotion Recognition Using Shared Self-Supervised Representations with Integrated Speech Enhancement
    Jing-Tong Tzeng, Seong-Gyun Leem, Ali N Salman, and 2 more authors
    In IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2025
  4. JMIR AI
    RespEnh.png
    Improving the Robustness and Clinical Applicability of Automatic Respiratory Sound Classification Using Deep Learning–Based Audio Enhancement: Algorithm Development and Validation
    Jing-Tong Tzeng, Jeng-Lin Li, Huan-Yu Chen, and 5 more authors
    Journal of Medical Internet Research Artificial Intelligence (JMIR AI), 2025