Qiaolin Wang

qw2443@columbia.edu | Columbia EE | CV

prof_pic.jpg Profile picture 2

I am Qiaolin Wang, a second-year Masterโ€™s student in Electrical Engineering at Columbia University, advised by Professor Nima Mesgarani.

I am passionate about building models that can Preceive, Reason, and Speak as naturally as humans do. My current research investigates the fundamental capabilities of Large Audio Language Models (LALMs), from their internal representations about syntax and context, to their capacity for complex reasoning across modalities. My future work aims to advance this by pioneering Audio-Visual Understanding and unified models for Reasoning and Generation.

Before joining Columbia, I earned my B.Eng. in Computer Science from Wuhan University. I also had an enriching experience as a Research Intern at Wiz.AI, where I developed a SOTA Speech Emotion Recognition LLM.

News

Nov 10, 2025 ๐Ÿ† Our paper was awarded EMNLP SAC Highlight!
Nov 07, 2025 ๐ŸŽค I will be presenting our work at the SANE 2025 workshop at Google NYC!
Sep 19, 2025 ๐Ÿ“„ SightSound-R1 is now available on arXiv! In this work, we propose a framework to transfer reasoning from VLMs to LALMs.
Aug 07, 2025 ๐ŸŽ‰ Thrilled that our paper was accepted for an oral presentation at EMNLP 2025!

Selected Publications

  1. EMNLP
    layerwise_probing.jpg
    Layer-wise Minimal Pair Probing Reveals Contextual Grammatical-Conceptual Hierarchy in Speech Representations
    Linyang He*, Qiaolin Wang*, Xilin Jiang, and 1 more author
    In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, 2025
    Oral presentation
  2. arXiv
    sightsound_r1.png
    SightSound-R1: Cross-Modal Reasoning Distillation from Vision to Audio Language Models
    Qiaolin Wang, Xilin Jiang, Linyang He, and 2 more authors
    arXiv preprint arXiv:2509.15661, 2025