Atlas / Reports / Detail
Qwen2.5-Omni Technical Report
Multimodal Models
Connected researchers
Yale Song
Qwen
Yale Song is an assistant professor in artificial intelligence at Yonsei University and is also affiliated with the Stanford AI Lab while working part-time with Adobe Research.
Shilong Liu
Qwen
Researcher whose public homepage focuses on computer vision, multimodal foundation models, and embodied AI; publication context connects Shilong Liu to the Qwen2.5-Omni technical report.
Jianwei Niu
Qwen
Jianwei Niu is a tenure-track research assistant professor in the School of Data Science at Lingnan University, Hong Kong. His research focuses on multimodal learning, computer vision, and embodied AI.
Jingren Zhou
MiniMax / Moonshot AI / Qwen / Z.ai
Alibaba senior technology leader and researcher associated with Qwen. Public profiles list him with Alibaba Group, and official Alibaba Cloud coverage identifies him as a chief technology officer leading large-model work.
Qinyu Chen
DeepSeek / Qwen
Research scientist at Qwen, Alibaba, whose public OpenReview profile lists work on vision-language models and large language models.
Linjie Li
Qwen
Linjie Li is a research scientist at Alibaba Group and a contributor to the Qwen2.5-Omni Technical Report.
Xiaoyu Hu
Qwen
Research engineer at Alibaba Group working on audio and multimodal foundation models, multimodal RL, and speech processing; coauthor of Qwen2.5-Omni.
Yang Song
OpenAI / Qwen
Research scientist and GPT-4 coauthor known for work on generative modeling, diffusion methods, and machine learning systems.
Jianfei Chen
Qwen
Jianfei Chen is an assistant professor at Monash University. His research spans computer vision, machine learning, multimodality, and trustworthy AI.
Linjun Yang
Qwen
Research scientist in Tongyi Lab and technical lead of Qwen2.5-Omni, with public work on end-to-end speech understanding and generation.