China's Navy Deploys AI to Eliminate Air Defense Blind Spots on New FrigateDeepSeek V4 to Launch on Huawei Chips With One Trillion ParametersFoxconn Posts Record Q1 Revenue as AI Server Demand Surges 30 PercentAsia's AI Boom Faces Its First Real Stress Test as Iran War Disrupts Energy and ChipsThe Physical AI Era Is Here: Why Robots Are Moving From Simulation to Factory FloorsAI Captured 80 Percent of Global Venture Funding in Q1 2026 — What That Means for Everything ElseAI Virtual Try-On Startups Take On Retail's Multibillion-Dollar Returns ProblemEclipse Raises $1.3 Billion to Build the 'Physical AI' EconomyChina's Navy Deploys AI to Eliminate Air Defense Blind Spots on New FrigateDeepSeek V4 to Launch on Huawei Chips With One Trillion ParametersFoxconn Posts Record Q1 Revenue as AI Server Demand Surges 30 PercentAsia's AI Boom Faces Its First Real Stress Test as Iran War Disrupts Energy and ChipsThe Physical AI Era Is Here: Why Robots Are Moving From Simulation to Factory FloorsAI Captured 80 Percent of Global Venture Funding in Q1 2026 — What That Means for Everything ElseAI Virtual Try-On Startups Take On Retail's Multibillion-Dollar Returns ProblemEclipse Raises $1.3 Billion to Build the 'Physical AI' EconomyChina's Navy Deploys AI to Eliminate Air Defense Blind Spots on New FrigateDeepSeek V4 to Launch on Huawei Chips With One Trillion ParametersFoxconn Posts Record Q1 Revenue as AI Server Demand Surges 30 PercentAsia's AI Boom Faces Its First Real Stress Test as Iran War Disrupts Energy and ChipsThe Physical AI Era Is Here: Why Robots Are Moving From Simulation to Factory FloorsAI Captured 80 Percent of Global Venture Funding in Q1 2026 — What That Means for Everything ElseAI Virtual Try-On Startups Take On Retail's Multibillion-Dollar Returns ProblemEclipse Raises $1.3 Billion to Build the 'Physical AI' Economy
Alibaba Qwen 3.5-Omni model capabilities diagram
Alibaba Cloud
Brief

Alibaba's Qwen 3.5-Omni Pushes Open-Source Multimodal AI to New Heights

Alibaba releases Qwen 3.5-Omni, a native omnimodal model processing 10+ hours of audio and 400 seconds of video, advancing the open-source multimodal frontier.

R
Rina ChandraTech Reporter
2 min read

Alibaba has released Qwen 3.5-Omni, a native omnimodal large language model that pushes the boundaries of what open-source multimodal AI can do. The model can process over ten hours of audio and 400 seconds of 720p video natively — capabilities that were previously limited to closed-source frontier models.

The "omnimodal" designation means Qwen 3.5-Omni processes text, images, audio, and video through a unified architecture rather than separate modules stitched together. This approach generally produces more coherent cross-modal reasoning — understanding a video's visual content in relation to its audio track, for example.

The release intensifies competition in the open-source multimodal space, where Google's Gemma 4 (with native vision and audio) and Meta's Llama models are also vying for developer adoption. Alibaba's approach of offering maximum multimodal capability at zero licensing cost is designed to drive adoption across Asia's developer ecosystem, particularly in China and Southeast Asia where Alibaba Cloud has significant market presence.

Qwen 3.5-Omni arrives as Alibaba continues to expand its enterprise AI offerings, including the Qwen-based agentic AI platform for business customers announced earlier this year.

Newsletter

Get Lanceum in your inbox

Weekly insights on AI and technology in Asia.

Share

More in Brief

Lanceum

Independent coverage of AI and technology across Asia. We go beyond headlines to explain what matters.

Colophon

Typeset in Space Grotesk & DM Serif Display. Built with Nuxt & Tailwind. Powered by curiosity.

© 2026 Lanceum. All rights reserved.

Independent • Rigorous • Asia-Focused