[2511.00279] LongCat-Flash-Omni Technical Report
View PDF of the article LongCat-Flash-Omni Technical Report, by Meituan LongCat Team and 132 other authors
View PDF HTML (beta)
a summary:We introduce LongCat-Flash-Omni, a state-of-the-art, open source, multimedia model with 560 billion parameters that excels at real-time audio and visual interaction. By adopting a progressive, curriculum-inspired training strategy that moves from simpler to more complex method sequence modeling tasks, LongCat-Flash-Omni achieves comprehensive multi-modal capabilities while maintaining strong unimodal capability. Based on LongCat-Flash, which adopts a high-performance shortcut-connected Mixed-of-Experts (MoE) architecture with non-computational experts, LongCat-Flash-Omni integrates efficient multimodal perception and speech reconstruction modules. Despite its massive size of 560 bytes (with 27 bytes of activation), LongCat-Flash-Omni achieves low-latency, real-time audio-visual interaction. For the training infrastructure, we developed a method-decoupled parallel scheme specifically designed to manage data and model heterogeneity inherent in large-scale multimodal training. This innovative approach demonstrates exceptional efficiency by maintaining over 90% of the throughput achieved by text-only training. Comprehensive evaluations show that LongCat-Flash-Omni achieves cutting-edge performance on multimedia benchmarks among open source models. Furthermore, it delivers highly competitive results across a wide range of modality-specific tasks, including text, image and video comprehension, as well as audio comprehension and generation. We provide a comprehensive overview of the model architecture design, training procedures, and data strategies, and open the model to promote future research and development in the community.
Submission date
From: Songxiang Liu [view email]
[v1]
Friday, October 31, 2025, 21:58:15 UTC (3,723 KB)
[v2]
Friday, 28 November 2025 09:10:07 UTC (3,707 KB)
Don’t miss more hot News like this! AI/" target="_blank" rel="noopener">Click here to discover the latest in AI news!
2025-12-01 05:00:00



