LongCat Models
Comprehensive overview of all LongCat AI model variants and their capabilities.
Model Variants
LongCat-Flash-Chat
Released: September 1, 2025
Foundation dialogue model with 560B parameters in a Mixture-of-Experts (MoE) architecture. Activates approximately 18.6B–31.3B parameters per token (averaging ~27B) through Zero-Computation Experts.
- Supports up to 128K context length
- Achieves 100+ tokens/s on H800 GPUs
- Strong instruction following, reasoning, and coding
LongCat-Flash-Thinking
Released: September 22, 2025
Enhanced reasoning model focusing on "Agentic Reasoning" and "Formal Reasoning". Features a dual-path reasoning framework and DORA asynchronous training system.
- 64.5% token savings in tool-call scenarios
- Improved tool-call efficiency
- Formal and agentic reasoning capabilities
LongCat-Video
Released: October 27, 2025
Video generation model based on Diffusion Transformer (DiT) architecture. Unified support for text-to-video, image-to-video, and video continuation tasks.
- Generates 5-minute coherent videos at 720p/30fps
- Long temporal sequences and cross-frame consistency
- Physical motion plausibility
LongCat-Flash-Omni
Released: November 2025
First open-source real-time all-modality interaction model. Unifies text, image, audio, and video with a single end-to-end ScMoE backbone.
- Open-source SOTA on Omni-Bench and WorldSense
- Low-latency, streaming multi-modal IO
- 128K context with multi-turn dialogue
LongCat-Audio-Codec
Audio processing module providing low-bitrate, real-time streaming audio tokenization and detokenization for speech LLMs, enabling efficient audio encoding and decoding.
Model Comparison
| Model | Parameters | Key Feature | Use Case |
|---|---|---|---|
| Flash-Chat | 560B (MoE) | High-throughput dialogue | General conversation, coding |
| Flash-Thinking | 560B (MoE) | Enhanced reasoning | Tool use, formal reasoning |
| Video | DiT-based | Video generation | Text/image-to-video, continuation |
| Flash-Omni | ScMoE | All-modality | Multi-modal interaction |
Get Started
Choose a model to explore detailed documentation, benchmarks, and deployment guides: