Search results for "MOE"
2026-03-26
01:51

Meituan Open-Source LongCat-Next: 3B Parameters for Unified Visual Understanding, Generation, and Speech

Meituan's Longcat team open-sourced LongCat-Next, a multimodal model based on MoE architecture that integrates five capabilities: text, visual understanding, image generation, and speech. Its core design, DiNA, achieves unified task processing through discrete tokens, while the dNaViT architecture in the visual domain enables excellent image generation performance. Compared to similar models, LongCat-Next demonstrates leading benchmark performance across various metrics, showcasing its advantages in multimodal understanding and generation.
More
06:27

Cursor Releases Composer 2 Technical Report, Foundation Model Score Improves by 70%

Cursor released the Composer 2 technical report on March 25, revealing the training scheme of the Kimi K2.5 model, which adopts a MoE architecture with parameters reaching 10.4 billion. Training is divided into two stages, using real-world scenario simulation for reinforcement learning. The model ultimately achieved 61.3 points on the CursorBench benchmark, a 70% improvement, with inference costs lower than other large language model APIs.
More
02:27

Meituan open-sources a 560B parameter theorem-proving model, achieving a 97.1% success rate over 72 inferences, setting a new open-source SOTA.

Meituan's LongCat team open-sourced LongCat-Flash-Prover on March 21st, a 560 billion parameter MoE model focused on Lean4 formal theorem proving. The model comprises three capabilities: automatic formalization, sketch generation, and complete proof generation, combined with reasoning tools and the Lean4 compiler to achieve real-time verification. Training utilizes the Hybrid-Experts Iteration Framework and HisPO algorithm to prevent reward hacking. Benchmark results show that the model sets new records for open-source weight models in automatic formalization and theorem proving.
More