meituan
@meituan
1
Published Tools
0
Total Stars
0
Weekly Downloads
Published Tools
1 Agentacross 1 categoryLongCat-Flash-Chat is a large-scale Mixture-of-Experts (MoE) model with 560B total parameters, of which 18.6Bβ31.3B (β27B on average) are dynamically activated per input. It introduces a shortcut-connected MoE design to reduce communication overhead and achieve high throughput while maintaining training stability through advanced scaling strategies such as hyperparameter transfer, deterministic computation, and multi-stage optimization. This release, LongCat-Flash-Chat, is a non-thinking founda
AgentLLM Model
1 dir