Available Models
Deepseek R1 0528
In this latest update, DeepSeek R1 has significantly improved its depth of reasoning and inference capabilities by leveraging increased computational resources and introducing algorithmic optimization mechanisms during post-training. The model has demonstrated outstanding performance across various benchmark evaluations, including mathematics, programming, and general logic. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro.
Deepseek V3.2
DeepSeek-V3.2 is a model that harmonizes high computational efficiency with superior reasoning and agent performance. Our approach is built upon three key technical breakthroughs: Deepseek Sparse Attention, Scalable Reinforcement Learning Framework, Large-Scale Agentic Task Synthesis Pipeline
GLM 4.7
GLM 4.7 is your new coding partner coming with the following features: Core Coding, Vibe Coding, Tool Using, and Complex Reasoning.
GLM 5
We are launching GLM-5, targeting complex systems engineering and long-horizon agentic tasks. Scaling is still one of the most important ways to improve the intelligence efficiency of Artificial General Intelligence (AGI). Compared to GLM-4.5, GLM-5 scales from 355B parameters (32B active) to 744B parameters (40B active), and increases pre-training data from 23T to 28.5T tokens. GLM-5 also integrates DeepSeek Sparse Attention (DSA), largely reducing deployment cost while preserving long-context capacity.
GPT-OSS-120b
OpenAI's open-weight models are designed for powerful reasoning, agentic tasks, and versatile developer use cases.
Kimi K2.5
Kimi K2.5 is an open-source, native multimodal agentic model built through continual pretraining on approximately 15 trillion mixed visual and text tokens atop Kimi-K2-Base. It seamlessly integrates vision and language understanding with advanced agentic capabilities, instant and thinking modes, as well as conversational and agentic paradigms.
Llama 4 Maverick 17B 128E
The Llama 4 collection of models are natively multimodal AI models that enable text and multimodal experiences. These models leverage a mixture-of-experts architecture to offer industry-leading performance in text and image understanding. These Llama 4 models mark the beginning of a new era for the Llama ecosystem. We are launching two efficient models in the Llama 4 series, Llama 4 Scout, a 17 billion parameter model with 16 experts, and Llama 4 Maverick, a 17 billion parameter model with 128 experts.
Llama 3.3 70B
The Meta Llama 3.3 multilingual large language model (LLM) is an instruction tuned generative model in 70B (text in/text out). The Llama 3.3 instruction tuned text only model is optimized for multilingual dialogue use cases and outperforms many of the available open source and closed chat models on common industry benchmarks.
MiniMax M2.5
MiniMax-M2.5 is extensively trained with reinforcement learning in hundreds of thousands of complex real-world environments, M2.5 is SOTA in coding, agentic tool use and search, office work, and a range of other economically valuable tasks, boasting scores of 80.2% in SWE-Bench Verified, 51.3% in Multi-SWE-Bench, and 76.3% in BrowseComp (with context management).
Qwen3.5 397B A17B
Qwen3.5 features a Unified Vision-Language Foundation, Efficient Hybrid Architecture, Scalable RL Generalization, Global Linguistic Coverage, and Next-Generation Training Infrastructure.
Qwen3.5 35B A3B
A smaller but faster version of the Qwen3.5 397B A17B model
Qwen3 Coder Next 80B
A coding focused variant of the Qwen3 model series: Super efficient with significant performance, Advanced agentic capabilities, Versatile integration with real-world IDE's