| MiniMax M2.5 | MiniMax M2.5 is a frontier mixture-of-experts model with 230B total / 10B active params, trained via large-scale reinforcement learning. Scores 80.2% on SWE-Bench Verified and 76.3% on BrowseComp. | LLM | Ready | MiniMax | MiniMax-M2.5 | FP8 | FP16 | FP16 | Contact Sales | Supported | 197K | Not Supported | Text | Text |
| Kimi K2.5 | Kimi K2.5 is a native multimodal MoE with 1T total / 32B active params, supporting instant and thinking modes, agent swarm coordination of up to 100 sub-agents. | LLM | Ready | Moonshot AI | Kimi-K2.5 | BF16 | BF16 | BF16 | Contact Sales | Supported | 262K | Not Supported | Text / Image | Text |
| GLM 5 | GLM-5 is a MoE model from Z.ai with 744B total / 40B active params, trained on 28.5T tokens. Scores 73.3% on SWE-bench Verified. | LLM | Contact Sales | Z.ai (Zhipu) | GLM-5 | BF16 | BF16 | BF16 | Contact Sales | Supported | 203K | Not Supported | Text | Text |
| DeepSeek V3.2 | DeepSeek V3.2 is a MoE with 671B total / 37B active params. Introduces ‘thinking with tools’ capability; achieves gold-medal results at 2025 IMO and IOI. | LLM | Ready | DeepSeek | DeepSeek-V3.2 | FP8 | FP16 | FP16 | Contact Sales | Supported | 164K | Not Supported | Text | Text |
| gpt-oss-120b | Open-weight MoE from OpenAI with 117B total / 5.1B active params per forward pass, runs on a single H100 or AMD MI300X GPU using MXFP4 quantization. | LLM | Ready | OpenAI | gpt-oss-120b | BF16 | BF16 | BF16 | Contact Sales | Supported | 131K | Not Supported | Text | Text |
| gpt-oss-20b | Open-weight MoE from OpenAI with 21B total / 3.6B active params, designed for low-latency inference on consumer or single-GPU hardware. | LLM | Contact Sales | OpenAI | gpt-oss-20b | BF16 | BF16 | BF16 | Contact Sales | Supported | 131K | Not Supported | Text | Text |
| Qwen3 Instruct | MoE LLM with 235B total / 22B active params, optimized for instruction following across 100+ languages. 256K token context window. | LLM | Ready | Alibaba / Qwen | Qwen3-235B-A22B | BF16 | BF16 | BF16 | Contact Sales | Supported | 256K | Not Supported | Text | Text |
| Qwen3 Thinking | MoE reasoning model with 235B total / 22B active params, designed for deep logic, math, science, and complex multi-step coding tasks. | LLM | Ready | Alibaba / Qwen | Qwen3-235B-A22B | BF16 | BF16 | BF16 | Contact Sales | Supported | 256K | Not Supported | Text | Text |
| Qwen3 Coder | MoE model with 480B total / 35B active params, purpose-built for agentic coding. Native 256K context window, scalable to 1M tokens. | LLM | Ready | Alibaba / Qwen | Qwen3-Coder-480B-A35B-Instruct | BF16 | BF16 | BF16 | Contact Sales | Supported | 262K | Not Supported | Text | Text |
| Qwen3.5 | Multimodal MoE with 397B total / 17B active params and hybrid Gated Delta Network architecture enabling 1M token context window. | LLM | Ready | Alibaba / Qwen | Qwen3.5-397B-A17B | BF16 | BF16 | BF16 | Contact Sales | Supported | 1M | Not Supported | Text / Image | Text |
| Qwen3 VL Instruct | Vision-language MoE with 30B total / 3B active params, supporting megapixel-level inputs, multilingual OCR, visual grounding, and GUI automation. | LLM | Ready | Alibaba / Qwen | Qwen3-VL-235B-A22B-Instruct | BF16 | BF16 | BF16 | Contact Sales | Supported | 256K | Not Supported | Text / Image | Text |
| Qwen3 ASR | Advanced speech recognition with 1.7B params, supporting 52 languages, streaming and offline inference. Built on Qwen3-Omni architecture. | Audio | Ready | Alibaba / Qwen | Qwen3-ASR / Qwen3-TTS | BF16 | BF16 | BF16 | Contact Sales | Supported | N/A | Not Supported | Audio | Text |
| Qwen-Image | Open-weight diffusion transformer from Alibaba with strong text rendering accuracy and multilingual typography. | Image | Ready | Alibaba / Qwen | Qwen-Image-2512 | BF16 | BF16 | N/A | Contact Sales | Supported | N/A | Not Supported | Text | Image |
| Qwen-Image-Edit | Dedicated image editing variant of Qwen-Image series, supporting style transfer, object insertion, and detail enhancement. | Image | Ready | Alibaba / Qwen | Qwen-Image-Edit-2511 | BF16 | BF16 | N/A | Contact Sales | Supported | N/A | Not Supported | Text / Image | Image |
| Flux2 | Production-grade image generation from Black Forest Labs with 9B parameter flow transformer, using rectified flow matching for efficient inference. | Image | Ready | Black Forest Labs | FLUX.2-Klein-9B / FLUX.2-Klien-4B | BF16 | BF16 | N/A | Contact Sales | Supported | N/A | Not Supported | Text | Image |
| Stable Diffusion 3.5 | Diffusion transformer from Stability AI with the most mature open-source tooling ecosystem (ComfyUI, Automatic1111, Forge). | Image | Contact Sales | Stability AI | stable-diffusion-3.5-large | BF16 | BF16 | N/A | Contact Sales | Supported | N/A | Not Supported | Text | Image |
| Hunyuan Image | Large-scale multimodal autoregressive image generation from Tencent with 80B total params across 64 MoE experts, trained on 5B image-text pairs. | Image | Ready | Tencent | HunyuanImage-3.0 | BF16 | BF16 | N/A | Contact Sales | Supported | N/A | Not Supported | Text | Image |
| Z-Image | Open-weight diffusion transformer (Apache 2.0) for ultra-fast inference with accurate bilingual text rendering in English and Chinese. | Image | Contact Sales | Tongyi-MAI / Alibaba | Z-Image-Turbo | BF16 | BF16 | N/A | Contact Sales | Supported | N/A | Not Supported | Text | Image |
| Wan2.2-I2V | Open-source video diffusion model from Alibaba for image-to-video generation with strong motion consistency and scene coherence. | Video | Ready | Alibaba / Wan | Wan2.2-I2V | FP16 | FP16 | N/A | Contact Sales | Supported | N/A | Not Supported | Text / Image | Video |
| Wan2.2-T2V | Open-source video diffusion model from Alibaba for text-to-video generation with improved motion dynamics over Wan2.1. | Video | Ready | Alibaba / Wan | Wan2.2-T2V | FP16 | FP16 | N/A | Contact Sales | Supported | N/A | Not Supported | Text | Video |