The best LLMs for your use case:
Hybrid instruct + reasoning model (232Bx22B MoE) optimized for high-throughput, cost-efficient inference and distillation.
Speed:
Intelligence:
Price: (1M Tokens)
$0.20Inputs:
JSON Mode:
Function Calling:
Benchmarks:
Multilingual MMLU
Multilingual
MGSM
Multilingual
LiveBench
General Knowledge
EQBench
Creative Writing
LiveCodeBench
Code
Aider Polyglot
Code
BFCL
Agents and Function Calling
GPQA-Diamond
General Knowledge
MMLU-Pro
General Knowledge
LongBenchv2
Summarization
WebDevArena
Code
LMArena
Chat
Multilingual MMLU
Multilingual
MGSM
Multilingual
LiveBench
General Knowledge
EQBench
Creative Writing
LiveCodeBench
Code
Aider Polyglot
Code
BFCL
Agents and Function Calling
GPQA-Diamond
General Knowledge
MMLU-Pro
General Knowledge
LongBenchv2
Summarization
WebDevArena
Code
LMArena
Chat
SOTA 128-expert MoE powerhouse for multilingual image/text understanding, creative writing, and enterprise-scale applications.
Speed:
Intelligence:
Price: (1M Tokens)
$0.27Inputs:
JSON Mode:
Function Calling:
Benchmarks:
MGSM
Multilingual
Multilingual MMLU
Multilingual
MMMU
Multimodal - Vision
DocVQA
Multimodal - Vision
ChartQA
Multimodal - Vision
GPQA-Diamond
General Knowledge
LMArena
Chat
WebDevArena
Code
LiveBench
General Knowledge
Aider Polyglot
Code
EQBench
Creative Writing
LiveCodeBench
Code
BFCL
Agents and Function Calling
MMLU-Pro
General Knowledge
MGSM
Multilingual
Multilingual MMLU
Multilingual
MMMU
Multimodal - Vision
DocVQA
Multimodal - Vision
ChartQA
Multimodal - Vision
GPQA-Diamond
General Knowledge
LMArena
Chat
WebDevArena
Code
LiveBench
General Knowledge
Aider Polyglot
Code
EQBench
Creative Writing
LiveCodeBench
Code
BFCL
Agents and Function Calling
MMLU-Pro
General Knowledge
Use case:
Multilingual