The best LLMs for your use case:
Qwen's native multimodal MoE model with 397B total parameters and 17B active, featuring hybrid Gated Delta Networks for strong reasoning and vision capabilities.
Speed:
Intelligence:
Price: (1M Tokens)
$0.60 / 3.60Inputs:
JSON Mode:
Function Calling:
Benchmarks:
Multilingual MMLU
Multilingual
GPQA-Diamond
General Knowledge
MMLU-Pro
General Knowledge
LongBenchv2
Summarization
MMMU
Multimodal - Vision
BFCL
Agents and Function Calling
LMArena
Chat
LiveCodeBench
Code
Multilingual MMLU
Multilingual
GPQA-Diamond
General Knowledge
MMLU-Pro
General Knowledge
LongBenchv2
Summarization
MMMU
Multimodal - Vision
BFCL
Agents and Function Calling
LMArena
Chat
LiveCodeBench
Code
SOTA 128-expert MoE powerhouse for multilingual image/text understanding, creative writing, and enterprise-scale applications.
Speed:
Intelligence:
Price: (1M Tokens)
$0.27Inputs:
JSON Mode:
Function Calling:
Benchmarks:
Multilingual MMLU
Multilingual
MMMU
Multimodal - Vision
GPQA-Diamond
General Knowledge
WebDevArena
Code
EQBench
Creative Writing
LMArena
Chat
BFCL
Agents and Function Calling
LiveCodeBench
Code
MMLU-Pro
General Knowledge
Multilingual MMLU
Multilingual
MMMU
Multimodal - Vision
GPQA-Diamond
General Knowledge
WebDevArena
Code
EQBench
Creative Writing
LMArena
Chat
BFCL
Agents and Function Calling
LiveCodeBench
Code
MMLU-Pro
General Knowledge
Use case:
Multilingual