A sophisticated text-based Mixture-of-Experts (MoE) model featuring 21B total parameters with 3B activated per token, delivering exceptional multimodal understanding and generation through heterogeneous MoE structures and modality-isolated routing. Supporting an extensive 131K token context length, the model achieves efficient inference via multi-expert parallel collaboration and quantization, while advanced post-training techniques including SFT, DPO, and UPO ensure optimized performance across diverse applications with specialized routing and balancing losses for superior task handling.
| Router | Input / 1M | Output / 1M | Cached Input / 1M |
|---|---|---|---|
| OpenRouter | $0.07 | $0.28 | — |
| Martian | $0.07 | $0.28 | — |
baidu/ernie-4.5-21b-a3bRanked by provider, pricing, capabilities, and arena performance
Same provider · Similar price
Similar price · Both support tools
Similar price · Both support tools
Similar price · Both support tools
Same provider · Similar price
Similar price · Both support tools