Llama-3.1-Nemotron-Ultra-253B-v1 is a large language model (LLM) optimized for advanced reasoning, human-interactive chat, retrieval-augmented generation (RAG), and tool-calling tasks. Derived from Meta’s Llama-3.1-405B-Instruct, it has been significantly customized using Neural Architecture Search (NAS), resulting in enhanced efficiency, reduced memory usage, and improved inference latency. The model supports a context length of up to 128K tokens and can operate efficiently on an 8x NVIDIA H100 node. Note: you must include `detailed thinking on` in the system prompt to enable reasoning. Please see Usage Recommendations for more.
| Router | Input / 1M | Output / 1M | Cached Input / 1M |
|---|---|---|---|
| OpenRouter | $0.60 | $1.80 | — |
| Martian | $0.60 | $1.80 | — |
nvidia/llama-3.1-nemotron-ultra-253b-v1Ranked by provider, pricing, capabilities, and arena performance
Same family · Similar price
Same family · Similar price
Same provider · Similar price
Same family · Similar price
Same provider · Similar price
Same provider · Similar price