Llama-3_1-Nemotron-Ultra-253B-v1
by nvidia
Llama-3_1-Nemotron-Ultra-253B-v1 is an open-source AI model by nvidia
Technical Specifications
View Config (4 entries)
{
"architectures": [
"DeciLMForCausalLM"
],
"auto_map": {
"AutoConfig": "configuration_decilm.DeciLMConfig",
"AutoModelForCausalLM": "modeling_decilm.DeciLMForCausalLM"
},
"model_type": "nemotron-nas",
"tokenizer_config": {
"bos_token": "<|begin_of_text|>",
"chat_template": "{{- bos_token }}{%- if messages[0]['role'] == 'system' %}{%- set system_message = messages[0]['content']|trim %}{%- set messages = messages[1:] %}{%- else %}{%- set system_message = \"detailed thinking on\" %}{%- endif %}{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}{{- system_message }}{{- \"<|eot_id|>\" }}{%- for message in messages %}{%- if message['role'] == 'assistant' and '</think>' in message['content'] %}{%- set content = message['content'].split('</think>')[-1].lstrip() %}{%- else %}{%- set content = message['content'] %}{%- endif %}{{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n' + content | trim + '<|eot_id|>' }}{%- endfor %}{%- if add_generation_prompt %}{{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}{%- endif %}",
"eos_token": "<|eot_id|>"
}
}
Est. VRAM Required
~155 GB
Estimation Formula
VRAM = params Γ 0.6 + 2 GB
Based on FP16 precision.
β οΈ Does not account for KV cache or parallel overhead.
π Estimate only. Actual requirements may vary.
Based on open-source metadata snapshot. Last synced: Dec 31, 2025
π§ Architecture Explorer
Neural network architecture
Technical Specifications
View Config (4 entries)
{
"architectures": [
"DeciLMForCausalLM"
],
"auto_map": {
"AutoConfig": "configuration_decilm.DeciLMConfig",
"AutoModelForCausalLM": "modeling_decilm.DeciLMForCausalLM"
},
"model_type": "nemotron-nas",
"tokenizer_config": {
"bos_token": "<|begin_of_text|>",
"chat_template": "{{- bos_token }}{%- if messages[0]['role'] == 'system' %}{%- set system_message = messages[0]['content']|trim %}{%- set messages = messages[1:] %}{%- else %}{%- set system_message = \"detailed thinking on\" %}{%- endif %}{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}{{- system_message }}{{- \"<|eot_id|>\" }}{%- for message in messages %}{%- if message['role'] == 'assistant' and '</think>' in message['content'] %}{%- set content = message['content'].split('</think>')[-1].lstrip() %}{%- else %}{%- set content = message['content'] %}{%- endif %}{{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n' + content | trim + '<|eot_id|>' }}{%- endfor %}{%- if add_generation_prompt %}{{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}{%- endif %}",
"eos_token": "<|eot_id|>"
}
}
π Limitations & Considerations
- β’ Benchmark scores may vary based on evaluation methodology and hardware configuration.
- β’ VRAM requirements are estimates; actual usage depends on quantization and batch size.
- β’ FNI scores are relative rankings and may change as new models are added.
- β License Unknown: Verify licensing terms before commercial use.
- β’ Source: Huggingface
π Related Resources
π Related Papers
No related papers linked yet. Check the model's official documentation for research papers.
π Training Datasets
Training data information not available. Refer to the original model card for details.
π Related Models
Data unavailable