🧠 Model

meta-llama-Llama-3.2-3B-Instruct-FP16

by context-labs

--- language: - en - de - fr - it - pt - hi - es - th library_name: transformers pipeline_tag: text-generation tags: - facebook - meta - pytorch - llama - llama

πŸ• Updated 12/21/2025
Compare This Model

Technical Specifications

Tasktext-generation
πŸ”„ Daily sync (11:00 Beijing)

Based on open-source metadata snapshot. Last synced: Dec 21, 2025

πŸ“Š FNI Methodology πŸ“š Knowledge Baseℹ️ Verify with original source

🧠 Architecture Explorer

Neural network architecture

1 Input Layer
2 Hidden Layers
3 Attention
4 Output Layer

Technical Specifications

Tasktext-generation

Model Card

--- language: - en - de - fr - it - pt - hi - es - th library_name: transformers pipeline_tag: text-generation tags: - facebook - meta - pytorch - llama - llama-3 license: llama3.2 extra_gated_prompt: >- ### LLAMA 3.2 COMMUNITY LICENSE AGREEMENT Llama 3.2 Version Release Date: September 25, 2024 β€œAgreement” means the terms and conditions for use, reproduction, distribution and modification of the Llama Materials set forth herein. β€œDocumentation” means the specifications, manuals and documenta...

πŸ“ Limitations & Considerations

  • β€’ Benchmark scores may vary based on evaluation methodology and hardware configuration.
  • β€’ VRAM requirements are estimates; actual usage depends on quantization and batch size.
  • β€’ FNI scores are relative rankings and may change as new models are added.
  • β€’ Source: Unknown

πŸ“š Related Resources

πŸ“„ Related Papers

No related papers linked yet. Check the model's official documentation for research papers.

πŸ“Š Training Datasets

Training data information not available. Refer to the original model card for details.

πŸ”— Related Models

Data unavailable

πŸš€ What's Next?