Llama 3.3 70B
by Meta · llama-3 family
70B
parameters
text-generation code-generation reasoning multilingual tool-use math creative-writing summarization
Llama 3.3 70B is Meta's most capable model in the Llama 3 series at the 70B parameter scale. It delivers performance competitive with much larger models like Llama 3.1 405B on many benchmarks, particularly in reasoning, coding, and multilingual tasks. This model represents a significant efficiency improvement, offering near-frontier performance in a size that can run on high-end consumer hardware with appropriate quantization. It supports 128K context and excels at instruction following, tool use, and complex reasoning.
Quick Start with Ollama
ollama run 70b-instruct-q4_K_M | Creator | Meta |
| Parameters | 70B |
| Architecture | transformer-decoder |
| Context Length | 128K tokens |
| License | Llama 3.3 Community License |
| Released | Dec 6, 2024 |
| Ollama | llama3.3 |
Quantization Options
| Format | File Size | VRAM Required | Quality | Ollama Tag |
|---|---|---|---|---|
| Q4_K_M recommended | 34.9 GB | 43.5 GB |
★
★
★
★
★
| 70b-instruct-q4_K_M |
| Q5_K_M | 40.8 GB | 50.5 GB |
★
★
★
★
★
| 70b-instruct-q5_K_M |
| Q8_0 | 64.8 GB | 72 GB |
★
★
★
★
★
| 70b-instruct-q8_0 |
Compatible Hardware for Q4_K_M
Showing compatibility for the recommended quantization (Q4_K_M, 43.5 GB VRAM).
Compatible Hardware
| Hardware | VRAM | Type | Fit |
|---|---|---|---|
| Mac Pro M2 Ultra 192GB | 192 GB | mac | Runs |
| Mac Studio M4 Ultra 192GB | 192 GB | mac | Runs |
| Mac Studio M4 Max 128GB | 128 GB | mac | Runs |
| MacBook Pro M4 Max 128GB | 128 GB | mac | Runs |
| Mac Studio M4 Max 64GB | 64 GB | mac | Runs |
| MacBook Pro M4 Max 64GB | 64 GB | mac | Runs |
| Mac mini M4 Pro 48GB | 48 GB | mac | Runs (tight) |
| MacBook Pro M4 Max 48GB | 48 GB | mac | Runs (tight) |
| MacBook Pro M4 Pro 48GB | 48 GB | mac | Runs (tight) |
| NVIDIA GeForce RTX 5090 | 32 GB | gpu | CPU Offload |
| Mac mini M4 32GB | 32 GB | mac | CPU Offload |
25 hardware
device(s) cannot run this model configuration.
Benchmark Scores
86.0
mmlu