MacBook Air M4 32GB
Apple · M4 · 32GB Unified Memory · Can run 86 models
Buy Apple
| Manufacturer | Apple |
| Unified Mem | 32 GB |
| Chip | M4 |
| CPU Cores | 10 |
| GPU Cores | 10 |
| Neural Engine | 16 |
| Bandwidth | 120 GB/s |
| MSRP | $1,299 |
| Released | Mar 12, 2025 |
AI Notes
The MacBook Air M4 32GB is the best ultrabook for local AI. With 32GB unified memory, it handles 13B models comfortably and can run 30B models with Q4 quantization. The fanless design runs completely silent. An excellent portable development machine for AI work.
Compatible Models
| Model | Parameters | Best Quant | VRAM Used | Fit | Est. Speed |
|---|---|---|---|---|---|
| Qwen 3 0.6B | 600M | Q4_K_M | 2.5 GB | Runs | ~48 tok/s |
| Qwen 3.5 0.8B | 800M | Q4_K_M | 1.5 GB | Runs | ~80 tok/s |
| Gemma 3 1B | 1B | Q8_0 | 2 GB | Runs | ~60 tok/s |
| Llama 3.2 1B | 1B | Q8_0 | 3 GB | Runs | ~40 tok/s |
| DeepSeek R1 1.5B | 1.5B | Q8_0 | 3 GB | Runs | ~40 tok/s |
| SmolLM2 1.7B | 1.7B | Q8_0 | 2.7 GB | Runs | ~44 tok/s |
| Gemma 2 2B | 2B | Q8_0 | 4 GB | Runs | ~30 tok/s |
| Gemma 3n E2B | 2B | Q4_K_M | 3.3 GB | Runs | ~36 tok/s |
| Gemma 4 E2B | 2B | Q4_K_M | 4 GB | Runs | ~30 tok/s |
| Qwen 3.5 2B | 2B | Q4_K_M | 3 GB | Runs | ~40 tok/s |
| Llama 3.2 3B | 3B | Q8_0 | 5 GB | Runs | ~24 tok/s |
| StarCoder2 3B | 3B | Q4_K_M | 3.5 GB | Runs | ~34 tok/s |
| Phi-3 Mini 3.8B | 3.8B | Q8_0 | 5.8 GB | Runs | ~21 tok/s |
| Phi-4 Mini 3.8B | 3.8B | Q4_K_M | 4.5 GB | Runs | ~27 tok/s |
| Gemma 3 4B | 4B | Q4_K_M | 5 GB | Runs | ~24 tok/s |
| Gemma 3n E4B | 4B | Q4_K_M | 4.5 GB | Runs | ~27 tok/s |
| Gemma 4 E4B | 4B | Q4_K_M | 6 GB | Runs | ~20 tok/s |
| Qwen 3 4B | 4B | Q4_K_M | 4.5 GB | Runs | ~27 tok/s |
| Qwen 3.5 4B | 4B | Q4_K_M | 4.5 GB | Runs | ~27 tok/s |
| Yi 1.5 6B | 6B | Q4_K_M | 5 GB | Runs | ~24 tok/s |
| Codestral Mamba 7B | 7B | Q4_K_M | 6.9 GB | Runs | ~17 tok/s |
| DeepSeek R1 7B | 7B | Q8_0 | 9 GB | Runs | ~13 tok/s |
| Falcon 3 7B | 7B | Q4_K_M | 6.8 GB | Runs | ~18 tok/s |
| InternLM 2.5 7B | 7B | Q4_K_M | 5.5 GB | Runs | ~22 tok/s |
| Mistral 7B | 7B | Q8_0 | 9 GB | Runs | ~13 tok/s |
| OpenChat 3.5 7B | 7B | Q4_K_M | 6.9 GB | Runs | ~17 tok/s |
| Qwen 2.5 7B | 7B | Q8_0 | 9 GB | Runs | ~13 tok/s |
| Qwen 2.5 Coder 7B | 7B | Q8_0 | 9 GB | Runs | ~13 tok/s |
| Qwen 2.5 VL 7B | 7B | Q4_K_M | 7 GB | Runs | ~17 tok/s |
| StarCoder2 7B | 7B | Q4_K_M | 5.5 GB | Runs | ~22 tok/s |
| WizardLM 2 7B | 7B | Q4_K_M | 6.9 GB | Runs | ~17 tok/s |
| Aya Expanse 8B | 8B | Q4_K_M | 6.5 GB | Runs | ~18 tok/s |
| Cogito 8B | 8B | Q4_K_M | 7.5 GB | Runs | ~16 tok/s |
| DeepSeek R1 8B | 8B | Q4_K_M | 7.5 GB | Runs | ~16 tok/s |
| Dolphin 3 8B | 8B | Q4_K_M | 6 GB | Runs | ~20 tok/s |
| Granite 3.3 8B | 8B | Q8_0 | 10 GB | Runs | ~12 tok/s |
| Llama 3.1 8B | 8B | Q8_0 | 10 GB | Runs | ~12 tok/s |
| Nemotron 3 Nano 8B | 8B | Q4_K_M | 7.5 GB | Runs | ~16 tok/s |
| Nous Hermes 2 8B | 8B | Q4_K_M | 6 GB | Runs | ~20 tok/s |
| Qwen 3 8B | 8B | Q4_K_M | 7.5 GB | Runs | ~16 tok/s |
| Gemma 2 9B | 9B | Q8_0 | 11 GB | Runs | ~11 tok/s |
| Qwen 3.5 9B | 9B | Q4_K_M | 7.5 GB | Runs | ~16 tok/s |
| Yi 1.5 9B | 9B | Q4_K_M | 6.5 GB | Runs | ~18 tok/s |
| Yi Coder 9B | 9B | Q4_K_M | 8 GB | Runs | ~15 tok/s |
| Falcon 3 10B | 10B | Q4_K_M | 8.5 GB | Runs | ~14 tok/s |
| Llama 3.2 Vision 11B | 11B | Q4_K_M | 8.5 GB | Runs | ~14 tok/s |
| Gemma 3 12B | 12B | Q4_K_M | 10.5 GB | Runs | ~11 tok/s |
| Mistral Nemo 12B | 12B | Q4_K_M | 9.5 GB | Runs | ~13 tok/s |
| DeepSeek R1 14B | 14B | Q4_K_M | 9.9 GB | Runs | ~12 tok/s |
| Phi-4 14B | 14B | Q4_K_M | 9.9 GB | Runs | ~12 tok/s |
| Phi-4 Reasoning 14B | 14B | Q4_K_M | 11 GB | Runs | ~11 tok/s |
| Qwen 2.5 14B | 14B | Q4_K_M | 9.9 GB | Runs | ~12 tok/s |
| Qwen 2.5 Coder 14B | 14B | Q4_K_M | 12 GB | Runs | ~10 tok/s |
| Qwen 3 14B | 14B | Q4_K_M | 12 GB | Runs | ~10 tok/s |
| StarCoder2 15B | 15B | Q8_0 | 17 GB | Runs | ~7 tok/s |
| InternLM 2.5 20B | 20B | Q4_K_M | 12 GB | Runs | ~10 tok/s |
| Codestral 22B | 22B | Q4_K_M | 14.7 GB | Runs | ~8 tok/s |
| Devstral 24B | 24B | Q4_K_M | 17 GB | Runs | ~7 tok/s |
| Magistral Small 24B | 24B | Q4_K_M | 17 GB | Runs | ~7 tok/s |
| Mistral Small 3.1 24B | 24B | Q4_K_M | 18 GB | Runs | ~7 tok/s |
| Gemma 4 26B | 26B | Q4_K_M | 20 GB | Runs | ~6 tok/s |
| Gemma 2 27B | 27B | Q4_K_M | 17.7 GB | Runs | ~7 tok/s |
| Gemma 3 27B | 27B | Q4_K_M | 20 GB | Runs | ~6 tok/s |
| Qwen 3.5 27B | 27B | Q4_K_M | 19 GB | Runs | ~6 tok/s |
| Qwen 3 30B-A3B (MoE) | 30B | Q4_K_M | 22 GB | Runs | ~5 tok/s |
| Gemma 4 31B | 31B | Q4_K_M | 22 GB | Runs | ~5 tok/s |
| Aya Expanse 32B | 32B | Q4_K_M | 22 GB | Runs | ~5 tok/s |
| Cogito 32B | 32B | Q4_K_M | 21.5 GB | Runs | ~6 tok/s |
| DeepSeek R1 32B | 32B | Q4_K_M | 20.7 GB | Runs | ~6 tok/s |
| Qwen 2.5 32B | 32B | Q4_K_M | 20.7 GB | Runs | ~6 tok/s |
| Qwen 2.5 Coder 32B | 32B | Q4_K_M | 23 GB | Runs | ~5 tok/s |
| Qwen 3 32B | 32B | Q4_K_M | 23 GB | Runs | ~5 tok/s |
| QwQ 32B | 32B | Q4_K_M | 21.5 GB | Runs | ~6 tok/s |
| WizardCoder 33B | 33B | Q4_K_M | 22 GB | Runs | ~5 tok/s |
| Nous Hermes 2 34B | 34B | Q4_K_M | 19 GB | Runs | ~6 tok/s |
| Yi 1.5 34B | 34B | Q4_K_M | 21 GB | Runs | ~6 tok/s |
| Command R 35B | 35B | Q4_K_M | 22.5 GB | Runs | ~5 tok/s |
| Qwen 3.5 35B A3B | 35B | Q4_K_M | 12 GB | Runs | ~10 tok/s |
| Dolphin Mixtral 8x7B | 47B | Q4_K_M | 26 GB | Runs | ~5 tok/s |
| Mixtral 8x7B | 47B | Q4_K_M | 29.7 GB | Runs (tight) | ~4 tok/s |
| Cogito 70B | 70B | Q4_K_M | 43 GB | CPU Offload | ~1 tok/s |
| DeepSeek R1 70B | 70B | Q4_K_M | 43.5 GB | CPU Offload | ~1 tok/s |
| Llama 3.1 70B | 70B | Q4_K_M | 43.5 GB | CPU Offload | ~1 tok/s |
| Llama 3.3 70B | 70B | Q4_K_M | 43.5 GB | CPU Offload | ~1 tok/s |
| Qwen 2.5 72B | 72B | Q4_K_M | 44.7 GB | CPU Offload | ~1 tok/s |
| Qwen 2.5 VL 72B | 72B | Q4_K_M | 41 GB | CPU Offload | ~1 tok/s |
19
model(s) are too large for this hardware.