Mac Mini M4 (32GB)
$799
The 32GB Mac Mini M4 doubles the memory headroom, enabling 30B parameter models at Q4 quantization. The sweet spot for casual local LLM users.
Specifications
Memory 32GB UNIFIED
Memory Bandwidth 120 GB/s
GPU Cores 10
CPU Cores 10
TDP 22W
Max Model (Q4) 60B parameters
Max Model (Q8) 30B parameters
Performance Tier Budget
Category Apple Silicon
Performance Benchmarks
Llama 8B Q4 (tok/s) 40
SDXL 1024px (seconds) 18s
Flux 1024px (seconds) 45s
Pros
- Great value for 32GB unified memory
- Runs 13B-30B models at Q4
- Silent and power-efficient
Cons
- Same GPU performance as 16GB variant
- Not upgradeable after purchase
- Limited to smaller models at FP16
Compatible Models (Q4)
Models that fit in 32GB at Q4 quantization
Llama 3.2 8B Instruct 8B
6GB required Llama 3.2 3B Instruct 3B
3.5GB required Mistral 7B Instruct v0.3 7B
5.5GB required Gemma 3 27B Instruct 27B
15.5GB required Gemma 3 12B Instruct 12B
8GB required DeepSeek Coder V2 Instruct 236B
12.5GB required Qwen2.5 Coder 32B Instruct 32B
18GB required Qwen2.5 Coder 7B Instruct 7B
5.5GB required CodeLlama 34B Instruct 34B
19GB required StarCoder2 15B 15B
9.5GB required DeepSeek Coder 6.7B Instruct 6.7B
5.35GB required DeepSeek R1 671B
20.5GB required DeepSeek R1 Distill Qwen 32B 32B
18GB required DeepSeek R1 Distill Qwen 7B 7B
5.5GB required QwQ 32B 32B
18GB required Phi-4 14B
9GB required FLUX.1 Schnell 12B
6GB required FLUX.1 Dev 12B
6GB required Stable Diffusion XL Base 1.0 3.5B
3GB required Stable Diffusion 3.5 Large 8B
5GB required + 8 more models
Compatible at Q8
23 models can run at Q8 quantization