Apple M1 Max vs Apple M1 Pro
Head-to-head AI inference comparison across 109 popular models. Each model is graded against both cards using its highest-quality quantization that still fits in VRAM. Bigger grade and faster tokens-per-second wins.
Spec
- VRAM
- Architecture
- Vendor
- MSRP
- Models running
- Wins (grade)
Apple M1 Max
- 64GB
- m1
- apple
- —
- 109 of 109
- 2 models
Apple M1 Pro
- 32GB
- m1
- apple
- —
- 109 of 109
- 0 models
Where Apple M1 Max pulls ahead
Where Apple M1 Pro pulls ahead
- No standout wins.
Language Models47 models
Max
tok/s
Model
tok/s
Pro
B
—
Llama 3.1 70B Instruct
70B · Meta
—
D
S
17
Qwen 2.5 32B
32B · Alibaba
42
A
S
17
Gemma 3 27B
27B · Google
42
S
S
17
Mistral Small 22B
22B · Mistral AI
42
S
S
26
Phi-4
14B · Microsoft
78
S
S
26
Qwen 2.5 14B
14B · Alibaba
78
S
S
26
Gemma 3 12B
12B · Google
78
S
S
26
Mistral Nemo 12B
12B · Mistral AI
78
S
S
26
Solar 10.7B
10.7B · Upstage
78
S
S
38
Falcon 3 10B
10B · TII
114
S
S
38
Gemma 2 9B Instruct
9.2B · Google
114
S
S
38
Yi 1.5 9B Chat
9B · 01.AI
114
S
S
38
DeepSeek R1 Distill 8B
8B · DeepSeek
114
S
S
38
Llama 3.1 8B Instruct
8B · Meta
114
S
S
38
Granite 3.3 8B
8B · IBM
114
S
S
38
EXAONE 3.5 7.8B
7.8B · LG AI
114
S
S
38
InternLM 2.5 7B
7.7B · Shanghai AI Lab
114
S
S
38
Qwen 2.5 7B Instruct
7.6B · Alibaba
114
S
S
38
Mistral 7B Instruct v0.3
7.3B · Mistral AI
114
S
S
38
Falcon 3 7B
7B · TII
114
S
S
38
OLMo 2 7B
7B · Allen AI
114
S
S
38
OpenChat 3.5 7B
7B · OpenChat
114
S
S
38
Yi 1.5 6B Chat
6B · 01.AI
114
S
S
62
Gemma 3 4B
4B · Google
168
S
S
62
Nemotron Mini 4B
4B · NVIDIA
168
S
S
62
Danube 3 4B
4B · H2O.ai
168
S
S
62
Phi-3.5 Mini 3.8B
3.8B · Microsoft
168
S
S
62
Phi-4 Mini 3.8B
3.8B · Microsoft
168
S
S
62
Llama 3.2 3B Instruct
3.2B · Meta
168
S
S
62
Qwen 2.5 3B
3B · Alibaba
168
S
S
62
Falcon 3 3B
3B · TII
168
S
S
62
StableLM Zephyr 3B
3B · Stability AI
168
S
S
62
Rocket 3B
3B · Pansophic
168
S
S
62
Gemma 2 2B
2.6B · Google
168
S
S
62
EXAONE 3.5 2.4B
2.4B · LG AI
168
S
S
90
Granite 3.3 2B
2B · IBM
216
S
S
90
SmolLM2 1.7B
1.7B · HuggingFace
216
S
S
90
Qwen 2.5 1.5B
1.5B · Alibaba
216
S
S
90
DeepSeek R1 Distill 1.5B
1.5B · DeepSeek
216
S
S
90
Llama 3.2 1B Instruct
1.24B · Meta
216
S
S
90
TinyLlama 1.1B
1.1B · TinyLlama
216
S
S
90
Gemma 3 1B
1B · Google
216
S
S
90
Falcon 3 1B
1B · TII
216
S
S
90
Qwen 2.5 0.5B
0.5B · Alibaba
216
S
S
90
Danube 3 500M
0.5B · H2O.ai
216
S
S
90
SmolLM2 360M
0.36B · HuggingFace
216
S
S
90
SmolLM2 135M
0.135B · HuggingFace
216
S
Code Models16 models
Max
tok/s
Model
tok/s
Pro
S
26
Qwen 2.5 Coder 14B
14B · Alibaba
78
S
S
26
Code Llama 13B Instruct
13B · Meta
78
S
S
38
Yi Coder 9B
9B · 01.AI
114
S
S
38
CodeGemma 7B
8.5B · Google
114
S
S
38
Qwen 2.5 Coder 7B
7.6B · Alibaba
114
S
S
38
StarCoder2 7B
7B · BigCode
114
S
S
38
Code Llama 7B
7B · Meta
114
S
S
38
DeepSeek Coder 6.7B
6.7B · DeepSeek
114
S
S
62
Qwen 2.5 Coder 3B
3B · Alibaba
168
S
S
62
StarCoder2 3B
3B · BigCode
168
S
S
62
Stable Code 3B
3B · Stability AI
168
S
S
90
CodeGemma 2B
2B · Google
216
S
S
90
Qwen 2.5 Coder 1.5B
1.5B · Alibaba
216
S
S
90
Yi Coder 1.5B
1.5B · 01.AI
216
S
S
90
DeepSeek Coder 1.3B
1.3B · DeepSeek
216
S
S
90
Qwen 2.5 Coder 0.5B
0.5B · Alibaba
216
S
Multimodal & Vision6 models
Max
tok/s
Model
tok/s
Pro
Image Generation9 models
Max
tok/s
Model
tok/s
Pro
S
26
FLUX.1 Schnell (GGUF)
12B · Black Forest Labs
78
S
S
26
FLUX.1 Dev (GGUF)
12B · Black Forest Labs
78
S
S
62
Stable Diffusion XL (CoreML)
3.5B · Stability AI
168
S
S
62
SDXL Turbo (GGUF)
3.5B · Stability AI
168
S
S
62
Stable Diffusion 3 Medium (GGUF)
2.5B · Stability AI
168
S
S
90
Stable Diffusion 2.1 Base (CoreML)
0.86B · Stability AI / Apple
216
S
S
90
Stable Diffusion 1.5 (CoreML)
0.86B · Runway
216
S
S
90
Stable Diffusion 1.5 (GGUF)
0.86B · Runway / GPUStack
216
S
S
90
Stable Diffusion 2.1 (GGUF)
0.86B · Stability AI
216
S
Speech9 models
Max
tok/s
Model
tok/s
Pro
S
90
Whisper Large v3
1.55B · OpenAI
216
S
S
90
Whisper Large v3 Turbo
0.81B · OpenAI
216
S
S
90
Whisper Medium
0.77B · OpenAI
216
S
S
90
Distil-Whisper Large v3
0.76B · HuggingFace
216
S
S
90
Whisper Small
0.24B · OpenAI
216
S
S
90
Whisper Base
0.074B · OpenAI
216
S
S
90
Whisper Base English
0.074B · OpenAI
216
S
S
90
Whisper Tiny English (Quantized)
0.039B · OpenAI
216
S
S
90
Whisper Tiny
0.039B · OpenAI
216
S
Text-to-Speech14 models
Max
tok/s
Model
tok/s
Pro
S
90
Kokoro 82M TTS
0.082B · Kokoro
216
S
S
90
Piper TTS - Amy (English)
0.02B · Rhasspy
216
S
S
90
Piper TTS - Lessac (English)
0.02B · Rhasspy
216
S
S
90
Piper TTS - LibriTTS-R (English)
0.02B · Rhasspy
216
S
S
90
Piper TTS - Spanish (MLS)
0.02B · Rhasspy
216
S
S
90
Piper TTS - French (Siwis)
0.02B · Rhasspy
216
S
S
90
Piper TTS - German (Thorsten)
0.02B · Rhasspy
216
S
S
90
Piper TTS - Chinese (Huayan)
0.02B · Rhasspy
216
S
S
90
Piper TTS - Japanese (Kokoro)
0.02B · Rhasspy
216
S
S
90
Piper TTS - Korean
0.02B · Rhasspy
216
S
S
90
Piper TTS - Russian (Irina)
0.02B · Rhasspy
216
S
S
90
Piper TTS - Portuguese (Faber)
0.02B · Rhasspy
216
S
S
90
Piper TTS - Italian (Riccardo)
0.02B · Rhasspy
216
S
S
90
Piper TTS - Arabic (Kareem)
0.02B · Rhasspy
216
S
Embeddings5 models
Max
tok/s
Model
tok/s
Pro
Rerankers2 models
Max
tok/s
Model
tok/s
Pro