AI Model Rankings
Best models for every use case, ranked by quality and sorted by minimum VRAM. Find the best model your hardware can run.
💬
Chat & General
General-purpose language models for conversation, writing, and reasoning
💻
Coding
Specialized models for code generation, completion, and debugging
| # | Model | Author | Params | Min VRAM | Action |
|---|---|---|---|---|---|
| 1 | Qwen 2.5 Coder 0.5B | Alibaba | 0.5B | 0.9GB | Check |
| 2 | DeepSeek Coder 1.3B | DeepSeek | 1.3B | 1.3GB | Check |
| 3 | Qwen 2.5 Coder 1.5B | Alibaba | 1.5B | 1.5GB | Check |
| 4 | Yi Coder 1.5B | 01.AI | 1.5B | 1.5GB | Check |
| 5 | CodeGemma 2B | 2B | 1.9GB | Check | |
| 6 | Stable Code 3B | Stability AI | 3B | 2.3GB | Check |
| 7 | StarCoder2 3B | BigCode | 3B | 2.4GB | Check |
| 8 | Qwen 2.5 Coder 3B | Alibaba | 3B | 2.5GB | Check |
| 9 | Code Llama 7B | Meta | 7B | 4.7GB | Check |
| 10 | DeepSeek Coder 6.7B | DeepSeek | 6.7B | 4.7GB | Check |
| 11 | StarCoder2 7B | BigCode | 7B | 5GB | Check |
| 12 | Qwen 2.5 Coder 7B | Alibaba | 7.6B | 5.3GB | Check |
| 13 | CodeGemma 7B | 8.5B | 5.6GB | Check | |
| 14 | Yi Coder 9B | 01.AI | 9B | 6.2GB | Check |
| 15 | Code Llama 13B Instruct | Meta | 13B | 8.7GB | Check |
| 16 | Qwen 2.5 Coder 14B | Alibaba | 14B | 10GB | Check |
🎨
Image Generation
Text-to-image models for art, photos, and design
| # | Model | Author | Params | Min VRAM | Action |
|---|---|---|---|---|---|
| 1 | Stable Diffusion 2.1 Base (CoreML) | Stability AI / Apple | 0.86B | 2GB | Check |
| 2 | Stable Diffusion 1.5 (CoreML) | Runway | 0.86B | 2.5GB | Check |
| 3 | Stable Diffusion 1.5 (GGUF) | Runway / GPUStack | 0.86B | 2.5GB | Check |
| 4 | Stable Diffusion 2.1 (GGUF) | Stability AI | 0.86B | 2.5GB | Check |
| 5 | Stable Diffusion XL (CoreML) | Stability AI | 3.5B | 5GB | Check |
| 6 | SDXL Turbo (GGUF) | Stability AI | 3.5B | 5GB | Check |
| 7 | Stable Diffusion 3 Medium (GGUF) | Stability AI | 2.5B | 5.5GB | Check |
| 8 | FLUX.1 Schnell (GGUF) | Black Forest Labs | 12B | 14GB | Check |
| 9 | FLUX.1 Dev (GGUF) | Black Forest Labs | 12B | 14GB | Check |
🎤
Speech-to-Text
Transcription and speech recognition models
| # | Model | Author | Params | Min VRAM | Action |
|---|---|---|---|---|---|
| 1 | Whisper Tiny English (Quantized) | OpenAI | 0.039B | 0.1GB | Check |
| 2 | Whisper Tiny | OpenAI | 0.039B | 0.2GB | Check |
| 3 | Whisper Base | OpenAI | 0.074B | 0.3GB | Check |
| 4 | Whisper Base English | OpenAI | 0.074B | 0.3GB | Check |
| 5 | Whisper Small | OpenAI | 0.24B | 0.6GB | Check |
| 6 | Distil-Whisper Large v3 | HuggingFace | 0.76B | 1GB | Check |
| 7 | Whisper Medium | OpenAI | 0.77B | 1.8GB | Check |
| 8 | Whisper Large v3 Turbo | OpenAI | 0.81B | 1.9GB | Check |
| 9 | Whisper Large v3 | OpenAI | 1.55B | 3.5GB | Check |
🔊
Text-to-Speech
Voice synthesis and text-to-speech models
| # | Model | Author | Params | Min VRAM | Action |
|---|---|---|---|---|---|
| 1 | Piper TTS - Amy (English) | Rhasspy | 0.02B | 0.15GB | Check |
| 2 | Piper TTS - Lessac (English) | Rhasspy | 0.02B | 0.15GB | Check |
| 3 | Piper TTS - LibriTTS-R (English) | Rhasspy | 0.02B | 0.15GB | Check |
| 4 | Piper TTS - Spanish (MLS) | Rhasspy | 0.02B | 0.15GB | Check |
| 5 | Piper TTS - French (Siwis) | Rhasspy | 0.02B | 0.15GB | Check |
| 6 | Piper TTS - German (Thorsten) | Rhasspy | 0.02B | 0.15GB | Check |
| 7 | Piper TTS - Chinese (Huayan) | Rhasspy | 0.02B | 0.15GB | Check |
| 8 | Piper TTS - Japanese (Kokoro) | Rhasspy | 0.02B | 0.15GB | Check |
| 9 | Piper TTS - Korean | Rhasspy | 0.02B | 0.15GB | Check |
| 10 | Piper TTS - Russian (Irina) | Rhasspy | 0.02B | 0.15GB | Check |
| 11 | Piper TTS - Portuguese (Faber) | Rhasspy | 0.02B | 0.15GB | Check |
| 12 | Piper TTS - Italian (Riccardo) | Rhasspy | 0.02B | 0.15GB | Check |
| 13 | Piper TTS - Arabic (Kareem) | Rhasspy | 0.02B | 0.15GB | Check |
| 14 | Kokoro 82M TTS | Kokoro | 0.082B | 0.2GB | Check |
🎵
Audio Generation
AI music and audio creation
| # | Model | Author | Params | Min VRAM | Action |
|---|---|---|---|---|---|
| 1 | MusicGen Small | Meta | 0.3B | 1.5GB | Check |
👁️
Multimodal / Vision
Models that understand both images and text
| # | Model | Author | Params | Min VRAM | Action |
|---|---|---|---|---|---|
| 1 | Moondream 2 | Moondream | 1.8B | 1.5GB | Check |
| 2 | Qwen2-VL 2B | Alibaba | 2.2B | 2GB | Check |
| 3 | MiniCPM-V 2.6 | OpenBMB | 2B | 2.1GB | Check |
| 4 | PaliGemma 3B | 3B | 2.5GB | Check | |
| 5 | Phi-3.5 Vision | Microsoft | 4.2B | 3.2GB | Check |
| 6 | LLaVA 1.6 7B | LLaVA | 7B | 5GB | Check |
🔗
Embedding
Text embedding models for search and retrieval
| # | Model | Author | Params | Min VRAM | Action |
|---|---|---|---|---|---|
| 1 | BGE Small EN v1.5 | BAAI | 0.033B | 0.1GB | Check |
| 2 | Snowflake Arctic Embed S | Snowflake | 0.033B | 0.1GB | Check |
| 3 | all-MiniLM-L6-v2 | Sentence Transformers | 0.023B | 0.1GB | Check |
| 4 | Nomic Embed Text v1.5 | Nomic AI | 0.137B | 0.3GB | Check |
| 5 | BGE Large EN v1.5 | BAAI | 0.335B | 0.5GB | Check |
Can't Run the Model You Want?
Cloud GPUs give you instant access to any model, any size.
Need a GPU upgrade?
See our Best GPU for AI buying guide