🔮All Our Models
Unsloth model catalog for all our Dynamic GGUF, 4-bit, 16-bit models on Hugging Face.
GGUFs let you run models in tools like Ollama, Open WebUI, and llama.cpp. Instruct (4-bit) safetensors can be used for inference or fine-tuning.
New & recommended models:
Model
Variant
GGUF
Instruct (4-bit)
DeepSeek models:
Model
Variant
GGUF
Instruct (4-bit)
Llama models:
Gemma models:
Qwen models:
Model
Variant
GGUF
Instruct (4-bit)
Mistral models:
Model
Variant
GGUF
Instruct (4-bit)
Phi models:
Other (GLM, Orpheus, Smol, Llava etc.) models:
Model
Variant
GGUF
Instruct (4-bit)
16-bit and 8-bit Instruct models are used for inference or fine-tuning:
New models:
Model
Variant
Instruct (16-bit)
DeepSeek models
Llama models
Gemma models:
Qwen models:
Family
Variant
Instruct (16-bit)
QVQ (preview)
72 B
—
Mistral models:
Phi models:
Text-to-Speech (TTS) models:
Base models are usually used for fine-tuning purposes:
New models:
Llama models:
Qwen models:
Llama models:
Gemma models
Mistral models:
Other (TTS, TinyLlama) models:
Last updated
Was this helpful?

