Models#
In general, I would always recommend you to use the GGUF quants. Will provide much more “bang for the buck” than the original weights.
Recommended quants are Q4_K_M
, Q5_K_M
and Q6_K
Accessing them in Ollama#
As you saw previously, the template is quite straightforward:
ollama run hc.co/<user>/<model_id>:<quant>
Recommended models#
Gemma-2 (9B/27B)
Phi-4 (14b)
Llama-3.2 (3B/1B)
Mistral Small 3 (2501) (24B)