Models#

In general, I would always recommend you to use the GGUF quants. Will provide much more “bang for the buck” than the original weights.

Recommended quants are Q4_K_M, Q5_K_M and Q6_K

Accessing them in Ollama#

As you saw previously, the template is quite straightforward:

ollama run hc.co/<user>/<model_id>:<quant>