This repo contains a GGUF f16 export for use with llama.cpp / Ollama / LM Studio.
f16
model-f16.gguf
q4_k_m
quantize
16-bit