Mistral Small 3 24B GGUF quantization Evaluation results

https://preview.redd.it/ontcp7qk5cge1.png?width=790&format=png&auto=webp&s=867fa635acedb4047fe1b1a0a77f20d5eaa3534c

https://preview.redd.it/h92f0kol5cge1.png?width=1605&format=png&auto=webp&s=bc5d097366612440247bc260fd5c4bf2f4c10ce1

https://preview.redd.it/hzo2smfm5cge1.png?width=2321&format=png&auto=webp&s=df8455553ec547e9c17cd69022d1a6f86be766ab

Please note that the purpose of this test is to check if the model's intelligence will be significantly affected at low quantization levels, rather than evaluating which gguf is the best.

Regarding Q6_K-lmstudio: This model was downloaded from the lmstudio hf repo and uploaded by bartowski. However, this one is a static quantization model, while others are dynamic quantization models from bartowski's own repo.

gguf: https://huggingface.co/bartowski/Mistral-Small-24B-Instruct-2501-GGUF

Backend: https://www.ollama.com/

evaluation tool: https://github.com/chigkim/Ollama-MMLU-Pro

evaluation config: https://pastebin.com/mqWZzxaH