Files
ik_llama.cpp/ggml
Kawrakow 190e7866db Quantization improvements (2) (#302)
* iq3_k: slightly better quantization

Not much of a difference for most models, but this change
avoids what it looks like a catastrophic failure for DeepSeek-Lite
(PPL is now 7.041 vs 7.314 on main).

* Small improvement for type-1 quants

---------

Co-authored-by: Iwan Kawrakow <iwan.kawrakow@gmail.com>
2025-04-01 10:31:06 +02:00
..
2024-07-27 07:55:01 +02:00
2025-04-01 10:31:06 +02:00
2024-07-27 07:55:01 +02:00