Files
ik_llama.cpp/github-data/pull_requests/75 - Fix Q5_0 flash attention.md
2025-07-23 13:31:53 +02:00

455 B

🐛 #75 - Fix Q5_0 flash attention

Author ikawrakow
State Closed
Created 2024-10-01
Updated 2024-10-01

Description

When I changed iqk_mul_mat to use type-1 dot products for type-0 legacy quants, I forgot to also change the vec_dot_type when the dot product is done via ggml as in flash attention. This PR fixes it.