mirror of
https://github.com/ikawrakow/ik_llama.cpp.git
synced 2026-03-08 21:10:06 +00:00
Fix DeepSeek FA (#382)
Co-authored-by: Iwan Kawrakow <iwan.kawrakow@gmail.com>
This commit is contained in:
@@ -521,6 +521,7 @@ void ggml_cuda_flash_attn_ext(ggml_backend_cuda_context & ctx, ggml_tensor * dst
|
|||||||
// K and V head sizes.
|
// K and V head sizes.
|
||||||
if (K->ne[0] != V->ne[0]) {
|
if (K->ne[0] != V->ne[0]) {
|
||||||
ggml_cuda_flash_attn_ext_wmma_f16(ctx, dst);
|
ggml_cuda_flash_attn_ext_wmma_f16(ctx, dst);
|
||||||
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
ggml_cuda_flash_attn_ext_mma_f16(ctx, dst);
|
ggml_cuda_flash_attn_ext_mma_f16(ctx, dst);
|
||||||
|
|||||||
Reference in New Issue
Block a user