mirror of
https://github.com/ikawrakow/ik_llama.cpp.git
synced 2026-02-26 16:14:10 +00:00
Whitespace
This commit is contained in:
@@ -3298,7 +3298,7 @@ static bool llama_kv_cache_init(
|
|||||||
ggml_format_name(v, "cache_v_l%d", i);
|
ggml_format_name(v, "cache_v_l%d", i);
|
||||||
cache.k_l.push_back(k);
|
cache.k_l.push_back(k);
|
||||||
cache.v_l.push_back(v);
|
cache.v_l.push_back(v);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if (cparams.mla_attn && n_mla < n_layer && n_mla > 0) {
|
if (cparams.mla_attn && n_mla < n_layer && n_mla > 0) {
|
||||||
LLAMA_LOG_ERROR("%s: unexpected situation with %d out of %d layers having MLA enabled\n", __func__, n_mla, int(n_layer));
|
LLAMA_LOG_ERROR("%s: unexpected situation with %d out of %d layers having MLA enabled\n", __func__, n_mla, int(n_layer));
|
||||||
|
|||||||
Reference in New Issue
Block a user