Files
ik_llama.cpp/tests
Iwan Kawrakow a0849e49f9 Merging mainline - WIP
AVX2 and CUDA appear to work.
CUDA performance seems slightly (~1-2%) lower as it is so often
the case with llama.cpp/ggml after some "improvements" have been made.
2024-07-26 17:16:21 +03:00
..
2024-03-09 14:17:11 +02:00
2024-07-26 16:32:40 +03:00
2024-07-26 16:32:40 +03:00
2024-01-29 15:50:50 -05:00
2024-07-26 16:32:40 +03:00
2024-07-26 16:32:40 +03:00
2024-07-26 16:32:40 +03:00
2024-07-26 17:16:21 +03:00
2024-07-26 16:32:40 +03:00
2024-07-26 16:32:40 +03:00
2024-07-26 16:32:40 +03:00