Aleksey Nikiforov
f5353047ef
Ported kimi-k2 support from llama.cpp ( #609 )
...
Original patch by @gabriellarson:
https://github.com/ggml-org/llama.cpp/pull/14654
Co-authored-by: anikifoss <anikifoss>
2025-07-14 18:43:52 +02:00
ubergarm
5446ccc8ac
add hunyuan moe support for 561 ( #565 )
...
* add hunyuan moe
* Don't reshape Vcur
* Apply chat template fix from mainline PR14584
2025-07-09 10:29:40 +02:00
Fizz~
6f3a3ba7e2
Special handling of Seed Coder FIM tokens ( #585 )
...
* Special handling of Seed Coder FIM tokens
* vocab: Add Seed Coder pretokenizer
* Formatting fix
* Update llama.h
2025-07-06 12:13:55 +02:00
Kawrakow
8e5106b20f
Add Falcon-Edge support ( #555 )
...
Co-authored-by: Iwan Kawrakow <iwan.kawrakow@gmail.com >
2025-06-26 08:48:52 +02:00
firecoperana
3f111ad7bb
add dry sampler ( #513 )
...
* add dry sampler
* use vocab instead of model in dry_init function
* fix compile error for build test
---------
Co-authored-by: firecoperana <firecoperana>
2025-06-19 10:24:53 +03:00
Kawrakow
474435f58b
LlaMA-4 support (text only) ( #321 )
...
* llama4: WIP
* llama4: this seems to be working
---------
Co-authored-by: Iwan Kawrakow <iwan.kawrakow@gmail.com >
2025-04-10 09:05:21 +02:00
saood06
2195632581
Deepseek V3 support added ( #176 )
...
Co-authored-by: Stanisław Szymczyk <sszymczy@gmail.com >
2025-01-23 18:24:10 +02:00
Kawrakow
8f43e55103
Merge mainline - Aug 12 2024 ( #17 )
...
* Merge mainline
* Fix after merge
* Remove CI check
---------
Co-authored-by: Iwan Kawrakow <iwan.kawrakow@gmail.com >
2024-08-12 15:14:32 +02:00
Kawrakow
154e0d75fc
Merge mainline llama.cpp ( #3 )
...
* Merging mainline - WIP
* Merging mainline - WIP
AVX2 and CUDA appear to work.
CUDA performance seems slightly (~1-2%) lower as it is so often
the case with llama.cpp/ggml after some "improvements" have been made.
* Merging mainline - fix Metal
* Remove check
---------
Co-authored-by: Iwan Kawrakow <iwan.kawrakow@gmail.com >
2024-07-27 07:55:01 +02:00