firecoperana
15159a87d4
Add vision support in llama-server ( #901 )
...
* server: add support for vision model
webui: add support for vision model
* server : remove hack for extra parallel slot#10187
* llama : fix KV shift for qwen2vl #13870
* add no-context-shift parameter
---------
Co-authored-by: firecoperana <firecoperana>
2025-11-05 10:43:46 +02:00
firecoperana
0f9ecaec04
Tool calls support from mainline ( #723 )
...
* Tool calls support from mainline
* update cmake
* revert api for /completions
* Fix broken thinking process for gpt-oss
* add missing args and fix webui bugs
* add missing args and fix webui bugs2
* Fix reasoning format error
* add usage
* change default post_sampling_probs to true
* add back generated_text
* Remove server endpoints tests
* add log
* Chat fixes
* Remove logs
* webui: revert extra handling of thinking process
---------
Co-authored-by: firecoperana <firecoperana>
Co-authored-by: Iwan Kawrakow <iwan.kawrakow@gmail.com >
2025-09-01 08:38:49 +03:00
firecoperana
df170c83a5
Webui improvement ( #481 )
...
* update webui
* add token/s in webui
* add webui files
* fix webui first message disappear in some browser
* add missing html files
---------
Co-authored-by: firecoperana <firecoperana>
2025-06-08 14:38:47 +03:00