Files
tabbyAPI/common
kingbri 2096c9bad2 Model: Default max_seq_len to 4096
A common problem in TabbyAPI is that users who want to get up and
running with a model always had issues with max_seq_len causing OOMs.
This is because model devs set max context values in the millions which
requires a lot of VRAM.

To idiot-proof first time setup, make the fallback default 4096 so
users can run their models. If a user still wants to use the model's
max_seq_len, set it to -1.

Signed-off-by: kingbri <8082010+kingbri1@users.noreply.github.com>
2025-06-13 14:57:24 -04:00
..
2024-09-18 20:36:17 -04:00
2025-05-02 21:33:25 -04:00
2025-04-22 21:14:45 -04:00
2025-06-13 14:57:24 -04:00
2025-02-07 18:03:33 -05:00
2024-09-11 18:00:29 +01:00
2025-04-26 02:14:30 -04:00