Files
tabbyAPI/endpoints/OAI/types
DocShotgun 156b74f3f0 Revision to paged attention checks (#133)
* Model: Clean up paged attention checks

* Model: Move cache_size checks after paged attn checks
Cache size is only relevant in paged mode

* Model: Fix no_flash_attention

* Model: Remove no_flash_attention
Ability to use flash attention is auto-detected, so this flag is unneeded. Uninstall flash attention to disable it on supported hardware.
2024-06-09 17:28:11 +02:00
..
2024-03-18 00:53:27 -04:00
2024-05-28 00:52:30 -04:00
2024-03-18 15:59:28 -04:00
2024-03-12 23:59:30 -04:00