133 Commits

Author SHA1 Message Date
turboderp
ca4f36feb2 Merge pull request #75 from LlamaEnjoyer/model_search_and_add_tweaks
feat: Improve model search and add debouncing
2025-02-05 23:50:35 +01:00
Llama Enjoyer
531c688eeb Make DRY sampling affected by the 'App defaults' button 2025-02-05 21:24:22 +01:00
Llama Enjoyer
5e669e491d feat: Improve model search and add debouncing
- Refined model search box appearance:
  - Set width and adjusted padding.
  - Added space for clear button.
- Added debouncing to model search input for performance optimization.
- Automatically set model name based on directory when creating new models.
- Added general-purpose debounce utility function.
2025-02-05 11:02:27 +01:00
turboderp
6906b144ee Adjust layout 2025-02-04 22:01:47 +01:00
turboderp
ce57e81982 Add DRY settings in chat mode 2025-02-04 22:01:10 +01:00
turboderp
42e4e871e0 Merge remote-tracking branch 'LlamaEnjoyer/read_sampling_params_from_model_config' 2025-02-04 21:23:48 +01:00
turboderp
dd48d472f1 Handle escape key in model search box 2025-02-04 21:20:51 +01:00
Llama Enjoyer
dfe827b864 minor css cleanup 2025-02-04 13:51:08 +01:00
Llama Enjoyer
8a08bd5580 minor css cleanup 2025-02-04 13:39:04 +01:00
Llama Enjoyer
664fb57fa2 fix: don't show the token counter if no model loaded 2025-02-02 22:28:04 +01:00
Llama Enjoyer
477c0ae2ac feat: add token counter and model search
Added /api/count_tokens endpoint to count tokens using model tokenizer
Implemented token counter in chat UI
Added model search functionality
Updated .gitignore to include .vscode/
2025-02-02 11:15:02 +01:00
Llama Enjoyer
8524c8cc43 feat: Add model parameter management and UI controls
Add support for reading generation_config.json for model parameters
Implement model_loaded_callback for parameter updates
Add UI controls to switch between model and app defaults
Add visual highlighting for modified parameters
Add new API endpoints for parameter management
Improve session handling of model parameters
Update UI styling for parameter controls
2025-02-01 00:01:13 +01:00
turboderp
f4d9478812 Merge pull request #64 from Downtown-Case/master
Add DRY and Banned Strings to Notepad
2025-01-24 20:28:43 +01:00
turboderp
a05e8e004a Merge pull request #70 from LlamaEnjoyer/qol_tweaks
Add auto session naming and context tracking qol features.
2025-01-24 20:25:08 +01:00
Llama Enjoyer
1f1de65b0d Merge branch 'more_qol_features' into qol_tweaks 2025-01-24 08:39:51 +01:00
Llama Enjoyer
3e0f293e94 Add cancel button to model loading overlay 2025-01-24 08:39:02 +01:00
Llama Enjoyer
7c577b448f Add auto session naming and context tracking qol features. 2025-01-14 22:51:05 +01:00
turboderp
36e488c183 Add Mistral-specific templates 2024-12-27 00:09:07 +01:00
turboderp
95c63efbae Merge remote-tracking branch 'github/master' 2024-10-10 00:17:43 +02:00
turboderp
206f25eb47 Bump to exllamav2 0.2.3 and fix for API changes 2024-10-10 00:17:32 +02:00
Downtown-Case
a2b8cf16f6 Add DRY and Banned Strings 2024-09-28 02:25:33 -04:00
Downtown-Case
729ebb933a Add DRY and Banned Strings 2024-09-28 02:24:10 -04:00
turboderp
bb017b72f5 Merge pull request #63 from LlamaEnjoyer/copy_to_clipboard
Add option to copy block of text to clipboard.
2024-09-23 18:19:01 +02:00
Llama Enjoyer
6b55c54e00 Add option to copy block of text to clipboard. 2024-09-22 11:57:42 +02:00
turboderp
cb0982152d Update reqs 2024-08-30 20:55:36 +02:00
turboderp
0d58ca8b31 Add switch to enable TP mode 2024-08-30 20:55:09 +02:00
turboderp
396a56f037 Add Q6 and Q8 cache modes 2024-06-09 21:56:20 +02:00
turboderp
47f3831cae Update exllamav2 to 0.1.5 2024-06-09 21:56:03 +02:00
turboderp
a529749be0 Include MarkedJS 2024-06-01 15:08:22 +02:00
turboderp
7f6f307a81 Assume subdirectories in app path rather than current path 2024-05-31 00:11:11 +02:00
turboderp
5e002a6a94 Add Granite prompt format 2024-05-30 20:55:49 +02:00
turboderp
81d9638a8f Raise maximum min/max/chunk tokens to 8k 2024-05-17 09:06:44 +02:00
turboderp
ee27aa01a4 Fix regression 2024-05-12 00:30:10 +02:00
turboderp
73d63761bf Add min tokens and banned strings 2024-05-11 23:21:05 +02:00
turboderp
cc9c9e3074 Bump exllamav2 to 0.0.21 2024-05-11 23:20:47 +02:00
turboderp
11aec2abf4 Add BOS to ChatML template 2024-05-05 17:09:01 +02:00
turboderp
426fbca344 Add fallback stop condition for Phi3 2024-04-24 22:24:31 +02:00
turboderp
af6021d147 Add Phi3 template 2024-04-24 21:15:43 +02:00
turboderp
b33d45d45b Add Llama3 template 2024-04-19 17:41:01 +02:00
turboderp
0e919cb1a1 Bump exllamav2 to 0.0.19 2024-04-19 14:04:06 +02:00
turboderp
444457af26 Add Cohere template 2024-04-06 09:46:02 +02:00
turboderp
984097f93c fix for some broken tokenizers 2024-03-30 11:06:38 +01:00
turboderp
86b6352461 ctrl+enter shortcut to regenerate last response 2024-03-27 06:55:53 +01:00
turboderp
2a23cb83f0 Add context options: delete-from-here, regenerate, complete 2024-03-27 06:46:24 +01:00
turboderp
a04c201ea8 Add save button to code blocks. 2024-03-26 05:28:25 +01:00
turboderp
84be9a4b4f Escape HTML in user blocks 2024-03-26 04:57:19 +01:00
turboderp
94ba8817d2 Prevent HTML tags from rendering in code blocks 2024-03-26 04:47:36 +01:00
turboderp
d3e2ccc782 Fix typo in Gemma prompt format 2024-03-26 04:35:30 +01:00
turboderp
415f49ffa4 Remove some console output 2024-03-21 02:00:24 +01:00
turboderp
64c201d66a Temp fix for exllamav2 tokenizer bug 2024-03-20 16:36:37 +01:00