add new talkinghead options to README

This commit is contained in:
Juha Jeronen
2024-01-10 00:25:39 +02:00
parent bac0d48f6c
commit a1f36f3547

View File

@@ -173,7 +173,9 @@ cd SillyTavern-extras
| `--mps` or `--m1` | Run the models on Apple Silicon. Only for M1 and M2 processors. |
| `--cuda` | Uses CUDA (GPU+VRAM) to run modules if it is available. Otherwise, falls back to using CPU. |
| `--cuda-device` | Specifies a CUDA device to use. Defaults to `cuda:0` (first available GPU). |
| `--talkinghead-gpu` | Uses GPU for talkinghead (10x FPS increase in animation). |
| `--talkinghead-gpu` | Uses GPU for talkinghead (10x FPS increase in animation). |
| `--talkinghead-model` | Load a specific THA3 model variant for talkinghead.<br>Default: `auto` (which is `separable_half` on GPU, `separable_float` on CPU). |
| `--talkinghead-models` | If THA3 models are not yet installed, downloads and installs them. Expects a HuggingFace model ID.<br>Default: [OktayAlpk/talking-head-anime-3](https://huggingface.co/OktayAlpk/talking-head-anime-3). |
| `--coqui-gpu` | Uses GPU for coqui TTS (if available). |
| `--coqui-model` | If provided, downloads and preloads a coqui TTS model. Default: none.<br>Example: `tts_models/multilingual/multi-dataset/bark` |
| `--summarization-model` | Load a custom summarization model.<br>Expects a HuggingFace model ID.<br>Default: [Qiliang/bart-large-cnn-samsum-ChatGPT_v3](https://huggingface.co/Qiliang/bart-large-cnn-samsum-ChatGPT_v3) |