mirror of
https://github.com/kvcache-ai/ktransformers.git
synced 2026-03-14 18:37:23 +00:00
Update SFT Installation Guide for KimiK2
Added installation instructions and usage examples for KimiK2.
This commit is contained in:
@@ -42,7 +42,7 @@ pip install custom_flashinfer/
|
||||
|
||||
## Download Model
|
||||
|
||||
Download the official KIMI weights. If the weights are in FP8 format, please refer to https://github.com/kvcache-ai/ktransformers/pull/1559 to convert them to BF16 weights.
|
||||
Download the official KIMI weights. If the weights are in FP8 format, please refer to [convert_kimi_k2_fp8_to_bf16_cpu.py](https://github.com/kvcache-ai/ktransformers/blob/main/kt-kernel/scripts/convert_kimi_k2_fp8_to_bf16_cpu.py) to convert them to BF16 weights.
|
||||
|
||||
## How to start
|
||||
|
||||
@@ -131,4 +131,5 @@ use_kt: true # use KTransformers as LoRA sft backend to inference
|
||||
kt_optimize_rule: examples/kt_optimize_rules/DeepSeek-V3-Chat-sft-amx-multi-gpu.yaml
|
||||
cpu_infer: 32
|
||||
chunk_size: 8192
|
||||
```
|
||||
|
||||
```
|
||||
|
||||
Reference in New Issue
Block a user