mirror of
https://github.com/kvcache-ai/ktransformers.git
synced 2026-03-14 18:37:23 +00:00
Update Kimi-K2-Thinking-Native.md (#1673)
This commit is contained in:
@@ -14,6 +14,7 @@ This tutorial demonstrates how to run Kimi-K2 model inference using SGLang integ
|
||||
|
||||
**Minimum Configuration:**
|
||||
- **GPU**: NVIDIA RTX 4090 48GB (or equivalent with at least 48GB VRAM available)
|
||||
- **CPU**: Intel Xeon with AMX support (e.g., Sapphire Rapids)
|
||||
- **RAM**: At least 650GB system memory
|
||||
- **Storage**: ~600GB for model weights (native INT4 weight, same weight dir for CPU and GPU)
|
||||
|
||||
@@ -40,8 +41,6 @@ git checkout kimi_k2
|
||||
pip install -e "python[all]"
|
||||
```
|
||||
|
||||
|
||||
|
||||
3. **CUDA toolkit** - Compatible with your GPU (CUDA 11.8+ recommended)
|
||||
4. **Hugging Face CLI** - For downloading models:
|
||||
```bash
|
||||
|
||||
Reference in New Issue
Block a user