From 1d62ac21f709d8c005ea126294e9df8ef3f1c96b Mon Sep 17 00:00:00 2001 From: Jiaqi Liao <30439460+SkqLiao@users.noreply.github.com> Date: Fri, 5 Dec 2025 23:08:02 +0800 Subject: [PATCH] Update Kimi-K2-Thinking-Native.md (#1673) --- doc/en/Kimi-K2-Thinking-Native.md | 3 +-- 1 file changed, 1 insertion(+), 2 deletions(-) diff --git a/doc/en/Kimi-K2-Thinking-Native.md b/doc/en/Kimi-K2-Thinking-Native.md index 8ce8d27..446b7e0 100644 --- a/doc/en/Kimi-K2-Thinking-Native.md +++ b/doc/en/Kimi-K2-Thinking-Native.md @@ -14,6 +14,7 @@ This tutorial demonstrates how to run Kimi-K2 model inference using SGLang integ **Minimum Configuration:** - **GPU**: NVIDIA RTX 4090 48GB (or equivalent with at least 48GB VRAM available) +- **CPU**: Intel Xeon with AMX support (e.g., Sapphire Rapids) - **RAM**: At least 650GB system memory - **Storage**: ~600GB for model weights (native INT4 weight, same weight dir for CPU and GPU) @@ -40,8 +41,6 @@ git checkout kimi_k2 pip install -e "python[all]" ``` - - 3. **CUDA toolkit** - Compatible with your GPU (CUDA 11.8+ recommended) 4. **Hugging Face CLI** - For downloading models: ```bash