Revert "[update] Reorganize documentation/README"

This commit is contained in:
Azure
2025-02-15 03:43:48 +08:00
committed by GitHub
parent 19d4a50b1c
commit 4f4ed36442
7 changed files with 241 additions and 420 deletions

View File

@@ -1,6 +1,6 @@
# Tutorial: Heterogeneous and Local MoE Inference
# Tutorial: Heterogeneous and Local DeepSeek-V2 Inference
DeepSeek-(Code)-V2 is a series of strong mixture-of-experts (MoE) models, featuring a total of 236 billion parameters, with 21 billion parameters activated per token. This model has demonstrated remarkable reasoning capabilities across various benchmarks, positioning it as one of the SOTA open models and nearly comparable in performance to GPT-4. DeepSeek-R1 uses a similar architecture to DeepSeek-V2, but with a bigger number of parameters.
DeepSeek-(Code)-V2 is a series of strong mixture-of-experts (MoE) models, featuring a total of 236 billion parameters, with 21 billion parameters activated per token. This model has demonstrated remarkable reasoning capabilities across various benchmarks, positioning it as one of the SOTA open models and nearly comparable in performance to GPT-4.
<p align="center">
<picture>