mirror of
https://github.com/kvcache-ai/ktransformers.git
synced 2026-05-05 13:41:35 +00:00
Revert "[update] Reorganize documentation/README"
This commit is contained in:
@@ -1,6 +1,6 @@
|
||||
# Tutorial: Heterogeneous and Local MoE Inference
|
||||
# Tutorial: Heterogeneous and Local DeepSeek-V2 Inference
|
||||
|
||||
DeepSeek-(Code)-V2 is a series of strong mixture-of-experts (MoE) models, featuring a total of 236 billion parameters, with 21 billion parameters activated per token. This model has demonstrated remarkable reasoning capabilities across various benchmarks, positioning it as one of the SOTA open models and nearly comparable in performance to GPT-4. DeepSeek-R1 uses a similar architecture to DeepSeek-V2, but with a bigger number of parameters.
|
||||
DeepSeek-(Code)-V2 is a series of strong mixture-of-experts (MoE) models, featuring a total of 236 billion parameters, with 21 billion parameters activated per token. This model has demonstrated remarkable reasoning capabilities across various benchmarks, positioning it as one of the SOTA open models and nearly comparable in performance to GPT-4.
|
||||
|
||||
<p align="center">
|
||||
<picture>
|
||||
|
||||
Reference in New Issue
Block a user