Update README.md

This commit is contained in:
inesggg 2025-08-06 15:49:08 +08:00 committed by GitHub
parent f6e34dd267
commit 2af264d674
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

View File

@ -42,7 +42,7 @@
9. [Contact](#9-contact)
## 1. Introduction
1. Introduction
We present DeepSeek-V3, a strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token.
To achieve efficient inference and cost-effective training, DeepSeek-V3 adopts Multi-head Latent Attention (MLA) and DeepSeekMoE architectures, which were thoroughly validated in DeepSeek-V2.