In April, DeepSeek officially launched the public beta testing of V4, with the context window suddenly raised from 128K Tokens to 1M (million) Tokens. Equipped with a new Ultra-MoE sparse activation architecture, the total parameter scale reaches 1 trillion, with only 13-37 billion parameters activated per inference. Inference speed is 35 times higher than V3, energy consumption reduced by 40%, and inference cost is only 1/70 of the GPT-4 series. Fully localized domestically, it deeply adapts to Huawei Ascend chips. Developers can process entire technical documents and large code libraries at once, and code accuracy is improved by more than 60% compared to V3.
#DeepSeekV4 #ChineseAI #LargeModel