Overview
- The V3.1 update adopts a UE8M0 FP8 precision format that DeepSeek says is tailored for next‑generation homegrown chips expected soon, though no vendors were named.
- Built from a V3 checkpoint, the model consolidates reasoning and non‑reasoning into one system, with a new “deep thinking” toggle in DeepSeek’s app and web platform.
- Context capacity doubles to 131,072 tokens and the company reports stronger tool and function calling, including big gains on autonomous browsing benchmarks.
- Reporting indicates earlier attempts to train a next‑gen model on Huawei Ascend hardware struggled, leading DeepSeek to revert training to Nvidia H20 while evaluating Ascend for inference.
- Model weights are available on Hugging Face and ModeScope, and DeepSeek plans to change API pricing for developers on Sept. 6 as Chinese chip and foundry shares rally on the news.