DeepSeek-R1 Update: Enhanced Thinking Depth, Improved Reasoning Ability
DeepSeek has released a minor update to its R1 model, DeepSeek-R1-0528, based on the DeepSeek V3 Base, which significantly enhances the model's depth of thinking and reasoning ability with enhanced post-training. The new version excels in multiple benchmark tests for mathematics, programming, and general logic, with a substantial increase in accuracy in complex reasoning tasks such as AIME 2025. The article mentions that the model uses more tokens for in-depth thinking during problem-solving. In addition, the new model optimizes hallucination issues, reducing the hallucination rate by about 45-50%, and enhances creative writing capabilities. The API has also been updated to include support for tool calling and JSON output. The release includes an open-source model based on Qwen3-8B, trained using a distilled Chain of Thought from R1. Model weights are available for download on ModelScope and Huggingface under the MIT license.