AI research firm DeepSeek has introduced DeepSeek-V3.2, a next-generation large language model designed to combine computational efficiency with advanced reasoning and agentic capabilities.
The model is built on three core technical innovations. First, DeepSeek Sparse Attention (DSA) significantly reduces computational complexity while preserving performance, particularly in long-context scenarios. The company says this makes V3.2 more efficient without sacrificing accuracy in complex tasks.
Second, DeepSeek has implemented a scalable reinforcement learning (RL) framework that expands post-training compute to enhance reasoning performance. According to the company, DeepSeek-V3.2 performs comparably to GPT-5, while its high-compute variant, DeepSeek-V3.2-Speciale, surpasses GPT-5 and demonstrates reasoning proficiency on par with Gemini-3.0-Pro. The model reportedly achieved gold-medal-level performance in the 2025 International Mathematical Olympiad (IMO) and International Olympiad in Informatics (IOI).
Third, the company developed a large-scale agentic task synthesis pipeline to better integrate reasoning into tool-use environments. This pipeline systematically generates training data at scale, enabling improved compliance, generalisation, and interactive task performance in complex real-world scenarios.
To support transparency and community validation, DeepSeek has released its final submissions for IOI 2025, ICPC World Finals, IMO 2025 and CMO 2025, allowing secondary verification of results.
DeepSeek-V3.2 also introduces an updated chat template featuring a revised tool-calling format and a new “thinking with tools” capability. To ease adoption, the company has provided a dedicated encoding toolkit, including Python scripts and test cases that demonstrate how to convert OpenAI-compatible messages into model inputs and parse outputs effectively.
With these advancements, DeepSeek positions V3.2 as a high-performance, AI-native reasoning system optimised for both efficiency and scalable agent deployment.


