<RETURN_TO_BASE

DeepSeek Unveils V3.2 for Long Context Reasoning

Discover DeepSeek-V3.2, a model designed to enhance reasoning in long-context workloads with reduced costs.

Overview of DeepSeek-V3.2

How do you achieve GPT-5-level reasoning on long-context, tool-using workloads without incurring exorbitant costs? DeepSeek introduces DeepSeek-V3.2 and DeepSeek-V3.2-Speciale, two models focused on high-quality reasoning and adaptable workflows with open weights and production APIs.

Technology Behind DeepSeek

The models utilize DeepSeek Sparse Attention (DSA), a GRPO reinforcement learning stack, and a tool-native protocol. They report performance comparable to GPT-5, with DeepSeek-V3.2-Speciale reaching Gemini 3.0 Pro reasoning levels in benchmarks.

Sparse Attention Mechanism

Both models build on the DeepSeek-V3 Mixture of Experts transformer. With 671B total parameters and 37B active parameters per token, DSA enhances efficiency.

  • The attention complexity shifts from O(L²) to O(kL), improving cost-efficiency significantly.
  • Performance benchmarks show a 50% reduction in long-context inference costs.

Continued Pre-Training

DeepSeek Sparse Attention (DSA) originates from continued training on DeepSeek-V3.2 Terminus. In the initial stage, low-precision models optimize relevance scores through a limited number of steps. The sparse phase enables broader training with 944B tokens, leveraging calculated alignment losses.

GRPO and Reinforcement Learning

DeepSeek-V3.2 employs Group Relative Policy Optimization (GRPO) for reinforcement learning, with computing costs exceeding 10% of pre-training. Individual training methods target specific domains like mathematics and programming.

Agent Data and Protocols

The research team has developed an extensive synthetic agent dataset to support tool usage and reasoning. At inference, DeepSeek-V3.2 supports thinking and non-thinking modes, allowing flexible reasoning management

Competitions and Performance

DeepSeek-V3.2 and particularly Speciale have shown formidable performance, achieving gold-level results in top international math and programming competitions.

Key Takeaways

  1. DeepSeek-V3.2 introduces Sparse Attention, enabling more efficient long-context reasoning.
  2. The models retain a robust 671B parameter MoE backbone, accommodating practical long documents and workflows.
  3. GRPO reinforces the platform's learning capability, enhancing performance across multiple domains.
  4. Integrated thinking into tool use allows internal reasonings to persist during operations.

For further details, check the DeepSeek Paper and explore Model Weights.

🇷🇺

Сменить язык

Читать эту статью на русском

Переключить на Русский