TAU2
Measured May 14, 2026Source
Score
0.34
An experimental reasoning-focused variant of the DeepSeek V3.2 model, optimized for complex logical deduction and chain-of-thought tasks. It likely maintains strong coding and general capabilities while enhancing performance on problems requiring multi-step reasoning.
Benchmark history
Score
0.34
Score
0.31
Score
0.69
Score
0.54
Score
0.88
Score
0.38
Score
0.79
Score
0.14
Score
0.8
Score
0.85
Score
87.7
Score
33.3
Score
32.9
CODPL speed
Plan availability

Thinking... Make sure you are connected to GitHub server