Skip to content
AI Viewer
DeepSeek Released September 22, 2025 Synced Mar 10, 2026

DeepSeek: DeepSeek V3.1 Terminus (exacto)

DeepSeek-V3.1 Terminus is an update to [DeepSeek V3.1](/deepseek/deepseek-chat-v3.1) that maintains the model's original capabilities while addressing issues reported by users, including language consistency and agent capabilities, further optimizing the model's performance in coding and search agents. It is a large hybrid reasoning model (671B parameters, 37B active) that supports both thinking and non-thinking modes. It extends the DeepSeek-V3 base with a two-phase long-context training process, reaching up to 128K tokens, and uses FP8 microscaling for efficient inference. Users can control the reasoning behaviour with the `reasoning` `enabled` boolean. [Learn more in our docs](https://openrouter.ai/docs/use-cases/reasoning-tokens#enable-reasoning-with-default-config) The model improves tool use, code generation, and reasoning efficiency, achieving performance comparable to DeepSeek-R1 on difficult benchmarks while responding more quickly. It supports structured tool calling, code agents, and search agents, making it suitable for research, coding, and agentic workflows.

Tool useReasoning

Why it stands out

164K-token context window handles longer documents and multi-turn conversations without truncation.
Combines tool use with reasoning — a strong baseline for agentic and multi-step workflows.
$0.21/M input makes it practical for always-on agents, batch processing, or high-volume classification.

What to watch

Text-only input — image or audio workflows require a separate model in the pipeline.
No benchmark score currently tracked — evaluate using task-specific testing alongside pricing and capability data.

Release timeline

Tracked events for DeepSeek: DeepSeek V3.1 Terminus (exacto).

Back to model tracker

release

DeepSeek: DeepSeek V3.1 Terminus (exacto) entered the tracked catalog

September 22, 2025

DeepSeek-V3.1 Terminus is an update to [DeepSeek V3.1](/deepseek/deepseek-chat-v3.1) that maintains the model's original capabilities while addressing issues reported by users, including language consistency and agent capabilities, further optimizing the model's performance in coding and search agents. It is a large hybrid reasoning model (671B parameters, 37B active) that supports both thinking and non-thinking modes. It extends the DeepSeek-V3 base with a two-phase long-context training process, reaching up to 128K tokens, and uses FP8 microscaling for efficient inference. Users can control the reasoning behaviour with the `reasoning` `enabled` boolean. [Learn more in our docs](https://openrouter.ai/docs/use-cases/reasoning-tokens#enable-reasoning-with-default-config) The model improves tool use, code generation, and reasoning efficiency, achieving performance comparable to DeepSeek-R1 on difficult benchmarks while responding more quickly. It supports structured tool calling, code agents, and search agents, making it suitable for research, coding, and agentic workflows.

View source

Nearby alternatives

Other DeepSeek models worth checking.

Need a recommendation instead?

Recent changes

LaunchSep 22

DeepSeek launched DeepSeek: DeepSeek V3.1 Terminus (exacto)

Compare

See how DeepSeek: DeepSeek V3.1 Terminus (exacto) stacks up.

All comparisons