Deep Cogito: Cogito V2 Preview Deepseek 671B

Text input Text output
Author's Description

Cogito v2 is a multilingual, instruction-tuned Mixture of Experts (MoE) large language model with 671 billion parameters. It supports both standard and reasoning-based generation modes. The model introduces hybrid reasoning via Iterated Distillation and Amplification (IDA)—an iterative self-improvement strategy designed to scale alignment with general intelligence. Cogito v2 has been optimized for STEM, programming, instruction following, and tool use. It supports 128k context length and offers strong performance in both multilingual and code-heavy environments. Users can control the reasoning behaviour with the `reasoning` `enabled` boolean. [Learn more in our docs](https://openrouter.ai/docs/use-cases/reasoning-tokens#enable-reasoning-with-default-config)

Key Specifications
Cost
$$$$
Context
163K
Parameters
671B
Released
Sep 02, 2025
Ability
Reliability
Supported Parameters

This model supports the following parameters:

Stop Top P Min P Frequency Penalty Max Tokens Reasoning Presence Penalty Include Reasoning Logit Bias Temperature
Features

This model supports the following features:

Reasoning
Performance Summary

Deep Cogito: Cogito V2 Preview Deepseek 671B, a multilingual, instruction-tuned Mixture of Experts model, demonstrates exceptional speed, consistently ranking among the fastest models across six benchmarks. Pricing information is currently unavailable, suggesting potential free tier usage. However, the model exhibits severe reliability issues, with a 0% success rate across all benchmarks due to frequent technical failures, rendering it unusable in its current state. Across all evaluated categories—Ethics, Instruction Following, Coding, Email Classification, Reasoning, and General Knowledge—the model achieved a 0.0% accuracy. This indicates a complete inability to provide correct or even usable responses, likely a direct consequence of the pervasive technical failures. While the model's architecture and stated optimizations for STEM, programming, instruction following, and tool use are promising, its current operational instability completely overshadows these potential strengths. The primary weakness is its critical lack of reliability, preventing any meaningful assessment of its intended capabilities.

Model Pricing

Current Pricing

Feature Price (per 1M tokens)
Prompt $1.25
Completion $1.25

Price History

Available Endpoints
Provider Endpoint Name Context Length Pricing (Input) Pricing (Output)
Together
Together | deepcogito/cogito-v2-preview-deepseek-671b 163K $1.25 / 1M tokens $1.25 / 1M tokens
Benchmark Results
Benchmark Category Reasoning Free Executions Accuracy Cost Duration
Other Models by deepcogito