Together AI has launched the DeepSeek V4 Pro, a substantial 1.6 trillion parameter Mixture-of-Experts (MoE) model, now accessible via its platform. This move brings advanced long-context reasoning capabilities to developers without the overhead of self-hosting.
The model boasts a 512K token context window on Together AI, expandable to a full 1 million tokens on dedicated infrastructure. This capacity is designed for complex tasks like analyzing entire code repositories or large document sets.
Controllable Reasoning Modes
DeepSeek V4 Pro introduces three distinct reasoning modes: Non-Think for rapid, low-complexity tasks; Think High for deeper analysis and multi-step reasoning; and Think Max for maximum effort in challenging scenarios like deep debugging.
