DeepSeek-V4 Pro now available on Together AI
Blog post from Together AI
DeepSeek V4 Pro on Together AI offers advanced features for handling long-context reasoning workloads with a 512K-token context window and a large-scale 1.6T-parameter Mixture-of-Experts architecture, activating 49B parameters. It provides three controllable reasoning modes—Non-Think, Think High, and Think Max—allowing teams to tailor the depth of reasoning to specific tasks. The platform's serverless pricing model makes it accessible, with options to switch to dedicated infrastructure for larger contexts and production control. DeepSeek V4 Pro is designed for complex workloads, such as code agents, document intelligence, and research synthesis, where it can manage entire repositories and large document sets within its context window without compressing them into summaries. Cached input pricing offers significant cost savings for repeated queries, and the forthcoming DeepSeek-V4 Flash will provide additional options focusing on speed and cost-efficiency.