Announcing General Availability of Together Instant Clusters, offering ready to use, self-service NVIDIA GPUs
Blog post from Together AI
Together Instant Clusters are now generally available, providing an API-first, self-service platform for AI infrastructure that automates the setup and management of GPU clusters, ranging from single-node to large multi-node configurations with hundreds of interconnected GPUs. This service is designed to streamline AI workflows by enabling rapid provisioning and scaling without the need for lengthy procurement processes or manual approvals. The clusters come pre-configured with essential components such as NVIDIA drivers, CUDA versions, and networking operators, ensuring they are production-ready and optimized for low-latency inference and high-throughput distributed training. Ideal for AI Native companies facing variable demand, Together Instant Clusters allow for quick scaling to accommodate intense training runs and inference traffic. They offer robust reliability measures to ensure cluster stability and performance, with continuous monitoring and real-time anomaly detection. Pricing is straightforward, with options for hourly, daily, or longer-term usage, and includes free data transfer and reasonably priced shared storage. This innovation aims to enhance productivity and research velocity for AI teams by facilitating faster deployment and operation of high-performance GPU clusters.