Home / Companies / Together AI / Blog / Post Details
Content Deep Dive

Introducing the Together AI Batch API: Process Thousands of LLM Requests at 50% Lower Cost

Blog post from Together AI

Post Details
Company
Date Published
Author
TOGETHER AI
Word Count
637
Language
English
Hacker News Points
-
Summary

The Together AI Batch API offers businesses and developers a cost-effective solution for processing large volumes of LLM requests efficiently. By using batch processing, users can process non-urgent workloads at half the cost of real-time inference, with most batches completing within 24 hours. The API supports up to 50,000 requests in a single batch file, has simple integration with JSONL files, and provides real-time progress tracking. With the Batch API, users can scale their AI inference without scaling their budget, and pricing is based on successful completions at an introductory 50% discount.