Company
Date Published
Author
TOGETHER AI
Word count
637
Language
English
Hacker News points
None

Summary

The Together AI Batch API offers businesses and developers a cost-effective solution for processing large volumes of LLM requests efficiently. By using batch processing, users can process non-urgent workloads at half the cost of real-time inference, with most batches completing within 24 hours. The API supports up to 50,000 requests in a single batch file, has simple integration with JSONL files, and provides real-time progress tracking. With the Batch API, users can scale their AI inference without scaling their budget, and pricing is based on successful completions at an introductory 50% discount.