Home / Companies / AssemblyAI / Blog / Post Details
Content Deep Dive

LLM Gateway: The easiest and most reliable way to call multiple LLMs

Blog post from AssemblyAI

Post Details
Company
Date Published
Author
Madison Bernstein
Word Count
1,678
Language
English
Hacker News Points
-
Summary

LLM Gateway has been relaunched as a streamlined solution for developers to efficiently interact with multiple large language models (LLMs) through a single OpenAI-compatible endpoint. This update introduces features such as automatic fallbacks, real-time streaming with tool calling, structured outputs, and prompt caching, alongside new models from Qwen and Moonshot. It aims to simplify the process for developers who otherwise manage multiple provider accounts and face issues like added fees and potential outages. The gateway supports a curated catalog of models and offers automatic fallback routing to ensure reliability without added latency. Its integration with AssemblyAI's infrastructure enables seamless operation within real-time voice agent pipelines, offering low latency and no extra network hops. The service emphasizes cost-effectiveness by passing through provider costs without additional markup, contrasting with competitors like OpenRouter.