LangCache public preview: Get fully managed semantic caching
Blog post from Redis
LangCache, a fully managed semantic caching service by Redis, is now available for public preview to all Redis Cloud users as part of the Fall Release. Designed to address the challenges faced by AI-powered applications, such as high costs and latency associated with large language models (LLMs), LangCache eliminates redundant queries by semantically caching requests and responses, significantly reducing operational expenses and speeding up response times. With a cache hit rate potentially saving up to 70% of token usage and yielding responses up to 15 times faster, LangCache integrates seamlessly with Redis Cloud, offering effortless setup and comprehensive monitoring features. The service allows users to manage privacy and control settings, configure scopes, and attach custom attributes, while supporting both OpenAI and Redis's own embedding models. LangCache is particularly beneficial for AI applications with repetitive queries, such as customer support chatbots and retrieval augmented generation (RAG) systems, providing substantial cost savings and improved efficiency.