Home / Companies / Redis / Blog / Post Details
Content Deep Dive

LangCache public preview: Get fully managed semantic caching

Blog post from Redis

Post Details
Company
Date Published
Author
Jim Allen Wallace
Word Count
818
Language
English
Hacker News Points
-
Summary

LangCache, a fully managed semantic caching service by Redis, is now available for public preview to all Redis Cloud users as part of the Fall Release. Designed to address the challenges faced by AI-powered applications, such as high costs and latency associated with large language models (LLMs), LangCache eliminates redundant queries by semantically caching requests and responses, significantly reducing operational expenses and speeding up response times. With a cache hit rate potentially saving up to 70% of token usage and yielding responses up to 15 times faster, LangCache integrates seamlessly with Redis Cloud, offering effortless setup and comprehensive monitoring features. The service allows users to manage privacy and control settings, configure scopes, and attach custom attributes, while supporting both OpenAI and Redis's own embedding models. LangCache is particularly beneficial for AI applications with repetitive queries, such as customer support chatbots and retrieval augmented generation (RAG) systems, providing substantial cost savings and improved efficiency.