LLM observability has become increasingly crucial for understanding and optimizing AI-driven applications, with tools like ClickStack providing robust solutions for monitoring performance, usage, and costs. This discussion focuses on using ClickStack, an open-source observability stack built on ClickHouse, to instrument LibreChat, an AI chat platform that integrates models from providers such as OpenAI and Google, and the ClickHouse MCP Server. By leveraging OpenTelemetry for capturing traces and metrics with minimal code changes, developers can gain insights into interactions between LibreChat, LLMs, and MCP services, allowing for real-time database queries and natural conversation interfaces. The implementation demonstrates how to deploy and configure ClickStack, enabling teams to monitor token consumption, usage patterns, and performance bottlenecks effectively. This observability framework not only enhances reliability and efficiency at scale but also facilitates proactive alerting on key properties to prevent unexpected cost surges or latency issues, ensuring a scalable, flexible, and cost-effective AI application management.