As software teams increasingly rely on large language models (LLMs) for various tasks, monitoring these models has become crucial for maintaining consistent user experiences and ensuring system performance. LLM monitoring involves tracking performance indicators to detect changes or issues, and is typically visualized through dashboards that summarize model performance with charts and graphs. Effective monitoring highlights inefficiencies, provides early warnings for performance degradation, and supports regulatory compliance and responsible AI practices. Unlike traditional monitoring for deterministic systems, LLM monitoring accounts for the probabilistic and dynamic nature of LLMs, focusing on metrics like latency, token usage, correctness, and conversation turns to evaluate performance. Additionally, LLM monitoring is distinct from LLM observability, which delves into diagnosing root causes of performance shifts. As AI systems evolve, integrating platforms like Opik can help operationalize monitoring processes, enabling continuous improvement through real-world interaction data, structured logging, and combined automated and human evaluations.