Guardrails has released a new version featuring the highly anticipated Guardrails Server, designed to streamline the deployment of Large Language Models (LLMs) by offering API access through an OpenAI SDK-compatible endpoint, allowing for cross-language support. The server facilitates easy cloud deployments, with guides available for platforms like AWS, and will soon cover GCP and Azure. The update includes Guardrails Watch for real-time monitoring and telemetry updates compatible with popular tools like Grafana and Datadog. JSON generation for open-source models, like those from Huggingface, is introduced through constrained decoding, expanding Guardrails’ applicability beyond closed-source models. Additionally, preview inference endpoints offer sub-second latency for tasks such as checking for profanity and personal information, with plans for self-hosting options in the future. Users are encouraged to contribute to the project and engage with the community via platforms like Discord and GitHub.