Ray Serve has introduced several new features to enhance its flexibility and scalability for modern AI inference workloads, accommodating the needs of teams handling multimodal AI tasks. These features include Async Inference, which allows for the safe and efficient management of long-running workloads by integrating asynchronous processing into the serving layer, eliminating the need for additional infrastructure. Custom Request Routing provides precise control over request distribution, enabling domain-specific routing logic that can optimize system performance. Custom Autoscaling allows developers to define scaling policies using custom metrics, offering granular control to balance throughput, cost, and latency, while External Scaling permits programmatic adjustments to replica counts via external data sources. Collectively, these advancements make Ray Serve more adaptable and programmable, streamlining the process of deploying complex AI systems in production environments by ensuring safe execution, improved latency, and optimized resource utilization.