As the AI landscape evolves, traditional "golden dataset" evaluations are becoming inadequate for complex tasks, particularly those using reinforcement learning (RL), due to their limited scope and inability to assess nuanced or partially correct responses. Rubric-based evaluations are increasingly favored, offering a granular, multi-dimensional approach that scores AI outputs across several criteria, providing detailed feedback essential for refining AI systems. This method is particularly valuable for RL, where rubrics help define precise reward signals, addressing challenges of reward sparsity and misspecification. Rubrics, traditionally used in education, are now applied to diverse AI tasks, enabling a holistic assessment of AI-generated code, chatbot responses, creative writing, and complex reasoning. Companies like Labelbox are pioneering rubric-based evaluations, partnering with leading AI labs to develop custom rubrics that leverage expert and AI evaluators to systematically assess AI outputs, providing actionable insights for model improvement. This approach marks a significant advancement in AI development, emphasizing the importance of nuanced, detailed feedback over binary correctness.