AI safety research is urgently important and should be supported by a wide range of public and private actors. Rapid AI progress is expected to lead to transformative AI systems with potentially large impacts on society, but the development of safe, reliable, and steerable systems remains a significant challenge. Anthropic's approach to AI safety research prioritizes empiricism, focusing on releasing a steady stream of safety-oriented research that has broad value for the AI community. The organization is working on developing techniques for scalable oversight, mechanistic interpretability, process-oriented learning, and understanding generalization to mitigate potential risks associated with advanced AI systems. Anthropic's goal is to develop a "portfolio" approach to AI safety research, addressing multiple angles and scenarios to ensure that their work can help succeed across different cases.