DeepSeek-R1-0528, an open-source model, has been released on Lambda's Inference API, challenging the dominance of OpenAI's o3 and Google's Gemini 2.5 Pro in complex tasks. The latest release builds upon the deepseek_v3 architecture, employing FP8 quantization to enhance its capabilities. At its core lies a robust architecture based on the DeepSeek-V3 backbone, utilizing a mixture-of-experts (MoE) model with multi-headed latent attention (MLA) and multi-token prediction (MTP). This approach enables efficient handling of complex reasoning tasks and allows the model to learn and improve through trial and error. R1-0528 demonstrates notable improvements over its predecessor across various benchmarks, achieving an impressive 87.5% accuracy in the AIME 2025 benchmark and scoring 73.3% on LiveCodeBench. The model now supports JSON output and function calling, enhancing its utility in various applications. It has significantly suppressed hallucination issues present in the legacy R1 version, leading to reliable and consistent outputs. With its sophisticated architecture, increased token utilization, and reduced dependence on supervised datasets, DeepSeek-R1-0528 is positioned to lead the next wave of AI advancements.