NVIDIA Nemotron 3 Nano, an advanced reasoning model, has been launched with Day-0 support on Fireworks, promising to enhance next-generation AI agents with its cutting-edge hybrid Mixture-of-Experts (MoE) architecture. This model, building on the Nemotron 2 Nano, combines a new MoE design with a hybrid transformer-mamba architecture, optimizing compute efficiency and accuracy, especially for applications like financial fraud detection and cybersecurity threat triaging. The Nemotron 3 Nano features 30 billion parameters but activates only 3 billion for inference, ensuring streamlined performance with a long context length of 1 million. Fireworks, known for its high-performance AI Inference Cloud powered by NVIDIA's latest GPU architectures, provides proprietary optimizations and custom kernel techniques to maximize throughput while maintaining model quality. The platform supports the deployment of Nemotron 3 Nano, enabling developers to efficiently handle tasks such as code summarization with a hands-on cookbook to guide them through setup and use. This model is particularly suitable for edge deployments and interactive workflows, offering a robust solution for extracting structure from code and improving the efficiency of internal tools and documentation systems.