Qwen 3-235B-A22B, a model hosted on Fireworks, introduces a unique architecture that combines reasoning with function calls in a single pass, allowing users to choose between transparency and efficiency by adjusting the "reasoning_effort" parameter. This model, which operates serverlessly with 235 billion total parameters and 22 billion active parameters, is fully compatible with OpenAI's client, enabling seamless integration without the need for custom SDKs. It employs a Mixture-of-Experts approach, utilizing only eight experts at a time, which provides high-quality outputs with reduced computational demands. By segregating the reasoning component, users can selectively log or audit the reasoning process independently, and the model supports extensive context windows, making it suitable for large-scale applications under the Apache-2.0 license.