Home / Companies / Fireworks AI / Blog / Post Details
Content Deep Dive

Qwen 3 on Fireworks AI: Controllable Chain-of-Thought and Tool Calling at Frontier Scale

Blog post from Fireworks AI

Post Details
Company
Date Published
Author
-
Word Count
869
Language
English
Hacker News Points
-
Summary

Qwen 3-235B-A22B, a model hosted on Fireworks, introduces a unique architecture that combines reasoning with function calls in a single pass, allowing users to choose between transparency and efficiency by adjusting the "reasoning_effort" parameter. This model, which operates serverlessly with 235 billion total parameters and 22 billion active parameters, is fully compatible with OpenAI's client, enabling seamless integration without the need for custom SDKs. It employs a Mixture-of-Experts approach, utilizing only eight experts at a time, which provides high-quality outputs with reduced computational demands. By segregating the reasoning component, users can selectively log or audit the reasoning process independently, and the model supports extensive context windows, making it suitable for large-scale applications under the Apache-2.0 license.