Alright, this Qwen3 announcement draft is packed with energy and great info! Let’s take this “HOLD UP! π’” and transform it into a flowing, engaging news article, keeping all that excitement but ditching the bullet points for smooth paragraphs.
This one definitely has the “meat” for a ~600-word piece, given all the different models, features, and implications.
Here we go!

HOLD UP! π’ Qwen Team Unleashes Qwen3 AI Family β And a TON are OPEN SOURCE! π
Get ready, AI world, because the Qwen team just dropped a bombshell β their brand new AI family, Qwen3, is here! And the best part? They’re not keeping these powerhouses locked away; a massive number of these models are being released as OPEN SOURCE under the permissive Apache 2.0 license! This is huge news for developers, researchers, and anyone looking to get their hands on some seriously capable AI. So, let’s meet the crew and see what all the fuss is about!
Meet the Qwen3 Crew: From Monster MoEs to Efficient Minis!
The Qwen3 family isn’t just one model; it’s a whole lineup designed to tackle a range of tasks. Leading the pack is The Big Boss: Qwen3-235B-A22B. This absolute monster is a Mixture of Experts (MoE) model boasting a staggering 235 BILLION total parameters, though it cleverly only activates 22 billion at a time for efficiency. The Qwen team claims this heavyweight can throw punches with the likes of DeepSeek-R1, Gemini 2.5 Pro, and Grok-3! π₯ Then there’s The Scrappy MoE: Qwen3-30B-A3B. Don’t let its mere 3 billion active parameters fool you; this lean, mean machine supposedly smacks down older 32B dense models. Talk about punching above its weight! πͺ
But it doesn’t stop there. Qwen3 also offers a whole suite of Dense Models Galore, with sizes ranging from a hefty 32B all the way down to a tiny 0.6B parameter model. And here’s where it gets really interesting: even the smaller models, like the 4B version, are apparently rivaling the performance of much older 72B beasts! π€― Efficiency is clearly the name of the game with the Qwen3 family, aiming to deliver maximum bang for your computational buck.

Hybrid Thinking Modes: Ponder Deeply or Answer Instantly! π€β‘
One of the coolest and most practical innovations with Qwen3 is its Hybrid Thinking Modes. Need your AI to really dig in and reason step-by-step on those tough, complex problems β think intricate math, challenging code generation, or deep logical puzzles? You can flip its π€ Thinking Mode ON! In this mode, Qwen3 takes its sweet time to ponder, analyze, and (hopefully!) arrive at a more accurate and well-reasoned answer.
But what if you just need a quick, straightforward answer to a simple question? No problem! Tell it β‘ Non-Thinking Mode is a go! Switch off the deep pondering, and you can get near-instant replies. BRRRR! The best part is that you control this “thinking budget.” Developers can use a simple enable_thinking=True/False
flag in their code, or even use /think
and /no_think
tags directly in chat interfaces to switch modes on the fly. This level of flexibility is super handy, allowing you to optimize for either depth of thought or speed of response depending on the task at hand. β

Localization to the Max & Agent Power-Ups! ππ€
The Qwen team is clearly aiming for global reach because Qwen3 supports a whopping 119 languages and dialects! We’re talking Chinese, English, Spanish, Arabic, Japanese, Swahiliβ¦ basically, if there’s a significant language group on the planet, Qwen3 is likely trying to speak it. π€― This is massive for building truly international applications.
And for all you agent builders out there, Qwen3 has been juiced up with enhanced skills for coding and acting like an intelligent agent. This means better tool use, improved planning capabilities, and it even plays nice with the Model Context Protocol (MCP). To make leveraging these agent skills even easier, they’ve built a Qwen-Agent library. The goal? Less hassle, more doing! π
Get Your Hands On It! Where to Find and Run Qwen3 π»π
Ready to dive in and start playing with these new models? The Qwen team is making them widely available. You can find them hitting the streets on popular platforms like Hugging Face π€, Github, Kaggle, and Modelscope.
Wondering how to run them? For some Local Fun, you can use tools like OLLAMA (e.g., ollama run qwen3:30b-a3b
), LMStudio, MLX, and KTransformers. For more robust, Recommended Deployment, they suggest frameworks like SGLang and vLLM. And if you just want to see it in action quickly, you can try the Qwen Chat web/app demo over at chat.qwen.ai
.
Open Source FTW! π What’s Next for Qwen? π
Seriously, a massive round of applause for the Qwen team for open-sourcing so much of this powerful new family under the Apache 2.0 license! π This is how the whole AI community moves forward. The Qwen team isn’t shy about their ambitions, hinting at big AI plans (AGI/ASI vibes, anyone?) and talking about exciting future developments: even more data, bigger and more capable models, longer context windows, more senses (hello, increased modalities!), and even smarter agents. It sounds like we should all buckle up, because the AI ride is only getting wilder! π’
So, what are you waiting for? Go Try Qwen3! π₯ This is a fantastic opportunity to experiment with some cutting-edge open-source AI.
Get Your Hands On It! π»π
Ready to play? These models are hitting the streets on: