Alibaba launches AI reasoning model to rival OpenAI’s o1

Alibaba’s Qwen team has introduced QwQ-32B-Preview, a reasoning AI model designed to challenge OpenAI’s o1 reasoning series, and the first available for download under a permissive license.

With 32.5 billion parameters and the ability to process prompts of up to 32,000 words, this model outperforms OpenAI’s o1-preview and o1-mini in several benchmarks, including the AIME and MATH tests. 

While OpenAI doesn’t disclose its parameter counts, models with higher parameter numbers generally exhibit superior problem-solving capabilities.

Reasoning models like QwQ-32B-Preview stand out for their ability to solve complex logic puzzles and math problems by planning and performing iterative actions. They also incorporate self-fact-checking mechanisms to enhance accuracy. 

However, these benefits come at the cost of slower response times, and like its peers, QwQ-32B-Preview has limitations—it may switch languages mid-response, get stuck in loops, or struggle with common-sense reasoning tasks.

QwQ-32B-Preview is available under an Apache 2.0 license, allowing for commercial use, but its release stops short of full transparency. Key components remain undisclosed, preventing replication or deeper exploration of the model’s workings. This positions the model as “open” but not fully open—a midpoint on the spectrum of AI accessibility.

Available for download via Hugging Face, QwQ-32B-Preview mirrors China’s AI regulatory stance. Like many Chinese-developed systems, it avoids politically sensitive topics and aligns with government narratives. For instance, it describes Taiwan as an “inalienable” part of China, reflecting state policy. Questions about events like Tiananmen Square yield no response.

QwQ-32B-Preview arrives as the scaling laws of AI—where larger datasets and computing power reliably yield improvements—face growing scrutiny. Reports indicate that AI labs like OpenAI, Google, and Anthropic are encountering diminishing returns with larger models, sparking a pivot toward new strategies like test-time compute. This technique allows models like o1 and QwQ-32B-Preview to take extra processing time during tasks, enabling them to work through problems more thoroughly and improve accuracy.

Aside from OpenAI and Chinese AI companies, other tech giants are looking into test-time compute as the future. According to a recent report from The Information, Google has expanded efforts in reasoning models, reportedly growing its internal team to 200 people and ramping up compute resources. 

Share this Post:

Accessibility Toolbar