Alibaba’s Qwen Team Releases QwQ-32B Open-Source Reasoning Model, Said to Perform Similar to DeepSeek-R1

Date:

Alibaba’s Qwen Team, a division tasked with developing artificial intelligence (AI) models, released the QwQ-32B AI model on Wednesday. It is a reasoning model based on extended test time compute with visible chain-of-thought (CoT). The developers claim that despite being smaller in size compared to the DeepSeek-R1, the model can match its performance based on benchmark scores. Like other AI models released by the Qwen Team, the QwQ-32B is also an open-source AI model, however, it is not fully open-sourced.

QwQ-32B Reasoning AI Model Released

In a blog post, Alibaba’s Qwen Team detailed the QwQ-32B reasoning model. QwQ (short for Qwen with Questions) series AI models were first introduced by the company in November 2024. These reasoning models were designed to offer an open-source alternative for the likes of OpenAI’s o1 series. The QwQ-32B is a 32 billion parameter model developed by scaling reinforcement learning (RL) techniques.

Explaining the training process, the developers said that the RL scaling approach was added to a cold-start checkpoint. Initially, RL was used only for coding and mathematics-related tasks, and the responses were verified to ensure accuracy. Later the technique was used for general capabilities along with rule-based verifiers. The Qwen Team found that this method increased general capabilities of the model without reducing its math and coding performance.

qwq 32b benchmark QwQ 32B AI Model benchmarks

QwQ-32B AI Model benchmarks
Photo Credit: Alibaba

 

The developers claim that these training structures enabled the QwQ-32B to perform at similar levels to the DeepSeek-R1 despite the latter being a 671-billion-parameter model (with 37 billion activated). Based on internal testing, the team claimed that QwQ-32B outperforms DeepSeek-R1 in the LiveBench (coding), IFEval (chat or instruction fine-tuned language), and the Berkeley Function Calling Leaderboard V3 or BFCL (ability to call functions) benchmarks.

See also  WhatsApp Rolls Out Voice Message Transcripts With Support for English and Other Languages

Developers and AI enthusiasts can find the open weights of the model on Hugging Face listing and Modelscope. The model is available under the Apache 2.0 licence which allows academic and research-related usage but forbids commercial use cases. Additionally, since the full training details and datasets are not available, the model is also not replicable or can be deconstructed. DeepSeek-R1 was also available under the same licence.

In case one lacks the right hardware to run the AI model locally, they can also access its capabilities via Qwen Chat. The model picker menu at the top-left of the page will let users select the QwQ-32B-preview model.

Alibaba’s Qwen Team, a division tasked with developing artificial intelligence (AI) models, released the QwQ-32B AI model on Wednesday. It is a reasoning model based on extended test time compute with visible chain-of-thought (CoT). The developers claim that despite being smaller in size compared to the DeepSeek-R1, the model can match its performance based on benchmark scores. Like other AI models released by the Qwen Team, the QwQ-32B is also an open-source AI model, however, it is not fully open-sourced.

QwQ-32B Reasoning AI Model Released

In a blog post, Alibaba’s Qwen Team detailed the QwQ-32B reasoning model. QwQ (short for Qwen with Questions) series AI models were first introduced by the company in November 2024. These reasoning models were designed to offer an open-source alternative for the likes of OpenAI’s o1 series. The QwQ-32B is a 32 billion parameter model developed by scaling reinforcement learning (RL) techniques.

See also  DJI Flip All-in-One Vlog Camera Drone With 4K Video Support, AI-Powered Subject Tracking Launched

Explaining the training process, the developers said that the RL scaling approach was added to a cold-start checkpoint. Initially, RL was used only for coding and mathematics-related tasks, and the responses were verified to ensure accuracy. Later the technique was used for general capabilities along with rule-based verifiers. The Qwen Team found that this method increased general capabilities of the model without reducing its math and coding performance.

qwq 32b benchmark QwQ 32B AI Model benchmarks

QwQ-32B AI Model benchmarks
Photo Credit: Alibaba

 

The developers claim that these training structures enabled the QwQ-32B to perform at similar levels to the DeepSeek-R1 despite the latter being a 671-billion-parameter model (with 37 billion activated). Based on internal testing, the team claimed that QwQ-32B outperforms DeepSeek-R1 in the LiveBench (coding), IFEval (chat or instruction fine-tuned language), and the Berkeley Function Calling Leaderboard V3 or BFCL (ability to call functions) benchmarks.

Developers and AI enthusiasts can find the open weights of the model on Hugging Face listing and Modelscope. The model is available under the Apache 2.0 licence which allows academic and research-related usage but forbids commercial use cases. Additionally, since the full training details and datasets are not available, the model is also not replicable or can be deconstructed. DeepSeek-R1 was also available under the same licence.

In case one lacks the right hardware to run the AI model locally, they can also access its capabilities via Qwen Chat. The model picker menu at the top-left of the page will let users select the QwQ-32B-preview model.

 

Share post:

Subscribe

spot_imgspot_img

Popular

More like this
Related

South Carolina prepares for second firing squad execution

A firing squad is set to kill a South...

RRB ALP Recruitment 2025: Apply for 9,970 vacancies from April 12; check selection process and other details here

The RRB ALP Recruitment 2025 application process for 9,970...

‘Gauti (Gautam Gambhir) bhai has helped me understand my potential’

Washington Sundar, a versatile all-rounder, faces the challenge of...

Apple is left without a life raft as Trump’s China trade war intensifies, analysts warn

Apple remains stranded without a life raft, experts say,...