Models

Four public model sizes, mirrored across base, chat, and quantized variants

The upstream README exposes a clear public catalog: 1.8B, 7B, 14B, and 72B, plus chat, Int4, and Int8 variants where available.

1.8B to 72BBase and chatQuantized variants

Model comparison

ModelReleaseMax lengthSystem promptPretrained tokensQ-LoRA memory2048-token Int4 generationTool use
Qwen-1.8B23.11.3032KYes2.2T5.8GB2.9GBYes
Qwen-7B23.08.0332KNo2.4T11.5GB8.2GBYes
Qwen-14B23.09.258KNo3.0T18.7GB13.0GBYes
Qwen-72B23.11.3032KYes3.0T61.4GB48.9GBYes

Values are mirrored from the upstream README tables and should be treated as source-cited historical product data.

Representative checkpoint paths

Variant cards

32K context · 2.2T tokens

Qwen-1.8B

The smallest family member still ships 32K context and system prompt support in the chat variant.

Open link

32K context · 2.4T tokens

Qwen-7B

The most practical open deployment target in the original line, with base, chat, Int4, and Int8 checkpoints.

Open link

8K context · 3.0T tokens

Qwen-14B

The 14B release pushed the original line deeper into coding and Chinese knowledge while retaining tool-use support.

Open link

32K context · 3.0T tokens

Qwen-72B

The flagship open release in the original repo, combining 32K context, stronger system prompts, and the top benchmark results.

Open link

Model distribution and ecosystem

ModelScope

Model hubs

Mirror the official model cards across both ModelScope and Hugging Face so download paths are visible in both English and Chinese contexts.

Open link

Managed API

DashScope API

The upstream README points to DashScope when you need a managed API surface instead of local model serving.

Open link

Agent framework

Qwen-Agent

The tool-use and code-interpreter sections connect directly to Qwen-Agent for evaluation and agent workflows.

Open link

Edge runtime

qwen.cpp

The original README highlights qwen.cpp as a lighter runtime path for the historical model line.

Open link

Source anchors

Models and Variants | Qwen Code