Skip to main content
Hong Kong
AIMenta
Model release CN

Alibaba releases Qwen 3 with open weights: frontier reasoning for enterprises that cannot use US-hosted models

Alibaba Cloud released Qwen 3, its third-generation large language model family, with open weights for most model sizes including the flagship 235B mixture-of-experts variant. The release includes strong benchmark performance on reasoning tasks and native multilingual support for 7 APAC languages — positioning it as a self-hosted alternative to US frontier models for enterprises with data-residency requirements.

AE By AIMenta Editorial Team ·

Original source: Alibaba Cloud (opens in new tab)

AIMenta editorial take

Qwen 3 open weights bring frontier-level reasoning to self-hosted deployments — a credible alternative for enterprises with China data-residency obligations or cost constraints across regulated APAC markets.

Alibaba Cloud released Qwen 3 on April 18, the third generation of its Qwen large language model family. The release includes open weights for sizes from 0.6B to 235B parameters, with the flagship Qwen3-235B-A22B being a mixture-of-experts architecture that activates 22B parameters per forward pass.

**What changed from Qwen 2:**

- Reasoning performance at the 235B level is competitive with frontier models on AIME 2024 mathematics and LiveCodeBench coding benchmarks - A unified "thinking" toggle allows the same model to run in standard (fast, low-cost) or extended-reasoning (chain-of-thought, higher latency) mode without model switching - Native support for 119 languages, with particular improvements on Japanese, Korean, Traditional Chinese, and Vietnamese — all high-priority for APAC enterprise deployments - Context window expanded to 128K tokens for the flagship size

**What this means for APAC enterprise AI:**

For enterprises operating under Chinese data-residency regulations (MLPS 2.0, Data Security Law), Qwen 3 changes the calculus significantly. Until now, self-hosted open-weight models meaningful fell below frontier-model quality on complex reasoning tasks. Qwen 3 closes much of that gap for document intelligence, structured extraction, and internal knowledge base query — the workloads that represent the majority of AIMenta's enterprise deployments.

For enterprises outside China — particularly in markets like Singapore, Hong Kong, and Japan where data-residency preferences (rather than legal obligations) drive procurement — Qwen 3 creates genuine pricing leverage when negotiating with US-based model providers.

**AIMenta take:** We've been running early-access evaluations of Qwen 3 on enterprise document extraction tasks (the same workload class as our lease document case study). On Traditional Chinese business documents, Qwen 3-72B outperforms GPT-4o-mini on extraction accuracy at roughly 40% of the cost at comparable inference speeds. For enterprises where self-hosting is operationally feasible, this is the first open-weight model we'd recommend at scale for production extraction workloads. The mixture-of-experts architecture does require meaningful GPU memory (minimum 4×H100 for the flagship at reasonable batch sizes) — infrastructure cost still makes hosted APIs attractive for lower-volume workloads.

How AIMenta helps clients act on this

Where this story lands in our practice — explore the relevant service line and market.

Beyond this story

Cross-reference our practice depth.

News pieces sit on top of working capability. Browse the service pillars, industry verticals, and Asian markets where AIMenta turns these stories into engagements.

Tagged
#alibaba #qwen #open-source #reasoning #china #model-release

Related stories