Qwen API: Access Alibaba Cloud’s Qwen Models
AIsa is an Alibaba Cloud Qwen Key Account Partner, giving you production access to the complete Qwen model family — from the 1M-context Qwen3.6 Plus flagship to three specialised coder variants — through a single OpenAI-compatible API key at discounted partner pricing. No Alibaba Cloud account. No Aliyun registration. No RMB billing. One AIsa key covers every Qwen model alongside every other LLM in the AIsa catalogue.Supported Qwen models
| Model | Context window | Best for | Input price* | Output price* |
|---|---|---|---|---|
qwen3.6-plus | 1,000,000 tokens | Ultra-long context, frontier reasoning | $0.276/M | $1.651/M |
qwen3-max | 262,144 tokens | Balanced capability and cost | $0.72/M | $3.60/M |
qwen3-coder-plus | 262,144 tokens | Code generation and completion | $0.70/M | $3.50/M |
qwen3-coder-flash | 262,144 tokens | Fast code tasks, high-throughput coding | $0.21/M | $1.05/M |
qwen3-coder-480b-a35b-instruct | 262,144 tokens | Maximum coding capability (480B MoE) | $1.05/M | $5.25/M |
- Prices shown are reference market rates. AIsa Key Account partner pricing may be lower — see aisa.one/models for your actual rate.
Qwen3.6 Plus is currently in preview and available at no cost through AIsa during the preview period.
Quickstart
Python
Node.js
Streaming
Model guide
Qwen3.6 Plus — 1M context flagship
Qwen3.6 Plus is Alibaba Cloud’s latest frontier model, combining hybrid linear attention with sparse mixture-of-experts routing. Its 1,000,000-token context window makes it practical for entire codebases, book-length documents, or hours-long conversation history. Use when you need:- Context windows beyond 128K (legal documents, large repos, research corpora)
- Strong general reasoning with low latency relative to model capability
- The absolute latest Qwen release with ongoing improvements
Qwen3 Max — balanced workhorse
Qwen3 Max offers 262K context with strong performance across reasoning, writing, and multilingual tasks. It hits the sweet spot between capability and cost for production workloads. Use when you need:- Reliable general-purpose performance at predictable cost
- 262K context for most enterprise document tasks
- Stable, production-hardened model behaviour
Qwen3 Coder — three variants for every coding workload
AIsa supports three Qwen3 Coder variants, each targeting a different point on the speed/capability curve:| Model | Best for |
|---|---|
qwen3-coder-plus | Best quality-per-dollar for everyday coding tasks |
qwen3-coder-flash | Fast, high-throughput coding — lower latency than Plus |
qwen3-coder-480b-a35b-instruct | Maximum capability — full 480B MoE, activating 35B per request |
Switching from OpenAI
If you already use the OpenAI SDK, switching to Qwen takes one line:Function calling with Qwen
Qwen3 models support OpenAI-compatible function/tool calling:Data privacy
All Qwen requests through AIsa are processed under AIsa’s Alibaba Cloud Key Account enterprise agreement. Customer data is not used for model training or shared outside the processing pipeline. For compliance documentation, contact us.What’s next
- All Chinese AI models — compare Qwen vs DeepSeek vs Kimi vs ByteDance Seed
- DeepSeek models — DeepSeek V3.2 at industry-leading price-performance
- Kimi K2.5 — 1T parameter MoE with agentic tool-calling