Qwen3.6-35B-A3B:智能体编程利器,现已开源

3
分类业界资讯
作者Qwen Team
来源跳转
发表时间

内容

Qwen3.6-Plus 发布之后,我们非常高兴地宣布开源 Qwen3.6-35B-A3B —— 一个稀疏但能力出色的混合专家(MoE)模型,总参数量为350亿,激活参数仅30亿。尽管高效轻量,Qwen3.6-35B-A3B 在智能体编程方面表现卓越,大幅超越前代模型 Qwen3.5-35B-A3B,并可与 Qwen3.5-27B 和 Gemma4-31B 等稠密模型一较高下。该模型依然支持多模态思考与非思考模式,是当前最具通用性的开源模型之一。现在,Qwen3.6-35B-A3B 已在 Qwen Studio 上线,可通过 API 调用,并以开源权重的形式向社区发布。

  • Qwen3.6-35B-A3B 是一个完全开源的 MoE 模型(总参数 35B / 激活参数 3B),主要特性包括: 卓越的智能体编程能力,可与大得多的模型相媲美
  • 强大的多模态感知与推理能力
  • 您可以在 Qwen Studio 进行交互对话, 通过 阿里云百炼 以 Qwen3.6-Flash 的名称调用 API (即将到来), 或从 Hugging FaceModelScope 下载模型权重。 image

模型表现

下文将全面展示 Qwen3.6-35B-A3B 与同规模模型在各类任务和模态上的评测对比结果。

自然语言

仅凭30亿激活参数,Qwen3.6-35B-A3B 在多项关键编程基准上超越了270亿参数的稠密模型 Qwen3.5-27B,并在智能体编程和推理任务上大幅超越其直接前代 Qwen3.5-35B-A3B。

Qwen3.5-27BGemma4-31BQwen3.5-35BA3BGemma4-26BA4BQwen3.6-35BA3B
Coding Agent
SWE-bench Verified75.052.070.017.473.4
SWE-bench Multilingual69.351.760.317.367.2
SWE-bench Pro51.235.744.613.849.5
Terminal-Bench 2.041.642.940.534.251.5
Claw-Eval Avg64.348.565.458.868.7
Claw-Eval Pass^346.225.051.028.050.0
SkillsBench Avg527.223.64.412.328.7
QwenClawBench52.241.747.738.752.6
NL2Repo27.315.520.511.629.4
QwenWebBench1068119797811781397
General Agent
TAU3-Bench68.467.568.959.067.2
VITA-Bench41.843.029.136.935.6
DeepPlanning22.624.022.816.225.9
Tool Decathlon31.521.228.712.026.9
MCPMark36.318.127.014.237.0
MCP-Atlas68.457.262.450.062.8
WideSearch66.435.259.138.360.1
Knowledge
MMLU-Pro86.185.285.382.685.2
MMLU-Redux93.293.793.392.793.3
SuperGPQA65.665.763.461.464.7
C-Eval90.582.690.282.590.0
STEM & Reasoning
GPQA85.584.384.282.386.0
HLE24.319.522.48.721.4
LiveCodeBench v680.780.074.677.180.4
HMMT Feb 2592.088.789.091.790.7
HMMT Nov 2589.887.589.287.589.1
HMMT Feb 2684.377.278.779.083.6
IMOAnswerBench79.974.576.874.378.9
AIME2692.689.291.088.392.7
  • SWE-Bench Series: Internal agent scaffold (bash + file-edit tools); temp=1.0, top_p=0.95, 200K context window. We correct some problematic tasks in the public set of SWE-bench Pro and evaluate all baselines on the refined benchmark.
  • Terminal-Bench 2.0: Harbor/Terminus-2 harness; 3h timeout, 32 CPU/48 GB RAM; temp=1.0, top_p=0.95, top_k=20, max_tokens=80K, 256K ctx; avg of 5 runs.
  • SkillsBench: Evaluated via OpenCode on 78 tasks (self-contained subset, excluding API-dependent tasks); avg of 5 runs.
  • NL2Repo: Others are evaluated via Claude Code (temp=1.0, top_p=0.95, max_turns=900).
  • QwenClawBench: An internal real-user-distribution Claw agent benchmark (open-sourcing soon); temp=0.6, 256K ctx.
  • QwenWebBench: An internal front-end code generation benchmark; bilingual (EN/CN), 7 categories (Web Design, Web Apps, Games, SVG, Data Visualization, Animation, and 3D); auto-render + multimodal judge (code/visual correctness); BT/Elo rating system.
  • TAU3-Bench: We use the official user model (gpt-5.2, low reasoning effort) + default BM25 retrieval.
  • VITA-Bench: Avg subdomain scores; using claude-4-sonnet as judger, as the official judger (claude-3.7-sonnet) is no longer available.
  • MCPMark: GitHub MCP v0.30.3; Playwright responses truncated at 32K tokens.
  • MCP-Atlas: Public set score; gemini-2.5-pro judger.
  • AIME 26: We use the full AIME 2026 (I & II), where the scores may differ from Qwen 3.5 notes.

视觉语言

Qwen3.6 原生支持多模态,Qwen3.6-35B-A3B 以仅约 30 亿激活参数,展现出远超其体量的感知与多模态推理能力。在大多数视觉语言基准上,它的表现已与 Claude Sonnet 4.5 持平,甚至在部分任务上实现超越。其在空间智能上的优势尤为突出:RefCOCO 92.0、ODInW13 50.8。

Qwen3.5-27BClaude-Sonnet-4.5Gemma4-31BGemma4-26BA4BQwen3.5-35B-A3BQwen3.6-35B-A3B
STEM and Puzzle
MMMU82.379.680.478.481.481.7
MMMU-Pro75.068.476.9*73.8*75.175.3
Mathvista(mini)87.879.879.379.486.286.4
ZEROBench_sub36.226.326.026.334.134.4
General VQA
RealWorldQA83.770.372.372.284.185.3
MMBench EN-DEV-v1.192.688.390.989.091.592.8
SimpleVQA56.057.652.952.258.358.9
HallusionBench70.059.967.466.167.969.8
Text Recognition and Document Understanding
OmniDocBench1.588.985.880.174.489.389.9
CharXiv(RQ)79.567.267.969.077.578.0
CC-OCR81.068.175.774.580.781.9
AI2D_TEST92.987.089.088.392.692.7
Spatial Intelligence
RefCOCO(avg)90.9------89.292.0
ODInW1341.1------42.650.8
EmbSpatialBench84.571.8----83.184.3
RefSpatialBench67.7------63.564.3
Video Understanding
VideoMME (w sub.)87.081.1----86.686.6
VideoMME (w/o sub.)82.875.3----82.582.5
VideoMMMU82.377.681.676.080.483.7
MLVU85.972.8----85.686.2
MVBench74.6------74.874.6
LVBench73.6------71.471.4
  • Empty cells (--) indicate scores not available or not applicable.

开始使用 Qwen3.6-35B-A3B

Qwen3.6-35B-A3B 即将登陆阿里云 Model Studio。我们正在全力筹备中,请耐心等待。

Qwen3.6-35B-A3B 的开源权重已在 Hugging FaceModelScope 上提供,支持本地部署;也可通过 阿里云百炼 API 以 qwen3.6-flash 的名称调用。此外,您还可以在 Qwen Studio 上即时体验。

该模型可以无缝集成到流行的第三方编程助手中,包括 OpenClaw、Claude Code 和 Qwen Code,从而简化开发流程,实现高效且具备上下文感知能力的编码体验。

API 使用方式

本次发布支持 preserve_thinking 功能:在消息中保留所有前序轮次的思维内容,推荐用于智能体任务

阿里云百炼

阿里云百炼支持行业标准协议,包括兼容 OpenAI 规范的聊天补全(chat completions)和响应(responses)API,以及兼容 Anthropic 的 API 接口。

以下是聊天补全 API 的代码示例:

"""
Environment variables (per official docs):
  DASHSCOPE_API_KEY: Your API Key from https://bailian.console.aliyun.com/
  DASHSCOPE_BASE_URL: (optional) Base URL for compatible-mode API.
    - Beijing: https://dashscope.aliyuncs.com/compatible-mode/v1
    - Singapore: https://dashscope-intl.aliyuncs.com/compatible-mode/v1
    - US (Virginia): https://dashscope-us.aliyuncs.com/compatible-mode/v1
  DASHSCOPE_MODEL: (optional) Model name; override for different models.
"""
from openai import OpenAI
import os

api_key = os.environ.get("DASHSCOPE_API_KEY")
if not api_key:
    raise ValueError(
        "DASHSCOPE_API_KEY is required. "
        "Set it via: export DASHSCOPE_API_KEY='your-api-key'"
    )

client = OpenAI(
    api_key=api_key,
    base_url=os.environ.get(
        "DASHSCOPE_BASE_URL",
        "https://dashscope.aliyuncs.com/compatible-mode/v1",
    ),
)

messages = [{"role": "user", "content": "Introduce vibe coding."}]

model = os.environ.get(
    "DASHSCOPE_MODEL",
    "qwen3.6-flash",
)
completion = client.chat.completions.create(
    model=model,
    messages=messages,
    extra_body={
        "enable_thinking": True,
        # "preserve_thinking": True,
    },
    stream=True
)

reasoning_content = ""  # Full reasoning trace
answer_content = ""  # Full response
is_answering = False  # Whether we have entered the answer phase
print("\n" + "=" * 20 + "Reasoning" + "=" * 20 + "\n")

for chunk in completion:
    if not chunk.choices:
        print("\nUsage:")
        print(chunk.usage)
        continue

    delta = chunk.choices[0].delta

    # Collect reasoning content only
    if hasattr(delta, "reasoning_content") and delta.reasoning_content is not None:
        if not is_answering:
            print(delta.reasoning_content, end="", flush=True)
        reasoning_content += delta.reasoning_content

    # Received content, start answer phase
    if hasattr(delta, "content") and delta.content:
        if not is_answering:
            print("\n" + "=" * 20 + "Answer" + "=" * 20 + "\n")
            is_answering = True
        print(delta.content, end="", flush=True)
        answer_content += delta.content

更多信息请访问 API 文档

代码及智能体

Qwen3.6-35B-A3B 具备出色的智能体编程能力,可以无缝集成到流行的第三方编程助手中,包括 OpenClaw、Claude Code 和 Qwen Code。

OpenClaw

Qwen3.6-35B-A3B 兼容 OpenClaw(原名 Moltbot / Clawdbot),这是一款可自托管的开源 AI 编码智能体。将其连接至 百炼,即可在终端中获得完整的智能体编码体验。请使用以下脚本开始:

# Node.js 22+
curl -fsSL https://molt.bot/install.sh | bash   # macOS / Linux

# Set your API key
export DASHSCOPE_API_KEY=<your_api_key>

# Launch OpenClaw
openclaw dashboard # web browser
# openclaw tui # Open a new terminal and start the TUI

首次使用时,请编辑 ~/.openclaw/openclaw.json 文件,将 OpenClaw 指向百炼。 找到或创建以下字段并合并它们——切勿覆盖整个文件,以保留您现有的设置:

{
  "models": {
    "mode": "merge",
    "providers": {
      "bailian": {
        "baseUrl": "https://dashscope.aliyuncs.com/compatible-mode/v1",
        "apiKey": "DASHSCOPE_API_KEY",
        "api": "openai-completions",
        "models": [{
          "id": "qwen3.6-flash",
          "name": "qwen3.6-flash",
          "reasoning": true,
          "input": ["text", "image"],
          "contextWindow": 131072,
          "maxTokens": 16384
        }]
      }
    }
  },
  "agents": {
    "defaults": {
      "model": {
        "primary": "bailian/qwen3.6-flash"
      },
      "models": {
        "bailian/qwen3.6-flash": {}
      }
    }
  }
}

Qwen Code

Qwen3.6-35B-A3B 适配 Qwen Code,这是一款专为终端设计的开源 AI 智能体,针对 Qwen 系列进行了深度优化。请使用以下脚本开始:

# Node.js 20+
npm install -g @qwen-code/qwen-code@latest

# Start Qwen Code (interactive)
qwen

# Then, in the session:
/help
/auth

首次使用时,系统会提示您登录。您可以随时运行 /auth 来切换认证方式。

Claude Code

Qwen API 也支持 Anthropic API 协议,这意味着您可以将其与 Claude Code 等工具配合使用,以获得更优质的编码体验:

# Install Claude Code
npm install -g @anthropic-ai/claude-code

# Configure environment
export ANTHROPIC_MODEL="qwen3.6-flash"
export ANTHROPIC_SMALL_FAST_MODEL="qwen3.6-flash"
export ANTHROPIC_BASE_URL=https://dashscope.aliyuncs.com/apps/anthropic
export ANTHROPIC_AUTH_TOKEN=<your_api_key>

# Launch the CLI
claude

总结

Qwen3.6-35B-A3B 表明,稀疏 MoE 模型可以实现卓越的智能体编程和推理能力。仅凭30亿激活参数,它便能够交付与数倍于其激活规模的稠密模型相当的性能,同时在多模态基准上同样表现出色。作为完全开源的模型权重,它为该规模下的模型能力树立了新的标杆。

展望未来,我们将继续扩展 Qwen3.6 开源家族,并不断拓展高效开源模型所能实现的边界。我们由衷感谢社区的宝贵反馈,并期待看到大家利用 Qwen3.6-35B-A3B 创造出的精彩成果。Qwen3.6 开源家族正在持续壮大,敬请关注我们的后续发布!

评论

(0)
未配置登录方式
暂无评论