AI/EXPLORER
ToolsCategoriesSitesLLMsCompareAI QuizAlternativesPremium
—AI Tools
—Sites & Blogs
—LLMs & Models
—Categories
AI Explorer

Find and compare the best artificial intelligence tools for your projects.

Made within France

Explore

  • ›All tools
  • ›Sites & Blogs
  • ›LLMs & Models
  • ›Compare
  • ›Chatbots
  • ›AI Images
  • ›Code & Dev

Company

  • ›Premium
  • ›About
  • ›Contact
  • ›Blog

Legal

  • ›Legal notice
  • ›Privacy
  • ›Terms

© 2026 AI Explorer·All rights reserved.

HomeLLMsQwen3 Coder Next AWQ 4bit

Qwen3 Coder Next AWQ 4bit

by bullpoint

Open source · 82k downloads · 24 likes

1.7
(24 reviews)CodeAPI & Local
About

The Qwen3 Coder Next AWQ 4bit model is an optimized and quantized (4-bit) version of the Qwen3 Coder Next model, specifically designed for coding agents and local development. With only 3 billion active parameters out of a total of 80 billion, it delivers performance comparable to much larger models while remaining highly resource-efficient. It stands out for its advanced long-term reasoning capabilities, complex tool usage, and resilience to execution failures, making it ideal for dynamic and autonomous programming tasks. With a context length of 256,000 tokens and flexible integration with various development environments (IDE, CLI), it easily adapts to existing workflows like Claude Code or Qwen Code. Its hybrid approach combining managed attention and multiple experts (MoE) allows it to handle diverse projects while minimizing memory usage, making it a strong choice for developers seeking performance and autonomy without hardware overhead.

Documentation

Qwen3-Coder-Next-AWQ-4bit

Model Size: 80B total parameters, 3B activated | Quantization: 4-bit AWQ | VRAM: ~45GB

Highlights

This is a 4-bit AWQ quantized version of Qwen3-Coder-Next, an open-weight language model designed specifically for coding agents and local development. It features the following key enhancements:

  • Super Efficient with Significant Performance: With only 3B activated parameters (80B total parameters), it achieves performance comparable to models with 10–20x more active parameters, making it highly cost-effective for agent deployment.
  • Advanced Agentic Capabilities: Through an elaborate training recipe, it excels at long-horizon reasoning, complex tool usage, and recovery from execution failures, ensuring robust performance in dynamic coding tasks.
  • Versatile Integration with Real-World IDE: Its 256k context length, combined with adaptability to various scaffold templates, enables seamless integration with different CLI/IDE platforms (e.g., Claude Code, Qwen Code, Qoder, Kilo, Trae, Cline, etc.), supporting diverse development environments.

Quantization Details

This model was quantized using llm-compressor with AWQ (Activation-aware Weight Quantization).

PropertyValue
Base ModelQwen/Qwen3-Coder-Next
Quantization MethodAWQ
Quantization Toolllm-compressor
Calibration Datasetnvidia/Llama-Nemotron-Post-Training-Dataset
Bits4
Group Size32
SymmetricYes
StrategyGroup
ObserverMSE
Formatpack-quantized
Quant Methodcompressed-tensors

Memory Usage

TypeSize
Original (BF16)~151 GB
Quantized (4-bit)~45 GB

Selective Quantization

To preserve model quality, the following components are kept at higher precision:

  • Embedding layers (model.embed_tokens)
  • LM head (lm_head)
  • All normalization layers (*norm*, *RMSNorm*, *input_layernorm, *post_attention_layernorm)
  • Gated Attention projections (self_attn.q_proj, self_attn.k_proj, self_attn.v_proj, self_attn.o_proj)
  • Gated DeltaNet components (linear_attn.in_proj_qkvz, linear_attn.in_proj_ba, linear_attn.out_proj, linear_attn.norm, linear_attn.conv1d, linear_attn.A_log, linear_attn.dt_bias)
  • MoE routing gates (mlp.gate, shared_expert_gate)
  • Shared expert layers (shared_expert.gate_proj, shared_expert.up_proj, shared_expert.down_proj)
  • MTP layers (mtp.*)

AWQ Smoothing Mappings

The quantization uses duo scaling with the following activation smoothing:

  • post_attention_layernorm → mlp.experts.*.gate_proj, mlp.experts.*.up_proj
  • mlp.experts.*.up_proj → mlp.experts.*.down_proj

Model Overview

Qwen3-Coder-Next has the following features:

  • Type: Causal Language Models
  • Training Stage: Pretraining & Post-training
  • Number of Parameters: 80B in total and 3B activated
  • Number of Parameters (Non-Embedding): 79B
  • Hidden Dimension: 2048
  • Number of Layers: 48
    • Hybrid Layout: 12 × (3 × (Gated DeltaNet → MoE) → 1 × (Gated Attention → MoE))
  • Gated Attention:
    • Number of Attention Heads: 16 for Q and 2 for KV
    • Head Dimension: 256
    • Rotary Position Embedding Dimension: 64
  • Gated DeltaNet:
    • Number of Linear Attention Heads: 32 for V and 16 for QK
    • Head Dimension: 128
  • Mixture of Experts:
    • Number of Experts: 512
    • Number of Activated Experts: 10
    • Number of Shared Experts: 1
    • Expert Intermediate Dimension: 512
  • Context Length: 262,144 natively

NOTE: This model supports only non-thinking mode and does not generate <think></think> blocks in its output. Meanwhile, specifying enable_thinking=False is no longer required.

For more details, including benchmark evaluation, hardware requirements, and inference performance, please refer to the blog, GitHub, and Documentation.

Quickstart

We advise you to use the latest version of transformers.

The following contains a code snippet illustrating how to use the model to generate content based on given inputs.

Python
from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "bullpoint/Qwen3-Coder-Next-AWQ-4bit"

# load the tokenizer and the model
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(
    model_name,
    torch_dtype="auto",
    device_map="auto"
)

# prepare the model input
prompt = "Write a quick sort algorithm."
messages = [
    {"role": "user", "content": prompt}
]
text = tokenizer.apply_chat_template(
    messages,
    tokenize=False,
    add_generation_prompt=True,
)
model_inputs = tokenizer([text], return_tensors="pt").to(model.device)

# conduct text completion
generated_ids = model.generate(
    **model_inputs,
    max_new_tokens=65536
)
output_ids = generated_ids[0][len(model_inputs.input_ids[0]):].tolist()

content = tokenizer.decode(output_ids, skip_special_tokens=True)

print("content:", content)

Note: If you encounter out-of-memory (OOM) issues, consider reducing the context length to a shorter value, such as 32,768.

For local use, applications such as Ollama, LMStudio, MLX-LM, llama.cpp, and KTransformers have also supported Qwen3.

Deployment

For deployment, you can use the latest sglang or vllm to create an OpenAI-compatible API endpoint.

SGLang

SGLang is a fast serving framework for large language models and vision language models. SGLang could be used to launch a server with OpenAI-compatible API service.

sglang>=v0.5.8 is required for Qwen3-Coder-Next, which can be installed using:

Shell
pip install 'sglang[all]>=v0.5.8'

See its documentation for more details.

The following command can be used to create an API endpoint at http://localhost:30000/v1 with maximum context length 256K tokens using tensor parallel on 2 GPUs.

Shell
python -m sglang.launch_server --model bullpoint/Qwen3-Coder-Next-AWQ-4bit --port 30000 --tp-size 2 --tool-call-parser qwen3_coder

[!Note] The default context length is 256K. Consider reducing the context length to a smaller value, e.g., 32768, if the server fails to start.

vLLM

vLLM is a high-throughput and memory-efficient inference and serving engine for LLMs. vLLM could be used to launch a server with OpenAI-compatible API service.

vllm>=0.15.0 is required for Qwen3-Coder-Next, which can be installed using:

Shell
pip install 'vllm>=0.15.0'

See its documentation for more details.

The following command can be used to create an API endpoint at http://localhost:8000/v1 with maximum context length 256K tokens using tensor parallel on 2 GPUs.

Shell
vllm serve bullpoint/Qwen3-Coder-Next-AWQ-4bit --port 8000 --tensor-parallel-size 2 --enable-auto-tool-choice --tool-call-parser qwen3_coder

[!Note] The default context length is 256K. Consider reducing the context length to a smaller value, e.g., 32768, if the server fails to start.

Agentic Coding

Qwen3-Coder-Next excels in tool calling capabilities.

You can simply define or use any tools as following example.

Python
# Your tool implementation
def square_the_number(num: float) -> dict:
    return num ** 2

# Define Tools
tools = [
    {
        "type": "function",
        "function": {
            "name": "square_the_number",
            "description": "output the square of the number.",
            "parameters": {
                "type": "object",
                "required": ["input_num"],
                "properties": {
                    "input_num": {
                        "type": "number",
                        "description": "input_num is a number that will be squared"
                    }
                },
            }
        }
    }
]

from openai import OpenAI
# Define LLM
client = OpenAI(
    # Use a custom endpoint compatible with OpenAI API
    base_url="http://localhost:8000/v1",  # api_base
    api_key="EMPTY"
)

messages = [{"role": "user", "content": "square the number 1024"}]

completion = client.chat.completions.create(
    messages=messages,
    model="Qwen3-Coder-Next-AWQ-4bit",
    max_tokens=65536,
    tools=tools,
)

print(completion.choices[0])

Best Practices

To achieve optimal performance, we recommend the following sampling parameters: temperature=1.0, top_p=0.95, top_k=40.

Citation

If you find our work helpful, feel free to give us a cite.

Bibtex
@techreport{qwen_qwen3_coder_next_tech_report,
  title        = {Qwen3-Coder-Next Technical Report},
  author       = {{Qwen Team}},
  url          = {https://github.com/QwenLM/Qwen3-Coder/blob/main/qwen3_coder_next_tech_report.pdf},
  note         = {Accessed: 2025}
}
Capabilities & Tags
transformerssafetensorsqwen3_nexttext-generationqwenqwen3qwen3-coderqwen3-nextmoemixture-of-experts
Links & Resources
Specifications
CategoryCode
AccessAPI & Local
LicenseOpen Source
PricingOpen Source
Parameters4B parameters
Rating
1.7

Try Qwen3 Coder Next AWQ 4bit

Access the model directly