AI/EXPLORER
ToolsCategoriesSitesLLMsCompareAI QuizAlternativesPremium
—AI Tools
—Sites & Blogs
—LLMs & Models
—Categories
AI Explorer

Find and compare the best artificial intelligence tools for your projects.

Made within France

Explore

  • ›All tools
  • ›Sites & Blogs
  • ›LLMs & Models
  • ›Compare
  • ›Chatbots
  • ›AI Images
  • ›Code & Dev

Company

  • ›Premium
  • ›About
  • ›Contact
  • ›Blog

Legal

  • ›Legal notice
  • ›Privacy
  • ›Terms

© 2026 AI Explorer·All rights reserved.

HomeLLMsJan v3 4B base instruct gguf

Jan v3 4B base instruct gguf

by janhq

Open source · 111k downloads · 52 likes

2.2
(52 reviews)ChatAPI & Local
About

Jan v3 4B base instruct is a compact language model with 4 billion parameters, designed as an optimized base for fine-tuning and lightweight programming assistance. Derived from post-training distillation from a larger teacher model, it maintains strong general performance while offering improved instruction-following capabilities right out of the box. Thanks to its reduced size and efficient architecture, it stands out as an ideal starting point for custom adaptations or tasks requiring code assistance. The model particularly excels in environments where lightness and responsiveness are critical, while remaining versatile for a wide range of applications. It is accessible via the Jan application, which simplifies local deployment or online use.

Documentation

Jan-v3-4B-base-instruct: a 4B baseline model for fine-tuning

GitHub License Jan App

image

Overview

Jan-v3-4B-base-instruct is a 4B-parameter model obtained via post-training distillation from a larger teacher, transferring capabilities while preserving general-purpose performance on standard benchmarks. The result is a compact, ownable base that is straightforward to fine-tune, broadly applicable and minimizing the usual capacity–capability trade-offs.

Building on this base, Jan-Code, a code-tuned variant, will be released soon.

Model Overview

This repo contains the BF16 version of Jan-v3-4B-base-instruct, which has the following features:

  • Type: Causal Language Models
  • Training Stage: Pretraining & Post-training
  • Number of Parameters: 4B in total
  • Number of Layers: 36
  • Number of Attention Heads (GQA): 32 for Q and 8 for KV
  • Context Length: 262,144 natively.

Intended Use

  • A better small base for downstream work: improved instruction following out of the box, strong starting point for fine-tuning, and effective lightweight coding assistance.

Performance

image

Quick Start

Integration with Jan Apps

Jan-v3 demo is hosted on Jan Browser at chat.jan.ai. It is also optimized for direct integration with Jan Desktop, select the model in the app to start using it.

Local Deployment

Using vLLM:

Bash
vllm serve janhq/Jan-v3-4B-base-instruct \
    --host 0.0.0.0 \
    --port 1234 \
    --enable-auto-tool-choice \
    --tool-call-parser hermes 
    

Using llama.cpp:

Bash
llama-server --model Jan-v3-4B-base-instruct-Q8_0.gguf \
    --host 0.0.0.0 \
    --port 1234 \
    --jinja \
    --no-context-shift

Recommended Parameters

For optimal performance in agentic and general tasks, we recommend the following inference parameters:

YAML
temperature: 0.7
top_p: 0.8
top_k: 20

🤝 Community & Support

  • Discussions: Hugging Face Community
  • Jan App: Learn more about the Jan App at jan.ai

📄 Citation

Bibtex
Updated Soon
Capabilities & Tags
transformersggufcodetext-generationenendpoints_compatibleconversational
Links & Resources
Specifications
CategoryChat
AccessAPI & Local
LicenseOpen Source
PricingOpen Source
Parameters4B parameters
Rating
2.2

Try Jan v3 4B base instruct gguf

Access the model directly