docs/src/pages/docs/desktop/jan-models/jan-code-4b.mdx
import { Callout } from 'nextra/components'
Jan-Code-4B is a lightweight, code-tuned 4B parameter model built for fast local inference. Fine-tuned on Jan-v3-4B-base-instruct, it is designed for practical coding tasks with an emphasis on handling well-scoped subtasks reliably while keeping latency and compute requirements low.
<a href="jan://models/huggingface/janhq/Jan-code-4b-gguf" style={{ display: 'inline-flex', alignItems: 'center', gap: '8px', padding: '10px 20px', backgroundColor: '#000', color: '#fff', borderRadius: '8px', fontWeight: '600', fontSize: '15px', textDecoration: 'none', marginTop: '8px', }}
Open in Jan </a>
| Property | Value |
|---|---|
| Parameters | 4B |
| Base Model | Jan-v3-4B-base-instruct (Qwen3-4B-Instruct-2507) |
| Fine-tuning focus | Code generation, editing, refactoring, debugging |
| License | Apache 2.0 |
Jan-Code-4B leads Jan-v3-base-instruct and the Qwen3-4B-Instruct-2507 base model across all three coding and reasoning benchmarks:
Using vLLM:
vllm serve janhq/Jan-code-4b \
--host 0.0.0.0 \
--port 1234 \
--enable-auto-tool-choice \
--tool-call-parser hermes
Using llama.cpp:
llama-server --model Jan-code-4b-Q8_0.gguf \
--host 0.0.0.0 \
--port 1234 \
--jinja \
--no-context-shift
temperature: 0.7
top_p: 0.8
top_k: 20