Jan-code-4b / README.md
jan-ai's picture
Update README.md
1e8f95d verified
|
raw
history blame
2.48 kB
metadata
license: apache-2.0
language:
  - en
base_model:
  - janhq/Jan-v3-4B-base-instruct
pipeline_tag: text-generation
library_name: transformers
tags:
  - agent

Jan-Code-4B: a small code-tuned model

GitHub License Jan App

image

Overview

Jan-Code-4B is a code-tuned model built on top of Jan-v3-4B-base-instruct. It’s designed to be a practical coding model you can run locally and iterate on quickly—useful for everyday code tasks and as a lightweight “worker” model in agentic workflows.

Compared to larger coding models, Jan-Code focuses on handling well-scoped subtasks reliably while keeping latency and compute requirements small.

Intended Use

  • Lightweight coding assistant for generation, editing, refactoring, and debugging
  • A small, fast worker model for agent setups (e.g., as a sub-agent that produces patches/tests while a larger model plans)

Performance

image

Quick Start

Integration with Jan Apps

Jan-v3 demo is hosted on Jan Browser at chat.jan.ai. It is also optimized for direct integration with Jan Desktop, select the model in the app to start using it.

Local Deployment

Using vLLM:

vllm serve janhq/Jan-code \
    --host 0.0.0.0 \
    --port 1234 \
    --enable-auto-tool-choice \
    --tool-call-parser hermes 
    

Using llama.cpp:

llama-server --model Jan-code-Q8_0.gguf \
    --host 0.0.0.0 \
    --port 1234 \
    --jinja \
    --no-context-shift

Recommended Parameters

For optimal performance in agentic and general tasks, we recommend the following inference parameters:

temperature: 0.7
top_p: 0.8
top_k: 20

🤝 Community & Support

📄 Citation

Updated Soon