Bob-the-Koala commited on
Commit
8e04052
·
verified ·
1 Parent(s): be0f84a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +141 -6
README.md CHANGED
@@ -8,14 +8,149 @@ tags:
8
  license: apache-2.0
9
  language:
10
  - en
 
 
 
 
 
 
 
11
  ---
 
12
 
13
- # Uploaded finetuned model
14
 
15
- - **Developed by:** Bob-the-Koala
16
- - **License:** apache-2.0
17
- - **Finetuned from model :** unsloth/Qwen3-8B-Base-unsloth-bnb-4bit
18
 
19
- This qwen3 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
 
21
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
 
8
  license: apache-2.0
9
  language:
10
  - en
11
+ datasets:
12
+ - TeichAI/MiniMax-M2.1-Code-SFT
13
+ - TeichAI/MiniMax-M2.1-8800x
14
+ - TeichAI/convo-v1
15
+ - AlicanKiraz0/Agentic-Chain-of-Thought-Coding-SFT-Dataset-v1.1
16
+ - TeichAI/claude-4.5-opus-high-reasoning-250x
17
+ pipeline_tag: text-generation
18
  ---
19
+ # LocalCodeViber
20
 
21
+ **LocalCodeViber** is a local-first agentic coding model built on [Qwen3-8B](https://huggingface.co/Qwen/Qwen3-8B), fine-tuned for tool-calling, multi-step code generation, and autonomous error recovery. Designed to run entirely on consumer hardware — no API, no cloud, no cost per token.
22
 
23
+ This is the SFT foundation model. Reinforcement learning is ongoing.
 
 
24
 
25
+ ---
26
+
27
+ ## What it does
28
+
29
+ LocalCodeViber was trained to operate as a coding agent — not just generate code, but use tools to read files, write files, run commands, search the web, and recover from failures just like a real developer would.
30
+
31
+ It can:
32
+
33
+ - Read and edit files in a workspace
34
+ - Write complete, working code from a single prompt
35
+ - Execute shell commands and interpret the output
36
+ - Recover from failed tool calls without giving up
37
+ - Create pull requests on GitHub repositories
38
+ - Think through problems step by step using native `<think>` tags before acting
39
+
40
+ ---
41
+
42
+ ## Model Details
43
+
44
+ | | |
45
+ |---|---|
46
+ | **Base Model** | Qwen3-8B-Base |
47
+ | **Architecture** | Qwen3 transformer, 36 layers |
48
+
49
+ ## Training Data
50
+
51
+ LocalCodeViber was trained on a curated mix of 14,837 examples across 5 datasets:
52
+
53
+ | Dataset | Examples | Focus |
54
+ |---|---|---|
55
+ | [TeichAI/convo-v1](https://huggingface.co/datasets/TeichAI/convo-v1) | 777 | Conversational format, instruction following |
56
+ | [AlicanKiraz0/Agentic-Chain-of-Thought-Coding-SFT-Dataset-v1.1](https://huggingface.co/datasets/AlicanKiraz0/Agentic-Chain-of-Thought-Coding-SFT-Dataset-v1.1) | ~3,700 | Agentic reasoning and tool use |
57
+ | [TeichAI/MiniMax-M2.1-Code-SFT](https://huggingface.co/datasets/TeichAI/MiniMax-M2.1-Code-SFT) | ~1,300 | Agentic Code generation |
58
+ | [TeichAI/MiniMax-M2.1-8800x](https://huggingface.co/datasets/TeichAI/MiniMax-M2.1-8800x) | 8,800 | Diverse coding tasks |
59
+ | [TeichAI/claude-4.5-opus-high-reasoning-250x](https://huggingface.co/datasets/TeichAI/claude-4.5-opus-high-reasoning-250x) | 250 | High-quality reasoning traces |
60
+
61
+ The dataset mix emphasises real agentic tool-use patterns including failed tool calls that are identified, diagnosed, and corrected — giving the model genuine error recovery capability rather than just pattern matching on success cases.
62
+
63
+ ---
64
+
65
+ ## Tools
66
+
67
+ LocalCodeViber understands the following tool schema out of the box:
68
+
69
+ ```json
70
+ ["read_file", "write_file", "edit_file", "list_directory", "search_code", "run_command", "web_search"]
71
+ ```
72
+
73
+ These match the tools in the training data. Pass them via the standard OpenAI tool calling API.
74
+
75
+ ---
76
+
77
+ ## Usage
78
+
79
+ ### LM Studio (Recommended)
80
+
81
+ 1. Download the GGUF version: [Bob-the-Koala/LocalCodeViber-GGUF](https://huggingface.co/Bob-the-Koala/LocalCodeViber-GGUF)
82
+ 2. Load in LM Studio and break free from API costs!
83
+
84
+ ### Ollama
85
+
86
+ ```bash
87
+ ollama run hf.co/Bob-the-Koala/LocalCodeViber-GGUF:Q4_K_M
88
+ ```
89
+
90
+ ### Transformers
91
+
92
+ ```python
93
+ from transformers import AutoModelForCausalLM, AutoTokenizer
94
+
95
+ model = AutoModelForCausalLM.from_pretrained(
96
+ "Bob-the-Koala/LocalCodeViber",
97
+ torch_dtype="auto",
98
+ device_map="auto"
99
+ )
100
+ tokenizer = AutoTokenizer.from_pretrained("Bob-the-Koala/LocalCodeViber")
101
+ ```
102
+
103
+ ---
104
+
105
+ ## GGUF Versions
106
+
107
+ Available in [Bob-the-Koala/LocalCodeViber-GGUF](https://huggingface.co/Bob-the-Koala/LocalCodeViber-GGUF):
108
+
109
+ | Quantization | Size | Use case |
110
+ |---|---|---|
111
+ | `Q4_K_M` | ~4.8 GB | Everyday use, best balance |
112
+
113
+ ---
114
+
115
+ ## System Prompt
116
+
117
+ For best results, use this system prompt:
118
+
119
+ ```
120
+ You are a helpful coding assistant with access to file operations and code analysis tools.
121
+ Complete the user's task thoroughly and efficiently.
122
+ When given a coding task, create working code files in the workspace.
123
+ ```
124
+
125
+ ---
126
+
127
+ ## Limitations
128
+
129
+ - Base model started from bnb-4bit weights — quality ceiling is below a full precision 8B model
130
+ - SFT only — reinforcement learning is in progress and will significantly improve reasoning quality
131
+ - Not suitable for tasks requiring knowledge past Qwen3's training cutoff
132
+
133
+ ---
134
+
135
+ ## Roadmap
136
+
137
+ - [ ] **LocalCodeViber-RL** — reinforcement learning on top of this SFT base, optimising for code correctness and task completion
138
+ - [ ] **LocalCodeViber-Claw** — fine-tuned specifically for [OpenClaw](https://github.com/openclaw/openclaw) skill schemas, channel routing, extra safety, and memory system
139
+ - [ ] **LocalCodeViber-14B** — same training recipe on Qwen3-14B for substantially higher capability
140
+
141
+ ---
142
+
143
+ ## Acknowledgements
144
+
145
+ LocalCodeViber was trained using [Unsloth](https://github.com/unslothai/unsloth) and would not exist without the datasets provided by [TeichAI](https://huggingface.co/TeichAI) and [AlicanKiraz0](https://huggingface.co/AlicanKiraz0).
146
+
147
+ ---
148
+
149
+ ## License
150
+
151
+ This model is released under the Apache 2.0 license
152
+
153
+ ---
154
 
155
+ *Built by [Bob-the-Koala](https://huggingface.co/Bob-the-Koala)*
156
+ [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)