ereniko commited on
Commit
4cf7d78
·
verified ·
1 Parent(s): 29d3597

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +165 -3
README.md CHANGED
@@ -1,3 +1,165 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ language:
4
+ - en
5
+ tags:
6
+ - linux
7
+ - bash
8
+ - terminal
9
+ - experimental
10
+ - from-scratch
11
+ pipeline_tag: text-generation
12
+ ---
13
+
14
+ # LaaLM-v2
15
+
16
+ **A small language model trained from scratch to be a Linux terminal. That's it.**
17
+
18
+ ---
19
+
20
+ ## What is this?
21
+
22
+ LaaLM-v2 is a ~100M parameter transformer trained from random weights to simulate bash commands. No pretrained base model, no finetuning - just straight training on terminal data.
23
+
24
+ Why? Because LaaLM-exp-v1 used a 3B pretrained model and knew way too much stuff it didn't need (like `sudo`, `apt`, networking). That's 30x more parameters than necessary. This version is built to only know what it needs to know.
25
+
26
+ **Current status:** In training. Things might change.
27
+
28
+ ---
29
+
30
+ ## What it can do
31
+
32
+ Basic file stuff:
33
+ - `pwd`, `ls`, `cd` (including `cd ..`)
34
+ - `touch`, `cat`, `echo`
35
+ - `echo text > file`, `echo text >> file`
36
+ - `mv`, `cp`, `rm` (with `-r` and `-rf`)
37
+ - `mkdir` (with `-p`)
38
+
39
+ Text processing:
40
+ - `grep pattern file`
41
+ - `head`, `tail`, `wc`
42
+ - `find`
43
+
44
+ Pipes work too:
45
+ - `cat file | grep something`
46
+ - `cat file | head -5`
47
+ - `cat file | wc`
48
+
49
+ And it shows its reasoning for complex commands:
50
+ ```bash
51
+ $ cat log.txt | grep error
52
+ <REASON>
53
+ STEP1: execute(cat log.txt) -> output="..."
54
+ STEP2: pipe_to(grep error)
55
+ -> filtered=3 lines
56
+ </REASON>
57
+ error: connection failed
58
+ error: timeout
59
+ error: invalid input
60
+ ```
61
+ These are currently what the model data looks like. The actual model may have different styles, but don't really expect a major difference.
62
+
63
+ ---
64
+
65
+ ## Training
66
+
67
+ Training of LaaLM-v2 is a bit rough.
68
+
69
+ For the current status as of 19 February 2026, LaaLM had 3 training rounds with about 17.5k conversation examples.
70
+
71
+ But the model suffers from extreme hallucination, which we are trying to solve.
72
+
73
+ And experimentation ate a lot of budget from us, so it caused slowdowns.
74
+
75
+ For the current training hardware, originally we thought of using the Nvidia RTX PRO 6000, but because of software errors, we used an MI300X. Though we have plans of using TPUs also.
76
+
77
+ Though our new training code shows good improvements, we hope for good results after the budget is solved.
78
+
79
+ But the current versions like we said are pretty rough for now.
80
+
81
+ Our latest tests resulted in this:
82
+
83
+ $ ls
84
+ touch log_4
85
+ cat log_4 | grep data
86
+ STEP1: execute(cat log_4) -> output=""
87
+ STEP2: pipe_to(grep data, input="")
88
+ -> filtered=0 lines
89
+ cat log_4 | grep data
90
+ STEP1: execute(cat log_4) -> output=""
91
+ STEP2: pipe_to(grep data, input="")
92
+ -> filtered=0 lines
93
+ cat log_4 | head
94
+
95
+ This is from LaaLM-v2.
96
+
97
+ ---
98
+
99
+ ## How to use it
100
+
101
+ *(Code examples coming soon - model is still training and the API might change)*
102
+
103
+ Basic idea:
104
+ 1. Load the model
105
+ 2. Start with a system message about the initial state
106
+ 3. Send commands as user messages
107
+ 4. Get outputs as assistant messages
108
+ 5. Model remembers what you did (files you created, directories you made, etc.)
109
+
110
+ Example conversation:
111
+ ```bash
112
+ $ touch myfile.txt
113
+ (empty output)
114
+
115
+ $ echo hello > myfile.txt
116
+ (empty output)
117
+
118
+ $ cat myfile.txt
119
+ hello
120
+
121
+ $ ls
122
+ myfile.txt
123
+
124
+ $ mv myfile.txt renamed.txt
125
+ (empty output)
126
+
127
+ $ cat renamed.txt
128
+ hello
129
+ ```
130
+
131
+ ---
132
+
133
+ ## Why train from scratch?
134
+
135
+ **LaaLM-v1** was a T5 model (encoder-decoder). Worked okay but couldn't really do conversations.
136
+
137
+ **LaaLM-exp-v1** was Qwen 3B finetuned. It worked great but it was 3B parameters! It knew about sudo, apt, networking, all this stuff it didn't need. Felt wasteful.
138
+
139
+ **LaaLM-v2** is trained from scratch with just what it needs. ~100M parameters, no bloat. Every parameter is there for bash commands, nothing else.
140
+
141
+ Yeah it's harder to train from scratch (no pretrained stopping behavior, needs way more data), but the purity is worth it.
142
+
143
+ ---
144
+
145
+ ## What it doesn't do
146
+
147
+ This isn't a real terminal. It simulates bash behavior but doesn't actually run commands.
148
+
149
+ No networking (ssh, wget, curl), no package managers (apt, pip), no system admin stuff (sudo, systemctl), no scripting (loops, variables), no permissions (chmod, users).
150
+
151
+ We thought of basic bash but we don't know if we can handle it.
152
+
153
+ ---
154
+
155
+ ## Status
156
+
157
+ Currently training. If you're seeing this, the model might not be uploaded yet or might change significantly.
158
+
159
+ Check back later or watch the repo for updates.
160
+
161
+ ---
162
+
163
+ ## License
164
+
165
+ MIT