LH-Tech-AI commited on
Commit
c5e469a
·
verified ·
1 Parent(s): 8e195fa

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +35 -3
README.md CHANGED
@@ -1,3 +1,35 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - bigcode/the-stack-smol
5
+ - ttbui/html_alpaca
6
+ language:
7
+ - en
8
+ tags:
9
+ - code
10
+ - coding
11
+ - small
12
+ - tiny
13
+ ---
14
+
15
+ # Welcome to htmLLM v2 124M!
16
+
17
+ With this LLM, we wanted to see, how well tiny LLMs with just 124 million parameters can perform on coding tasks.
18
+
19
+ This model is also a bit finetuned using html_alpaca directly in the pretraining.
20
+
21
+ If you want to try it, you can use htmllm.ipynb in the HF model files and download the model weight from this HF model.
22
+
23
+ # Code
24
+ All code can be accessed via the file **htmllm_v2_124m.ipynb** in this HF model.
25
+
26
+ # Weights
27
+ The final **base** model checkpoint can be downloaded here in the files list as **ckpt.pt**. It will be available soon!
28
+
29
+ # Training
30
+ We trained our model on a single Kaggle T4 GPU.
31
+
32
+ # Thanks to:
33
+ - Andrej Karpathy and his nanoGPT code
34
+ - Kaggle for the free GPU hours for training on the T4
35
+ - You all for your support on my reddit.