File size: 885 Bytes
c5e469a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
eac7da3
c5e469a
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
---
license: apache-2.0
datasets:
- bigcode/the-stack-smol
- ttbui/html_alpaca
language:
- en
tags:
- code
- coding
- small
- tiny
---

# Welcome to htmLLM v2 124M!

With this LLM, we wanted to see, how well tiny LLMs with just 124 million parameters can perform on coding tasks.

This model is also a bit finetuned using html_alpaca directly in the pretraining.

If you want to try it, you can use htmllm.ipynb in the HF model files and download the model weight from this HF model.

# Code
All code can be accessed via the file **htmllm_v2_124m.ipynb** in this HF model.

# Weights
The final **base** model checkpoint can be downloaded here in the files list as **ckpt.pt**.

# Training
We trained our model on a single Kaggle T4 GPU.

# Thanks to:
- Andrej Karpathy and his nanoGPT code
- Kaggle for the free GPU hours for training on the T4
- You all for your support on my reddit.