HDTenEightyP commited on
Commit
c5e4dba
·
verified ·
1 Parent(s): 9666d33

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -3
README.md CHANGED
@@ -1,3 +1,14 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ language:
4
+ - en
5
+ tags:
6
+ - text-generation-inference
7
+ pipeline_tag: text-generation
8
+ ---
9
+ An 81-million parameter LLM using GPT-2 encodings.
10
+ Trained using 10GB of USENET posts along with over 1 GB of miscellaneous BBS posts, digitized books, and text documents.
11
+
12
+ This model has 10 layers, 10 heads and 640 embeddings, with a context window of 1024 tokens.
13
+ It was able to achieve a training loss of 2.3256 and validation loss of 2.3651.
14
+ Supervised fine-tuning should be performed before use.