GPT-USENET-2 / README.md
HDTenEightyP's picture
Update README.md
27b6c70 verified
---
license: mit
language:
- en
tags:
- text-generation-inference
pipeline_tag: text-generation
---
![GPTUsenet2](https://cdn-uploads.huggingface.co/production/uploads/64b7618e2f5a966b972e9978/rl907PheZr8k0URxR-UVQ.png)
## GPT-Usenet-2
An 81-million parameter LLM using GPT-2 encodings.
Trained using 10GB of USENET posts along with over 1 GB of miscellaneous BBS posts, digitized books, and text documents.
Supervised fine-tuning should be performed before use.
## Purpose of GPT-Usenet-2
LLMs are all currently focused on becoming larger and larger, able to do more and more. However, this just makes them jack of all trades, master of none. GPT-Usenet takes a different approach. Instead of trying to do everything perfectly, GPT-Usenet offers a digital stem cell, which can then be finetuned into a single, specialized role and run in parallel with copies of itself.
## Technical Information
| | |
|---------------------------------|----:|
|Layers |10|
|Heads |10|
|Embeddings |640|
|Context Window |1024 tokens|
|Tokenizer |GPT-2 BPE|
## Training Information
| | |
|---------------------------------|----:|
|Training Loss |around 2.0254|
|Validation Loss |around 1.9795|
|Device |Google Colab L4, Google Colab A100|
|Training Time |16 Hours|
## Example Syntax
| | |
|---------------------------------|----:|
|From:|The username who sent this message|
|Sender:|The group that username belongs to|
|Newsgroups:|The broad subject field of the email.|
|Subject:|The subject of the message.|
|Write the SFT response here. First, Prefix the first sentence with > to signify that it is a Reasoning sentence.||
|--|The stop tokens|
```
From:user
Sender:usergroup
Newsgroups:motorskills.papercraft
Subject:Paper airplanes
>Provide detailed steps on building a paper airplane.
Instructions: ...
--
```
For finetuning, your data should be in the .mbox format.