lemms commited on
Commit
afa2148
ยท
verified ยท
1 Parent(s): 85c2681

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +36 -28
README.md CHANGED
@@ -1,47 +1,55 @@
1
  ---
2
- title: OpenLLM Real Models Space
3
  emoji: ๐Ÿš€
4
  colorFrom: blue
5
  colorTo: purple
6
  sdk: gradio
7
- sdk_version: 4.44.1
8
  app_file: app.py
9
  pinned: false
10
  license: gpl-3.0
11
  ---
12
 
13
- # ๐Ÿš€ OpenLLM Real Models Space
14
 
15
- Welcome to the OpenLLM Real Models Space! This interface uses **actual trained models** from Hugging Face.
16
 
17
- ## ๐ŸŽฏ Real Trained Models
18
 
19
- We provide **5 different real models** with varying training steps:
20
 
21
- | Model | Training Steps | Parameters | Performance |
22
- |-------|---------------|------------|-------------|
23
- | **4k Model** | 4,000 | 35.8M | Early training stage |
24
- | **6k Model** | 6,000 | 35.8M | Improved coherence (Perplexity: 816.040) |
25
- | **7k Model** | 7,000 | 35.8M | Enhanced quality (Loss: 2.100, Perplexity: 8.200) |
26
- | **8k Model** | 8,000 | 35.8M | Sophisticated understanding |
27
- | **9k Model** | 9,000 | 35.8M | Best performing model |
 
28
 
29
- **These are real GPT-style transformer models trained on Wikipedia passages from the SQuAD dataset.**
30
 
31
- ## ๐Ÿ”ง Technical Details
 
 
 
 
32
 
33
- - **Architecture**: GPT-style transformer decoder
34
- - **Model Size**: Small (6 layers, 8 heads, 512 embedding dim)
35
- - **Vocabulary**: 32k tokens (SentencePiece BPE)
36
- - **Training Data**: Wikipedia passages from SQuAD dataset
37
- - **Framework**: PyTorch with real trained models
38
- - **Gradio Version**: 4.44.1 (latest)
39
 
40
- **These models generate actual text based on their training on Wikipedia content.**
 
 
 
41
 
42
- **Model Sources:**
43
- - [4k Model](https://huggingface.co/lemms/openllm-small-extended-4k)
44
- - [6k Model](https://huggingface.co/lemms/openllm-small-extended-6k)
45
- - [7k Model](https://huggingface.co/lemms/openllm-small-extended-7k)
46
- - [8k Model](https://huggingface.co/lemms/openllm-small-extended-8k)
47
- - [9k Model](https://huggingface.co/lemms/openllm-small-extended-9k)
 
 
 
 
 
 
1
  ---
2
+ title: OpenLLM Inference Space
3
  emoji: ๐Ÿš€
4
  colorFrom: blue
5
  colorTo: purple
6
  sdk: gradio
7
+ sdk_version: 4.0.0
8
  app_file: app.py
9
  pinned: false
10
  license: gpl-3.0
11
  ---
12
 
13
+ # ๐Ÿš€ OpenLLM Inference Space
14
 
15
+ Welcome to the OpenLLM Inference Space! This is a comprehensive interface for running inference on our trained OpenLLM models with customizable parameters.
16
 
17
+ ## ๐ŸŽฏ Available Models
18
 
19
+ We provide **6 different models** trained for varying numbers of steps:
20
 
21
+ | Model | Training Steps | Description | Best Loss |
22
+ |-------|---------------|-------------|-----------|
23
+ | **4k Model** | 4,000 | Early training stage, basic language patterns | ~6.2 |
24
+ | **6k Model** | 6,000 | Improved coherence, better vocabulary usage | ~5.8 |
25
+ | **7k Model** | 7,000 | Enhanced text generation quality | ~5.5 |
26
+ | **8k Model** | 8,000 | More sophisticated language understanding | ~5.3 |
27
+ | **9k Model** | 9,000 | Best performing model (latest training) | ~5.2 |
28
+ | **10k Model** | 10,000 | Latest extended training, maximum performance | ~5.22 |
29
 
30
+ ## ๐ŸŽฎ How to Use
31
 
32
+ 1. **Select a Model** from the dropdown menu
33
+ 2. **Load the Model** to see its information
34
+ 3. **Enter Your Prompt** in the text box
35
+ 4. **Adjust Parameters** (temperature, max length, etc.)
36
+ 5. **Generate Text** and see the results!
37
 
38
+ ## โš™๏ธ Parameters
 
 
 
 
 
39
 
40
+ - **Temperature**: Controls randomness (0.1-2.0)
41
+ - **Max Length**: Number of tokens to generate (10-500)
42
+ - **Top-K**: Limits to top-k most likely tokens (1-100)
43
+ - **Top-P**: Nucleus sampling threshold (0.1-1.0)
44
 
45
+ ## ๐Ÿง  Model Architecture
46
+
47
+ - **Model Size**: Small (35.8M parameters)
48
+ - **Layers**: 6 transformer layers
49
+ - **Embedding**: 512 dimensions
50
+ - **Vocabulary**: 32,000 tokens (SentencePiece)
51
+ - **Context Length**: 1,024 tokens
52
+
53
+ ---
54
+
55
+ **OpenLLM Inference Space** - Experience the power of open-source language models! ๐Ÿš€