File size: 1,910 Bytes
3b3f04b
afa2148
3b3f04b
 
 
 
4e3beb2
3b3f04b
 
 
 
 
afa2148
3b3f04b
afa2148
3b3f04b
afa2148
3b3f04b
5ae897b
3b3f04b
afa2148
 
 
 
 
 
 
 
5ae897b
898dc6a
afa2148
3b3f04b
afa2148
 
 
 
 
3b3f04b
afa2148
3b3f04b
afa2148
 
 
 
898dc6a
afa2148
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
---
title: OpenLLM Inference Space
emoji: ๐Ÿš€
colorFrom: blue
colorTo: purple
sdk: gradio
sdk_version: 4.44.1
app_file: app.py
pinned: false
license: gpl-3.0
---

# ๐Ÿš€ OpenLLM Inference Space

Welcome to the OpenLLM Inference Space! This is a comprehensive interface for running inference on our trained OpenLLM models with customizable parameters.

## ๐ŸŽฏ Available Models

We provide **7 different models** trained for varying numbers of steps:

| Model | Training Steps | Description | Best Loss |
|-------|---------------|-------------|-----------|
| **4k Model** | 4,000 | Early training stage, basic language patterns | ~6.2 |
| **6k Model** | 6,000 | Improved coherence, better vocabulary usage | ~5.8 |
| **7k Model** | 7,000 | Enhanced text generation quality | ~5.5 |
| **8k Model** | 8,000 | More sophisticated language understanding | ~5.3 |
| **9k Model** | 9,000 | Best performing model (latest training) | ~5.2 |
| **10k Model** | 10,000 | Latest extended training, maximum performance | ~5.22 |
| **10k Improved** | 10,000 | Improved training process, proper checkpoint format | ~5.1774 |

## ๐ŸŽฎ How to Use

1. **Select a Model** from the dropdown menu
2. **Load the Model** to see its information
3. **Enter Your Prompt** in the text box
4. **Adjust Parameters** (temperature, max length, etc.)
5. **Generate Text** and see the results!

## โš™๏ธ Parameters

- **Temperature**: Controls randomness (0.1-2.0)
- **Max Length**: Number of tokens to generate (10-500)
- **Top-K**: Limits to top-k most likely tokens (1-100)
- **Top-P**: Nucleus sampling threshold (0.1-1.0)

## ๐Ÿง  Model Architecture

- **Model Size**: Small (35.8M parameters)
- **Layers**: 6 transformer layers
- **Embedding**: 512 dimensions
- **Vocabulary**: 32,000 tokens (SentencePiece)
- **Context Length**: 1,024 tokens

---

**OpenLLM Inference Space** - Experience the power of open-source language models! ๐Ÿš€