llm / README.md
lemms's picture
Upload README.md with huggingface_hub
5ae897b verified
---
title: OpenLLM Inference Space
emoji: ๐Ÿš€
colorFrom: blue
colorTo: purple
sdk: gradio
sdk_version: 4.44.1
app_file: app.py
pinned: false
license: gpl-3.0
---
# ๐Ÿš€ OpenLLM Inference Space
Welcome to the OpenLLM Inference Space! This is a comprehensive interface for running inference on our trained OpenLLM models with customizable parameters.
## ๐ŸŽฏ Available Models
We provide **7 different models** trained for varying numbers of steps:
| Model | Training Steps | Description | Best Loss |
|-------|---------------|-------------|-----------|
| **4k Model** | 4,000 | Early training stage, basic language patterns | ~6.2 |
| **6k Model** | 6,000 | Improved coherence, better vocabulary usage | ~5.8 |
| **7k Model** | 7,000 | Enhanced text generation quality | ~5.5 |
| **8k Model** | 8,000 | More sophisticated language understanding | ~5.3 |
| **9k Model** | 9,000 | Best performing model (latest training) | ~5.2 |
| **10k Model** | 10,000 | Latest extended training, maximum performance | ~5.22 |
| **10k Improved** | 10,000 | Improved training process, proper checkpoint format | ~5.1774 |
## ๐ŸŽฎ How to Use
1. **Select a Model** from the dropdown menu
2. **Load the Model** to see its information
3. **Enter Your Prompt** in the text box
4. **Adjust Parameters** (temperature, max length, etc.)
5. **Generate Text** and see the results!
## โš™๏ธ Parameters
- **Temperature**: Controls randomness (0.1-2.0)
- **Max Length**: Number of tokens to generate (10-500)
- **Top-K**: Limits to top-k most likely tokens (1-100)
- **Top-P**: Nucleus sampling threshold (0.1-1.0)
## ๐Ÿง  Model Architecture
- **Model Size**: Small (35.8M parameters)
- **Layers**: 6 transformer layers
- **Embedding**: 512 dimensions
- **Vocabulary**: 32,000 tokens (SentencePiece)
- **Context Length**: 1,024 tokens
---
**OpenLLM Inference Space** - Experience the power of open-source language models! ๐Ÿš€