|
|
--- |
|
|
language: |
|
|
- en |
|
|
pipeline_tag: text-generation |
|
|
tags: |
|
|
- pytorch |
|
|
- causal-lm |
|
|
- custom |
|
|
- general |
|
|
--- |
|
|
|
|
|
# Sky610TX |
|
|
|
|
|
|
|
|
## Model Details |
|
|
- **Architecture:** GPT-2 Style (Custom Ascendant Config) |
|
|
- **Parameters:** ~389 Million |
|
|
- **Training tokens:** 1.3 Billion |
|
|
- **Context Window:** 1024 Tokens |
|
|
- **50k iterations** |
|
|
|
|
|
## The Future |
|
|
|
|
|
**Work has started on a new, 1.2B parameter model. It will be much better at coding, reasoning, facts, and conversation with over 10B tokens! It is currently in development and is expected to release soon** |
|
|
|
|
|
|
|
|
## How to Use |
|
|
```python |
|
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
|
|
|
|
model = AutoModelForCausalLM.from_pretrained("8BitStudio/Sky610TX") |
|
|
tokenizer = AutoTokenizer.from_pretrained("8BitStudio/Sky610TX") |
|
|
|
|
|
input_text = "User: Hello\nAssistant:" |
|
|
inputs = tokenizer(input_text, return_tensors="pt") |
|
|
outputs = model.generate(**inputs, max_new_tokens=50) |
|
|
print(tokenizer.decode(outputs[0])) |