Zira-Z.1 / README.md
HyperX-Sen's picture
Update README.md
2ae3211 verified
---
license: apache-2.0
tags:
- trl
- sft
- hinglish
language:
- en
- hi
base_model:
- Qwen/Qwen2.5-7B
pipeline_tag: text-generation
---
# ๐Ÿš€ Zira-Z.1 ๐ŸŒŸ
### *The Bilingual Beast Built on Qwen 2.5 (7B)*
![Zira-Z.1 Banner](img/banner.png) <!-- Add an epic banner image -->
---
## ๐Ÿง  Model Highlights
> **Zira-Z.1** isn't just a model โ€” it's a revolution in understanding *both* English and Hinglish.
> Born from the powerful DNA of **Qwen 2.5 (7B)**, this multilingual marvel was fine-tuned for raw text generation across two of the most widely spoken languages in the world.
- ๐Ÿ’ฅ **Base**: Qwen 2.5 - 7B (One of the finest open LLMs out there)
- ๐Ÿ—ฃ๏ธ **Languages**: English ๐Ÿ‡ฌ๐Ÿ‡ง + Hinglish ๐Ÿ‡ฎ๐Ÿ‡ณ (Code-mixed, no pure Hindi)
- ๐Ÿ”ง **Training**: Fine-tuned on diverse bilingual corpora โ€” clean, simple text format (non-instruct)
- ๐Ÿฆพ **Purpose**: General-purpose **text generation**, especially where English and Hinglish blend naturally
**Please NOTE that this is a basic text generation model and lacks coherence in its output; the release of the new instruct model has been delayed due to resource constraints, with an expected launch in approximately 5 days.**
---
## ๐Ÿ” Why Zira-Z.1?
Because **multilingual LLMs** are cool.
But **Zira-Z.1** is cooler. ๐Ÿ˜Ž
- ๐Ÿ”— Code-switching? Natural.
- โœ๏ธ Generates culturally fluent, relatable Hinglish.
- ๐Ÿ“š Handles casual text, commentary, social chatter, and more.
- ๐ŸŽฏ Perfect for early-stage Indic bilingual applications and experimentation
---
## ๐Ÿ“‰ Training Curve
> *She trained hard, and it shows...*
![Insert loss curve here](img/Figure_1.png) <!-- Add your actual training curve image here -->
---
## ๐Ÿ› ๏ธ Usage
```import transformers
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("HyperX-Sen/Zira-Z.1")
model = AutoModelForCausalLM.from_pretrained("HyperX-Sen/Zira-Z.1")
inputs = tokenizer("Tum kya soch rahe ho about AI?", return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=50)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))'
```
---
## ๐Ÿงฌ License & Contribution
- ๐Ÿ“œ **License**: Open for research & commercial use (see LICENSE)
- ๐Ÿค Contributions: Welcomed with open arms (and open pull requests)
---
Made with โค๏ธ, logic, and a lot of chai โ˜•