Safetensors
Serbian
mistral
YugoGPT-Florida / README.md
datatab's picture
Update README.md
dcf670e verified
---
license: mit
datasets:
- datatab/SrpWikiDataset
- datatab/open-orca-slim-serbian
language:
- sr
base_model:
- datatab/Yugo55A-GPT
---
![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/629dff8f6daf7662067b81d7/qEuLoNYUhPcBhEusyvsKv.jpeg)
- **Developed by:** datatab
- **License:** mit
## 🏆 Results
> Results obtained through the [**Serbian LLM Evaluation Benchmark**](https://huggingface.co/datasets/datatab/serbian-llm-benchmark)
<table>
<tr>
<th><strong>MODEL<strong></th>
<th><strong>ARC-E<strong></th>
<th><strong>ARC-C<strong></th>
<th><strong>Hellaswag<strong></th>
<th><strong>PiQA<strong></th>
<th><strong>Winogrande<strong></th>
<th><strong>BoolQ<strong></th>
<th><strong>OpenbookQA<strong></th>
<th><strong>OZ_EVAL<strong></th>
<th>SCORE</th>
</tr>
<tr>
<td><a href="https://huggingface.co/datatab/YugoGPT-Florida/">YugoGPT-Florida</a></td>
<td><strong>0.6918<strong></td>
<td><strong>0.5766<strong></td>
<td><strong>0.4037<strong></td>
<td><strong>0.7374<strong></td>
<td><strong>0.5782<strong></td>
<td><strong>0.8685<strong></td>
<td><strong>0.5918<strong></td>
<td><strong>0.7407<strong></td>
<td><strong>64,85875<strong></td>
</tr>
<tr>
<td><a href="https://huggingface.co/datatab/Yugo55A-GPT/">Yugo55A-GPT</a></td>
<td>0.5846</td>
<td>0.5185</td>
<td>0.3686</td>
<td>0.7076</td>
<td>0.5277</td>
<td>0.8584</td>
<td>0.5485</td>
<td>0.6883</td>
<td>60,0275</td>
</tr>
<tr>
<td><a href="https://huggingface.co/datatab/Yugo60-GPT">Yugo60-GPT</a></td>
<td>0.4948</td>
<td>0.4542</td>
<td>0.3342</td>
<td>0.6897</td>
<td>0.5138</td>
<td>0.8212</td>
<td>0.5155</td>
<td>0.6379</td>
<td>55,76625</td>
</tr>
<tr>
<td><a href="https://huggingface.co/datatab/Yugo45-GPT">Yugo45-GPT</a></td>
<td>0.4049</td>
<td>0.3900</td>
<td>0.2812</td>
<td>0.6055</td>
<td>0.4992</td>
<td>0.5793</td>
<td>0.4433</td>
<td>0.6111</td>
<td>47,68125</td>
</tr>
</table>
![image/png](https://cdn-uploads.huggingface.co/production/uploads/629dff8f6daf7662067b81d7/CsMbGawHDqUsL-iFfini0.png)
![image/png](https://cdn-uploads.huggingface.co/production/uploads/629dff8f6daf7662067b81d7/eIThLiw4AJZdtkLHEOhy2.png)
![image/png](https://cdn-uploads.huggingface.co/production/uploads/629dff8f6daf7662067b81d7/i_J2-WAuRFGnzMFrm6Py-.png)
# 🏋️ Training Stats
![image/png](https://cdn-uploads.huggingface.co/production/uploads/629dff8f6daf7662067b81d7/rdwLicOUMSwNjbaajHhOe.png)
![image/png](https://cdn-uploads.huggingface.co/production/uploads/629dff8f6daf7662067b81d7/noi6p0oEvYLLRIKKg1923.png)
![image/png](https://cdn-uploads.huggingface.co/production/uploads/629dff8f6daf7662067b81d7/a_spZs0R-72cVrcerWADb.png)
![image/png](https://cdn-uploads.huggingface.co/production/uploads/629dff8f6daf7662067b81d7/juVLei2jy7ZNvXQ1tHRPN.png)
## 💻 Usage
```terminal
!pip -q install git+https://github.com/huggingface/transformers
```
```python
from IPython.display import HTML, display
def set_css():
display(HTML('''
<style>
pre {
white-space: pre-wrap;
}
</style>
'''))
get_ipython().events.register('pre_run_cell', set_css)
```
```python
import torch
import transformers
from transformers import AutoTokenizer, MistralForCausalLM
device = "cuda" if torch.cuda.is_available() else "cpu"
model = MistralForCausalLM.from_pretrained(
"datatab/YugoGPT-Florida",
torch_dtype="auto"
).to(device)
tokenizer = AutoTokenizer.from_pretrained("datatab/YugoGPT-Florida")
```
```python
from typing import Optional
from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer
def generate(
user_content: str, system_content: Optional[str] = ""
) -> str:
system_content = """Ispod se nalazi uputstvo koje definiše zadatak, zajedno sa unosom koji pruža dodatni kontekst.
Na osnovu ovih informacija, napišite odgovor koji precizno i tačno ispunjava zahtev.
"""
messages = [
{
"role": "system",
"content": system_content,
},
{"role": "user", "content": user_content},
]
tokenized_chat = tokenizer.apply_chat_template(
messages, tokenize=True, add_generation_prompt=True, return_tensors="pt"
).to("cuda")
text_streamer = TextStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True)
output = model.generate(
tokenized_chat,
streamer=text_streamer,
max_new_tokens=2048,
temperature=0.1,
repetition_penalty=1.11,
top_p=0.92,
top_k=1000,
pad_token_id=tokenizer.pad_token_id,
eos_token_id=tokenizer.eos_token_id,
do_sample=True,
)
generated_text = tokenizer.decode(output[0], skip_special_tokens=True)
```
```python
generate("Nabroj mi sve planete suncevog sistemai reci mi koja je najveca planeta?")
```
```terminal
Sunčev sistem sadrži osam planeta: Merkur, Venera, Zemlja, Mars, Jupiter, Saturn, Uran i Neptun. Najveća planeta u Sunčevom sistemu je Jupiter.
```
## 💡 Contributions Welcome!
Have ideas, bug fixes, or want to add a custom model? We'd love for you to be part of the journey! Contributions help grow and enhance the capabilities of the **YugoGPT-Florida**.
## 📜 Citation
Thanks for using **YugoGPT-Florida** — where language learning models meet Serbian precision and creativity! Let's build smarter models together. 🚀�
If you find this model useful in your research, please cite it as follows:
```bibtex
@article{YugoGPT-Florida},
title={YugoGPT-Florida},
author={datatab},
year={2024},
url={https://huggingface.co/datatab/YugoGPT-Florida}
}
```
<div id="zastava">
<div class="grb">
<img src="https://www.ai.gov.rs/img/logo_60x120-2.png" style="position:relative; left:30px; z-index:10; height:85px">
</div>
<table width=100% style="border:0px">
<<tr style="background-color:#C6363C;width:100%;border:0px;height:30px"><td style="width:100vw"></td></tr>
<tr style="background-color:#0C4076;width:100%;border:0px;height:30px"><td></td></tr>
<tr style="background-color:#ffffff;width:100%;border:0px;height:30px"><td></td></tr>
</table>
</div>
<style>
.ffeat: {
color:red
}
.cover {
width: 100%;
margin-bottom: 5pt
}
.highlight-container, .highlight {
position: relative;
text-decoration:none
}
.highlight-container {
display: inline-block;
}
.highlight{
color:white;
text-transform:uppercase;
font-size: 16pt;
}
.highlight-container{
padding:5px 10px
}
.highlight-container:before {
content: " ";
display: block;
height: 100%;
width: 100%;
margin-left: 0px;
margin-right: 0px;
position: absolute;
background: #e80909;
transform: rotate(2deg);
top: -1px;
left: -1px;
border-radius: 20% 25% 20% 24%;
padding: 10px 18px 18px 10px;
}
div.grb, #zastava>table {
position:absolute;
top:0px;
left: 0px;
margin:0px
}
div.grb>img, #zastava>table{
margin:0px
}
#zastava {
position: relative;
margin-bottom:120px
}
p {
font-size:14pt
}
</style>