Update README.md
Browse files
README.md
CHANGED
|
@@ -33,6 +33,20 @@ datasets:
|
|
| 33 |
|
| 34 |
## Introduction
|
| 35 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 36 |
## How to run (transformers)
|
| 37 |
|
| 38 |
### Install transformers
|
|
@@ -64,4 +78,12 @@ gen_tokens = model.generate(
|
|
| 64 |
|
| 65 |
gen_text = tokenizer.decode(gen_tokens[0])
|
| 66 |
print(gen_text)
|
| 67 |
-
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 33 |
|
| 34 |
## Introduction
|
| 35 |
|
| 36 |
+
This model is an effort in order to make a multi-lingual and _on device_ models which can be executed on the consumer hardware. The model follows the steps used in training _DeepSeek_ model. However, the model is _not a reasoning model_ and a generic question answering, conversational and _uncensored_ model which has been made with a cost of around $4000 USD.
|
| 37 |
+
|
| 38 |
+
If you're curious about the model you also can see our [GitHub](https://github.com/mann-e/hormoz) and learn more about the benchmarks and costs.
|
| 39 |
+
|
| 40 |
+
Also, this model is based on _Command R_'s architecture, since that architecture gave us the best results in multilingual chat. Specially with languages such as _Persian_ and _Arabic_. This way, you can consider this model like a commercially useaeble version of _aya expanse_ as well.
|
| 41 |
+
|
| 42 |
+
### The name
|
| 43 |
+
|
| 44 |
+
<p align="center">
|
| 45 |
+
<img src="https://github.com/Mann-E/hormoz/blob/main/hormoz-logo.png?raw=true" width=768px />
|
| 46 |
+
</p>
|
| 47 |
+
|
| 48 |
+
The name __Hormoz__ comes from the Persian word "هرمز" which has multiple meanings. It can point to the _strait of Hormoz_ in Persian Gulf or _Hormoz Island_ which is part of the Hormozgan Province in the south of Iran. Also it may point to "اورمزد" or _Ourmozd_ which is middle/ancient Persian name for the planet _Jupiter_ and derived from the term _Ahura Mazda_ or the Avestan term for God.
|
| 49 |
+
|
| 50 |
## How to run (transformers)
|
| 51 |
|
| 52 |
### Install transformers
|
|
|
|
| 78 |
|
| 79 |
gen_text = tokenizer.decode(gen_tokens[0])
|
| 80 |
print(gen_text)
|
| 81 |
+
```
|
| 82 |
+
|
| 83 |
+
## License
|
| 84 |
+
|
| 85 |
+
This model is published under _MIT_ license.
|
| 86 |
+
|
| 87 |
+
### Commercial Use
|
| 88 |
+
|
| 89 |
+
Since this model is MIT licensed, you're free to do whatever you want with the model. However since we're a relatively small startup, we recommend you if you are a big corporate and you host this model, give us a capacity of your API as well. This way, we both can benefit from the model.
|