Improve language tag

#1
by lbourdois - opened
Files changed (1) hide show
  1. README.md +88 -77
README.md CHANGED
@@ -1,78 +1,89 @@
1
- ---
2
- license: apache-2.0
3
- tags:
4
- - trl
5
- - sft
6
- - hinglish
7
- language:
8
- - en
9
- - hi
10
- base_model:
11
- - Qwen/Qwen2.5-7B
12
- pipeline_tag: text-generation
13
- ---
14
- # ๐Ÿš€ Zira-Z.1 ๐ŸŒŸ
15
- ### *The Bilingual Beast Built on Qwen 2.5 (7B)*
16
-
17
- ![Zira-Z.1 Banner](img/banner.png) <!-- Add an epic banner image -->
18
-
19
- ---
20
-
21
- ## ๐Ÿง  Model Highlights
22
-
23
- > **Zira-Z.1** isn't just a model โ€” it's a revolution in understanding *both* English and Hinglish.
24
- > Born from the powerful DNA of **Qwen 2.5 (7B)**, this multilingual marvel was fine-tuned for raw text generation across two of the most widely spoken languages in the world.
25
-
26
- - ๐Ÿ’ฅ **Base**: Qwen 2.5 - 7B (One of the finest open LLMs out there)
27
- - ๐Ÿ—ฃ๏ธ **Languages**: English ๐Ÿ‡ฌ๐Ÿ‡ง + Hinglish ๐Ÿ‡ฎ๐Ÿ‡ณ (Code-mixed, no pure Hindi)
28
- - ๐Ÿ”ง **Training**: Fine-tuned on diverse bilingual corpora โ€” clean, simple text format (non-instruct)
29
- - ๐Ÿฆพ **Purpose**: General-purpose **text generation**, especially where English and Hinglish blend naturally
30
-
31
- **Please NOTE that this is a basic text generation model and lacks coherence in its output; the release of the new instruct model has been delayed due to resource constraints, with an expected launch in approximately 5 days.**
32
-
33
- ---
34
-
35
- ## ๐Ÿ” Why Zira-Z.1?
36
-
37
- Because **multilingual LLMs** are cool.
38
- But **Zira-Z.1** is cooler. ๐Ÿ˜Ž
39
-
40
- - ๐Ÿ”— Code-switching? Natural.
41
- - โœ๏ธ Generates culturally fluent, relatable Hinglish.
42
- - ๐Ÿ“š Handles casual text, commentary, social chatter, and more.
43
- - ๐ŸŽฏ Perfect for early-stage Indic bilingual applications and experimentation
44
-
45
- ---
46
-
47
- ## ๐Ÿ“‰ Training Curve
48
-
49
- > *She trained hard, and it shows...*
50
-
51
- ![Insert loss curve here](img/Figure_1.png) <!-- Add your actual training curve image here -->
52
-
53
- ---
54
-
55
- ## ๐Ÿ› ๏ธ Usage
56
-
57
- ```import transformers
58
-
59
- from transformers import AutoTokenizer, AutoModelForCausalLM
60
-
61
- tokenizer = AutoTokenizer.from_pretrained("HyperX-Sen/Zira-Z.1")
62
- model = AutoModelForCausalLM.from_pretrained("HyperX-Sen/Zira-Z.1")
63
-
64
- inputs = tokenizer("Tum kya soch rahe ho about AI?", return_tensors="pt")
65
- outputs = model.generate(**inputs, max_new_tokens=50)
66
- print(tokenizer.decode(outputs[0], skip_special_tokens=True))'
67
- ```
68
-
69
- ---
70
-
71
- ## ๐Ÿงฌ License & Contribution
72
-
73
- - ๐Ÿ“œ **License**: Open for research & commercial use (see LICENSE)
74
- - ๐Ÿค Contributions: Welcomed with open arms (and open pull requests)
75
-
76
- ---
77
-
 
 
 
 
 
 
 
 
 
 
 
78
  Made with โค๏ธ, logic, and a lot of chai โ˜•
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - trl
5
+ - sft
6
+ - hinglish
7
+ language:
8
+ - zho
9
+ - eng
10
+ - fra
11
+ - spa
12
+ - por
13
+ - deu
14
+ - ita
15
+ - rus
16
+ - jpn
17
+ - kor
18
+ - vie
19
+ - tha
20
+ - ara
21
+ base_model:
22
+ - Qwen/Qwen2.5-7B
23
+ pipeline_tag: text-generation
24
+ ---
25
+ # ๐Ÿš€ Zira-Z.1 ๐ŸŒŸ
26
+ ### *The Bilingual Beast Built on Qwen 2.5 (7B)*
27
+
28
+ ![Zira-Z.1 Banner](img/banner.png) <!-- Add an epic banner image -->
29
+
30
+ ---
31
+
32
+ ## ๐Ÿง  Model Highlights
33
+
34
+ > **Zira-Z.1** isn't just a model โ€” it's a revolution in understanding *both* English and Hinglish.
35
+ > Born from the powerful DNA of **Qwen 2.5 (7B)**, this multilingual marvel was fine-tuned for raw text generation across two of the most widely spoken languages in the world.
36
+
37
+ - ๐Ÿ’ฅ **Base**: Qwen 2.5 - 7B (One of the finest open LLMs out there)
38
+ - ๐Ÿ—ฃ๏ธ **Languages**: English ๐Ÿ‡ฌ๐Ÿ‡ง + Hinglish ๐Ÿ‡ฎ๐Ÿ‡ณ (Code-mixed, no pure Hindi)
39
+ - ๐Ÿ”ง **Training**: Fine-tuned on diverse bilingual corpora โ€” clean, simple text format (non-instruct)
40
+ - ๐Ÿฆพ **Purpose**: General-purpose **text generation**, especially where English and Hinglish blend naturally
41
+
42
+ **Please NOTE that this is a basic text generation model and lacks coherence in its output; the release of the new instruct model has been delayed due to resource constraints, with an expected launch in approximately 5 days.**
43
+
44
+ ---
45
+
46
+ ## ๐Ÿ” Why Zira-Z.1?
47
+
48
+ Because **multilingual LLMs** are cool.
49
+ But **Zira-Z.1** is cooler. ๐Ÿ˜Ž
50
+
51
+ - ๐Ÿ”— Code-switching? Natural.
52
+ - โœ๏ธ Generates culturally fluent, relatable Hinglish.
53
+ - ๐Ÿ“š Handles casual text, commentary, social chatter, and more.
54
+ - ๐ŸŽฏ Perfect for early-stage Indic bilingual applications and experimentation
55
+
56
+ ---
57
+
58
+ ## ๐Ÿ“‰ Training Curve
59
+
60
+ > *She trained hard, and it shows...*
61
+
62
+ ![Insert loss curve here](img/Figure_1.png) <!-- Add your actual training curve image here -->
63
+
64
+ ---
65
+
66
+ ## ๐Ÿ› ๏ธ Usage
67
+
68
+ ```import transformers
69
+
70
+ from transformers import AutoTokenizer, AutoModelForCausalLM
71
+
72
+ tokenizer = AutoTokenizer.from_pretrained("HyperX-Sen/Zira-Z.1")
73
+ model = AutoModelForCausalLM.from_pretrained("HyperX-Sen/Zira-Z.1")
74
+
75
+ inputs = tokenizer("Tum kya soch rahe ho about AI?", return_tensors="pt")
76
+ outputs = model.generate(**inputs, max_new_tokens=50)
77
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))'
78
+ ```
79
+
80
+ ---
81
+
82
+ ## ๐Ÿงฌ License & Contribution
83
+
84
+ - ๐Ÿ“œ **License**: Open for research & commercial use (see LICENSE)
85
+ - ๐Ÿค Contributions: Welcomed with open arms (and open pull requests)
86
+
87
+ ---
88
+
89
  Made with โค๏ธ, logic, and a lot of chai โ˜•