Edraky commited on
Commit
721eda9
ยท
verified ยท
1 Parent(s): 025311d

Create Edraky AI!

Browse files

<p align="center">
<img src="https://cdn-uploads.huggingface.co/production/uploads/686e726239f003427404a1be/uuB7LFKDX1C5B28DGJyZN.png" alt="Edraky Logo" width="200"/>
</p>

# ๐Ÿง  ุฅุฏุฑุงูƒูŠ (Edraky) โ€” Multilingual Educational AI Assistant

**ุฅุฏุฑุงูƒูŠ (Edraky)** is a multilingual AI model designed to support students with interactive learning, question-answering, and intelligent content generation across **Arabic**, **English**, and **Hebrew**. It is especially focused on **educational content** for middle-school learners, including subjects like Social Studies, Arabic Language, and more.

---

## ๐Ÿงพ Model Details

- **Languages**: Arabic, English, Hebrew
- **License**: Apache 2.0
- **Base Model**: `Qwen/Qwen2-1.5B-Instruct`
- **Library**: `transformers`
- **Pipeline Tag**: `text-generation`
- **Trained on**:
- `rosoft/rStar-Coder`
- `gsm8k-rerun/Qwen_Qwen2.5-1.5B-Instruct`
- **Tags**: `multilingual`, `educational`, `fine-tuned`, `arabic`, `hebrew`, `qwen`, `student-assistant`

---

## ๐ŸŽฏ Use Cases

- ๐Ÿ“š **AI Study Assistant**: Helps students with curriculum-based questions
- ๐Ÿง  **Educational Content Creation**: Generates quiz questions, summaries, and explanations
- ๐ŸŒ **Language Support**: Understands and responds in Arabic, English, and Hebrew
- ๐Ÿ‘จโ€๐Ÿซ **Teacher Support**: Generates model answers, revision notes, and more

---

## ๐Ÿ” Evaluation Metrics

| Metric | Value |
|------------|-------|
| Accuracy | โœ… High (on educational QA datasets) |
| BLEU Score | ๐Ÿงช Coming Soon |
| F1 Score | ๐Ÿงช Coming Soon |

---

## ๐Ÿ“š Training Datasets

The model was fine-tuned on high-quality datasets including:

- [`rosoft/rStar-Coder`](https://huggingface.co/datasets/rosoft/rStar-Coder)
- [`gsm8k-rerun/Qwen_Qwen2.5-1.5B-Instruct`](https://huggingface.co/datasets/gsm8k-rerun/Qwen_Qwen2.5-1.5B-Instruct)

Additional in-house educational data tailored for Egyptian curricula was used to optimize performance.

---

## ๐Ÿš€ How to Use

You can try it with the Hugging Face Inference API or use it directly via Transformers:

```python
from transformers import AutoModelForCausalLM, AutoTokenizer

model_id = "your-username/edraky-ai"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id)

inputs = tokenizer("ุงุดุฑุญ ุฏุฑุณ ุงู„ุซูˆุฑุฉ ุงู„ุนุฑุงุจูŠุฉ", return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=200)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
๐Ÿ’ก Intended Uses
This model is designed to be:

๐Ÿง‘โ€๐ŸŽ“ Student-friendly

๐Ÿ‘ฉโ€๐Ÿซ Useful for teachers and educators

๐ŸŒ Ready for multilingual educational tasks

โš ๏ธ Limitations
This model is not a replacement for a certified teacher.

Knowledge is limited to the data it was trained on.

May hallucinate or generate incorrect facts; always verify educational material.

๐Ÿค Contributing
We're open to collaboration, dataset contributions, or feedback to improve this model further for educational use in the Arab world and beyond.

๐Ÿงพ License
Apache License 2.0 โ€” Free to use, share, and modify with attribution.

Files changed (1) hide show
  1. README.md +26 -3
README.md CHANGED
@@ -1,3 +1,26 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - fka/awesome-chatgpt-prompts
5
+ - microsoft/rStar-Coder
6
+ - gsm8k-rerun/Qwen_Qwen2.5-1.5B-Instruct
7
+ language:
8
+ - ar
9
+ - en
10
+ - he
11
+ metrics:
12
+ - accuracy
13
+ - perplexity
14
+ - wer
15
+ base_model:
16
+ - Qwen/Qwen2-1.5B-Instruct
17
+ pipeline_tag: text-generation
18
+ library_name: transformers
19
+ tags:
20
+ - multilingual
21
+ - arabic
22
+ - hebrew
23
+ - qwen
24
+ - educational
25
+ - fine-tuned
26
+ ---