Create Edraky AI!
Browse files<p align="center">
<img src="https://cdn-uploads.huggingface.co/production/uploads/686e726239f003427404a1be/uuB7LFKDX1C5B28DGJyZN.png" alt="Edraky Logo" width="200"/>
</p>
# ๐ง ุฅุฏุฑุงูู (Edraky) โ Multilingual Educational AI Assistant
**ุฅุฏุฑุงูู (Edraky)** is a multilingual AI model designed to support students with interactive learning, question-answering, and intelligent content generation across **Arabic**, **English**, and **Hebrew**. It is especially focused on **educational content** for middle-school learners, including subjects like Social Studies, Arabic Language, and more.
---
## ๐งพ Model Details
- **Languages**: Arabic, English, Hebrew
- **License**: Apache 2.0
- **Base Model**: `Qwen/Qwen2-1.5B-Instruct`
- **Library**: `transformers`
- **Pipeline Tag**: `text-generation`
- **Trained on**:
- `rosoft/rStar-Coder`
- `gsm8k-rerun/Qwen_Qwen2.5-1.5B-Instruct`
- **Tags**: `multilingual`, `educational`, `fine-tuned`, `arabic`, `hebrew`, `qwen`, `student-assistant`
---
## ๐ฏ Use Cases
- ๐ **AI Study Assistant**: Helps students with curriculum-based questions
- ๐ง **Educational Content Creation**: Generates quiz questions, summaries, and explanations
- ๐ **Language Support**: Understands and responds in Arabic, English, and Hebrew
- ๐จโ๐ซ **Teacher Support**: Generates model answers, revision notes, and more
---
## ๐ Evaluation Metrics
| Metric | Value |
|------------|-------|
| Accuracy | โ
High (on educational QA datasets) |
| BLEU Score | ๐งช Coming Soon |
| F1 Score | ๐งช Coming Soon |
---
## ๐ Training Datasets
The model was fine-tuned on high-quality datasets including:
- [`rosoft/rStar-Coder`](https://huggingface.co/datasets/rosoft/rStar-Coder)
- [`gsm8k-rerun/Qwen_Qwen2.5-1.5B-Instruct`](https://huggingface.co/datasets/gsm8k-rerun/Qwen_Qwen2.5-1.5B-Instruct)
Additional in-house educational data tailored for Egyptian curricula was used to optimize performance.
---
## ๐ How to Use
You can try it with the Hugging Face Inference API or use it directly via Transformers:
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model_id = "your-username/edraky-ai"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id)
inputs = tokenizer("ุงุดุฑุญ ุฏุฑุณ ุงูุซูุฑุฉ ุงูุนุฑุงุจูุฉ", return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=200)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
๐ก Intended Uses
This model is designed to be:
๐งโ๐ Student-friendly
๐ฉโ๐ซ Useful for teachers and educators
๐ Ready for multilingual educational tasks
โ ๏ธ Limitations
This model is not a replacement for a certified teacher.
Knowledge is limited to the data it was trained on.
May hallucinate or generate incorrect facts; always verify educational material.
๐ค Contributing
We're open to collaboration, dataset contributions, or feedback to improve this model further for educational use in the Arab world and beyond.
๐งพ License
Apache License 2.0 โ Free to use, share, and modify with attribution.
|
@@ -1,3 +1,26 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: apache-2.0
|
| 3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
datasets:
|
| 4 |
+
- fka/awesome-chatgpt-prompts
|
| 5 |
+
- microsoft/rStar-Coder
|
| 6 |
+
- gsm8k-rerun/Qwen_Qwen2.5-1.5B-Instruct
|
| 7 |
+
language:
|
| 8 |
+
- ar
|
| 9 |
+
- en
|
| 10 |
+
- he
|
| 11 |
+
metrics:
|
| 12 |
+
- accuracy
|
| 13 |
+
- perplexity
|
| 14 |
+
- wer
|
| 15 |
+
base_model:
|
| 16 |
+
- Qwen/Qwen2-1.5B-Instruct
|
| 17 |
+
pipeline_tag: text-generation
|
| 18 |
+
library_name: transformers
|
| 19 |
+
tags:
|
| 20 |
+
- multilingual
|
| 21 |
+
- arabic
|
| 22 |
+
- hebrew
|
| 23 |
+
- qwen
|
| 24 |
+
- educational
|
| 25 |
+
- fine-tuned
|
| 26 |
+
---
|