AI & ML interests

Arabic & bilingual LLM pretraining • Fine-tuning (domain & cultural alignment) • Preference alignment (SFT, DPO, PPO, RLHF/RLAIF) • Reinforcement learning • Evaluation harnesses (Arabic & multilingual) • Efficient RAG (re-ranking & guardrails) • ASR • TTS • Vision-language models (VLMs)

Recent Activity

Articles

MohamedRashad 
posted an update 5 months ago
MohamedRashad 
posted an update 5 months ago
MohamedRashad 
posted an update 7 months ago
MohamedRashad 
posted an update 7 months ago
view post
Post
385
I just updated an old (non working) space i had with the implementation of a cool research paper named UniRig

The idea is that you upload any 3d model and it rigs it for you with correct armature and the skinning process to give you the final model fully rigged and ready to be used.

Check it out here:
MohamedRashad/UniRig
MohamedRashad 
posted an update 8 months ago
view post
Post
1090
I have processed and cleaned the famous SADA2022 dataset from SADIA for Arabic ASR and other related tasks and uploaded it here:
MohamedRashad/SADA22

Edit:
I also added another dataset from SADIA named SCC22
MohamedRashad/SCC22
MohamedRashad 
posted an update 9 months ago
view post
Post
2721
I collected the recitations of the holy quran from 20 different reciters and uploaded the full dataset here:
MohamedRashad/Quran-Recitations

Check it out 🥷
  • 1 reply
·
MohamedRashad 
posted an update 9 months ago
MohamedRashad 
posted an update 10 months ago
view post
Post
3538
I think we have released the best Arabic model under 25B at least based on https://huggingface.co/spaces/inceptionai/AraGen-Leaderboard

Yehia = https://huggingface.co/ALLaM-AI/ALLaM-7B-Instruct-preview + GRPO

and its ranked number one model under the 25B parameter size mark.

Now, i said "i think" not "i am sure" because this model used the same metric of evaluation the AraGen developers use (the 3C3H) as a reward model to improve its responses and this sparks the question. Is this something good for users or is it another type of overfitting that we don't want ?

I don't know if this is a good thing or a bad thing but what i know is that you can try it from here:
Navid-AI/Yehia-7B-preview

or Download it for your personal experiments from here:
Navid-AI/Yehia-7B-preview

Ramadan Kareem 🌙
  • 1 reply
·
MohamedRashad 
posted an update 11 months ago
MohamedRashad 
posted an update 12 months ago
view post
Post
2108
The winners of Best Paper Award in NeurIPs2024 (FoundationVision) Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction (2404.02905) has just released a new paper called infinty:
Infinity: Scaling Bitwise AutoRegressive Modeling for High-Resolution Image Synthesis (2412.04431)

And i managed to build a space for it so anyone can try it out: MohamedRashad/Infinity

The idea of a text to image model using autoregressive archticture is quite interesting in my opinion.
MohamedRashad 
posted an update about 1 year ago
MohamedRashad 
posted an update about 1 year ago
MohamedRashad 
posted an update about 1 year ago
MohamedRashad 
posted an update over 1 year ago
MohamedRashad 
posted an update over 1 year ago
view post
Post
3574
For all the Muslims out there who are interested in Quran and its tafsir (explanations). This humble dataset consists of 84 different books of tafsir for nearly all the ayat in the Quran:
MohamedRashad/Quran-Tafseer

I hope it helps someone to build something nice and useful with it ^_^
MohamedRashad 
posted an update over 1 year ago
MohamedRashad 
posted an update over 1 year ago
MohamedRashad 
posted an update over 1 year ago
MohamedRashad 
posted an update over 1 year ago