| | import gradio as gr |
| | import torch |
| | from transformers import AutoModelForCausalLM, AutoTokenizer |
| | import os |
| |
|
| | |
| | |
| | MODEL_ID = "TinyLlama/TinyLlama-1.1B-Chat-v1.0" |
| |
|
| | |
| |
|
| | |
| | def load_model(): |
| | """تحميل نموذج Phi-3-Mini المكمم (أقل استهلاكاً للذاكرة).""" |
| | try: |
| | |
| | device = torch.device("cpu") |
| | print(f"✅ سيتم تشغيل النموذج على: {device}") |
| |
|
| | |
| | model = AutoModelForCausalLM.from_pretrained( |
| | MODEL_ID, |
| | torch_dtype=torch.float32, |
| | load_in_4bit=True, |
| | device_map="auto" |
| | ).to(device) |
| | |
| | tokenizer = AutoTokenizer.from_pretrained(MODEL_ID) |
| | |
| | return model, tokenizer, device |
| | |
| | except Exception as e: |
| | print(f"❌ فشل تحميل نموذج Phi-3-Mini: {e}") |
| | return None, None, None |
| |
|
| | |
| | model, tokenizer, device = load_model() |
| |
|
| | |
| | |