File size: 1,194 Bytes
933ed0f 5970a64 3125d16 4df8b3c 3125d16 4df8b3c bdd8e93 7646c36 2ccbe9d 7646c36 bdd8e93 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 | ---
language:
- ar
library_name: transformers
pipeline_tag: text-generation
datasets:
- IBB-University/DATA_FATAWA
metrics:
- accuracy
- bleu
- bertscore
widget:
- text: ' اركان الاسلام'
- text: ما حكم الاحتفال بالمولد النبوي
- text: ما هي الكتب السماوية
---
## Testing the model using `transformers`:
```python
from transformers import GPT2TokenizerFast, pipeline
#for base and medium
from transformers import GPT2LMHeadModel
#for large and mega
# pip install arabert
from arabert.aragpt2.grover.modeling_gpt2 import GPT2LMHeadModel
from arabert.preprocess import ArabertPreprocessor
MODEL_NAME='IBB-University/ghadeer_question_answer'
arabert_prep = ArabertPreprocessor(model_name=MODEL_NAME)
text=""
text_clean = arabert_prep.preprocess(text)
model = GPT2LMHeadModel.from_pretrained(MODEL_NAME)
tokenizer = GPT2TokenizerFast.from_pretrained(MODEL_NAME)
generation_pipeline = pipeline("text-generation",model=model,tokenizer=tokenizer)
#feel free to try different decoding settings
generation_pipeline(text,
pad_token_id=tokenizer.eos_token_id,
max_length=512,
penalty_alpha=0.6,
top_k=4 )[0]['generated_text']
``` |