File size: 1,194 Bytes
933ed0f
 
 
 
 
5970a64
 
 
 
3125d16
 
4df8b3c
3125d16
 
 
4df8b3c
bdd8e93
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7646c36
2ccbe9d
7646c36
bdd8e93
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
---
language:
- ar
library_name: transformers
pipeline_tag: text-generation
datasets:
- IBB-University/DATA_FATAWA
metrics:
- accuracy
- bleu
- bertscore
widget:
- text: ' اركان الاسلام'
- text: ما حكم الاحتفال بالمولد النبوي
- text: ما هي الكتب السماوية
---
## Testing the model using `transformers`:

```python
from transformers import GPT2TokenizerFast, pipeline
#for base and medium
from transformers import GPT2LMHeadModel
#for large and mega
# pip install arabert
from arabert.aragpt2.grover.modeling_gpt2 import GPT2LMHeadModel

from arabert.preprocess import ArabertPreprocessor

MODEL_NAME='IBB-University/ghadeer_question_answer'
arabert_prep = ArabertPreprocessor(model_name=MODEL_NAME)

text=""
text_clean = arabert_prep.preprocess(text)

model = GPT2LMHeadModel.from_pretrained(MODEL_NAME)
tokenizer = GPT2TokenizerFast.from_pretrained(MODEL_NAME)
generation_pipeline = pipeline("text-generation",model=model,tokenizer=tokenizer)

#feel free to try different decoding settings
generation_pipeline(text,
    pad_token_id=tokenizer.eos_token_id,
     max_length=512,
    penalty_alpha=0.6,
    top_k=4 )[0]['generated_text']
```