|
|
--- |
|
|
base_model: Qwen/Qwen3-VL-8B-Instruct |
|
|
library_name: peft |
|
|
pipeline_tag: text-generation |
|
|
tags: |
|
|
- base_model:adapter:Qwen/Qwen3-VL-8B-Instruct |
|
|
- lora |
|
|
- sft |
|
|
- transformers |
|
|
- trl |
|
|
--- |
|
|
|
|
|
# Model Card for ar-ms-baseline |
|
|
|
|
|
## Model Summary |
|
|
|
|
|
This model is the baseline system for the NAKBA NLP 2026: Arabic Manuscript Understanding Shared Task (Systems Track). It fine-tunes Qwen3-VL-8B-Instruct with LoRA to transcribe Arabic manuscript line images into text. |
|
|
|
|
|
## Model Details |
|
|
|
|
|
### Description |
|
|
|
|
|
- **Model type:** Vision-language OCR/HTR model (LoRA-adapted) |
|
|
- **Finetuned from model:** Qwen/Qwen3-VL-8B-Instruct |
|
|
|
|
|
### Sources |
|
|
|
|
|
- **Repository:** https://github.com/U4RASD/ar-ms-baseline |
|
|
- **Shared Task:** https://acrps.ai/nakba-nlp-manu-understanding-2026 |
|
|
|
|
|
## Training Details |
|
|
|
|
|
### Training Data |
|
|
|
|
|
- NAKBA NLP 2026 Shared Task (Subtask 2) training split from the Omar Al-Saleh memoir collection. |
|
|
- Dataset includes line images with gold transcriptions. |
|
|
|
|
|
### Training Procedure |
|
|
|
|
|
- Supervised fine-tuning with LoRA adapters on Qwen/Qwen3-VL-8B-Instruct. |
|
|
|
|
|
#### Training Hyperparameters |
|
|
|
|
|
- **Config reference:** Hyperparameters are listed in `configs/default.json` |
|
|
|
|
|
## Evaluation |
|
|
|
|
|
### Testing Data, Factors & Metrics |
|
|
|
|
|
#### Testing Data |
|
|
|
|
|
- NAKBA NLP 2026 Shared Task (Subtask 2) released test set of line images. |
|
|
|
|
|
#### Metrics |
|
|
|
|
|
- **CER (Character Error Rate)** |
|
|
- **WER (Word Error Rate)** |
|
|
|
|
|
### Results |
|
|
|
|
|
On released test set: |
|
|
- CER: 0.2297 |
|
|
- WER: 0.4998 |
|
|
|
|
|
- **Hardware:** NVIDIA H100 SXM |
|
|
|
|
|
## Contact |
|
|
|
|
|
- ar-ms@dohainstitute.edu.qa |
|
|
|