Datasets:

Modalities:
Text
Formats:
parquet
Languages:
Serbian
License:
JovanaR commited on
Commit
6b295d5
·
verified ·
1 Parent(s): ff5167d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -60,7 +60,7 @@ This work presents the development of two question–answering (QA) datasets for
60
 
61
  The **SQuAD-sr-md** dataset contains around 7k examples and represents a manually refined subset of the original SQuAD-sr dataset.While SQuAD-sr is a significant resource for Serbian QA research, it was created through a translation-based approach using an adapted Translate–Align–Retrieve method,
62
  which introduced various linguistic inconsistencies. The manual revision process, although time-consuming, substantially improved the grammatical correctness, terminological consistency, and overall reliability of the dataset, making it more suitable for training extractive QA models.
63
- We fine-tuned a several Serbian base encoder models on (i) our SQuAD-sr-md dataset and (ii) a subset of SQuAD-sr that corresponds exactly to the articles and QA pairs thatwere refined to create the SQuAD-sr-md. Evaluation conducted of the models highlighted the TeslaXLM base model as a particularly promising candidate for future extractive QA training. The results also indicate that the refinement of the dataset positively impacts model performance.
64
 
65
  ### SQuAD-sr subset
66
 
 
60
 
61
  The **SQuAD-sr-md** dataset contains around 7k examples and represents a manually refined subset of the original SQuAD-sr dataset.While SQuAD-sr is a significant resource for Serbian QA research, it was created through a translation-based approach using an adapted Translate–Align–Retrieve method,
62
  which introduced various linguistic inconsistencies. The manual revision process, although time-consuming, substantially improved the grammatical correctness, terminological consistency, and overall reliability of the dataset, making it more suitable for training extractive QA models.
63
+ We fine-tuned a several Serbian base encoder models on (i) our SQuAD-sr-md dataset and (ii) a subset of SQuAD-sr that corresponds exactly to the articles and QA pairs thatwere refined to create the SQuAD-sr-md. Evaluation of these models highlighted the TeslaXLM base model as a particularly promising candidate for future extractive QA training. The results also indicate that the refinement of the dataset positively impacts model performance.
64
 
65
  ### SQuAD-sr subset
66