Model Card for cobald-parser-pretrain-en
A transformer-based multihead parser for CoBaLD annotation.
This model parses a pre-tokenized CoNLL-U text and jointly labels each token with three tiers of tags:
- Grammatical tags (lemma, UPOS, XPOS, morphological features),
- Syntactic tags (basic and enhanced Universal Dependencies),
- Semantic tags (deep slot and semantic class).
Model Sources
- Repository: https://github.com/CobaldAnnotation/CobaldParser
- Paper: https://dialogue-conf.org/wp-content/uploads/2025/04/BaiukIBaiukAPetrovaM.009.pdf
Citation
@inproceedings{baiuk2025cobald,
title={CoBaLD Parser: Joint Morphosyntactic and Semantic Annotation},
author={Baiuk, Ilia and Baiuk, Alexandra and Petrova, Maria},
booktitle={Proceedings of the International Conference "Dialogue"},
volume={I},
year={2025}
}
- Downloads last month
- 8
Model tree for CoBaLD/cobald-parser-pretrain-en
Base model
FacebookAI/xlm-roberta-baseDataset used to train CoBaLD/cobald-parser-pretrain-en
Evaluation results
- Null F1 on enhanced-ud-syntaxvalidation set self-reported0.250
- Ud Jaccard on enhanced-ud-syntaxvalidation set self-reported0.843
- Eud Jaccard on enhanced-ud-syntaxvalidation set self-reported0.786