Add library_name and update pipeline_tag
#1
by
nielsr HF Staff - opened
README.md
CHANGED
|
@@ -1,8 +1,11 @@
|
|
| 1 |
---
|
|
|
|
| 2 |
language:
|
| 3 |
- tr
|
| 4 |
- en
|
| 5 |
license: apache-2.0
|
|
|
|
|
|
|
| 6 |
tags:
|
| 7 |
- fill-mask
|
| 8 |
- turkish
|
|
@@ -12,12 +15,12 @@ tags:
|
|
| 12 |
- modernbert
|
| 13 |
- TRUBA
|
| 14 |
- MN5
|
| 15 |
-
base_model: ModernBERT-large
|
| 16 |
-
pipeline_tag: fill-mask
|
| 17 |
---
|
| 18 |
|
| 19 |
# Mursit-Large
|
| 20 |
|
|
|
|
|
|
|
| 21 |
[](https://github.com/newmindai/mecellem-models) [](https://huggingface.co/spaces/newmindai/Mizan) [](https://opensource.org/licenses/Apache-2.0)
|
| 22 |
|
| 23 |
## Model Description
|
|
@@ -85,7 +88,7 @@ The following table presents MLM accuracy scores (averaged across the 80-10-10 s
|
|
| 85 |
| KocLab-Bilkent/BERTurk-Legal | 54.10 |
|
| 86 |
| ytu-ce-cosmos/turkish-base-bert-uncased | 52.69 |
|
| 87 |
|
| 88 |
-
*MLM accuracy averaged across the 80-10-10 masking strategy.
|
| 89 |
|
| 90 |
## Performance on MTEB-Turkish Benchmark
|
| 91 |
|
|
@@ -178,6 +181,7 @@ with torch.no_grad():
|
|
| 178 |
score = predictions[0][idx].item()
|
| 179 |
print(f"{token}: {score:.4f}")
|
| 180 |
```
|
|
|
|
| 181 |
# ONNX Model Inference - Masked Language Modeling (MLM)
|
| 182 |
|
| 183 |
This script demonstrates how to use the ONNX model from Hugging Face for masked language modeling tasks.
|
|
@@ -253,12 +257,6 @@ for p in predictions:
|
|
| 253 |
- Question answering
|
| 254 |
- Feature extraction for downstream tasks
|
| 255 |
|
| 256 |
-
## Reproducibility
|
| 257 |
-
|
| 258 |
-
To reproduce the MLM benchmark results for this model, please refer to:
|
| 259 |
-
|
| 260 |
-
- **MLM Benchmark Results:** [github.com/newmindai/mecellem-models/benchmark/mlm](https://github.com/newmindai/mecellem-models/tree/main/benchmark/mlm) - Contains code and evaluation configurations for reproducing MLM accuracy scores on Turkish datasets using the 80-10-10 masking strategy.
|
| 261 |
-
|
| 262 |
## Acknowledgments
|
| 263 |
|
| 264 |
This work was supported by the EuroHPC Joint Undertaking through project etur46 with access to the MareNostrum 5 supercomputer, hosted by Barcelona Supercomputing Center (BSC), Spain. MareNostrum 5 is owned by EuroHPC JU and operated by BSC. We are grateful to the BSC support team for their assistance with job scheduling, environment configuration, and technical guidance throughout the project.
|
|
@@ -272,7 +270,7 @@ If you use this model, please cite our paper:
|
|
| 272 |
```bibtex
|
| 273 |
@article{mecellem2026,
|
| 274 |
title={Mecellem Models: Turkish Models Trained from Scratch and Continually Pre-trained for the Legal Domain},
|
| 275 |
-
author={Uğur, Özgür and Göksu, Mahmut and Çimen, Mahmut and Yılmaz, Musa and Şavirdi, Esra and Demir, Alp Talha and Güllüce, Rumeysa and
|
| 276 |
journal={arXiv preprint arXiv:2601.16018},
|
| 277 |
year={2026},
|
| 278 |
month={January},
|
|
@@ -283,6 +281,7 @@ If you use this model, please cite our paper:
|
|
| 283 |
primaryClass={cs.CL}
|
| 284 |
}
|
| 285 |
```
|
|
|
|
| 286 |
### Base Model References
|
| 287 |
|
| 288 |
```bibtex
|
|
@@ -292,6 +291,4 @@ If you use this model, please cite our paper:
|
|
| 292 |
booktitle={Proceedings of the 2025 Conference on Language Models},
|
| 293 |
year={2025}
|
| 294 |
}
|
| 295 |
-
```
|
| 296 |
-
|
| 297 |
-
<!-- Updated: 2026-01-15 09:38:24 -->
|
|
|
|
| 1 |
---
|
| 2 |
+
base_model: ModernBERT-large
|
| 3 |
language:
|
| 4 |
- tr
|
| 5 |
- en
|
| 6 |
license: apache-2.0
|
| 7 |
+
pipeline_tag: feature-extraction
|
| 8 |
+
library_name: transformers
|
| 9 |
tags:
|
| 10 |
- fill-mask
|
| 11 |
- turkish
|
|
|
|
| 15 |
- modernbert
|
| 16 |
- TRUBA
|
| 17 |
- MN5
|
|
|
|
|
|
|
| 18 |
---
|
| 19 |
|
| 20 |
# Mursit-Large
|
| 21 |
|
| 22 |
+
This model was introduced in the paper [Mecellem Models: Turkish Models Trained from Scratch and Continually Pre-trained for the Legal Domain](https://huggingface.co/papers/2601.16018).
|
| 23 |
+
|
| 24 |
[](https://github.com/newmindai/mecellem-models) [](https://huggingface.co/spaces/newmindai/Mizan) [](https://opensource.org/licenses/Apache-2.0)
|
| 25 |
|
| 26 |
## Model Description
|
|
|
|
| 88 |
| KocLab-Bilkent/BERTurk-Legal | 54.10 |
|
| 89 |
| ytu-ce-cosmos/turkish-base-bert-uncased | 52.69 |
|
| 90 |
|
| 91 |
+
*MLM accuracy averaged across the 80-10-10 masking strategy. Evaluation datasets: blackerx/turkish_v2, fthbrmnby/turkish_product_reviews, hazal/Turkish-Biomedical-corpus-trM, newmindai/EuroHPC-Legal. All experiments are reproducible (see Section A.2 in the paper).*
|
| 92 |
|
| 93 |
## Performance on MTEB-Turkish Benchmark
|
| 94 |
|
|
|
|
| 181 |
score = predictions[0][idx].item()
|
| 182 |
print(f"{token}: {score:.4f}")
|
| 183 |
```
|
| 184 |
+
|
| 185 |
# ONNX Model Inference - Masked Language Modeling (MLM)
|
| 186 |
|
| 187 |
This script demonstrates how to use the ONNX model from Hugging Face for masked language modeling tasks.
|
|
|
|
| 257 |
- Question answering
|
| 258 |
- Feature extraction for downstream tasks
|
| 259 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 260 |
## Acknowledgments
|
| 261 |
|
| 262 |
This work was supported by the EuroHPC Joint Undertaking through project etur46 with access to the MareNostrum 5 supercomputer, hosted by Barcelona Supercomputing Center (BSC), Spain. MareNostrum 5 is owned by EuroHPC JU and operated by BSC. We are grateful to the BSC support team for their assistance with job scheduling, environment configuration, and technical guidance throughout the project.
|
|
|
|
| 270 |
```bibtex
|
| 271 |
@article{mecellem2026,
|
| 272 |
title={Mecellem Models: Turkish Models Trained from Scratch and Continually Pre-trained for the Legal Domain},
|
| 273 |
+
author={Uğur, Özgür and Göksu, Mahmut and Çimen, Mahmut and Yılmaz, Musa and Şavirdi, Esra and Demir, Alp Talha and Güllüce, Rumeysa and İclal Çetin and Ömer Can Sağbaş},
|
| 274 |
journal={arXiv preprint arXiv:2601.16018},
|
| 275 |
year={2026},
|
| 276 |
month={January},
|
|
|
|
| 281 |
primaryClass={cs.CL}
|
| 282 |
}
|
| 283 |
```
|
| 284 |
+
|
| 285 |
### Base Model References
|
| 286 |
|
| 287 |
```bibtex
|
|
|
|
| 291 |
booktitle={Proceedings of the 2025 Conference on Language Models},
|
| 292 |
year={2025}
|
| 293 |
}
|
| 294 |
+
```
|
|
|
|
|
|