Add library_name and update pipeline_tag

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +10 -13
README.md CHANGED
@@ -1,8 +1,11 @@
1
  ---
 
2
  language:
3
  - tr
4
  - en
5
  license: apache-2.0
 
 
6
  tags:
7
  - fill-mask
8
  - turkish
@@ -12,12 +15,12 @@ tags:
12
  - modernbert
13
  - TRUBA
14
  - MN5
15
- base_model: ModernBERT-large
16
- pipeline_tag: fill-mask
17
  ---
18
 
19
  # Mursit-Large
20
 
 
 
21
  [![GitHub](https://img.shields.io/badge/GitHub-NewMindAI-black?logo=github)](https://github.com/newmindai/mecellem-models) [![HuggingFace Space](https://img.shields.io/badge/HF%20Space-Mizan-blue?logo=huggingface)](https://huggingface.co/spaces/newmindai/Mizan) [![License](https://img.shields.io/badge/License-Apache%202.0-blue.svg)](https://opensource.org/licenses/Apache-2.0)
22
 
23
  ## Model Description
@@ -85,7 +88,7 @@ The following table presents MLM accuracy scores (averaged across the 80-10-10 s
85
  | KocLab-Bilkent/BERTurk-Legal | 54.10 |
86
  | ytu-ce-cosmos/turkish-base-bert-uncased | 52.69 |
87
 
88
- *MLM accuracy averaged across the 80-10-10 masking strategy. turkish-base-bert-uncased was evaluated only on uncased datasets. Evaluation datasets: blackerx/turkish_v2, fthbrmnby/turkish_product_reviews, hazal/Turkish-Biomedical-corpus-trM, newmindai/EuroHPC-Legal. All experiments are reproducible (see Section A.2 in the paper).*
89
 
90
  ## Performance on MTEB-Turkish Benchmark
91
 
@@ -178,6 +181,7 @@ with torch.no_grad():
178
  score = predictions[0][idx].item()
179
  print(f"{token}: {score:.4f}")
180
  ```
 
181
  # ONNX Model Inference - Masked Language Modeling (MLM)
182
 
183
  This script demonstrates how to use the ONNX model from Hugging Face for masked language modeling tasks.
@@ -253,12 +257,6 @@ for p in predictions:
253
  - Question answering
254
  - Feature extraction for downstream tasks
255
 
256
- ## Reproducibility
257
-
258
- To reproduce the MLM benchmark results for this model, please refer to:
259
-
260
- - **MLM Benchmark Results:** [github.com/newmindai/mecellem-models/benchmark/mlm](https://github.com/newmindai/mecellem-models/tree/main/benchmark/mlm) - Contains code and evaluation configurations for reproducing MLM accuracy scores on Turkish datasets using the 80-10-10 masking strategy.
261
-
262
  ## Acknowledgments
263
 
264
  This work was supported by the EuroHPC Joint Undertaking through project etur46 with access to the MareNostrum 5 supercomputer, hosted by Barcelona Supercomputing Center (BSC), Spain. MareNostrum 5 is owned by EuroHPC JU and operated by BSC. We are grateful to the BSC support team for their assistance with job scheduling, environment configuration, and technical guidance throughout the project.
@@ -272,7 +270,7 @@ If you use this model, please cite our paper:
272
  ```bibtex
273
  @article{mecellem2026,
274
  title={Mecellem Models: Turkish Models Trained from Scratch and Continually Pre-trained for the Legal Domain},
275
- author={Uğur, Özgür and Göksu, Mahmut and Çimen, Mahmut and Yılmaz, Musa and Şavirdi, Esra and Demir, Alp Talha and Güllüce, Rumeysa and Çetin, İclal and Sağbaş, Ömer Can},
276
  journal={arXiv preprint arXiv:2601.16018},
277
  year={2026},
278
  month={January},
@@ -283,6 +281,7 @@ If you use this model, please cite our paper:
283
  primaryClass={cs.CL}
284
  }
285
  ```
 
286
  ### Base Model References
287
 
288
  ```bibtex
@@ -292,6 +291,4 @@ If you use this model, please cite our paper:
292
  booktitle={Proceedings of the 2025 Conference on Language Models},
293
  year={2025}
294
  }
295
- ```
296
-
297
- <!-- Updated: 2026-01-15 09:38:24 -->
 
1
  ---
2
+ base_model: ModernBERT-large
3
  language:
4
  - tr
5
  - en
6
  license: apache-2.0
7
+ pipeline_tag: feature-extraction
8
+ library_name: transformers
9
  tags:
10
  - fill-mask
11
  - turkish
 
15
  - modernbert
16
  - TRUBA
17
  - MN5
 
 
18
  ---
19
 
20
  # Mursit-Large
21
 
22
+ This model was introduced in the paper [Mecellem Models: Turkish Models Trained from Scratch and Continually Pre-trained for the Legal Domain](https://huggingface.co/papers/2601.16018).
23
+
24
  [![GitHub](https://img.shields.io/badge/GitHub-NewMindAI-black?logo=github)](https://github.com/newmindai/mecellem-models) [![HuggingFace Space](https://img.shields.io/badge/HF%20Space-Mizan-blue?logo=huggingface)](https://huggingface.co/spaces/newmindai/Mizan) [![License](https://img.shields.io/badge/License-Apache%202.0-blue.svg)](https://opensource.org/licenses/Apache-2.0)
25
 
26
  ## Model Description
 
88
  | KocLab-Bilkent/BERTurk-Legal | 54.10 |
89
  | ytu-ce-cosmos/turkish-base-bert-uncased | 52.69 |
90
 
91
+ *MLM accuracy averaged across the 80-10-10 masking strategy. Evaluation datasets: blackerx/turkish_v2, fthbrmnby/turkish_product_reviews, hazal/Turkish-Biomedical-corpus-trM, newmindai/EuroHPC-Legal. All experiments are reproducible (see Section A.2 in the paper).*
92
 
93
  ## Performance on MTEB-Turkish Benchmark
94
 
 
181
  score = predictions[0][idx].item()
182
  print(f"{token}: {score:.4f}")
183
  ```
184
+
185
  # ONNX Model Inference - Masked Language Modeling (MLM)
186
 
187
  This script demonstrates how to use the ONNX model from Hugging Face for masked language modeling tasks.
 
257
  - Question answering
258
  - Feature extraction for downstream tasks
259
 
 
 
 
 
 
 
260
  ## Acknowledgments
261
 
262
  This work was supported by the EuroHPC Joint Undertaking through project etur46 with access to the MareNostrum 5 supercomputer, hosted by Barcelona Supercomputing Center (BSC), Spain. MareNostrum 5 is owned by EuroHPC JU and operated by BSC. We are grateful to the BSC support team for their assistance with job scheduling, environment configuration, and technical guidance throughout the project.
 
270
  ```bibtex
271
  @article{mecellem2026,
272
  title={Mecellem Models: Turkish Models Trained from Scratch and Continually Pre-trained for the Legal Domain},
273
+ author={Uğur, Özgür and Göksu, Mahmut and Çimen, Mahmut and Yılmaz, Musa and Şavirdi, Esra and Demir, Alp Talha and Güllüce, Rumeysa and İclal Çetin and Ömer Can Sağbaş},
274
  journal={arXiv preprint arXiv:2601.16018},
275
  year={2026},
276
  month={January},
 
281
  primaryClass={cs.CL}
282
  }
283
  ```
284
+
285
  ### Base Model References
286
 
287
  ```bibtex
 
291
  booktitle={Proceedings of the 2025 Conference on Language Models},
292
  year={2025}
293
  }
294
+ ```