nielsr HF Staff commited on
Commit
6d3977b
·
verified ·
1 Parent(s): 70c8494

Add library_name metadata

Browse files

Hi! I'm Niels, part of the community science team at Hugging Face.

This PR improves the model card by adding the `library_name: transformers` metadata tag. This ensures that the model is correctly identified as a Transformers model, enabling the automated code snippets and the inference widget on the Hub. I've also updated the citation to reflect the published arXiv ID.

Files changed (1) hide show
  1. README.md +15 -12
README.md CHANGED
@@ -1,15 +1,16 @@
1
  ---
2
- license: mit
 
3
  datasets:
4
  - MultiCoNER/multiconer_v2
5
  language:
6
  - fr
 
 
7
  metrics:
8
  - f1
9
  - precision
10
  - recall
11
- base_model:
12
- - FacebookAI/xlm-roberta-large
13
  pipeline_tag: token-classification
14
  tags:
15
  - NER
@@ -69,6 +70,16 @@ print(result)
69
  If you use this model, please cite the following papers:
70
 
71
  ```bibtex
 
 
 
 
 
 
 
 
 
 
72
  @inproceedings{fetahu2023multiconer,
73
  title={MultiCoNER v2: a Large Multilingual dataset for Fine-grained and Noisy Named Entity Recognition},
74
  author={Fetahu, Besnik and Chen, Zhiyu and Kar, Sudipta and Rokhlenko, Oleg and Malmasi, Shervin},
@@ -77,15 +88,6 @@ If you use this model, please cite the following papers:
77
  year={2023}
78
  }
79
 
80
- @misc{kaushik2026awedfiner,
81
- title = {AWED-FiNER: Agents, Web Applications, and Expert Detectors for Fine-grained Named Entity Recognition across 36 Languages for 6.6 Billion Speakers},
82
- author = {Kaushik, Prachuryya and Anand, Ashish},
83
- year = {2026},
84
- note = {arXiv preprint, submitted},
85
- archivePrefix= {arXiv},
86
- eprint = {submit/7163987}
87
- }
88
-
89
  @inproceedings{kaushik2026sampurner,
90
  title={SampurNER: Fine-grained Named Entity Recognition Dataset for 22 Indian Languages},
91
  author={Kaushik, Prachuryya and Anand, Ashish},
@@ -93,3 +95,4 @@ If you use this model, please cite the following papers:
93
  volume={40},
94
  year={2026}
95
  }
 
 
1
  ---
2
+ base_model:
3
+ - FacebookAI/xlm-roberta-large
4
  datasets:
5
  - MultiCoNER/multiconer_v2
6
  language:
7
  - fr
8
+ license: mit
9
+ library_name: transformers
10
  metrics:
11
  - f1
12
  - precision
13
  - recall
 
 
14
  pipeline_tag: token-classification
15
  tags:
16
  - NER
 
70
  If you use this model, please cite the following papers:
71
 
72
  ```bibtex
73
+ @misc{kaushik2026awedfiner,
74
+ title={AWED-FiNER: Agents, Web applications, and Expert Detectors for Fine-grained Named Entity Recognition across 36 Languages for 6.6 Billion Speakers},
75
+ author={Prachuryya Kaushik and Ashish Anand},
76
+ year={2026},
77
+ eprint={2601.10161},
78
+ archivePrefix={arXiv},
79
+ primaryClass={cs.CL},
80
+ url={https://arxiv.org/abs/2601.10161},
81
+ }
82
+
83
  @inproceedings{fetahu2023multiconer,
84
  title={MultiCoNER v2: a Large Multilingual dataset for Fine-grained and Noisy Named Entity Recognition},
85
  author={Fetahu, Besnik and Chen, Zhiyu and Kar, Sudipta and Rokhlenko, Oleg and Malmasi, Shervin},
 
88
  year={2023}
89
  }
90
 
 
 
 
 
 
 
 
 
 
91
  @inproceedings{kaushik2026sampurner,
92
  title={SampurNER: Fine-grained Named Entity Recognition Dataset for 22 Indian Languages},
93
  author={Kaushik, Prachuryya and Anand, Ashish},
 
95
  volume={40},
96
  year={2026}
97
  }
98
+ ```