Add library_name to metadata and improve model card documentation
#1
by
nielsr HF Staff - opened
README.md
CHANGED
|
@@ -1,24 +1,27 @@
|
|
| 1 |
---
|
| 2 |
-
|
|
|
|
| 3 |
datasets:
|
| 4 |
- MultiCoNER/multiconer_v2
|
| 5 |
language:
|
| 6 |
- de
|
|
|
|
| 7 |
metrics:
|
| 8 |
- f1
|
| 9 |
- precision
|
| 10 |
- recall
|
| 11 |
-
base_model:
|
| 12 |
-
- FacebookAI/xlm-roberta-large
|
| 13 |
pipeline_tag: token-classification
|
| 14 |
tags:
|
| 15 |
- NER
|
| 16 |
- Named_Entity_Recognition
|
| 17 |
pretty_name: MultiCoNER2 German XLM-RoBERTa
|
|
|
|
| 18 |
---
|
| 19 |
|
| 20 |
**XLM-RoBERTa is fine-tuned on German [MultiCoNER2](https://huggingface.co/datasets/MultiCoNER/multiconer_v2) dataset for Fine-grained Named Entity Recognition.**
|
| 21 |
|
|
|
|
|
|
|
| 22 |
The tagset of [MultiCoNER2](https://huggingface.co/datasets/MultiCoNER/multiconer_v2) is a fine-grained tagset. The fine to coarse level mapping of the tags are as follows:
|
| 23 |
|
| 24 |
* Location (LOC) : Facility, OtherLOC, HumanSettlement, Station
|
|
@@ -93,4 +96,5 @@ If you use this model, please cite the following papers:
|
|
| 93 |
booktitle={Proceedings of the AAAI Conference on Artificial Intelligence},
|
| 94 |
volume={40},
|
| 95 |
year={2026}
|
| 96 |
-
}
|
|
|
|
|
|
| 1 |
---
|
| 2 |
+
base_model:
|
| 3 |
+
- FacebookAI/xlm-roberta-large
|
| 4 |
datasets:
|
| 5 |
- MultiCoNER/multiconer_v2
|
| 6 |
language:
|
| 7 |
- de
|
| 8 |
+
license: mit
|
| 9 |
metrics:
|
| 10 |
- f1
|
| 11 |
- precision
|
| 12 |
- recall
|
|
|
|
|
|
|
| 13 |
pipeline_tag: token-classification
|
| 14 |
tags:
|
| 15 |
- NER
|
| 16 |
- Named_Entity_Recognition
|
| 17 |
pretty_name: MultiCoNER2 German XLM-RoBERTa
|
| 18 |
+
library_name: transformers
|
| 19 |
---
|
| 20 |
|
| 21 |
**XLM-RoBERTa is fine-tuned on German [MultiCoNER2](https://huggingface.co/datasets/MultiCoNER/multiconer_v2) dataset for Fine-grained Named Entity Recognition.**
|
| 22 |
|
| 23 |
+
This model is part of the [AWED-FiNER](https://huggingface.co/papers/2601.10161) framework, providing fine-grained NER solutions.
|
| 24 |
+
|
| 25 |
The tagset of [MultiCoNER2](https://huggingface.co/datasets/MultiCoNER/multiconer_v2) is a fine-grained tagset. The fine to coarse level mapping of the tags are as follows:
|
| 26 |
|
| 27 |
* Location (LOC) : Facility, OtherLOC, HumanSettlement, Station
|
|
|
|
| 96 |
booktitle={Proceedings of the AAAI Conference on Artificial Intelligence},
|
| 97 |
volume={40},
|
| 98 |
year={2026}
|
| 99 |
+
}
|
| 100 |
+
```
|