Datasets:
Link to AWED-FiNER paper and GitHub repository, add sample usage (#2)
Browse files- Link to AWED-FiNER paper and GitHub repository, add sample usage (fe249c2dc0d21f4e974ba1032bc32cfd710d988d)
Co-authored-by: Niels Rogge <nielsr@users.noreply.huggingface.co>
README.md
CHANGED
|
@@ -1,21 +1,24 @@
|
|
| 1 |
---
|
| 2 |
-
license: mit
|
| 3 |
-
task_categories:
|
| 4 |
-
- token-classification
|
| 5 |
language:
|
| 6 |
- as
|
| 7 |
- brx
|
| 8 |
- mr
|
| 9 |
- ne
|
| 10 |
- sa
|
| 11 |
-
|
| 12 |
size_categories:
|
| 13 |
- 10M<n<100M
|
|
|
|
|
|
|
|
|
|
|
|
|
| 14 |
---
|
| 15 |
|
| 16 |
-
#
|
|
|
|
|
|
|
| 17 |
|
| 18 |
-
**
|
| 19 |
|
| 20 |
Utilizing CLASSER, fine-grained named entity recognition dataset is created in five languages: Assamese (as), Bodo (brx), Marathi (mr), Nepali (ne) and Sanskrit (sa).
|
| 21 |
|
|
@@ -25,6 +28,25 @@ Utilizing CLASSER, fine-grained named entity recognition dataset is created in f
|
|
| 25 |
|
| 26 |
*Figure: Overview of the CLASSER framework.*
|
| 27 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 28 |
## CLASSER Dataset Statistics
|
| 29 |
|
| 30 |
<table>
|
|
@@ -79,9 +101,18 @@ Utilizing CLASSER, fine-grained named entity recognition dataset is created in f
|
|
| 79 |
|
| 80 |
## Citation
|
| 81 |
|
| 82 |
-
If you use this dataset, please cite the following
|
| 83 |
|
| 84 |
```bibtex
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 85 |
@inproceedings{kaushik2025classer,
|
| 86 |
title = {{CLASSER}: Cross-lingual Annotation Projection enhancement through Script Similarity for Fine-grained Named Entity Recognition},
|
| 87 |
author = {Kaushik, Prachuryya and Anand, Ashish},
|
|
@@ -90,3 +121,4 @@ If you use this dataset, please cite the following paper:
|
|
| 90 |
publisher = {Association for Computational Linguistics},
|
| 91 |
note = {Main conference paper}
|
| 92 |
}
|
|
|
|
|
|
| 1 |
---
|
|
|
|
|
|
|
|
|
|
| 2 |
language:
|
| 3 |
- as
|
| 4 |
- brx
|
| 5 |
- mr
|
| 6 |
- ne
|
| 7 |
- sa
|
| 8 |
+
license: mit
|
| 9 |
size_categories:
|
| 10 |
- 10M<n<100M
|
| 11 |
+
task_categories:
|
| 12 |
+
- token-classification
|
| 13 |
+
- other
|
| 14 |
+
pretty_name: CLASSER
|
| 15 |
---
|
| 16 |
|
| 17 |
+
# CLASSER: Cross-lingual Annotation Projection enhancement through Script Similarity for Fine-grained Named Entity Recognition
|
| 18 |
+
|
| 19 |
+
**CLASSER** is a framework for cross-lingual annotation projection with script-similarity-based refinement to create high-quality fine-grained named entity recognition datasets. It is part of the [AWED-FiNER](https://github.com/PrachuryyaKaushik/AWED-FiNER) ecosystem.
|
| 20 |
|
| 21 |
+
[**Paper**](https://huggingface.co/papers/2601.10161) | [**GitHub**](https://github.com/PrachuryyaKaushik/AWED-FiNER) | [**Interactive Demo**](https://huggingface.co/spaces/prachuryyaIITG/AWED-FiNER)
|
| 22 |
|
| 23 |
Utilizing CLASSER, fine-grained named entity recognition dataset is created in five languages: Assamese (as), Bodo (brx), Marathi (mr), Nepali (ne) and Sanskrit (sa).
|
| 24 |
|
|
|
|
| 28 |
|
| 29 |
*Figure: Overview of the CLASSER framework.*
|
| 30 |
|
| 31 |
+
## Sample Usage
|
| 32 |
+
|
| 33 |
+
You can use the AWED-FiNER agentic tool to interact with expert models trained using this framework. Below is an example using the `smolagents` library:
|
| 34 |
+
|
| 35 |
+
```python
|
| 36 |
+
from smolagents import CodeAgent, HfApiModel
|
| 37 |
+
from tool import AWEDFiNERTool
|
| 38 |
+
|
| 39 |
+
# Initialize the expert tool
|
| 40 |
+
ner_tool = AWEDFiNERTool()
|
| 41 |
+
|
| 42 |
+
# Initialize the agent (using a model of your choice)
|
| 43 |
+
agent = CodeAgent(tools=[ner_tool], model=HfApiModel())
|
| 44 |
+
|
| 45 |
+
# The agent will automatically use AWED-FiNER for specialized NER
|
| 46 |
+
# Case: Processing a vulnerable language (Bodo)
|
| 47 |
+
agent.run("Recognize the named entities in this Bodo sentence: 'बिथाङा दिल्लियाव थाङो।'")
|
| 48 |
+
```
|
| 49 |
+
|
| 50 |
## CLASSER Dataset Statistics
|
| 51 |
|
| 52 |
<table>
|
|
|
|
| 101 |
|
| 102 |
## Citation
|
| 103 |
|
| 104 |
+
If you use this dataset, please cite the following papers:
|
| 105 |
|
| 106 |
```bibtex
|
| 107 |
+
@misc{kaushik2026awedfiner,
|
| 108 |
+
title = {AWED-FiNER: Agents, Web Applications, and Expert Detectors for Fine-grained Named Entity Recognition across 36 Languages for 6.6 Billion Speakers},
|
| 109 |
+
author = {Kaushik, Prachuryya and Anand, Ashish},
|
| 110 |
+
year = {2026},
|
| 111 |
+
note = {arXiv preprint, submitted},
|
| 112 |
+
archivePrefix= {arXiv},
|
| 113 |
+
eprint = {submit/7163987}
|
| 114 |
+
}
|
| 115 |
+
|
| 116 |
@inproceedings{kaushik2025classer,
|
| 117 |
title = {{CLASSER}: Cross-lingual Annotation Projection enhancement through Script Similarity for Fine-grained Named Entity Recognition},
|
| 118 |
author = {Kaushik, Prachuryya and Anand, Ashish},
|
|
|
|
| 121 |
publisher = {Association for Computational Linguistics},
|
| 122 |
note = {Main conference paper}
|
| 123 |
}
|
| 124 |
+
```
|