MuRIL is fine-tuned on Tamil APTFiNER dataset for Fine-grained Named Entity Recognition.

The tagset of MultiCoNER2 is a fine-grained tagset. The fine to coarse level mapping of the tags are as follows:

  • Location (LOC) : Facility, OtherLOC, HumanSettlement, Station
  • Creative Work (CW) : VisualWork, MusicalWork, WrittenWork, ArtWork, Software
  • Group (GRP) : MusicalGRP, PublicCORP, PrivateCORP, AerospaceManufacturer, SportsGRP, CarManufacturer, ORG
  • Person (PER) : Scientist, Artist, Athlete, Politician, Cleric, SportsManager, OtherPER
  • Product (PROD) : Clothing, Vehicle, Food, Drink, OtherPROD
  • Medical (MED) : Medication/Vaccine, MedicalProcedure, AnatomicalStructure, Symptom, Disease

Model performance:

Precision: 58.96
Recall: 65.32
F1: 61.98

Training Parameters:

Epochs: 6
Optimizer: AdamW
Learning Rate: 5e-5
Weight Decay: 0.01
Batch Size: 64

Contributors

Prachuryya Kaushik
Adittya Gupta
Ajanta Maurya
Gautam Sharma
Prof. V Vijaya Saradhi
Prof. Ashish Anand

APTFiNER is a part of the AWED-FiNER ecosystem: Paper | GitHub | Interactive Demo

Sample Usage

You can use the AWED-FiNER agentic tool to interact with expert models trained using this framework. Below is an example using the smolagents library:

from smolagents import CodeAgent, HfApiModel
from tool import AWEDFiNERTool

# Initialize the expert tool
ner_tool = AWEDFiNERTool()

# Initialize the agent (using a model of your choice)
agent = CodeAgent(tools=[ner_tool], model=HfApiModel())

# The agent will automatically use AWED-FiNER for specialized NER
# Case: Processing a vulnerable language (Bodo)
agent.run("Recognize the named entities in this Bodo sentence: 'बिथाङा दिल्लियाव थाङो।'")

Citation

If you use this model, please cite the following papers:

@inproceedings{kaushik2026aptfiner,
  title={APTFiNER: Annotation Preserving Translation for Fine-grained Named Entity Recognition},
  author={Kaushik, Prachuryya and Gupta, Adittya and Maurya, Ajanta and Sharma, Gautam and Saradhi, Vijaya V and Anand, Ashish},
  booktitle={Proceedings of the Fifteenth Language Resources and Evaluation Conference},
  volume={15},
  year={2026}
}

@inproceedings{kaushik2026sampurner,
  title={SampurNER: Fine-grained Named Entity Recognition Dataset for 22 Indian Languages},
  author={Kaushik, Prachuryya and Anand, Ashish},
  booktitle={Proceedings of the AAAI Conference on Artificial Intelligence},
  volume={40},
  year={2026}
}

@inproceedings{fetahu2023multiconer,
  title={MultiCoNER v2: a Large Multilingual dataset for Fine-grained and Noisy Named Entity Recognition},
  author={Fetahu, Besnik and Chen, Zhiyu and Kar, Sudipta and Rokhlenko, Oleg and Malmasi, Shervin},
  booktitle={Findings of the Association for Computational Linguistics: EMNLP 2023},
  pages={2027--2051},
  year={2023}
}
Downloads last month
27
Safetensors
Model size
0.6B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for prachuryyaIITG/APTFiNER_Tamil_MuRIL

Finetuned
(25)
this model

Dataset used to train prachuryyaIITG/APTFiNER_Tamil_MuRIL

Space using prachuryyaIITG/APTFiNER_Tamil_MuRIL 1

Collections including prachuryyaIITG/APTFiNER_Tamil_MuRIL

Paper for prachuryyaIITG/APTFiNER_Tamil_MuRIL