MuRIL is fine-tuned on Telugu APTFiNER dataset for Fine-grained Named Entity Recognition.

The tagset of MultiCoNER2 is a fine-grained tagset. The fine to coarse level mapping of the tags are as follows:

  • Location (LOC) : Facility, OtherLOC, HumanSettlement, Station
  • Creative Work (CW) : VisualWork, MusicalWork, WrittenWork, ArtWork, Software
  • Group (GRP) : MusicalGRP, PublicCORP, PrivateCORP, AerospaceManufacturer, SportsGRP, CarManufacturer, ORG
  • Person (PER) : Scientist, Artist, Athlete, Politician, Cleric, SportsManager, OtherPER
  • Product (PROD) : Clothing, Vehicle, Food, Drink, OtherPROD
  • Medical (MED) : Medication/Vaccine, MedicalProcedure, AnatomicalStructure, Symptom, Disease

Model performance:

Precision: 75.72
Recall: 77.26
F1: 76.49

Training Parameters:

Epochs: 6
Optimizer: AdamW
Learning Rate: 5e-5
Weight Decay: 0.01
Batch Size: 64

Citation

If you use this model, please cite the following papers:

@inproceedings{kaushik2026sampurner,
  title={SampurNER: Fine-grained Named Entity Recognition Dataset for 22 Indian Languages},
  author={Kaushik, Prachuryya and Anand, Ashish},
  booktitle={Proceedings of the AAAI Conference on Artificial Intelligence},
  volume={40},
  year={2026}
}

@inproceedings{fetahu2023multiconer,
  title={MultiCoNER v2: a Large Multilingual dataset for Fine-grained and Noisy Named Entity Recognition},
  author={Fetahu, Besnik and Chen, Zhiyu and Kar, Sudipta and Rokhlenko, Oleg and Malmasi, Shervin},
  booktitle={Findings of the Association for Computational Linguistics: EMNLP 2023},
  pages={2027--2051},
  year={2023}
}
Downloads last month
13
Safetensors
Model size
0.5B params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for prachuryyaIITG/APTFiNER_Telugu_MuRIL

Finetuned
(20)
this model

Space using prachuryyaIITG/APTFiNER_Telugu_MuRIL 1