library_name: transformers
license: apache-2.0
base_model: bert-base-cased
tags:
- generated_from_trainer
metrics:
- precision
- recall
- f1
- accuracy
model-index:
- name: bert-finetuned-ner
results: []
pipeline_tag: token-classification
bert-finetuned-ner
This model is a fine-tuned version of bert-base-cased on unimelb-nlp/wikiann dataset English Language . It achieves the following results on the evaluation set:
- Loss: 0.2904
- Precision: 0.8249
- Recall: 0.8498
- F1: 0.8372
- Accuracy: 0.9311
Model description
This model is a BERT-based Named Entity Recognition (NER) system fine-tuned from bert-base-cased for English token classification.
It identifies and classifies named entities using the BIO tagging scheme across the following entity types:
PER (Person)
ORG (Organization)
LOC (Location)
O (Outside)
The model processes tokenized text and outputs entity spans using contextualized embeddings learned through transformer self-attention mechanisms.
Intended uses & limitations
Intended Uses
Information extraction from English text
Named entity recognition in NLP pipelines
Academic research and educational projects
Preprocessing step for downstream tasks (e.g., relation extraction, QA)
Limitations
Trained only on English data
Performance may degrade on domain-specific text (medical, legal, informal)
Limited to PER, ORG, LOC entity types
Sensitive to tokenization artifacts in noisy or misspelled text
Training and evaluation data
The model was trained and evaluated using the WikiAnn (PAN-X) dataset for English.
Dataset Details
Multilingual, Wikipedia-based NER corpus
Automatically annotated
BIO labeling scheme
Final Data Split
Training: 30,000 sentences
Validation: 5,000 sentences
Test: 5,000 sentences
Entity Labels
O, B-PER, I-PER, B-ORG, I-ORG, B-LOC, I-LOC
Training procedure
Base Model: bert-base-cased
Framework: Hugging Face Transformers
Task: Token Classification (NER)
Epochs: 3
Learning Rate: 2e-5
Optimizer: AdamW
Weight Decay: 0.01
Evaluation Metric: SeqEval (Precision, Recall, F1, Accuracy)
Label Alignment: Subword-aware BIO label propagation
Trainer API: Hugging Face Trainer
The model was evaluated after each epoch and achieved strong overall performance on the held-out test set.
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: linear
- num_epochs: 3
Training results
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|---|---|---|---|---|---|---|---|
| 0.2821 | 1.0 | 3750 | 0.2421 | 0.7914 | 0.8387 | 0.8143 | 0.9259 |
| 0.1919 | 2.0 | 7500 | 0.2524 | 0.8163 | 0.8433 | 0.8296 | 0.9289 |
| 0.1307 | 3.0 | 11250 | 0.2904 | 0.8249 | 0.8498 | 0.8372 | 0.9311 |
Framework versions
- Transformers 4.57.3
- Pytorch 2.9.0+cu126
- Datasets 4.0.0
- Tokenizers 0.22.1