# Custom BERT NER Model This repository contains a BERT-based Named Entity Recognition (NER) model fine-tuned on the CoNLL-2003 dataset. The model is trained to identify common named entity types such as persons, organizations, locations, and miscellaneous entities. --- ## Model Details - **Model architecture:** BERT (bert-base-cased) - **Task:** Token classification / Named Entity Recognition (NER) - **Training data:** CoNLL-2003 dataset (~14,000 training samples) - **Number of epochs:** 5 - **Framework:** Hugging Face Transformers + Datasets - **Device:** CUDA-enabled GPU for training and inference - **WandB:** Disabled during training --- ## Usage You can use this model for token classification to identify named entities in your text. ### Installation ```python pip install transformers datasets torch ``` ## Load the model and tokenizer ```pyhton from transformers import BertTokenizerFast, BertForTokenClassification import torch model_name_or_path = "AventIQ-AI/Custom-BERT-NER-Model" tokenizer = BertTokenizerFast.from_pretrained(model_name_or_path) model = BertForTokenClassification.from_pretrained(model_name_or_path) model.to("cuda") # or "cpu" model.eval() ``` ## Example inference ```python text = "Hi, I am Deepak and I am living in Delhi." tokens = tokenizer(text, return_tensors="pt").to(model.device) outputs = model(**tokens) predictions = torch.argmax(outputs.logits, dim=2) labels = [model.config.id2label[p.item()] for p in predictions[0]] for token, label in zip(tokenizer.tokenize(text), labels): print(f"{token}: {label}") ``` ## Training Details - Dataset: CoNLL-2003, loaded via the Hugging Face datasets library - Optimizer: AdamW - Learning Rate: 5e-5 - Batch Size: 16 - Max Sequence Length: 128 - Epochs: 5 - Evaluation: Performed on validation split (if applicable) - Quantization: Applied post-training for model size reduction (optional) ## Limitations - The model may not generalize well to unseen entity types or domains outside CoNLL-2003. - It can occasionally mislabel entities, especially for rare or new names. - A CUDA-enabled GPU is required for efficient training and inference.