bniladridas commited on
Commit
35b9137
·
1 Parent(s): 6b3acc4

Rebrand to harpertokenNER and clean up README

Browse files
Files changed (1) hide show
  1. README.md +3 -18
README.md CHANGED
@@ -13,7 +13,7 @@ pipeline_tag: token-classification
13
  library_name: transformers
14
  base_model: bert-base-uncased
15
  model-index:
16
- - name: token-classification-ai-fine-tune
17
  results:
18
  - task:
19
  type: token-classification
@@ -29,15 +29,13 @@ widget:
29
  - text: "Apple is buying a U.K. startup for $1 billion"
30
  ---
31
 
32
- # token-classification-ai-fine-tune
33
-
34
- [![Hugging Face](https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-Model-blue)](https://huggingface.co/bniladridas/token-classification-ai-fine-tune)
35
 
36
  This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the [CoNLL-2003](https://huggingface.co/datasets/conll2003) dataset. It achieves a validation loss of **0.0474** on the evaluation set.
37
 
38
  ## Model Description
39
 
40
- This is a token classification model fine-tuned for **Named Entity Recognition (NER)**, built on the `bert-base-uncased` architecture. It’s crafted to identify entities (like people, organizations, and locations) in text, optimized here for CPU accessibility. Uploaded by [bniladridas](https://huggingface.co/bniladridas), it delivers strong NER performance on the CoNLL-2003 benchmark. For a GPU-accelerated version with CUDA support, see the [GitHub repository](https://github.com/bniladridas/token-classification-ai-fine-tune).
41
 
42
  ## Intended Uses & Limitations
43
 
@@ -84,17 +82,4 @@ The following hyperparameters were used during training:
84
  - **Datasets**: 1.18.3
85
  - **Tokenizers**: 0.13.3
86
 
87
- ### Additional Notes
88
- This version is optimized for CPU use with these intentional adjustments:
89
- 1. **Full-precision training**: Swapped out fp16 for broader compatibility
90
- 2. **Streamlined batch sizes**: Set to 8 for efficient CPU processing
91
- 3. **Simplified workflow**: Skipped gradient accumulation for smoother CPU runs
92
- 4. **Full feature set**: Retained all monitoring (e.g., TensorBoard) and saving capabilities
93
-
94
- For the GPU version with CUDA, mixed precision, and gradient accumulation, check out the [GitHub repository](https://github.com/bniladridas/token-classification-ai-fine-tune). To clone it, run:
95
-
96
- ```bash
97
- git clone https://github.com/bniladridas/token-classification-ai-fine-tune.git
98
- ```
99
 
100
- This model was pushed to the Hugging Face Hub for easy CPU-based deployment.
 
13
  library_name: transformers
14
  base_model: bert-base-uncased
15
  model-index:
16
+ - name: harpertokenNER
17
  results:
18
  - task:
19
  type: token-classification
 
29
  - text: "Apple is buying a U.K. startup for $1 billion"
30
  ---
31
 
32
+ # harpertokenNER
 
 
33
 
34
  This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the [CoNLL-2003](https://huggingface.co/datasets/conll2003) dataset. It achieves a validation loss of **0.0474** on the evaluation set.
35
 
36
  ## Model Description
37
 
38
+ This is a token classification model fine-tuned for **Named Entity Recognition (NER)** on the CoNLL-2003 dataset, built on the `bert-base-uncased` architecture. It identifies entities like people, organizations, and locations in text. Optimized for CPU use. Uploaded by [harpertoken](https://huggingface.co/harpertoken).
39
 
40
  ## Intended Uses & Limitations
41
 
 
82
  - **Datasets**: 1.18.3
83
  - **Tokenizers**: 0.13.3
84
 
 
 
 
 
 
 
 
 
 
 
 
 
85