bniladridas
commited on
Commit
·
35b9137
1
Parent(s):
6b3acc4
Rebrand to harpertokenNER and clean up README
Browse files
README.md
CHANGED
|
@@ -13,7 +13,7 @@ pipeline_tag: token-classification
|
|
| 13 |
library_name: transformers
|
| 14 |
base_model: bert-base-uncased
|
| 15 |
model-index:
|
| 16 |
-
- name:
|
| 17 |
results:
|
| 18 |
- task:
|
| 19 |
type: token-classification
|
|
@@ -29,15 +29,13 @@ widget:
|
|
| 29 |
- text: "Apple is buying a U.K. startup for $1 billion"
|
| 30 |
---
|
| 31 |
|
| 32 |
-
#
|
| 33 |
-
|
| 34 |
-
[](https://huggingface.co/bniladridas/token-classification-ai-fine-tune)
|
| 35 |
|
| 36 |
This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the [CoNLL-2003](https://huggingface.co/datasets/conll2003) dataset. It achieves a validation loss of **0.0474** on the evaluation set.
|
| 37 |
|
| 38 |
## Model Description
|
| 39 |
|
| 40 |
-
This is a token classification model fine-tuned for **Named Entity Recognition (NER)
|
| 41 |
|
| 42 |
## Intended Uses & Limitations
|
| 43 |
|
|
@@ -84,17 +82,4 @@ The following hyperparameters were used during training:
|
|
| 84 |
- **Datasets**: 1.18.3
|
| 85 |
- **Tokenizers**: 0.13.3
|
| 86 |
|
| 87 |
-
### Additional Notes
|
| 88 |
-
This version is optimized for CPU use with these intentional adjustments:
|
| 89 |
-
1. **Full-precision training**: Swapped out fp16 for broader compatibility
|
| 90 |
-
2. **Streamlined batch sizes**: Set to 8 for efficient CPU processing
|
| 91 |
-
3. **Simplified workflow**: Skipped gradient accumulation for smoother CPU runs
|
| 92 |
-
4. **Full feature set**: Retained all monitoring (e.g., TensorBoard) and saving capabilities
|
| 93 |
-
|
| 94 |
-
For the GPU version with CUDA, mixed precision, and gradient accumulation, check out the [GitHub repository](https://github.com/bniladridas/token-classification-ai-fine-tune). To clone it, run:
|
| 95 |
-
|
| 96 |
-
```bash
|
| 97 |
-
git clone https://github.com/bniladridas/token-classification-ai-fine-tune.git
|
| 98 |
-
```
|
| 99 |
|
| 100 |
-
This model was pushed to the Hugging Face Hub for easy CPU-based deployment.
|
|
|
|
| 13 |
library_name: transformers
|
| 14 |
base_model: bert-base-uncased
|
| 15 |
model-index:
|
| 16 |
+
- name: harpertokenNER
|
| 17 |
results:
|
| 18 |
- task:
|
| 19 |
type: token-classification
|
|
|
|
| 29 |
- text: "Apple is buying a U.K. startup for $1 billion"
|
| 30 |
---
|
| 31 |
|
| 32 |
+
# harpertokenNER
|
|
|
|
|
|
|
| 33 |
|
| 34 |
This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the [CoNLL-2003](https://huggingface.co/datasets/conll2003) dataset. It achieves a validation loss of **0.0474** on the evaluation set.
|
| 35 |
|
| 36 |
## Model Description
|
| 37 |
|
| 38 |
+
This is a token classification model fine-tuned for **Named Entity Recognition (NER)** on the CoNLL-2003 dataset, built on the `bert-base-uncased` architecture. It identifies entities like people, organizations, and locations in text. Optimized for CPU use. Uploaded by [harpertoken](https://huggingface.co/harpertoken).
|
| 39 |
|
| 40 |
## Intended Uses & Limitations
|
| 41 |
|
|
|
|
| 82 |
- **Datasets**: 1.18.3
|
| 83 |
- **Tokenizers**: 0.13.3
|
| 84 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 85 |
|
|
|