castelao's picture
Update README.md
b3d59c9 verified
|
raw
history blame
1.67 kB
---
author: National Laboratory of the Rockies
license: apache-2.0
language:
- en
base_model:
- google-bert/bert-base-uncased
tags:
- text-classification
- bert
---
# BERT Fine-tuned on PV-Inspection Reports
This model, developed at NLR, is a fine-tuned version of `bert-base-uncased` on multiple PV inspection reports in the US.
**Developed by:**
- Jeff Cook, Strategic Energy Analysis Center
- Guilherme Castelão, Strategic Energy Analysis Center
- Danny Chang, Strategic Energy Analysis Center
- Sertac Akar, Accelerated Deployment and Decision Support Center
- James Elsworth, Resilient Infrastructure and Security Center
## Model Description
- **Base model:** bert-base-uncased
- **Task:** classify type of issue reported
- **Training data:** PV inspection reports
## Intended Uses
Classify the reason(s) of failure from PV installations inspections reports.
## How to Use
```python
from transformers import AutoModelForSequenceClassification, AutoTokenizer
model = AutoModelForSequenceClassification.from_pretrained("NatLabRockies/pvi-issue-classification")
tokenizer = AutoTokenizer.from_pretrained("NatLabRockies/pvi-issue-classification")
# Example usage
inputs = tokenizer("NEC 110.26 specifies working space clearance about electrical equipment.", return_tensors="pt")
outputs = model(**inputs)
logits = outputs.logits
predicted_class_id = logits.argmax().item()
predicted_class = model.config.id2label[predicted_class_id]
print(f"Predicted issue: {predicted_class}")
```
## Training Details
- **Training epochs:** [number]
- **Learning rate:** [value]
- **Batch size:** [value]
## Evaluation Results
[Add your metrics: accuracy, F1, etc.]