File size: 1,667 Bytes
c7b5387 cc65dbd c7b5387 4b51d6f b3d59c9 4b51d6f cc65dbd 7a2d97d cc65dbd 4b51d6f b282951 4b51d6f e3fdd2f 4b51d6f 80990e4 a3e2799 4b51d6f | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 | ---
author: National Laboratory of the Rockies
license: apache-2.0
language:
- en
base_model:
- google-bert/bert-base-uncased
tags:
- text-classification
- bert
---
# BERT Fine-tuned on PV-Inspection Reports
This model, developed at NLR, is a fine-tuned version of `bert-base-uncased` on multiple PV inspection reports in the US.
**Developed by:**
- Jeff Cook, Strategic Energy Analysis Center
- Guilherme Castelão, Strategic Energy Analysis Center
- Danny Chang, Strategic Energy Analysis Center
- Sertac Akar, Accelerated Deployment and Decision Support Center
- James Elsworth, Resilient Infrastructure and Security Center
## Model Description
- **Base model:** bert-base-uncased
- **Task:** classify type of issue reported
- **Training data:** PV inspection reports
## Intended Uses
Classify the reason(s) of failure from PV installations inspections reports.
## How to Use
```python
from transformers import AutoModelForSequenceClassification, AutoTokenizer
model = AutoModelForSequenceClassification.from_pretrained("NatLabRockies/pvi-issue-classification")
tokenizer = AutoTokenizer.from_pretrained("NatLabRockies/pvi-issue-classification")
# Example usage
inputs = tokenizer("NEC 110.26 specifies working space clearance about electrical equipment.", return_tensors="pt")
outputs = model(**inputs)
logits = outputs.logits
predicted_class_id = logits.argmax().item()
predicted_class = model.config.id2label[predicted_class_id]
print(f"Predicted issue: {predicted_class}")
```
## Training Details
- **Training epochs:** [number]
- **Learning rate:** [value]
- **Batch size:** [value]
## Evaluation Results
[Add your metrics: accuracy, F1, etc.]
|