Umsakwa's picture
Update README.md
fb8ce5d verified
---
library_name: transformers
tags:
- image-classification
- vit
- pytorch
license: apache-2.0
language:
- en
metrics:
- accuracy
- f1
datasets:
- AI-Lab-Makerere/beans
---
# Umsakwa/Uddayvit-image-classification-model
This Vision Transformer (ViT) model has been fine-tuned for image classification tasks on the [Beans Dataset](https://huggingface.co/datasets/beans), which consists of images of beans categorized into three classes:
- **Angular Leaf Spot**
- **Bean Rust**
- **Healthy**
## Model Details
- **Architecture**: Vision Transformer (ViT)
- **Base Model**: `google/vit-base-patch16-224-in21k`
- **Framework**: PyTorch
- **Task**: Image Classification
- **Labels**: 3 (angular_leaf_spot, bean_rust, healthy)
- **Input Shape**: 224x224 RGB images
- **Training Dataset**: [Beans Dataset](https://huggingface.co/datasets/beans)
- **Fine-Tuning**: The model was fine-tuned on the Beans dataset to classify plant diseases in beans.
### Model Description
The model uses the ViT architecture, which processes image patches using a transformer-based approach. It has been trained to classify bean diseases with high accuracy. This makes it particularly useful for agricultural applications, such as early disease detection and plant health monitoring.
- **Developed by**: Udday (Umsakwa)
- **Language(s)**: N/A (Image-based)
- **License**: Apache-2.0
- **Finetuned from**: `google/vit-base-patch16-224-in21k`
### Model Sources
- **Repository**: [Umsakwa/Uddayvit-image-classification-model](https://huggingface.co/Umsakwa/Uddayvit-image-classification-model)
## Uses
### Direct Use
This model can be directly used for classifying bean leaf images into one of three categories: angular leaf spot, bean rust, or healthy.
### Downstream Use
The model may also be fine-tuned further for similar agricultural image classification tasks or integrated into larger plant health monitoring systems.
### Out-of-Scope Use
- The model is not suitable for non-agricultural image classification tasks without further fine-tuning.
- Not robust to extreme distortions, occlusions, or very low-resolution images.
## Bias, Risks, and Limitations
- **Bias**: The dataset may contain biases due to specific environmental or geographic conditions of the sampled plants.
- **Limitations**: Performance may degrade on datasets significantly different from the training dataset.
### Recommendations
- Users should ensure the model is evaluated on their specific dataset before deployment.
- Additional fine-tuning may be required for domain-specific applications.
## How to Get Started with the Model
To use this model for inference:
```python
from transformers import ViTForImageClassification, ViTImageProcessor
# Load model and processor
model = ViTForImageClassification.from_pretrained("Umsakwa/Uddayvit-image-classification-model")
processor = ViTImageProcessor.from_pretrained("Umsakwa/Uddayvit-image-classification-model")
# Prepare an image
image = processor(images="path_to_image.jpg", return_tensors="pt")
# Run inference
outputs = model(**image)
predictions = outputs.logits.argmax(-1)