ysfad's picture
Upload README.md with huggingface_hub
c46d32e verified
---
library_name: pytorch
tags:
- waste-classification
- mae
- vision-transformer
- environmental
- recycling
license: mit
datasets:
- RealWaste
metrics:
- accuracy
model-index:
- name: MAE Waste Classifier
results:
- task:
type: image-classification
name: Waste Classification
dataset:
type: RealWaste
name: RealWaste Dataset
metrics:
- type: accuracy
value: 0.9327
name: Validation Accuracy
---
# MAE Waste Classifier
A finetuned MAE (Masked Autoencoder) ViT-Base model for waste classification achieving **93.27% validation accuracy** on 9 waste categories.
## Model Details
- **Architecture**: Vision Transformer (ViT-Base) with MAE pretraining
- **Parameters**: ~86M
- **Input Size**: 224x224 RGB images
- **Classes**: 9 waste categories
- **Validation Accuracy**: 93.27%
## Categories
1. **Cardboard** - Flatten and place in recycling bin. Remove any tape or staples.
2. **Food Organics** - Compost in organic waste bin or home composter.
3. **Glass** - Rinse and place in glass recycling. Remove lids and caps.
4. **Metal** - Rinse aluminum/steel cans and place in recycling bin.
5. **Miscellaneous Trash** - Dispose in general waste bin. Cannot be recycled.
6. **Paper** - Place clean paper in recycling. Remove plastic windows from envelopes.
7. **Plastic** - Check recycling number. Rinse containers before recycling.
8. **Textile Trash** - Donate if reusable, otherwise dispose in textile recycling.
9. **Vegetation** - Compost in organic waste or use for mulch in garden.
## Usage
```python
import torch
import timm
from PIL import Image
from torchvision import transforms
# Load model
model = timm.create_model('vit_base_patch16_224', pretrained=False, num_classes=9)
checkpoint = torch.load('best_model.pth', map_location='cpu')
model.load_state_dict(checkpoint['model_state_dict'])
model.eval()
# Preprocessing
transform = transforms.Compose([
transforms.Resize((224, 224)),
transforms.ToTensor(),
transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225])
])
# Inference
image = Image.open('waste_item.jpg').convert('RGB')
input_tensor = transform(image).unsqueeze(0)
with torch.no_grad():
outputs = model(input_tensor)
probabilities = torch.nn.functional.softmax(outputs, dim=1)
predicted_class = torch.argmax(probabilities, dim=1).item()
categories = ['Cardboard', 'Food Organics', 'Glass', 'Metal', 'Miscellaneous Trash', 'Paper', 'Plastic', 'Textile Trash', 'Vegetation']
print(f"Predicted: {categories[predicted_class]}")
```
## Training Details
- **Dataset**: RealWaste (4,752 images)
- **Pretraining**: MAE on ImageNet
- **Finetuning**: 15 epochs on RealWaste
- **Optimizer**: AdamW
- **Hardware**: NVIDIA RTX 3080 Ti
## Performance
- **Validation Accuracy**: 93.27%
- **Training Accuracy**: 99.89%
- **Model Size**: ~350MB
- **Inference Speed**: ~50ms per image (GPU)
## Environmental Impact
This model helps improve recycling efficiency by providing accurate waste classification and proper disposal instructions.