import gradio as gr import torch import torch.nn as nn from torchvision import transforms from torchvision.models import swin_t from PIL import Image # 🔧 Model definition class MMIM(nn.Module): def __init__(self, num_classes=36): super(MMIM, self).__init__() self.backbone = swin_t(weights='IMAGENET1K_V1') self.backbone.head = nn.Identity() self.classifier = nn.Sequential( nn.Linear(768, 512), nn.ReLU(), nn.Dropout(0.3), nn.Linear(512, num_classes) ) def forward(self, x): features = self.backbone(x) return self.classifier(features) # ✅ Load model device = torch.device("cuda" if torch.cuda.is_available() else "cpu") model = MMIM(num_classes=36) # 🧠 Load only matching weights from checkpoint (skip classifier mismatch) checkpoint = torch.load("MMIM_best.pth", map_location=device) filtered_checkpoint = { k: v for k, v in checkpoint.items() if k in model.state_dict() and model.state_dict()[k].shape == v.shape } model.load_state_dict(filtered_checkpoint, strict=False) model.to(device) model.eval() # ✅ class_names mapped according to confusion matrix order class_names = [ "Chinee apple", # class1 "Black grass", # class14 "Charlock", # class15 "Cleavers", # class16 "Common Chickweed", # class17 "Common Wheat", # class18 "Fat Hen", # class19 "Lanthana", # class2 "Loose Silky bent", # class20 "Maize", # class21 "Scentless Mayweed", # class22 "Shepherds Purse", # class23 "Small-Flowered Cranesbill", # class24 "Sugar beet", # class25 "Carpetweeds", # class26 "Crabgrass",# class27 "Eclipta", # class28 "Goosegrass", # class29 "Negative", # class3 "Morningglory", # class30 "Nutsedge", # class31 "Palmer Amarnath", # class32 "Prickly Sida", # class33 "Purslane", # class34 "Ragweed", # class35 "Sicklepod", # class36 "SpottedSpurge", # class37 "SpurredAnoda", # class38 "Swinecress", # class39 "Parkinsonia", # class4 "Waterhemp", # class40 "Parthenium", # class5 "Prickly acacia", # class6 "Rubber vine", # class7 "Siam weed", # class8 "Snake weed", # class9 ] # 🔁 Image transform transform = transforms.Compose([ transforms.Resize((224, 224)), transforms.ToTensor() ]) # 🔍 Prediction function def predict(img): img = img.convert('RGB') img_tensor = transform(img).unsqueeze(0).to(device) with torch.no_grad(): outputs = model(img_tensor) probs = torch.softmax(outputs, dim=1) conf, pred = torch.max(probs, 1) predicted_class = class_names[pred.item()] confidence = conf.item() * 100 if predicted_class.lower() == "negative": return f"⚠️ This image is predicted as Negative.\nConfidence: {confidence:.2f}%" return f"✅ Predicted class: {predicted_class}\nConfidence: {confidence:.2f}%" # 🎨 Gradio Interface interface = gr.Interface( fn=predict, inputs=gr.Image(type="pil"), outputs="text", title="Weed Image Classifier", description="Upload a weed image to predict its class." ) interface.launch()