Update app.py
Browse files
app.py
CHANGED
|
@@ -18,7 +18,7 @@ model.load_state_dict(torch.load(weights_path, map_location="cpu"))
|
|
| 18 |
model.eval()
|
| 19 |
|
| 20 |
# Preprocessing: same as training
|
| 21 |
-
IMG_HEIGHT, IMG_WIDTH =
|
| 22 |
transform = transforms.Compose([
|
| 23 |
transforms.Resize((IMG_HEIGHT, IMG_WIDTH)),
|
| 24 |
transforms.ToTensor()
|
|
@@ -26,7 +26,7 @@ transform = transforms.Compose([
|
|
| 26 |
|
| 27 |
def predict(image):
|
| 28 |
orig_w, orig_h = image.size # original size of uploaded image
|
| 29 |
-
img = transform(image).unsqueeze(0)
|
| 30 |
with torch.no_grad():
|
| 31 |
pred = model(img)
|
| 32 |
|
|
|
|
| 18 |
model.eval()
|
| 19 |
|
| 20 |
# Preprocessing: same as training
|
| 21 |
+
IMG_HEIGHT, IMG_WIDTH = 448, 448
|
| 22 |
transform = transforms.Compose([
|
| 23 |
transforms.Resize((IMG_HEIGHT, IMG_WIDTH)),
|
| 24 |
transforms.ToTensor()
|
|
|
|
| 26 |
|
| 27 |
def predict(image):
|
| 28 |
orig_w, orig_h = image.size # original size of uploaded image
|
| 29 |
+
img = transform(image).unsqueeze(0)
|
| 30 |
with torch.no_grad():
|
| 31 |
pred = model(img)
|
| 32 |
|