kadzon commited on
Commit
140414b
·
1 Parent(s): 1c4e900

add imports

Browse files
Files changed (1) hide show
  1. app.py +15 -11
app.py CHANGED
@@ -5,9 +5,16 @@ import torchvision
5
  import numpy as np
6
  from PIL import Image
7
 
8
- # Load MegaDetector v5a model
 
 
 
 
 
 
 
9
 
10
- # models = ["model_weights/md_v5a.0.0.pt","model_weights/md_v5b.0.0.pt"]
11
  model = torch.hub.load('ultralytics/yolov5', 'custom', "model_weights/md_v5a.0.0.pt")
12
 
13
  def yolo(im, size=640):
@@ -20,18 +27,15 @@ def yolo(im, size=640):
20
  results.render() # updates results.imgs with boxes and labels
21
  return Image.fromarray(results.imgs[0])
22
 
23
- #image = gr.inputs.Image(type="pil", label="Input Image")
24
- #chosen_model = gr.inputs.Dropdown(choices = models, value = "model_weights/md_v5a.0.0.pt",type = "value", label="Model Weight")
25
- #size = 640
26
-
27
- #inputs = [image, chosen_model, size]
28
- inputs = gr.inputs.Image(type="pil", label="Input Image")
29
- outputs = gr.outputs.Image(type="pil", label="Output Image")
30
-
31
  title = "MegaDetector and DeepLabcutLive"
32
  description = "Interact with MegaDetector and DeeplabCutLive for pose estimation"
33
- article = "<p style='text-align: center'>This app makes predictions using a YOLOv5x6 model that was trained to detect animals, humans, and vehicles in camera trap images; find out more about the project on <a href='https://github.com/microsoft/CameraTraps'>GitHub</a>. This app was built by Henry Lydecker but really depends on code and models developed by <a href='http://ecologize.org/'>Ecologize</a> and <a href='http://aka.ms/aiforearth'>Microsoft AI for Earth</a>. Find out more about the YOLO model from the original creator, <a href='https://pjreddie.com/darknet/yolo/'>Joseph Redmon</a>. YOLOv5 is a family of compound-scaled object detection models trained on the COCO dataset and developed by Ultralytics, and includes simple functionality for Test Time Augmentation (TTA), model ensembling, hyperparameter evolution, and export to ONNX, CoreML and TFLite. <a href='https://github.com/ultralytics/yolov5'>Source code</a> | <a href='https://pytorch.org/hub/ultralytics_yolov5'>PyTorch Hub</a></p>"
34
 
 
 
 
35
 
 
36
  examples = [['data/owl.jpg'], ['data/snake.jpg'],['data/beluga.jpg'],['data/rhino.jpg']]
37
  gr.Interface(yolo, inputs, outputs, title=title, description=description, article=article, examples=examples, theme="huggingface").launch(enable_queue=True)
 
5
  import numpy as np
6
  from PIL import Image
7
 
8
+ #script load
9
+ import json
10
+ import os
11
+ import numpy as np
12
+ import tensorflow.compat.v1 as tf
13
+ tf.disable_v2_behavior()
14
+ from dlclive import DLCLive, Processor
15
+ from numpy import savetxt
16
 
17
+ # Load MegaDetector v5a model
18
  model = torch.hub.load('ultralytics/yolov5', 'custom', "model_weights/md_v5a.0.0.pt")
19
 
20
  def yolo(im, size=640):
 
27
  results.render() # updates results.imgs with boxes and labels
28
  return Image.fromarray(results.imgs[0])
29
 
30
+ #Layouts and descriptions
 
 
 
 
 
 
 
31
  title = "MegaDetector and DeepLabcutLive"
32
  description = "Interact with MegaDetector and DeeplabCutLive for pose estimation"
33
+ article = "<p style='text-align: center'>This app uses MegaDetector YOLOv5x6 model that was trained to detect animals, humans, and vehicles in camera trap images; find out more about the project on <a href='https://github.com/microsoft/CameraTraps'>GitHub</a>. We have also integrated DeepLabCut Live for pose estimation <a href='https://github.com/DeepLabCut/DeepLabCut-live'></a>.</p>"
34
 
35
+ # input image and output image
36
+ inputs = gr.inputs.Image(type="pil", label="Input Image")
37
+ outputs = gr.outputs.Image(type="pil", label="Output Image")
38
 
39
+ #data images stored
40
  examples = [['data/owl.jpg'], ['data/snake.jpg'],['data/beluga.jpg'],['data/rhino.jpg']]
41
  gr.Interface(yolo, inputs, outputs, title=title, description=description, article=article, examples=examples, theme="huggingface").launch(enable_queue=True)