Spaces:
Build error
Build error
embedding choice
Browse files
TEST_sculley.jpg
ADDED
|
TEST_spindler.jpg
ADDED
|
__pycache__/app.cpython-38.pyc
CHANGED
|
Binary files a/__pycache__/app.cpython-38.pyc and b/__pycache__/app.cpython-38.pyc differ
|
|
|
app.py
CHANGED
|
@@ -25,6 +25,20 @@ dbackends = [
|
|
| 25 |
['Fast Multi-task Cascade Convolutional Neural Network (PyTorch)','fastmtcnn']
|
| 26 |
]
|
| 27 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 28 |
dbackendinfo = 'Detectors with 🌈 require a color image.'
|
| 29 |
|
| 30 |
|
|
@@ -91,9 +105,10 @@ with gr.Blocks() as demo:
|
|
| 91 |
|
| 92 |
|
| 93 |
with gr.Tab("Identify People in One Image"):
|
|
|
|
| 94 |
oneimageannotations = []
|
| 95 |
-
def identify_in_one_image(imgfile):
|
| 96 |
-
oneimageresults = DeepFace.find(img_path=imgfile, db_path="db")
|
| 97 |
oneimageresults = pd.concat(oneimageresults)
|
| 98 |
for i, found_face in oneimageresults.iterrows():
|
| 99 |
face_coordinates = (found_face["source_x"],found_face["source_y"], (found_face["source_x"] + found_face["source_w"]),(found_face["source_y"] + found_face["source_h"]))
|
|
@@ -107,14 +122,14 @@ with gr.Blocks() as demo:
|
|
| 107 |
)
|
| 108 |
return results, oneimageannotations
|
| 109 |
|
| 110 |
-
oneimage_input_image = gr.Image()
|
| 111 |
|
| 112 |
found_faces=gr.AnnotatedImage()
|
| 113 |
debug_output = gr.Textbox()
|
| 114 |
gr.Interface(
|
| 115 |
allow_flagging = "never",
|
| 116 |
fn=identify_in_one_image,
|
| 117 |
-
inputs=oneimage_input_image,
|
| 118 |
outputs=[found_faces, debug_output]
|
| 119 |
)
|
| 120 |
|
|
|
|
| 25 |
['Fast Multi-task Cascade Convolutional Neural Network (PyTorch)','fastmtcnn']
|
| 26 |
]
|
| 27 |
|
| 28 |
+
embedding_backends = [
|
| 29 |
+
"VGG-Face",
|
| 30 |
+
"Facenet",
|
| 31 |
+
"Facenet512",
|
| 32 |
+
"OpenFace",
|
| 33 |
+
"DeepFace",
|
| 34 |
+
"DeepID",
|
| 35 |
+
"ArcFace",
|
| 36 |
+
"Dlib",
|
| 37 |
+
"SFace",
|
| 38 |
+
"GhostFaceNet",
|
| 39 |
+
]
|
| 40 |
+
|
| 41 |
+
|
| 42 |
dbackendinfo = 'Detectors with 🌈 require a color image.'
|
| 43 |
|
| 44 |
|
|
|
|
| 105 |
|
| 106 |
|
| 107 |
with gr.Tab("Identify People in One Image"):
|
| 108 |
+
embedding_backendchoice = gr.Radio(choices=embedding_backends,label='Embedding Backend:',container=True,value='ArcFace')
|
| 109 |
oneimageannotations = []
|
| 110 |
+
def identify_in_one_image(imgfile, embedding_backendchoice):
|
| 111 |
+
oneimageresults = DeepFace.find(img_path=imgfile, db_path="db", model_name=embedding_backendchoice)
|
| 112 |
oneimageresults = pd.concat(oneimageresults)
|
| 113 |
for i, found_face in oneimageresults.iterrows():
|
| 114 |
face_coordinates = (found_face["source_x"],found_face["source_y"], (found_face["source_x"] + found_face["source_w"]),(found_face["source_y"] + found_face["source_h"]))
|
|
|
|
| 122 |
)
|
| 123 |
return results, oneimageannotations
|
| 124 |
|
| 125 |
+
oneimage_input_image = gr.Image(value="TEST_spindler.jpg")
|
| 126 |
|
| 127 |
found_faces=gr.AnnotatedImage()
|
| 128 |
debug_output = gr.Textbox()
|
| 129 |
gr.Interface(
|
| 130 |
allow_flagging = "never",
|
| 131 |
fn=identify_in_one_image,
|
| 132 |
+
inputs=[oneimage_input_image, embedding_backendchoice],
|
| 133 |
outputs=[found_faces, debug_output]
|
| 134 |
)
|
| 135 |
|
db/ds_model_arcface_detector_opencv_aligned_normalization_base_expand_0.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8c493963c11bc05abf1668a4c009d8a7259009b13c4c81721e8f1a3c77176683
|
| 3 |
+
size 28405
|