File size: 4,680 Bytes
960b635 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 |
import gradio as gr
# ---- IMPORT BACKENDS ----
from image_backend import predict_image_pil
from audio_backend import predict_audio
# =========================
# IMAGE LOGIC (UNCHANGED)
# =========================
def analyze_image(image):
label, confidence, heatmap = predict_image_pil(image)
if label == "Fake":
if confidence >= 90:
risk = "π¨ High likelihood of Deepfake"
elif confidence >= 60:
risk = "β οΈ Possibly Deepfake"
else:
risk = "β οΈ Uncertain Deepfake"
else:
if confidence >= 90:
risk = "β
Likely Real"
elif confidence >= 60:
risk = "β οΈ Possibly Real"
else:
risk = "β οΈ Uncertain β Needs Review"
return label, f"{confidence} %", risk, heatmap
# =========================
# AUDIO LOGIC (UNCHANGED)
# =========================
def analyze_audio(audio_path):
label, confidence = predict_audio(audio_path)
if label == "fake":
if confidence >= 90:
risk = "π¨ High likelihood of Deepfake"
elif confidence >= 60:
risk = "β οΈ Possibly Deepfake"
else:
risk = "β οΈ Uncertain β Needs Review"
else:
if confidence >= 90:
risk = "β
Likely Real"
elif confidence >= 60:
risk = "β οΈ Possibly Real"
else:
risk = "β οΈ Uncertain β Needs Review"
return label.capitalize(), f"{confidence} %", risk
# =========================
# UI (EXACT SAME)
# =========================
with gr.Blocks() as demo:
gr.Markdown("# π§ Unified Deepfake Detection System")
with gr.Tabs():
# HOME TAB
with gr.Tab("π Home"):
gr.Markdown(
"""
## Welcome π
Select the type of media you want to analyze:
"""
)
gr.Markdown("### π Choose Detection Mode")
gr.Markdown("- πΌ **Image Deepfake Detection**\n- π§ **Audio Deepfake Detection**")
gr.Markdown(
"""
π Use the tabs above to switch between Image and Audio detection.
"""
)
# IMAGE TAB
with gr.Tab("πΌ Image Deepfake"):
gr.Markdown("# πΌ Deepfake Image Detection System")
with gr.Row():
with gr.Column(scale=1):
image_input = gr.Image(
label="Upload Image",
type="pil",
height=280
)
img_submit = gr.Button("Submit")
img_clear = gr.Button("Clear")
with gr.Column(scale=2):
img_pred = gr.Text(label="Prediction")
img_conf = gr.Text(label="Confidence")
img_risk = gr.Text(label="Risk Assessment")
img_heatmap = gr.Image(
label="Explainability Heatmap",
height=280
)
img_submit.click(
fn=analyze_image,
inputs=image_input,
outputs=[img_pred, img_conf, img_risk, img_heatmap]
)
img_clear.click(
fn=lambda: (None, "", "", None),
inputs=None,
outputs=[image_input, img_pred, img_conf, img_risk]
)
# AUDIO TAB
with gr.Tab("π§ Audio Deepfake"):
gr.Markdown("# π§ Deepfake Audio Detection System")
with gr.Row():
with gr.Column(scale=1):
audio_input = gr.Audio(
label="Upload Audio (.wav)",
type="filepath"
)
aud_submit = gr.Button("Submit")
aud_clear = gr.Button("Clear")
with gr.Column(scale=2):
aud_pred = gr.Text(label="Prediction")
aud_conf = gr.Text(label="Confidence")
aud_risk = gr.Text(label="Risk Assessment")
aud_submit.click(
fn=analyze_audio,
inputs=audio_input,
outputs=[aud_pred, aud_conf, aud_risk]
)
aud_clear.click(
fn=lambda: (None, "", ""),
inputs=None,
outputs=[audio_input, aud_pred, aud_conf]
)
demo.launch()
|