|
|
import json |
|
|
import logging |
|
|
import os |
|
|
import shutil |
|
|
import subprocess |
|
|
import tempfile |
|
|
from dataclasses import dataclass |
|
|
from glob import glob |
|
|
from pathlib import Path |
|
|
from typing import Any, Dict, List, Optional, Tuple, Union |
|
|
|
|
|
import cv2 |
|
|
import ffmpeg |
|
|
import numpy as np |
|
|
import torch |
|
|
from scipy import signal |
|
|
from scipy.interpolate import interp1d |
|
|
from scenedetect import ContentDetector, SceneManager, StatsManager |
|
|
from scenedetect.video_manager import VideoManager |
|
|
|
|
|
from detectors.s3fd import S3FD |
|
|
from detectors.s3fd.nets import S3FDNet |
|
|
from SyncNetInstance import SyncNetInstance |
|
|
from SyncNetModel import S |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
@dataclass |
|
|
class PipelineConfig: |
|
|
|
|
|
facedet_scale: float = 0.25 |
|
|
crop_scale: float = 0.40 |
|
|
min_track: int = 50 |
|
|
frame_rate: int = 25 |
|
|
num_failed_det: int = 25 |
|
|
min_face_size: int = 100 |
|
|
|
|
|
|
|
|
batch_size: int = 20 |
|
|
vshift: int = 15 |
|
|
|
|
|
|
|
|
s3fd_weights: str = "sfd_face.pth" |
|
|
syncnet_weights: str = "syncnet_v2.model" |
|
|
|
|
|
|
|
|
ffmpeg_bin: str = "ffmpeg" |
|
|
audio_sample_rate: int = 16000 |
|
|
|
|
|
@classmethod |
|
|
def from_dict(cls, d: Dict[str, Any]): |
|
|
return cls(**{k: v for k, v in d.items() if k in cls.__annotations__}) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
class SyncNetPipeline: |
|
|
def __init__( |
|
|
self, |
|
|
cfg: Union[PipelineConfig, Dict[str, Any], None] = None, |
|
|
*, |
|
|
device: str = "cuda", |
|
|
**override, |
|
|
): |
|
|
base = cfg if isinstance(cfg, PipelineConfig) else PipelineConfig.from_dict(cfg or {}) |
|
|
for k, v in override.items(): |
|
|
if hasattr(base, k): |
|
|
setattr(base, k, v) |
|
|
self.cfg = base |
|
|
self.device = device |
|
|
|
|
|
self.s3fd = self._load_s3fd(self.cfg.s3fd_weights) |
|
|
self.syncnet = self._load_syncnet(self.cfg.syncnet_weights) |
|
|
|
|
|
|
|
|
def _load_s3fd(self, path: str) -> S3FD: |
|
|
logging.info(f"Loading S3FD from {path}") |
|
|
net = S3FDNet(device=self.device) |
|
|
net.load_state_dict(torch.load(path, map_location=self.device)) |
|
|
net.eval() |
|
|
return S3FD(net=net, device=self.device) |
|
|
|
|
|
def _load_syncnet(self, path: str) -> SyncNetInstance: |
|
|
logging.info(f"Loading SyncNet from {path}") |
|
|
model = S() |
|
|
model.load_state_dict(torch.load(path, map_location=self.device)) |
|
|
model.eval() |
|
|
return SyncNetInstance(net=model, device=self.device) |
|
|
|
|
|
|
|
|
@staticmethod |
|
|
def _iou(a, b): |
|
|
xA, yA = max(a[0], b[0]), max(a[1], b[1]) |
|
|
xB, yB = min(a[2], b[2]), min(a[3], b[3]) |
|
|
inter = max(0, xB - xA) * max(0, yB - yA) |
|
|
areaA = (a[2] - a[0]) * (a[3] - a[1]) |
|
|
areaB = (b[2] - b[0]) * (b[3] - b[1]) |
|
|
return inter / (areaA + areaB - inter + 1e-8) |
|
|
|
|
|
def _track(self, dets, *, min_track_override: Optional[int] = None): |
|
|
cfg = self.cfg |
|
|
min_track = max( |
|
|
1, min_track_override if min_track_override is not None else cfg.min_track |
|
|
) |
|
|
tracks = [] |
|
|
while True: |
|
|
t = [] |
|
|
for faces in dets: |
|
|
for f in faces: |
|
|
if not t: |
|
|
t.append(f) |
|
|
faces.remove(f) |
|
|
elif ( |
|
|
f["frame"] - t[-1]["frame"] <= cfg.num_failed_det |
|
|
and self._iou(f["bbox"], t[-1]["bbox"]) > 0.5 |
|
|
): |
|
|
t.append(f) |
|
|
faces.remove(f) |
|
|
continue |
|
|
else: |
|
|
break |
|
|
if not t: |
|
|
break |
|
|
if len(t) >= min_track: |
|
|
fr = np.array([d["frame"] for d in t]) |
|
|
bb = np.array([d["bbox"] for d in t]) |
|
|
full_f = np.arange(fr[0], fr[-1] + 1) |
|
|
bb_i = np.stack([interp1d(fr, bb[:, i])(full_f) for i in range(4)], 1) |
|
|
if max( |
|
|
np.mean(bb_i[:, 2] - bb_i[:, 0]), |
|
|
np.mean(bb_i[:, 3] - bb_i[:, 1]), |
|
|
) > cfg.min_face_size: |
|
|
tracks.append({"frame": full_f, "bbox": bb_i}) |
|
|
return tracks |
|
|
|
|
|
def _crop(self, track, frames, audio_wav, base): |
|
|
cfg = self.cfg |
|
|
base.parent.mkdir(parents=True, exist_ok=True) |
|
|
tmp_avi = f"{base}t.avi" |
|
|
vw = cv2.VideoWriter(tmp_avi, cv2.VideoWriter_fourcc(*"XVID"), cfg.frame_rate, (224, 224)) |
|
|
|
|
|
s, x, y = [], [], [] |
|
|
for b in track["bbox"]: |
|
|
s.append(max(b[3] - b[1], b[2] - b[0]) / 2) |
|
|
x.append((b[0] + b[2]) / 2) |
|
|
y.append((b[1] + b[3]) / 2) |
|
|
s, x, y = map(lambda v: signal.medfilt(v, 13), (s, x, y)) |
|
|
|
|
|
for i, fidx in enumerate(track["frame"]): |
|
|
img = cv2.imread(frames[fidx]) |
|
|
if img is None: |
|
|
continue |
|
|
bs = s[i] |
|
|
cs = cfg.crop_scale |
|
|
pad = int(bs * (1 + 2 * cs)) |
|
|
img_p = cv2.copyMakeBorder( |
|
|
img, pad, pad, pad, pad, cv2.BORDER_CONSTANT, value=(110, 110, 110) |
|
|
) |
|
|
my, mx = y[i] + pad, x[i] + pad |
|
|
y1, y2 = int(my - bs), int(my + bs * (1 + 2 * cs)) |
|
|
x1, x2 = int(mx - bs * (1 + cs)), int(mx + bs * (1 + cs)) |
|
|
crop = cv2.resize(img_p[y1:y2, x1:x2], (224, 224)) |
|
|
vw.write(crop) |
|
|
vw.release() |
|
|
|
|
|
slice_wav = f"{base}.wav" |
|
|
ss = track["frame"][0] / cfg.frame_rate |
|
|
to = (track["frame"][-1] + 1) / cfg.frame_rate |
|
|
subprocess.call( |
|
|
f'{cfg.ffmpeg_bin} -y -i "{audio_wav}" -ss {ss:.3f} -to {to:.3f} "{slice_wav}"', |
|
|
shell=True, |
|
|
) |
|
|
|
|
|
final_avi = f"{base}.avi" |
|
|
subprocess.call( |
|
|
f'{cfg.ffmpeg_bin} -y -i "{tmp_avi}" -i "{slice_wav}" -c:v copy -c:a copy "{final_avi}"', |
|
|
shell=True, |
|
|
) |
|
|
os.remove(tmp_avi) |
|
|
return final_avi |
|
|
|
|
|
|
|
|
def inference( |
|
|
self, |
|
|
video_path: str, |
|
|
audio_path: str, |
|
|
*, |
|
|
cache_dir: Optional[str] = None, |
|
|
) -> Tuple[List[int], List[float], List[float], float, float, str, bool]: |
|
|
cfg = self.cfg |
|
|
work = Path(cache_dir) if cache_dir else Path(tempfile.mkdtemp()) |
|
|
if cache_dir: |
|
|
work.mkdir(parents=True, exist_ok=True) |
|
|
|
|
|
try: |
|
|
|
|
|
avi = work / "video.avi" |
|
|
( |
|
|
ffmpeg.input(video_path) |
|
|
.output(str(avi), **{"q:v": 2}, r=cfg.frame_rate, **{"async": 1}) |
|
|
.overwrite_output() |
|
|
.run() |
|
|
) |
|
|
|
|
|
|
|
|
frames_dir = work / "frames" |
|
|
frames_dir.mkdir(exist_ok=True) |
|
|
( |
|
|
ffmpeg.input(str(avi)) |
|
|
.output(str(frames_dir / "%06d.jpg"), **{"q:v": 2}, f="image2", threads=1) |
|
|
.overwrite_output() |
|
|
.run() |
|
|
) |
|
|
frames = sorted(glob(str(frames_dir / "*.jpg"))) |
|
|
|
|
|
|
|
|
audio_wav = work / "speech.wav" |
|
|
( |
|
|
ffmpeg.input(audio_path) |
|
|
.output(str(audio_wav), ac=1, ar=cfg.audio_sample_rate, format="wav") |
|
|
.overwrite_output() |
|
|
.run() |
|
|
) |
|
|
|
|
|
|
|
|
detections = [] |
|
|
for i, fp in enumerate(frames): |
|
|
img = cv2.imread(fp) |
|
|
boxes = ( |
|
|
self.s3fd.detect_faces( |
|
|
cv2.cvtColor(img, cv2.COLOR_BGR2RGB), |
|
|
conf_th=0.9, |
|
|
scales=[cfg.facedet_scale], |
|
|
) |
|
|
if img is not None |
|
|
else [] |
|
|
) |
|
|
detections.append( |
|
|
[ |
|
|
{"frame": i, "bbox": b[:-1].tolist(), "conf": float(b[-1])} |
|
|
for b in boxes |
|
|
] |
|
|
) |
|
|
|
|
|
flat = [f for fs in detections for f in fs] |
|
|
s3fd_json = json.dumps(flat) if flat else "" |
|
|
has_face = bool(flat) |
|
|
|
|
|
|
|
|
vm = VideoManager([str(avi)]) |
|
|
sm = SceneManager(StatsManager()) |
|
|
sm.add_detector(ContentDetector()) |
|
|
vm.start() |
|
|
sm.detect_scenes(frame_source=vm) |
|
|
scenes = sm.get_scene_list(vm.get_base_timecode()) or [ |
|
|
(vm.get_base_timecode(), vm.get_current_timecode()) |
|
|
] |
|
|
total_frames = len(detections) |
|
|
|
|
|
|
|
|
tracks = [] |
|
|
for sc in scenes: |
|
|
s = min(max(sc[0].frame_num, 0), total_frames) |
|
|
end_tc = sc[1] if sc[1] is not None else vm.get_current_timecode() |
|
|
e = min(max(end_tc.frame_num, s), total_frames) |
|
|
if e <= s: |
|
|
continue |
|
|
|
|
|
scene_detections = [lst.copy() for lst in detections[s:e]] |
|
|
if not scene_detections: |
|
|
continue |
|
|
|
|
|
min_track_len = max(1, min(cfg.min_track, len(scene_detections))) |
|
|
tracks.extend( |
|
|
self._track( |
|
|
scene_detections, |
|
|
min_track_override=min_track_len, |
|
|
) |
|
|
) |
|
|
|
|
|
|
|
|
crops = [ |
|
|
self._crop(t, frames, str(audio_wav), Path(work) / "cropped" / f"{i:05d}") for i, t in enumerate(tracks) |
|
|
] |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
offsets, confs, dists = [], [], [] |
|
|
class Opt: ... |
|
|
for i, cp in enumerate(crops): |
|
|
crop_dir = work / "cropped" / f"crop_{i:05d}" |
|
|
frames_dir = crop_dir |
|
|
frames_dir.mkdir(parents=True, exist_ok=True) |
|
|
audio_path = crop_dir / "audio.wav" |
|
|
|
|
|
|
|
|
( |
|
|
ffmpeg.input(cp) |
|
|
.output(str(frames_dir / "%06d.jpg"), f="image2", threads=1) |
|
|
.overwrite_output() |
|
|
.run() |
|
|
) |
|
|
|
|
|
|
|
|
( |
|
|
ffmpeg.input(cp) |
|
|
.output( |
|
|
str(audio_path), |
|
|
ac=1, |
|
|
vn=None, |
|
|
acodec="pcm_s16le", |
|
|
ar=16000, |
|
|
af="aresample=async=1", |
|
|
) |
|
|
.overwrite_output() |
|
|
.run() |
|
|
) |
|
|
|
|
|
opt = Opt() |
|
|
opt.tmp_dir = str(crop_dir) |
|
|
opt.batch_size = cfg.batch_size |
|
|
opt.vshift = cfg.vshift |
|
|
|
|
|
off, conf, dist = self.syncnet.evaluate(opt=opt) |
|
|
offsets.append(off) |
|
|
confs.append(conf) |
|
|
dists.append(dist) |
|
|
|
|
|
if not offsets: |
|
|
return ([], [], [], 0.0, 0.0, "", False) |
|
|
|
|
|
return offsets, confs, dists, max(confs), min(dists), s3fd_json, has_face |
|
|
|
|
|
finally: |
|
|
if not cache_dir: |
|
|
shutil.rmtree(work, ignore_errors=True) |
|
|
|