|
|
import torch |
|
|
import numpy as np |
|
|
from pathlib import Path |
|
|
from transnetv2_pytorch import TransNetV2 |
|
|
import json |
|
|
import re |
|
|
|
|
|
|
|
|
K = 3 |
|
|
MIN_DURATION_FRAMES = 2 |
|
|
MIN_CONFIDENCE = 0.02 |
|
|
|
|
|
data_dir = Path("data/animations") |
|
|
files = sorted(data_dir.glob("sample-*.webp")) |
|
|
print(f"Found {len(files)} files to process.") |
|
|
|
|
|
def get_best_device(): |
|
|
if torch.cuda.is_available(): |
|
|
return torch.device("cuda") |
|
|
elif torch.backends.mps.is_available(): |
|
|
return torch.device("mps") |
|
|
|
|
|
else: |
|
|
return torch.device("cpu") |
|
|
|
|
|
def load_original_frames(filepath): |
|
|
"""Load original frames from an animated webp file as PIL Images.""" |
|
|
from PIL import Image |
|
|
im = Image.open(filepath) |
|
|
frames = [] |
|
|
try: |
|
|
while True: |
|
|
frames.append(im.convert("RGB")) |
|
|
im.seek(im.tell() + 1) |
|
|
except EOFError: |
|
|
pass |
|
|
return frames |
|
|
|
|
|
def save_prediction_plot(single_frame_pred, original_frames, filename, interval=5, title=None): |
|
|
""" |
|
|
Save a plot of single frame predictions with thumbnails annotated at regular intervals. |
|
|
""" |
|
|
import matplotlib.pyplot as plt |
|
|
from matplotlib.offsetbox import OffsetImage, AnnotationBbox |
|
|
import numpy as np |
|
|
|
|
|
plt.figure(figsize=(12, 4)) |
|
|
if title: |
|
|
plt.title(title) |
|
|
plt.plot(single_frame_pred) |
|
|
ax = plt.gca() |
|
|
|
|
|
for idx in range(0, len(original_frames), interval): |
|
|
thumb = original_frames[idx].resize((64, 64)) |
|
|
imagebox = OffsetImage(np.array(thumb), zoom=0.5) |
|
|
ab = AnnotationBbox(imagebox, (idx, single_frame_pred[idx]), frameon=False, box_alignment=(0.5, -0.1)) |
|
|
ax.add_artist(ab) |
|
|
plt.xlabel("Frame") |
|
|
plt.ylabel("Prediction") |
|
|
plt.tight_layout() |
|
|
plt.savefig(filename) |
|
|
plt.close() |
|
|
|
|
|
def save_timeline_jpg(frames, scene_change_indices, filename, interval=5, roi_radius=2, title=None, single_frame_pred=None): |
|
|
""" |
|
|
Save a timeline JPG with thumbnails every `interval` frames and every frame near scene changes. |
|
|
Scene change regions are highlighted. Each thumbnail is annotated with its frame index. |
|
|
""" |
|
|
import matplotlib.pyplot as plt |
|
|
from matplotlib.offsetbox import OffsetImage, AnnotationBbox |
|
|
import matplotlib.patches as mpatches |
|
|
import numpy as np |
|
|
|
|
|
|
|
|
frames_to_render = set(range(0, len(frames), interval)) |
|
|
for idx in scene_change_indices: |
|
|
for offset in range(-roi_radius, roi_radius+1): |
|
|
fidx = idx + offset |
|
|
if 0 <= fidx < len(frames): |
|
|
frames_to_render.add(fidx) |
|
|
frames_to_render = sorted(frames_to_render) |
|
|
|
|
|
|
|
|
n = len(frames_to_render) |
|
|
x_positions = list(range(n)) |
|
|
|
|
|
fig, ax = plt.subplots(figsize=(max(8, n*0.5), 3)) |
|
|
ax.set_xlim(-1, n) |
|
|
ax.set_ylim(0, 1) |
|
|
ax.axis('off') |
|
|
|
|
|
|
|
|
for idx in scene_change_indices: |
|
|
region = [i for i, fidx in enumerate(frames_to_render) if abs(fidx-idx) <= roi_radius] |
|
|
if region: |
|
|
start, end = region[0], region[-1] |
|
|
rect = mpatches.Rectangle((start-0.5, 0.05), end-start+1, 0.9, color='yellow', alpha=0.2) |
|
|
ax.add_patch(rect) |
|
|
|
|
|
|
|
|
last_frames = set(scene_change_indices) |
|
|
first_frames = set(idx + 1 for idx in scene_change_indices if idx + 1 < len(frames)) |
|
|
|
|
|
|
|
|
for i, fidx in enumerate(frames_to_render): |
|
|
thumb = frames[fidx].resize((32, 32)) |
|
|
imagebox = OffsetImage(np.array(thumb), zoom=0.7) |
|
|
|
|
|
if fidx in last_frames: |
|
|
bboxprops = dict(edgecolor='red', linewidth=2) |
|
|
elif fidx in first_frames: |
|
|
bboxprops = dict(edgecolor='green', linewidth=2) |
|
|
else: |
|
|
bboxprops = None |
|
|
ab = AnnotationBbox( |
|
|
imagebox, |
|
|
(x_positions[i], 0.6), |
|
|
frameon=True, |
|
|
box_alignment=(0.5, 0.5), |
|
|
bboxprops=bboxprops |
|
|
) |
|
|
ax.add_artist(ab) |
|
|
|
|
|
ax.text(x_positions[i], 0.32, str(fidx), ha='center', va='center', fontsize=9, color='black', bbox=dict(facecolor='white', edgecolor='none', alpha=0.8, boxstyle='round,pad=0.2')) |
|
|
|
|
|
if single_frame_pred is not None: |
|
|
pred_val = single_frame_pred[fidx] |
|
|
|
|
|
if isinstance(pred_val, np.ndarray): |
|
|
pred_val = float(pred_val.squeeze()) |
|
|
ax.text(x_positions[i], 0.18, f"{pred_val:.2f}", ha='center', va='center', fontsize=8, color='blue', bbox=dict(facecolor='white', edgecolor='none', alpha=0.7, boxstyle='round,pad=0.2')) |
|
|
|
|
|
if title: |
|
|
ax.text(0, 0.95, title, fontsize=12, ha='left', va='top', color='navy') |
|
|
plt.tight_layout() |
|
|
plt.savefig(filename, dpi=150) |
|
|
plt.close(fig) |
|
|
|
|
|
def frames_to_video_tensor(frames): |
|
|
"""Convert a list of PIL frames to a torch tensor of shape (num_frames, 27, 48, 3) and dtype uint8.""" |
|
|
import numpy as np |
|
|
from PIL import Image |
|
|
processed = [] |
|
|
for frame in frames: |
|
|
arr = np.array(frame.resize((48, 27), resample=Image.Resampling.BILINEAR), dtype=np.uint8) |
|
|
processed.append(torch.from_numpy(arr)) |
|
|
return torch.stack(processed) |
|
|
|
|
|
def detect_scene_changes(frames): |
|
|
video_tensor = frames_to_video_tensor(frames) |
|
|
video_tensor = video_tensor.unsqueeze(0).to(device) |
|
|
with torch.no_grad(): |
|
|
single_frame_logits, all_frame_logits = model(video_tensor) |
|
|
|
|
|
single_frame_logits_np = single_frame_logits.cpu().numpy().squeeze() |
|
|
all_frame_logits_np = all_frame_logits["many_hot"].cpu().numpy().squeeze() |
|
|
single_frame_pred = torch.sigmoid(single_frame_logits).cpu().numpy().squeeze() |
|
|
all_frame_pred_np = torch.sigmoid(all_frame_logits["many_hot"]).cpu().numpy().squeeze() |
|
|
return { |
|
|
"single_frame_pred": single_frame_pred, |
|
|
"all_frame_pred": all_frame_pred_np, |
|
|
"single_frame_logits": single_frame_logits_np, |
|
|
"all_frame_logits": all_frame_logits_np, |
|
|
} |
|
|
|
|
|
def cached_detect_scene_changes(file, original_frames): |
|
|
"""Detect scene changes with caching to avoid redundant computation.""" |
|
|
match = re.search(r"sample-(\d+)", file.name) |
|
|
sample_num = match.group(1) if match else "unknown" |
|
|
transnetv2_json = file.parent / f"sample-{sample_num}.transnetv2.json" |
|
|
if transnetv2_json.exists(): |
|
|
with open(transnetv2_json, "r") as f: |
|
|
result = json.load(f) |
|
|
result["single_frame_pred"] = np.array(result["single_frame_pred"]) |
|
|
result["all_frame_pred"] = np.array(result["all_frame_pred"]) |
|
|
result["single_frame_logits"] = np.array(result["single_frame_logits"]) |
|
|
result["all_frame_logits"] = np.array(result["all_frame_logits"]) |
|
|
else: |
|
|
result = detect_scene_changes(original_frames) |
|
|
|
|
|
with open(transnetv2_json, "w") as f: |
|
|
json.dump({ |
|
|
"single_frame_pred": result["single_frame_pred"].tolist(), |
|
|
"all_frame_pred": result["all_frame_pred"].tolist(), |
|
|
"single_frame_logits": result["single_frame_logits"].tolist(), |
|
|
"all_frame_logits": result["all_frame_logits"].tolist() |
|
|
}, f, indent=2) |
|
|
return result |
|
|
|
|
|
if __name__ == "__main__": |
|
|
device = get_best_device() |
|
|
print(f"Using device: {device}") |
|
|
model = TransNetV2() |
|
|
state_dict = torch.load("transnetv2-pytorch-weights.pth") |
|
|
model.load_state_dict(state_dict) |
|
|
model.eval().to(device) |
|
|
|
|
|
for file in files: |
|
|
match = re.search(r"sample-(\d+)", file.name) |
|
|
sample_num = match.group(1) if match else "unknown" |
|
|
original_frames = load_original_frames(file) |
|
|
result = cached_detect_scene_changes(file, original_frames) |
|
|
|
|
|
|
|
|
single_frame_pred = result["single_frame_pred"] |
|
|
|
|
|
|
|
|
valid_indices = np.arange(1, len(single_frame_pred) - 1) |
|
|
valid_preds = single_frame_pred[1:-1] |
|
|
|
|
|
sorted_indices = valid_indices[np.argsort(valid_preds)[::-1]] |
|
|
scene_change_indices = [] |
|
|
scene_cut_confidences = [] |
|
|
for idx in sorted_indices: |
|
|
if all(abs(idx - prev) >= MIN_DURATION_FRAMES for prev in scene_change_indices): |
|
|
scene_change_indices.append(int(idx)) |
|
|
scene_cut_confidences.append(float(single_frame_pred[idx])) |
|
|
if len(scene_change_indices) >= (K - 1): |
|
|
break |
|
|
|
|
|
|
|
|
failed = any(conf < MIN_CONFIDENCE for conf in scene_cut_confidences) |
|
|
|
|
|
print(f"File: {file.name}, Frames: {len(original_frames)}, Scene Changes: {len(scene_change_indices)}, Success: {not failed}") |
|
|
|
|
|
json_filename = file.parent / f"sample-{sample_num}.json" |
|
|
with open(json_filename, "w") as f: |
|
|
json.dump({ |
|
|
"num_frames": len(original_frames), |
|
|
"scene_change_indices": scene_change_indices, |
|
|
"scene_cut_confidences": scene_cut_confidences, |
|
|
|
|
|
"params": { |
|
|
"k": K, |
|
|
"min_duration_frames": MIN_DURATION_FRAMES, |
|
|
"min_confidence": MIN_CONFIDENCE, |
|
|
}, |
|
|
"success": not failed |
|
|
}, f, indent=2) |
|
|
|
|
|
timeline_filename = file.parent / f"sample-{sample_num}.timeline.jpg" |
|
|
plot_filename = file.parent / f"sample-{sample_num}.plot.jpg" |
|
|
save_timeline_jpg( |
|
|
frames=original_frames, |
|
|
scene_change_indices=scene_change_indices, |
|
|
filename=timeline_filename, |
|
|
interval=10, |
|
|
roi_radius=2, |
|
|
title=f"Timeline: {file.name}", |
|
|
single_frame_pred=result["single_frame_pred"] |
|
|
) |
|
|
|
|
|
save_prediction_plot( |
|
|
single_frame_pred=result["single_frame_pred"], |
|
|
original_frames=original_frames, |
|
|
filename=plot_filename, |
|
|
interval=5, |
|
|
title=f"Single Frame Predictions: {file.name}" |
|
|
) |
|
|
|
|
|
|