|
|
|
|
|
""" |
|
|
Attention weight analysis and visualization helpers for SignX. |
|
|
|
|
|
Capabilities: |
|
|
1. Parse attention weight tensors |
|
|
2. Map each generated gloss to video frame ranges |
|
|
3. Render visual assets (heatmaps, alignment plots, timelines) |
|
|
4. Write detailed analysis reports |
|
|
|
|
|
Example: |
|
|
from eval.attention_analysis import AttentionAnalyzer |
|
|
|
|
|
analyzer = AttentionAnalyzer( |
|
|
attentions=attention_weights, # [time, batch, beam, src_len] |
|
|
translation="WORD1 WORD2 WORD3", |
|
|
video_frames=100 |
|
|
) |
|
|
|
|
|
analyzer.generate_all_visualizations(output_dir="results/") |
|
|
""" |
|
|
|
|
|
import os |
|
|
import io |
|
|
import json |
|
|
import shutil |
|
|
import subprocess |
|
|
import numpy as np |
|
|
from pathlib import Path |
|
|
from datetime import datetime |
|
|
|
|
|
|
|
|
class AttentionAnalyzer: |
|
|
"""Analyze attention tensors and generate visual/debug artifacts.""" |
|
|
|
|
|
def __init__(self, attentions, translation, video_frames, beam_sequences=None, beam_scores=None, |
|
|
video_path=None, original_video_fps=30, original_video_total_frames=None): |
|
|
""" |
|
|
Args: |
|
|
attentions: numpy array, shape [time_steps, batch, beam, src_len] |
|
|
or [time_steps, src_len] (best beam already selected) |
|
|
translation: str, BPE-removed gloss sequence |
|
|
video_frames: int, number of SMKD feature frames |
|
|
beam_sequences: list, optional beam texts |
|
|
beam_scores: list, optional beam scores |
|
|
video_path: str, optional path to original video (for frame grabs) |
|
|
original_video_fps: int, FPS of original video (default 30) |
|
|
original_video_total_frames: optional exact frame count |
|
|
""" |
|
|
self.attentions = attentions |
|
|
self.translation = translation |
|
|
self.words = translation.split() |
|
|
self.video_frames = video_frames |
|
|
self.beam_sequences = beam_sequences |
|
|
self.beam_scores = beam_scores |
|
|
|
|
|
|
|
|
self.video_path = video_path |
|
|
self.original_video_fps = original_video_fps |
|
|
self.original_video_total_frames = original_video_total_frames |
|
|
self._cv2_module = None |
|
|
self._cv2_checked = False |
|
|
|
|
|
|
|
|
if video_path and original_video_total_frames is None: |
|
|
metadata = self._read_video_metadata() |
|
|
if metadata: |
|
|
self.original_video_total_frames = metadata.get('frames') |
|
|
if metadata.get('fps'): |
|
|
self.original_video_fps = metadata['fps'] |
|
|
elif video_path: |
|
|
print(f"Warning: failed to parse video metadata; gloss-to-frame visualization may be misaligned ({video_path})") |
|
|
|
|
|
|
|
|
if len(attentions.shape) == 4: |
|
|
self.attn_best = attentions[:, 0, 0, :] |
|
|
elif len(attentions.shape) == 3: |
|
|
self.attn_best = attentions[:, 0, :] |
|
|
else: |
|
|
self.attn_best = attentions |
|
|
|
|
|
|
|
|
self.word_frame_ranges = self._compute_word_frame_ranges() |
|
|
self.frame_attention_strength = self._compute_frame_attention_strength() |
|
|
|
|
|
def _compute_word_frame_ranges(self): |
|
|
""" |
|
|
Compute the dominant video frame range for each generated word. |
|
|
|
|
|
Returns: |
|
|
list of dict entries containing word, frame range, peak, and confidence. |
|
|
""" |
|
|
word_ranges = [] |
|
|
|
|
|
for word_idx, word in enumerate(self.words): |
|
|
if word_idx >= self.attn_best.shape[0]: |
|
|
|
|
|
word_ranges.append({ |
|
|
'word': word, |
|
|
'start_frame': 0, |
|
|
'end_frame': 0, |
|
|
'peak_frame': 0, |
|
|
'avg_attention': 0.0, |
|
|
'confidence': 'unknown' |
|
|
}) |
|
|
continue |
|
|
|
|
|
attn_weights = self.attn_best[word_idx, :] |
|
|
|
|
|
|
|
|
peak_frame = int(np.argmax(attn_weights)) |
|
|
peak_weight = attn_weights[peak_frame] |
|
|
|
|
|
|
|
|
threshold = peak_weight * 0.9 |
|
|
significant_frames = np.where(attn_weights >= threshold)[0] |
|
|
|
|
|
if len(significant_frames) > 0: |
|
|
start_frame = int(significant_frames[0]) |
|
|
end_frame = int(significant_frames[-1]) |
|
|
avg_weight = float(attn_weights[significant_frames].mean()) |
|
|
else: |
|
|
start_frame = peak_frame |
|
|
end_frame = peak_frame |
|
|
avg_weight = float(peak_weight) |
|
|
|
|
|
|
|
|
if avg_weight > 0.5: |
|
|
confidence = 'high' |
|
|
elif avg_weight > 0.2: |
|
|
confidence = 'medium' |
|
|
else: |
|
|
confidence = 'low' |
|
|
|
|
|
word_ranges.append({ |
|
|
'word': word, |
|
|
'start_frame': start_frame, |
|
|
'end_frame': end_frame, |
|
|
'peak_frame': peak_frame, |
|
|
'avg_attention': avg_weight, |
|
|
'confidence': confidence |
|
|
}) |
|
|
|
|
|
return word_ranges |
|
|
|
|
|
def _compute_frame_attention_strength(self): |
|
|
"""Compute average attention per feature frame (normalized 0-1).""" |
|
|
if self.attn_best.size == 0: |
|
|
return np.zeros(self.video_frames, dtype=np.float32) |
|
|
|
|
|
if self.attn_best.ndim == 1: |
|
|
frame_strength = self.attn_best.copy() |
|
|
else: |
|
|
frame_strength = self.attn_best.mean(axis=0) |
|
|
|
|
|
if frame_strength.shape[0] != self.video_frames: |
|
|
frame_strength = np.resize(frame_strength, self.video_frames) |
|
|
|
|
|
max_val = frame_strength.max() |
|
|
if max_val > 0: |
|
|
frame_strength = frame_strength / max_val |
|
|
return frame_strength |
|
|
|
|
|
def _map_strength_to_original_frames(self, mapping_list, original_frame_count): |
|
|
"""Map latent attention strength to original video frame resolution.""" |
|
|
if not mapping_list or original_frame_count <= 0: |
|
|
return None |
|
|
|
|
|
orig_strength = np.zeros(original_frame_count, dtype=np.float32) |
|
|
counts = np.zeros(original_frame_count, dtype=np.float32) |
|
|
|
|
|
for feat_idx, mapping in enumerate(mapping_list): |
|
|
if feat_idx >= len(self.frame_attention_strength): |
|
|
break |
|
|
start = int(mapping.get('frame_start', 0)) |
|
|
end = int(mapping.get('frame_end', start)) |
|
|
end = max(end, start + 1) |
|
|
start = max(start, 0) |
|
|
end = min(end, original_frame_count) |
|
|
if start >= end: |
|
|
continue |
|
|
orig_strength[start:end] += self.frame_attention_strength[feat_idx] |
|
|
counts[start:end] += 1 |
|
|
|
|
|
mask = counts > 0 |
|
|
if mask.any(): |
|
|
orig_strength[mask] = orig_strength[mask] / counts[mask] |
|
|
|
|
|
max_val = orig_strength.max() |
|
|
if max_val > 0: |
|
|
orig_strength = orig_strength / max_val |
|
|
return orig_strength |
|
|
|
|
|
def generate_all_visualizations(self, output_dir): |
|
|
""" |
|
|
Generate every visualization artifact to the provided directory. |
|
|
""" |
|
|
output_dir = Path(output_dir) |
|
|
output_dir.mkdir(parents=True, exist_ok=True) |
|
|
|
|
|
print(f"\nGenerating visualization assets in: {output_dir}") |
|
|
|
|
|
|
|
|
self.plot_attention_heatmap(output_dir / "attention_heatmap.png") |
|
|
|
|
|
|
|
|
self.plot_frame_alignment(output_dir / "frame_alignment.png") |
|
|
|
|
|
|
|
|
self.save_alignment_data(output_dir / "frame_alignment.json") |
|
|
|
|
|
|
|
|
self.save_text_report(output_dir / "analysis_report.txt") |
|
|
|
|
|
|
|
|
np.save(output_dir / "attention_weights.npy", self.attentions) |
|
|
|
|
|
|
|
|
|
|
|
debug_file = output_dir / "debug_video_path.txt" |
|
|
with open(debug_file, 'w') as f: |
|
|
f.write(f"video_path = {repr(self.video_path)}\n") |
|
|
f.write(f"video_path type = {type(self.video_path)}\n") |
|
|
f.write(f"video_path is None: {self.video_path is None}\n") |
|
|
f.write(f"bool(video_path): {bool(self.video_path)}\n") |
|
|
|
|
|
print(f"[DEBUG] video_path = {self.video_path}") |
|
|
if self.video_path: |
|
|
print(f"[DEBUG] Generating gloss-to-frames visualization with video: {self.video_path}") |
|
|
try: |
|
|
self.generate_gloss_to_frames_visualization(output_dir / "gloss_to_frames.png") |
|
|
print(f"[DEBUG] Successfully generated gloss_to_frames.png") |
|
|
except Exception as e: |
|
|
print(f"[DEBUG] Failed to generate gloss_to_frames.png: {e}") |
|
|
import traceback |
|
|
traceback.print_exc() |
|
|
else: |
|
|
print("[DEBUG] Skipping gloss-to-frames visualization (no video path provided)") |
|
|
|
|
|
print(f"✓ Wrote {len(list(output_dir.glob('*')))} file(s)") |
|
|
|
|
|
def plot_attention_heatmap(self, output_path): |
|
|
"""Render the attention heatmap (image + PDF copy).""" |
|
|
try: |
|
|
import matplotlib |
|
|
matplotlib.use('Agg') |
|
|
import matplotlib.pyplot as plt |
|
|
except ImportError: |
|
|
print(" Skipping heatmap: matplotlib is not available") |
|
|
return |
|
|
|
|
|
fig, ax = plt.subplots(figsize=(14, 8)) |
|
|
|
|
|
|
|
|
im = ax.imshow(self.attn_best.T, cmap='hot', aspect='auto', |
|
|
interpolation='nearest', origin='lower') |
|
|
|
|
|
|
|
|
ax.set_xlabel('Generated Word Index', fontsize=13) |
|
|
ax.set_ylabel('Video Frame Index', fontsize=13) |
|
|
ax.set_title('Cross-Attention Weights\n(Decoder → Video Frames)', |
|
|
fontsize=15, pad=20, fontweight='bold') |
|
|
|
|
|
|
|
|
if len(self.words) <= self.attn_best.shape[0]: |
|
|
ax.set_xticks(range(len(self.words))) |
|
|
ax.set_xticklabels(self.words, rotation=45, ha='right', fontsize=10) |
|
|
|
|
|
|
|
|
cbar = plt.colorbar(im, ax=ax, label='Attention Weight', fraction=0.046, pad=0.04) |
|
|
cbar.ax.tick_params(labelsize=10) |
|
|
|
|
|
plt.tight_layout() |
|
|
plt.savefig(output_path, dpi=150, bbox_inches='tight') |
|
|
|
|
|
pdf_path = Path(output_path).with_suffix('.pdf') |
|
|
plt.savefig(str(pdf_path), format='pdf', bbox_inches='tight') |
|
|
plt.close() |
|
|
|
|
|
print(f" ✓ {output_path.name} (PDF copy saved)") |
|
|
|
|
|
def plot_frame_alignment(self, output_path): |
|
|
"""Render the frame-alignment charts (full + compact).""" |
|
|
try: |
|
|
import matplotlib |
|
|
matplotlib.use('Agg') |
|
|
import matplotlib.pyplot as plt |
|
|
import matplotlib.patches as patches |
|
|
from matplotlib.gridspec import GridSpec |
|
|
except ImportError: |
|
|
print(" Skipping alignment plot: matplotlib is not available") |
|
|
return |
|
|
|
|
|
output_path = Path(output_path) |
|
|
|
|
|
|
|
|
feature_mapping = None |
|
|
output_dir = output_path.parent |
|
|
mapping_file = output_dir / "feature_frame_mapping.json" |
|
|
if mapping_file.exists(): |
|
|
try: |
|
|
with open(mapping_file, 'r') as f: |
|
|
feature_mapping = json.load(f) |
|
|
except Exception as e: |
|
|
print(f" Warning: Failed to load feature mapping: {e}") |
|
|
|
|
|
if self.word_frame_ranges: |
|
|
max_feat_end = max(w['end_frame'] for w in self.word_frame_ranges) |
|
|
else: |
|
|
max_feat_end = self.video_frames - 1 |
|
|
latent_full_limit = self.video_frames + 2 |
|
|
latent_short_limit = max(min(latent_full_limit, max_feat_end + 2), 5) |
|
|
|
|
|
original_frame_count = None |
|
|
mapping_list = None |
|
|
orig_full_limit = None |
|
|
orig_short_limit = None |
|
|
pixel_strength_curve = None |
|
|
if feature_mapping: |
|
|
original_frame_count = feature_mapping.get('original_frame_count', self.video_frames) |
|
|
mapping_list = feature_mapping.get('mapping', []) |
|
|
orig_full_limit = original_frame_count + 2 |
|
|
if mapping_list: |
|
|
idx = min(max_feat_end, len(mapping_list) - 1) |
|
|
orig_short_limit = mapping_list[idx]['frame_end'] + 2 |
|
|
pixel_strength_curve = self._map_strength_to_original_frames(mapping_list, original_frame_count) |
|
|
|
|
|
def render_alignment(out_path, latent_xlim_end, orig_xlim_end=None): |
|
|
if feature_mapping: |
|
|
fig = plt.figure(figsize=(18, 9)) |
|
|
gs = GridSpec(3, 1, height_ratios=[4, 1, 1], hspace=0.32) |
|
|
else: |
|
|
fig = plt.figure(figsize=(18, 7.5)) |
|
|
gs = GridSpec(2, 1, height_ratios=[4, 1], hspace=0.32) |
|
|
|
|
|
|
|
|
ax1 = fig.add_subplot(gs[0]) |
|
|
colors = plt.cm.tab20(np.linspace(0, 1, max(len(self.words), 20))) |
|
|
|
|
|
for i, word_info in enumerate(self.word_frame_ranges): |
|
|
start = word_info['start_frame'] |
|
|
end = word_info['end_frame'] |
|
|
word = word_info['word'] |
|
|
confidence = word_info['confidence'] |
|
|
alpha = 0.9 if confidence == 'high' else 0.7 if confidence == 'medium' else 0.5 |
|
|
|
|
|
rect = patches.Rectangle( |
|
|
(start, i), end - start + 1, 0.8, |
|
|
linewidth=2, edgecolor='black', |
|
|
facecolor=colors[i % 20], alpha=alpha |
|
|
) |
|
|
ax1.add_patch(rect) |
|
|
|
|
|
ax1.text(start + (end - start) / 2, i + 0.4, word, |
|
|
ha='center', va='center', fontsize=11, |
|
|
fontweight='bold', color='white', |
|
|
bbox=dict(boxstyle='round,pad=0.3', facecolor='black', alpha=0.5)) |
|
|
|
|
|
peak = word_info['peak_frame'] |
|
|
ax1.plot(peak, i + 0.4, 'r*', markersize=15, markeredgecolor='yellow', |
|
|
markeredgewidth=1.5) |
|
|
|
|
|
ax1.set_xlim(-2, latent_xlim_end) |
|
|
ax1.set_ylim(-0.5, len(self.words)) |
|
|
ax1.set_xlabel('') |
|
|
ax1.set_ylabel('') |
|
|
ax1.set_title('Word-to-Frame Alignment\n(based on attention peaks, ★ = peak frame)', |
|
|
fontsize=15, pad=15, fontweight='bold') |
|
|
ax1.grid(True, alpha=0.3, axis='x', linestyle='--') |
|
|
ax1.set_yticks(range(len(self.words))) |
|
|
ax1.set_yticklabels([w['word'] for w in self.word_frame_ranges], fontsize=10) |
|
|
|
|
|
|
|
|
ax2 = fig.add_subplot(gs[1]) |
|
|
ax2.barh(0, self.video_frames, height=0.6, color='lightgray', |
|
|
edgecolor='black', linewidth=2) |
|
|
for i, word_info in enumerate(self.word_frame_ranges): |
|
|
start = word_info['start_frame'] |
|
|
end = word_info['end_frame'] |
|
|
confidence = word_info['confidence'] |
|
|
alpha = 0.9 if confidence == 'high' else 0.7 if confidence == 'medium' else 0.5 |
|
|
ax2.barh(0, end - start + 1, left=start, height=0.6, |
|
|
color=colors[i % 20], alpha=alpha, edgecolor='black', linewidth=0.5) |
|
|
|
|
|
ax2.set_xlim(-2, latent_xlim_end) |
|
|
ax2.set_ylim(-0.4, 0.4) |
|
|
ax2.set_xlabel('') |
|
|
ax2.set_yticks([0]) |
|
|
ax2.set_yticklabels(['Latent Space'], fontsize=11, fontweight='bold') |
|
|
ax2.tick_params(axis='y', length=0) |
|
|
ax2.set_title('Latent Feature Timeline', fontsize=13, fontweight='bold') |
|
|
ax2.grid(True, alpha=0.3, axis='x', linestyle='--') |
|
|
|
|
|
if self.frame_attention_strength is not None and len(self.frame_attention_strength) >= self.video_frames: |
|
|
latent_curve_x = np.arange(self.video_frames) |
|
|
latent_curve_y = self.frame_attention_strength[:self.video_frames] * 0.6 - 0.3 |
|
|
ax2.plot(latent_curve_x, latent_curve_y, color='#E53935', linewidth=1.5, alpha=0.9) |
|
|
|
|
|
timeline_axes = [ax2] |
|
|
|
|
|
if feature_mapping: |
|
|
ax3 = fig.add_subplot(gs[2]) |
|
|
ax3.barh(0, original_frame_count, height=0.6, color='lightgray', |
|
|
edgecolor='black', linewidth=2) |
|
|
|
|
|
for i, word_info in enumerate(self.word_frame_ranges): |
|
|
feat_start = word_info['start_frame'] |
|
|
feat_end = word_info['end_frame'] |
|
|
confidence = word_info['confidence'] |
|
|
alpha = 0.9 if confidence == 'high' else 0.7 if confidence == 'medium' else 0.5 |
|
|
if mapping_list and feat_start < len(mapping_list) and feat_end < len(mapping_list): |
|
|
orig_start = mapping_list[feat_start]['frame_start'] |
|
|
orig_end = mapping_list[feat_end]['frame_end'] |
|
|
ax3.barh(0, orig_end - orig_start, left=orig_start, height=0.6, |
|
|
color=colors[i % 20], alpha=alpha, edgecolor='black', linewidth=0.5) |
|
|
|
|
|
ax3_xlim = orig_xlim_end if orig_xlim_end is not None else original_frame_count + 2 |
|
|
ax3.set_xlim(-2, ax3_xlim) |
|
|
ax3.set_ylim(-0.4, 0.4) |
|
|
ax3.set_xlabel('') |
|
|
ax3.set_yticks([0]) |
|
|
ax3.set_yticklabels(['Pixel Space'], fontsize=11, fontweight='bold') |
|
|
ax3.tick_params(axis='y', length=0) |
|
|
ax3.set_title(f'Original Video Timeline ({original_frame_count} frames, ' |
|
|
f'{feature_mapping["downsampling_ratio"]:.2f}x downsampling)', |
|
|
fontsize=13, fontweight='bold') |
|
|
ax3.grid(True, alpha=0.3, axis='x', linestyle='--') |
|
|
|
|
|
if pixel_strength_curve is not None and len(pixel_strength_curve) >= original_frame_count: |
|
|
pixel_curve_x = np.arange(original_frame_count) |
|
|
pixel_curve_y = pixel_strength_curve[:original_frame_count] * 0.6 - 0.3 |
|
|
ax3.plot(pixel_curve_x, pixel_curve_y, color='#E53935', linewidth=1.5, alpha=0.9) |
|
|
timeline_axes.append(ax3) |
|
|
|
|
|
plt.tight_layout() |
|
|
fig.canvas.draw() |
|
|
|
|
|
ax1_pos = ax1.get_position() |
|
|
renderer = fig.canvas.get_renderer() |
|
|
ytick_extents = [tick.get_window_extent(renderer) for tick in ax1.get_yticklabels() if tick.get_text()] |
|
|
fig_width_px = fig.get_size_inches()[0] * fig.dpi |
|
|
if ytick_extents: |
|
|
min_x_px = min(ext.x0 for ext in ytick_extents) |
|
|
else: |
|
|
min_x_px = ax1_pos.x0 * fig_width_px |
|
|
line_x = max(0.01, (min_x_px / fig_width_px) - 0.01) |
|
|
gw_center = 0.5 * (ax1_pos.y0 + ax1_pos.y1) |
|
|
timeline_bounds = [ax.get_position() for ax in timeline_axes] |
|
|
timeline_center = 0.5 * (min(pos.y0 for pos in timeline_bounds) + max(pos.y1 for pos in timeline_bounds)) |
|
|
fig.text(line_x, gw_center, 'Generated Word', rotation='vertical', |
|
|
ha='right', va='center', fontsize=12, fontweight='bold') |
|
|
fig.text(line_x, timeline_center, 'Timeline', rotation='vertical', |
|
|
ha='right', va='center', fontsize=12, fontweight='bold') |
|
|
|
|
|
png_path = Path(out_path) |
|
|
plt.savefig(str(png_path), dpi=150, bbox_inches='tight') |
|
|
pdf_path = png_path.with_suffix('.pdf') |
|
|
plt.savefig(str(pdf_path), format='pdf', bbox_inches='tight') |
|
|
plt.close() |
|
|
|
|
|
print(f" ✓ {png_path.name} (PDF copy saved)") |
|
|
|
|
|
render_alignment(output_path, latent_full_limit, orig_full_limit) |
|
|
|
|
|
if latent_short_limit < latent_full_limit - 1e-6: |
|
|
short_path = output_path.with_name("frame_alignment_short.png") |
|
|
render_alignment(short_path, latent_short_limit, orig_short_limit if orig_short_limit else orig_full_limit) |
|
|
|
|
|
def save_alignment_data(self, output_path): |
|
|
"""Persist frame-alignment metadata to JSON.""" |
|
|
data = { |
|
|
'translation': self.translation, |
|
|
'words': self.words, |
|
|
'total_video_frames': self.video_frames, |
|
|
'frame_ranges': self.word_frame_ranges, |
|
|
'statistics': { |
|
|
'avg_confidence': np.mean([w['avg_attention'] for w in self.word_frame_ranges]), |
|
|
'high_confidence_words': sum(1 for w in self.word_frame_ranges if w['confidence'] == 'high'), |
|
|
'medium_confidence_words': sum(1 for w in self.word_frame_ranges if w['confidence'] == 'medium'), |
|
|
'low_confidence_words': sum(1 for w in self.word_frame_ranges if w['confidence'] == 'low'), |
|
|
} |
|
|
} |
|
|
|
|
|
with open(output_path, 'w', encoding='utf-8') as f: |
|
|
json.dump(data, f, indent=2, ensure_ascii=False) |
|
|
|
|
|
print(f" ✓ {output_path.name}") |
|
|
|
|
|
def save_text_report(self, output_path): |
|
|
"""Write a plain-text report (used for analysis_report.txt).""" |
|
|
with open(output_path, 'w', encoding='utf-8') as f: |
|
|
f.write("=" * 80 + "\n") |
|
|
f.write(" Sign Language Recognition - Attention Analysis Report\n") |
|
|
f.write("=" * 80 + "\n\n") |
|
|
|
|
|
f.write(f"Generated at: {datetime.now().strftime('%Y-%m-%d %H:%M:%S')}\n\n") |
|
|
|
|
|
f.write("Translation:\n") |
|
|
f.write("-" * 80 + "\n") |
|
|
f.write(f"{self.translation}\n\n") |
|
|
|
|
|
f.write("Video info:\n") |
|
|
f.write("-" * 80 + "\n") |
|
|
f.write(f"Total feature frames: {self.video_frames}\n") |
|
|
f.write(f"Word count: {len(self.words)}\n\n") |
|
|
|
|
|
f.write("Attention tensor:\n") |
|
|
f.write("-" * 80 + "\n") |
|
|
f.write(f"Shape: {self.attentions.shape}\n") |
|
|
f.write(f" - Decoder steps: {self.attentions.shape[0]}\n") |
|
|
if len(self.attentions.shape) >= 3: |
|
|
f.write(f" - Batch size: {self.attentions.shape[1]}\n") |
|
|
if len(self.attentions.shape) >= 4: |
|
|
f.write(f" - Beam size: {self.attentions.shape[2]}\n") |
|
|
f.write(f" - Source length: {self.attentions.shape[3]}\n") |
|
|
f.write("\n") |
|
|
|
|
|
f.write("Word-to-frame details:\n") |
|
|
f.write("=" * 80 + "\n") |
|
|
f.write(f"{'No.':<5} {'Word':<20} {'Frames':<15} {'Peak':<8} {'Attn':<8} {'Conf':<10}\n") |
|
|
f.write("-" * 80 + "\n") |
|
|
|
|
|
for i, w in enumerate(self.word_frame_ranges): |
|
|
frame_range = f"{w['start_frame']}-{w['end_frame']}" |
|
|
f.write(f"{i+1:<5} {w['word']:<20} {frame_range:<15} " |
|
|
f"{w['peak_frame']:<8} {w['avg_attention']:<8.3f} {w['confidence']:<10}\n") |
|
|
|
|
|
f.write("\n" + "=" * 80 + "\n") |
|
|
|
|
|
|
|
|
stats = { |
|
|
'avg_confidence': np.mean([w['avg_attention'] for w in self.word_frame_ranges]), |
|
|
'high': sum(1 for w in self.word_frame_ranges if w['confidence'] == 'high'), |
|
|
'medium': sum(1 for w in self.word_frame_ranges if w['confidence'] == 'medium'), |
|
|
'low': sum(1 for w in self.word_frame_ranges if w['confidence'] == 'low'), |
|
|
} |
|
|
|
|
|
f.write("\nSummary:\n") |
|
|
f.write("-" * 80 + "\n") |
|
|
f.write(f"Average attention weight: {stats['avg_confidence']:.3f}\n") |
|
|
f.write(f"High-confidence words: {stats['high']} ({stats['high']/len(self.words)*100:.1f}%)\n") |
|
|
f.write(f"Medium-confidence words: {stats['medium']} ({stats['medium']/len(self.words)*100:.1f}%)\n") |
|
|
f.write(f"Low-confidence words: {stats['low']} ({stats['low']/len(self.words)*100:.1f}%)\n") |
|
|
f.write("\n" + "=" * 80 + "\n") |
|
|
|
|
|
print(f" ✓ {output_path.name}") |
|
|
|
|
|
|
|
|
def _map_feature_frame_to_original(self, feature_frame_idx): |
|
|
""" |
|
|
Map a SMKD feature frame index back to the original video frame index. |
|
|
|
|
|
Args: |
|
|
feature_frame_idx: Zero-based feature frame index |
|
|
|
|
|
Returns: |
|
|
int: Original frame index, or None if unavailable. |
|
|
""" |
|
|
if self.original_video_total_frames is None: |
|
|
return None |
|
|
|
|
|
|
|
|
downsample_ratio = self.original_video_total_frames / self.video_frames |
|
|
|
|
|
|
|
|
original_frame_idx = int(feature_frame_idx * downsample_ratio) |
|
|
|
|
|
return min(original_frame_idx, self.original_video_total_frames - 1) |
|
|
|
|
|
def _extract_video_frames(self, frame_indices): |
|
|
""" |
|
|
Extract the requested original video frames (best-effort). |
|
|
|
|
|
Args: |
|
|
frame_indices: list[int] of original frame IDs to load |
|
|
|
|
|
Returns: |
|
|
dict mapping frame index to numpy array (BGR). |
|
|
""" |
|
|
if not self.video_path: |
|
|
return {} |
|
|
|
|
|
cv2 = self._get_cv2_module() |
|
|
if cv2 is not None: |
|
|
return self._extract_frames_with_cv2(cv2, frame_indices) |
|
|
|
|
|
return self._extract_frames_with_ffmpeg(frame_indices) |
|
|
|
|
|
def _get_cv2_module(self): |
|
|
"""Lazy-load cv2 and cache the import outcome.""" |
|
|
if self._cv2_checked: |
|
|
return self._cv2_module |
|
|
|
|
|
try: |
|
|
import cv2 |
|
|
self._cv2_module = cv2 |
|
|
except ImportError: |
|
|
self._cv2_module = None |
|
|
finally: |
|
|
self._cv2_checked = True |
|
|
|
|
|
if self._cv2_module is None: |
|
|
print("Warning: opencv-python is missing; falling back to ffmpeg grabs") |
|
|
return self._cv2_module |
|
|
|
|
|
def _extract_frames_with_cv2(self, cv2, frame_indices): |
|
|
"""Extract frames via OpenCV if available.""" |
|
|
frames = {} |
|
|
cap = cv2.VideoCapture(self.video_path) |
|
|
|
|
|
if not cap.isOpened(): |
|
|
print(f"Warning: Cannot open video file: {self.video_path}") |
|
|
return {} |
|
|
|
|
|
for frame_idx in sorted(frame_indices): |
|
|
cap.set(cv2.CAP_PROP_POS_FRAMES, frame_idx) |
|
|
ret, frame = cap.read() |
|
|
if ret: |
|
|
frame_rgb = cv2.cvtColor(frame, cv2.COLOR_BGR2RGB) |
|
|
frames[frame_idx] = frame_rgb |
|
|
|
|
|
cap.release() |
|
|
return frames |
|
|
|
|
|
def _extract_frames_with_ffmpeg(self, frame_indices): |
|
|
"""Extract frames via ffmpeg + Pillow (OpenCV fallback).""" |
|
|
if shutil.which("ffmpeg") is None: |
|
|
print("Warning: ffmpeg not found; cannot extract frames") |
|
|
return {} |
|
|
|
|
|
try: |
|
|
from PIL import Image |
|
|
except ImportError: |
|
|
print("Warning: Pillow not installed; cannot decode ffmpeg output") |
|
|
return {} |
|
|
|
|
|
frames = {} |
|
|
for frame_idx in sorted(frame_indices): |
|
|
cmd = [ |
|
|
"ffmpeg", |
|
|
"-v", "error", |
|
|
"-i", str(self.video_path), |
|
|
"-vf", f"select=eq(n\\,{frame_idx})", |
|
|
"-vframes", "1", |
|
|
"-f", "image2pipe", |
|
|
"-vcodec", "png", |
|
|
"-" |
|
|
] |
|
|
try: |
|
|
result = subprocess.run( |
|
|
cmd, check=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE |
|
|
) |
|
|
if not result.stdout: |
|
|
continue |
|
|
image = Image.open(io.BytesIO(result.stdout)).convert("RGB") |
|
|
frames[frame_idx] = np.array(image) |
|
|
except subprocess.CalledProcessError as e: |
|
|
print(f"Warning: ffmpeg failed to extract frame {frame_idx}: {e}") |
|
|
except Exception as ex: |
|
|
print(f"Warning: failed to decode frame {frame_idx}: {ex}") |
|
|
|
|
|
if frames: |
|
|
print(f" ✓ Extracted {len(frames)} frame(s) via ffmpeg") |
|
|
else: |
|
|
print(" ⓘ ffmpeg did not return any frames") |
|
|
return frames |
|
|
|
|
|
def generate_gloss_to_frames_visualization(self, output_path): |
|
|
""" |
|
|
Create the gloss-to-frames visualization: |
|
|
Column 1: gloss text |
|
|
Column 2: relative time + frame indices |
|
|
Column 3: representative video thumbnails |
|
|
""" |
|
|
if not self.video_path: |
|
|
print(" ⓘ Skipping gloss-to-frames visualization (no video path provided)") |
|
|
return |
|
|
|
|
|
try: |
|
|
import matplotlib.pyplot as plt |
|
|
import matplotlib.gridspec as gridspec |
|
|
except ImportError: |
|
|
print("Warning: matplotlib not installed") |
|
|
return |
|
|
|
|
|
|
|
|
feature_mapping = None |
|
|
output_dir = Path(output_path).parent |
|
|
mapping_file = output_dir / "feature_frame_mapping.json" |
|
|
if mapping_file.exists(): |
|
|
try: |
|
|
with open(mapping_file, 'r') as f: |
|
|
mapping_data = json.load(f) |
|
|
feature_mapping = mapping_data['mapping'] |
|
|
except Exception as e: |
|
|
print(f" Warning: Failed to load feature mapping: {e}") |
|
|
|
|
|
|
|
|
all_original_frames = set() |
|
|
for word_info in self.word_frame_ranges: |
|
|
|
|
|
start_feat = word_info['start_frame'] |
|
|
end_feat = word_info['end_frame'] |
|
|
|
|
|
|
|
|
if feature_mapping: |
|
|
|
|
|
for feat_idx in range(start_feat, end_feat + 1): |
|
|
if feat_idx < len(feature_mapping): |
|
|
|
|
|
feat_info = feature_mapping[feat_idx] |
|
|
for orig_idx in range(feat_info['frame_start'], feat_info['frame_end']): |
|
|
all_original_frames.add(orig_idx) |
|
|
else: |
|
|
|
|
|
for feat_idx in range(start_feat, end_feat + 1): |
|
|
orig_idx = self._map_feature_frame_to_original(feat_idx) |
|
|
if orig_idx is not None: |
|
|
all_original_frames.add(orig_idx) |
|
|
|
|
|
|
|
|
print(f" Extracting {len(all_original_frames)} original video frame(s)...") |
|
|
video_frames_dict = self._extract_video_frames(list(all_original_frames)) |
|
|
|
|
|
if not video_frames_dict: |
|
|
print(" ⓘ No video frames extracted, skipping visualization") |
|
|
return |
|
|
|
|
|
|
|
|
n_words = len(self.words) |
|
|
fig = plt.figure(figsize=(28, 3 * n_words)) |
|
|
gs = gridspec.GridSpec(n_words, 4, width_ratios=[1.5, 1.5, 2.5, 8], hspace=0.3, wspace=0.2) |
|
|
|
|
|
for row_idx, (word, word_info) in enumerate(zip(self.words, self.word_frame_ranges)): |
|
|
|
|
|
ax_gloss = fig.add_subplot(gs[row_idx, 0]) |
|
|
ax_gloss.text(0.5, 0.5, word, fontsize=24, weight='bold', |
|
|
ha='center', va='center', wrap=True) |
|
|
ax_gloss.axis('off') |
|
|
|
|
|
|
|
|
ax_feature = fig.add_subplot(gs[row_idx, 1]) |
|
|
|
|
|
|
|
|
feat_start = word_info['start_frame'] |
|
|
feat_end = word_info['end_frame'] |
|
|
feat_peak = word_info['peak_frame'] |
|
|
|
|
|
feature_text = f"SMKD Feature Index\n" |
|
|
feature_text += f"{'='*20}\n\n" |
|
|
feature_text += f"Range:\n {feat_start} → {feat_end}\n\n" |
|
|
feature_text += f"Peak:\n {feat_peak}\n\n" |
|
|
feature_text += f"Count:\n {feat_end - feat_start + 1} features" |
|
|
|
|
|
ax_feature.text(0.5, 0.5, feature_text, fontsize=11, family='monospace', |
|
|
va='center', ha='center', |
|
|
bbox=dict(boxstyle='round,pad=0.8', facecolor='lightblue', |
|
|
edgecolor='darkblue', linewidth=2, alpha=0.7)) |
|
|
ax_feature.axis('off') |
|
|
|
|
|
|
|
|
ax_peak_frames = fig.add_subplot(gs[row_idx, 2]) |
|
|
|
|
|
peak_frames_to_show = [] |
|
|
orig_peak_start, orig_peak_end = None, None |
|
|
if feature_mapping and feat_peak is not None and feat_peak < len(feature_mapping): |
|
|
|
|
|
peak_info = feature_mapping[feat_peak] |
|
|
orig_peak_start = peak_info['frame_start'] |
|
|
orig_peak_end = peak_info['frame_end'] |
|
|
|
|
|
|
|
|
for orig_idx in range(orig_peak_start, orig_peak_end): |
|
|
if orig_idx in video_frames_dict: |
|
|
peak_frames_to_show.append(video_frames_dict[orig_idx]) |
|
|
|
|
|
if peak_frames_to_show: |
|
|
|
|
|
combined_peak = np.hstack(peak_frames_to_show) |
|
|
ax_peak_frames.imshow(combined_peak) |
|
|
|
|
|
|
|
|
ax_peak_frames.text(0.5, -0.05, f"Peak Feature {feat_peak}\nFrames {orig_peak_start}-{orig_peak_end-1} ({len(peak_frames_to_show)} frames)", |
|
|
ha='center', va='top', transform=ax_peak_frames.transAxes, |
|
|
fontsize=10, weight='bold', color='red', |
|
|
bbox=dict(boxstyle='round,pad=0.3', facecolor='yellow', alpha=0.7)) |
|
|
else: |
|
|
ax_peak_frames.text(0.5, 0.5, "No peak frames", |
|
|
ha='center', va='center', transform=ax_peak_frames.transAxes) |
|
|
|
|
|
ax_peak_frames.axis('off') |
|
|
|
|
|
|
|
|
ax_all_frames = fig.add_subplot(gs[row_idx, 3]) |
|
|
|
|
|
all_frames_to_show = [] |
|
|
orig_start, orig_end = None, None |
|
|
if feature_mapping: |
|
|
|
|
|
if feat_start < len(feature_mapping) and feat_end < len(feature_mapping): |
|
|
orig_start = feature_mapping[feat_start]['frame_start'] |
|
|
orig_end = feature_mapping[feat_end]['frame_end'] |
|
|
|
|
|
|
|
|
for orig_idx in range(orig_start, orig_end): |
|
|
if orig_idx in video_frames_dict: |
|
|
all_frames_to_show.append(video_frames_dict[orig_idx]) |
|
|
|
|
|
if all_frames_to_show: |
|
|
|
|
|
combined_all = np.hstack(all_frames_to_show) |
|
|
ax_all_frames.imshow(combined_all) |
|
|
|
|
|
|
|
|
frame_count = len(all_frames_to_show) |
|
|
ax_all_frames.text(0.5, -0.05, f"All Frames ({frame_count} frames)\nRange: {orig_start}-{orig_end-1}", |
|
|
ha='center', va='top', transform=ax_all_frames.transAxes, |
|
|
fontsize=10, weight='bold', color='blue', |
|
|
bbox=dict(boxstyle='round,pad=0.3', facecolor='lightblue', alpha=0.7)) |
|
|
else: |
|
|
ax_all_frames.text(0.5, 0.5, "No frames available", |
|
|
ha='center', va='center', transform=ax_all_frames.transAxes) |
|
|
|
|
|
ax_all_frames.axis('off') |
|
|
|
|
|
plt.suptitle(f"Three-Layer Alignment: Gloss ↔ Feature Index ↔ Original Frames\nTranslation: {self.translation}", |
|
|
fontsize=16, weight='bold', y=0.995) |
|
|
|
|
|
plt.savefig(output_path, dpi=150, bbox_inches='tight', facecolor='white') |
|
|
plt.close() |
|
|
|
|
|
print(f" ✓ {Path(output_path).name}") |
|
|
|
|
|
def _read_video_metadata(self): |
|
|
"""Attempt to read the original video's frame count and FPS.""" |
|
|
metadata = self._read_metadata_with_cv2() |
|
|
if metadata: |
|
|
return metadata |
|
|
return self._read_metadata_with_ffprobe() |
|
|
|
|
|
def _read_metadata_with_cv2(self): |
|
|
cv2 = self._get_cv2_module() |
|
|
if cv2 is None: |
|
|
return None |
|
|
|
|
|
cap = cv2.VideoCapture(self.video_path) |
|
|
if not cap.isOpened(): |
|
|
return None |
|
|
|
|
|
total_frames = int(cap.get(cv2.CAP_PROP_FRAME_COUNT)) |
|
|
fps = cap.get(cv2.CAP_PROP_FPS) |
|
|
cap.release() |
|
|
|
|
|
if total_frames <= 0: |
|
|
return None |
|
|
|
|
|
return {'frames': total_frames, 'fps': fps or self.original_video_fps} |
|
|
|
|
|
def _read_metadata_with_ffprobe(self): |
|
|
if shutil.which("ffprobe") is None: |
|
|
return None |
|
|
|
|
|
cmd = [ |
|
|
"ffprobe", |
|
|
"-v", "error", |
|
|
"-select_streams", "v:0", |
|
|
"-show_entries", "stream=nb_frames,r_frame_rate,avg_frame_rate,duration", |
|
|
"-of", "json", |
|
|
str(self.video_path) |
|
|
] |
|
|
|
|
|
try: |
|
|
result = subprocess.run( |
|
|
cmd, check=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE, text=True |
|
|
) |
|
|
except subprocess.CalledProcessError: |
|
|
return None |
|
|
|
|
|
try: |
|
|
info = json.loads(result.stdout) |
|
|
except json.JSONDecodeError: |
|
|
return None |
|
|
|
|
|
streams = info.get("streams") or [] |
|
|
if not streams: |
|
|
return None |
|
|
|
|
|
stream = streams[0] |
|
|
total_frames = stream.get("nb_frames") |
|
|
fps = stream.get("avg_frame_rate") or stream.get("r_frame_rate") |
|
|
duration = stream.get("duration") |
|
|
|
|
|
fps_value = self._parse_ffprobe_fps(fps) |
|
|
total_frames_value = None |
|
|
|
|
|
if isinstance(total_frames, str) and total_frames.isdigit(): |
|
|
total_frames_value = int(total_frames) |
|
|
|
|
|
if total_frames_value is None and duration and fps_value: |
|
|
try: |
|
|
total_frames_value = int(round(float(duration) * fps_value)) |
|
|
except ValueError: |
|
|
total_frames_value = None |
|
|
|
|
|
if total_frames_value is None: |
|
|
return None |
|
|
|
|
|
return {'frames': total_frames_value, 'fps': fps_value or self.original_video_fps} |
|
|
|
|
|
@staticmethod |
|
|
def _parse_ffprobe_fps(rate_str): |
|
|
"""Parse an ffprobe frame-rate string such as '30000/1001'.""" |
|
|
if not rate_str or rate_str in ("0/0", "0"): |
|
|
return None |
|
|
|
|
|
if "/" in rate_str: |
|
|
num, denom = rate_str.split("/", 1) |
|
|
try: |
|
|
num = float(num) |
|
|
denom = float(denom) |
|
|
if denom == 0: |
|
|
return None |
|
|
return num / denom |
|
|
except ValueError: |
|
|
return None |
|
|
|
|
|
try: |
|
|
return float(rate_str) |
|
|
except ValueError: |
|
|
return None |
|
|
|
|
|
|
|
|
def analyze_from_numpy_file(attention_file, translation, video_frames, output_dir): |
|
|
""" |
|
|
Load attention weights from a .npy file and generate visualization assets. |
|
|
|
|
|
Args: |
|
|
attention_file: Path to the numpy file |
|
|
translation: Clean translation string |
|
|
video_frames: Number of SMKD feature frames |
|
|
output_dir: Destination directory for outputs |
|
|
""" |
|
|
attentions = np.load(attention_file) |
|
|
analyzer = AttentionAnalyzer(attentions, translation, video_frames) |
|
|
analyzer.generate_all_visualizations(output_dir) |
|
|
return analyzer |
|
|
|