ASLLRP_utterances_results / SignX /eval /attention_analysis.py
FangSen9000
The reasoning has been converted into English.
9f9e779
#!/usr/bin/env python3
"""
Attention weight analysis and visualization helpers for SignX.
Capabilities:
1. Parse attention weight tensors
2. Map each generated gloss to video frame ranges
3. Render visual assets (heatmaps, alignment plots, timelines)
4. Write detailed analysis reports
Example:
from eval.attention_analysis import AttentionAnalyzer
analyzer = AttentionAnalyzer(
attentions=attention_weights, # [time, batch, beam, src_len]
translation="WORD1 WORD2 WORD3",
video_frames=100
)
analyzer.generate_all_visualizations(output_dir="results/")
"""
import os
import io
import json
import shutil
import subprocess
import numpy as np
from pathlib import Path
from datetime import datetime
class AttentionAnalyzer:
"""Analyze attention tensors and generate visual/debug artifacts."""
def __init__(self, attentions, translation, video_frames, beam_sequences=None, beam_scores=None,
video_path=None, original_video_fps=30, original_video_total_frames=None):
"""
Args:
attentions: numpy array, shape [time_steps, batch, beam, src_len]
or [time_steps, src_len] (best beam already selected)
translation: str, BPE-removed gloss sequence
video_frames: int, number of SMKD feature frames
beam_sequences: list, optional beam texts
beam_scores: list, optional beam scores
video_path: str, optional path to original video (for frame grabs)
original_video_fps: int, FPS of original video (default 30)
original_video_total_frames: optional exact frame count
"""
self.attentions = attentions
self.translation = translation
self.words = translation.split()
self.video_frames = video_frames
self.beam_sequences = beam_sequences
self.beam_scores = beam_scores
# Video metadata
self.video_path = video_path
self.original_video_fps = original_video_fps
self.original_video_total_frames = original_video_total_frames
self._cv2_module = None
self._cv2_checked = False
# Auto-read metadata if only video path is given
if video_path and original_video_total_frames is None:
metadata = self._read_video_metadata()
if metadata:
self.original_video_total_frames = metadata.get('frames')
if metadata.get('fps'):
self.original_video_fps = metadata['fps']
elif video_path:
print(f"Warning: failed to parse video metadata; gloss-to-frame visualization may be misaligned ({video_path})")
# Always operate on the best path (batch=0, beam=0)
if len(attentions.shape) == 4:
self.attn_best = attentions[:, 0, 0, :] # [time, src_len]
elif len(attentions.shape) == 3:
self.attn_best = attentions[:, 0, :] # [time, src_len]
else:
self.attn_best = attentions # [time, src_len]
# Pre-compute gloss-to-frame ranges
self.word_frame_ranges = self._compute_word_frame_ranges()
self.frame_attention_strength = self._compute_frame_attention_strength()
def _compute_word_frame_ranges(self):
"""
Compute the dominant video frame range for each generated word.
Returns:
list of dict entries containing word, frame range, peak, and confidence.
"""
word_ranges = []
for word_idx, word in enumerate(self.words):
if word_idx >= self.attn_best.shape[0]:
# Out of range
word_ranges.append({
'word': word,
'start_frame': 0,
'end_frame': 0,
'peak_frame': 0,
'avg_attention': 0.0,
'confidence': 'unknown'
})
continue
attn_weights = self.attn_best[word_idx, :]
# Peak frame for this word
peak_frame = int(np.argmax(attn_weights))
peak_weight = attn_weights[peak_frame]
# Frames whose weight >= 90% of the peak
threshold = peak_weight * 0.9
significant_frames = np.where(attn_weights >= threshold)[0]
if len(significant_frames) > 0:
start_frame = int(significant_frames[0])
end_frame = int(significant_frames[-1])
avg_weight = float(attn_weights[significant_frames].mean())
else:
start_frame = peak_frame
end_frame = peak_frame
avg_weight = float(peak_weight)
# Qualitative confidence bucket
if avg_weight > 0.5:
confidence = 'high'
elif avg_weight > 0.2:
confidence = 'medium'
else:
confidence = 'low'
word_ranges.append({
'word': word,
'start_frame': start_frame,
'end_frame': end_frame,
'peak_frame': peak_frame,
'avg_attention': avg_weight,
'confidence': confidence
})
return word_ranges
def _compute_frame_attention_strength(self):
"""Compute average attention per feature frame (normalized 0-1)."""
if self.attn_best.size == 0:
return np.zeros(self.video_frames, dtype=np.float32)
if self.attn_best.ndim == 1:
frame_strength = self.attn_best.copy()
else:
frame_strength = self.attn_best.mean(axis=0)
if frame_strength.shape[0] != self.video_frames:
frame_strength = np.resize(frame_strength, self.video_frames)
max_val = frame_strength.max()
if max_val > 0:
frame_strength = frame_strength / max_val
return frame_strength
def _map_strength_to_original_frames(self, mapping_list, original_frame_count):
"""Map latent attention strength to original video frame resolution."""
if not mapping_list or original_frame_count <= 0:
return None
orig_strength = np.zeros(original_frame_count, dtype=np.float32)
counts = np.zeros(original_frame_count, dtype=np.float32)
for feat_idx, mapping in enumerate(mapping_list):
if feat_idx >= len(self.frame_attention_strength):
break
start = int(mapping.get('frame_start', 0))
end = int(mapping.get('frame_end', start))
end = max(end, start + 1)
start = max(start, 0)
end = min(end, original_frame_count)
if start >= end:
continue
orig_strength[start:end] += self.frame_attention_strength[feat_idx]
counts[start:end] += 1
mask = counts > 0
if mask.any():
orig_strength[mask] = orig_strength[mask] / counts[mask]
max_val = orig_strength.max()
if max_val > 0:
orig_strength = orig_strength / max_val
return orig_strength
def generate_all_visualizations(self, output_dir):
"""
Generate every visualization artifact to the provided directory.
"""
output_dir = Path(output_dir)
output_dir.mkdir(parents=True, exist_ok=True)
print(f"\nGenerating visualization assets in: {output_dir}")
# 1. Attention heatmap
self.plot_attention_heatmap(output_dir / "attention_heatmap.png")
# 2. Frame alignment
self.plot_frame_alignment(output_dir / "frame_alignment.png")
# 3. JSON metadata
self.save_alignment_data(output_dir / "frame_alignment.json")
# 4. Text report
self.save_text_report(output_dir / "analysis_report.txt")
# 5. Raw numpy dump (for downstream tooling)
np.save(output_dir / "attention_weights.npy", self.attentions)
# 6. Gloss-to-Frames visualization (if video is available)
# Write debug info to file
debug_file = output_dir / "debug_video_path.txt"
with open(debug_file, 'w') as f:
f.write(f"video_path = {repr(self.video_path)}\n")
f.write(f"video_path type = {type(self.video_path)}\n")
f.write(f"video_path is None: {self.video_path is None}\n")
f.write(f"bool(video_path): {bool(self.video_path)}\n")
print(f"[DEBUG] video_path = {self.video_path}")
if self.video_path:
print(f"[DEBUG] Generating gloss-to-frames visualization with video: {self.video_path}")
try:
self.generate_gloss_to_frames_visualization(output_dir / "gloss_to_frames.png")
print(f"[DEBUG] Successfully generated gloss_to_frames.png")
except Exception as e:
print(f"[DEBUG] Failed to generate gloss_to_frames.png: {e}")
import traceback
traceback.print_exc()
else:
print("[DEBUG] Skipping gloss-to-frames visualization (no video path provided)")
print(f"✓ Wrote {len(list(output_dir.glob('*')))} file(s)")
def plot_attention_heatmap(self, output_path):
"""Render the attention heatmap (image + PDF copy)."""
try:
import matplotlib
matplotlib.use('Agg')
import matplotlib.pyplot as plt
except ImportError:
print(" Skipping heatmap: matplotlib is not available")
return
fig, ax = plt.subplots(figsize=(14, 8))
# Heatmap
im = ax.imshow(self.attn_best.T, cmap='hot', aspect='auto',
interpolation='nearest', origin='lower')
# Axis labels
ax.set_xlabel('Generated Word Index', fontsize=13)
ax.set_ylabel('Video Frame Index', fontsize=13)
ax.set_title('Cross-Attention Weights\n(Decoder → Video Frames)',
fontsize=15, pad=20, fontweight='bold')
# Word labels on the x-axis
if len(self.words) <= self.attn_best.shape[0]:
ax.set_xticks(range(len(self.words)))
ax.set_xticklabels(self.words, rotation=45, ha='right', fontsize=10)
# Color bar
cbar = plt.colorbar(im, ax=ax, label='Attention Weight', fraction=0.046, pad=0.04)
cbar.ax.tick_params(labelsize=10)
plt.tight_layout()
plt.savefig(output_path, dpi=150, bbox_inches='tight')
# also save PDF copy for high-res usage
pdf_path = Path(output_path).with_suffix('.pdf')
plt.savefig(str(pdf_path), format='pdf', bbox_inches='tight')
plt.close()
print(f" ✓ {output_path.name} (PDF copy saved)")
def plot_frame_alignment(self, output_path):
"""Render the frame-alignment charts (full + compact)."""
try:
import matplotlib
matplotlib.use('Agg')
import matplotlib.pyplot as plt
import matplotlib.patches as patches
from matplotlib.gridspec import GridSpec
except ImportError:
print(" Skipping alignment plot: matplotlib is not available")
return
output_path = Path(output_path)
# Try to load feature-to-frame mapping
feature_mapping = None
output_dir = output_path.parent
mapping_file = output_dir / "feature_frame_mapping.json"
if mapping_file.exists():
try:
with open(mapping_file, 'r') as f:
feature_mapping = json.load(f)
except Exception as e:
print(f" Warning: Failed to load feature mapping: {e}")
if self.word_frame_ranges:
max_feat_end = max(w['end_frame'] for w in self.word_frame_ranges)
else:
max_feat_end = self.video_frames - 1
latent_full_limit = self.video_frames + 2
latent_short_limit = max(min(latent_full_limit, max_feat_end + 2), 5)
original_frame_count = None
mapping_list = None
orig_full_limit = None
orig_short_limit = None
pixel_strength_curve = None
if feature_mapping:
original_frame_count = feature_mapping.get('original_frame_count', self.video_frames)
mapping_list = feature_mapping.get('mapping', [])
orig_full_limit = original_frame_count + 2
if mapping_list:
idx = min(max_feat_end, len(mapping_list) - 1)
orig_short_limit = mapping_list[idx]['frame_end'] + 2
pixel_strength_curve = self._map_strength_to_original_frames(mapping_list, original_frame_count)
def render_alignment(out_path, latent_xlim_end, orig_xlim_end=None):
if feature_mapping:
fig = plt.figure(figsize=(18, 9))
gs = GridSpec(3, 1, height_ratios=[4, 1, 1], hspace=0.32)
else:
fig = plt.figure(figsize=(18, 7.5))
gs = GridSpec(2, 1, height_ratios=[4, 1], hspace=0.32)
# === Top plot: word-to-frame alignment ===
ax1 = fig.add_subplot(gs[0])
colors = plt.cm.tab20(np.linspace(0, 1, max(len(self.words), 20)))
for i, word_info in enumerate(self.word_frame_ranges):
start = word_info['start_frame']
end = word_info['end_frame']
word = word_info['word']
confidence = word_info['confidence']
alpha = 0.9 if confidence == 'high' else 0.7 if confidence == 'medium' else 0.5
rect = patches.Rectangle(
(start, i), end - start + 1, 0.8,
linewidth=2, edgecolor='black',
facecolor=colors[i % 20], alpha=alpha
)
ax1.add_patch(rect)
ax1.text(start + (end - start) / 2, i + 0.4, word,
ha='center', va='center', fontsize=11,
fontweight='bold', color='white',
bbox=dict(boxstyle='round,pad=0.3', facecolor='black', alpha=0.5))
peak = word_info['peak_frame']
ax1.plot(peak, i + 0.4, 'r*', markersize=15, markeredgecolor='yellow',
markeredgewidth=1.5)
ax1.set_xlim(-2, latent_xlim_end)
ax1.set_ylim(-0.5, len(self.words))
ax1.set_xlabel('')
ax1.set_ylabel('')
ax1.set_title('Word-to-Frame Alignment\n(based on attention peaks, ★ = peak frame)',
fontsize=15, pad=15, fontweight='bold')
ax1.grid(True, alpha=0.3, axis='x', linestyle='--')
ax1.set_yticks(range(len(self.words)))
ax1.set_yticklabels([w['word'] for w in self.word_frame_ranges], fontsize=10)
# === Middle plot: latent timeline ===
ax2 = fig.add_subplot(gs[1])
ax2.barh(0, self.video_frames, height=0.6, color='lightgray',
edgecolor='black', linewidth=2)
for i, word_info in enumerate(self.word_frame_ranges):
start = word_info['start_frame']
end = word_info['end_frame']
confidence = word_info['confidence']
alpha = 0.9 if confidence == 'high' else 0.7 if confidence == 'medium' else 0.5
ax2.barh(0, end - start + 1, left=start, height=0.6,
color=colors[i % 20], alpha=alpha, edgecolor='black', linewidth=0.5)
ax2.set_xlim(-2, latent_xlim_end)
ax2.set_ylim(-0.4, 0.4)
ax2.set_xlabel('')
ax2.set_yticks([0])
ax2.set_yticklabels(['Latent Space'], fontsize=11, fontweight='bold')
ax2.tick_params(axis='y', length=0)
ax2.set_title('Latent Feature Timeline', fontsize=13, fontweight='bold')
ax2.grid(True, alpha=0.3, axis='x', linestyle='--')
if self.frame_attention_strength is not None and len(self.frame_attention_strength) >= self.video_frames:
latent_curve_x = np.arange(self.video_frames)
latent_curve_y = self.frame_attention_strength[:self.video_frames] * 0.6 - 0.3
ax2.plot(latent_curve_x, latent_curve_y, color='#E53935', linewidth=1.5, alpha=0.9)
timeline_axes = [ax2]
if feature_mapping:
ax3 = fig.add_subplot(gs[2])
ax3.barh(0, original_frame_count, height=0.6, color='lightgray',
edgecolor='black', linewidth=2)
for i, word_info in enumerate(self.word_frame_ranges):
feat_start = word_info['start_frame']
feat_end = word_info['end_frame']
confidence = word_info['confidence']
alpha = 0.9 if confidence == 'high' else 0.7 if confidence == 'medium' else 0.5
if mapping_list and feat_start < len(mapping_list) and feat_end < len(mapping_list):
orig_start = mapping_list[feat_start]['frame_start']
orig_end = mapping_list[feat_end]['frame_end']
ax3.barh(0, orig_end - orig_start, left=orig_start, height=0.6,
color=colors[i % 20], alpha=alpha, edgecolor='black', linewidth=0.5)
ax3_xlim = orig_xlim_end if orig_xlim_end is not None else original_frame_count + 2
ax3.set_xlim(-2, ax3_xlim)
ax3.set_ylim(-0.4, 0.4)
ax3.set_xlabel('')
ax3.set_yticks([0])
ax3.set_yticklabels(['Pixel Space'], fontsize=11, fontweight='bold')
ax3.tick_params(axis='y', length=0)
ax3.set_title(f'Original Video Timeline ({original_frame_count} frames, '
f'{feature_mapping["downsampling_ratio"]:.2f}x downsampling)',
fontsize=13, fontweight='bold')
ax3.grid(True, alpha=0.3, axis='x', linestyle='--')
if pixel_strength_curve is not None and len(pixel_strength_curve) >= original_frame_count:
pixel_curve_x = np.arange(original_frame_count)
pixel_curve_y = pixel_strength_curve[:original_frame_count] * 0.6 - 0.3
ax3.plot(pixel_curve_x, pixel_curve_y, color='#E53935', linewidth=1.5, alpha=0.9)
timeline_axes.append(ax3)
plt.tight_layout()
fig.canvas.draw()
ax1_pos = ax1.get_position()
renderer = fig.canvas.get_renderer()
ytick_extents = [tick.get_window_extent(renderer) for tick in ax1.get_yticklabels() if tick.get_text()]
fig_width_px = fig.get_size_inches()[0] * fig.dpi
if ytick_extents:
min_x_px = min(ext.x0 for ext in ytick_extents)
else:
min_x_px = ax1_pos.x0 * fig_width_px
line_x = max(0.01, (min_x_px / fig_width_px) - 0.01)
gw_center = 0.5 * (ax1_pos.y0 + ax1_pos.y1)
timeline_bounds = [ax.get_position() for ax in timeline_axes]
timeline_center = 0.5 * (min(pos.y0 for pos in timeline_bounds) + max(pos.y1 for pos in timeline_bounds))
fig.text(line_x, gw_center, 'Generated Word', rotation='vertical',
ha='right', va='center', fontsize=12, fontweight='bold')
fig.text(line_x, timeline_center, 'Timeline', rotation='vertical',
ha='right', va='center', fontsize=12, fontweight='bold')
png_path = Path(out_path)
plt.savefig(str(png_path), dpi=150, bbox_inches='tight')
pdf_path = png_path.with_suffix('.pdf')
plt.savefig(str(pdf_path), format='pdf', bbox_inches='tight')
plt.close()
print(f" ✓ {png_path.name} (PDF copy saved)")
render_alignment(output_path, latent_full_limit, orig_full_limit)
if latent_short_limit < latent_full_limit - 1e-6:
short_path = output_path.with_name("frame_alignment_short.png")
render_alignment(short_path, latent_short_limit, orig_short_limit if orig_short_limit else orig_full_limit)
def save_alignment_data(self, output_path):
"""Persist frame-alignment metadata to JSON."""
data = {
'translation': self.translation,
'words': self.words,
'total_video_frames': self.video_frames,
'frame_ranges': self.word_frame_ranges,
'statistics': {
'avg_confidence': np.mean([w['avg_attention'] for w in self.word_frame_ranges]),
'high_confidence_words': sum(1 for w in self.word_frame_ranges if w['confidence'] == 'high'),
'medium_confidence_words': sum(1 for w in self.word_frame_ranges if w['confidence'] == 'medium'),
'low_confidence_words': sum(1 for w in self.word_frame_ranges if w['confidence'] == 'low'),
}
}
with open(output_path, 'w', encoding='utf-8') as f:
json.dump(data, f, indent=2, ensure_ascii=False)
print(f" ✓ {output_path.name}")
def save_text_report(self, output_path):
"""Write a plain-text report (used for analysis_report.txt)."""
with open(output_path, 'w', encoding='utf-8') as f:
f.write("=" * 80 + "\n")
f.write(" Sign Language Recognition - Attention Analysis Report\n")
f.write("=" * 80 + "\n\n")
f.write(f"Generated at: {datetime.now().strftime('%Y-%m-%d %H:%M:%S')}\n\n")
f.write("Translation:\n")
f.write("-" * 80 + "\n")
f.write(f"{self.translation}\n\n")
f.write("Video info:\n")
f.write("-" * 80 + "\n")
f.write(f"Total feature frames: {self.video_frames}\n")
f.write(f"Word count: {len(self.words)}\n\n")
f.write("Attention tensor:\n")
f.write("-" * 80 + "\n")
f.write(f"Shape: {self.attentions.shape}\n")
f.write(f" - Decoder steps: {self.attentions.shape[0]}\n")
if len(self.attentions.shape) >= 3:
f.write(f" - Batch size: {self.attentions.shape[1]}\n")
if len(self.attentions.shape) >= 4:
f.write(f" - Beam size: {self.attentions.shape[2]}\n")
f.write(f" - Source length: {self.attentions.shape[3]}\n")
f.write("\n")
f.write("Word-to-frame details:\n")
f.write("=" * 80 + "\n")
f.write(f"{'No.':<5} {'Word':<20} {'Frames':<15} {'Peak':<8} {'Attn':<8} {'Conf':<10}\n")
f.write("-" * 80 + "\n")
for i, w in enumerate(self.word_frame_ranges):
frame_range = f"{w['start_frame']}-{w['end_frame']}"
f.write(f"{i+1:<5} {w['word']:<20} {frame_range:<15} "
f"{w['peak_frame']:<8} {w['avg_attention']:<8.3f} {w['confidence']:<10}\n")
f.write("\n" + "=" * 80 + "\n")
# Summary
stats = {
'avg_confidence': np.mean([w['avg_attention'] for w in self.word_frame_ranges]),
'high': sum(1 for w in self.word_frame_ranges if w['confidence'] == 'high'),
'medium': sum(1 for w in self.word_frame_ranges if w['confidence'] == 'medium'),
'low': sum(1 for w in self.word_frame_ranges if w['confidence'] == 'low'),
}
f.write("\nSummary:\n")
f.write("-" * 80 + "\n")
f.write(f"Average attention weight: {stats['avg_confidence']:.3f}\n")
f.write(f"High-confidence words: {stats['high']} ({stats['high']/len(self.words)*100:.1f}%)\n")
f.write(f"Medium-confidence words: {stats['medium']} ({stats['medium']/len(self.words)*100:.1f}%)\n")
f.write(f"Low-confidence words: {stats['low']} ({stats['low']/len(self.words)*100:.1f}%)\n")
f.write("\n" + "=" * 80 + "\n")
print(f" ✓ {output_path.name}")
def _map_feature_frame_to_original(self, feature_frame_idx):
"""
Map a SMKD feature frame index back to the original video frame index.
Args:
feature_frame_idx: Zero-based feature frame index
Returns:
int: Original frame index, or None if unavailable.
"""
if self.original_video_total_frames is None:
return None
# Approximate downsampling ratio between latent frames and original frames
downsample_ratio = self.original_video_total_frames / self.video_frames
# Map latent index to original frame index
original_frame_idx = int(feature_frame_idx * downsample_ratio)
return min(original_frame_idx, self.original_video_total_frames - 1)
def _extract_video_frames(self, frame_indices):
"""
Extract the requested original video frames (best-effort).
Args:
frame_indices: list[int] of original frame IDs to load
Returns:
dict mapping frame index to numpy array (BGR).
"""
if not self.video_path:
return {}
cv2 = self._get_cv2_module()
if cv2 is not None:
return self._extract_frames_with_cv2(cv2, frame_indices)
return self._extract_frames_with_ffmpeg(frame_indices)
def _get_cv2_module(self):
"""Lazy-load cv2 and cache the import outcome."""
if self._cv2_checked:
return self._cv2_module
try:
import cv2
self._cv2_module = cv2
except ImportError:
self._cv2_module = None
finally:
self._cv2_checked = True
if self._cv2_module is None:
print("Warning: opencv-python is missing; falling back to ffmpeg grabs")
return self._cv2_module
def _extract_frames_with_cv2(self, cv2, frame_indices):
"""Extract frames via OpenCV if available."""
frames = {}
cap = cv2.VideoCapture(self.video_path)
if not cap.isOpened():
print(f"Warning: Cannot open video file: {self.video_path}")
return {}
for frame_idx in sorted(frame_indices):
cap.set(cv2.CAP_PROP_POS_FRAMES, frame_idx)
ret, frame = cap.read()
if ret:
frame_rgb = cv2.cvtColor(frame, cv2.COLOR_BGR2RGB)
frames[frame_idx] = frame_rgb
cap.release()
return frames
def _extract_frames_with_ffmpeg(self, frame_indices):
"""Extract frames via ffmpeg + Pillow (OpenCV fallback)."""
if shutil.which("ffmpeg") is None:
print("Warning: ffmpeg not found; cannot extract frames")
return {}
try:
from PIL import Image
except ImportError:
print("Warning: Pillow not installed; cannot decode ffmpeg output")
return {}
frames = {}
for frame_idx in sorted(frame_indices):
cmd = [
"ffmpeg",
"-v", "error",
"-i", str(self.video_path),
"-vf", f"select=eq(n\\,{frame_idx})",
"-vframes", "1",
"-f", "image2pipe",
"-vcodec", "png",
"-"
]
try:
result = subprocess.run(
cmd, check=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE
)
if not result.stdout:
continue
image = Image.open(io.BytesIO(result.stdout)).convert("RGB")
frames[frame_idx] = np.array(image)
except subprocess.CalledProcessError as e:
print(f"Warning: ffmpeg failed to extract frame {frame_idx}: {e}")
except Exception as ex:
print(f"Warning: failed to decode frame {frame_idx}: {ex}")
if frames:
print(f" ✓ Extracted {len(frames)} frame(s) via ffmpeg")
else:
print(" ⓘ ffmpeg did not return any frames")
return frames
def generate_gloss_to_frames_visualization(self, output_path):
"""
Create the gloss-to-frames visualization:
Column 1: gloss text
Column 2: relative time + frame indices
Column 3: representative video thumbnails
"""
if not self.video_path:
print(" ⓘ Skipping gloss-to-frames visualization (no video path provided)")
return
try:
import matplotlib.pyplot as plt
import matplotlib.gridspec as gridspec
except ImportError:
print("Warning: matplotlib not installed")
return
# Load feature-to-frame mapping if available
feature_mapping = None
output_dir = Path(output_path).parent
mapping_file = output_dir / "feature_frame_mapping.json"
if mapping_file.exists():
try:
with open(mapping_file, 'r') as f:
mapping_data = json.load(f)
feature_mapping = mapping_data['mapping']
except Exception as e:
print(f" Warning: Failed to load feature mapping: {e}")
# Collect every original frame we need to grab
all_original_frames = set()
for word_info in self.word_frame_ranges:
# Feature frame range
start_feat = word_info['start_frame']
end_feat = word_info['end_frame']
# Map the feature range onto original video frames
if feature_mapping:
# Use the precomputed mapping data
for feat_idx in range(start_feat, end_feat + 1):
if feat_idx < len(feature_mapping):
# Pull every original frame for that feature segment
feat_info = feature_mapping[feat_idx]
for orig_idx in range(feat_info['frame_start'], feat_info['frame_end']):
all_original_frames.add(orig_idx)
else:
# Fallback: assume uniform downsampling
for feat_idx in range(start_feat, end_feat + 1):
orig_idx = self._map_feature_frame_to_original(feat_idx)
if orig_idx is not None:
all_original_frames.add(orig_idx)
# Extract the necessary frames
print(f" Extracting {len(all_original_frames)} original video frame(s)...")
video_frames_dict = self._extract_video_frames(list(all_original_frames))
if not video_frames_dict:
print(" ⓘ No video frames extracted, skipping visualization")
return
# Create figure (4 columns: Gloss | Feature Index | Peak Frame | Full Span)
n_words = len(self.words)
fig = plt.figure(figsize=(28, 3 * n_words))
gs = gridspec.GridSpec(n_words, 4, width_ratios=[1.5, 1.5, 2.5, 8], hspace=0.3, wspace=0.2)
for row_idx, (word, word_info) in enumerate(zip(self.words, self.word_frame_ranges)):
# Column 1: Gloss label
ax_gloss = fig.add_subplot(gs[row_idx, 0])
ax_gloss.text(0.5, 0.5, word, fontsize=24, weight='bold',
ha='center', va='center', wrap=True)
ax_gloss.axis('off')
# Column 2: Feature index info
ax_feature = fig.add_subplot(gs[row_idx, 1])
# Feature frame details
feat_start = word_info['start_frame']
feat_end = word_info['end_frame']
feat_peak = word_info['peak_frame']
feature_text = f"SMKD Feature Index\n"
feature_text += f"{'='*20}\n\n"
feature_text += f"Range:\n {feat_start}{feat_end}\n\n"
feature_text += f"Peak:\n {feat_peak}\n\n"
feature_text += f"Count:\n {feat_end - feat_start + 1} features"
ax_feature.text(0.5, 0.5, feature_text, fontsize=11, family='monospace',
va='center', ha='center',
bbox=dict(boxstyle='round,pad=0.8', facecolor='lightblue',
edgecolor='darkblue', linewidth=2, alpha=0.7))
ax_feature.axis('off')
# Column 3: Original frames for the peak feature
ax_peak_frames = fig.add_subplot(gs[row_idx, 2])
peak_frames_to_show = []
orig_peak_start, orig_peak_end = None, None
if feature_mapping and feat_peak is not None and feat_peak < len(feature_mapping):
# Use detailed mapping to determine the original frame span
peak_info = feature_mapping[feat_peak]
orig_peak_start = peak_info['frame_start']
orig_peak_end = peak_info['frame_end']
# Show each original frame linked to the peak feature range
for orig_idx in range(orig_peak_start, orig_peak_end):
if orig_idx in video_frames_dict:
peak_frames_to_show.append(video_frames_dict[orig_idx])
if peak_frames_to_show:
# Horizontally stitch frames
combined_peak = np.hstack(peak_frames_to_show)
ax_peak_frames.imshow(combined_peak)
# Add caption
ax_peak_frames.text(0.5, -0.05, f"Peak Feature {feat_peak}\nFrames {orig_peak_start}-{orig_peak_end-1} ({len(peak_frames_to_show)} frames)",
ha='center', va='top', transform=ax_peak_frames.transAxes,
fontsize=10, weight='bold', color='red',
bbox=dict(boxstyle='round,pad=0.3', facecolor='yellow', alpha=0.7))
else:
ax_peak_frames.text(0.5, 0.5, "No peak frames",
ha='center', va='center', transform=ax_peak_frames.transAxes)
ax_peak_frames.axis('off')
# Column 4: All frames covered by the gloss span
ax_all_frames = fig.add_subplot(gs[row_idx, 3])
all_frames_to_show = []
orig_start, orig_end = None, None
if feature_mapping:
# Determine range via mapping
if feat_start < len(feature_mapping) and feat_end < len(feature_mapping):
orig_start = feature_mapping[feat_start]['frame_start']
orig_end = feature_mapping[feat_end]['frame_end']
# Collect every frame in the span
for orig_idx in range(orig_start, orig_end):
if orig_idx in video_frames_dict:
all_frames_to_show.append(video_frames_dict[orig_idx])
if all_frames_to_show:
# Stitch all frames horizontally
combined_all = np.hstack(all_frames_to_show)
ax_all_frames.imshow(combined_all)
# Add caption showing total
frame_count = len(all_frames_to_show)
ax_all_frames.text(0.5, -0.05, f"All Frames ({frame_count} frames)\nRange: {orig_start}-{orig_end-1}",
ha='center', va='top', transform=ax_all_frames.transAxes,
fontsize=10, weight='bold', color='blue',
bbox=dict(boxstyle='round,pad=0.3', facecolor='lightblue', alpha=0.7))
else:
ax_all_frames.text(0.5, 0.5, "No frames available",
ha='center', va='center', transform=ax_all_frames.transAxes)
ax_all_frames.axis('off')
plt.suptitle(f"Three-Layer Alignment: Gloss ↔ Feature Index ↔ Original Frames\nTranslation: {self.translation}",
fontsize=16, weight='bold', y=0.995)
plt.savefig(output_path, dpi=150, bbox_inches='tight', facecolor='white')
plt.close()
print(f" ✓ {Path(output_path).name}")
def _read_video_metadata(self):
"""Attempt to read the original video's frame count and FPS."""
metadata = self._read_metadata_with_cv2()
if metadata:
return metadata
return self._read_metadata_with_ffprobe()
def _read_metadata_with_cv2(self):
cv2 = self._get_cv2_module()
if cv2 is None:
return None
cap = cv2.VideoCapture(self.video_path)
if not cap.isOpened():
return None
total_frames = int(cap.get(cv2.CAP_PROP_FRAME_COUNT))
fps = cap.get(cv2.CAP_PROP_FPS)
cap.release()
if total_frames <= 0:
return None
return {'frames': total_frames, 'fps': fps or self.original_video_fps}
def _read_metadata_with_ffprobe(self):
if shutil.which("ffprobe") is None:
return None
cmd = [
"ffprobe",
"-v", "error",
"-select_streams", "v:0",
"-show_entries", "stream=nb_frames,r_frame_rate,avg_frame_rate,duration",
"-of", "json",
str(self.video_path)
]
try:
result = subprocess.run(
cmd, check=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE, text=True
)
except subprocess.CalledProcessError:
return None
try:
info = json.loads(result.stdout)
except json.JSONDecodeError:
return None
streams = info.get("streams") or []
if not streams:
return None
stream = streams[0]
total_frames = stream.get("nb_frames")
fps = stream.get("avg_frame_rate") or stream.get("r_frame_rate")
duration = stream.get("duration")
fps_value = self._parse_ffprobe_fps(fps)
total_frames_value = None
if isinstance(total_frames, str) and total_frames.isdigit():
total_frames_value = int(total_frames)
if total_frames_value is None and duration and fps_value:
try:
total_frames_value = int(round(float(duration) * fps_value))
except ValueError:
total_frames_value = None
if total_frames_value is None:
return None
return {'frames': total_frames_value, 'fps': fps_value or self.original_video_fps}
@staticmethod
def _parse_ffprobe_fps(rate_str):
"""Parse an ffprobe frame-rate string such as '30000/1001'."""
if not rate_str or rate_str in ("0/0", "0"):
return None
if "/" in rate_str:
num, denom = rate_str.split("/", 1)
try:
num = float(num)
denom = float(denom)
if denom == 0:
return None
return num / denom
except ValueError:
return None
try:
return float(rate_str)
except ValueError:
return None
def analyze_from_numpy_file(attention_file, translation, video_frames, output_dir):
"""
Load attention weights from a .npy file and generate visualization assets.
Args:
attention_file: Path to the numpy file
translation: Clean translation string
video_frames: Number of SMKD feature frames
output_dir: Destination directory for outputs
"""
attentions = np.load(attention_file)
analyzer = AttentionAnalyzer(attentions, translation, video_frames)
analyzer.generate_all_visualizations(output_dir)
return analyzer