FangSen9000
Optimize display logic (PDF saving, good samples, good display)
eaf4dff
# coding: utf-8
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import time
import re
import numpy as np
import tensorflow as tf
from utils import queuer, util, metric
from pathlib import Path
def strip_leading_index_tokens(tokens):
if not tokens:
return tokens
return tokens[1:] if tokens[0].isdigit() else tokens
def _extract_numpy_from_nested(obj):
"""Extract first numpy array from nested list/tuple structures."""
if isinstance(obj, np.ndarray):
return obj
if isinstance(obj, (list, tuple)) and len(obj) > 0:
for item in obj:
extracted = _extract_numpy_from_nested(item)
if extracted is not None:
return extracted
return None
def _flatten_attention_batches(attention_batches):
"""
Convert raw attention fetches (list per eval batch, each possibly nested per GPU)
into a per-sample list aligned with translation order.
"""
flattened = []
if not attention_batches:
return flattened
for batch_idx, batch in enumerate(attention_batches):
arr = _extract_numpy_from_nested(batch)
if arr is None:
tf.logging.warning(f"[ATTN] Unable to extract numpy array from batch {batch_idx}")
continue
if arr.ndim == 4:
# [time, batch, beam, src_len]
time_steps, batch_size = arr.shape[0], arr.shape[1]
for b in range(batch_size):
flattened.append(arr[:, b, :, :])
elif arr.ndim in (2, 3):
flattened.append(arr)
else:
tf.logging.warning(f"[ATTN] Unexpected attention ndim={arr.ndim} for batch {batch_idx}")
return flattened
def _load_gloss_mapping():
"""Lazy-load ASLLRP gloss mapping (video_id -> text)."""
cache = getattr(_load_gloss_mapping, "_cache", None)
if cache is not None:
return cache
candidate_paths = [
Path(__file__).resolve().parent.parent / "../ASLLRP_utterances_mapping.txt",
Path(__file__).resolve().parent.parent / "ASLLRP_utterances_mapping.txt",
Path(__file__).resolve().parent.parent / "output/ASLLRP_utterances_mapping.txt",
Path(__file__).resolve().parent.parent / "../output/ASLLRP_utterances_mapping.txt",
]
mapping = {}
for cand in candidate_paths:
cand = cand.resolve()
if not cand.exists():
continue
try:
with cand.open('r', encoding='utf-8') as f:
for line in f:
line = line.strip()
if not line or ':' not in line:
continue
key, text = line.split(':', 1)
mapping[key.strip()] = text.strip()
tf.logging.info(f"[GT] Loaded gloss mapping from {cand}")
break
except Exception as exc:
tf.logging.warning(f"[GT] Failed to load mapping from {cand}: {exc}")
mapping = {}
_load_gloss_mapping._cache = mapping if mapping else None
return _load_gloss_mapping._cache
def _lookup_ground_truth_text(video_path):
"""Return GT gloss string for the given video path if available."""
if not video_path:
return None
mapping = _load_gloss_mapping()
if not mapping:
return None
stem = Path(video_path).stem
candidates = [stem]
if '_' in stem:
candidates.append(stem.split('_')[0])
for candidate in candidates:
if candidate in mapping:
return mapping[candidate]
# also try stripping leading zeros
stripped = candidate.lstrip('0')
if stripped and stripped in mapping:
return mapping[stripped]
return None
def decode_target_token(id_seq, vocab):
"""Convert sequence ids into tokens"""
valid_id_seq = []
for tok_id in id_seq:
if tok_id == vocab.eos() \
or tok_id == vocab.pad():
break
valid_id_seq.append(tok_id)
return vocab.to_tokens(valid_id_seq)
def decode_hypothesis(seqs, scores, params, mask=None):
"""Generate decoded sequence from seqs"""
if mask is None:
mask = [1.] * len(seqs)
hypoes = []
marks = []
for _seqs, _scores, _m in zip(seqs, scores, mask):
if _m < 1.: continue
for seq, score in zip(_seqs, _scores):
# Temporarily, Use top-1 decoding
best_seq = seq[0]
best_score = score[0]
hypo = decode_target_token(best_seq, params.tgt_vocab)
mark = best_score
hypoes.append(hypo)
marks.append(mark)
return hypoes, marks
def decoding(session, features, out_seqs, out_scores, dataset, params, out_attention=None):
"""Performing decoding with exising information"""
tf.logging.info(f"[DEBUG] decoding called with out_attention={out_attention is not None}")
if out_attention is not None:
tf.logging.info(f"[DEBUG] out_attention type: {type(out_attention)}")
translations = []
scores = []
indices = []
attentions = [] if out_attention is not None else None
eval_queue = queuer.EnQueuer(
dataset.batcher(params.eval_batch_size,
buffer_size=params.buffer_size,
shuffle=False,
train=False),
lambda x: x,
worker_processes_num=params.process_num,
input_queue_size=params.input_queue_size,
output_queue_size=params.output_queue_size,
)
def _predict_one_batch(_data_on_gpu):
feed_dicts = {}
_step_indices = []
for fidx, shard_data in enumerate(_data_on_gpu):
# define feed_dict
_feed_dict = {
features[fidx]["image"]: shard_data['img'],
features[fidx]["mask"]: shard_data['mask'],
features[fidx]["source"]: shard_data['src'],
}
feed_dicts.update(_feed_dict)
# collect data indices
_step_indices.extend(shard_data['index'])
# pick up valid outputs
data_size = len(_data_on_gpu)
valid_out_seqs = out_seqs[:data_size]
valid_out_scores = out_scores[:data_size]
# Prepare outputs to fetch
fetch_list = [valid_out_seqs, valid_out_scores]
if out_attention is not None:
valid_out_attention = out_attention[:data_size]
fetch_list.append(valid_out_attention)
# Run session
fetch_results = session.run(fetch_list, feed_dict=feed_dicts)
_decode_seqs, _decode_scores = fetch_results[0], fetch_results[1]
_decode_attention = fetch_results[2] if out_attention is not None else None
# DEBUG: Check what we got from session.run
if _decode_attention is not None and bidx == 0: # Only log first batch to avoid spam
tf.logging.info(f"[DEBUG] _decode_attention type: {type(_decode_attention)}")
if isinstance(_decode_attention, list):
tf.logging.info(f"[DEBUG] _decode_attention is list, len: {len(_decode_attention)}")
for i, item in enumerate(_decode_attention):
if item is not None:
tf.logging.info(f"[DEBUG] item[{i}] type: {type(item)}, shape: {item.shape if hasattr(item, 'shape') else 'no shape'}")
_step_translations, _step_scores = decode_hypothesis(
_decode_seqs, _decode_scores, params
)
return _step_translations, _step_scores, _step_indices, _decode_attention
very_begin_time = time.time()
data_on_gpu = []
for bidx, data in enumerate(eval_queue):
if bidx == 0:
# remove the data reading time
very_begin_time = time.time()
data_on_gpu.append(data)
# use multiple gpus, and data samples is not enough
if len(params.gpus) > 0 and len(data_on_gpu) < len(params.gpus):
continue
start_time = time.time()
step_outputs = _predict_one_batch(data_on_gpu)
data_on_gpu = []
translations.extend(step_outputs[0])
scores.extend(step_outputs[1])
indices.extend(step_outputs[2])
if attentions is not None and step_outputs[3] is not None:
attentions.append(step_outputs[3])
tf.logging.info(
"Decoding Batch {} using {:.3f} s, translating {} "
"sentences using {:.3f} s in total".format(
bidx, time.time() - start_time,
len(translations), time.time() - very_begin_time
)
)
if len(data_on_gpu) > 0:
start_time = time.time()
step_outputs = _predict_one_batch(data_on_gpu)
translations.extend(step_outputs[0])
scores.extend(step_outputs[1])
indices.extend(step_outputs[2])
if attentions is not None and step_outputs[3] is not None:
attentions.append(step_outputs[3])
tf.logging.info(
"Decoding Batch {} using {:.3f} s, translating {} "
"sentences using {:.3f} s in total".format(
'final', time.time() - start_time,
len(translations), time.time() - very_begin_time
)
)
return translations, scores, indices, attentions
def eval_metric(trans, target_file, indices=None, remove_bpe=False):
"""BLEU Evaluate """
target_valid_files = util.fetch_valid_ref_files(target_file)
if target_valid_files is None:
return 0.0
if indices is not None:
trans = [data[1] for data in sorted(zip(indices, trans), key=lambda x: x[0])]
references = []
for ref_file in target_valid_files:
cur_refs = tf.gfile.Open(ref_file).readlines()
processed = []
for line in cur_refs:
tokens = line.strip().split()
tokens = tokens[1:] if tokens and tokens[0].isdigit() else tokens
if remove_bpe:
cleaned = (' '.join(tokens)).replace("@@ ", "").split()
processed.append(cleaned)
else:
processed.append(tokens)
references.append(processed)
references = list(zip(*references))
new_trans = []
for line in trans:
tokens = line
tokens = tokens[1:] if tokens and tokens[0].isdigit() else tokens
if remove_bpe:
tokens = (' '.join(tokens)).replace("@@ ", "").split()
new_trans.append(tokens)
return metric.bleu(new_trans, references)
def dump_tanslation(tranes, output, indices=None, attentions=None, video_path=None):
"""save translation"""
if indices is not None:
tranes = [data[1] for data in
sorted(zip(indices, tranes), key=lambda x: x[0])]
with tf.gfile.Open(output, 'w') as writer:
for hypo in tranes:
if isinstance(hypo, list):
writer.write(' '.join(hypo) + "\n")
else:
writer.write(str(hypo) + "\n")
tf.logging.info("Saving translations into {}".format(output))
# DEBUG: Check attention status
tf.logging.info(f"[DEBUG] attentions is None: {attentions is None}")
if attentions is not None:
tf.logging.info(f"[DEBUG] attentions type: {type(attentions)}, len: {len(attentions)}")
# Save detailed attention analysis if available
if attentions is not None and len(attentions) > 0:
tf.logging.info("[DEBUG] Calling dump_detailed_attention_output")
try:
dump_detailed_attention_output(tranes, output, indices, attentions, video_path)
except Exception as e:
tf.logging.warning(f"Failed to save detailed attention output: {e}")
import traceback
tf.logging.warning(traceback.format_exc())
else:
tf.logging.info("[DEBUG] Skipping attention analysis (attentions is None or empty)")
def dump_translation_with_reference(tranes, output, ref_file, indices=None, remove_bpe=False,
attentions=None, dataset=None):
"""
Save translation with reference for easy comparison
Args:
tranes: Translation results
output: Output file path
ref_file: Reference file path
indices: Sample indices
remove_bpe: Whether to remove BPE
attentions: Attention weights (list of numpy arrays) for frame alignment
dataset: Dataset object for getting video frame counts
Returns:
dict: Coverage metrics {'complete_coverage': float, 'avg_iou': float}
"""
import tensorflow as tf
import numpy as np
import json
import os
import sys
from pathlib import Path
per_sample_attn = None
if attentions is not None and len(attentions) > 0:
per_sample_attn = _flatten_attention_batches(attentions)
if len(per_sample_attn) == 0:
per_sample_attn = None
# Align attention list with translations
if per_sample_attn is None:
per_sample_attn = [None] * len(tranes)
elif len(per_sample_attn) < len(tranes):
per_sample_attn.extend([None] * (len(tranes) - len(per_sample_attn)))
# Sort translations (and attentions) by sample index if provided
if indices is not None:
sorted_data = sorted(
zip(indices, tranes, per_sample_attn),
key=lambda x: x[0]
)
tranes = [data[1] for data in sorted_data]
per_sample_attn = [data[2] for data in sorted_data]
sorted_indices = [data[0] for data in sorted_data]
else:
sorted_indices = list(range(len(tranes)))
# Load references
references = []
if tf.gfile.Exists(ref_file):
refs = tf.gfile.Open(ref_file).readlines()
for line in refs:
tokens = strip_leading_index_tokens(line.strip().split())
ref = ' '.join(tokens)
if remove_bpe:
ref = ref.replace("@@ ", "")
references.append(ref)
# Initialize coverage metrics
all_complete_coverage = []
all_iou = []
# Try to load GT annotations and test info for ASLLRP
gt_annotations = None
test_info = None
try:
# Determine repo root from absolute output path
output_abs = os.path.abspath(output)
output_parts = output_abs.split(os.sep)
if 'SignX' in output_parts:
signx_idx = output_parts.index('SignX')
repo_root = os.sep.join(output_parts[:signx_idx+1])
else:
# Fallback: search parents for directory named SignX
repo_root = None
current_path = Path(output_abs).parent
for parent in [current_path] + list(current_path.parents):
if parent.name == 'SignX':
repo_root = str(parent)
break
if repo_root is None:
raise RuntimeError("Cannot locate SignX repo root from output path.")
# Load GT annotations
gt_path = os.path.join(os.path.dirname(repo_root), 'ASLLRP_utterances_with_frames.json')
if os.path.exists(gt_path):
with open(gt_path, 'r') as f:
gt_annotations = json.load(f)
tf.logging.info(f"Loaded GT annotations from {gt_path}")
# Load test info
test_info_path = os.path.join(os.path.dirname(repo_root), 'CorrNet_Plus_modified/dataset/ASLLRP/test_info.npy')
if os.path.exists(test_info_path):
test_info = np.load(test_info_path, allow_pickle=True)
tf.logging.info(f"Loaded test info from {test_info_path}")
except Exception as e:
tf.logging.warning(f"Could not load GT annotations or test info: {e}")
# Import AttentionAnalyzer if we have attentions
AttentionAnalyzer = None
if attentions is not None and len(attentions) > 0 and gt_annotations is not None:
try:
# Add eval directory to path
eval_dir = os.path.join(os.path.dirname(__file__), '..', 'eval')
if os.path.exists(eval_dir) and eval_dir not in sys.path:
sys.path.insert(0, eval_dir)
from attention_analysis import AttentionAnalyzer as AA
AttentionAnalyzer = AA
except ImportError as e:
tf.logging.warning(f"Could not import AttentionAnalyzer: {e}")
# Save comparison file
comparison_file = output.replace('.trans.txt', '.comparison.txt')
with tf.gfile.Open(comparison_file, 'w') as writer:
writer.write("=" * 100 + "\n")
writer.write("TRANSLATION COMPARISON (Hypothesis vs Reference)\n")
writer.write("=" * 100 + "\n\n")
for i, hypo in enumerate(tranes):
# Format hypothesis
if isinstance(hypo, list):
hypo_str = ' '.join(hypo)
hypo_tokens = hypo
else:
hypo_str = str(hypo)
hypo_tokens = hypo_str.split()
if remove_bpe:
hypo_str = hypo_str.replace("@@ ", "")
hypo_tokens = [t.replace("@@", "").strip() for t in hypo_tokens]
hypo_tokens = [t for t in hypo_tokens if t] # Remove empty
# Get reference
ref_str = references[i] if i < len(references) else "[NO REFERENCE]"
ref_tokens = strip_leading_index_tokens(ref_str.split())
ref_str = ' '.join(ref_tokens)
# Calculate frame alignment if we have attention weights
pred_frame_info = None
gt_frame_info = None
coverage_metrics = None
sample_attn = per_sample_attn[i] if per_sample_attn and i < len(per_sample_attn) else None
if (AttentionAnalyzer is not None and sample_attn is not None and
test_info is not None and gt_annotations is not None):
try:
sample_idx = sorted_indices[i]
# Get video ID from test_info
video_id = test_info[sample_idx]['video_id']
# Get GT annotations for this video
if video_id in gt_annotations:
gt_glosses = gt_annotations[video_id]['glosses']
attn = sample_attn
# Get feature frame count from attention shape
if isinstance(attn, np.ndarray) and len(attn.shape) >= 2:
# Reduce dimensions to [time, src_len]
if len(attn.shape) == 4:
attn = attn[:, 0, 0, :]
elif len(attn.shape) == 3:
attn = attn[:, 0, :]
feature_frames = attn.shape[1]
# Use AttentionAnalyzer to compute frame ranges
analyzer = AttentionAnalyzer(
attentions=attn,
translation=' '.join(hypo_tokens),
video_frames=feature_frames
)
# Get predicted frame ranges
pred_frame_ranges = analyzer.word_frame_ranges
# Calculate coverage metrics
complete_count = 0
iou_scores = []
# Build GT gloss-to-frame mapping
gt_frame_map = {gt['gloss']: (gt['start_24fps'], gt['end_24fps'])
for gt in gt_glosses}
for pred_info in pred_frame_ranges:
pred_gloss = pred_info['word']
pred_start = pred_info['start_frame']
pred_end = pred_info['end_frame']
if pred_gloss in gt_frame_map:
gt_start, gt_end = gt_frame_map[pred_gloss]
# Complete coverage: prediction contains GT
if pred_start <= gt_start and pred_end >= gt_end:
complete_count += 1
# IoU calculation
intersection_start = max(pred_start, gt_start)
intersection_end = min(pred_end, gt_end)
intersection = max(0, intersection_end - intersection_start + 1)
union_start = min(pred_start, gt_start)
union_end = max(pred_end, gt_end)
union = union_end - union_start + 1
iou = intersection / union if union > 0 else 0.0
iou_scores.append(iou)
# Calculate rates
num_matched = len(iou_scores)
if num_matched > 0:
complete_coverage_rate = complete_count / num_matched
avg_iou = np.mean(iou_scores)
all_complete_coverage.append(complete_coverage_rate)
all_iou.append(avg_iou)
coverage_metrics = {
'complete_coverage_rate': complete_coverage_rate,
'avg_iou': avg_iou,
'num_matched': num_matched
}
pred_frame_info = pred_frame_ranges
gt_frame_info = gt_frame_map
except Exception as e:
tf.logging.warning(f"Error calculating frame alignment for sample {i}: {e}")
# Write comparison with frame information
writer.write(f"[Sample {i}]\n")
# Write HYP with frames if available
if pred_frame_info:
hyp_with_frames = []
for pred_info in pred_frame_info:
gloss = pred_info['word']
start = pred_info['start_frame']
end = pred_info['end_frame']
hyp_with_frames.append(f"{gloss}[{start}-{end}]")
writer.write(f"HYP: {' '.join(hyp_with_frames)}\n")
else:
writer.write(f"HYP: {hypo_str}\n")
# Write REF with frames if available
if gt_frame_info:
ref_with_frames = []
for token in ref_tokens:
if token in gt_frame_info:
start, end = gt_frame_info[token]
ref_with_frames.append(f"{token}[{start}-{end}]")
else:
ref_with_frames.append(token)
writer.write(f"REF: {' '.join(ref_with_frames)}\n")
else:
writer.write(f"REF: {ref_str}\n")
# Write coverage metrics if available
if coverage_metrics:
writer.write(f"Complete Coverage: {coverage_metrics['complete_coverage_rate']:.2%} "
f"({coverage_metrics['num_matched']} matched glosses)\n")
writer.write(f"IoU Coverage: {coverage_metrics['avg_iou']:.4f}\n")
writer.write("-" * 100 + "\n\n")
tf.logging.info("Saving comparison into {}".format(comparison_file))
# Return average coverage metrics
result_metrics = {}
if all_complete_coverage:
result_metrics['complete_coverage'] = float(np.mean(all_complete_coverage))
result_metrics['avg_iou'] = float(np.mean(all_iou))
tf.logging.info(f"Coverage Metrics: Complete={result_metrics['complete_coverage']:.2%}, "
f"IoU={result_metrics['avg_iou']:.4f}")
return result_metrics
def dump_detailed_attention_output(tranes, output, indices, attentions, video_path=None):
"""
保存详细的attention分析结果
Args:
tranes: 翻译结果列表
output: 输出文件路径
indices: 样本索引
attentions: attention权重数据(list of numpy arrays)
video_path: 视频文件路径(可选,用于提取视频帧)
"""
import os
import sys
import re
from datetime import datetime
from pathlib import Path
# 获取输出目录和文件名
output_path = Path(output)
base_dir = output_path.parent
base_name = output_path.stem # 不带扩展名
# 创建带时间戳的详细分析目录
timestamp = datetime.now().strftime("%Y%m%d_%H%M%S")
detail_dir = base_dir / f"detailed_{base_name}_{timestamp}"
detail_dir.mkdir(parents=True, exist_ok=True)
tf.logging.info(f"Saving detailed attention analysis to: {detail_dir}")
# 重排序翻译结果
if indices is not None:
sorted_items = sorted(zip(indices, tranes), key=lambda x: x[0])
tranes = [item[1] for item in sorted_items]
# 合并所有batch的attention数据
try:
import numpy as np
flattened_attn = _flatten_attention_batches(attentions)
if len(flattened_attn) == 0:
tf.logging.warning("No valid attention data found")
return
tf.logging.info(f"[DEBUG] Found {len(flattened_attn)} valid attention samples")
# If只推理单个视频,使用文件名作为样本目录,便于定位
video_based_name = None
if video_path:
try:
candidate = Path(video_path).stem
if candidate:
sanitized = re.sub(r'[^0-9A-Za-z._-]+', '_', candidate).strip('_')
if sanitized:
video_based_name = sanitized
except Exception as exc:
tf.logging.warning(f"Failed to derive video-based sample name: {exc}")
use_video_name = video_based_name is not None and len(flattened_attn) == 1
gt_text = _lookup_ground_truth_text(video_path) if video_path else None
for sample_idx, sample_attn in enumerate(flattened_attn):
if sample_idx >= len(tranes):
break
# Ensure shape is [time, src_len]
if sample_attn.ndim == 4:
sample_attn = sample_attn[:, 0, 0, :]
elif sample_attn.ndim == 3:
sample_attn = sample_attn[:, 0, :]
# 获取翻译结果
trans = tranes[sample_idx]
if isinstance(trans, list):
trans = ' '.join(trans)
trans_clean = trans.replace('@@ ', '')
# 创建样本专属目录(单视频推理时用文件名,更易记忆)
if use_video_name and sample_idx == 0:
sample_name = video_based_name
else:
sample_name = f"sample_{sample_idx:03d}"
sample_dir = detail_dir / sample_name
sample_dir.mkdir(exist_ok=True)
# 保存numpy数据
np.save(sample_dir / "attention_weights.npy", sample_attn)
# 保存翻译结果
with open(sample_dir / "translation.txt", 'w', encoding='utf-8') as f:
f.write(f"With BPE: {trans}\n")
f.write(f"Clean: {trans_clean}\n")
if gt_text:
f.write(f"Ground Truth: {gt_text}\n")
else:
f.write("Ground Truth: [NOT FOUND]\n")
# Calculate and save feature-to-frame mapping
if video_path and os.path.exists(video_path):
try:
import cv2
import json
cap = cv2.VideoCapture(video_path)
original_frame_count = int(cap.get(cv2.CAP_PROP_FRAME_COUNT))
cap.release()
feature_count = sample_attn.shape[1]
frame_mapping = []
for feat_idx in range(feature_count):
start_frame = int(feat_idx * original_frame_count / feature_count)
end_frame = int((feat_idx + 1) * original_frame_count / feature_count)
frame_mapping.append({
"feature_index": feat_idx,
"frame_start": start_frame,
"frame_end": end_frame,
"frame_count": end_frame - start_frame
})
mapping_data = {
"original_frame_count": original_frame_count,
"feature_count": feature_count,
"downsampling_ratio": original_frame_count / feature_count,
"mapping": frame_mapping
}
with open(sample_dir / "feature_frame_mapping.json", 'w') as f:
json.dump(mapping_data, f, indent=2)
tf.logging.info(f" ✓ Feature-to-frame mapping saved ({original_frame_count} frames → {feature_count} features)")
except Exception as e:
tf.logging.warning(f"Failed to generate feature-to-frame mapping: {e}")
# 使用attention_analysis模块生成可视化
try:
script_dir = Path(__file__).parent.parent
eval_dir = script_dir / "eval"
if str(eval_dir) not in sys.path:
sys.path.insert(0, str(eval_dir))
from attention_analysis import AttentionAnalyzer
analyzer = AttentionAnalyzer(
attentions=sample_attn,
translation=trans_clean,
video_frames=sample_attn.shape[1],
video_path=video_path
)
analyzer.generate_all_visualizations(sample_dir)
tf.logging.info(f" ✓ Sample {sample_idx}: {sample_dir.name}")
except Exception as e:
tf.logging.warning(f"Failed to generate visualizations for sample {sample_idx}: {e}")
tf.logging.info(f"Detailed attention analysis complete: {detail_dir}")
tf.logging.info(f" - Analyzed {min(len(flattened_attn), len(tranes))} samples")
tf.logging.info(f" - Output directory: {detail_dir}")
except Exception as e:
import traceback
tf.logging.error(f"Error in dump_detailed_attention_output: {e}")
tf.logging.error(traceback.format_exc())