|
|
|
|
|
|
|
|
from __future__ import absolute_import |
|
|
from __future__ import division |
|
|
from __future__ import print_function |
|
|
|
|
|
import time |
|
|
import re |
|
|
import numpy as np |
|
|
import tensorflow as tf |
|
|
|
|
|
from utils import queuer, util, metric |
|
|
from pathlib import Path |
|
|
|
|
|
|
|
|
def strip_leading_index_tokens(tokens): |
|
|
if not tokens: |
|
|
return tokens |
|
|
return tokens[1:] if tokens[0].isdigit() else tokens |
|
|
|
|
|
|
|
|
def _extract_numpy_from_nested(obj): |
|
|
"""Extract first numpy array from nested list/tuple structures.""" |
|
|
if isinstance(obj, np.ndarray): |
|
|
return obj |
|
|
if isinstance(obj, (list, tuple)) and len(obj) > 0: |
|
|
for item in obj: |
|
|
extracted = _extract_numpy_from_nested(item) |
|
|
if extracted is not None: |
|
|
return extracted |
|
|
return None |
|
|
|
|
|
|
|
|
def _flatten_attention_batches(attention_batches): |
|
|
""" |
|
|
Convert raw attention fetches (list per eval batch, each possibly nested per GPU) |
|
|
into a per-sample list aligned with translation order. |
|
|
""" |
|
|
flattened = [] |
|
|
if not attention_batches: |
|
|
return flattened |
|
|
|
|
|
for batch_idx, batch in enumerate(attention_batches): |
|
|
arr = _extract_numpy_from_nested(batch) |
|
|
if arr is None: |
|
|
tf.logging.warning(f"[ATTN] Unable to extract numpy array from batch {batch_idx}") |
|
|
continue |
|
|
|
|
|
if arr.ndim == 4: |
|
|
|
|
|
time_steps, batch_size = arr.shape[0], arr.shape[1] |
|
|
for b in range(batch_size): |
|
|
flattened.append(arr[:, b, :, :]) |
|
|
elif arr.ndim in (2, 3): |
|
|
flattened.append(arr) |
|
|
else: |
|
|
tf.logging.warning(f"[ATTN] Unexpected attention ndim={arr.ndim} for batch {batch_idx}") |
|
|
|
|
|
return flattened |
|
|
|
|
|
|
|
|
def _load_gloss_mapping(): |
|
|
"""Lazy-load ASLLRP gloss mapping (video_id -> text).""" |
|
|
cache = getattr(_load_gloss_mapping, "_cache", None) |
|
|
if cache is not None: |
|
|
return cache |
|
|
|
|
|
candidate_paths = [ |
|
|
Path(__file__).resolve().parent.parent / "../ASLLRP_utterances_mapping.txt", |
|
|
Path(__file__).resolve().parent.parent / "ASLLRP_utterances_mapping.txt", |
|
|
Path(__file__).resolve().parent.parent / "output/ASLLRP_utterances_mapping.txt", |
|
|
Path(__file__).resolve().parent.parent / "../output/ASLLRP_utterances_mapping.txt", |
|
|
] |
|
|
|
|
|
mapping = {} |
|
|
for cand in candidate_paths: |
|
|
cand = cand.resolve() |
|
|
if not cand.exists(): |
|
|
continue |
|
|
try: |
|
|
with cand.open('r', encoding='utf-8') as f: |
|
|
for line in f: |
|
|
line = line.strip() |
|
|
if not line or ':' not in line: |
|
|
continue |
|
|
key, text = line.split(':', 1) |
|
|
mapping[key.strip()] = text.strip() |
|
|
tf.logging.info(f"[GT] Loaded gloss mapping from {cand}") |
|
|
break |
|
|
except Exception as exc: |
|
|
tf.logging.warning(f"[GT] Failed to load mapping from {cand}: {exc}") |
|
|
mapping = {} |
|
|
|
|
|
_load_gloss_mapping._cache = mapping if mapping else None |
|
|
return _load_gloss_mapping._cache |
|
|
|
|
|
|
|
|
def _lookup_ground_truth_text(video_path): |
|
|
"""Return GT gloss string for the given video path if available.""" |
|
|
if not video_path: |
|
|
return None |
|
|
mapping = _load_gloss_mapping() |
|
|
if not mapping: |
|
|
return None |
|
|
|
|
|
stem = Path(video_path).stem |
|
|
candidates = [stem] |
|
|
if '_' in stem: |
|
|
candidates.append(stem.split('_')[0]) |
|
|
|
|
|
for candidate in candidates: |
|
|
if candidate in mapping: |
|
|
return mapping[candidate] |
|
|
|
|
|
stripped = candidate.lstrip('0') |
|
|
if stripped and stripped in mapping: |
|
|
return mapping[stripped] |
|
|
return None |
|
|
|
|
|
|
|
|
def decode_target_token(id_seq, vocab): |
|
|
"""Convert sequence ids into tokens""" |
|
|
valid_id_seq = [] |
|
|
for tok_id in id_seq: |
|
|
if tok_id == vocab.eos() \ |
|
|
or tok_id == vocab.pad(): |
|
|
break |
|
|
valid_id_seq.append(tok_id) |
|
|
return vocab.to_tokens(valid_id_seq) |
|
|
|
|
|
|
|
|
def decode_hypothesis(seqs, scores, params, mask=None): |
|
|
"""Generate decoded sequence from seqs""" |
|
|
if mask is None: |
|
|
mask = [1.] * len(seqs) |
|
|
|
|
|
hypoes = [] |
|
|
marks = [] |
|
|
for _seqs, _scores, _m in zip(seqs, scores, mask): |
|
|
if _m < 1.: continue |
|
|
|
|
|
for seq, score in zip(_seqs, _scores): |
|
|
|
|
|
best_seq = seq[0] |
|
|
best_score = score[0] |
|
|
|
|
|
hypo = decode_target_token(best_seq, params.tgt_vocab) |
|
|
mark = best_score |
|
|
|
|
|
hypoes.append(hypo) |
|
|
marks.append(mark) |
|
|
|
|
|
return hypoes, marks |
|
|
|
|
|
|
|
|
def decoding(session, features, out_seqs, out_scores, dataset, params, out_attention=None): |
|
|
"""Performing decoding with exising information""" |
|
|
tf.logging.info(f"[DEBUG] decoding called with out_attention={out_attention is not None}") |
|
|
if out_attention is not None: |
|
|
tf.logging.info(f"[DEBUG] out_attention type: {type(out_attention)}") |
|
|
|
|
|
translations = [] |
|
|
scores = [] |
|
|
indices = [] |
|
|
attentions = [] if out_attention is not None else None |
|
|
|
|
|
eval_queue = queuer.EnQueuer( |
|
|
dataset.batcher(params.eval_batch_size, |
|
|
buffer_size=params.buffer_size, |
|
|
shuffle=False, |
|
|
train=False), |
|
|
lambda x: x, |
|
|
worker_processes_num=params.process_num, |
|
|
input_queue_size=params.input_queue_size, |
|
|
output_queue_size=params.output_queue_size, |
|
|
) |
|
|
|
|
|
def _predict_one_batch(_data_on_gpu): |
|
|
feed_dicts = {} |
|
|
|
|
|
_step_indices = [] |
|
|
for fidx, shard_data in enumerate(_data_on_gpu): |
|
|
|
|
|
_feed_dict = { |
|
|
features[fidx]["image"]: shard_data['img'], |
|
|
features[fidx]["mask"]: shard_data['mask'], |
|
|
features[fidx]["source"]: shard_data['src'], |
|
|
} |
|
|
feed_dicts.update(_feed_dict) |
|
|
|
|
|
|
|
|
_step_indices.extend(shard_data['index']) |
|
|
|
|
|
|
|
|
data_size = len(_data_on_gpu) |
|
|
valid_out_seqs = out_seqs[:data_size] |
|
|
valid_out_scores = out_scores[:data_size] |
|
|
|
|
|
|
|
|
fetch_list = [valid_out_seqs, valid_out_scores] |
|
|
if out_attention is not None: |
|
|
valid_out_attention = out_attention[:data_size] |
|
|
fetch_list.append(valid_out_attention) |
|
|
|
|
|
|
|
|
fetch_results = session.run(fetch_list, feed_dict=feed_dicts) |
|
|
_decode_seqs, _decode_scores = fetch_results[0], fetch_results[1] |
|
|
_decode_attention = fetch_results[2] if out_attention is not None else None |
|
|
|
|
|
|
|
|
if _decode_attention is not None and bidx == 0: |
|
|
tf.logging.info(f"[DEBUG] _decode_attention type: {type(_decode_attention)}") |
|
|
if isinstance(_decode_attention, list): |
|
|
tf.logging.info(f"[DEBUG] _decode_attention is list, len: {len(_decode_attention)}") |
|
|
for i, item in enumerate(_decode_attention): |
|
|
if item is not None: |
|
|
tf.logging.info(f"[DEBUG] item[{i}] type: {type(item)}, shape: {item.shape if hasattr(item, 'shape') else 'no shape'}") |
|
|
|
|
|
_step_translations, _step_scores = decode_hypothesis( |
|
|
_decode_seqs, _decode_scores, params |
|
|
) |
|
|
|
|
|
return _step_translations, _step_scores, _step_indices, _decode_attention |
|
|
|
|
|
very_begin_time = time.time() |
|
|
data_on_gpu = [] |
|
|
for bidx, data in enumerate(eval_queue): |
|
|
if bidx == 0: |
|
|
|
|
|
very_begin_time = time.time() |
|
|
|
|
|
data_on_gpu.append(data) |
|
|
|
|
|
if len(params.gpus) > 0 and len(data_on_gpu) < len(params.gpus): |
|
|
continue |
|
|
|
|
|
start_time = time.time() |
|
|
step_outputs = _predict_one_batch(data_on_gpu) |
|
|
data_on_gpu = [] |
|
|
|
|
|
translations.extend(step_outputs[0]) |
|
|
scores.extend(step_outputs[1]) |
|
|
indices.extend(step_outputs[2]) |
|
|
if attentions is not None and step_outputs[3] is not None: |
|
|
attentions.append(step_outputs[3]) |
|
|
|
|
|
tf.logging.info( |
|
|
"Decoding Batch {} using {:.3f} s, translating {} " |
|
|
"sentences using {:.3f} s in total".format( |
|
|
bidx, time.time() - start_time, |
|
|
len(translations), time.time() - very_begin_time |
|
|
) |
|
|
) |
|
|
|
|
|
if len(data_on_gpu) > 0: |
|
|
|
|
|
start_time = time.time() |
|
|
step_outputs = _predict_one_batch(data_on_gpu) |
|
|
|
|
|
translations.extend(step_outputs[0]) |
|
|
scores.extend(step_outputs[1]) |
|
|
indices.extend(step_outputs[2]) |
|
|
if attentions is not None and step_outputs[3] is not None: |
|
|
attentions.append(step_outputs[3]) |
|
|
|
|
|
tf.logging.info( |
|
|
"Decoding Batch {} using {:.3f} s, translating {} " |
|
|
"sentences using {:.3f} s in total".format( |
|
|
'final', time.time() - start_time, |
|
|
len(translations), time.time() - very_begin_time |
|
|
) |
|
|
) |
|
|
|
|
|
return translations, scores, indices, attentions |
|
|
|
|
|
|
|
|
def eval_metric(trans, target_file, indices=None, remove_bpe=False): |
|
|
"""BLEU Evaluate """ |
|
|
target_valid_files = util.fetch_valid_ref_files(target_file) |
|
|
if target_valid_files is None: |
|
|
return 0.0 |
|
|
|
|
|
if indices is not None: |
|
|
trans = [data[1] for data in sorted(zip(indices, trans), key=lambda x: x[0])] |
|
|
|
|
|
references = [] |
|
|
for ref_file in target_valid_files: |
|
|
cur_refs = tf.gfile.Open(ref_file).readlines() |
|
|
processed = [] |
|
|
for line in cur_refs: |
|
|
tokens = line.strip().split() |
|
|
tokens = tokens[1:] if tokens and tokens[0].isdigit() else tokens |
|
|
if remove_bpe: |
|
|
cleaned = (' '.join(tokens)).replace("@@ ", "").split() |
|
|
processed.append(cleaned) |
|
|
else: |
|
|
processed.append(tokens) |
|
|
references.append(processed) |
|
|
|
|
|
references = list(zip(*references)) |
|
|
|
|
|
new_trans = [] |
|
|
for line in trans: |
|
|
tokens = line |
|
|
tokens = tokens[1:] if tokens and tokens[0].isdigit() else tokens |
|
|
if remove_bpe: |
|
|
tokens = (' '.join(tokens)).replace("@@ ", "").split() |
|
|
new_trans.append(tokens) |
|
|
|
|
|
return metric.bleu(new_trans, references) |
|
|
|
|
|
|
|
|
def dump_tanslation(tranes, output, indices=None, attentions=None, video_path=None): |
|
|
"""save translation""" |
|
|
if indices is not None: |
|
|
tranes = [data[1] for data in |
|
|
sorted(zip(indices, tranes), key=lambda x: x[0])] |
|
|
with tf.gfile.Open(output, 'w') as writer: |
|
|
for hypo in tranes: |
|
|
if isinstance(hypo, list): |
|
|
writer.write(' '.join(hypo) + "\n") |
|
|
else: |
|
|
writer.write(str(hypo) + "\n") |
|
|
tf.logging.info("Saving translations into {}".format(output)) |
|
|
|
|
|
|
|
|
tf.logging.info(f"[DEBUG] attentions is None: {attentions is None}") |
|
|
if attentions is not None: |
|
|
tf.logging.info(f"[DEBUG] attentions type: {type(attentions)}, len: {len(attentions)}") |
|
|
|
|
|
|
|
|
if attentions is not None and len(attentions) > 0: |
|
|
tf.logging.info("[DEBUG] Calling dump_detailed_attention_output") |
|
|
try: |
|
|
dump_detailed_attention_output(tranes, output, indices, attentions, video_path) |
|
|
except Exception as e: |
|
|
tf.logging.warning(f"Failed to save detailed attention output: {e}") |
|
|
import traceback |
|
|
tf.logging.warning(traceback.format_exc()) |
|
|
else: |
|
|
tf.logging.info("[DEBUG] Skipping attention analysis (attentions is None or empty)") |
|
|
|
|
|
|
|
|
def dump_translation_with_reference(tranes, output, ref_file, indices=None, remove_bpe=False, |
|
|
attentions=None, dataset=None): |
|
|
""" |
|
|
Save translation with reference for easy comparison |
|
|
|
|
|
Args: |
|
|
tranes: Translation results |
|
|
output: Output file path |
|
|
ref_file: Reference file path |
|
|
indices: Sample indices |
|
|
remove_bpe: Whether to remove BPE |
|
|
attentions: Attention weights (list of numpy arrays) for frame alignment |
|
|
dataset: Dataset object for getting video frame counts |
|
|
|
|
|
Returns: |
|
|
dict: Coverage metrics {'complete_coverage': float, 'avg_iou': float} |
|
|
""" |
|
|
import tensorflow as tf |
|
|
import numpy as np |
|
|
import json |
|
|
import os |
|
|
import sys |
|
|
from pathlib import Path |
|
|
|
|
|
per_sample_attn = None |
|
|
if attentions is not None and len(attentions) > 0: |
|
|
per_sample_attn = _flatten_attention_batches(attentions) |
|
|
if len(per_sample_attn) == 0: |
|
|
per_sample_attn = None |
|
|
|
|
|
|
|
|
if per_sample_attn is None: |
|
|
per_sample_attn = [None] * len(tranes) |
|
|
elif len(per_sample_attn) < len(tranes): |
|
|
per_sample_attn.extend([None] * (len(tranes) - len(per_sample_attn))) |
|
|
|
|
|
|
|
|
if indices is not None: |
|
|
sorted_data = sorted( |
|
|
zip(indices, tranes, per_sample_attn), |
|
|
key=lambda x: x[0] |
|
|
) |
|
|
tranes = [data[1] for data in sorted_data] |
|
|
per_sample_attn = [data[2] for data in sorted_data] |
|
|
sorted_indices = [data[0] for data in sorted_data] |
|
|
else: |
|
|
sorted_indices = list(range(len(tranes))) |
|
|
|
|
|
|
|
|
references = [] |
|
|
if tf.gfile.Exists(ref_file): |
|
|
refs = tf.gfile.Open(ref_file).readlines() |
|
|
for line in refs: |
|
|
tokens = strip_leading_index_tokens(line.strip().split()) |
|
|
ref = ' '.join(tokens) |
|
|
if remove_bpe: |
|
|
ref = ref.replace("@@ ", "") |
|
|
references.append(ref) |
|
|
|
|
|
|
|
|
all_complete_coverage = [] |
|
|
all_iou = [] |
|
|
|
|
|
|
|
|
gt_annotations = None |
|
|
test_info = None |
|
|
try: |
|
|
|
|
|
output_abs = os.path.abspath(output) |
|
|
output_parts = output_abs.split(os.sep) |
|
|
if 'SignX' in output_parts: |
|
|
signx_idx = output_parts.index('SignX') |
|
|
repo_root = os.sep.join(output_parts[:signx_idx+1]) |
|
|
else: |
|
|
|
|
|
repo_root = None |
|
|
current_path = Path(output_abs).parent |
|
|
for parent in [current_path] + list(current_path.parents): |
|
|
if parent.name == 'SignX': |
|
|
repo_root = str(parent) |
|
|
break |
|
|
if repo_root is None: |
|
|
raise RuntimeError("Cannot locate SignX repo root from output path.") |
|
|
|
|
|
|
|
|
gt_path = os.path.join(os.path.dirname(repo_root), 'ASLLRP_utterances_with_frames.json') |
|
|
if os.path.exists(gt_path): |
|
|
with open(gt_path, 'r') as f: |
|
|
gt_annotations = json.load(f) |
|
|
tf.logging.info(f"Loaded GT annotations from {gt_path}") |
|
|
|
|
|
|
|
|
test_info_path = os.path.join(os.path.dirname(repo_root), 'CorrNet_Plus_modified/dataset/ASLLRP/test_info.npy') |
|
|
if os.path.exists(test_info_path): |
|
|
test_info = np.load(test_info_path, allow_pickle=True) |
|
|
tf.logging.info(f"Loaded test info from {test_info_path}") |
|
|
except Exception as e: |
|
|
tf.logging.warning(f"Could not load GT annotations or test info: {e}") |
|
|
|
|
|
|
|
|
AttentionAnalyzer = None |
|
|
if attentions is not None and len(attentions) > 0 and gt_annotations is not None: |
|
|
try: |
|
|
|
|
|
eval_dir = os.path.join(os.path.dirname(__file__), '..', 'eval') |
|
|
if os.path.exists(eval_dir) and eval_dir not in sys.path: |
|
|
sys.path.insert(0, eval_dir) |
|
|
from attention_analysis import AttentionAnalyzer as AA |
|
|
AttentionAnalyzer = AA |
|
|
except ImportError as e: |
|
|
tf.logging.warning(f"Could not import AttentionAnalyzer: {e}") |
|
|
|
|
|
|
|
|
comparison_file = output.replace('.trans.txt', '.comparison.txt') |
|
|
with tf.gfile.Open(comparison_file, 'w') as writer: |
|
|
writer.write("=" * 100 + "\n") |
|
|
writer.write("TRANSLATION COMPARISON (Hypothesis vs Reference)\n") |
|
|
writer.write("=" * 100 + "\n\n") |
|
|
|
|
|
for i, hypo in enumerate(tranes): |
|
|
|
|
|
if isinstance(hypo, list): |
|
|
hypo_str = ' '.join(hypo) |
|
|
hypo_tokens = hypo |
|
|
else: |
|
|
hypo_str = str(hypo) |
|
|
hypo_tokens = hypo_str.split() |
|
|
|
|
|
if remove_bpe: |
|
|
hypo_str = hypo_str.replace("@@ ", "") |
|
|
hypo_tokens = [t.replace("@@", "").strip() for t in hypo_tokens] |
|
|
hypo_tokens = [t for t in hypo_tokens if t] |
|
|
|
|
|
|
|
|
ref_str = references[i] if i < len(references) else "[NO REFERENCE]" |
|
|
ref_tokens = strip_leading_index_tokens(ref_str.split()) |
|
|
ref_str = ' '.join(ref_tokens) |
|
|
|
|
|
|
|
|
pred_frame_info = None |
|
|
gt_frame_info = None |
|
|
coverage_metrics = None |
|
|
|
|
|
sample_attn = per_sample_attn[i] if per_sample_attn and i < len(per_sample_attn) else None |
|
|
|
|
|
if (AttentionAnalyzer is not None and sample_attn is not None and |
|
|
test_info is not None and gt_annotations is not None): |
|
|
try: |
|
|
sample_idx = sorted_indices[i] |
|
|
|
|
|
|
|
|
video_id = test_info[sample_idx]['video_id'] |
|
|
|
|
|
|
|
|
if video_id in gt_annotations: |
|
|
gt_glosses = gt_annotations[video_id]['glosses'] |
|
|
|
|
|
attn = sample_attn |
|
|
|
|
|
|
|
|
if isinstance(attn, np.ndarray) and len(attn.shape) >= 2: |
|
|
|
|
|
if len(attn.shape) == 4: |
|
|
attn = attn[:, 0, 0, :] |
|
|
elif len(attn.shape) == 3: |
|
|
attn = attn[:, 0, :] |
|
|
|
|
|
feature_frames = attn.shape[1] |
|
|
|
|
|
|
|
|
analyzer = AttentionAnalyzer( |
|
|
attentions=attn, |
|
|
translation=' '.join(hypo_tokens), |
|
|
video_frames=feature_frames |
|
|
) |
|
|
|
|
|
|
|
|
pred_frame_ranges = analyzer.word_frame_ranges |
|
|
|
|
|
|
|
|
complete_count = 0 |
|
|
iou_scores = [] |
|
|
|
|
|
|
|
|
gt_frame_map = {gt['gloss']: (gt['start_24fps'], gt['end_24fps']) |
|
|
for gt in gt_glosses} |
|
|
|
|
|
for pred_info in pred_frame_ranges: |
|
|
pred_gloss = pred_info['word'] |
|
|
pred_start = pred_info['start_frame'] |
|
|
pred_end = pred_info['end_frame'] |
|
|
|
|
|
if pred_gloss in gt_frame_map: |
|
|
gt_start, gt_end = gt_frame_map[pred_gloss] |
|
|
|
|
|
|
|
|
if pred_start <= gt_start and pred_end >= gt_end: |
|
|
complete_count += 1 |
|
|
|
|
|
|
|
|
intersection_start = max(pred_start, gt_start) |
|
|
intersection_end = min(pred_end, gt_end) |
|
|
intersection = max(0, intersection_end - intersection_start + 1) |
|
|
|
|
|
union_start = min(pred_start, gt_start) |
|
|
union_end = max(pred_end, gt_end) |
|
|
union = union_end - union_start + 1 |
|
|
|
|
|
iou = intersection / union if union > 0 else 0.0 |
|
|
iou_scores.append(iou) |
|
|
|
|
|
|
|
|
num_matched = len(iou_scores) |
|
|
if num_matched > 0: |
|
|
complete_coverage_rate = complete_count / num_matched |
|
|
avg_iou = np.mean(iou_scores) |
|
|
|
|
|
all_complete_coverage.append(complete_coverage_rate) |
|
|
all_iou.append(avg_iou) |
|
|
|
|
|
coverage_metrics = { |
|
|
'complete_coverage_rate': complete_coverage_rate, |
|
|
'avg_iou': avg_iou, |
|
|
'num_matched': num_matched |
|
|
} |
|
|
|
|
|
pred_frame_info = pred_frame_ranges |
|
|
gt_frame_info = gt_frame_map |
|
|
|
|
|
except Exception as e: |
|
|
tf.logging.warning(f"Error calculating frame alignment for sample {i}: {e}") |
|
|
|
|
|
|
|
|
writer.write(f"[Sample {i}]\n") |
|
|
|
|
|
|
|
|
if pred_frame_info: |
|
|
hyp_with_frames = [] |
|
|
for pred_info in pred_frame_info: |
|
|
gloss = pred_info['word'] |
|
|
start = pred_info['start_frame'] |
|
|
end = pred_info['end_frame'] |
|
|
hyp_with_frames.append(f"{gloss}[{start}-{end}]") |
|
|
writer.write(f"HYP: {' '.join(hyp_with_frames)}\n") |
|
|
else: |
|
|
writer.write(f"HYP: {hypo_str}\n") |
|
|
|
|
|
|
|
|
if gt_frame_info: |
|
|
ref_with_frames = [] |
|
|
for token in ref_tokens: |
|
|
if token in gt_frame_info: |
|
|
start, end = gt_frame_info[token] |
|
|
ref_with_frames.append(f"{token}[{start}-{end}]") |
|
|
else: |
|
|
ref_with_frames.append(token) |
|
|
writer.write(f"REF: {' '.join(ref_with_frames)}\n") |
|
|
else: |
|
|
writer.write(f"REF: {ref_str}\n") |
|
|
|
|
|
|
|
|
if coverage_metrics: |
|
|
writer.write(f"Complete Coverage: {coverage_metrics['complete_coverage_rate']:.2%} " |
|
|
f"({coverage_metrics['num_matched']} matched glosses)\n") |
|
|
writer.write(f"IoU Coverage: {coverage_metrics['avg_iou']:.4f}\n") |
|
|
|
|
|
writer.write("-" * 100 + "\n\n") |
|
|
|
|
|
tf.logging.info("Saving comparison into {}".format(comparison_file)) |
|
|
|
|
|
|
|
|
result_metrics = {} |
|
|
if all_complete_coverage: |
|
|
result_metrics['complete_coverage'] = float(np.mean(all_complete_coverage)) |
|
|
result_metrics['avg_iou'] = float(np.mean(all_iou)) |
|
|
tf.logging.info(f"Coverage Metrics: Complete={result_metrics['complete_coverage']:.2%}, " |
|
|
f"IoU={result_metrics['avg_iou']:.4f}") |
|
|
|
|
|
return result_metrics |
|
|
|
|
|
|
|
|
def dump_detailed_attention_output(tranes, output, indices, attentions, video_path=None): |
|
|
""" |
|
|
保存详细的attention分析结果 |
|
|
|
|
|
Args: |
|
|
tranes: 翻译结果列表 |
|
|
output: 输出文件路径 |
|
|
indices: 样本索引 |
|
|
attentions: attention权重数据(list of numpy arrays) |
|
|
video_path: 视频文件路径(可选,用于提取视频帧) |
|
|
""" |
|
|
import os |
|
|
import sys |
|
|
import re |
|
|
from datetime import datetime |
|
|
from pathlib import Path |
|
|
|
|
|
|
|
|
output_path = Path(output) |
|
|
base_dir = output_path.parent |
|
|
base_name = output_path.stem |
|
|
|
|
|
|
|
|
timestamp = datetime.now().strftime("%Y%m%d_%H%M%S") |
|
|
detail_dir = base_dir / f"detailed_{base_name}_{timestamp}" |
|
|
detail_dir.mkdir(parents=True, exist_ok=True) |
|
|
|
|
|
tf.logging.info(f"Saving detailed attention analysis to: {detail_dir}") |
|
|
|
|
|
|
|
|
if indices is not None: |
|
|
sorted_items = sorted(zip(indices, tranes), key=lambda x: x[0]) |
|
|
tranes = [item[1] for item in sorted_items] |
|
|
|
|
|
|
|
|
try: |
|
|
import numpy as np |
|
|
|
|
|
flattened_attn = _flatten_attention_batches(attentions) |
|
|
if len(flattened_attn) == 0: |
|
|
tf.logging.warning("No valid attention data found") |
|
|
return |
|
|
|
|
|
tf.logging.info(f"[DEBUG] Found {len(flattened_attn)} valid attention samples") |
|
|
|
|
|
|
|
|
video_based_name = None |
|
|
if video_path: |
|
|
try: |
|
|
candidate = Path(video_path).stem |
|
|
if candidate: |
|
|
sanitized = re.sub(r'[^0-9A-Za-z._-]+', '_', candidate).strip('_') |
|
|
if sanitized: |
|
|
video_based_name = sanitized |
|
|
except Exception as exc: |
|
|
tf.logging.warning(f"Failed to derive video-based sample name: {exc}") |
|
|
|
|
|
use_video_name = video_based_name is not None and len(flattened_attn) == 1 |
|
|
gt_text = _lookup_ground_truth_text(video_path) if video_path else None |
|
|
|
|
|
for sample_idx, sample_attn in enumerate(flattened_attn): |
|
|
if sample_idx >= len(tranes): |
|
|
break |
|
|
|
|
|
|
|
|
if sample_attn.ndim == 4: |
|
|
sample_attn = sample_attn[:, 0, 0, :] |
|
|
elif sample_attn.ndim == 3: |
|
|
sample_attn = sample_attn[:, 0, :] |
|
|
|
|
|
|
|
|
trans = tranes[sample_idx] |
|
|
if isinstance(trans, list): |
|
|
trans = ' '.join(trans) |
|
|
trans_clean = trans.replace('@@ ', '') |
|
|
|
|
|
|
|
|
if use_video_name and sample_idx == 0: |
|
|
sample_name = video_based_name |
|
|
else: |
|
|
sample_name = f"sample_{sample_idx:03d}" |
|
|
sample_dir = detail_dir / sample_name |
|
|
sample_dir.mkdir(exist_ok=True) |
|
|
|
|
|
|
|
|
np.save(sample_dir / "attention_weights.npy", sample_attn) |
|
|
|
|
|
|
|
|
with open(sample_dir / "translation.txt", 'w', encoding='utf-8') as f: |
|
|
f.write(f"With BPE: {trans}\n") |
|
|
f.write(f"Clean: {trans_clean}\n") |
|
|
if gt_text: |
|
|
f.write(f"Ground Truth: {gt_text}\n") |
|
|
else: |
|
|
f.write("Ground Truth: [NOT FOUND]\n") |
|
|
|
|
|
|
|
|
if video_path and os.path.exists(video_path): |
|
|
try: |
|
|
import cv2 |
|
|
import json |
|
|
|
|
|
cap = cv2.VideoCapture(video_path) |
|
|
original_frame_count = int(cap.get(cv2.CAP_PROP_FRAME_COUNT)) |
|
|
cap.release() |
|
|
|
|
|
feature_count = sample_attn.shape[1] |
|
|
frame_mapping = [] |
|
|
for feat_idx in range(feature_count): |
|
|
start_frame = int(feat_idx * original_frame_count / feature_count) |
|
|
end_frame = int((feat_idx + 1) * original_frame_count / feature_count) |
|
|
frame_mapping.append({ |
|
|
"feature_index": feat_idx, |
|
|
"frame_start": start_frame, |
|
|
"frame_end": end_frame, |
|
|
"frame_count": end_frame - start_frame |
|
|
}) |
|
|
|
|
|
mapping_data = { |
|
|
"original_frame_count": original_frame_count, |
|
|
"feature_count": feature_count, |
|
|
"downsampling_ratio": original_frame_count / feature_count, |
|
|
"mapping": frame_mapping |
|
|
} |
|
|
|
|
|
with open(sample_dir / "feature_frame_mapping.json", 'w') as f: |
|
|
json.dump(mapping_data, f, indent=2) |
|
|
|
|
|
tf.logging.info(f" ✓ Feature-to-frame mapping saved ({original_frame_count} frames → {feature_count} features)") |
|
|
|
|
|
except Exception as e: |
|
|
tf.logging.warning(f"Failed to generate feature-to-frame mapping: {e}") |
|
|
|
|
|
|
|
|
try: |
|
|
script_dir = Path(__file__).parent.parent |
|
|
eval_dir = script_dir / "eval" |
|
|
if str(eval_dir) not in sys.path: |
|
|
sys.path.insert(0, str(eval_dir)) |
|
|
|
|
|
from attention_analysis import AttentionAnalyzer |
|
|
|
|
|
analyzer = AttentionAnalyzer( |
|
|
attentions=sample_attn, |
|
|
translation=trans_clean, |
|
|
video_frames=sample_attn.shape[1], |
|
|
video_path=video_path |
|
|
) |
|
|
analyzer.generate_all_visualizations(sample_dir) |
|
|
tf.logging.info(f" ✓ Sample {sample_idx}: {sample_dir.name}") |
|
|
|
|
|
except Exception as e: |
|
|
tf.logging.warning(f"Failed to generate visualizations for sample {sample_idx}: {e}") |
|
|
|
|
|
tf.logging.info(f"Detailed attention analysis complete: {detail_dir}") |
|
|
tf.logging.info(f" - Analyzed {min(len(flattened_attn), len(tranes))} samples") |
|
|
tf.logging.info(f" - Output directory: {detail_dir}") |
|
|
|
|
|
except Exception as e: |
|
|
import traceback |
|
|
tf.logging.error(f"Error in dump_detailed_attention_output: {e}") |
|
|
tf.logging.error(traceback.format_exc()) |
|
|
|