FangSen9000
Successfully completed the attention-based segmentation inference
9803b71
# coding: utf-8
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import tensorflow as tf
from utils import util, dtype
from collections import namedtuple
from tensorflow.python.util import nest
class BeamSearchState(namedtuple("BeamSearchState",
("inputs", "state", "finish", "attention_history"))):
pass
def beam_search(features, encoding_fn, decoding_fn, params):
decode_length = params.decode_length
beam_size = params.beam_size
alpha = params.decode_alpha
eos_id = params.tgt_vocab.eos()
pad_id = params.tgt_vocab.pad()
eval_task = params.eval_task
# Check if attention collection is enabled
collect_attention = getattr(params, 'collect_attention_weights', False)
tf.logging.info(f"[DEBUG] beam_search: collect_attention_weights={collect_attention}")
batch_size = tf.shape(features["image"])[0]
beam_i32 = tf.constant(beam_size, dtype=tf.int32)
one_i32 = tf.constant(1, dtype=tf.int32)
batch_beam_shape = tf.stack([batch_size, beam_i32])
batch_beam_seq_shape = tf.stack([batch_size, beam_i32, one_i32])
if eval_task in ('sign2text', 'sign2gloss'):
model_state = encoding_fn(features["image"], features["mask"])
else:
model_state = encoding_fn(features["source"], None)
src_mask = features["mask"]
source_length = tf.reduce_sum(src_mask, -1)
max_target_length = source_length + decode_length
model_state = nest.map_structure(
lambda x: util.expand_tile_dims(x, beam_size, axis=1),
model_state
)
# in our mixed precision mode, we finally convert logits into tf.float32
# tfdtype = tf.as_dtype(dtype.floatx())
tfdtype = tf.float32
# [batch, beam]
init_log_probs = tf.constant([[0.] + [tfdtype.min] * (beam_size - 1)], dtype=tfdtype)
init_log_probs = tf.tile(init_log_probs, tf.stack([batch_size, one_i32]))
init_scores = tf.zeros_like(init_log_probs)
# [batch, beam, 1], begin-of-sequence
init_seq = tf.fill(batch_beam_seq_shape, params.tgt_vocab.pad())
init_finish_seq = tf.zeros_like(init_seq)
# [batch, beam]
init_finish_scores = tf.fill(batch_beam_shape, tfdtype.min)
init_finish_flags = tf.zeros(batch_beam_shape, tf.bool)
def cache_init(prev_seq, state):
# used to initialize some caches
# this is because pre-compute these caches is to hard,
# so let's use one dummy run to obtain them.
flat_prev_seqs = util.merge_neighbor_dims(prev_seq, axis=0)
flat_prev_state = nest.map_structure(
lambda x: util.merge_neighbor_dims(x, axis=0),
state
)
_, step_state = decoding_fn(
flat_prev_seqs[:, -1:], flat_prev_state, 0)
new_state = nest.map_structure(
lambda x: util.unmerge_neighbor_dims(x, batch_size, axis=0),
step_state
)
new_state = util.dict_update(new_state, state)
return new_state
model_state = cache_init(init_seq, model_state)
# Remove cross_attention from initial state (it's not part of the recurrent state)
# It will be computed fresh at each step and collected separately
if 'cross_attention' in model_state:
model_state = {k: v for k, v in model_state.items() if k != 'cross_attention'}
# Always initialize attention history TensorArray (for while_loop compatibility)
# But only write to it if collection is enabled
init_attention_history = tf.TensorArray(
dtype=tfdtype,
size=0,
dynamic_size=True,
clear_after_read=False,
element_shape=tf.TensorShape([None, None, None]) # [batch, beam, src_len]
)
bsstate = BeamSearchState(
inputs=(init_seq, init_log_probs, init_scores),
state=model_state,
finish=(init_finish_seq, init_finish_scores, init_finish_flags),
attention_history=init_attention_history
)
def _not_finished(time, bsstate):
# if the maximum time step is reached, or
# all samples in one batch satisfy that the worst finished sequence
# score is not better than the best alive sequence score
alive_log_probs = bsstate.inputs[1]
finish_scores = bsstate.finish[1]
finish_flags = bsstate.finish[2]
# upper bound of length penality
max_length_penality = tf.pow(
(5. + tf.cast(max_target_length, tfdtype)) / 6., alpha)
best_alive_score = alive_log_probs[:, 0] / max_length_penality
# minimum score among finished sequences alone
worst_finish_score = tf.reduce_min(
finish_scores * tf.cast(finish_flags, tfdtype), 1)
# deal with unfinished instances, which is set to `tf.float32.min`
unfinish_mask = 1. - tf.cast(tf.reduce_any(finish_flags, 1), tfdtype)
worst_finish_score += unfinish_mask * tfdtype.min
# boundary
bound_is_met = tf.reduce_all(tf.greater(worst_finish_score,
best_alive_score))
# length constraint
length_is_met = tf.reduce_any(
tf.less(time, tf.cast(max_target_length, tf.int32)))
return tf.logical_and(tf.logical_not(bound_is_met), length_is_met)
def _step_fn(time, bsstate):
"""one expansion step of beam search process"""
# 1. feed previous predictions, and get the next probabilities
# generating beam * vocab_size predictions
prev_seq, prev_log_probs, prev_scores = bsstate.inputs
flat_prev_seqs = util.merge_neighbor_dims(prev_seq, axis=0)
flat_prev_state = nest.map_structure(
lambda x: util.merge_neighbor_dims(x, axis=0),
bsstate.state
)
# curr_logits: [batch * beam, vocab_size]
step_logits, step_state = decoding_fn(
flat_prev_seqs[:, -1:], flat_prev_state, time)
# add gumbel noise into the logits, simulate gumbel top-k sampling without replacement
if params.enable_noise_beam_search:
step_logits += util.gumbel_noise(util.shape_list(step_logits))
# apply temperature decoding
step_logits /= params.beam_search_temperature
step_log_probs = util.log_prob_from_logits(step_logits)
vocab_size = util.shape_list(step_log_probs)[-1]
# force decoding: disable empty output
eos_mask = tf.cast(tf.equal(tf.range(vocab_size), eos_id), tfdtype)
step_log_probs = tf.cond(dtype.tf_to_float(time) < dtype.tf_to_float(1.),
lambda: step_log_probs + tf.expand_dims(eos_mask, 0) * - dtype.inf(),
lambda: step_log_probs)
# expand to [batch, beam, vocab_size]
step_log_probs = util.unmerge_neighbor_dims(step_log_probs,
batch_size, axis=0)
step_state = nest.map_structure(
lambda x: util.unmerge_neighbor_dims(x, batch_size, axis=0),
step_state
)
# 2. compute top-k scored next predictions
# reducing beam * vocab_size to 2 * beam
# [batch, beam, 1] + [batch, beam, vocab_size]
curr_log_probs = tf.expand_dims(prev_log_probs, 2) + step_log_probs
length_penality = tf.pow((5.0 + tf.cast(time + 1, tfdtype)) / 6., alpha)
curr_scores = curr_log_probs / length_penality
# [batch, beam * vocab_size]
curr_flat_scores = util.merge_neighbor_dims(curr_scores, axis=1)
# [batch, 2 * beam]
topk_scores, topk_indices = tf.nn.top_k(
curr_flat_scores, 2 * beam_size)
# index manipulation, [batch, 2 * beam]
curr_beam_indices = topk_indices // vocab_size
curr_symbol_indices = topk_indices % vocab_size
beam2_pos = util.batch_coordinates(batch_size, 2 * beam_size)
curr_coordinates = tf.stack([beam2_pos, curr_beam_indices], axis=2)
# extract candidate sequences
# [batch, 2 * beam, time + 1]
curr_seq = tf.gather_nd(prev_seq, curr_coordinates)
curr_seq = tf.concat([curr_seq,
tf.expand_dims(curr_symbol_indices, 2)], 2)
# 3. handling alive sequences
# reducing 2 * beam to beam
curr_fin_flags = tf.logical_or(
tf.equal(curr_symbol_indices, eos_id),
# if time step exceeds the maximum decoding length, should stop
tf.expand_dims(
tf.greater_equal(time, tf.cast(max_target_length, tf.int32)), 1)
)
alive_scores = topk_scores + tf.cast(curr_fin_flags, tfdtype) * tfdtype.min
# [batch, 2 * beam] -> [batch, beam]
alive_scores, alive_indices = tf.nn.top_k(alive_scores, beam_size)
beam_pos = util.batch_coordinates(batch_size, beam_size)
alive_coordinates = tf.stack([beam_pos, alive_indices], axis=2)
alive_seq = tf.gather_nd(curr_seq, alive_coordinates)
alive_beam_indices = tf.gather_nd(curr_beam_indices, alive_coordinates)
beam_coordinates = tf.stack([beam_pos, alive_beam_indices], axis=2)
alive_state = nest.map_structure(
lambda x: tf.gather_nd(x, beam_coordinates),
step_state
)
alive_log_probs = alive_scores * length_penality
# Collect cross-attention weights if collection is enabled
# Also remove cross_attention from alive_state to maintain consistent structure
updated_attention_history = bsstate.attention_history
if 'cross_attention' in alive_state:
if collect_attention:
# step_state['cross_attention']: [batch, beam, 1, src_len] (already unmerged)
# Squeeze the tgt_len dimension: [batch, beam, src_len]
attention_weights = step_state['cross_attention'][:, :, 0, :]
# Reorder according to alive beams
attention_weights = tf.gather_nd(attention_weights, beam_coordinates)
# Write to TensorArray
updated_attention_history = bsstate.attention_history.write(time, attention_weights)
# Remove cross_attention from alive_state (not part of recurrent state)
alive_state = {k: v for k, v in alive_state.items() if k != 'cross_attention'}
# 4. handle finished sequences
# reducing 3 * beam to beam
prev_fin_seq, prev_fin_scores, prev_fin_flags = bsstate.finish
# [batch, 2 * beam]
curr_fin_scores = topk_scores + (1.0 - tf.cast(curr_fin_flags, tfdtype)) * tfdtype.min
# [batch, 3 * beam]
fin_flags = tf.concat([prev_fin_flags, curr_fin_flags], axis=1)
fin_scores = tf.concat([prev_fin_scores, curr_fin_scores], axis=1)
# [batch, beam]
fin_scores, fin_indices = tf.nn.top_k(fin_scores, beam_size)
fin_coordinates = tf.stack([beam_pos, fin_indices], axis=2)
fin_flags = tf.gather_nd(fin_flags, fin_coordinates)
pad_seq = tf.fill([batch_size, beam_size, 1],
tf.constant(pad_id, tf.int32))
prev_fin_seq = tf.concat([prev_fin_seq, pad_seq], axis=2)
fin_seq = tf.concat([prev_fin_seq, curr_seq], axis=1)
fin_seq = tf.gather_nd(fin_seq, fin_coordinates)
next_state = BeamSearchState(
inputs=(alive_seq, alive_log_probs, alive_scores),
state=alive_state,
finish=(fin_seq, fin_scores, fin_flags),
attention_history=updated_attention_history
)
return time + 1, next_state
time = tf.constant(0, tf.int32, name="time")
shape_invariants = BeamSearchState(
inputs=(tf.TensorShape([None, None, None]),
tf.TensorShape([None, None]),
tf.TensorShape([None, None])),
state=nest.map_structure(
lambda x: util.get_shape_invariants(x),
bsstate.state
),
finish=(tf.TensorShape([None, None, None]),
tf.TensorShape([None, None]),
tf.TensorShape([None, None])),
attention_history=tf.TensorShape(None) # TensorArray shape
)
outputs = tf.while_loop(_not_finished, _step_fn, [time, bsstate],
shape_invariants=[tf.TensorShape([]),
shape_invariants],
parallel_iterations=32,
back_prop=False)
final_state = outputs[1]
alive_seqs = final_state.inputs[0]
init_scores = final_state.inputs[2]
final_seqs = final_state.finish[0]
final_scores = final_state.finish[1]
final_flags = final_state.finish[2]
alive_seqs.set_shape([None, beam_size, None])
final_seqs.set_shape([None, beam_size, None])
final_seqs = tf.where(tf.reduce_any(final_flags, 1), final_seqs,
alive_seqs)
final_scores = tf.where(tf.reduce_any(final_flags, 1), final_scores,
init_scores)
result = {
'seq': final_seqs[:, :, 1:],
'score': final_scores
}
# Only include attention history if collection was enabled
if collect_attention:
# Stack attention history from TensorArray
# Returns [time_steps, batch, beam, src_len]
attention_history_tensor = final_state.attention_history.stack()
result['attention_history'] = attention_history_tensor
return result