|
|
import copy |
|
|
import os |
|
|
from decimal import Decimal, ROUND_HALF_UP |
|
|
from typing import Any, Callable, Dict, Optional, Tuple, Union, TYPE_CHECKING |
|
|
|
|
|
import numpy as np |
|
|
import torch |
|
|
import torch.utils.checkpoint |
|
|
import torch.utils.checkpoint |
|
|
from torch import nn |
|
|
from torch.nn.utils.rnn import pad_sequence |
|
|
from transformers import PreTrainedModel |
|
|
from transformers.generation.configuration_utils import GenerationConfig, GenerationMode |
|
|
from transformers.generation.logits_process import ( |
|
|
LogitsProcessorList, |
|
|
SuppressTokensAtBeginLogitsProcessor, |
|
|
SuppressTokensLogitsProcessor, ) |
|
|
from transformers.generation.logits_process import WhisperNoSpeechDetection |
|
|
from transformers.generation.stopping_criteria import ( |
|
|
StoppingCriteriaList, |
|
|
) |
|
|
from transformers.generation.utils import GenerateNonBeamOutput, \ |
|
|
GenerateEncoderDecoderOutput, GenerateDecoderOnlyOutput, GenerateBeamOutput, GenerateBeamDecoderOnlyOutput, \ |
|
|
GenerateBeamEncoderDecoderOutput |
|
|
from transformers.modeling_outputs import BaseModelOutput |
|
|
from transformers.models.whisper.modeling_whisper import ( |
|
|
WhisperForConditionalGeneration, |
|
|
) |
|
|
from transformers.utils import logging |
|
|
from .decoding import CTCRescorerLogitsProcessor, LogSoftmaxProcessor |
|
|
from .utils import WhisperTimeStampLogitsProcessorCustom |
|
|
|
|
|
if TYPE_CHECKING: |
|
|
from transformers.generation.streamers import BaseStreamer |
|
|
|
|
|
logging.set_verbosity_debug() |
|
|
logger = logging.get_logger("transformers") |
|
|
|
|
|
|
|
|
class DiCoWGenerationMixin(WhisperForConditionalGeneration): |
|
|
|
|
|
def _prepare_encoder_decoder_kwargs_for_generation( |
|
|
self, inputs_tensor: torch.Tensor, model_kwargs, model_input_name, generation_config, |
|
|
) -> Dict[str, Any]: |
|
|
|
|
|
model_kwargs = super()._prepare_encoder_decoder_kwargs_for_generation( |
|
|
inputs_tensor, model_kwargs, model_input_name, generation_config |
|
|
) |
|
|
|
|
|
if hasattr(generation_config, "ctc_weight") and generation_config.ctc_weight > 0: |
|
|
self.encoder_logits = self.get_enc_logits(model_kwargs["encoder_outputs"].last_hidden_state) |
|
|
|
|
|
return model_kwargs |
|
|
|
|
|
def _prepare_decoder_input_ids_for_generation( |
|
|
self, |
|
|
batch_size: int, |
|
|
model_input_name: str, |
|
|
model_kwargs: Dict[str, torch.Tensor], |
|
|
decoder_start_token_id: torch.Tensor, |
|
|
device: torch.device = None, |
|
|
) -> Tuple[torch.LongTensor, Dict[str, torch.Tensor]]: |
|
|
batch_size = model_kwargs['decoder_input_ids'].shape[0] |
|
|
out = super()._prepare_decoder_input_ids_for_generation( |
|
|
batch_size, |
|
|
model_input_name, |
|
|
model_kwargs, |
|
|
decoder_start_token_id, |
|
|
device, |
|
|
) |
|
|
return out |
|
|
|
|
|
def prepare_kwargs_for_generate(self, |
|
|
max_frames, |
|
|
cur_bsz, |
|
|
batch_idx_map, |
|
|
seek, |
|
|
kwargs, |
|
|
attention_mask): |
|
|
"""This method also prepares STNO masks and other kwargs for generation.""" |
|
|
|
|
|
seek_vad = seek // 2 |
|
|
input_stride = self.model.encoder.conv1.stride[0] * self.model.encoder.conv2.stride[0] |
|
|
num_segment_frames = input_stride * self.config.max_source_positions |
|
|
num_frames_vad = num_segment_frames // 2 |
|
|
max_frames_vad = max_frames // 2 |
|
|
seek_num_frames = (max_frames_vad - seek_vad).clamp(max=num_frames_vad) |
|
|
|
|
|
stno_masks = [] |
|
|
for i in range(cur_bsz): |
|
|
prev_i = batch_idx_map[i] |
|
|
segment_input_slice = kwargs["stno_mask"][prev_i: prev_i + 1, :, |
|
|
seek_vad[prev_i]: seek_vad[prev_i] + seek_num_frames[prev_i]] |
|
|
|
|
|
if segment_input_slice.shape[-1] < num_frames_vad: |
|
|
orig_len = segment_input_slice.shape[-1] |
|
|
|
|
|
segment_input_slice = torch.nn.functional.pad( |
|
|
segment_input_slice, pad=(0, num_frames_vad - orig_len) |
|
|
) |
|
|
|
|
|
segment_input_slice[0, 0, orig_len:] = 1.0 |
|
|
|
|
|
stno_masks.append(segment_input_slice) |
|
|
kwargs["stno_mask"] = torch.cat(stno_masks, dim=0) |
|
|
self.stno_mask_seek = kwargs["stno_mask"] |
|
|
|
|
|
if self.config.use_enrollments and "enrollments" in kwargs: |
|
|
for key in kwargs["enrollments"]: |
|
|
kwargs["enrollments"][key] = kwargs["enrollments"][key][batch_idx_map] |
|
|
|
|
|
if attention_mask is not None: |
|
|
attention_mask = attention_mask[batch_idx_map] |
|
|
|
|
|
if "labels" in kwargs: |
|
|
kwargs['labels'] = kwargs["labels"][batch_idx_map] |
|
|
kwargs['upp_labels'] = kwargs["upp_labels"][batch_idx_map] |
|
|
return kwargs, attention_mask |
|
|
|
|
|
|
|
|
def _retrieve_init_tokens(self, input_features, batch_size, generation_config, config, num_segment_frames, kwargs): |
|
|
task = getattr(generation_config, "task", None) |
|
|
language = getattr(generation_config, "language", None) |
|
|
|
|
|
if "enrollments" in kwargs: |
|
|
self.enrollments = kwargs["enrollments"] |
|
|
|
|
|
forced_decoder_ids = generation_config.forced_decoder_ids if hasattr(generation_config, "forced_decoder_ids") else None |
|
|
if forced_decoder_ids is not None: |
|
|
if language is None and task is None and forced_decoder_ids[0][1] is None: |
|
|
logger.warning_once( |
|
|
"Due to a bug fix in https://github.com/huggingface/transformers/pull/28687 transcription using a multilingual Whisper will default to language detection followed by transcription instead of translation to English." |
|
|
"This might be a breaking change for your use case. If you want to instead always translate your audio to English, make sure to pass `language='en'`." |
|
|
) |
|
|
elif hasattr(config, "forced_decoder_ids") and config.forced_decoder_ids is not None: |
|
|
forced_decoder_ids = config.forced_decoder_ids |
|
|
|
|
|
elif forced_decoder_ids is not None and language is not None: |
|
|
logger.info( |
|
|
f"You have passed language={language}, but also have set `forced_decoder_ids` to {forced_decoder_ids} which creates a conflict. `forced_decoder_ids` will be ignored in favor of language={language}." |
|
|
) |
|
|
forced_decoder_ids = None |
|
|
|
|
|
if forced_decoder_ids is not None: |
|
|
return forced_decoder_ids |
|
|
|
|
|
init_tokens = super()._retrieve_init_tokens(input_features, batch_size, generation_config, config, num_segment_frames, kwargs) |
|
|
del self.enrollments |
|
|
return init_tokens |
|
|
|
|
|
def detect_language( |
|
|
self, |
|
|
input_features: Optional[torch.FloatTensor] = None, |
|
|
encoder_outputs: Optional[Union[torch.FloatTensor, BaseModelOutput]] = None, |
|
|
generation_config: Optional[GenerationConfig] = None, |
|
|
num_segment_frames: int = 3000, |
|
|
) -> torch.Tensor: |
|
|
""" |
|
|
Detects language from log-mel input features or encoder_outputs |
|
|
|
|
|
Parameters: |
|
|
input_features (`torch.Tensor` of shape `(batch_size, feature_size, sequence_length)`, *optional*): |
|
|
Float values of log-mel features extracted from the raw speech waveform. The raw speech waveform can be obtained by |
|
|
loading a `.flac` or `.wav` audio file into an array of type `list[float]`, a `numpy.ndarray` or a `torch.Tensor`, *e.g.* via |
|
|
the soundfile library (`pip install soundfile`). To prepare the array into `input_features`, the |
|
|
[`AutoFeatureExtractor`] should be used for extracting the mel features, padding and conversion into a |
|
|
tensor of type `torch.FloatTensor`. See [`~WhisperFeatureExtractor.__call__`] for details. |
|
|
encoder_outputs (`tuple(tuple(torch.FloatTensor)`, *optional*): |
|
|
Tuple consists of (`last_hidden_state`, *optional*: `hidden_states`, *optional*: `attentions`) |
|
|
`last_hidden_state` of shape `(batch_size, sequence_length, hidden_size)`, *optional*) is a sequence of |
|
|
hidden-states at the output of the last layer of the encoder. Used in the cross-attention of the decoder. |
|
|
generation_config (`~generation.GenerationConfig`, *optional*): |
|
|
The generation configuration to be used as base parametrization for the generation call. `**kwargs` |
|
|
passed to generate matching the attributes of `generation_config` will override them. If |
|
|
`generation_config` is not provided, the default will be used, which had the following loading |
|
|
priority: 1) from the `generation_config.json` model file, if it exists; 2) from the model |
|
|
configuration. Please note that unspecified parameters will inherit [`~generation.GenerationConfig`]'s |
|
|
default values, whose documentation should be checked to parameterize generation. |
|
|
num_segment_frames (`int`, *optional*, defaults to 3000): |
|
|
The number of log-mel frames the model expects |
|
|
|
|
|
Return: |
|
|
A `torch.LongTensor` representing the detected language ids. |
|
|
""" |
|
|
if input_features is None and encoder_outputs is None: |
|
|
raise ValueError("You have to specify either `input_features` or `encoder_outputs`") |
|
|
elif input_features is not None and encoder_outputs is not None: |
|
|
raise ValueError("Make sure to specify only one of `input_features` or `encoder_outputs` - not both!") |
|
|
elif input_features is not None: |
|
|
inputs = {"input_features": input_features[:, :, :num_segment_frames]} |
|
|
batch_size = input_features.shape[0] |
|
|
elif encoder_outputs is not None: |
|
|
inputs = {"encoder_outputs": encoder_outputs} |
|
|
batch_size = ( |
|
|
encoder_outputs[0].shape[0] if isinstance(encoder_outputs, BaseModelOutput) else encoder_outputs[0] |
|
|
) |
|
|
|
|
|
generation_config = generation_config or self.generation_config |
|
|
decoder_input_ids = ( |
|
|
torch.ones((batch_size, 1), device=self.device, dtype=torch.long) |
|
|
* generation_config.decoder_start_token_id |
|
|
) |
|
|
|
|
|
with torch.no_grad(): |
|
|
|
|
|
"""<DiCoW CODE>""" |
|
|
if hasattr(self, "enrollments"): |
|
|
inputs["enrollments"] = self.enrollments |
|
|
|
|
|
logits = self(**inputs, decoder_input_ids=decoder_input_ids, use_cache=False, |
|
|
stno_mask=self.stno_mask[:, :, :num_segment_frames // 2]).logits[:, -1] |
|
|
"""</DiCoW CODE>""" |
|
|
|
|
|
non_lang_mask = torch.ones_like(logits[0], dtype=torch.bool) |
|
|
non_lang_mask[list(generation_config.lang_to_id.values())] = False |
|
|
|
|
|
logits[:, non_lang_mask] = -np.inf |
|
|
|
|
|
lang_ids = logits.argmax(-1) |
|
|
|
|
|
return lang_ids |
|
|
|
|
|
def _get_logits_processor( |
|
|
self, |
|
|
generation_config: GenerationConfig, |
|
|
input_ids_seq_length: Optional[int] = None, |
|
|
encoder_input_ids: Optional[torch.LongTensor] = None, |
|
|
prefix_allowed_tokens_fn: Optional[Callable[[int, torch.Tensor], list[int]]] = None, |
|
|
logits_processor: Optional[LogitsProcessorList] = None, |
|
|
device: Optional[str] = None, |
|
|
model_kwargs: Optional[dict[str, Any]] = None, |
|
|
negative_prompt_ids: Optional[torch.Tensor] = None, |
|
|
negative_prompt_attention_mask: Optional[torch.Tensor] = None, |
|
|
) -> LogitsProcessorList: |
|
|
|
|
|
gen_config_copy = copy.deepcopy(generation_config) |
|
|
gen_config_copy.forced_decoder_ids = None |
|
|
processors = super()._get_logits_processor( |
|
|
gen_config_copy, |
|
|
input_ids_seq_length, |
|
|
encoder_input_ids, |
|
|
prefix_allowed_tokens_fn, |
|
|
logits_processor, |
|
|
device, |
|
|
model_kwargs, |
|
|
negative_prompt_ids, |
|
|
negative_prompt_attention_mask, |
|
|
) |
|
|
if hasattr(generation_config, "ctc_weight") and generation_config.ctc_weight > 0: |
|
|
enc_logits = self.encoder_logits |
|
|
if generation_config.num_beams <= 1: |
|
|
processors.append(LogSoftmaxProcessor()) |
|
|
else: |
|
|
enc_logits = enc_logits.repeat_interleave(generation_config.num_beams, dim=0) |
|
|
self.ctc_rescorer = CTCRescorerLogitsProcessor( |
|
|
enc_logits, |
|
|
torch.full((enc_logits.shape[0],), fill_value=enc_logits.shape[1], |
|
|
device=enc_logits.device), |
|
|
enc_logits.shape[-1] - 1, |
|
|
generation_config.pad_token_id, |
|
|
generation_config.eos_token_id, |
|
|
generation_config.decoder_start_token_id, |
|
|
self.tokenizer, |
|
|
0, |
|
|
generation_config.ctc_weight, |
|
|
generation_config.num_beams, |
|
|
False, |
|
|
) |
|
|
processors.append(self.ctc_rescorer) |
|
|
return processors |
|
|
|
|
|
def _retrieve_logit_processors(self, generation_config, logits_processor, begin_index, num_beams, device): |
|
|
if generation_config.return_timestamps is True: |
|
|
"""<DiCoW CODE>""" |
|
|
timestamp_processor = WhisperTimeStampLogitsProcessorCustom(generation_config, begin_index=begin_index) |
|
|
"""</DiCoW CODE>""" |
|
|
logits_processor = ( |
|
|
[timestamp_processor] if logits_processor is None else [timestamp_processor] + logits_processor |
|
|
) |
|
|
|
|
|
if generation_config.suppress_tokens is not None: |
|
|
suppress_tokens_processor = SuppressTokensLogitsProcessor(generation_config.suppress_tokens, device=device) |
|
|
logits_processor = ( |
|
|
[suppress_tokens_processor] |
|
|
if logits_processor is None |
|
|
else [suppress_tokens_processor] + logits_processor |
|
|
) |
|
|
generation_config.suppress_tokens = None |
|
|
|
|
|
if generation_config.begin_suppress_tokens is not None: |
|
|
begin_suppress_processor = SuppressTokensAtBeginLogitsProcessor( |
|
|
generation_config.begin_suppress_tokens, begin_index=begin_index, device=device |
|
|
) |
|
|
logits_processor = ( |
|
|
[begin_suppress_processor] |
|
|
if logits_processor is None |
|
|
else [begin_suppress_processor] + logits_processor |
|
|
) |
|
|
generation_config.begin_suppress_tokens = None |
|
|
|
|
|
if generation_config.no_speech_threshold is not None: |
|
|
no_speech_detector = WhisperNoSpeechDetection( |
|
|
no_speech_token=generation_config.no_timestamps_token_id - 1, |
|
|
begin_index=begin_index, |
|
|
scores_is_logprobs=num_beams > 1, |
|
|
) |
|
|
logits_processor = ( |
|
|
[no_speech_detector] if logits_processor is None else [no_speech_detector] + logits_processor |
|
|
) |
|
|
no_speech_detector.set_model(self) |
|
|
|
|
|
return logits_processor |
|
|
|
|
|
@staticmethod |
|
|
def round_to_nearest_0_02(x): |
|
|
d = Decimal(str(x)) |
|
|
step = Decimal('0.02') |
|
|
|
|
|
rounded = (d / step).to_integral_value(rounding=ROUND_HALF_UP) * step |
|
|
return rounded |
|
|
|
|
|
def _fix_timestamps_from_segmentation(self, sequences): |
|
|
""" |
|
|
Adjusts token sequences with global timestamps to fit within Whisper's 0–30s timestamp token range. |
|
|
""" |
|
|
|
|
|
first_timestamp_token = self.tokenizer.get_vocab()["<|0.00|>"] |
|
|
empty_text_token = self.tokenizer.get_vocab()["Ġ"] |
|
|
results = [] |
|
|
|
|
|
|
|
|
for idx, sequence_segs in enumerate(sequences['segments']): |
|
|
sequences['segments'][idx] = [ |
|
|
seg for seg in sequence_segs |
|
|
if len(seg['tokens']) > 0 and (len(seg['tokens']) != 1 or seg['tokens'][0] != first_timestamp_token) |
|
|
] |
|
|
|
|
|
|
|
|
for idx, sequence_segs in enumerate(sequences['segments']): |
|
|
result = [] |
|
|
prev_segment_end_time = None |
|
|
correction = Decimal(0.0) |
|
|
|
|
|
for i, seg in enumerate(sequence_segs): |
|
|
|
|
|
start_time = self.round_to_nearest_0_02(seg['start'].item()) |
|
|
end_time = self.round_to_nearest_0_02(seg['end'].item()) |
|
|
tokens = seg['tokens'] |
|
|
|
|
|
|
|
|
current_block = (start_time + correction) // 30 |
|
|
|
|
|
if prev_segment_end_time is not None: |
|
|
|
|
|
|
|
|
|
|
|
prev_block = (prev_segment_end_time - Decimal("0.001")) // 30 |
|
|
|
|
|
num_dummies = current_block - prev_block - 1 |
|
|
|
|
|
|
|
|
if current_block > prev_block: |
|
|
result.append((30, [empty_text_token], 30)) |
|
|
|
|
|
|
|
|
for _ in range(int(num_dummies)): |
|
|
result.append((0, [empty_text_token], 30)) |
|
|
else: |
|
|
|
|
|
for _ in range(int(start_time // 30)): |
|
|
result.append((0, [empty_text_token], 30)) |
|
|
|
|
|
|
|
|
if ((start_time + correction) // 30 == (end_time + correction) // 30): |
|
|
|
|
|
result.append(((start_time + correction) % 30, tokens, (end_time + correction) % 30)) |
|
|
elif (end_time + correction) % 30 == 0: |
|
|
result.append(((start_time + correction) % 30, tokens, 30)) |
|
|
|
|
|
correction = Decimal(0.0) |
|
|
else: |
|
|
|
|
|
new_seg_start = (correction + start_time) % 30 |
|
|
seg_duration = end_time - start_time |
|
|
new_end_time = (end_time + correction) % 30 |
|
|
|
|
|
if seg_duration == 30.0: |
|
|
if float(new_seg_start) % 30.0 == 0.0: |
|
|
new_end_time = Decimal(30.0) |
|
|
correction = Decimal(0.0) |
|
|
else: |
|
|
correction = Decimal(-0.02) |
|
|
new_end_time += Decimal(correction) |
|
|
else: |
|
|
correction = Decimal(0.0) |
|
|
result.append((new_seg_start, tokens, new_end_time)) |
|
|
|
|
|
|
|
|
prev_segment_end_time = end_time + correction |
|
|
|
|
|
|
|
|
encoded = self.tokenizer( |
|
|
"".join([f"<|{seg[0]:.2f}|>{self.tokenizer.decode(seg[1])}<|{seg[2]:.2f}|>" for seg in result]) |
|
|
)['input_ids'] |
|
|
results.append(encoded) |
|
|
|
|
|
|
|
|
sequences = pad_sequence( |
|
|
[torch.tensor(res, device=sequences['sequences'].device) for res in results], |
|
|
batch_first=True, |
|
|
padding_value=self.tokenizer.pad_token_id |
|
|
) |
|
|
return sequences |
|
|
|
|
|
@staticmethod |
|
|
def _retrieve_segment( |
|
|
seek_sequence, |
|
|
seek_outputs, |
|
|
time_offset, |
|
|
timestamp_begin, |
|
|
seek_num_frames, |
|
|
time_precision, |
|
|
time_precision_features, |
|
|
input_stride, |
|
|
prev_idx, |
|
|
idx, |
|
|
return_token_timestamps, |
|
|
decoder_input_ids, |
|
|
): |
|
|
|
|
|
|
|
|
timestamp_tokens: torch.Tensor = seek_sequence.ge(timestamp_begin) |
|
|
single_timestamp_ending = timestamp_tokens[-2:].tolist() == [False, True] |
|
|
timestamp_segment_indices = torch.where(timestamp_tokens[:-1] & timestamp_tokens[1:])[0] |
|
|
timestamp_segment_indices.add_(1) |
|
|
token_timestamps = seek_outputs[idx]["token_timestamps"] if return_token_timestamps else [] |
|
|
idx_offset = decoder_input_ids.shape[-1] |
|
|
device = seek_sequence.device |
|
|
|
|
|
|
|
|
|
|
|
if len(timestamp_segment_indices) > 0: |
|
|
|
|
|
slices = timestamp_segment_indices.tolist() |
|
|
segments = [] |
|
|
if single_timestamp_ending: |
|
|
slices.append(len(seek_sequence)) |
|
|
else: |
|
|
|
|
|
slices[-1] += 1 |
|
|
|
|
|
last_slice = 0 |
|
|
|
|
|
for i, current_slice in enumerate(slices): |
|
|
is_last_slice = i == len(slices) - 1 |
|
|
sliced_tokens = seek_sequence[last_slice:current_slice] |
|
|
start_timestamp_pos = sliced_tokens[0] - timestamp_begin |
|
|
idx_sliced_tokens = -1 if not is_last_slice or single_timestamp_ending else -2 |
|
|
end_timestamp_pos = sliced_tokens[idx_sliced_tokens] - timestamp_begin |
|
|
segments.append( |
|
|
{ |
|
|
"start": time_offset[prev_idx] |
|
|
+ start_timestamp_pos.to(torch.float32 if device.type == "mps" else torch.float64) |
|
|
* time_precision, |
|
|
"end": time_offset[prev_idx] |
|
|
+ end_timestamp_pos.to(torch.float32 if device.type == "mps" else torch.float64) |
|
|
* time_precision, |
|
|
"tokens": sliced_tokens, |
|
|
"idxs": (idx_offset + last_slice, idx_offset + current_slice), |
|
|
"result": seek_outputs[idx], |
|
|
} |
|
|
) |
|
|
if return_token_timestamps: |
|
|
segments[-1]["token_timestamps"] = ( |
|
|
token_timestamps[idx_offset + last_slice: idx_offset + current_slice] + time_offset[ |
|
|
prev_idx] |
|
|
) |
|
|
last_slice = current_slice |
|
|
|
|
|
if single_timestamp_ending: |
|
|
|
|
|
segment_offset = seek_num_frames[prev_idx] |
|
|
else: |
|
|
|
|
|
|
|
|
|
|
|
last_timestamp_pos = seek_sequence[last_slice - 2].item() - timestamp_begin |
|
|
segment_offset = last_timestamp_pos * input_stride |
|
|
else: |
|
|
|
|
|
|
|
|
timestamps = seek_sequence[timestamp_tokens.nonzero().flatten()] |
|
|
start_timestamp_pos = 0.0 |
|
|
last_timestamp_pos = seek_num_frames[prev_idx] // 2 |
|
|
skip = False |
|
|
segment_offset = seek_num_frames[prev_idx] |
|
|
|
|
|
if timestamps.numel() > 1: |
|
|
start_timestamp_pos = timestamps[-2].item() - timestamp_begin |
|
|
last_timestamp_pos = timestamps[-1].item() - timestamp_begin |
|
|
elif timestamps.numel() == 1: |
|
|
|
|
|
start_timestamp_pos = timestamps[-1].item() - timestamp_begin |
|
|
if start_timestamp_pos > 200: |
|
|
|
|
|
segment_offset = start_timestamp_pos * input_stride - 100 |
|
|
skip = True |
|
|
elif timestamps.numel() == 0 and len(seek_sequence) > 1: |
|
|
|
|
|
pass |
|
|
else: |
|
|
|
|
|
skip = True |
|
|
|
|
|
if skip: |
|
|
segments = [] |
|
|
else: |
|
|
segments = [ |
|
|
{ |
|
|
"start": time_offset[prev_idx] + start_timestamp_pos * time_precision, |
|
|
"end": time_offset[prev_idx] + last_timestamp_pos * time_precision, |
|
|
"tokens": seek_sequence, |
|
|
"result": seek_outputs[idx], |
|
|
} |
|
|
] |
|
|
if return_token_timestamps: |
|
|
segments[-1]["token_timestamps"] = token_timestamps + time_offset[prev_idx] |
|
|
segment_offset = seek_num_frames[prev_idx] |
|
|
|
|
|
if segment_offset <= 0: |
|
|
msg = f"Timestamps: {timestamps}, Segments: {segments}" |
|
|
raise ValueError(f"Segment offset: {segment_offset} <= 0. This should not happen!\n{msg}") |
|
|
|
|
|
return segments, segment_offset |
|
|
|
|
|
def generate( |
|
|
self, |
|
|
generation_config: Optional[GenerationConfig] = None, |
|
|
condition_on_prev_tokens: Optional[bool] = None, |
|
|
assistant_model: Optional["PreTrainedModel"] = None, |
|
|
**kwargs, |
|
|
): |
|
|
if condition_on_prev_tokens: |
|
|
raise NotImplementedError("Current version does not support conditioning") |
|
|
|
|
|
gen_c, _ = self._prepare_generation_config(generation_config, **kwargs) |
|
|
gen_mode = gen_c.get_generation_mode(assistant_model) |
|
|
|
|
|
if gen_mode not in [GenerationMode.GREEDY_SEARCH, GenerationMode.BEAM_SEARCH]: |
|
|
raise ValueError( |
|
|
f"Provided generation mode {gen_mode} is not supported" |
|
|
f" for WhisperForConditionalGeneration with joint CTC decoding") |
|
|
|
|
|
if "stno_mask" in kwargs: |
|
|
self.stno_mask = kwargs["stno_mask"] |
|
|
|
|
|
output = super().generate(**kwargs, return_segments=True) |
|
|
|
|
|
self.encoder_logits = None |
|
|
|
|
|
if isinstance(output, dict): |
|
|
output = self._fix_timestamps_from_segmentation(output) |
|
|
|
|
|
return output |
|
|
|
|
|
|
|
|
def generate_with_fallback( |
|
|
self, |
|
|
segment_input, |
|
|
decoder_input_ids, |
|
|
cur_bsz, |
|
|
seek, |
|
|
batch_idx_map, |
|
|
temperatures, |
|
|
generation_config, |
|
|
logits_processor, |
|
|
stopping_criteria, |
|
|
prefix_allowed_tokens_fn, |
|
|
synced_gpus, |
|
|
return_token_timestamps, |
|
|
do_condition_on_prev_tokens, |
|
|
is_shortform, |
|
|
batch_size, |
|
|
attention_mask, |
|
|
kwargs, |
|
|
): |
|
|
kwargs_local = copy.deepcopy(kwargs) |
|
|
max_frames = attention_mask.sum(-1).cpu().to(torch.long) |
|
|
kwargs_local, attention_mask = self.prepare_kwargs_for_generate(max_frames, cur_bsz, batch_idx_map, seek, kwargs_local, attention_mask) |
|
|
seek_sequences, seek_outputs, should_skip, do_condition_on_prev_tokens, model_output_type = super().generate_with_fallback( |
|
|
segment_input, |
|
|
decoder_input_ids, |
|
|
cur_bsz, |
|
|
seek, |
|
|
batch_idx_map, |
|
|
temperatures, |
|
|
generation_config, |
|
|
logits_processor, |
|
|
stopping_criteria, |
|
|
prefix_allowed_tokens_fn, |
|
|
synced_gpus, |
|
|
return_token_timestamps, |
|
|
do_condition_on_prev_tokens, |
|
|
is_shortform, |
|
|
batch_size, |
|
|
attention_mask, |
|
|
kwargs_local, |
|
|
) |
|
|
self.stno_mask_seek = None |
|
|
|
|
|
return seek_sequences, seek_outputs, should_skip, do_condition_on_prev_tokens, model_output_type |
|
|
|
|
|
|
|
|
def _sample( |
|
|
self, |
|
|
input_ids: torch.LongTensor, |
|
|
logits_processor: LogitsProcessorList, |
|
|
stopping_criteria: StoppingCriteriaList, |
|
|
generation_config: GenerationConfig, |
|
|
synced_gpus: bool = False, |
|
|
streamer: Optional["BaseStreamer"] = None, |
|
|
**model_kwargs, |
|
|
) -> Union[GenerateNonBeamOutput, torch.LongTensor]: |
|
|
r""" |
|
|
Generates sequences of token ids for models with a language modeling head using **multinomial sampling** and |
|
|
can be used for text-decoder, text-to-text, speech-to-text, and vision-to-text models. |
|
|
|
|
|
Parameters: |
|
|
input_ids (`torch.LongTensor` of shape `(batch_size, sequence_length)`): |
|
|
The sequence used as a prompt for the generation. |
|
|
logits_processor (`LogitsProcessorList`): |
|
|
An instance of [`LogitsProcessorList`]. List of instances of class derived from [`LogitsProcessor`] |
|
|
used to modify the prediction scores of the language modeling head applied at each generation step. |
|
|
stopping_criteria (`StoppingCriteriaList`): |
|
|
An instance of [`StoppingCriteriaList`]. List of instances of class derived from [`StoppingCriteria`] |
|
|
used to tell if the generation loop should stop. |
|
|
generation_config ([`~generation.GenerationConfig`]): |
|
|
The generation configuration to be used as parametrization of the decoding method. |
|
|
synced_gpus (`bool`): |
|
|
Whether to continue running the while loop until max_length (needed to avoid deadlocking with |
|
|
`FullyShardedDataParallel` and DeepSpeed ZeRO Stage 3). |
|
|
streamer (`BaseStreamer`, *optional*): |
|
|
Streamer object that will be used to stream the generated sequences. Generated tokens are passed |
|
|
through `streamer.put(token_ids)` and the streamer is responsible for any further processing. |
|
|
model_kwargs: |
|
|
Additional model specific kwargs will be forwarded to the `forward` function of the model. If model is |
|
|
an encoder-decoder model the kwargs should include `encoder_outputs`. |
|
|
|
|
|
Return: |
|
|
[`~generation.GenerateDecoderOnlyOutput`], [`~generation.GenerateEncoderDecoderOutput`] or `torch.LongTensor`: |
|
|
A `torch.LongTensor` containing the generated tokens (default behaviour) or a |
|
|
[`~generation.GenerateDecoderOnlyOutput`] if `model.config.is_encoder_decoder=False` and |
|
|
`return_dict_in_generate=True` or a [`~generation.GenerateEncoderDecoderOutput`] if |
|
|
`model.config.is_encoder_decoder=True`. |
|
|
""" |
|
|
|
|
|
pad_token_id = generation_config._pad_token_tensor |
|
|
output_attentions = generation_config.output_attentions |
|
|
output_hidden_states = generation_config.output_hidden_states |
|
|
output_scores = generation_config.output_scores |
|
|
output_logits = generation_config.output_logits |
|
|
return_dict_in_generate = generation_config.return_dict_in_generate |
|
|
has_eos_stopping_criteria = any(hasattr(criteria, "eos_token_id") for criteria in stopping_criteria) |
|
|
do_sample = generation_config.do_sample |
|
|
|
|
|
|
|
|
scores = () if (return_dict_in_generate and output_scores) else None |
|
|
raw_logits = () if (return_dict_in_generate and output_logits) else None |
|
|
decoder_attentions = () if (return_dict_in_generate and output_attentions) else None |
|
|
cross_attentions = () if (return_dict_in_generate and output_attentions) else None |
|
|
decoder_hidden_states = () if (return_dict_in_generate and output_hidden_states) else None |
|
|
|
|
|
|
|
|
if return_dict_in_generate and self.config.is_encoder_decoder: |
|
|
encoder_attentions = model_kwargs["encoder_outputs"].get("attentions") if output_attentions else None |
|
|
encoder_hidden_states = ( |
|
|
model_kwargs["encoder_outputs"].get("hidden_states") if output_hidden_states else None |
|
|
) |
|
|
|
|
|
|
|
|
batch_size, cur_len = input_ids.shape[:2] |
|
|
this_peer_finished = False |
|
|
unfinished_sequences = torch.ones(batch_size, dtype=torch.long, device=input_ids.device) |
|
|
model_kwargs = self._get_initial_cache_position(cur_len, input_ids.device, model_kwargs) |
|
|
|
|
|
model_forward = self.__call__ |
|
|
compile_forward = self._valid_auto_compile_criteria(model_kwargs, generation_config) |
|
|
if compile_forward: |
|
|
os.environ["TOKENIZERS_PARALLELISM"] = "0" |
|
|
|
|
|
if self.config._attn_implementation == "flash_attention_2": |
|
|
|
|
|
if generation_config.compile_config is not None and generation_config.compile_config.fullgraph: |
|
|
logger.warning_once( |
|
|
"When using Flash Attention 2 and a static cache, you cannot use the option `CompileConfig(fullgraph=True)` as " |
|
|
"FA2 introduces graph breaks. We overrode the option with `fullgraph=False`." |
|
|
) |
|
|
generation_config.compile_config.fullgraph = False |
|
|
model_forward = self.get_compiled_call(generation_config.compile_config) |
|
|
|
|
|
if generation_config.prefill_chunk_size is not None: |
|
|
model_kwargs = self._prefill_chunking(input_ids, generation_config, **model_kwargs) |
|
|
is_prefill = False |
|
|
else: |
|
|
is_prefill = True |
|
|
|
|
|
while self._has_unfinished_sequences(this_peer_finished, synced_gpus, device=input_ids.device): |
|
|
|
|
|
model_inputs = self.prepare_inputs_for_generation(input_ids, **model_kwargs) |
|
|
|
|
|
if is_prefill: |
|
|
outputs = self(**model_inputs, return_dict=True) |
|
|
is_prefill = False |
|
|
else: |
|
|
outputs = model_forward(**model_inputs, return_dict=True) |
|
|
|
|
|
|
|
|
model_kwargs = self._update_model_kwargs_for_generation( |
|
|
outputs, |
|
|
model_kwargs, |
|
|
is_encoder_decoder=self.config.is_encoder_decoder, |
|
|
) |
|
|
if synced_gpus and this_peer_finished: |
|
|
continue |
|
|
|
|
|
|
|
|
|
|
|
next_token_logits = outputs.logits[:, -1, :].to(copy=True, dtype=torch.float32, device=input_ids.device) |
|
|
|
|
|
|
|
|
next_token_scores = logits_processor(input_ids, next_token_logits) |
|
|
|
|
|
|
|
|
if return_dict_in_generate: |
|
|
if output_scores: |
|
|
scores += (next_token_scores,) |
|
|
if output_logits: |
|
|
raw_logits += (next_token_logits,) |
|
|
if output_attentions: |
|
|
decoder_attentions += ( |
|
|
(outputs.decoder_attentions,) if self.config.is_encoder_decoder else (outputs.attentions,) |
|
|
) |
|
|
if self.config.is_encoder_decoder: |
|
|
cross_attentions += (outputs.cross_attentions,) |
|
|
|
|
|
if output_hidden_states: |
|
|
decoder_hidden_states += ( |
|
|
(outputs.decoder_hidden_states,) |
|
|
if self.config.is_encoder_decoder |
|
|
else (outputs.hidden_states,) |
|
|
) |
|
|
|
|
|
|
|
|
if do_sample: |
|
|
probs = nn.functional.softmax(next_token_scores, dim=-1) |
|
|
|
|
|
next_tokens = torch.multinomial(probs, num_samples=1).squeeze(1) |
|
|
else: |
|
|
next_tokens = torch.argmax(next_token_scores, dim=-1) |
|
|
|
|
|
|
|
|
if has_eos_stopping_criteria: |
|
|
next_tokens = next_tokens * unfinished_sequences + pad_token_id * (1 - unfinished_sequences) |
|
|
|
|
|
"""<DiCoW CODE>""" |
|
|
|
|
|
if hasattr(self, "ctc_rescorer"): |
|
|
self.ctc_rescorer.update_state(next_tokens, torch.arange(next_tokens.shape[0])) |
|
|
"""</DiCoW CODE>""" |
|
|
|
|
|
|
|
|
input_ids = torch.cat([input_ids, next_tokens[:, None]], dim=-1) |
|
|
if streamer is not None: |
|
|
streamer.put(next_tokens.cpu()) |
|
|
|
|
|
unfinished_sequences = unfinished_sequences & ~stopping_criteria(input_ids, scores) |
|
|
this_peer_finished = unfinished_sequences.max() == 0 |
|
|
cur_len += 1 |
|
|
|
|
|
|
|
|
|
|
|
del outputs |
|
|
|
|
|
if streamer is not None: |
|
|
streamer.end() |
|
|
|
|
|
if return_dict_in_generate: |
|
|
if self.config.is_encoder_decoder: |
|
|
return GenerateEncoderDecoderOutput( |
|
|
sequences=input_ids, |
|
|
scores=scores, |
|
|
logits=raw_logits, |
|
|
encoder_attentions=encoder_attentions, |
|
|
encoder_hidden_states=encoder_hidden_states, |
|
|
decoder_attentions=decoder_attentions, |
|
|
cross_attentions=cross_attentions, |
|
|
decoder_hidden_states=decoder_hidden_states, |
|
|
past_key_values=model_kwargs.get("past_key_values"), |
|
|
) |
|
|
else: |
|
|
return GenerateDecoderOnlyOutput( |
|
|
sequences=input_ids, |
|
|
scores=scores, |
|
|
logits=raw_logits, |
|
|
attentions=decoder_attentions, |
|
|
hidden_states=decoder_hidden_states, |
|
|
past_key_values=model_kwargs.get("past_key_values"), |
|
|
) |
|
|
else: |
|
|
return input_ids |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
def _beam_search( |
|
|
self, |
|
|
input_ids: torch.LongTensor, |
|
|
logits_processor: LogitsProcessorList, |
|
|
stopping_criteria: StoppingCriteriaList, |
|
|
generation_config: GenerationConfig, |
|
|
synced_gpus: bool, |
|
|
**model_kwargs, |
|
|
) -> Union[GenerateBeamOutput, torch.LongTensor]: |
|
|
r""" |
|
|
Generates sequences of token ids for models with a language modeling head using **beam search decoding** and |
|
|
can be used for text-decoder, text-to-text, speech-to-text, and vision-to-text models. |
|
|
|
|
|
If it's the first time you're diving into Beam Search, we recommend you read the following blog post: |
|
|
https://huggingface.co/blog/how-to-generate (especially the beam search section). |
|
|
|
|
|
You can recompute the sequence scores from the individual scores using the `compute_transition_scores` function |
|
|
(https://huggingface.co/docs/transformers/main_classes/text_generation#transformers.GenerationMixin.compute_transition_scores) |
|
|
|
|
|
Parameters: |
|
|
input_ids (`torch.LongTensor` of shape `(batch_size*num_beams, sequence_length)`): |
|
|
The sequence used as a prompt for the generation. |
|
|
logits_processor (`LogitsProcessorList`): |
|
|
An instance of [`LogitsProcessorList`]. List of instances of class derived from [`LogitsProcessor`] |
|
|
used to modify the prediction scores of the language modeling head applied at each generation step. |
|
|
stopping_criteria (`StoppingCriteriaList`: |
|
|
An instance of [`StoppingCriteriaList`]. List of instances of class derived from [`StoppingCriteria`] |
|
|
used to tell if the generation loop should stop. |
|
|
generation_config ([`~generation.GenerationConfig`]): |
|
|
The generation configuration to be used as parametrization of the decoding method. |
|
|
synced_gpus (`bool`): |
|
|
Whether to continue running the while loop until max_length (needed to avoid deadlocking with |
|
|
`FullyShardedDataParallel` and DeepSpeed ZeRO Stage 3). |
|
|
model_kwargs: |
|
|
Additional model specific kwargs will be forwarded to the `forward` function of the model. If model is |
|
|
an encoder-decoder model the kwargs should include `encoder_outputs`. |
|
|
|
|
|
Return: |
|
|
[`generation.GenerateBeamDecoderOnlyOutput`], [`~generation.GenerateBeamEncoderDecoderOutput`] or |
|
|
`torch.LongTensor`: A `torch.LongTensor` containing the generated tokens (default behaviour) or a |
|
|
[`~generation.GenerateBeamDecoderOnlyOutput`] if `model.config.is_encoder_decoder=False` and |
|
|
`return_dict_in_generate=True` or a [`~generation.GenerateBeamEncoderDecoderOutput`] if |
|
|
`model.config.is_encoder_decoder=True`. |
|
|
""" |
|
|
|
|
|
|
|
|
pad_token_id = generation_config._pad_token_tensor |
|
|
eos_token_id = generation_config._eos_token_tensor |
|
|
output_attentions = generation_config.output_attentions |
|
|
output_hidden_states = generation_config.output_hidden_states |
|
|
output_scores = generation_config.output_scores |
|
|
output_logits = generation_config.output_logits |
|
|
return_dict_in_generate = generation_config.return_dict_in_generate |
|
|
do_sample = generation_config.do_sample |
|
|
early_stopping = generation_config.early_stopping |
|
|
length_penalty = generation_config.length_penalty |
|
|
max_length = generation_config.max_length |
|
|
num_beams = generation_config.num_beams |
|
|
num_return_sequences = generation_config.num_return_sequences |
|
|
|
|
|
batch_size_unflattened, cur_len = input_ids.shape[:2] |
|
|
batch_size = batch_size_unflattened // num_beams |
|
|
|
|
|
if self.__class__.__name__ == "MoshiDepthDecoder": |
|
|
vocab_size = self.config.audio_vocab_size |
|
|
elif self.__class__.__name__ == "ImageGPTForCausalImageModeling": |
|
|
vocab_size = self.get_output_embeddings().out_features |
|
|
else: |
|
|
vocab_size = self.config.get_text_config().vocab_size |
|
|
decoder_prompt_len = cur_len |
|
|
this_peer_finished = False |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
n_eos_tokens = eos_token_id.shape[0] if eos_token_id is not None else 0 |
|
|
beams_to_keep = max(2, 1 + n_eos_tokens) * num_beams |
|
|
top_num_beam_mask = torch.cat( |
|
|
(torch.ones((num_beams), dtype=torch.bool), torch.zeros((beams_to_keep - num_beams), dtype=torch.bool)), |
|
|
dim=0, |
|
|
).to(input_ids.device) |
|
|
|
|
|
model_kwargs = self._get_initial_cache_position(cur_len, input_ids.device, model_kwargs) |
|
|
|
|
|
|
|
|
|
|
|
sequential = generation_config.low_memory |
|
|
if sequential: |
|
|
raise ValueError( |
|
|
"`low_memory=True` is not supported after the beam search refactor. Please check the discussion in " |
|
|
"#35802 *after the PR got merged*, and add a comment there if your questions are not yet answered." |
|
|
) |
|
|
|
|
|
|
|
|
all_scores = () if (return_dict_in_generate and output_scores) else None |
|
|
raw_logits = () if (return_dict_in_generate and output_logits) else None |
|
|
beam_indices = () if (return_dict_in_generate and output_logits) else None |
|
|
decoder_attentions = () if (return_dict_in_generate and output_attentions) else None |
|
|
cross_attentions = () if (return_dict_in_generate and output_attentions) else None |
|
|
decoder_hidden_states = () if (return_dict_in_generate and output_hidden_states) else None |
|
|
|
|
|
|
|
|
if return_dict_in_generate and self.config.is_encoder_decoder: |
|
|
encoder_attentions = model_kwargs["encoder_outputs"].get("attentions") if output_attentions else None |
|
|
encoder_hidden_states = ( |
|
|
model_kwargs["encoder_outputs"].get("hidden_states") if output_hidden_states else None |
|
|
) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
output_fill_value = pad_token_id or eos_token_id[0] if eos_token_id is not None else -1 |
|
|
running_sequences = torch.full( |
|
|
(batch_size, num_beams, max_length), |
|
|
fill_value=output_fill_value, |
|
|
dtype=torch.int64, |
|
|
device=input_ids.device, |
|
|
) |
|
|
running_sequences[:, :, :cur_len] = self._unflatten_beam_dim(input_ids, batch_size, num_beams) |
|
|
sequences = running_sequences.detach().clone() |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
running_beam_scores = torch.zeros((batch_size, num_beams), dtype=torch.float, device=input_ids.device) |
|
|
running_beam_scores[:, 1:] = -1e9 |
|
|
beam_scores = torch.full((batch_size, num_beams), fill_value=-1e9, dtype=torch.float, device=input_ids.device) |
|
|
|
|
|
|
|
|
is_sent_finished = torch.zeros((batch_size, num_beams), dtype=torch.bool, device=input_ids.device) |
|
|
|
|
|
|
|
|
is_early_stop_heuristic_unsatisfied = torch.ones((batch_size, 1), dtype=torch.bool, device=input_ids.device) |
|
|
|
|
|
|
|
|
next_token_hits_stopping_criteria = torch.zeros( |
|
|
(batch_size, num_beams), dtype=torch.bool, device=input_ids.device |
|
|
) |
|
|
|
|
|
|
|
|
running_beam_indices = torch.full( |
|
|
(batch_size, num_beams, max_length - cur_len), fill_value=-1, dtype=torch.int32, device=input_ids.device |
|
|
) |
|
|
beam_indices = running_beam_indices.detach().clone() |
|
|
|
|
|
|
|
|
while self._has_unfinished_sequences(this_peer_finished, synced_gpus, device=input_ids.device): |
|
|
|
|
|
flat_running_sequences = self._flatten_beam_dim(running_sequences[:, :, :cur_len]) |
|
|
model_inputs = self.prepare_inputs_for_generation(flat_running_sequences, **model_kwargs) |
|
|
|
|
|
|
|
|
model_inputs.update({"output_attentions": output_attentions} if output_attentions else {}) |
|
|
model_inputs.update({"output_hidden_states": output_hidden_states} if output_hidden_states else {}) |
|
|
|
|
|
model_outputs = self(**model_inputs, return_dict=True) |
|
|
|
|
|
|
|
|
model_kwargs = self._update_model_kwargs_for_generation( |
|
|
model_outputs, |
|
|
model_kwargs, |
|
|
is_encoder_decoder=self.config.is_encoder_decoder, |
|
|
) |
|
|
if synced_gpus and this_peer_finished: |
|
|
continue |
|
|
|
|
|
|
|
|
logits = model_outputs.logits[:, -1, :].to(copy=True, dtype=torch.float32, device=input_ids.device) |
|
|
|
|
|
|
|
|
|
|
|
log_probs = nn.functional.log_softmax(logits, dim=-1) |
|
|
log_probs = logits_processor(flat_running_sequences, log_probs) |
|
|
|
|
|
|
|
|
if return_dict_in_generate: |
|
|
if output_logits: |
|
|
raw_logits += (logits.clone(),) |
|
|
if return_dict_in_generate and output_scores: |
|
|
all_scores += (log_probs.clone(),) |
|
|
|
|
|
if output_attentions: |
|
|
decoder_attentions += ( |
|
|
(model_outputs.decoder_attentions,) |
|
|
if self.config.is_encoder_decoder |
|
|
else (model_outputs.attentions,) |
|
|
) |
|
|
if self.config.is_encoder_decoder: |
|
|
cross_attentions += (model_outputs.cross_attentions,) |
|
|
|
|
|
if output_hidden_states: |
|
|
decoder_hidden_states += ( |
|
|
(model_outputs.decoder_hidden_states,) |
|
|
if self.config.is_encoder_decoder |
|
|
else (model_outputs.hidden_states,) |
|
|
) |
|
|
|
|
|
|
|
|
|
|
|
del model_outputs |
|
|
|
|
|
log_probs = self._unflatten_beam_dim(log_probs, batch_size, num_beams) |
|
|
log_probs = log_probs + running_beam_scores[:, :, None] |
|
|
log_probs = torch.reshape(log_probs, (batch_size, num_beams * vocab_size)) |
|
|
|
|
|
|
|
|
|
|
|
topk_log_probs, topk_running_sequences, topk_running_beam_indices = self._get_top_k_continuations( |
|
|
accumulated_log_probs=log_probs, |
|
|
running_sequences=running_sequences, |
|
|
running_beam_indices=running_beam_indices, |
|
|
cur_len=cur_len, |
|
|
decoder_prompt_len=decoder_prompt_len, |
|
|
do_sample=do_sample, |
|
|
beams_to_keep=beams_to_keep, |
|
|
num_beams=num_beams, |
|
|
vocab_size=vocab_size, |
|
|
batch_size=batch_size, |
|
|
) |
|
|
|
|
|
|
|
|
next_token_hits_stopping_criteria = stopping_criteria( |
|
|
self._flatten_beam_dim(topk_running_sequences[:, :, : cur_len + 1]), |
|
|
all_scores, |
|
|
) |
|
|
next_token_hits_stopping_criteria = self._unflatten_beam_dim( |
|
|
next_token_hits_stopping_criteria, batch_size, beams_to_keep |
|
|
) |
|
|
|
|
|
|
|
|
running_sequences, running_beam_scores, running_beam_indices = self._get_running_beams_for_next_iteration( |
|
|
topk_log_probs=topk_log_probs, |
|
|
topk_running_sequences=topk_running_sequences, |
|
|
topk_running_beam_indices=topk_running_beam_indices, |
|
|
next_token_hits_stopping_criteria=next_token_hits_stopping_criteria, |
|
|
num_beams=num_beams, |
|
|
) |
|
|
|
|
|
|
|
|
sequences, beam_scores, beam_indices, is_sent_finished = self._update_finished_beams( |
|
|
sequences=sequences, |
|
|
topk_running_sequences=topk_running_sequences, |
|
|
beam_scores=beam_scores, |
|
|
topk_log_probs=topk_log_probs, |
|
|
beam_indices=beam_indices, |
|
|
topk_running_beam_indices=topk_running_beam_indices, |
|
|
is_early_stop_heuristic_unsatisfied=is_early_stop_heuristic_unsatisfied, |
|
|
is_sent_finished=is_sent_finished, |
|
|
next_token_hits_stopping_criteria=next_token_hits_stopping_criteria, |
|
|
top_num_beam_mask=top_num_beam_mask, |
|
|
num_beams=num_beams, |
|
|
cur_len=cur_len, |
|
|
decoder_prompt_len=decoder_prompt_len, |
|
|
length_penalty=length_penalty, |
|
|
early_stopping=early_stopping, |
|
|
) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
beam_idx = None |
|
|
|
|
|
|
|
|
if model_kwargs.get("past_key_values", None) is not None: |
|
|
beam_idx = self._flatten_beam_dim(running_beam_indices[..., cur_len - decoder_prompt_len]) |
|
|
if hasattr(self, "_reorder_cache"): |
|
|
model_kwargs["past_key_values"] = self._reorder_cache(model_kwargs["past_key_values"], beam_idx) |
|
|
else: |
|
|
model_kwargs["past_key_values"].reorder_cache(beam_idx) |
|
|
|
|
|
if hasattr(self, "ctc_rescorer"): |
|
|
self.ctc_rescorer.update_state(running_sequences.flatten(0,1)[:, cur_len], beam_idx) |
|
|
|
|
|
cur_len = cur_len + 1 |
|
|
is_early_stop_heuristic_unsatisfied = self._check_early_stop_heuristic( |
|
|
is_early_stop_heuristic_unsatisfied=is_early_stop_heuristic_unsatisfied, |
|
|
running_beam_scores=running_beam_scores, |
|
|
beam_scores=beam_scores, |
|
|
is_sent_finished=is_sent_finished, |
|
|
cur_len=cur_len, |
|
|
max_length=max_length, |
|
|
decoder_prompt_len=decoder_prompt_len, |
|
|
early_stopping=early_stopping, |
|
|
length_penalty=length_penalty, |
|
|
) |
|
|
this_peer_finished = not self._beam_search_has_unfinished_sequences( |
|
|
is_early_stop_heuristic_unsatisfied, |
|
|
is_sent_finished, |
|
|
next_token_hits_stopping_criteria, |
|
|
early_stopping, |
|
|
) |
|
|
|
|
|
|
|
|
|
|
|
sequences = self._flatten_beam_dim(sequences[:, :num_return_sequences, :]) |
|
|
beam_scores = self._flatten_beam_dim(beam_scores[:, :num_return_sequences]) |
|
|
beam_indices = self._flatten_beam_dim(beam_indices[:, :num_return_sequences, :]) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
max_generated_length = ((beam_indices + 1).bool()).sum(dim=1).max() |
|
|
output_length = decoder_prompt_len + max_generated_length |
|
|
sequences = sequences[:, :output_length] |
|
|
beam_indices = beam_indices[:, :max_generated_length] |
|
|
|
|
|
if return_dict_in_generate: |
|
|
if not output_scores: |
|
|
beam_scores = None |
|
|
|
|
|
if self.config.is_encoder_decoder: |
|
|
return GenerateBeamEncoderDecoderOutput( |
|
|
sequences=sequences, |
|
|
sequences_scores=beam_scores, |
|
|
scores=all_scores, |
|
|
logits=raw_logits, |
|
|
beam_indices=beam_indices, |
|
|
encoder_attentions=encoder_attentions, |
|
|
encoder_hidden_states=encoder_hidden_states, |
|
|
decoder_attentions=decoder_attentions, |
|
|
cross_attentions=cross_attentions, |
|
|
decoder_hidden_states=decoder_hidden_states, |
|
|
past_key_values=model_kwargs.get("past_key_values"), |
|
|
) |
|
|
else: |
|
|
return GenerateBeamDecoderOnlyOutput( |
|
|
sequences=sequences, |
|
|
sequences_scores=beam_scores, |
|
|
scores=all_scores, |
|
|
logits=raw_logits, |
|
|
beam_indices=beam_indices, |
|
|
attentions=decoder_attentions, |
|
|
hidden_states=decoder_hidden_states, |
|
|
past_key_values=model_kwargs.get("past_key_values"), |
|
|
) |
|
|
else: |
|
|
return sequences |