| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| from __future__ import annotations |
|
|
| import warnings |
| from dataclasses import dataclass |
| from typing import List, Optional, Tuple |
|
|
| import tensorflow as tf |
|
|
| from .utils import ModelOutput |
|
|
|
|
| @dataclass |
| class TFBaseModelOutput(ModelOutput): |
| """ |
| Base class for model's outputs, with potential hidden states and attentions. |
| |
| Args: |
| last_hidden_state (`tf.Tensor` of shape `(batch_size, sequence_length, hidden_size)`): |
| Sequence of hidden-states at the output of the last layer of the model. |
| hidden_states (`tuple(tf.FloatTensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings + one for the output of each layer) of shape |
| `(batch_size, sequence_length, hidden_size)`. |
| |
| Hidden-states of the model at the output of each layer plus the initial embedding outputs. |
| attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights after the attention softmax, used to compute the weighted average in the self-attention |
| heads. |
| """ |
|
|
| last_hidden_state: tf.Tensor = None |
| hidden_states: Tuple[tf.Tensor] | None = None |
| attentions: Tuple[tf.Tensor] | None = None |
|
|
|
|
| @dataclass |
| class TFBaseModelOutputWithNoAttention(ModelOutput): |
| """ |
| Base class for model's outputs, with potential hidden states. |
| |
| Args: |
| last_hidden_state (`tf.Tensor` shape `(batch_size, num_channels, height, width)`): |
| Sequence of hidden-states at the output of the last layer of the model. |
| hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings, if the model has an embedding layer, + one for |
| the output of each layer) of shape `(batch_size, num_channels, height, width)`. |
| |
| Hidden-states of the model at the output of each layer plus the optional initial embedding outputs. |
| """ |
|
|
| last_hidden_state: tf.Tensor = None |
| hidden_states: Optional[Tuple[tf.Tensor, ...]] = None |
|
|
|
|
| @dataclass |
| class TFBaseModelOutputWithPooling(ModelOutput): |
| """ |
| Base class for model's outputs that also contains a pooling of the last hidden states. |
| |
| Args: |
| last_hidden_state (`tf.Tensor` of shape `(batch_size, sequence_length, hidden_size)`): |
| Sequence of hidden-states at the output of the last layer of the model. |
| pooler_output (`tf.Tensor` of shape `(batch_size, hidden_size)`): |
| Last layer hidden-state of the first token of the sequence (classification token) further processed by a |
| Linear layer and a Tanh activation function. The Linear layer weights are trained from the next sentence |
| prediction (classification) objective during pretraining. |
| |
| This output is usually *not* a good summary of the semantic content of the input, you're often better with |
| averaging or pooling the sequence of hidden-states for the whole input sequence. |
| hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings + one for the output of each layer) of shape |
| `(batch_size, sequence_length, hidden_size)`. |
| |
| Hidden-states of the model at the output of each layer plus the initial embedding outputs. |
| attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights after the attention softmax, used to compute the weighted average in the self-attention |
| heads. |
| """ |
|
|
| last_hidden_state: tf.Tensor = None |
| pooler_output: tf.Tensor = None |
| hidden_states: Tuple[tf.Tensor] | None = None |
| attentions: Tuple[tf.Tensor] | None = None |
|
|
|
|
| @dataclass |
| class TFBaseModelOutputWithPoolingAndNoAttention(ModelOutput): |
| """ |
| Base class for model's outputs that also contains a pooling of the last hidden states. |
| |
| Args: |
| last_hidden_state (`tf.Tensor` of shape `(batch_size, num_channels, height, width)`): |
| Sequence of hidden-states at the output of the last layer of the model. |
| pooler_output (`tf.Tensor` of shape `(batch_size, hidden_size)`): |
| Last layer hidden-state after a pooling operation on the spatial dimensions. |
| hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings, if the model has an embedding layer, + one for |
| the output of each layer) of shape `(batch_size, num_channels, height, width)`. |
| |
| Hidden-states of the model at the output of each layer plus the optional initial embedding outputs. |
| """ |
|
|
| last_hidden_state: tf.Tensor = None |
| pooler_output: tf.Tensor = None |
| hidden_states: Optional[Tuple[tf.Tensor, ...]] = None |
|
|
|
|
| @dataclass |
| class TFBaseModelOutputWithPoolingAndCrossAttentions(ModelOutput): |
| """ |
| Base class for model's outputs that also contains a pooling of the last hidden states. |
| |
| Args: |
| last_hidden_state (`tf.Tensor` of shape `(batch_size, sequence_length, hidden_size)`): |
| Sequence of hidden-states at the output of the last layer of the model. |
| pooler_output (`tf.Tensor` of shape `(batch_size, hidden_size)`): |
| Last layer hidden-state of the first token of the sequence (classification token) further processed by a |
| Linear layer and a Tanh activation function. The Linear layer weights are trained from the next sentence |
| prediction (classification) objective during pretraining. |
| |
| This output is usually *not* a good summary of the semantic content of the input, you're often better with |
| averaging or pooling the sequence of hidden-states for the whole input sequence. |
| past_key_values (`List[tf.Tensor]`, *optional*, returned when `use_cache=True` is passed or when `config.use_cache=True`): |
| List of `tf.Tensor` of length `config.n_layers`, with each tensor of shape `(2, batch_size, num_heads, |
| sequence_length, embed_size_per_head)`). |
| |
| Contains pre-computed hidden-states (key and values in the attention blocks) that can be used (see |
| `past_key_values` input) to speed up sequential decoding. |
| hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings + one for the output of each layer) of shape |
| `(batch_size, sequence_length, hidden_size)`. |
| |
| Hidden-states of the model at the output of each layer plus the initial embedding outputs. |
| attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights after the attention softmax, used to compute the weighted average in the self-attention |
| heads. |
| cross_attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights of the decoder's cross-attention layer, after the attention softmax, used to compute the |
| weighted average in the cross-attention heads. |
| """ |
|
|
| last_hidden_state: tf.Tensor = None |
| pooler_output: tf.Tensor = None |
| past_key_values: List[tf.Tensor] | None = None |
| hidden_states: Tuple[tf.Tensor] | None = None |
| attentions: Tuple[tf.Tensor] | None = None |
| cross_attentions: Tuple[tf.Tensor] | None = None |
|
|
|
|
| @dataclass |
| class TFBaseModelOutputWithPast(ModelOutput): |
| """ |
| Base class for model's outputs that may also contain a past key/values (to speed up sequential decoding). |
| |
| Args: |
| last_hidden_state (`tf.Tensor` of shape `(batch_size, sequence_length, hidden_size)`): |
| Sequence of hidden-states at the output of the last layer of the model. |
| |
| If `past_key_values` is used only the last hidden-state of the sequences of shape `(batch_size, 1, |
| hidden_size)` is output. |
| past_key_values (`List[tf.Tensor]`, *optional*, returned when `use_cache=True` is passed or when `config.use_cache=True`): |
| List of `tf.Tensor` of length `config.n_layers`, with each tensor of shape `(2, batch_size, num_heads, |
| sequence_length, embed_size_per_head)`). |
| |
| Contains pre-computed hidden-states (key and values in the attention blocks) that can be used (see |
| `past_key_values` input) to speed up sequential decoding. |
| hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings + one for the output of each layer) of shape |
| `(batch_size, sequence_length, hidden_size)`. |
| |
| Hidden-states of the model at the output of each layer plus the initial embedding outputs. |
| attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights after the attention softmax, used to compute the weighted average in the self-attention |
| heads. |
| """ |
|
|
| last_hidden_state: tf.Tensor = None |
| past_key_values: List[tf.Tensor] | None = None |
| hidden_states: Tuple[tf.Tensor] | None = None |
| attentions: Tuple[tf.Tensor] | None = None |
|
|
|
|
| @dataclass |
| class TFBaseModelOutputWithCrossAttentions(ModelOutput): |
| """ |
| Base class for model's outputs, with potential hidden states and attentions. |
| |
| Args: |
| last_hidden_state (`tf.Tensor` of shape `(batch_size, sequence_length, hidden_size)`): |
| Sequence of hidden-states at the output of the last layer of the model. |
| hidden_states (`tuple(tf.FloatTensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings + one for the output of each layer) of shape |
| `(batch_size, sequence_length, hidden_size)`. |
| |
| Hidden-states of the model at the output of each layer plus the initial embedding outputs. |
| attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights after the attention softmax, used to compute the weighted average in the self-attention |
| heads. |
| cross_attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights of the decoder's cross-attention layer, after the attention softmax, used to compute the |
| weighted average in the cross-attention heads. |
| """ |
|
|
| last_hidden_state: tf.Tensor = None |
| hidden_states: Tuple[tf.Tensor] | None = None |
| attentions: Tuple[tf.Tensor] | None = None |
| cross_attentions: Tuple[tf.Tensor] | None = None |
|
|
|
|
| @dataclass |
| class TFBaseModelOutputWithPastAndCrossAttentions(ModelOutput): |
| """ |
| Base class for model's outputs that may also contain a past key/values (to speed up sequential decoding). |
| |
| Args: |
| last_hidden_state (`tf.Tensor` of shape `(batch_size, sequence_length, hidden_size)`): |
| Sequence of hidden-states at the output of the last layer of the model. |
| |
| If `past_key_values` is used only the last hidden-state of the sequences of shape `(batch_size, 1, |
| hidden_size)` is output. |
| past_key_values (`List[tf.Tensor]`, *optional*, returned when `use_cache=True` is passed or when `config.use_cache=True`): |
| List of `tf.Tensor` of length `config.n_layers`, with each tensor of shape `(2, batch_size, num_heads, |
| sequence_length, embed_size_per_head)`). |
| |
| Contains pre-computed hidden-states (key and values in the attention blocks) that can be used (see |
| `past_key_values` input) to speed up sequential decoding. |
| hidden_states (`tuple(tf.FloatTensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings + one for the output of each layer) of shape |
| `(batch_size, sequence_length, hidden_size)`. |
| |
| Hidden-states of the model at the output of each layer plus the initial embedding outputs. |
| attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights after the attention softmax, used to compute the weighted average in the self-attention |
| heads. |
| cross_attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights of the decoder's cross-attention layer, after the attention softmax, used to compute the |
| weighted average in the cross-attention heads. |
| """ |
|
|
| last_hidden_state: tf.Tensor = None |
| past_key_values: List[tf.Tensor] | None = None |
| hidden_states: Tuple[tf.Tensor] | None = None |
| attentions: Tuple[tf.Tensor] | None = None |
| cross_attentions: Tuple[tf.Tensor] | None = None |
|
|
|
|
| @dataclass |
| class TFSeq2SeqModelOutput(ModelOutput): |
| """ |
| Base class for model encoder's outputs that also contains : pre-computed hidden states that can speed up sequential |
| decoding. |
| |
| Args: |
| last_hidden_state (`tf.Tensor` of shape `(batch_size, sequence_length, hidden_size)`): |
| Sequence of hidden-states at the output of the last layer of the decoder of the model. |
| |
| If `past_key_values` is used only the last hidden-state of the sequences of shape `(batch_size, 1, |
| hidden_size)` is output. |
| past_key_values (`List[tf.Tensor]`, *optional*, returned when `use_cache=True` is passed or when `config.use_cache=True`): |
| List of `tf.Tensor` of length `config.n_layers`, with each tensor of shape `(2, batch_size, num_heads, |
| sequence_length, embed_size_per_head)`). |
| |
| Contains pre-computed hidden-states (key and values in the attention blocks) of the decoder that can be |
| used (see `past_key_values` input) to speed up sequential decoding. |
| decoder_hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings + one for the output of each layer) of shape |
| `(batch_size, sequence_length, hidden_size)`. |
| |
| Hidden-states of the decoder at the output of each layer plus the initial embedding outputs. |
| decoder_attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights of the decoder, after the attention softmax, used to compute the weighted average in the |
| self-attention heads. |
| cross_attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights of the decoder's cross-attention layer, after the attention softmax, used to compute the |
| weighted average in the cross-attention heads. |
| encoder_last_hidden_state (`tf.Tensor` of shape `(batch_size, sequence_length, hidden_size)`, *optional*): |
| Sequence of hidden-states at the output of the last layer of the encoder of the model. |
| encoder_hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings + one for the output of each layer) of shape |
| `(batch_size, sequence_length, hidden_size)`. |
| |
| Hidden-states of the encoder at the output of each layer plus the initial embedding outputs. |
| encoder_attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights of the encoder, after the attention softmax, used to compute the weighted average in the |
| self-attention heads. |
| """ |
|
|
| last_hidden_state: tf.Tensor = None |
| past_key_values: List[tf.Tensor] | None = None |
| decoder_hidden_states: Tuple[tf.Tensor] | None = None |
| decoder_attentions: Tuple[tf.Tensor] | None = None |
| cross_attentions: Tuple[tf.Tensor] | None = None |
| encoder_last_hidden_state: tf.Tensor | None = None |
| encoder_hidden_states: Tuple[tf.Tensor] | None = None |
| encoder_attentions: Tuple[tf.Tensor] | None = None |
|
|
|
|
| @dataclass |
| class TFCausalLMOutput(ModelOutput): |
| """ |
| Base class for causal language model (or autoregressive) outputs. |
| |
| Args: |
| loss (`tf.Tensor` of shape `(n,)`, *optional*, where n is the number of non-masked labels, returned when `labels` is provided): |
| Language modeling loss (for next-token prediction). |
| logits (`tf.Tensor` of shape `(batch_size, sequence_length, config.vocab_size)`): |
| Prediction scores of the language modeling head (scores for each vocabulary token before SoftMax). |
| hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings + one for the output of each layer) of shape |
| `(batch_size, sequence_length, hidden_size)`. |
| |
| Hidden-states of the model at the output of each layer plus the initial embedding outputs. |
| attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights after the attention softmax, used to compute the weighted average in the self-attention |
| heads. |
| """ |
|
|
| loss: tf.Tensor | None = None |
| logits: tf.Tensor = None |
| hidden_states: Tuple[tf.Tensor] | None = None |
| attentions: Tuple[tf.Tensor] | None = None |
|
|
|
|
| @dataclass |
| class TFCausalLMOutputWithPast(ModelOutput): |
| """ |
| Base class for causal language model (or autoregressive) outputs. |
| |
| Args: |
| loss (`tf.Tensor` of shape `(n,)`, *optional*, where n is the number of non-masked labels, returned when `labels` is provided): |
| Language modeling loss (for next-token prediction). |
| logits (`tf.Tensor` of shape `(batch_size, sequence_length, config.vocab_size)`): |
| Prediction scores of the language modeling head (scores for each vocabulary token before SoftMax). |
| past_key_values (`List[tf.Tensor]`, *optional*, returned when `use_cache=True` is passed or when `config.use_cache=True`): |
| List of `tf.Tensor` of length `config.n_layers`, with each tensor of shape `(2, batch_size, num_heads, |
| sequence_length, embed_size_per_head)`). |
| |
| Contains pre-computed hidden-states (key and values in the attention blocks) that can be used (see |
| `past_key_values` input) to speed up sequential decoding. |
| hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings + one for the output of each layer) of shape |
| `(batch_size, sequence_length, hidden_size)`. |
| |
| Hidden-states of the model at the output of each layer plus the initial embedding outputs. |
| attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights after the attention softmax, used to compute the weighted average in the self-attention |
| heads. |
| """ |
|
|
| loss: tf.Tensor | None = None |
| logits: tf.Tensor = None |
| past_key_values: List[tf.Tensor] | None = None |
| hidden_states: Tuple[tf.Tensor] | None = None |
| attentions: Tuple[tf.Tensor] | None = None |
|
|
|
|
| @dataclass |
| class TFCausalLMOutputWithCrossAttentions(ModelOutput): |
| """ |
| Base class for causal language model (or autoregressive) outputs. |
| |
| Args: |
| loss (`tf.Tensor` of shape `(n,)`, *optional*, where n is the number of non-masked labels, returned when `labels` is provided): |
| Language modeling loss (for next-token prediction). |
| logits (`tf.Tensor` of shape `(batch_size, sequence_length, config.vocab_size)`): |
| Prediction scores of the language modeling head (scores for each vocabulary token before SoftMax). |
| hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings + one for the output of each layer) of shape |
| `(batch_size, sequence_length, hidden_size)`. |
| |
| Hidden-states of the model at the output of each layer plus the initial embedding outputs. |
| attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights after the attention softmax, used to compute the weighted average in the self-attention |
| heads. |
| cross_attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights of the decoder's cross-attention layer, after the attention softmax, used to compute the |
| weighted average in the cross-attention heads. |
| past_key_values (`List[tf.Tensor]`, *optional*, returned when `use_cache=True` is passed or when `config.use_cache=True`): |
| List of `tf.Tensor` of length `config.n_layers`, with each tensor of shape `(2, batch_size, num_heads, |
| sequence_length, embed_size_per_head)`). |
| |
| Contains pre-computed hidden-states (key and values in the attention blocks) that can be used (see |
| `past_key_values` input) to speed up sequential decoding. |
| """ |
|
|
| loss: tf.Tensor | None = None |
| logits: tf.Tensor = None |
| past_key_values: List[tf.Tensor] | None = None |
| hidden_states: Tuple[tf.Tensor] | None = None |
| attentions: Tuple[tf.Tensor] | None = None |
| cross_attentions: Tuple[tf.Tensor] | None = None |
|
|
|
|
| @dataclass |
| class TFMaskedLMOutput(ModelOutput): |
| """ |
| Base class for masked language models outputs. |
| |
| Args: |
| loss (`tf.Tensor` of shape `(n,)`, *optional*, where n is the number of non-masked labels, returned when `labels` is provided): |
| Masked language modeling (MLM) loss. |
| logits (`tf.Tensor` of shape `(batch_size, sequence_length, config.vocab_size)`): |
| Prediction scores of the language modeling head (scores for each vocabulary token before SoftMax). |
| hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings + one for the output of each layer) of shape |
| `(batch_size, sequence_length, hidden_size)`. |
| |
| Hidden-states of the model at the output of each layer plus the initial embedding outputs. |
| attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights after the attention softmax, used to compute the weighted average in the self-attention |
| heads. |
| """ |
|
|
| loss: tf.Tensor | None = None |
| logits: tf.Tensor = None |
| hidden_states: Tuple[tf.Tensor] | None = None |
| attentions: Tuple[tf.Tensor] | None = None |
|
|
|
|
| @dataclass |
| class TFSeq2SeqLMOutput(ModelOutput): |
| """ |
| Base class for sequence-to-sequence language models outputs. |
| |
| Args: |
| loss (`tf.Tensor` of shape `(n,)`, *optional*, where n is the number of non-masked labels, returned when `labels` is provided): |
| Language modeling loss. |
| logits (`tf.Tensor` of shape `(batch_size, sequence_length, config.vocab_size)`): |
| Prediction scores of the language modeling head (scores for each vocabulary token before SoftMax). |
| past_key_values (`List[tf.Tensor]`, *optional*, returned when `use_cache=True` is passed or when `config.use_cache=True`): |
| List of `tf.Tensor` of length `config.n_layers`, with each tensor of shape `(2, batch_size, num_heads, |
| sequence_length, embed_size_per_head)`). |
| |
| Contains pre-computed hidden-states (key and values in the attention blocks) of the decoder that can be |
| used (see `past_key_values` input) to speed up sequential decoding. |
| decoder_hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings + one for the output of each layer) of shape |
| `(batch_size, sequence_length, hidden_size)`. |
| |
| Hidden-states of the decoder at the output of each layer plus the initial embedding outputs. |
| decoder_attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights of the decoder, after the attention softmax, used to compute the weighted average in the |
| self-attention heads. |
| cross_attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights of the decoder's cross-attention layer, after the attention softmax, used to compute the |
| weighted average in the cross-attention heads. |
| encoder_last_hidden_state (`tf.Tensor` of shape `(batch_size, sequence_length, hidden_size)`, *optional*): |
| Sequence of hidden-states at the output of the last layer of the encoder of the model. |
| encoder_hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings + one for the output of each layer) of shape |
| `(batch_size, sequence_length, hidden_size)`. |
| |
| Hidden-states of the encoder at the output of each layer plus the initial embedding outputs. |
| encoder_attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights of the encoder, after the attention softmax, used to compute the weighted average in the |
| self-attention heads. |
| """ |
|
|
| loss: tf.Tensor | None = None |
| logits: tf.Tensor = None |
| past_key_values: List[tf.Tensor] | None = None |
| decoder_hidden_states: Tuple[tf.Tensor] | None = None |
| decoder_attentions: Tuple[tf.Tensor] | None = None |
| cross_attentions: Tuple[tf.Tensor] | None = None |
| encoder_last_hidden_state: tf.Tensor | None = None |
| encoder_hidden_states: Tuple[tf.Tensor] | None = None |
| encoder_attentions: Tuple[tf.Tensor] | None = None |
|
|
|
|
| @dataclass |
| class TFNextSentencePredictorOutput(ModelOutput): |
| """ |
| Base class for outputs of models predicting if two sentences are consecutive or not. |
| |
| Args: |
| loss (`tf.Tensor` of shape `(n,)`, *optional*, where n is the number of non-masked labels, returned when `next_sentence_label` is provided): |
| Next sentence prediction loss. |
| logits (`tf.Tensor` of shape `(batch_size, 2)`): |
| Prediction scores of the next sequence prediction (classification) head (scores of True/False continuation |
| before SoftMax). |
| hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings + one for the output of each layer) of shape |
| `(batch_size, sequence_length, hidden_size)`. |
| |
| Hidden-states of the model at the output of each layer plus the initial embedding outputs. |
| attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights after the attention softmax, used to compute the weighted average in the self-attention |
| heads. |
| """ |
|
|
| loss: tf.Tensor | None = None |
| logits: tf.Tensor = None |
| hidden_states: Tuple[tf.Tensor] | None = None |
| attentions: Tuple[tf.Tensor] | None = None |
|
|
|
|
| @dataclass |
| class TFSequenceClassifierOutput(ModelOutput): |
| """ |
| Base class for outputs of sentence classification models. |
| |
| Args: |
| loss (`tf.Tensor` of shape `(batch_size, )`, *optional*, returned when `labels` is provided): |
| Classification (or regression if config.num_labels==1) loss. |
| logits (`tf.Tensor` of shape `(batch_size, config.num_labels)`): |
| Classification (or regression if config.num_labels==1) scores (before SoftMax). |
| hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings + one for the output of each layer) of shape |
| `(batch_size, sequence_length, hidden_size)`. |
| |
| Hidden-states of the model at the output of each layer plus the initial embedding outputs. |
| attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights after the attention softmax, used to compute the weighted average in the self-attention |
| heads. |
| """ |
|
|
| loss: tf.Tensor | None = None |
| logits: tf.Tensor = None |
| hidden_states: Tuple[tf.Tensor] | None = None |
| attentions: Tuple[tf.Tensor] | None = None |
|
|
|
|
| @dataclass |
| class TFSeq2SeqSequenceClassifierOutput(ModelOutput): |
| """ |
| Base class for outputs of sequence-to-sequence sentence classification models. |
| |
| Args: |
| loss (`tf.Tensor` of shape `(1,)`, *optional*, returned when `label` is provided): |
| Classification (or regression if config.num_labels==1) loss. |
| logits (`tf.Tensor` of shape `(batch_size, config.num_labels)`): |
| Classification (or regression if config.num_labels==1) scores (before SoftMax). |
| past_key_values (`List[tf.Tensor]`, *optional*, returned when `use_cache=True` is passed or when `config.use_cache=True`): |
| List of `tf.Tensor` of length `config.n_layers`, with each tensor of shape `(2, batch_size, num_heads, |
| sequence_length, embed_size_per_head)`). |
| |
| Contains pre-computed hidden-states (key and values in the attention blocks) of the decoder that can be |
| used (see `past_key_values` input) to speed up sequential decoding. |
| decoder_hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings + one for the output of each layer) of shape |
| `(batch_size, sequence_length, hidden_size)`. |
| |
| Hidden-states of the decoder at the output of each layer plus the initial embedding outputs. |
| decoder_attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights of the decoder, after the attention softmax, used to compute the weighted average in the |
| self-attention heads. |
| cross_attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)` |
| encoder_last_hidden_state (`tf.Tensor` of shape `(batch_size, sequence_length, hidden_size)`, *optional*): |
| Sequence of hidden-states at the output of the last layer of the encoder of the model. |
| encoder_hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings + one for the output of each layer) of shape |
| `(batch_size, sequence_length, hidden_size)`. |
| |
| Hidden-states of the encoder at the output of each layer plus the initial embedding outputs. |
| encoder_attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights of the encoder, after the attention softmax, used to compute the weighted average in the |
| self-attention heads. |
| """ |
|
|
| loss: tf.Tensor | None = None |
| logits: tf.Tensor = None |
| past_key_values: List[tf.Tensor] | None = None |
| decoder_hidden_states: Tuple[tf.Tensor] | None = None |
| decoder_attentions: Tuple[tf.Tensor] | None = None |
| cross_attentions: Tuple[tf.Tensor] | None = None |
| encoder_last_hidden_state: tf.Tensor | None = None |
| encoder_hidden_states: Tuple[tf.Tensor] | None = None |
| encoder_attentions: Tuple[tf.Tensor] | None = None |
|
|
|
|
| @dataclass |
| class TFSemanticSegmenterOutput(ModelOutput): |
| """ |
| Base class for outputs of semantic segmentation models. |
| |
| Args: |
| loss (`tf.Tensor` of shape `(1,)`, *optional*, returned when `labels` is provided): |
| Classification (or regression if config.num_labels==1) loss. |
| logits (`tf.Tensor` of shape `(batch_size, config.num_labels, logits_height, logits_width)`): |
| Classification scores for each pixel. |
| |
| <Tip warning={true}> |
| |
| The logits returned do not necessarily have the same size as the `pixel_values` passed as inputs. This is |
| to avoid doing two interpolations and lose some quality when a user needs to resize the logits to the |
| original image size as post-processing. You should always check your logits shape and resize as needed. |
| |
| </Tip> |
| |
| hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings, if the model has an embedding layer, + one for |
| the output of each layer) of shape `(batch_size, patch_size, hidden_size)`. |
| |
| Hidden-states of the model at the output of each layer plus the optional initial embedding outputs. |
| attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, patch_size, sequence_length)`. |
| |
| Attentions weights after the attention softmax, used to compute the weighted average in the self-attention |
| heads. |
| """ |
|
|
| loss: tf.Tensor | None = None |
| logits: tf.Tensor = None |
| hidden_states: Tuple[tf.Tensor] | None = None |
| attentions: Tuple[tf.Tensor] | None = None |
|
|
|
|
| @dataclass |
| class TFSemanticSegmenterOutputWithNoAttention(ModelOutput): |
| """ |
| Base class for outputs of semantic segmentation models that do not output attention scores. |
| |
| Args: |
| loss (`tf.Tensor` of shape `(1,)`, *optional*, returned when `labels` is provided): |
| Classification (or regression if config.num_labels==1) loss. |
| logits (`tf.Tensor` of shape `(batch_size, config.num_labels, logits_height, logits_width)`): |
| Classification scores for each pixel. |
| |
| <Tip warning={true}> |
| |
| The logits returned do not necessarily have the same size as the `pixel_values` passed as inputs. This is |
| to avoid doing two interpolations and lose some quality when a user needs to resize the logits to the |
| original image size as post-processing. You should always check your logits shape and resize as needed. |
| |
| </Tip> |
| |
| hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings, if the model has an embedding layer, + one for |
| the output of each layer) of shape `(batch_size, patch_size, hidden_size)`. |
| |
| Hidden-states of the model at the output of each layer plus the optional initial embedding outputs. |
| """ |
|
|
| loss: tf.Tensor | None = None |
| logits: tf.Tensor = None |
| hidden_states: Tuple[tf.Tensor] | None = None |
|
|
|
|
| @dataclass |
| class TFImageClassifierOutput(ModelOutput): |
| """ |
| Base class for outputs of image classification models. |
| |
| Args: |
| loss (`tf.Tensor` of shape `(1,)`, *optional*, returned when `labels` is provided): |
| Classification (or regression if config.num_labels==1) loss. |
| logits (`tf.Tensor` of shape `(batch_size, config.num_labels)`): |
| Classification (or regression if config.num_labels==1) scores (before SoftMax). |
| hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings, if the model has an embedding layer, + one for |
| the output of each stage) of shape `(batch_size, sequence_length, hidden_size)`. Hidden-states (also called |
| feature maps) of the model at the output of each stage. |
| attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, patch_size, sequence_length)`. |
| |
| Attentions weights after the attention softmax, used to compute the weighted average in the self-attention |
| heads. |
| """ |
|
|
| loss: tf.Tensor | None = None |
| logits: tf.Tensor = None |
| hidden_states: Tuple[tf.Tensor] | None = None |
| attentions: Tuple[tf.Tensor] | None = None |
|
|
|
|
| @dataclass |
| class TFMultipleChoiceModelOutput(ModelOutput): |
| """ |
| Base class for outputs of multiple choice models. |
| |
| Args: |
| loss (`tf.Tensor` of shape *(batch_size, )*, *optional*, returned when `labels` is provided): |
| Classification loss. |
| logits (`tf.Tensor` of shape `(batch_size, num_choices)`): |
| *num_choices* is the second dimension of the input tensors. (see *input_ids* above). |
| |
| Classification scores (before SoftMax). |
| hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings + one for the output of each layer) of shape |
| `(batch_size, sequence_length, hidden_size)`. |
| |
| Hidden-states of the model at the output of each layer plus the initial embedding outputs. |
| attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights after the attention softmax, used to compute the weighted average in the self-attention |
| heads. |
| """ |
|
|
| loss: tf.Tensor | None = None |
| logits: tf.Tensor = None |
| hidden_states: Tuple[tf.Tensor] | None = None |
| attentions: Tuple[tf.Tensor] | None = None |
|
|
|
|
| @dataclass |
| class TFTokenClassifierOutput(ModelOutput): |
| """ |
| Base class for outputs of token classification models. |
| |
| Args: |
| loss (`tf.Tensor` of shape `(n,)`, *optional*, where n is the number of unmasked labels, returned when `labels` is provided) : |
| Classification loss. |
| logits (`tf.Tensor` of shape `(batch_size, sequence_length, config.num_labels)`): |
| Classification scores (before SoftMax). |
| hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings + one for the output of each layer) of shape |
| `(batch_size, sequence_length, hidden_size)`. |
| |
| Hidden-states of the model at the output of each layer plus the initial embedding outputs. |
| attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights after the attention softmax, used to compute the weighted average in the self-attention |
| heads. |
| """ |
|
|
| loss: tf.Tensor | None = None |
| logits: tf.Tensor = None |
| hidden_states: Tuple[tf.Tensor] | None = None |
| attentions: Tuple[tf.Tensor] | None = None |
|
|
|
|
| @dataclass |
| class TFQuestionAnsweringModelOutput(ModelOutput): |
| """ |
| Base class for outputs of question answering models. |
| |
| Args: |
| loss (`tf.Tensor` of shape `(batch_size, )`, *optional*, returned when `start_positions` and `end_positions` are provided): |
| Total span extraction loss is the sum of a Cross-Entropy for the start and end positions. |
| start_logits (`tf.Tensor` of shape `(batch_size, sequence_length)`): |
| Span-start scores (before SoftMax). |
| end_logits (`tf.Tensor` of shape `(batch_size, sequence_length)`): |
| Span-end scores (before SoftMax). |
| hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings + one for the output of each layer) of shape |
| `(batch_size, sequence_length, hidden_size)`. |
| |
| Hidden-states of the model at the output of each layer plus the initial embedding outputs. |
| attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights after the attention softmax, used to compute the weighted average in the self-attention |
| heads. |
| """ |
|
|
| loss: tf.Tensor | None = None |
| start_logits: tf.Tensor = None |
| end_logits: tf.Tensor = None |
| hidden_states: Tuple[tf.Tensor] | None = None |
| attentions: Tuple[tf.Tensor] | None = None |
|
|
|
|
| @dataclass |
| class TFSeq2SeqQuestionAnsweringModelOutput(ModelOutput): |
| """ |
| Base class for outputs of sequence-to-sequence question answering models. |
| |
| Args: |
| loss (`tf.Tensor` of shape `(1,)`, *optional*, returned when `labels` is provided): |
| Total span extraction loss is the sum of a Cross-Entropy for the start and end positions. |
| start_logits (`tf.Tensor` of shape `(batch_size, sequence_length)`): |
| Span-start scores (before SoftMax). |
| end_logits (`tf.Tensor` of shape `(batch_size, sequence_length)`): |
| Span-end scores (before SoftMax). |
| past_key_values (`List[tf.Tensor]`, *optional*, returned when `use_cache=True` is passed or when `config.use_cache=True`): |
| List of `tf.Tensor` of length `config.n_layers`, with each tensor of shape `(2, batch_size, num_heads, |
| sequence_length, embed_size_per_head)`). |
| |
| Contains pre-computed hidden-states (key and values in the attention blocks) of the decoder that can be |
| used (see `past_key_values` input) to speed up sequential decoding. |
| decoder_hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings + one for the output of each layer) of shape |
| `(batch_size, sequence_length, hidden_size)`. |
| |
| Hidden-states of the decoder at the output of each layer plus the initial embedding outputs. |
| decoder_attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights of the decoder, after the attention softmax, used to compute the weighted average in the |
| self-attention heads. |
| encoder_last_hidden_state (`tf.Tensor` of shape `(batch_size, sequence_length, hidden_size)`, *optional*): |
| Sequence of hidden-states at the output of the last layer of the encoder of the model. |
| encoder_hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings + one for the output of each layer) of shape |
| `(batch_size, sequence_length, hidden_size)`. |
| |
| Hidden-states of the encoder at the output of each layer plus the initial embedding outputs. |
| encoder_attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights of the encoder, after the attention softmax, used to compute the weighted average in the |
| self-attention heads. |
| """ |
|
|
| loss: tf.Tensor | None = None |
| start_logits: tf.Tensor = None |
| end_logits: tf.Tensor = None |
| past_key_values: List[tf.Tensor] | None = None |
| decoder_hidden_states: Tuple[tf.Tensor] | None = None |
| decoder_attentions: Tuple[tf.Tensor] | None = None |
| encoder_last_hidden_state: tf.Tensor | None = None |
| encoder_hidden_states: Tuple[tf.Tensor] | None = None |
| encoder_attentions: Tuple[tf.Tensor] | None = None |
|
|
|
|
| @dataclass |
| class TFSequenceClassifierOutputWithPast(ModelOutput): |
| """ |
| Base class for outputs of sentence classification models. |
| |
| Args: |
| loss (`tf.Tensor` of shape `(batch_size, )`, *optional*, returned when `labels` is provided): |
| Classification (or regression if config.num_labels==1) loss. |
| logits (`tf.Tensor` of shape `(batch_size, config.num_labels)`): |
| Classification (or regression if config.num_labels==1) scores (before SoftMax). |
| past_key_values (`List[tf.Tensor]`, *optional*, returned when `use_cache=True` is passed or when `config.use_cache=True`): |
| List of `tf.Tensor` of length `config.n_layers`, with each tensor of shape `(2, batch_size, num_heads, |
| sequence_length, embed_size_per_head)`). |
| |
| Contains pre-computed hidden-states (key and values in the attention blocks) that can be used (see |
| `past_key_values` input) to speed up sequential decoding. |
| hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings + one for the output of each layer) of shape |
| `(batch_size, sequence_length, hidden_size)`. |
| |
| Hidden-states of the model at the output of each layer plus the initial embedding outputs. |
| attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, |
| sequence_length)`. |
| |
| Attentions weights after the attention softmax, used to compute the weighted average in the self-attention |
| heads. |
| """ |
|
|
| loss: tf.Tensor | None = None |
| logits: tf.Tensor = None |
| past_key_values: List[tf.Tensor] | None = None |
| hidden_states: Tuple[tf.Tensor] | None = None |
| attentions: Tuple[tf.Tensor] | None = None |
|
|
|
|
| @dataclass |
| class TFImageClassifierOutputWithNoAttention(ModelOutput): |
| """ |
| Base class for outputs of image classification models. |
| |
| Args: |
| loss (`tf.Tensor` of shape `(1,)`, *optional*, returned when `labels` is provided): |
| Classification (or regression if config.num_labels==1) loss. |
| logits (`tf.Tensor` of shape `(batch_size, config.num_labels)`): |
| Classification (or regression if config.num_labels==1) scores (before SoftMax). |
| hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings, if the model has an embedding layer, + one for |
| the output of each stage) of shape `(batch_size, num_channels, height, width)`. Hidden-states (also called |
| feature maps) of the model at the output of each stage. |
| """ |
|
|
| loss: tf.Tensor | None = None |
| logits: tf.Tensor = None |
| hidden_states: Optional[Tuple[tf.Tensor, ...]] = None |
|
|
|
|
| @dataclass |
| class TFMaskedImageModelingOutput(ModelOutput): |
| """ |
| Base class for outputs of masked image completion / in-painting models. |
| |
| Args: |
| loss (`tf.Tensor` of shape `(1,)`, *optional*, returned when `bool_masked_pos` is provided): |
| Reconstruction loss. |
| reconstruction (`tf.Tensor` of shape `(batch_size, num_channels, height, width)`): |
| Reconstructed / completed images. |
| hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when |
| `config.output_hidden_states=True`): |
| Tuple of `tf.Tensor` (one for the output of the embeddings, if the model has an embedding layer, + one for |
| the output of each stage) of shape `(batch_size, sequence_length, hidden_size)`. Hidden-states (also called |
| feature maps) of the model at the output of each stage. |
| attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when |
| `config.output_attentions=True`): |
| Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, patch_size, sequence_length)`. |
| Attentions weights after the attention softmax, used to compute the weighted average in the self-attention |
| heads. |
| """ |
|
|
| loss: tf.Tensor | None = None |
| reconstruction: tf.Tensor = None |
| hidden_states: Tuple[tf.Tensor] | None = None |
| attentions: Tuple[tf.Tensor] | None = None |
|
|
| @property |
| def logits(self): |
| warnings.warn( |
| "logits attribute is deprecated and will be removed in version 5 of Transformers." |
| " Please use the reconstruction attribute to retrieve the final output instead.", |
| FutureWarning, |
| ) |
| return self.reconstruction |
|
|