Update README.md
Browse filesAdd description of encoder output to readme.
README.md
CHANGED
|
@@ -128,6 +128,12 @@ with torch.no_grad():
|
|
| 128 |
output = model(input_values=inputs['input_values'],
|
| 129 |
attention_mask=inputs['attention_mask'],
|
| 130 |
output_hidden_states=True)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 131 |
```
|
| 132 |
|
| 133 |
## Downstream Use
|
|
|
|
| 128 |
output = model(input_values=inputs['input_values'],
|
| 129 |
attention_mask=inputs['attention_mask'],
|
| 130 |
output_hidden_states=True)
|
| 131 |
+
|
| 132 |
+
# output is a Wav2Vec2BaseModelOutput or tuple containing:
|
| 133 |
+
# last_hidden_state: torch.FloatTensor containing hidden states of the last layer of the model
|
| 134 |
+
# extract_features: torch.FloatTensor containing extracted features from the convolution downsampling layers
|
| 135 |
+
# hidden_states: tuple(torch.FloatTensor) containing hidden states of each layer of the model
|
| 136 |
+
# attentions: tuple(torch.FloatTensor) containing attention states of each layer of the model
|
| 137 |
```
|
| 138 |
|
| 139 |
## Downstream Use
|