Commit
·
b06b002
1
Parent(s):
ea64011
Update README.md
Browse files
README.md
CHANGED
|
@@ -70,7 +70,10 @@ output = tokenizer.decode(output1[0])
|
|
| 70 |
|
| 71 |
print(output)
|
| 72 |
|
| 73 |
-
|
|
|
|
|
|
|
|
|
|
| 74 |
Sure, I can help you with that!
|
| 75 |
|
| 76 |
Attention mechanisms in transformer models are typically implemented using the attention mechanism in the self-attention layer. Self-attention allows the model to focus on different parts of the input sequence when processing it. This is achieved by computing a set of attention weights, which are used to weigh the contribution of each input element to the output.
|
|
@@ -118,9 +121,8 @@ The `query`, `key`, and `value` tensors represent the input sequence to the tran
|
|
| 118 |
The output of the `attention_weights` function is a NumPy tensor that represents the attention weights for the input sequence. These weights are used by the transformer model to weigh the contribution of each input element to the output.
|
| 119 |
|
| 120 |
I hope this helps!</s>
|
| 121 |
-
|
| 122 |
-
|
| 123 |
-
|
| 124 |
## Finetuning details
|
| 125 |
The finetuning scripts will be available in our [RAIL Github Repository](https://github.com/vmware-labs/research-and-development-artificial-intelligence-lab/tree/main/instruction-tuning)
|
| 126 |
## Evaluation
|
|
|
|
| 70 |
|
| 71 |
print(output)
|
| 72 |
|
| 73 |
+
```
|
| 74 |
+
### Output
|
| 75 |
+
|
| 76 |
+
|
| 77 |
Sure, I can help you with that!
|
| 78 |
|
| 79 |
Attention mechanisms in transformer models are typically implemented using the attention mechanism in the self-attention layer. Self-attention allows the model to focus on different parts of the input sequence when processing it. This is achieved by computing a set of attention weights, which are used to weigh the contribution of each input element to the output.
|
|
|
|
| 121 |
The output of the `attention_weights` function is a NumPy tensor that represents the attention weights for the input sequence. These weights are used by the transformer model to weigh the contribution of each input element to the output.
|
| 122 |
|
| 123 |
I hope this helps!</s>
|
| 124 |
+
<hr>
|
| 125 |
+
|
|
|
|
| 126 |
## Finetuning details
|
| 127 |
The finetuning scripts will be available in our [RAIL Github Repository](https://github.com/vmware-labs/research-and-development-artificial-intelligence-lab/tree/main/instruction-tuning)
|
| 128 |
## Evaluation
|