Readme with Neuron
#4
by
michaelfeil
- opened
README.md
CHANGED
|
@@ -135,6 +135,27 @@ Whitepaper coming soon!
|
|
| 135 |
|
| 136 |
Gradient is accelerating AI transformation across industries. https://gradient.ai/
|
| 137 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 138 |
## Contact Us
|
| 139 |
|
| 140 |
Drop an email to [contact@gradient.ai](mailto:contact@gradient.ai)
|
|
|
|
| 135 |
|
| 136 |
Gradient is accelerating AI transformation across industries. https://gradient.ai/
|
| 137 |
|
| 138 |
+
## Usage with AWS Neuron
|
| 139 |
+
```
|
| 140 |
+
from transformers import AutoTokenizer
|
| 141 |
+
from optimum.neuron import NeuronModelForCausalLM
|
| 142 |
+
|
| 143 |
+
# Instantiate and convert to Neuron a PyTorch checkpoint
|
| 144 |
+
model = NeuronModelForCausalLM.from_pretrained("gradientai/v-alpha-tross")
|
| 145 |
+
|
| 146 |
+
tokenizer = AutoTokenizer.from_pretrained("gradientai/v-alpha-tross")
|
| 147 |
+
|
| 148 |
+
tokens = tokenizer("I really wish ", return_tensors="pt")
|
| 149 |
+
with torch.inference_mode():
|
| 150 |
+
sample_output = model.generate(
|
| 151 |
+
**tokens,
|
| 152 |
+
min_length=16,
|
| 153 |
+
max_length=32,
|
| 154 |
+
)
|
| 155 |
+
outputs = [tokenizer.decode(tok) for tok in sample_output]
|
| 156 |
+
print(outputs)
|
| 157 |
+
```
|
| 158 |
+
|
| 159 |
## Contact Us
|
| 160 |
|
| 161 |
Drop an email to [contact@gradient.ai](mailto:contact@gradient.ai)
|