Commit ·
96c2387
1
Parent(s): 5ee1c5f
Update README.md
Browse files
README.md
CHANGED
|
@@ -37,14 +37,14 @@ You can use the `transformers` library to load and utilize the T5-Reverse (T5R)
|
|
| 37 |
|
| 38 |
```python
|
| 39 |
# Import required libraries
|
| 40 |
-
import torch
|
| 41 |
-
from transformers import pipeline
|
| 42 |
|
| 43 |
# Load the model and tokenizer using the pipeline from Hugging Face Hub
|
| 44 |
-
inference = pipeline("text2text-generation", model="kargaranamir/T5R-base")
|
| 45 |
|
| 46 |
# Example instruction and prompt
|
| 47 |
-
sample = '''
|
| 48 |
Instruction: X
|
| 49 |
Output: 1- Base your meals on higher fibre starchy carbohydrates. 2- Eat lots of fruit and veg. 3- Eat more fish, including a portion of oily fish.
|
| 50 |
What kind of instruction could this be the answer to?
|
|
@@ -52,9 +52,9 @@ X:
|
|
| 52 |
'''
|
| 53 |
|
| 54 |
# Generate a response using the model
|
| 55 |
-
res = inference(sample)
|
| 56 |
|
| 57 |
# Print the generated response
|
| 58 |
-
print(res)
|
| 59 |
-
|
| 60 |
```
|
|
|
|
| 37 |
|
| 38 |
```python
|
| 39 |
# Import required libraries
|
| 40 |
+
>>> import torch
|
| 41 |
+
>>> from transformers import pipeline
|
| 42 |
|
| 43 |
# Load the model and tokenizer using the pipeline from Hugging Face Hub
|
| 44 |
+
>>> inference = pipeline("text2text-generation", model="kargaranamir/T5R-base")
|
| 45 |
|
| 46 |
# Example instruction and prompt
|
| 47 |
+
>>> sample = '''
|
| 48 |
Instruction: X
|
| 49 |
Output: 1- Base your meals on higher fibre starchy carbohydrates. 2- Eat lots of fruit and veg. 3- Eat more fish, including a portion of oily fish.
|
| 50 |
What kind of instruction could this be the answer to?
|
|
|
|
| 52 |
'''
|
| 53 |
|
| 54 |
# Generate a response using the model
|
| 55 |
+
>>> res = inference(sample)
|
| 56 |
|
| 57 |
# Print the generated response
|
| 58 |
+
>>> print(res)
|
| 59 |
+
[{'generated_text': 'Instruction: Generate three recommendations for a healthy diet.'}]
|
| 60 |
```
|