Update README.md
Browse files
README.md
CHANGED
|
@@ -138,6 +138,8 @@ For improved practical inference speed, we strongly recommend running Jellyfish
|
|
| 138 |
We provide two simple Python code examples for inference using the Jellyfish model.
|
| 139 |
|
| 140 |
#### Using Transformers and Torch Modules
|
|
|
|
|
|
|
| 141 |
```python
|
| 142 |
from transformers import AutoModelForCausalLM, AutoTokenizer, GenerationConfig
|
| 143 |
import torch
|
|
@@ -192,6 +194,7 @@ response = tokenizer.decode(
|
|
| 192 |
print(response)
|
| 193 |
|
| 194 |
```
|
|
|
|
| 195 |
|
| 196 |
#### Using vLLM
|
| 197 |
```python
|
|
|
|
| 138 |
We provide two simple Python code examples for inference using the Jellyfish model.
|
| 139 |
|
| 140 |
#### Using Transformers and Torch Modules
|
| 141 |
+
<div style="height: auto; max-height: 400px; overflow-y: scroll;">
|
| 142 |
+
|
| 143 |
```python
|
| 144 |
from transformers import AutoModelForCausalLM, AutoTokenizer, GenerationConfig
|
| 145 |
import torch
|
|
|
|
| 194 |
print(response)
|
| 195 |
|
| 196 |
```
|
| 197 |
+
</div>
|
| 198 |
|
| 199 |
#### Using vLLM
|
| 200 |
```python
|