debug
Browse files- README.md +1 -1
- data/models/llama3-1-70b.py +1 -1
README.md
CHANGED
|
@@ -4,7 +4,7 @@ emoji: 🐠
|
|
| 4 |
colorFrom: gray
|
| 5 |
colorTo: yellow
|
| 6 |
sdk: gradio
|
| 7 |
-
sdk_version: 4.
|
| 8 |
app_file: app.py
|
| 9 |
pinned: false
|
| 10 |
license: llama3
|
|
|
|
| 4 |
colorFrom: gray
|
| 5 |
colorTo: yellow
|
| 6 |
sdk: gradio
|
| 7 |
+
sdk_version: 4.40.0
|
| 8 |
app_file: app.py
|
| 9 |
pinned: false
|
| 10 |
license: llama3
|
data/models/llama3-1-70b.py
CHANGED
|
@@ -18,7 +18,7 @@ model = AutoModelForCausalLM.from_pretrained(
|
|
| 18 |
torch_dtype=torch.float16,
|
| 19 |
device_map="sequential",
|
| 20 |
offload_folder="offload",
|
| 21 |
-
offload_state_dict=True
|
| 22 |
)
|
| 23 |
|
| 24 |
@app.route('/chat', methods=['POST'])
|
|
|
|
| 18 |
torch_dtype=torch.float16,
|
| 19 |
device_map="sequential",
|
| 20 |
offload_folder="offload",
|
| 21 |
+
offload_state_dict=True
|
| 22 |
)
|
| 23 |
|
| 24 |
@app.route('/chat', methods=['POST'])
|