Spaces:
Sleeping
Sleeping
Commit
·
dcb2108
1
Parent(s):
4cf07ab
fix: set device_map to auto
Browse files
app.py
CHANGED
|
@@ -8,11 +8,11 @@ from transformers import (
|
|
| 8 |
import torch
|
| 9 |
|
| 10 |
chat_model_name = "sapienzanlp/Minerva-7B-instruct-v1.0"
|
| 11 |
-
chat_model = AutoModelForCausalLM.from_pretrained(chat_model_name, dtype=torch.bfloat16)
|
| 12 |
chat_tokenizer = AutoTokenizer.from_pretrained(chat_model_name)
|
| 13 |
|
| 14 |
moderator_model_name = "saiteki-kai/QA-DeBERTa-v3-large"
|
| 15 |
-
moderator_model = AutoModelForSequenceClassification.from_pretrained(moderator_model_name)
|
| 16 |
moderator_tokenizer = AutoTokenizer.from_pretrained(moderator_model_name)
|
| 17 |
|
| 18 |
@spaces.GPU()
|
|
|
|
| 8 |
import torch
|
| 9 |
|
| 10 |
chat_model_name = "sapienzanlp/Minerva-7B-instruct-v1.0"
|
| 11 |
+
chat_model = AutoModelForCausalLM.from_pretrained(chat_model_name, dtype=torch.bfloat16, device_map="auto")
|
| 12 |
chat_tokenizer = AutoTokenizer.from_pretrained(chat_model_name)
|
| 13 |
|
| 14 |
moderator_model_name = "saiteki-kai/QA-DeBERTa-v3-large"
|
| 15 |
+
moderator_model = AutoModelForSequenceClassification.from_pretrained(moderator_model_name, device_map="auto")
|
| 16 |
moderator_tokenizer = AutoTokenizer.from_pretrained(moderator_model_name)
|
| 17 |
|
| 18 |
@spaces.GPU()
|