Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -30,7 +30,7 @@ from transformers.utils import logging
|
|
| 30 |
from transformers import AutoTokenizer, AutoModelForCausalLM # isort: skip
|
| 31 |
|
| 32 |
logger = logging.get_logger(__name__)
|
| 33 |
-
model_name_or_path="
|
| 34 |
|
| 35 |
@dataclass
|
| 36 |
class GenerationConfig:
|
|
@@ -186,7 +186,7 @@ def load_model():
|
|
| 186 |
# trust_remote_code=True).to(torch.bfloat16).cuda())
|
| 187 |
model = AutoModelForCausalLM.from_pretrained(
|
| 188 |
model_name_or_path,
|
| 189 |
-
trust_remote_code=True).to(
|
| 190 |
tokenizer = AutoTokenizer.from_pretrained(model_name_or_path,
|
| 191 |
trust_remote_code=True)
|
| 192 |
return model, tokenizer
|
|
|
|
| 30 |
from transformers import AutoTokenizer, AutoModelForCausalLM # isort: skip
|
| 31 |
|
| 32 |
logger = logging.get_logger(__name__)
|
| 33 |
+
model_name_or_path="cloneQ/my_personal_assistant"
|
| 34 |
|
| 35 |
@dataclass
|
| 36 |
class GenerationConfig:
|
|
|
|
| 186 |
# trust_remote_code=True).to(torch.bfloat16).cuda())
|
| 187 |
model = AutoModelForCausalLM.from_pretrained(
|
| 188 |
model_name_or_path,
|
| 189 |
+
trust_remote_code=True).to("cpu")
|
| 190 |
tokenizer = AutoTokenizer.from_pretrained(model_name_or_path,
|
| 191 |
trust_remote_code=True)
|
| 192 |
return model, tokenizer
|