Spaces:
Sleeping
Sleeping
Deploy FastAPI Swagger Space (Docker)
Browse files
app.py
CHANGED
|
@@ -7,6 +7,7 @@ from pydantic import BaseModel, Field
|
|
| 7 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 8 |
|
| 9 |
MODEL_REPO = "aagzamov/search-query-parser"
|
|
|
|
| 10 |
MAX_NEW_TOKENS = 192
|
| 11 |
|
| 12 |
SortType = Literal["relevance", "price_asc", "price_desc", "newest"]
|
|
@@ -87,7 +88,7 @@ model = None
|
|
| 87 |
|
| 88 |
def load_model():
|
| 89 |
global tokenizer, model
|
| 90 |
-
tokenizer = AutoTokenizer.from_pretrained(
|
| 91 |
model = AutoModelForCausalLM.from_pretrained(
|
| 92 |
MODEL_REPO,
|
| 93 |
torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32,
|
|
|
|
| 7 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 8 |
|
| 9 |
MODEL_REPO = "aagzamov/search-query-parser"
|
| 10 |
+
TOKENIZER_REPO = "Qwen/Qwen2.5-1.5B-Instruct"
|
| 11 |
MAX_NEW_TOKENS = 192
|
| 12 |
|
| 13 |
SortType = Literal["relevance", "price_asc", "price_desc", "newest"]
|
|
|
|
| 88 |
|
| 89 |
def load_model():
|
| 90 |
global tokenizer, model
|
| 91 |
+
tokenizer = AutoTokenizer.from_pretrained(TOKENIZER_REPO, use_fast=False)
|
| 92 |
model = AutoModelForCausalLM.from_pretrained(
|
| 93 |
MODEL_REPO,
|
| 94 |
torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32,
|