..,
Browse files
__pycache__/create_app.cpython-312.pyc
CHANGED
|
Binary files a/__pycache__/create_app.cpython-312.pyc and b/__pycache__/create_app.cpython-312.pyc differ
|
|
|
create_app.py
CHANGED
|
@@ -31,7 +31,7 @@ def load_models():
|
|
| 31 |
model_name = 'Qwen/Qwen3-1.7B'
|
| 32 |
QWEN_TOKENIZER = AutoTokenizer.from_pretrained(model_name, device='auto')
|
| 33 |
QWEN_TOKENIZER.pad_token_id = QWEN_TOKENIZER.eos_token_id
|
| 34 |
-
QWEN_MODEL = AutoModelForCausalLM.from_pretrained(model_name, trust_remote_code=True, device_map="auto",
|
| 35 |
QWEN_MODEL = QWEN_MODEL.to(device)
|
| 36 |
MODELS_LOADED = True
|
| 37 |
|
|
|
|
| 31 |
model_name = 'Qwen/Qwen3-1.7B'
|
| 32 |
QWEN_TOKENIZER = AutoTokenizer.from_pretrained(model_name, device='auto')
|
| 33 |
QWEN_TOKENIZER.pad_token_id = QWEN_TOKENIZER.eos_token_id
|
| 34 |
+
QWEN_MODEL = AutoModelForCausalLM.from_pretrained(model_name, trust_remote_code=True, device_map="auto", torch_dtype=torch.float16).half()
|
| 35 |
QWEN_MODEL = QWEN_MODEL.to(device)
|
| 36 |
MODELS_LOADED = True
|
| 37 |
|
inference/__pycache__/infer_single.cpython-312.pyc
CHANGED
|
Binary files a/inference/__pycache__/infer_single.cpython-312.pyc and b/inference/__pycache__/infer_single.cpython-312.pyc differ
|
|
|
requirements.txt
CHANGED
|
@@ -10,5 +10,4 @@ flask==3.1.1
|
|
| 10 |
flask_login==0.6.3
|
| 11 |
werkzeug==3.1.3
|
| 12 |
flask_sqlalchemy==3.1.1
|
| 13 |
-
gunicorn
|
| 14 |
-
bitsandbytes-0.46.0
|
|
|
|
| 10 |
flask_login==0.6.3
|
| 11 |
werkzeug==3.1.3
|
| 12 |
flask_sqlalchemy==3.1.1
|
| 13 |
+
gunicorn
|
|
|