SFM2001 commited on
Commit
ed8bf25
·
1 Parent(s): b0f940f
__pycache__/create_app.cpython-312.pyc CHANGED
Binary files a/__pycache__/create_app.cpython-312.pyc and b/__pycache__/create_app.cpython-312.pyc differ
 
create_app.py CHANGED
@@ -31,7 +31,7 @@ def load_models():
31
  model_name = 'Qwen/Qwen3-1.7B'
32
  QWEN_TOKENIZER = AutoTokenizer.from_pretrained(model_name, device='auto')
33
  QWEN_TOKENIZER.pad_token_id = QWEN_TOKENIZER.eos_token_id
34
- QWEN_MODEL = AutoModelForCausalLM.from_pretrained(model_name, trust_remote_code=True, device_map="auto", load_in_8bit=True, torch_dtype=torch.float16).half()
35
  QWEN_MODEL = QWEN_MODEL.to(device)
36
  MODELS_LOADED = True
37
 
 
31
  model_name = 'Qwen/Qwen3-1.7B'
32
  QWEN_TOKENIZER = AutoTokenizer.from_pretrained(model_name, device='auto')
33
  QWEN_TOKENIZER.pad_token_id = QWEN_TOKENIZER.eos_token_id
34
+ QWEN_MODEL = AutoModelForCausalLM.from_pretrained(model_name, trust_remote_code=True, device_map="auto", torch_dtype=torch.float16).half()
35
  QWEN_MODEL = QWEN_MODEL.to(device)
36
  MODELS_LOADED = True
37
 
inference/__pycache__/infer_single.cpython-312.pyc CHANGED
Binary files a/inference/__pycache__/infer_single.cpython-312.pyc and b/inference/__pycache__/infer_single.cpython-312.pyc differ
 
requirements.txt CHANGED
@@ -10,5 +10,4 @@ flask==3.1.1
10
  flask_login==0.6.3
11
  werkzeug==3.1.3
12
  flask_sqlalchemy==3.1.1
13
- gunicorn
14
- bitsandbytes-0.46.0
 
10
  flask_login==0.6.3
11
  werkzeug==3.1.3
12
  flask_sqlalchemy==3.1.1
13
+ gunicorn