- create_app.py +4 -2
create_app.py
CHANGED
|
@@ -35,13 +35,15 @@ def load_models():
|
|
| 35 |
raise # Re-raise to see full traceback
|
| 36 |
# QWEN_MODEL = AutoModelForCausalLM.from_pretrained(model_name).half()
|
| 37 |
# QWEN_MODEL = QWEN_MODEL.to(device)
|
| 38 |
-
print(
|
|
|
|
| 39 |
LONGFORMER_TOKENIZER = LongformerTokenizer.from_pretrained('allenai/longformer-base-4096', device='auto')
|
| 40 |
config = LongformerConfig.from_json_file("Longformer_checkpoint/config.json")
|
| 41 |
print("LONGFORMER TOKENIZER LOADED")
|
| 42 |
LONGFORMER_MODEL = CustomLongformerForSequenceClassification(config).from_pretrained('SFM2001/LongFormerScorer')
|
| 43 |
LONGFORMER_MODEL = LONGFORMER_MODEL.to(device)
|
| 44 |
-
print(
|
|
|
|
| 45 |
LONGFORMER_MODEL.eval()
|
| 46 |
MODELS_LOADED = True
|
| 47 |
print("LOAD ENDED")
|
|
|
|
| 35 |
raise # Re-raise to see full traceback
|
| 36 |
# QWEN_MODEL = AutoModelForCausalLM.from_pretrained(model_name).half()
|
| 37 |
# QWEN_MODEL = QWEN_MODEL.to(device)
|
| 38 |
+
print(QWEN_MODEL)
|
| 39 |
+
print("QWEN MODEL LOADED", flush=True)
|
| 40 |
LONGFORMER_TOKENIZER = LongformerTokenizer.from_pretrained('allenai/longformer-base-4096', device='auto')
|
| 41 |
config = LongformerConfig.from_json_file("Longformer_checkpoint/config.json")
|
| 42 |
print("LONGFORMER TOKENIZER LOADED")
|
| 43 |
LONGFORMER_MODEL = CustomLongformerForSequenceClassification(config).from_pretrained('SFM2001/LongFormerScorer')
|
| 44 |
LONGFORMER_MODEL = LONGFORMER_MODEL.to(device)
|
| 45 |
+
print(LONGFORMER_MODEL)
|
| 46 |
+
print("LONGFORMER MODEL LOADED", flush=True)
|
| 47 |
LONGFORMER_MODEL.eval()
|
| 48 |
MODELS_LOADED = True
|
| 49 |
print("LOAD ENDED")
|