Update app.py
Browse files
app.py
CHANGED
|
@@ -2,7 +2,9 @@ import os
|
|
| 2 |
import gc
|
| 3 |
import torch
|
| 4 |
import gradio as gr
|
| 5 |
-
from transformers import
|
|
|
|
|
|
|
| 6 |
from huggingface_hub import login
|
| 7 |
import os
|
| 8 |
|
|
@@ -24,7 +26,7 @@ print(f"Loading model from {MODEL_PATH} on {device}...")
|
|
| 24 |
# =============================
|
| 25 |
# Load Tokenizer and Model
|
| 26 |
# =============================
|
| 27 |
-
tokenizer =
|
| 28 |
model = LlamaForCausalLM.from_pretrained(
|
| 29 |
MODEL_PATH,
|
| 30 |
device_map="auto",
|
|
|
|
| 2 |
import gc
|
| 3 |
import torch
|
| 4 |
import gradio as gr
|
| 5 |
+
from transformers import StoppingCriteria, StoppingCriteriaList
|
| 6 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM
|
| 7 |
+
|
| 8 |
from huggingface_hub import login
|
| 9 |
import os
|
| 10 |
|
|
|
|
| 26 |
# =============================
|
| 27 |
# Load Tokenizer and Model
|
| 28 |
# =============================
|
| 29 |
+
tokenizer = AutoTokenizer.from_pretrained(MODEL_PATH, legacy=False)
|
| 30 |
model = LlamaForCausalLM.from_pretrained(
|
| 31 |
MODEL_PATH,
|
| 32 |
device_map="auto",
|