Update app.py
Browse files
app.py
CHANGED
|
@@ -2,13 +2,12 @@ import os
|
|
| 2 |
import gc
|
| 3 |
import torch
|
| 4 |
import gradio as gr
|
| 5 |
-
from transformers import StoppingCriteria, StoppingCriteriaList
|
| 6 |
-
from transformers import AutoTokenizer, AutoModelForCausalLM
|
| 7 |
|
| 8 |
# =============================
|
| 9 |
# Configuration
|
| 10 |
# =============================
|
| 11 |
-
MODEL_PATH = r"
|
| 12 |
MAX_NEW_TOKENS = 200
|
| 13 |
TEMPERATURE = 0.5
|
| 14 |
TOP_K = 50
|
|
@@ -21,7 +20,7 @@ print(f"Loading model from {MODEL_PATH} on {device}...")
|
|
| 21 |
# =============================
|
| 22 |
# Load Tokenizer and Model
|
| 23 |
# =============================
|
| 24 |
-
tokenizer =
|
| 25 |
model = LlamaForCausalLM.from_pretrained(
|
| 26 |
MODEL_PATH,
|
| 27 |
device_map="auto",
|
|
|
|
| 2 |
import gc
|
| 3 |
import torch
|
| 4 |
import gradio as gr
|
| 5 |
+
from transformers import LlamaTokenizer, LlamaForCausalLM, StoppingCriteria, StoppingCriteriaList
|
|
|
|
| 6 |
|
| 7 |
# =============================
|
| 8 |
# Configuration
|
| 9 |
# =============================
|
| 10 |
+
MODEL_PATH = r"C:\Users\JAY\Downloads\Chatdoc\ChatDoctor\pretrained"
|
| 11 |
MAX_NEW_TOKENS = 200
|
| 12 |
TEMPERATURE = 0.5
|
| 13 |
TOP_K = 50
|
|
|
|
| 20 |
# =============================
|
| 21 |
# Load Tokenizer and Model
|
| 22 |
# =============================
|
| 23 |
+
tokenizer = LlamaTokenizer.from_pretrained(MODEL_PATH)
|
| 24 |
model = LlamaForCausalLM.from_pretrained(
|
| 25 |
MODEL_PATH,
|
| 26 |
device_map="auto",
|