Update app.py
Browse files
app.py
CHANGED
|
@@ -11,7 +11,7 @@ st.title("🩺 ChatDoctor - Medical Assistant")
|
|
| 11 |
# Load model and tokenizer
|
| 12 |
@st.cache_resource
|
| 13 |
def load_model():
|
| 14 |
-
model = AutoModelForCausalLM.from_pretrained("abhiyanta/chatDoctor").to("cpu")
|
| 15 |
tokenizer = AutoTokenizer.from_pretrained("abhiyanta/chatDoctor")
|
| 16 |
return model, tokenizer
|
| 17 |
|
|
@@ -20,7 +20,7 @@ model, tokenizer = load_model()
|
|
| 20 |
# Alpaca-style prompt template
|
| 21 |
alpaca_prompt = "### Instruction:\n{0}\n\n### Input:\n{1}\n\n### Output:\n{2}"
|
| 22 |
|
| 23 |
-
# Text input for the user
|
| 24 |
user_input = st.text_input("Ask your medical question:")
|
| 25 |
|
| 26 |
# Button to trigger response
|
|
|
|
| 11 |
# Load model and tokenizer
|
| 12 |
@st.cache_resource
|
| 13 |
def load_model():
|
| 14 |
+
model = AutoModelForCausalLM.from_pretrained("abhiyanta/chatDoctor", use_cache=True).to("cpu")
|
| 15 |
tokenizer = AutoTokenizer.from_pretrained("abhiyanta/chatDoctor")
|
| 16 |
return model, tokenizer
|
| 17 |
|
|
|
|
| 20 |
# Alpaca-style prompt template
|
| 21 |
alpaca_prompt = "### Instruction:\n{0}\n\n### Input:\n{1}\n\n### Output:\n{2}"
|
| 22 |
|
| 23 |
+
# Text input for the user
|
| 24 |
user_input = st.text_input("Ask your medical question:")
|
| 25 |
|
| 26 |
# Button to trigger response
|