Spaces:
Runtime error
Runtime error
Commit
·
b28acc9
1
Parent(s):
cd8cbfa
Update app.py
Browse files
app.py
CHANGED
|
@@ -1,8 +1,7 @@
|
|
| 1 |
import streamlit as st
|
| 2 |
-
from langchain.llms import
|
| 3 |
import os
|
| 4 |
-
from langchain.
|
| 5 |
-
from langchain.embeddings import OpenAIEmbeddings,CohereEmbeddings
|
| 6 |
from langchain.document_loaders import PyPDFLoader
|
| 7 |
from langchain.vectorstores import Chroma
|
| 8 |
from langchain.agents.agent_toolkits import (
|
|
@@ -17,12 +16,10 @@ from langchain.agents import Tool
|
|
| 17 |
from langchain.tools import DuckDuckGoSearchRun
|
| 18 |
# from langchain.utilities import WikipediaAPIWrapper
|
| 19 |
from langchain.chains import RetrievalQA
|
| 20 |
-
from langchain.chat_models import ChatOpenAI
|
| 21 |
from langchain.agents import Tool
|
| 22 |
from langchain.agents import initialize_agent
|
| 23 |
from langchain.agents import AgentType
|
| 24 |
from langchain.memory import ConversationBufferMemory
|
| 25 |
-
from langchain.chat_models import ChatOpenAI
|
| 26 |
from langchain.prompts import ChatPromptTemplate
|
| 27 |
from typing import Dict, Union, Any, List
|
| 28 |
|
|
@@ -30,7 +27,6 @@ from langchain.callbacks.base import BaseCallbackHandler
|
|
| 30 |
from langchain.schema import AgentAction
|
| 31 |
from langchain.agents import AgentType, initialize_agent, load_tools
|
| 32 |
from langchain.callbacks import tracing_enabled
|
| 33 |
-
from langchain.llms import OpenAI
|
| 34 |
from langchain.chat_models import ChatOpenAI
|
| 35 |
from langchain import PromptTemplate
|
| 36 |
from fpdf import FPDF
|
|
@@ -40,18 +36,18 @@ from fpdf import FPDF
|
|
| 40 |
st.title('MedTechAI')
|
| 41 |
|
| 42 |
persist_directory = ""
|
| 43 |
-
model = ""
|
| 44 |
-
with st.sidebar:
|
| 45 |
-
with st.form('Cohere/OpenAI'):
|
| 46 |
-
mod = st.radio('Choose OpenAI/Cohere', ('OpenAI', 'Cohere'))
|
| 47 |
-
api_key = st.text_input('Enter API key', type="password")
|
| 48 |
-
# model = st.radio('Choose Company', ('ArtisanAppetite foods', 'BMW','Titan Watches'))
|
| 49 |
-
submitted = st.form_submit_button("Submit")
|
| 50 |
-
|
| 51 |
|
| 52 |
|
| 53 |
embeddings = HuggingFaceEmbeddings(model_name="sentence-transformers/all-MiniLM-L6-v2", model_kwargs={'device': 'cpu'})
|
| 54 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 55 |
|
| 56 |
def report(Symptoms,history,age,gen,response):
|
| 57 |
prompt_updates=f"""
|
|
@@ -100,9 +96,6 @@ def report(Symptoms,history,age,gen,response):
|
|
| 100 |
```
|
| 101 |
"""
|
| 102 |
print("report")
|
| 103 |
-
model_name = "gpt-3.5-turbo-16k"
|
| 104 |
-
temperature = 0.0
|
| 105 |
-
#llm = OpenAI(model_name=model_name, temperature=temperature)
|
| 106 |
llm = CTransformers(model = "TheBloke/Llama-2-7B-Chat-GGML",model_type="llama",max_new_tokens = 2048,temperature = 0.5)
|
| 107 |
out = llm(prompt_updates)
|
| 108 |
|
|
|
|
| 1 |
import streamlit as st
|
| 2 |
+
from langchain.llms import CTransformers
|
| 3 |
import os
|
| 4 |
+
from langchain.embeddings import HuggingFaceEmbeddings
|
|
|
|
| 5 |
from langchain.document_loaders import PyPDFLoader
|
| 6 |
from langchain.vectorstores import Chroma
|
| 7 |
from langchain.agents.agent_toolkits import (
|
|
|
|
| 16 |
from langchain.tools import DuckDuckGoSearchRun
|
| 17 |
# from langchain.utilities import WikipediaAPIWrapper
|
| 18 |
from langchain.chains import RetrievalQA
|
|
|
|
| 19 |
from langchain.agents import Tool
|
| 20 |
from langchain.agents import initialize_agent
|
| 21 |
from langchain.agents import AgentType
|
| 22 |
from langchain.memory import ConversationBufferMemory
|
|
|
|
| 23 |
from langchain.prompts import ChatPromptTemplate
|
| 24 |
from typing import Dict, Union, Any, List
|
| 25 |
|
|
|
|
| 27 |
from langchain.schema import AgentAction
|
| 28 |
from langchain.agents import AgentType, initialize_agent, load_tools
|
| 29 |
from langchain.callbacks import tracing_enabled
|
|
|
|
| 30 |
from langchain.chat_models import ChatOpenAI
|
| 31 |
from langchain import PromptTemplate
|
| 32 |
from fpdf import FPDF
|
|
|
|
| 36 |
st.title('MedTechAI')
|
| 37 |
|
| 38 |
persist_directory = ""
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 39 |
|
| 40 |
|
| 41 |
embeddings = HuggingFaceEmbeddings(model_name="sentence-transformers/all-MiniLM-L6-v2", model_kwargs={'device': 'cpu'})
|
| 42 |
|
| 43 |
+
print("""
|
| 44 |
+
Duplicated space from WeekendAI/MedicalBot
|
| 45 |
+
https://huggingface.co/spaces/WeekendAI/MedicalBot
|
| 46 |
+
|
| 47 |
+
This version has been disconnected from OpenAI and needs no API key.
|
| 48 |
+
This uses the TheBloke/Llama-2-7B-Chat-GGML model and the sentence-transformers/all-MiniLM-L6-v2 HuggingFaceEmbeddings
|
| 49 |
+
|
| 50 |
+
""")
|
| 51 |
|
| 52 |
def report(Symptoms,history,age,gen,response):
|
| 53 |
prompt_updates=f"""
|
|
|
|
| 96 |
```
|
| 97 |
"""
|
| 98 |
print("report")
|
|
|
|
|
|
|
|
|
|
| 99 |
llm = CTransformers(model = "TheBloke/Llama-2-7B-Chat-GGML",model_type="llama",max_new_tokens = 2048,temperature = 0.5)
|
| 100 |
out = llm(prompt_updates)
|
| 101 |
|