Spaces:
Runtime error
Runtime error
Commit
·
cd8cbfa
1
Parent(s):
8ff95da
Update app.py
Browse files
app.py
CHANGED
|
@@ -35,6 +35,8 @@ from langchain.chat_models import ChatOpenAI
|
|
| 35 |
from langchain import PromptTemplate
|
| 36 |
from fpdf import FPDF
|
| 37 |
|
|
|
|
|
|
|
| 38 |
st.title('MedTechAI')
|
| 39 |
|
| 40 |
persist_directory = ""
|
|
@@ -47,16 +49,8 @@ with st.sidebar:
|
|
| 47 |
submitted = st.form_submit_button("Submit")
|
| 48 |
|
| 49 |
|
| 50 |
-
|
| 51 |
-
|
| 52 |
-
if(mod=='OpenAI'):
|
| 53 |
-
os.environ["OPENAI_API_KEY"] = api_key
|
| 54 |
-
llm = OpenAI(temperature=0.1, verbose=True)
|
| 55 |
-
embeddings = OpenAIEmbeddings()
|
| 56 |
-
if(mod=='Cohere'):
|
| 57 |
-
os.environ["Cohere_API_KEY"] = api_key
|
| 58 |
-
llm = Cohere(cohere_api_key=api_key)
|
| 59 |
-
embeddings = CohereEmbeddings(cohere_api_key=api_key)
|
| 60 |
|
| 61 |
|
| 62 |
def report(Symptoms,history,age,gen,response):
|
|
@@ -108,7 +102,8 @@ def report(Symptoms,history,age,gen,response):
|
|
| 108 |
print("report")
|
| 109 |
model_name = "gpt-3.5-turbo-16k"
|
| 110 |
temperature = 0.0
|
| 111 |
-
llm = OpenAI(model_name=model_name, temperature=temperature)
|
|
|
|
| 112 |
out = llm(prompt_updates)
|
| 113 |
|
| 114 |
return out
|
|
|
|
| 35 |
from langchain import PromptTemplate
|
| 36 |
from fpdf import FPDF
|
| 37 |
|
| 38 |
+
|
| 39 |
+
|
| 40 |
st.title('MedTechAI')
|
| 41 |
|
| 42 |
persist_directory = ""
|
|
|
|
| 49 |
submitted = st.form_submit_button("Submit")
|
| 50 |
|
| 51 |
|
| 52 |
+
|
| 53 |
+
embeddings = HuggingFaceEmbeddings(model_name="sentence-transformers/all-MiniLM-L6-v2", model_kwargs={'device': 'cpu'})
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 54 |
|
| 55 |
|
| 56 |
def report(Symptoms,history,age,gen,response):
|
|
|
|
| 102 |
print("report")
|
| 103 |
model_name = "gpt-3.5-turbo-16k"
|
| 104 |
temperature = 0.0
|
| 105 |
+
#llm = OpenAI(model_name=model_name, temperature=temperature)
|
| 106 |
+
llm = CTransformers(model = "TheBloke/Llama-2-7B-Chat-GGML",model_type="llama",max_new_tokens = 2048,temperature = 0.5)
|
| 107 |
out = llm(prompt_updates)
|
| 108 |
|
| 109 |
return out
|