Update app.py
Browse files
app.py
CHANGED
|
@@ -28,7 +28,7 @@ def func(user):
|
|
| 28 |
prompt = PromptTemplate(template=template, input_variables=["question"])
|
| 29 |
|
| 30 |
local_path = (
|
| 31 |
-
"./
|
| 32 |
)
|
| 33 |
|
| 34 |
|
|
@@ -36,7 +36,7 @@ def func(user):
|
|
| 36 |
# callbacks = [StreamingStdOutCallbackHandler()]
|
| 37 |
|
| 38 |
# Verbose is required to pass to the callback manager
|
| 39 |
-
llm = LlamaCpp(
|
| 40 |
llm_chain = LLMChain(prompt=prompt, llm=llm)
|
| 41 |
question = user
|
| 42 |
llm_chain.run(question)
|
|
|
|
| 28 |
prompt = PromptTemplate(template=template, input_variables=["question"])
|
| 29 |
|
| 30 |
local_path = (
|
| 31 |
+
"./ggml-gpt4all-j-v1.3-groovy.bin"
|
| 32 |
)
|
| 33 |
|
| 34 |
|
|
|
|
| 36 |
# callbacks = [StreamingStdOutCallbackHandler()]
|
| 37 |
|
| 38 |
# Verbose is required to pass to the callback manager
|
| 39 |
+
llm = LlamaCpp(model_path="./ggml-gpt4all-j-v1.3-groovy.bin")
|
| 40 |
llm_chain = LLMChain(prompt=prompt, llm=llm)
|
| 41 |
question = user
|
| 42 |
llm_chain.run(question)
|