Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -1,4 +1,7 @@
|
|
| 1 |
from tqdm import tqdm
|
|
|
|
|
|
|
|
|
|
| 2 |
from subprocess import Popen, PIPE as P
|
| 3 |
#b=Popen("wget 1https://huggingface.co/peterpeter8585/deepseek-llm-7b-chat-Q8_0-GGUF/resolve/main/deepseek-llm-7b-chat-q8_0.gguf -O ./model.gguf",shell=True,stderr=P,stdout=P)
|
| 4 |
#print(b.stdout.read()+b.stderr.read())
|
|
@@ -76,12 +79,7 @@ import multiprocessing
|
|
| 76 |
#from langchain.llms import GPT4All
|
| 77 |
|
| 78 |
#llm = GPT4All(model=("./model.gguf"))
|
| 79 |
-
from langchain_huggingface.llms import HuggingFacePipeline
|
| 80 |
|
| 81 |
-
llm = HuggingFacePipeline.from_model_id(
|
| 82 |
-
model_id="HuggingFaceH4/zephyr-7b-beta",
|
| 83 |
-
task="text-generation"
|
| 84 |
-
)
|
| 85 |
from langchain.retrievers import WikipediaRetriever as Wiki
|
| 86 |
import gradio as gr
|
| 87 |
chatbot = gr.Chatbot(
|
|
|
|
| 1 |
from tqdm import tqdm
|
| 2 |
+
from langchain_huggingface.llms import HuggingFaceEndpoint as HF
|
| 3 |
+
|
| 4 |
+
llm = HF(repo_id="HuggingFaceH4/zephyr-7b-beta")
|
| 5 |
from subprocess import Popen, PIPE as P
|
| 6 |
#b=Popen("wget 1https://huggingface.co/peterpeter8585/deepseek-llm-7b-chat-Q8_0-GGUF/resolve/main/deepseek-llm-7b-chat-q8_0.gguf -O ./model.gguf",shell=True,stderr=P,stdout=P)
|
| 7 |
#print(b.stdout.read()+b.stderr.read())
|
|
|
|
| 79 |
#from langchain.llms import GPT4All
|
| 80 |
|
| 81 |
#llm = GPT4All(model=("./model.gguf"))
|
|
|
|
| 82 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 83 |
from langchain.retrievers import WikipediaRetriever as Wiki
|
| 84 |
import gradio as gr
|
| 85 |
chatbot = gr.Chatbot(
|