stream
Browse files- .gitignore +2 -0
- chat_app.py +4 -0
.gitignore
CHANGED
|
@@ -172,3 +172,5 @@ cython_debug/
|
|
| 172 |
|
| 173 |
# PyPI configuration file
|
| 174 |
.pypirc
|
|
|
|
|
|
|
|
|
| 172 |
|
| 173 |
# PyPI configuration file
|
| 174 |
.pypirc
|
| 175 |
+
/.github/
|
| 176 |
+
/.github/
|
chat_app.py
CHANGED
|
@@ -3,6 +3,7 @@ import os
|
|
| 3 |
import gradio as gr
|
| 4 |
from typing import List
|
| 5 |
|
|
|
|
| 6 |
from langchain_huggingface import HuggingFaceEndpoint
|
| 7 |
|
| 8 |
from langchain.schema import BaseMessage
|
|
@@ -45,10 +46,13 @@ prompt = ChatPromptTemplate.from_messages([
|
|
| 45 |
])
|
| 46 |
|
| 47 |
model_id="mistralai/Mistral-7B-Instruct-v0.3"
|
|
|
|
| 48 |
llm = HuggingFaceEndpoint(
|
| 49 |
repo_id=model_id,
|
| 50 |
max_new_tokens=1024,
|
| 51 |
temperature=0.1,
|
|
|
|
|
|
|
| 52 |
huggingfacehub_api_token=os.getenv('HF_TOKEN'),
|
| 53 |
)
|
| 54 |
|
|
|
|
| 3 |
import gradio as gr
|
| 4 |
from typing import List
|
| 5 |
|
| 6 |
+
import langchain_core.callbacks
|
| 7 |
from langchain_huggingface import HuggingFaceEndpoint
|
| 8 |
|
| 9 |
from langchain.schema import BaseMessage
|
|
|
|
| 46 |
])
|
| 47 |
|
| 48 |
model_id="mistralai/Mistral-7B-Instruct-v0.3"
|
| 49 |
+
callbacks = [langchain_core.callbacks.StreamingStdOutCallbackHandler()]
|
| 50 |
llm = HuggingFaceEndpoint(
|
| 51 |
repo_id=model_id,
|
| 52 |
max_new_tokens=1024,
|
| 53 |
temperature=0.1,
|
| 54 |
+
callbacks=callbacks,
|
| 55 |
+
streaming=True,
|
| 56 |
huggingfacehub_api_token=os.getenv('HF_TOKEN'),
|
| 57 |
)
|
| 58 |
|