Spaces:
Sleeping
Sleeping
downgrade to use meta-llama/Meta-Llama-3-8B-Instruct
Browse files
app.py
CHANGED
|
@@ -293,7 +293,7 @@ def chat_actions():
|
|
| 293 |
result = ""
|
| 294 |
try:
|
| 295 |
llm = HuggingFaceHub(
|
| 296 |
-
repo_id="meta-llama/Meta-Llama-3
|
| 297 |
)
|
| 298 |
response_text = llm.invoke(prompt)
|
| 299 |
escaped_query = re.escape(query)
|
|
|
|
| 293 |
result = ""
|
| 294 |
try:
|
| 295 |
llm = HuggingFaceHub(
|
| 296 |
+
repo_id="meta-llama/Meta-Llama-3-8B-Instruct", model_kwargs={"temperature": 0.1, "max_new_tokens": 256, "task":"text-generation"}
|
| 297 |
)
|
| 298 |
response_text = llm.invoke(prompt)
|
| 299 |
escaped_query = re.escape(query)
|