Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -61,14 +61,14 @@
|
|
| 61 |
# demo.launch()
|
| 62 |
|
| 63 |
|
| 64 |
-
|
| 65 |
import gradio as gr
|
| 66 |
from huggingface_hub import InferenceClient
|
| 67 |
|
| 68 |
# Initialize the Hugging Face Inference Client
|
| 69 |
client = InferenceClient(
|
| 70 |
"meta-llama/Meta-Llama-3.1-8B-Instruct",
|
| 71 |
-
|
| 72 |
)
|
| 73 |
|
| 74 |
# Define a function to handle the chat input and get a response from the model
|
|
|
|
| 61 |
# demo.launch()
|
| 62 |
|
| 63 |
|
| 64 |
+
import os
|
| 65 |
import gradio as gr
|
| 66 |
from huggingface_hub import InferenceClient
|
| 67 |
|
| 68 |
# Initialize the Hugging Face Inference Client
|
| 69 |
client = InferenceClient(
|
| 70 |
"meta-llama/Meta-Llama-3.1-8B-Instruct",
|
| 71 |
+
token= os.getenv("HF_API_TOKEN"),# Replace with your actual token
|
| 72 |
)
|
| 73 |
|
| 74 |
# Define a function to handle the chat input and get a response from the model
|