Spaces:
Sleeping
Sleeping
Upload folder using huggingface_hub
Browse files
app.py
CHANGED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
import gradio as gr
|
| 2 |
from huggingface_hub import InferenceClient
|
| 3 |
|
| 4 |
-
# Use
|
| 5 |
-
client = InferenceClient("
|
| 6 |
|
| 7 |
def generate(prompt, temperature=0.8, max_tokens=256):
|
| 8 |
try:
|
|
|
|
| 1 |
import gradio as gr
|
| 2 |
from huggingface_hub import InferenceClient
|
| 3 |
|
| 4 |
+
# Use a model that supports text-generation
|
| 5 |
+
client = InferenceClient("meta-llama/Llama-3.2-1B-Instruct")
|
| 6 |
|
| 7 |
def generate(prompt, temperature=0.8, max_tokens=256):
|
| 8 |
try:
|