Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -14,7 +14,7 @@ else:
|
|
| 14 |
HF_TOKEN_ERROR = None
|
| 15 |
|
| 16 |
client = InferenceClient(token=HF_TOKEN)
|
| 17 |
-
PROMPT_IMPROVER_MODEL = "
|
| 18 |
|
| 19 |
def improve_prompt(original_prompt):
|
| 20 |
if HF_TOKEN_ERROR:
|
|
@@ -32,7 +32,7 @@ Improve this prompt: {original_prompt}
|
|
| 32 |
improved_prompt = client.text_generation(
|
| 33 |
prompt=prompt_for_llm,
|
| 34 |
model=PROMPT_IMPROVER_MODEL,
|
| 35 |
-
max_new_tokens=
|
| 36 |
temperature=0.7,
|
| 37 |
top_p=0.9,
|
| 38 |
repetition_penalty=1.2,
|
|
@@ -86,7 +86,7 @@ css = """
|
|
| 86 |
with gr.Blocks(css=css) as demo:
|
| 87 |
gr.Markdown(
|
| 88 |
"""
|
| 89 |
-
# Xylaria Iris
|
| 90 |
""",
|
| 91 |
elem_classes="title"
|
| 92 |
)
|
|
|
|
| 14 |
HF_TOKEN_ERROR = None
|
| 15 |
|
| 16 |
client = InferenceClient(token=HF_TOKEN)
|
| 17 |
+
PROMPT_IMPROVER_MODEL = "Qwen/Qwen2.5-Coder-32B-Instruct"
|
| 18 |
|
| 19 |
def improve_prompt(original_prompt):
|
| 20 |
if HF_TOKEN_ERROR:
|
|
|
|
| 32 |
improved_prompt = client.text_generation(
|
| 33 |
prompt=prompt_for_llm,
|
| 34 |
model=PROMPT_IMPROVER_MODEL,
|
| 35 |
+
max_new_tokens=1280,
|
| 36 |
temperature=0.7,
|
| 37 |
top_p=0.9,
|
| 38 |
repetition_penalty=1.2,
|
|
|
|
| 86 |
with gr.Blocks(css=css) as demo:
|
| 87 |
gr.Markdown(
|
| 88 |
"""
|
| 89 |
+
# Xylaria Iris v3
|
| 90 |
""",
|
| 91 |
elem_classes="title"
|
| 92 |
)
|