Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -2,7 +2,7 @@ from huggingface_hub import InferenceClient
|
|
| 2 |
import gradio as gr
|
| 3 |
import random
|
| 4 |
|
| 5 |
-
client = InferenceClient("
|
| 6 |
|
| 7 |
def format_prompt(message, history):
|
| 8 |
prompt = "<s>\n"
|
|
@@ -22,11 +22,11 @@ def generate(
|
|
| 22 |
|
| 23 |
generate_kwargs = dict(
|
| 24 |
temperature=temperature,
|
| 25 |
-
max_new_tokens=max_new_tokens,
|
| 26 |
top_p=top_p,
|
| 27 |
repetition_penalty=repetition_penalty,
|
| 28 |
do_sample=True,
|
| 29 |
-
seed=random.randint(0, 1000)
|
| 30 |
)
|
| 31 |
|
| 32 |
formatted_prompt = format_prompt(prompt, history)
|
|
|
|
| 2 |
import gradio as gr
|
| 3 |
import random
|
| 4 |
|
| 5 |
+
client = InferenceClient("THUDM/codegeex2-6b")
|
| 6 |
|
| 7 |
def format_prompt(message, history):
|
| 8 |
prompt = "<s>\n"
|
|
|
|
| 22 |
|
| 23 |
generate_kwargs = dict(
|
| 24 |
temperature=temperature,
|
| 25 |
+
# max_new_tokens=max_new_tokens,
|
| 26 |
top_p=top_p,
|
| 27 |
repetition_penalty=repetition_penalty,
|
| 28 |
do_sample=True,
|
| 29 |
+
# seed=random.randint(0, 1000)
|
| 30 |
)
|
| 31 |
|
| 32 |
formatted_prompt = format_prompt(prompt, history)
|