Update app.py
Browse files
app.py
CHANGED
|
@@ -95,7 +95,7 @@ def extract_pdf_md(pdf_dir):
|
|
| 95 |
return content
|
| 96 |
|
| 97 |
|
| 98 |
-
def openai_api(messages, model="claude-
|
| 99 |
print("use model:", model, "temperature:", temperature)
|
| 100 |
try:
|
| 101 |
completion = client.chat.completions.create(
|
|
@@ -160,7 +160,7 @@ Please pay attention to the pipe format as shown in the example below. This form
|
|
| 160 |
return response
|
| 161 |
|
| 162 |
|
| 163 |
-
def predict(prompt, file_content, model="claude-
|
| 164 |
file_content = del_references(file_content)
|
| 165 |
|
| 166 |
messages = [
|
|
@@ -383,7 +383,7 @@ with gr.Blocks(title="Automated Enzyme Kinetics Extractor") as demo:
|
|
| 383 |
exp = gr.Button("Example Prompt")
|
| 384 |
with gr.Row():
|
| 385 |
# 模型选择下拉菜单
|
| 386 |
-
model_choices = ["claude-
|
| 387 |
model_dropdown = gr.Dropdown(choices=model_choices, label="Select Model", value=model_choices[0])
|
| 388 |
|
| 389 |
# 温度选择滑块
|
|
|
|
| 95 |
return content
|
| 96 |
|
| 97 |
|
| 98 |
+
def openai_api(messages, model="claude-sonnet-4-5-20250929", temperature=0.1):
|
| 99 |
print("use model:", model, "temperature:", temperature)
|
| 100 |
try:
|
| 101 |
completion = client.chat.completions.create(
|
|
|
|
| 160 |
return response
|
| 161 |
|
| 162 |
|
| 163 |
+
def predict(prompt, file_content, model="claude-sonnet-4-5-20250929", temperature=0.1):
|
| 164 |
file_content = del_references(file_content)
|
| 165 |
|
| 166 |
messages = [
|
|
|
|
| 383 |
exp = gr.Button("Example Prompt")
|
| 384 |
with gr.Row():
|
| 385 |
# 模型选择下拉菜单
|
| 386 |
+
model_choices = ["claude-sonnet-4-5-20250929", "gpt-4o-2024-08-06"]
|
| 387 |
model_dropdown = gr.Dropdown(choices=model_choices, label="Select Model", value=model_choices[0])
|
| 388 |
|
| 389 |
# 温度选择滑块
|