jackkuo commited on
Commit
4bf2c6c
·
verified ·
1 Parent(s): 6b73056

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -3
app.py CHANGED
@@ -95,7 +95,7 @@ def extract_pdf_md(pdf_dir):
95
  return content
96
 
97
 
98
- def openai_api(messages, model="claude-3-5-sonnet-20240620", temperature=0.1):
99
  print("use model:", model, "temperature:", temperature)
100
  try:
101
  completion = client.chat.completions.create(
@@ -160,7 +160,7 @@ Please pay attention to the pipe format as shown in the example below. This form
160
  return response
161
 
162
 
163
- def predict(prompt, file_content, model="claude-3-5-sonnet-20240620", temperature=0.1):
164
  file_content = del_references(file_content)
165
 
166
  messages = [
@@ -383,7 +383,7 @@ with gr.Blocks(title="Automated Enzyme Kinetics Extractor") as demo:
383
  exp = gr.Button("Example Prompt")
384
  with gr.Row():
385
  # 模型选择下拉菜单
386
- model_choices = ["claude-3-5-sonnet-20240620", "gpt-4o-2024-08-06"]
387
  model_dropdown = gr.Dropdown(choices=model_choices, label="Select Model", value=model_choices[0])
388
 
389
  # 温度选择滑块
 
95
  return content
96
 
97
 
98
+ def openai_api(messages, model="claude-sonnet-4-5-20250929", temperature=0.1):
99
  print("use model:", model, "temperature:", temperature)
100
  try:
101
  completion = client.chat.completions.create(
 
160
  return response
161
 
162
 
163
+ def predict(prompt, file_content, model="claude-sonnet-4-5-20250929", temperature=0.1):
164
  file_content = del_references(file_content)
165
 
166
  messages = [
 
383
  exp = gr.Button("Example Prompt")
384
  with gr.Row():
385
  # 模型选择下拉菜单
386
+ model_choices = ["claude-sonnet-4-5-20250929", "gpt-4o-2024-08-06"]
387
  model_dropdown = gr.Dropdown(choices=model_choices, label="Select Model", value=model_choices[0])
388
 
389
  # 温度选择滑块