Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
|
@@ -4,7 +4,6 @@ from crew import run_crew
|
|
| 4 |
from utils import get_questions
|
| 5 |
|
| 6 |
QUESTION_FILE_PATH = "data/gaia_validation.jsonl"
|
| 7 |
-
QUESTION_LEVEL = 0
|
| 8 |
|
| 9 |
def _run(question, openai_api_key, gemini_api_key, file_name = ""):
|
| 10 |
"""
|
|
@@ -51,7 +50,9 @@ def _run(question, openai_api_key, gemini_api_key, file_name = ""):
|
|
| 51 |
|
| 52 |
gr.close_all()
|
| 53 |
|
| 54 |
-
|
|
|
|
|
|
|
| 55 |
|
| 56 |
with gr.Blocks() as gaia:
|
| 57 |
gr.Markdown("## General AI Assistant 🤖🤝🤖")
|
|
@@ -114,7 +115,21 @@ with gr.Blocks() as gaia:
|
|
| 114 |
)
|
| 115 |
|
| 116 |
gr.Examples(
|
| 117 |
-
examples=
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 118 |
inputs=[question, level, ground_truth, file_name, openai_api_key, gemini_api_key],
|
| 119 |
outputs=answer,
|
| 120 |
cache_examples=False
|
|
|
|
| 4 |
from utils import get_questions
|
| 5 |
|
| 6 |
QUESTION_FILE_PATH = "data/gaia_validation.jsonl"
|
|
|
|
| 7 |
|
| 8 |
def _run(question, openai_api_key, gemini_api_key, file_name = ""):
|
| 9 |
"""
|
|
|
|
| 50 |
|
| 51 |
gr.close_all()
|
| 52 |
|
| 53 |
+
examples_1 = get_questions(QUESTION_FILE_PATH, 1)
|
| 54 |
+
examples_2 = get_questions(QUESTION_FILE_PATH, 2)
|
| 55 |
+
examples_3 = get_questions(QUESTION_FILE_PATH, 3)
|
| 56 |
|
| 57 |
with gr.Blocks() as gaia:
|
| 58 |
gr.Markdown("## General AI Assistant 🤖🤝🤖")
|
|
|
|
| 115 |
)
|
| 116 |
|
| 117 |
gr.Examples(
|
| 118 |
+
examples=examples_1,
|
| 119 |
+
inputs=[question, level, ground_truth, file_name, openai_api_key, gemini_api_key],
|
| 120 |
+
outputs=answer,
|
| 121 |
+
cache_examples=False
|
| 122 |
+
)
|
| 123 |
+
|
| 124 |
+
gr.Examples(
|
| 125 |
+
examples=examples_2,
|
| 126 |
+
inputs=[question, level, ground_truth, file_name, openai_api_key, gemini_api_key],
|
| 127 |
+
outputs=answer,
|
| 128 |
+
cache_examples=False
|
| 129 |
+
)
|
| 130 |
+
|
| 131 |
+
gr.Examples(
|
| 132 |
+
examples=examples_3,
|
| 133 |
inputs=[question, level, ground_truth, file_name, openai_api_key, gemini_api_key],
|
| 134 |
outputs=answer,
|
| 135 |
cache_examples=False
|