File size: 5,862 Bytes
10e9b7d
527c93a
 
7ca40c9
 
527c93a
7ca40c9
 
527c93a
 
7ca40c9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
527c93a
 
10e9b7d
eccf8e4
7d65c66
3c4371f
b158950
10e9b7d
a786c3f
0d17c63
81f2b27
e910739
81f2b27
e80aab9
3db6293
e80aab9
a786c3f
 
7e4a06b
a786c3f
3c4371f
7e4a06b
3c4371f
7d65c66
3c4371f
7e4a06b
31243f4
 
36ed51a
3c4371f
a786c3f
31243f4
eccf8e4
31243f4
7d65c66
31243f4
 
a786c3f
7d65c66
a786c3f
e80aab9
a786c3f
7d65c66
 
41efa04
 
18d2418
 
31243f4
 
 
 
 
a786c3f
7d65c66
 
31243f4
a786c3f
b158950
 
 
41efa04
b158950
 
 
 
 
 
 
 
31243f4
 
 
7d65c66
31243f4
e80aab9
7d65c66
e80aab9
 
31243f4
e80aab9
 
3c4371f
 
a786c3f
 
7d65c66
a786c3f
e80aab9
a786c3f
e80aab9
31243f4
0ee0419
e514fd7
 
a786c3f
 
 
e514fd7
e80aab9
 
7e4a06b
31243f4
9088b99
7d65c66
e80aab9
a786c3f
e80aab9
 
0d17c63
 
 
 
 
31243f4
a786c3f
0d17c63
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
import os
import nltk

# 1) Asegurarnos de que la carpeta nltk_data exista
os.makedirs("nltk_data", exist_ok=True)

# 2) Apuntar NLTK a usar esa carpeta
os.environ["NLTK_DATA"] = "nltk_data"
nltk.data.path.insert(0, "nltk_data")

# 3) Verificar si 'punkt' y 'stopwords' ya están descargados; si no, descargarlos ahora.
#    Así evitamos el error y hacemos la descarga solo una vez en runtime.
def ensure_nltk_resource(res_name):
    try:
        nltk.data.find(res_name)
        return True
    except LookupError:
        return False

# Si falta 'punkt', lo descargamos
if not ensure_nltk_resource("tokenizers/punkt"):
    print("DEBUG: 'punkt' no encontrado en nltk_data, descargando...")
    nltk.download("punkt", download_dir="nltk_data", quiet=True)

# Si falta 'stopwords', lo descargamos
if not ensure_nltk_resource("corpora/stopwords"):
    print("DEBUG: 'stopwords' no encontrado en nltk_data, descargando...")
    nltk.download("stopwords", download_dir="nltk_data", quiet=True)

# ── Ahora SÍ traemos el resto de librerías que usan nltk y llama_index  ──────
import gradio as gr
import requests
import inspect
import pandas as pd
import time

# --- AGENTE SIMPLIFICADO CON FUNCIONALIDAD PERSONALIZADA ---
from my_tools import basic_agent_response



# --- Constants ---
DEFAULT_API_URL = "https://agents-course-unit4-scoring.hf.space"

def run_and_submit_all(profile: gr.OAuthProfile | None):
    space_id = os.getenv("SPACE_ID")
    if profile:
        username = f"{profile.username}"
        print(f"User logged in: {username}")
    else:
        print("User not logged in.")
        return "Please Login to Hugging Face with the button.", None

    api_url = DEFAULT_API_URL
    questions_url = f"{api_url}/questions"
    submit_url = f"{api_url}/submit"
    agent_code = f"https://huggingface.co/spaces/{space_id}/tree/main"

    # --- Fetch Questions ---
    print(f"Fetching questions from: {questions_url}")
    try:
        response = requests.get(questions_url, timeout=15)
        response.raise_for_status()
        questions_data = response.json()
        if not questions_data:
            return "Fetched questions list is empty or invalid format.", None
    except Exception as e:
        return f"Error fetching questions: {e}", None

    # --- Run Agent ---
    results_log = []
    answers_payload = []

    total_questions = len(questions_data)
    #for item in questions_data:
    for item_index, item in enumerate(questions_data):
        task_id = item.get("task_id")
        question_text = item.get("question")
        if not task_id or question_text is None:
            continue
        try:
            submitted_answer = basic_agent_response(question_text)
            answers_payload.append({"task_id": task_id, "submitted_answer": submitted_answer})
            results_log.append({"Task ID": task_id, "Question": question_text, "Submitted Answer": submitted_answer})
        except Exception as e:
            results_log.append({"Task ID": task_id, "Question": question_text, "Submitted Answer": f"AGENT ERROR: {e}"})
        
        # ----- LÍNEA DE TIME.SLEEP VA AQUÍ -----
        # Esperar después de procesar cada pregunta, excepto la última
        # item_index y total_questions AHORA ESTÁN DEFINIDOS CORRECTAMENTE en este alcance
        if item_index < total_questions - 1:
            wait_duration = 20  # Segundos de espera
            print(f"--- Question {item_index + 1} processed. Waiting {wait_duration} seconds before next question to manage API rate limits... ---")
            time.sleep(wait_duration)
        else:
            print(f"--- All {total_questions} questions processed. Proceeding to submission. ---")
        # ----- FIN DE LA SECCIÓN DE TIME.SLEEP -----
        
    if not answers_payload:
        return "Agent did not produce any answers to submit.", pd.DataFrame(results_log)

    submission_data = {"username": username.strip(), "agent_code": agent_code, "answers": answers_payload}
    print(f"Submitting {len(answers_payload)} answers to: {submit_url}")
    try:
        response = requests.post(submit_url, json=submission_data, timeout=60)
        response.raise_for_status()
        result_data = response.json()
        final_status = (
            f"Submission Successful!\n"
            f"User: {result_data.get('username')}\n"
            f"Overall Score: {result_data.get('score', 'N/A')}% "
            f"({result_data.get('correct_count', '?')}/{result_data.get('total_attempted', '?')} correct)\n"
            f"Message: {result_data.get('message', 'No message received.')}")
        return final_status, pd.DataFrame(results_log)
    except Exception as e:
        return f"Submission Failed: {e}", pd.DataFrame(results_log)

# --- Interfaz Gradio ---
with gr.Blocks() as demo:
    gr.Markdown("# Basic Agent Evaluation Runner")
    gr.Markdown(
        """
        **Instructions:**
        1. Clone this space and modify the code to define your agent's logic and tools.
        2. Log in to Hugging Face with the button below.
        3. Click 'Run Evaluation & Submit All Answers' to evaluate your agent.
        """
    )

    gr.LoginButton()
    run_button = gr.Button("Run Evaluation & Submit All Answers")
    status_output = gr.Textbox(label="Run Status / Submission Result", lines=5, interactive=False)
    results_table = gr.DataFrame(label="Questions and Agent Answers", wrap=True)

    run_button.click(fn=run_and_submit_all, outputs=[status_output, results_table])

if __name__ == "__main__":
    print("🔍 Prueba de pregunta GAIA manual")
    test_question = "¿Cuánto es 37 por 19?"
    print("Pregunta:", test_question)
    print("Respuesta del agente:", basic_agent_response(test_question))
    
    print("Launching Gradio Interface for Basic Agent Evaluation...")
    demo.launch(debug=True, share=False)