{ "cells": [ { "cell_type": "code", "execution_count": 1, "id": "2a8e89d4", "metadata": {}, "outputs": [ { "name": "stderr", "output_type": "stream", "text": [ "/gpfs/radev/home/tl688/.conda/envs/evoagentx/lib/python3.11/site-packages/PyPDF2/__init__.py:21: DeprecationWarning: PyPDF2 is deprecated. Please move to the pypdf library instead.\n", " warnings.warn(\n" ] } ], "source": [ "import os\n", "\n", "from dotenv import load_dotenv\n", "\n", "from evoagentx.agents.agent_manager import AgentManager\n", "from evoagentx.benchmark import HotPotQA\n", "from evoagentx.core.callbacks import suppress_logger_info\n", "from evoagentx.core.logging import logger\n", "from evoagentx.evaluators import Evaluator\n", "from evoagentx.models import OpenAILLM, OpenAILLMConfig\n", "from evoagentx.optimizers import TextGradOptimizer\n", "from evoagentx.prompts import StringTemplate\n", "from evoagentx.workflow import SequentialWorkFlowGraph\n", "from dotenv import load_dotenv\n", "\n", "from evoagentx.agents.agent_manager import AgentManager\n", "from evoagentx.benchmark import MBPP\n", "from evoagentx.core.callbacks import suppress_logger_info\n", "from evoagentx.core.logging import logger\n", "from evoagentx.evaluators import Evaluator\n", "from evoagentx.models import OpenAILLM, OpenAILLMConfig\n", "from evoagentx.optimizers import TextGradOptimizer\n", "from evoagentx.prompts import StringTemplate\n", "from evoagentx.workflow import SequentialWorkFlowGraph\n", "\n", "from evoagentx.models import OpenAILLMConfig, OpenAILLM\n", "from evoagentx.workflow import SEWWorkFlowGraph, STRUCTUREWorkFlowGraph\n", "from evoagentx.agents import AgentManager\n", "from evoagentx.benchmark import HumanEval,AFlowMBPP\n", "from evoagentx.evaluators import Evaluator \n", "from evoagentx.optimizers import SEWOptimizer, STRUCTUREOptimizer\n", "from evoagentx.optimizers.structure_optimizer import STRUCTUREWorkFlowScheme\n", "from evoagentx.core.callbacks import suppress_logger_info\n", "\n", "from evoagentx.models import OpenAILLMConfig, OpenAILLM,AzureOpenAIConfig,LiteLLMConfig,LiteLLM\n", "from evoagentx.workflow import SEWWorkFlowGraph \n", "from evoagentx.agents import AgentManager\n", "from evoagentx.benchmark import MBPPPLUS, AFlowMBPPPLUS\n", "from evoagentx.evaluators import Evaluator \n", "from evoagentx.optimizers import SEWOptimizer \n", "from evoagentx.core.callbacks import suppress_logger_info\n", "from evoagentx.benchmark import HumanEvalPLUS\n", "from evoagentx.benchmark import SciCode\n", "from evoagentx.benchmark import PubMedQA\n", "from copy import deepcopy\n", "\n", "import nest_asyncio\n", "nest_asyncio.apply()\n", "\n", "class PubMedQASplits(PubMedQA):\n", "\n", " def _load_data(self):\n", " # load the original test data \n", " super()._load_data()\n", " # split the data into train, dev and test\n", " import numpy as np \n", " np.random.seed(42)\n", " permutation = np.random.permutation(len(self._dev_data))\n", " full_test_data = self._dev_data \n", " # randomly select 10 samples for train, 40 for dev, and 100 for test\n", " self._train_data = [full_test_data[idx] for idx in permutation[:50]]\n", " self._dev_data = [full_test_data[idx] for idx in permutation[:50]]\n", " self._test_data =self._test_data[0:500]\n", " self._fulldata = full_test_data\n", "\n", "\n", "def collate_func(example: dict) -> dict:\n", " context_list = []\n", " paragraphs = example[\"context\"][\"contexts\"]\n", " context = \"\\n\".join(paragraphs)\n", " problem = \"Context: {}\\n\\nQuestion: {}\\n\\nAnswer:\".format(context, example[\"question\"])\n", " return {\"problem\": problem}\n", "\n", "\n", "hotpotqa_graph_data = {\n", " \"goal\": \"Answer the question based on the context. The answer should be a direct response to the question, without including explanations or reasoning.\",\n", " \"tasks\": [\n", " {\n", " \"name\": \"answer_generate\",\n", " \"description\": \"Answer the question based on the context.\",\n", " \"inputs\": [\n", " {\"name\": \"problem\", \"type\": \"str\", \"required\": True, \"description\": \"The problem to solve.\"}\n", " ],\n", " \"outputs\": [\n", " {\"name\": \"answer\", \"type\": \"str\", \"required\": True, \"description\": \"The answer to the problem.\"}\n", " ],\n", " \"prompt_template\": StringTemplate(instruction=\"Think step by step to answer the question. You should explain your thinking process in the 'thought' field, and provide the final answer in the 'answer' field.\\nFormat your output in xml format, such as xxx and xxx.\"),\n", " \"parse_mode\": \"xml\"\n", " }\n", " ] \n", "}\n", "\n", "os.environ[\"AZURE_OPENAI_DEPLOYMENT_NAME\"] = \"gpt-4o-mini\"\n", "os.environ[\"AZURE_OPENAI_ENDPOINT\"] = \"https://tianyuliu-hua-raredisea-resource.cognitiveservices.azure.com/\"\n", "os.environ[\"AZURE_OPENAI_KEY\"] = \"2pa9h2ZIN1lQepFWwYADlXIKIansa9KPhxMoumeGbRQ08f2uDTXiJQQJ99BKACHYHv6XJ3w3AAAAACOGsQIt\"\n", "os.environ[\"AZURE_OPENAI_API_VERSION\"] = \"2025-01-01-preview\"\n", "llm_config = LiteLLMConfig(model=\"azure/\" + os.getenv(\"AZURE_OPENAI_DEPLOYMENT_NAME\"), # Azure model format\n", " azure_endpoint=os.getenv(\"AZURE_OPENAI_ENDPOINT\"),\n", " azure_key=os.getenv(\"AZURE_OPENAI_KEY\"),\n", " api_version=os.getenv(\"AZURE_OPENAI_API_VERSION\", \"2024-12-01-preview\"), top_p=0.85, temperature=0.2, frequency_penalty=0.0, presence_penalty=0.0)\n", "\n", "executor_llm = LiteLLM(config=llm_config)\n", "optimizer_llm = LiteLLM(config=llm_config)" ] }, { "cell_type": "code", "execution_count": 2, "id": "20e078fa", "metadata": {}, "outputs": [ { "name": "stderr", "output_type": "stream", "text": [ "\u001b[32m2025-12-21 11:35:25.188\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.benchmark.pubmedqa\u001b[0m:\u001b[36m_load_data_from_file\u001b[0m:\u001b[36m51\u001b[0m - \u001b[1mloading HotPotQA data from /gpfs/radev/pi/ying_rex/tl688/selfevolve/EvoAgentX/examples/pubmedqa/pubmedqa_train.json ...\u001b[0m\n", "\u001b[32m2025-12-21 11:35:27.770\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.benchmark.pubmedqa\u001b[0m:\u001b[36m_load_data_from_file\u001b[0m:\u001b[36m51\u001b[0m - \u001b[1mloading HotPotQA data from /gpfs/radev/pi/ying_rex/tl688/selfevolve/EvoAgentX/examples/pubmedqa/pubmedqa_train.json ...\u001b[0m\n", "\u001b[32m2025-12-21 11:35:30.442\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.benchmark.pubmedqa\u001b[0m:\u001b[36m_load_data_from_file\u001b[0m:\u001b[36m51\u001b[0m - \u001b[1mloading HotPotQA data from /gpfs/radev/pi/ying_rex/tl688/selfevolve/EvoAgentX/examples/pubmedqa/pubmedqa_label.json ...\u001b[0m\n" ] } ], "source": [ "benchmark = PubMedQASplits()\n", "workflow_graph = SequentialWorkFlowGraph.from_dict(hotpotqa_graph_data)\n", "agent_manager = AgentManager()\n", "agent_manager.add_agents_from_workflow(workflow_graph, executor_llm.config)\n", "\n", "evaluator = Evaluator(\n", " llm=executor_llm, \n", " agent_manager=agent_manager, \n", " collate_func=collate_func, \n", " num_workers=20, \n", " verbose=True\n", ")\n", "\n", "textgrad_optimizer = TextGradOptimizer(\n", " graph=workflow_graph, \n", " optimize_mode=\"all\",\n", " executor_llm=executor_llm, \n", " optimizer_llm=optimizer_llm,\n", " batch_size=3,\n", " max_steps=20,\n", " evaluator=evaluator,\n", " eval_every_n_steps=1,\n", " eval_rounds=1,\n", " save_interval=None,\n", " save_path=\"./\",\n", " rollback=True,\n", " constraints=[]\n", ")\n" ] }, { "cell_type": "code", "execution_count": 6, "id": "78d5904e", "metadata": { "scrolled": true }, "outputs": [ { "data": { "text/plain": [ "500" ] }, "execution_count": 6, "metadata": {}, "output_type": "execute_result" } ], "source": [ "len(benchmark._test_data)" ] }, { "cell_type": "code", "execution_count": 7, "id": "d686ee20", "metadata": { "scrolled": true }, "outputs": [ { "name": "stderr", "output_type": "stream", "text": [ "\u001b[32m2025-12-21 11:35:45.504\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36m__main__\u001b[0m:\u001b[36m\u001b[0m:\u001b[36m1\u001b[0m - \u001b[1mEvaluating workflow on test set...\u001b[0m\n", "Evaluating workflow: 0%| | 1/500 [00:02<17:10, 2.07s/it]Task exception was never retrieved\n", "future: exception=RuntimeError('Event loop is closed')>\n", "Traceback (most recent call last):\n", " File \"/gpfs/radev/home/tl688/.conda/envs/evoagentx/lib/python3.11/asyncio/tasks.py\", line 277, in __step\n", " result = coro.send(None)\n", " ^^^^^^^^^^^^^^^\n", " File \"/gpfs/radev/home/tl688/.conda/envs/evoagentx/lib/python3.11/site-packages/litellm/utils.py\", line 873, in _client_async_logging_helper\n", " GLOBAL_LOGGING_WORKER.ensure_initialized_and_enqueue(\n", " File \"/gpfs/radev/home/tl688/.conda/envs/evoagentx/lib/python3.11/site-packages/litellm/litellm_core_utils/logging_worker.py\", line 322, in ensure_initialized_and_enqueue\n", " self.enqueue(async_coroutine)\n", " File \"/gpfs/radev/home/tl688/.conda/envs/evoagentx/lib/python3.11/site-packages/litellm/litellm_core_utils/logging_worker.py\", line 131, in enqueue\n", " self._queue.put_nowait(task)\n", " File \"/gpfs/radev/home/tl688/.conda/envs/evoagentx/lib/python3.11/asyncio/queues.py\", line 147, in put_nowait\n", " self._wakeup_next(self._getters)\n", " File \"/gpfs/radev/home/tl688/.conda/envs/evoagentx/lib/python3.11/asyncio/queues.py\", line 63, in _wakeup_next\n", " waiter.set_result(None)\n", " File \"/gpfs/radev/home/tl688/.conda/envs/evoagentx/lib/python3.11/asyncio/futures.py\", line 263, in set_result\n", " self.__schedule_callbacks()\n", " File \"/gpfs/radev/home/tl688/.conda/envs/evoagentx/lib/python3.11/asyncio/futures.py\", line 173, in __schedule_callbacks\n", " self._loop.call_soon(callback, self, context=ctx)\n", " File \"/gpfs/radev/home/tl688/.conda/envs/evoagentx/lib/python3.11/asyncio/base_events.py\", line 762, in call_soon\n", " self._check_closed()\n", " File \"/gpfs/radev/home/tl688/.conda/envs/evoagentx/lib/python3.11/asyncio/base_events.py\", line 520, in _check_closed\n", " raise RuntimeError('Event loop is closed')\n", "RuntimeError: Event loop is closed\n", "Evaluating workflow: 0%| | 2/500 [00:02<07:44, 1.07it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 1%| | 6/500 [00:02<02:07, 3.87it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 2%|▏ | 9/500 [00:02<01:22, 5.96it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 2%|▏ | 11/500 [00:03<01:14, 6.59it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 3%|▎ | 13/500 [00:03<01:24, 5.76it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 3%|▎ | 15/500 [00:04<01:39, 4.87it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 3%|▎ | 17/500 [00:05<02:11, 3.67it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 4%|▍ | 21/500 [00:05<01:11, 6.73it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 5%|▍ | 24/500 [00:05<00:53, 8.85it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 5%|▌ | 26/500 [00:05<00:45, 10.48it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 6%|▌ | 29/500 [00:07<01:55, 4.07it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 6%|▌ | 31/500 [00:07<01:30, 5.19it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 7%|▋ | 37/500 [00:07<00:45, 10.10it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 8%|▊ | 41/500 [00:07<00:45, 10.19it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 9%|▊ | 43/500 [00:08<00:45, 9.94it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 9%|▉ | 45/500 [00:08<01:01, 7.38it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 10%|▉ | 49/500 [00:09<01:20, 5.61it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 10%|█ | 52/500 [00:10<01:12, 6.15it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 12%|█▏ | 58/500 [00:10<00:51, 8.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 12%|█▏ | 60/500 [00:10<00:48, 9.15it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 12%|█▏ | 62/500 [00:11<00:59, 7.30it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 13%|█▎ | 63/500 [00:11<01:01, 7.12it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 13%|█▎ | 64/500 [00:11<01:16, 5.71it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 14%|█▍ | 70/500 [00:12<00:51, 8.43it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 14%|█▍ | 72/500 [00:12<00:45, 9.31it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 15%|█▍ | 74/500 [00:12<00:52, 8.19it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 16%|█▌ | 79/500 [00:13<00:51, 8.14it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 16%|█▌ | 81/500 [00:13<00:53, 7.76it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 17%|█▋ | 83/500 [00:14<00:58, 7.10it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 17%|█▋ | 85/500 [00:14<01:02, 6.61it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 17%|█▋ | 87/500 [00:14<01:09, 5.90it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 18%|█▊ | 89/500 [00:15<01:04, 6.34it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 19%|█▉ | 94/500 [00:15<00:33, 11.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 19%|█▉ | 96/500 [00:15<00:30, 13.45it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 20%|█▉ | 98/500 [00:16<00:50, 7.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 20%|██ | 100/500 [00:16<00:51, 7.75it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 20%|██ | 102/500 [00:16<00:57, 6.98it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 21%|██ | 103/500 [00:17<01:20, 4.96it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 21%|██ | 106/500 [00:17<01:07, 5.86it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 22%|██▏ | 108/500 [00:17<00:53, 7.36it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 22%|██▏ | 110/500 [00:17<00:47, 8.26it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 22%|██▏ | 112/500 [00:18<00:45, 8.59it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 23%|██▎ | 114/500 [00:18<00:56, 6.85it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 23%|██▎ | 116/500 [00:18<00:51, 7.50it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 24%|██▎ | 118/500 [00:18<00:46, 8.16it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 24%|██▍ | 120/500 [00:19<01:18, 4.86it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 25%|██▍ | 123/500 [00:20<00:55, 6.82it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 25%|██▌ | 125/500 [00:20<00:42, 8.85it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 25%|██▌ | 127/500 [00:20<00:46, 8.03it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 26%|██▌ | 131/500 [00:20<00:38, 9.48it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 27%|██▋ | 133/500 [00:21<00:54, 6.68it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 27%|██▋ | 136/500 [00:21<01:00, 6.03it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 28%|██▊ | 138/500 [00:22<00:52, 6.88it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 28%|██▊ | 140/500 [00:22<01:01, 5.84it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 29%|██▉ | 144/500 [00:22<00:33, 10.62it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 29%|██▉ | 146/500 [00:22<00:38, 9.21it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 30%|██▉ | 148/500 [00:23<00:43, 8.17it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 30%|███ | 152/500 [00:23<00:42, 8.17it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 31%|███ | 154/500 [00:24<00:53, 6.41it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 31%|███ | 155/500 [00:24<00:51, 6.69it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 32%|███▏ | 158/500 [00:24<00:48, 7.05it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 32%|███▏ | 160/500 [00:25<00:45, 7.44it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 32%|███▏ | 161/500 [00:25<00:54, 6.21it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 32%|███▏ | 162/500 [00:25<01:01, 5.47it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 33%|███▎ | 165/500 [00:26<01:10, 4.76it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 33%|███▎ | 167/500 [00:26<00:51, 6.48it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 34%|███▎ | 168/500 [00:26<00:52, 6.35it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 34%|███▍ | 171/500 [00:27<00:45, 7.21it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 34%|███▍ | 172/500 [00:27<00:43, 7.48it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 35%|███▌ | 177/500 [00:27<00:32, 10.09it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 36%|███▌ | 179/500 [00:27<00:35, 8.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 36%|███▌ | 181/500 [01:00<26:57, 5.07s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 37%|███▋ | 184/500 [01:01<14:50, 2.82s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 38%|███▊ | 191/500 [01:01<04:59, 1.03it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 39%|███▉ | 194/500 [01:01<03:24, 1.49it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2025-12-21 11:36:47.778\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 22720085, 'question': 'Does insulin resistance drive the association between hyperglycemia and cardiovascular risk?', 'context': {'contexts': ['Several studies have shown associations between hyperglycemia and risk of cardiovascular disease (CVD) and mortality, yet glucose-lowering treatment does little to mitigate this risk. We examined whether associations between hyperglycemia and CVD risk were explained by underlying insulin resistance.', 'In 60 middle-aged individuals without diabetes we studied the associations of fasting plasma glucose, 2-hour post oral glucose tolerance test plasma glucose, insulin sensitivity as well as body fat percentage with CVD risk. Insulin sensitivity was measured as the glucose infusion rate during a euglycemic hyperinsulinemic clamp, body fat percentage was measured by dual X-ray absorptiometry, and CVD risk was estimated using the Framingham risk score. Associations of fasting plasma glucose, 2-hour plasma glucose, insulin sensitivity and body fat percentage with the Framingham risk score were assessed in linear regression models.', 'Both fasting and 2-hour plasma glucose levels were associated with higher Framingham risk score (fasting glucose: r(2) = 0.21; 2-hour glucose: r(2) = 0.24; P<0.001 for both), and insulin sensitivity with lower Framingham risk score (r(2) = 0.36; P<0.001). However, adjustment for insulin sensitivity and 2-hour glucose made the effect of fasting glucose non-significant (P = 0.060). Likewise, when adjusting for insulin sensitivity and fasting glucose, the association between 2-hour glucose and Framingham risk score disappeared (P = 0.143). In contrast, insulin sensitivity was still associated with Framingham risk score after adjusting for glucose levels (P<0.001). Body fat was not associated with Framingham risk score when taking insulin sensitivity into account (P = 0.550).'], 'labels': ['BACKGROUND', 'METHODS', 'RESULTS'], 'meshes': ['Adult', 'Blood Glucose', 'Cardiovascular Diseases', 'Female', 'Humans', 'Hyperglycemia', 'Insulin Resistance', 'Male', 'Middle Aged', 'Risk Factors'], 'reasoning_required_pred': ['y', 'e', 's'], 'reasoning_free_pred': ['y', 'e', 's']}, 'long_answer': 'The association between plasma glucose levels and CVD risk is mainly explained by insulin resistance, which raises the question of whether glucose lowering per se without changes in the processes that underlie hyperglycemia should be the sole clinical paradigm in the treatment of type 2 diabetes or its prevention.', 'final_decision': 'yes'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2025-12-21 11:36:47.798\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 22211919, 'question': 'Actinobaculum schaalii, a cause of urinary tract infections in children?', 'context': {'contexts': ['Urine samples were examined by wet smear microscopy, incubated in 5% CO(2) for 1-2 days, and species-specific real-time polymerase chain reaction (PCR) for A. schaalii was performed.', 'In 5 of the 29 screened urines, A. schaalii was found only by real-time PCR in quantities equivalent to ≥ 10(4) -10(5) CFU/mL. In addition, A. schaalii was found in quantities equivalent to ≥ 10(6) CFU/mL by both culture and PCR in two children with a urinary tract infection and large numbers of leucocytes in the urine.'], 'labels': ['METHODS', 'RESULTS'], 'meshes': ['Adolescent', 'Child, Preschool', 'Female', 'Gram-Positive Bacteria', 'Gram-Positive Bacterial Infections', 'Humans', 'Infant', 'Male', 'Retrospective Studies', 'Urinary Tract Infections'], 'reasoning_required_pred': ['y', 'e', 's'], 'reasoning_free_pred': ['m', 'a', 'y', 'b', 'e']}, 'long_answer': 'Actinobaculum schaalii is CO(2)-dependent. Therefore, if there are clinical symptoms and/or a negative culture despite the presence of leucocytes in the urine, Gram staining and incubation in 5% CO(2) or species-specific real-time PCR should be performed to identify A. schaalii.', 'final_decision': 'maybe'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 40%|███▉ | 198/500 [01:02<02:32, 1.98it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 40%|███▉ | 199/500 [01:04<03:11, 1.57it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 40%|████ | 200/500 [01:04<02:52, 1.74it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 41%|████ | 203/500 [01:04<01:45, 2.82it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 41%|████ | 205/500 [01:04<01:24, 3.48it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 41%|████▏ | 207/500 [01:05<01:17, 3.76it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 42%|████▏ | 208/500 [01:05<01:19, 3.66it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 43%|████▎ | 213/500 [01:06<00:49, 5.79it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 43%|████▎ | 217/500 [01:06<00:47, 5.96it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 44%|████▍ | 219/500 [01:07<00:50, 5.55it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 44%|████▍ | 221/500 [01:07<00:44, 6.32it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 44%|████▍ | 222/500 [01:07<00:47, 5.88it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 45%|████▌ | 225/500 [01:08<00:48, 5.67it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 45%|████▌ | 226/500 [01:08<00:47, 5.75it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 46%|████▌ | 228/500 [01:08<00:43, 6.32it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 46%|████▌ | 231/500 [01:09<00:37, 7.26it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 47%|████▋ | 234/500 [01:09<00:34, 7.64it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 47%|████▋ | 236/500 [01:09<00:27, 9.45it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 48%|████▊ | 238/500 [01:10<00:40, 6.55it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 48%|████▊ | 239/500 [01:10<00:45, 5.75it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 49%|████▊ | 243/500 [01:10<00:33, 7.75it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 49%|████▉ | 244/500 [01:10<00:39, 6.46it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 49%|████▉ | 246/500 [01:11<00:36, 7.03it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 50%|████▉ | 248/500 [01:11<00:37, 6.74it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 50%|█████ | 251/500 [01:12<00:40, 6.11it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 51%|█████ | 254/500 [01:12<00:29, 8.29it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 51%|█████ | 255/500 [01:12<00:30, 8.10it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 51%|█████▏ | 257/500 [01:12<00:30, 8.08it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 52%|█████▏ | 259/500 [01:13<00:33, 7.16it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 53%|█████▎ | 265/500 [01:13<00:25, 9.15it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 54%|█████▍ | 269/500 [01:14<00:33, 6.82it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 54%|█████▍ | 271/500 [01:14<00:28, 8.05it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 55%|█████▍ | 273/500 [01:14<00:32, 7.00it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 55%|█████▌ | 275/500 [01:15<00:36, 6.14it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 56%|█████▌ | 278/500 [01:15<00:29, 7.63it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 56%|█████▌ | 280/500 [01:15<00:23, 9.37it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 56%|█████▋ | 282/500 [01:16<00:28, 7.59it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 57%|█████▋ | 284/500 [01:16<00:35, 6.11it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 57%|█████▋ | 286/500 [01:16<00:27, 7.87it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 57%|█████▋ | 287/500 [01:16<00:27, 7.65it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 58%|█████▊ | 290/500 [01:17<00:24, 8.70it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 58%|█████▊ | 292/500 [01:17<00:23, 8.72it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 59%|█████▉ | 294/500 [01:17<00:27, 7.57it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 59%|█████▉ | 296/500 [01:18<00:32, 6.27it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 60%|█████▉ | 298/500 [01:18<00:28, 6.99it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 60%|██████ | 300/500 [01:19<00:41, 4.78it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 61%|██████ | 304/500 [01:19<00:27, 7.03it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 61%|██████ | 306/500 [01:19<00:24, 7.78it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 62%|██████▏ | 312/500 [01:20<00:21, 8.60it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 63%|██████▎ | 315/500 [01:20<00:16, 10.90it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 63%|██████▎ | 317/500 [01:21<00:32, 5.68it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 64%|██████▍ | 319/500 [01:21<00:31, 5.71it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 64%|██████▍ | 322/500 [01:22<00:32, 5.54it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 65%|██████▌ | 325/500 [01:22<00:22, 7.74it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 65%|██████▌ | 327/500 [01:22<00:18, 9.31it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 66%|██████▌ | 329/500 [01:23<00:21, 8.10it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 66%|██████▌ | 331/500 [01:23<00:36, 4.60it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 67%|██████▋ | 333/500 [01:24<00:33, 4.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 67%|██████▋ | 334/500 [01:24<00:34, 4.80it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 67%|██████▋ | 336/500 [01:24<00:28, 5.83it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 68%|██████▊ | 338/500 [01:25<00:28, 5.64it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 68%|██████▊ | 341/500 [01:25<00:24, 6.41it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 68%|██████▊ | 342/500 [01:25<00:23, 6.82it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 69%|██████▉ | 344/500 [01:26<00:26, 5.83it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 69%|██████▉ | 347/500 [01:26<00:27, 5.59it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 70%|███████ | 350/500 [01:27<00:21, 7.08it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 71%|███████ | 353/500 [01:27<00:19, 7.68it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 71%|███████ | 355/500 [01:28<00:27, 5.24it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 71%|███████ | 356/500 [01:28<00:25, 5.73it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 71%|███████▏ | 357/500 [01:28<00:29, 4.86it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 72%|███████▏ | 359/500 [01:28<00:24, 5.72it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 72%|███████▏ | 361/500 [01:28<00:21, 6.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 73%|███████▎ | 364/500 [01:29<00:19, 7.07it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 73%|███████▎ | 366/500 [01:29<00:20, 6.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 73%|███████▎ | 367/500 [01:29<00:18, 7.07it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 74%|███████▍ | 369/500 [01:30<00:18, 6.94it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 74%|███████▍ | 370/500 [01:30<00:31, 4.07it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 75%|███████▍ | 374/500 [02:01<07:54, 3.76s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 76%|███████▌ | 379/500 [02:01<03:02, 1.51s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 76%|███████▌ | 381/500 [02:01<02:11, 1.10s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 77%|███████▋ | 383/500 [02:01<01:35, 1.22it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2025-12-21 11:37:47.947\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 20684175, 'question': 'Vitamin D supplementation and regulatory T cells in apparently healthy subjects: vitamin D treatment for autoimmune diseases?', 'context': {'contexts': ['Epidemiological data show significant associations of vitamin D deficiency and autoimmune diseases. Vitamin D may prevent autoimmunity by stimulating naturally occurring regulatory T cells.', 'To elucidate whether vitamin D supplementation increases Tregs frequency (%Tregs) within circulating CD4+ T cells.', 'We performed an uncontrolled vitamin D supplementation trial among 50 apparently healthy subjects including supplementation of 140,000 IU at baseline and after 4 weeks (visit 1). The final follow-up visit was performed 8 weeks after the baseline examination (visit 2). Blood was drawn at each study visit to determine 25-hydroxyvitamin D levels and %Tregs. Tregs were characterized as CD4+CD25++ T cells with expression of the transcription factor forkhead box P3 and low or absent expression of CD127.', 'Forty-six study participants (65% females, mean age +/- SD 31 +/- 8 years) completed the trial. 25(OH)D levels increased from 23.9 +/- 12.9 ng/ml at baseline to 45.9 +/- 14.0 ng/ml at visit 1 and 58.0 +/- 15.1 ng/ml at visit 2. %Tregs at baseline were 4.8 +/- 1.4. Compared to baseline levels we noticed a significant increase of %Tregs at study visit 1 (5.9 +/- 1.7, P<0.001) and 2 (5.6 +/- 1.6, P<0.001).'], 'labels': ['BACKGROUND', 'OBJECTIVES', 'METHODS', 'RESULTS'], 'meshes': ['Adult', 'C-Reactive Protein', 'CD4 Lymphocyte Count', 'CD4-Positive T-Lymphocytes', 'Calcium', 'Dietary Supplements', 'Female', 'Follow-Up Studies', 'Forkhead Transcription Factors', 'Humans', 'Immunologic Factors', 'Interleukin-2 Receptor alpha Subunit', 'Interleukin-7 Receptor alpha Subunit', 'Male', 'Pilot Projects', 'T-Lymphocytes, Regulatory', 'Vitamin D', 'Vitamins'], 'reasoning_required_pred': ['y', 'e', 's'], 'reasoning_free_pred': ['y', 'e', 's']}, 'long_answer': 'Vitamin D supplementation was associated with significantly increased %Tregs in apparently healthy individuals. This immunomodulatory effect of vitamin D might underlie the associations of vitamin D deficiency and autoimmune diseases. Hence, our finding provides a rationale for further studies to investigate vitamin D effects on autoimmunological processes.', 'final_decision': 'yes'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2025-12-21 11:37:47.951\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 25779009, 'question': 'Bactericidal activity of 3 cutaneous/mucosal antiseptic solutions in the presence of interfering substances: Improvement of the NF EN 13727 European Standard?', 'context': {'contexts': ['There is no standard protocol for the evaluation of antiseptics used for skin and mucous membranes in the presence of interfering substances. Our objective was to suggest trial conditions adapted from the NF EN 13727 standard, for the evaluation of antiseptics used in gynecology and dermatology.', 'Three antiseptic solutions were tested in vitro: a chlorhexidine-benzalkonium (CB) combination, a hexamidine-chlorhexidine-chlorocresol (HCC) combination, and povidone iodine (P). The adaptation of trial conditions to the standard involved choosing dilutions, solvent, and interfering substances. The activity of solutions was assessed on the recommended strains at concentrations of 97% (pure solution), 50%, and 10% (diluted solution), and 1%. A logarithmic reduction ≥ 5 was expected after 60seconds of contact, to meet requirements of bactericidal activity.', 'HCC did not present any bactericidal activity except on P. aeruginosa at a concentration of 97%. P was not bactericidal on E. hirae at any concentration and on S. aureus at 97%. CB had the most homogeneous bactericidal activity with a reduction>5 log on the 4 bacterial strains at concentrations of 97%, 50% and 10%.'], 'labels': ['OBJECTIVE', 'METHODS', 'RESULTS'], 'meshes': ['Animals', 'Anti-Infective Agents, Local', 'Benzalkonium Compounds', 'Benzamidines', 'Cattle', 'Chlorhexidine', 'Cresols', 'Dose-Response Relationship, Drug', 'Drug Combinations', 'Drug Interactions', 'Enterococcus', 'Erythrocytes', 'Escherichia coli', 'Europe', 'Hand Disinfection', 'Humans', 'Inorganic Chemicals', 'Microbial Sensitivity Tests', 'Mucous Membrane', 'Osmolar Concentration', 'Povidone-Iodine', 'Pseudomonas aeruginosa', 'Serum Albumin, Bovine', 'Skin', 'Solutions', 'Staphylococcus aureus'], 'reasoning_required_pred': ['n', 'o'], 'reasoning_free_pred': ['m', 'a', 'y', 'b', 'e']}, 'long_answer': 'Adapting the NF EN 13727 standard allowed assessing the 3 tested solutions: only CB was bactericidal in dirty conditions. This study proved the possibility of validating antiseptic choice in vitro, in current practice conditions, for adjunctive treatment of skin and mucous membranes disorders, primarily of bacterial origin or with a potential of superinfection.', 'final_decision': 'maybe'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 77%|███████▋ | 385/500 [02:02<01:15, 1.52it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 77%|███████▋ | 387/500 [02:02<01:00, 1.87it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 11:37:48.967\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 78%|███████▊ | 388/500 [02:03<00:59, 1.87it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 11:37:49.186\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 78%|███████▊ | 389/500 [02:03<00:52, 2.12it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 78%|███████▊ | 390/500 [02:04<00:56, 1.94it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 79%|███████▉ | 394/500 [02:04<00:28, 3.67it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 79%|███████▉ | 397/500 [02:05<00:24, 4.18it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 80%|███████▉ | 398/500 [02:05<00:21, 4.64it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 80%|███████▉ | 399/500 [02:05<00:21, 4.70it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 80%|████████ | 401/500 [02:06<00:20, 4.91it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 81%|████████ | 403/500 [02:06<00:17, 5.67it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 81%|████████ | 406/500 [02:06<00:11, 8.22it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 82%|████████▏ | 408/500 [02:06<00:09, 10.20it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 82%|████████▏ | 410/500 [02:07<00:18, 4.96it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 82%|████████▏ | 412/500 [02:08<00:17, 4.95it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 83%|████████▎ | 415/500 [02:08<00:13, 6.24it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 83%|████████▎ | 417/500 [02:08<00:11, 7.18it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 84%|████████▍ | 419/500 [02:09<00:15, 5.32it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 84%|████████▍ | 420/500 [02:09<00:13, 6.03it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 84%|████████▍ | 421/500 [02:09<00:17, 4.46it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 85%|████████▌ | 425/500 [02:10<00:11, 6.75it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 86%|████████▌ | 428/500 [02:10<00:06, 10.36it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 86%|████████▌ | 430/500 [02:10<00:07, 10.00it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 86%|████████▋ | 432/500 [02:11<00:12, 5.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 87%|████████▋ | 434/500 [02:11<00:11, 5.70it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 88%|████████▊ | 438/500 [02:11<00:08, 7.64it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 88%|████████▊ | 440/500 [02:12<00:07, 7.58it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 88%|████████▊ | 441/500 [02:12<00:09, 6.52it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 89%|████████▉ | 444/500 [02:12<00:06, 8.34it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 89%|████████▉ | 446/500 [02:12<00:07, 6.86it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 90%|████████▉ | 449/500 [02:13<00:08, 6.18it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 90%|█████████ | 451/500 [02:13<00:08, 6.11it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 91%|█████████ | 453/500 [02:13<00:05, 8.15it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 91%|█████████ | 455/500 [02:14<00:08, 5.59it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 92%|█████████▏| 459/500 [02:14<00:05, 7.20it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 92%|█████████▏| 462/500 [02:15<00:04, 7.83it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 93%|█████████▎| 463/500 [02:15<00:05, 6.99it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 93%|█████████▎| 465/500 [02:16<00:07, 4.49it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 94%|█████████▎| 468/500 [02:16<00:05, 5.64it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 94%|█████████▍| 469/500 [02:16<00:05, 5.91it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 95%|█████████▍| 473/500 [02:17<00:03, 7.20it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 95%|█████████▌| 477/500 [02:17<00:02, 10.03it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 96%|█████████▌| 479/500 [02:18<00:03, 5.37it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 96%|█████████▋| 482/500 [02:18<00:02, 6.35it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 97%|█████████▋| 483/500 [02:18<00:03, 5.38it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 98%|█████████▊| 488/500 [02:19<00:01, 9.04it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 98%|█████████▊| 490/500 [02:19<00:00, 10.36it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 98%|█████████▊| 492/500 [02:20<00:01, 4.64it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 99%|█████████▉| 496/500 [02:20<00:00, 6.94it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 100%|█████████▉| 498/500 [02:21<00:00, 4.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 100%|█████████▉| 499/500 [02:24<00:00, 1.55it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 100%|██████████| 500/500 [02:25<00:00, 3.44it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2025-12-21 11:38:10.797\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36m__main__\u001b[0m:\u001b[36m\u001b[0m:\u001b[36m4\u001b[0m - \u001b[1mEvaluation metrics (before optimization): {'f1': 0.0, 'em': 0.0, 'acc': 0.6754032258064516}\u001b[0m\n", "\u001b[32m2025-12-21 11:38:10.798\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36m__main__\u001b[0m:\u001b[36m\u001b[0m:\u001b[36m6\u001b[0m - \u001b[1mOptimizing workflow...\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\n", " 0%| | 0/20 [00:00\n", "Evaluating workflow: 36%|███▌ | 18/50 [00:03<00:02, 11.33it/s]Unclosed client session\n", "client_session: \n", "\u001b[AUnclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348311.707778698)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348311.647691463)])']\n", "connector: \n", "\n", "Evaluating workflow: 40%|████ | 20/50 [00:03<00:05, 5.75it/s]Unclosed connector\n", "connections: ['deque([(, 348312.23635049)])']\n", "connector: \n", "\u001b[AUnclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348311.148779468)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348311.630780795)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348312.00281991)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348312.052342569)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348311.840471983)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348316.745466147)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348316.165887551)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348316.752908318)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348314.724728949)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348314.537095086)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348313.91465534)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348314.085274149)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348313.272450519)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348312.427323949)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348312.681080957)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348313.156899846)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348311.786324125)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348311.288999718)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348317.19348374)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348311.834190148)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348316.695726268)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348314.086405555)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348314.405993578)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348316.216846468)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348314.617232865)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348314.288309083)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348315.181866199)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348316.021681599)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348316.358518434)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348316.374404212)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348316.649466415)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348316.564751447)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348316.619848032)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348317.0824557)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348317.786948359)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348314.781416639)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348316.7268057)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348317.300977994)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348314.380897667)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348314.656413519)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348319.311558957)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348317.676818207)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348316.630187401)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348320.461060864)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348321.68116078)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348318.749932897)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348318.905694949)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348319.245991914)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348319.503241205)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348319.8090812)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348318.895857763)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348317.795785771)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 348319.69797527)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348319.241657293)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348319.298395116)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348318.692069484)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348320.301868432)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348320.149600407)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348320.791557016)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348320.817759577)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348320.825029438)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348321.249912607)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348321.329881505)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348321.440685516)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348321.251109687)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348322.020546748)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348319.870334418)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348319.662293175)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348322.521886712)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348318.835151507)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348322.587861434)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348322.488747945)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348322.584561276)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348324.034301114)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348322.63879678)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348322.88054391)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348323.581023969)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348323.225957134)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348324.324220949)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348322.982648019)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348324.37832626)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348321.536386689)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348323.300775122)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348323.920203921)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348321.380082826)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348321.857863619)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348325.133350931)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348324.655766084)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348324.437947337)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348324.413166128)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348324.969619466)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348325.345403635)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348324.600110444)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348325.565754742)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348324.545399914)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348326.621218427)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348325.407093075)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348325.759203806)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348324.495640594)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348326.456424199)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348324.206928965)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348326.227347517)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348326.95208261)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348326.767500948)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348326.799068306)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348327.40820404)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348328.970231368)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348326.280893945)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348327.575970841)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 348327.018461223)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348326.704750131)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348328.005815702)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348327.16583929)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348327.658372019)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348327.800594396)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348329.701254189)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348327.864327609)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348328.610752634)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348328.768942915)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348329.329012404)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348329.099549652)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348329.657205054)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348329.498647269)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348329.205533748)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348329.864089231)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348333.147114607)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348329.020331424)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348329.09530974)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348329.967923159)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348329.844746844)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348330.712757343)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348330.356745478)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348331.002425996)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348330.911959533)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348331.425665391)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348331.579598057)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348331.733637452)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348331.638564865)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348331.651366908)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348331.122859506)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348332.45704708)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348331.709476199)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348331.605860442)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348332.799187236)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348332.293706299)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348330.427561187)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348332.752423073)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348331.999254919)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348332.86031685)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348333.331783163)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348333.450586334)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348333.778296414)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348333.910759251)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348336.504687065)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348334.050825826)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348335.558438878)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348370.533793954)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348334.649280165)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348334.404188232)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348333.800932985)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348334.161186118)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348332.311946656)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348336.969372243)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348335.326589105)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348335.339946032)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348335.471694384)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348336.02729734)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348336.689882936)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348335.601394266)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348336.137177763)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348336.254526528)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348336.588552425)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348336.305255541)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348335.862770269)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348337.057644845)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348371.911981286)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348335.769890517)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348336.493026596)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348336.863266558)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348370.272416711)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348371.290397595)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348370.309636487)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348370.444197937)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348371.311623169)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348371.825295548)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348370.27705807)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348370.762083339)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348370.219045227)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348370.621109689)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348370.460845434)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348369.840744447)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348370.159740042)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348370.59251977)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 348370.622182594)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348370.316282047)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348376.40489361)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348373.604415643)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348373.559569332)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348373.268168812)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348375.049931382)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348373.634495545)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348374.763897549)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348374.025815465)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348375.238273495)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348375.222327979)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348374.774335461)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348375.273792659)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348376.556131186)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348374.453766134)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348374.791003209)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348375.869836178)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348375.286910498)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348374.855033283)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348373.77685645)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348374.08200607)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348373.869891298)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348377.638837837)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348377.837476127)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348377.299303598)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348377.572978297)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348380.080686496)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 348377.237549573)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348376.778125122)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348378.174483491)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348377.855305944)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348377.40765142)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348378.438952866)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348378.545725492)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348377.871801517)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348379.454219444)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348378.765371286)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348376.209370031)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348376.331368589)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348379.450227248)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348379.159440079)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348380.068501595)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348379.477834162)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348380.296989134)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348380.675089206)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348379.425925641)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348379.77361815)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348381.545435231)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348380.50179014)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348381.245826401)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348378.586665685)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348378.657706945)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348380.630699492)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348381.38518665)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348378.222537421)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 348381.769892414)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348381.259979177)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348382.678736612)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348382.113722373)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348381.167751188)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348382.140731515)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348381.832298218)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348382.548632761)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348381.521057684)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348382.1751615)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348382.15550978)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348382.556982475)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348383.615563109)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348383.351046503)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348380.990081451)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348383.521413528)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348383.746822863)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348384.276354959)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348385.547422869)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348384.033740658)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348384.615802825)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348383.403487835)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348383.656695976)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348383.333962049)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348384.624221039)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348385.262521847)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348384.967270546)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348384.452177779)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348384.845679455)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348385.999544515)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348384.7750236)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348386.341535278)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348385.855552544)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348385.783811282)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348386.310395717)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348385.719604469)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348386.940703791)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348386.911268959)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348386.377808938)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348386.569723217)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348389.52102017)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348388.245910828)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348386.802831076)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348386.224197991)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348387.541931772)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348387.36961285)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348384.900102072)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348388.63662774)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348389.503143027)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348388.571578299)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348388.472301714)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348389.51214834)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348388.035549106)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348388.842413115)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348390.637677162)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348388.226048021)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 348388.849382444)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348389.702011194)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348390.983015517)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348388.652497247)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348389.532576764)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348389.633391746)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348387.31392138)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348390.179012145)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348391.673327884)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348391.418247943)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348394.227596881)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348391.851906267)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348391.919776661)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348391.740855165)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348390.831323205)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348391.792460791)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348391.096942568)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348397.887960351)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348389.765681143)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348389.707638842)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348393.11629428)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348392.489478495)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348391.565995687)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348393.077989515)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348393.40656893)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348394.486428197)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348393.940152574)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348393.64799848)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348393.848620331)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348393.741036773)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348394.500184751)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348394.772059853)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348396.366550645)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348392.233805254)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348395.925868301)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348395.280871145)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348392.213463777)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348396.958781009)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348395.965003762)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348395.788300511)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348397.544179229)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348430.72265447)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348395.660679405)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348396.353896837)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348396.117180526)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348398.300062369)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348397.251900362)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348396.511490421)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348395.201513403)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348399.802086013)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348394.793463919)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348430.661621516)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348394.662025784)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348397.61262434)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348397.126276163)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348398.355094196)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348398.640547533)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348398.975441572)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348430.954706764)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348398.920545244)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348430.145969252)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348430.346788972)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348398.811732248)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348399.215889588)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348431.459082768)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348398.474023871)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348397.81185508)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348398.043485307)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348430.446747988)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348431.955261054)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348430.340259066)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348430.240427358)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348430.174760745)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348431.463376596)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348430.776315065)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 348434.475572211)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348435.487050574)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348433.374250392)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348430.511249962)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348431.625946373)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348430.568759176)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348430.506041905)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348433.409843617)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348432.481248932)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348433.864017248)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348434.295694241)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348432.700553575)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348435.737176002)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348433.715787153)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348434.6069521)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348435.206284659)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348435.679222997)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348435.039113856)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348435.729195934)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348433.933095354)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348436.992538914)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348434.808898559)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348438.782003176)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348435.29328713)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348436.644326156)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348435.843246204)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348437.666736805)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348436.153095427)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 348437.455904)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348435.614545481)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348433.75688256)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348437.094009912)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348438.044821921)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348438.224419173)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348439.118008343)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348446.360481619)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348439.907220321)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348438.709232755)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348439.144159777)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348438.324846156)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348439.004758909)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348440.169699235)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348437.099824643)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348438.796811339)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348439.148790224)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348437.331039812)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348437.504272362)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348439.367331777)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348440.499537301)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348440.235410441)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348440.721247388)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348441.135766691)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348440.8672147)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348441.897786726)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348442.119566897)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348442.074705646)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348440.517154621)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348440.794744319)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348441.388955917)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348439.228115147)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348439.444166114)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348441.478028496)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348440.904781785)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348441.616701853)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348444.325038979)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348443.924706539)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348443.028803253)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348442.937080796)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348442.825473925)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348442.579038157)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348444.430971644)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348443.512246723)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348443.614291456)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348444.013839898)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348443.041836389)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348443.677761608)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348442.036906207)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348445.61320632)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348441.417864843)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348444.251326332)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348446.516723137)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348444.06536542)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348444.63541717)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348445.365184722)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348445.396403455)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348446.191586228)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348445.980712639)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348445.741933086)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348446.617672532)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348445.957650436)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348445.882019982)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348446.725807359)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348446.337447619)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348446.482546217)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348444.655725425)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348448.37348447)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348447.583297098)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348448.092036552)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348447.713882178)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348449.482011269)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348448.007271057)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348447.387419288)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348448.519782121)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348448.289944835)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348448.31750222)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348449.968441814)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348447.35351645)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348448.401624917)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348449.70563074)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348448.527697508)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348454.310760183)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348449.802575867)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 348449.835482175)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348453.225801952)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348449.720675506)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348449.520025252)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348450.505294027)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348515.174264451)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348514.800865367)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348517.40270991)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348514.730242748)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348515.787571175)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348514.679585524)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348518.412815046)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348517.553355301)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348514.791710661)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348515.000363652)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348513.917525162)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348514.67465588)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348515.704837944)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348514.991851295)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348514.595599606)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348514.262962484)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348515.971813661)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348514.776450036)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348515.452669966)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348515.004167815)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348515.440509489)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348516.605962962)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348517.643940123)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348517.954067349)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348518.260203437)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348518.021116311)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348517.008012942)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348517.298004859)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348517.638268178)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348518.193674366)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348517.518635839)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348517.449161614)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348518.181209229)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348518.259030957)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348519.199780706)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348518.952099689)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348519.512757509)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348519.407278576)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348519.476276248)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348519.53884114)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348517.707637834)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348519.421403826)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348517.614803249)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348521.116037515)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348520.02722732)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348520.744892481)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348520.068905467)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348519.735153824)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348519.145186392)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348519.90481572)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348582.48410115)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348586.44240352)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 348582.267304686)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348582.413849404)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348582.755102451)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348582.243095023)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348586.549061807)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348582.360024123)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348581.680684282)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348582.915840285)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348582.71762694)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348582.327627667)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348582.664848493)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348582.468348509)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348582.445987016)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348582.75270088)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348582.131744208)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348582.794621767)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348582.65957807)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348582.187087966)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348583.746501978)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348584.202464717)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348585.802504678)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348585.140905271)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348585.04603753)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348585.120547608)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348585.282891529)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348585.117127325)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348585.124352556)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348587.711466362)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348589.272074517)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348586.002731725)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348586.539777013)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348586.443556538)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348587.029610248)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348585.129474849)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348585.122281285)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348585.408691561)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348585.375689037)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348585.597019215)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348587.078410425)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348590.029160412)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348587.085599333)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348587.77048928)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348588.050236028)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348587.31033263)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348588.422815755)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348587.639344387)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348588.361585182)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348588.036068545)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348657.409656379)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348656.804978859)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348657.256327821)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348657.032332564)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348657.721341067)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348657.244291766)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348657.437448682)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348657.3970722)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348657.272682969)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348658.385925245)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348659.152389112)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348658.033303326)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348657.720218792)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348657.706139767)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348657.043894252)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348657.939760982)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348656.910766252)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348657.83012928)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348657.461377833)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348657.381445982)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348659.664684189)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348663.215599199)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348658.591840374)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348660.766239286)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348659.896464563)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348659.657927889)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348659.942541641)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348660.672122634)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348661.921534009)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348658.946352739)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348660.371542916)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348659.936060079)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348660.634664462)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348660.741936789)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348660.609819341)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348660.716208991)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348661.110889528)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348659.889019632)])']\n", "connector: \n", "\n", "Evaluating workflow: 46%|████▌ | 23/50 [00:04<00:05, 4.97it/s]\u001b[A\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 50%|█████ | 25/50 [00:04<00:04, 5.93it/s]\u001b[A" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\n", "Evaluating workflow: 54%|█████▍ | 27/50 [00:04<00:03, 7.37it/s]\u001b[A" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\n", "Evaluating workflow: 58%|█████▊ | 29/50 [00:04<00:02, 7.96it/s]\u001b[A" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\n", "Evaluating workflow: 62%|██████▏ | 31/50 [00:05<00:02, 6.46it/s]\u001b[A" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\n", "Evaluating workflow: 66%|██████▌ | 33/50 [00:05<00:02, 6.38it/s]\u001b[A\n", "Evaluating workflow: 70%|███████ | 35/50 [00:05<00:01, 7.92it/s]\u001b[A" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\n", "Evaluating workflow: 74%|███████▍ | 37/50 [00:06<00:01, 7.83it/s]\u001b[A" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\n", "Evaluating workflow: 78%|███████▊ | 39/50 [00:06<00:01, 6.99it/s]\u001b[A" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\n", "Evaluating workflow: 84%|████████▍ | 42/50 [00:06<00:00, 8.46it/s]\u001b[A" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\n", "Evaluating workflow: 88%|████████▊ | 44/50 [00:07<00:00, 7.10it/s]\u001b[A" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\n", "Evaluating workflow: 92%|█████████▏| 46/50 [00:07<00:00, 7.69it/s]\u001b[A" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\n", "Evaluating workflow: 96%|█████████▌| 48/50 [00:07<00:00, 6.65it/s]\u001b[A" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\n", "Evaluating workflow: 98%|█████████▊| 49/50 [00:07<00:00, 6.01it/s]\u001b[A" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\n", "Evaluating workflow: 100%|██████████| 50/50 [00:08<00:00, 5.86it/s]\u001b[A" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2025-12-21 11:46:07.740\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.optimizers.textgrad_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m275\u001b[0m - \u001b[1mStep 7 metrics: {'f1': 0.0, 'em': 0.0, 'acc': 0.96}\u001b[0m\n", "\u001b[32m2025-12-21 11:46:07.741\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.optimizers.textgrad_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m291\u001b[0m - \u001b[1mMetrics are worse than the best snapshot which has {'f1': 0.0, 'em': 0.0, 'acc': 0.98}. Rolling back to the best snapshot.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\n", "\r", " 35%|███▌ | 7/20 [07:56<14:38, 67.56s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2025-12-21 11:46:07.744\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.optimizers.textgrad_optimizer\u001b[0m:\u001b[36mstep\u001b[0m:\u001b[36m319\u001b[0m - \u001b[1mExecuting workflow...\u001b[0m\n", "\u001b[32m2025-12-21 11:46:17.589\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.optimizers.textgrad_optimizer\u001b[0m:\u001b[36mstep\u001b[0m:\u001b[36m347\u001b[0m - \u001b[1mComputing gradients...\u001b[0m\n", "\u001b[32m2025-12-21 11:47:00.072\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.optimizers.textgrad_optimizer\u001b[0m:\u001b[36mstep\u001b[0m:\u001b[36m349\u001b[0m - \u001b[1mUpdating agents...\u001b[0m\n", "\u001b[32m2025-12-21 11:47:03.095\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.optimizers.textgrad_optimizer\u001b[0m:\u001b[36mstep\u001b[0m:\u001b[36m353\u001b[0m - \u001b[1mAgents updated\u001b[0m\n", "\u001b[32m2025-12-21 11:47:03.096\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.optimizers.textgrad_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m271\u001b[0m - \u001b[1mEvaluating the workflow at step 8 ...\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\n", "Evaluating workflow: 0%| | 0/50 [00:00\u001b[0m:\u001b[36m10\u001b[0m - \u001b[1mEvaluating workflow on test set...\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\n", "Evaluating workflow: 0%| | 2/500 [00:01<05:04, 1.63it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 1%| | 6/500 [00:01<01:19, 6.19it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 2%|▏ | 8/500 [00:02<01:37, 5.04it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 2%|▏ | 10/500 [00:02<01:33, 5.23it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 2%|▏ | 12/500 [00:02<01:30, 5.38it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 4%|▍ | 19/500 [00:03<00:49, 9.74it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 5%|▍ | 23/500 [00:03<00:39, 12.16it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 5%|▌ | 25/500 [00:04<01:13, 6.45it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 6%|▌ | 30/500 [00:04<01:03, 7.43it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 7%|▋ | 33/500 [00:05<01:03, 7.36it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 7%|▋ | 35/500 [00:05<01:05, 7.10it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 8%|▊ | 39/500 [00:05<00:42, 10.83it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 8%|▊ | 41/500 [00:05<00:44, 10.43it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 9%|▊ | 43/500 [00:06<01:03, 7.22it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 9%|▉ | 46/500 [00:06<00:56, 8.07it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 10%|▉ | 48/500 [00:07<01:04, 7.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 10%|█ | 50/500 [00:07<00:52, 8.60it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 10%|█ | 52/500 [00:07<00:49, 9.07it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 11%|█ | 56/500 [00:07<00:48, 9.16it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 12%|█▏ | 61/500 [00:08<00:41, 10.53it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 13%|█▎ | 63/500 [00:08<00:53, 8.18it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 13%|█▎ | 65/500 [00:08<00:51, 8.45it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 13%|█▎ | 67/500 [00:09<00:46, 9.39it/s]Unclosed client session\n", "client_session: \n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 348660.062508542)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Evaluating workflow: 14%|█▍ | 69/500 [00:09<01:25, 5.05it/s]Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348663.302940201)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348662.574484717)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348662.759781788)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348663.361842076)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348666.822304902)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348663.025722723)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348661.718937247)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348661.784055457)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348729.388288885)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348732.134783533)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348729.580429592)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348729.519234933)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348729.819872975)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348728.854409988)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348729.024356929)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348730.568148792)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348729.584604258)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348729.983269049)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348729.073715614)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348730.21099916)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348730.486820573)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348732.470697744)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348733.005046906)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348734.690414721)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348735.122304742)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348735.279249382)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348734.732701454)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348735.326796656)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348793.987526499)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348793.77276326)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348794.061783576)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348794.40031248)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348794.498363205)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348794.665917)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348794.378033473)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348794.120801753)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348794.011917468)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348793.868129777)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348794.315303177)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348794.963310355)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348793.925770317)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348793.932817839)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348793.801556963)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348793.720147676)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348795.549710215)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348796.485713561)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348795.980156056)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348796.678865347)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348797.794578322)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348799.045754535)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348797.73238293)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348797.37471901)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348797.152134114)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 348797.119130024)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348799.395137223)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348799.228712129)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348799.495375421)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348799.760671156)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348799.371537404)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348799.411853772)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348852.544532173)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348852.337976418)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348852.434294964)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348852.221472224)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348853.042037786)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348852.058296881)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348852.335183673)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348852.459903983)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348853.020365738)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348852.710047081)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348852.36265311)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348852.120809078)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348852.910129004)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348851.974668243)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348852.380914037)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348851.608593502)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348852.321246984)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348854.886796725)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348854.658971912)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348855.520294685)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 348855.367378381)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348856.776297895)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348854.893948596)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348857.216624395)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348857.228820528)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348858.550079807)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348857.417374722)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348858.055480198)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348857.243088591)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348857.399981501)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348857.459237962)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348924.484607418)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348926.485608645)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348924.904326269)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348925.127487544)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348924.939880047)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348924.780915264)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348925.652835737)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348925.385003639)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348925.011347827)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348924.801578256)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348924.699349858)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348924.246536036)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348926.484100754)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348925.612744705)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348924.072665963)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348926.488404617)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 348926.487328651)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348927.693486134)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348927.492708533)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348928.576441006)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348925.474882754)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348927.474739823)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348925.362057151)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348927.38789704)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348927.568854011)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348927.304621485)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348928.468440443)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348928.079266)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348928.476356354)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348928.145673387)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348927.222169126)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348928.838314556)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348928.783547744)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348730.204957806)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348728.522730685)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348729.526149152)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348794.565021922)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348794.325115437)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348793.339721707)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348852.392170552)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348851.911225124)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348852.457858667)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348924.838038437)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348925.768365239)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348929.073365887)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348664.064217719)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348661.721516762)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348664.35527483)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348730.134950254)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348735.722818192)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348730.427579677)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348729.920065649)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348730.071971656)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348730.817995155)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348732.411016517)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348732.814378945)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348732.081516239)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348732.045199978)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348732.117035477)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348732.08672421)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348732.780954694)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348733.142553516)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 348732.872161458)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348731.714292116)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348732.515783296)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348733.059779282)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348733.122026499)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348733.435696103)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348733.704021541)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348733.586175388)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348733.965906435)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348735.639559723)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348733.585111333)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348734.805649973)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348735.377136422)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348794.645552716)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348795.328640816)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348797.159433844)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348796.65207041)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348796.714232069)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348796.720562888)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348797.771255791)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348797.787392519)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348797.663209338)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348797.987065976)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348798.127358134)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348797.183048024)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348797.19932018)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348800.253584892)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348853.463741104)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348854.170288006)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348854.951311811)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348854.676443611)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348855.126815867)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348855.445248603)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348854.713014659)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348855.852334291)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348855.593085962)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348855.877408569)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348856.314258816)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348855.702821222)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348856.405830653)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348855.754225882)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348855.288559678)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 348855.994906692)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348925.515769544)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348796.668883509)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348931.252740789)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348930.68951156)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348930.047885146)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348929.29067605)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348929.196917981)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348928.183437744)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348929.898643764)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348929.435086142)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348929.296822246)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348929.837853773)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348930.444634241)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348930.042117142)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348929.635923067)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348989.767882854)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348989.380294508)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348989.156385935)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348988.539391365)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348990.570043729)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348988.430164099)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348990.235769639)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348989.27227118)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348989.096227374)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348989.17695011)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348988.703231026)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348989.092202617)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348989.321349127)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348989.230829564)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348988.85933222)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348988.816545939)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348990.419124045)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348990.071703173)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348990.59114715)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348992.265113695)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348992.486941349)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348991.823952268)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348989.836538898)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348989.607466722)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348989.397593756)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348991.772266852)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348991.077709789)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348991.426404693)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348991.7747235)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348992.328607127)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348992.109072469)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348992.362712377)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348995.371555915)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348991.940961618)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348991.966582386)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348992.674859593)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348992.544708564)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348991.958601264)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348993.304028069)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348993.667378718)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348992.871080047)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348993.250057815)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348993.820848742)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 348993.085030898)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348995.396554711)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348996.38528802)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348994.322935677)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348993.958818545)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348994.451018387)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 348994.349596698)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349054.796921186)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349055.374041545)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349055.016403751)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349055.269600451)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349054.575214711)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349054.972120102)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349054.927487404)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349055.516270584)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349057.870234831)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349055.45033154)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349055.006830776)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349054.218504212)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349055.659430851)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349054.141729913)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349055.35996957)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349055.845881)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349055.378361875)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349055.640319821)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349055.492518794)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349054.867831699)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349055.671982512)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349057.272500371)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349057.969654053)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349057.592947935)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349057.533709031)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349057.438116452)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349055.766831902)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349058.091801828)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349057.653274039)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349058.319330203)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349058.94669408)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349058.037143909)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349058.645404827)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349058.094202246)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349059.033854712)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349059.075680975)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349058.813146737)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349058.58213202)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349057.916138494)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349057.308627105)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349057.463730997)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349059.70297056)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349060.147916141)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349059.994236971)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349059.199686838)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349060.765234188)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349059.24768258)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349060.450517319)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349058.748679308)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349061.232136114)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349129.996950763)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349128.701393565)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349129.204480426)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349128.556256797)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349128.520659412)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349128.877222719)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349128.447576208)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349129.255394738)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349128.435222931)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349129.033683384)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349128.706224116)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349128.486576598)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349129.554515345)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349127.656385644)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349128.439905893)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349128.686969502)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349128.905486843)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349128.962276114)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349130.472059498)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349131.48414393)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349131.890830348)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349131.807729692)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349131.530500094)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349129.427549859)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349128.255162657)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349128.366732514)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349131.915865829)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349131.125322077)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349130.93536479)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349131.812808196)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349132.058210723)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349130.24080908)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349130.936391379)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349131.605943119)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349130.558995855)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349132.655897603)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349132.300103593)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349131.805650927)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349132.028505419)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349133.049251401)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349132.830580821)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349132.443172321)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349134.389248851)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349133.673977077)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349133.629312679)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349132.552946971)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349133.509060948)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349131.767095493)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349134.376458636)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349133.640883133)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349192.820089773)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349192.136194061)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349191.93192297)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349191.488744812)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349191.789810949)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349191.782800663)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 349191.937428046)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349190.851652177)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349191.374582457)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349200.859332183)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349191.58618653)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349190.470947096)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349191.060640178)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349194.340237449)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349191.898661683)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349191.145359564)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349192.015310434)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349192.375365016)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349192.436954269)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349195.702987017)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349193.390833246)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349194.313511804)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349193.200539544)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349191.724397676)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349191.788243428)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349194.120101114)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349193.986705035)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349194.350465599)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349193.897368471)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349195.407107696)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349195.226953409)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349194.654181597)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349194.587192856)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349194.455540025)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349194.92791685)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 349194.666131004)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349194.432942341)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349194.406165251)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349197.4328064)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349195.326713604)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349196.072108846)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349197.297059742)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349195.906372515)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349201.15579572)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349196.77991312)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349197.014096316)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349196.350241852)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349196.079524933)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349196.01139178)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349197.169378739)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349263.964029998)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349265.263713957)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349263.728752826)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349262.3594854)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349262.667800005)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349262.251887474)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349261.418399258)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349266.554065119)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349263.455377271)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349262.107363579)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349262.145900049)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349262.333126632)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349262.258869237)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 349263.338918263)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349262.498109253)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349262.629842502)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349262.546089306)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349261.880459649)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349262.3398812)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349262.121623843)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349263.501786346)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349265.796240835)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349266.357856449)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349265.346595586)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349264.86010381)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349264.75461104)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349262.85305455)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349264.814936192)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349264.820040145)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349265.848053982)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349265.721537138)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349265.888456704)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349267.729589673)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349266.517011755)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349265.577384169)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349266.798685094)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349267.75805451)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349264.590604817)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349266.452948643)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349264.319005536)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349264.96166312)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349266.16643659)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349268.472711775)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349267.058378783)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349267.782917463)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349267.284340943)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349267.728455658)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349270.976384306)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349268.665115216)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349266.812530022)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349332.328851053)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349330.947525478)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349331.566682862)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349332.387173365)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349332.066404217)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349331.468284731)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349332.297047765)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349331.734843605)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349334.680753814)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349331.96568428)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349331.073237372)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349332.614498234)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349331.76582017)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349331.334716308)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349332.424268724)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349331.768160482)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349331.937322391)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349331.410816875)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349332.072101655)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349337.250162432)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349332.593601281)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349333.002717219)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349333.612012989)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349333.609055921)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349337.427674659)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349336.946263274)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349334.252710085)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349334.384559108)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349334.122414066)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349334.877759309)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349334.595543381)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349334.955486126)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349334.652105765)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349334.858434847)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349335.955443395)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349334.268483055)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349335.723484748)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349335.764715137)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349335.24478181)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349335.220670686)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349334.211873719)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349334.540834397)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349336.844749503)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349336.31858672)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349338.74848562)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349337.087444287)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349337.503908995)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349336.651519705)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349335.934613378)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 349337.484020364)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349400.213843314)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349400.090789792)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349400.72843933)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349400.053162419)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349399.361832829)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349402.168902338)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349400.435430813)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349401.701094356)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349399.690632497)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349401.010064193)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349399.903279122)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349400.754454317)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349399.707144986)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349400.003470427)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349401.900697799)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349400.487464509)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349400.569779173)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349400.066947957)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349400.558632389)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349400.594391631)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349400.871718208)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349401.087774486)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349401.838205694)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349402.067230182)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349402.963119571)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349402.682005731)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349403.182848151)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349403.104465723)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 349402.584760377)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349405.29799755)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349403.022757457)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349403.389525572)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349403.663511183)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349403.708392331)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349403.431780586)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349403.289217823)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349403.587564639)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349403.66640421)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349403.364801194)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349405.363876669)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349405.938742183)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349405.449335487)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349404.727140774)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349405.307041206)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349405.612794623)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349408.256485106)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349403.404323265)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349403.50413989)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349405.764594779)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349405.281109002)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349467.177337223)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349466.94937386)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349467.789043202)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349469.733102158)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 349467.712021743)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349467.94292541)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349467.987561944)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349466.340589492)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349467.135211876)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349467.180054844)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349466.078593401)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349467.553799324)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349467.052499983)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349467.54181674)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349468.460006134)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349466.626441723)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349467.506310518)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349467.5350404)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349467.76393552)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349469.891128872)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349470.371648208)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349467.970557528)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349467.001716497)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349468.441144058)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349467.370394636)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349471.368308394)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349469.415951252)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349469.432307711)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349469.706177377)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349470.763090888)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349471.022501601)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349469.14846494)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349469.701122095)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349470.617575519)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349470.580608963)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349469.492349362)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349471.764613034)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349470.579359466)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349470.576041194)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349471.168474819)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349472.958708508)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349471.673105243)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349471.581736076)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349472.16327479)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349472.249545227)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349471.883413543)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349471.748280892)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349470.500784435)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349470.125962681)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349470.683987071)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349537.635499341)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349537.609920946)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349538.125004657)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349536.956876484)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349536.999555711)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349536.308590346)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349536.889737772)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349537.650384125)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349537.224893574)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349537.156042394)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349537.908095561)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349536.934107693)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349538.03632862)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349537.454364666)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349537.142364004)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349536.879794653)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349537.370183728)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349536.656158964)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349538.156162798)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349538.201041495)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349539.932905776)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349539.479403684)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349539.279300529)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349536.741277401)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349537.329429949)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349538.719315445)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349538.514974969)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349539.484463297)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349539.148114686)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349540.732800718)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349540.671402079)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349540.006819359)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349539.657145774)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349540.908507102)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349541.492519397)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349540.880367755)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349541.177024801)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349540.734093687)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349540.001760063)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349541.746465915)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349539.56486207)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349543.042651884)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349539.793627569)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349540.041762622)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349540.490582486)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349541.783452334)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349542.012369301)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349541.629090781)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349541.646445502)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349542.273695646)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349588.877506181)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349590.219443888)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349588.98779625)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349588.864831426)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349590.603426531)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349592.232390338)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349589.689805685)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349589.481036869)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349589.668028282)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349588.807785736)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349589.584959333)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349589.407973177)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349590.65151291)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349588.449382767)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349589.066519521)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349589.68405536)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349588.211947415)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349589.843951525)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349592.77907654)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349589.250703692)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349590.995567309)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349591.948257132)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349591.631097625)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349591.155064102)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349591.353443147)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349591.146895862)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349590.941992461)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349589.778822443)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349593.486230345)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349589.966425571)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349594.120242493)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349592.590103738)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349592.340385627)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349592.720918878)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349592.796720017)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349591.937755908)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349593.70119915)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349593.720996992)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349593.637734716)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349592.447228157)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349592.597693424)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349592.399760399)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349592.722171051)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349594.569032959)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349593.833681446)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349593.627206178)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349593.95577732)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349594.240739103)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349595.156865522)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349594.083005052)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349652.684444419)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349651.313197055)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349652.791775673)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349652.906799293)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349653.844382866)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349652.833741766)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349652.666240847)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349652.535071266)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349651.629266533)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349653.192248383)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349652.719586001)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349652.260047796)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349651.759141012)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349653.335470199)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349652.845089725)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349652.203033823)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349653.14655578)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349652.226932654)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349652.722428516)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349653.169349536)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349652.782872108)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349655.17063511)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349654.751361212)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349655.277966855)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349655.150114827)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349653.862519141)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349655.308431609)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 349653.140586772)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349655.37419169)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349655.707063189)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349654.715623864)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349655.230620294)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349656.532751627)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349655.446579727)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349655.842874538)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349656.395208914)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349654.982876105)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349654.469168498)])']\n", "connector: \n", "Evaluating workflow: 15%|█▍ | 74/500 [00:10<01:10, 6.00it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 15%|█▌ | 76/500 [00:10<01:01, 6.91it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 16%|█▌ | 79/500 [00:11<00:49, 8.55it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 17%|█▋ | 83/500 [00:11<00:38, 10.70it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 17%|█▋ | 85/500 [00:11<00:33, 12.28it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 17%|█▋ | 87/500 [00:12<00:58, 7.02it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 18%|█▊ | 90/500 [00:12<01:02, 6.57it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 18%|█▊ | 92/500 [00:12<00:50, 8.00it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 19%|█▉ | 94/500 [00:13<01:01, 6.64it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 20%|██ | 100/500 [00:13<00:33, 12.02it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 20%|██ | 102/500 [00:14<00:54, 7.30it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 21%|██ | 105/500 [00:14<00:44, 8.85it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 21%|██▏ | 107/500 [00:16<02:48, 2.33it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2025-12-21 12:00:44.586\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 22042121, 'question': 'Perioperative care in an animal model for training in abdominal surgery: is it necessary a preoperative fasting?', 'context': {'contexts': ['Demonstrate that the rabbit may be used in the training of surgery, in addition to present its perioperative care.', 'Thirty two animals, with age and weight, respectively, from 3 to 5.5 months old and 3000 to 4200 grams, were undergone different periods of pre-operative fasting, exclusive intramuscular anesthesia (ketamine+xylazine), laparotomy with total gastrectomy and total splenectomy. It was dosed the pre-operative (initial) and post-surgical (end) serum blood glucose, in addition to quantify the gastric content after the resection of the part.', 'The anesthetical-surgical procedure presented a mortality rate of 3.125% (1:32) and a morbidity rate of 6.25% (2:32). It was evidenced an initial mean blood glucose = 199.4 mg/dl and the end = 326.1 mg/dl. In spite of extended fasting (minimum of 2 hours for the absolute fasting and maximum of 8.5 hours for liquids, and 20.5 hours for solids) all animals presented at the end of the surgical procedure any gastric content and a blood glucose increase. Those with fasting for liquids and solids when compared to the quantity of solid gastric content, presented a moderate negative degree of correlation.'], 'labels': ['PURPOSE', 'METHODS', 'RESULTS'], 'meshes': ['Animals', 'Blood Glucose', 'Fasting', 'Gastrectomy', 'Gastric Emptying', 'Laparotomy', 'Male', 'Models, Animal', 'Perioperative Care', 'Preoperative Care', 'Prospective Studies', 'Rabbits', 'Splenectomy'], 'reasoning_required_pred': ['n', 'o'], 'reasoning_free_pred': ['n', 'o']}, 'long_answer': 'The rabbit is a good model to be used in training of surgery, with a low morbi-mortality, able to be anesthetized intramuscularly, with no need of pre-operative fasting and does not present hypoglycemia even with the extended fasting period.', 'final_decision': 'no'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 22%|██▏ | 110/500 [00:21<04:51, 1.34it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2025-12-21 12:00:44.601\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 27184293, 'question': 'Does microbial contamination influence the success of the hematopoietic cell transplantation outcomes?', 'context': {'contexts': ['Microbial contamination can be a marker for faulty process and is assumed to play an important role in the collection of hematopoietic progenitor cell (HPC) and infusion procedure. We aimed to determine the microbial contamination rates and evaluate the success of hematopoietic cell transplantation (HCT) in patients who received contaminated products.PATIENTS-', 'We analyzed microbial contamination records of HPC grafts between 2012 and 2015, retrospectively. Contamination rates of autologous donors were evaluated for at three steps: at the end of mobilization, following processing with dimethyl sulfoxide, and just before stem cell infusion. Grafts of allogeneic donors were assessed only before HCT.', 'A total of 445 mobilization procedures were carried out on 333 (167 autologous and 166 allogeneic) donors. The microbiological contamination of peripheral blood (323/333 donations) and bone marrow (10/333 donations) products were analyzed. Bacterial contamination was detected in 18 of 1552 (1.15 %) culture bottles of 333 donors. During the study period 248 patients underwent HCT and among these patients microbial contamination rate on sample basis was 1.3 % (16/1212). Microbial contamination detected in nine patients (7 autologous; 2 allogeneic). In 8 of 9 patients, a febrile neutropenic attack was observed. The median day for the neutropenic fever was 4 days (0-9). None of the patients died within the post-transplant 30 days who received contaminated products.'], 'labels': ['INTRODUCTION', 'METHODS', 'RESULT'], 'meshes': ['Adolescent', 'Adult', 'Aged', 'Allografts', 'Autografts', 'Female', 'Hematologic Neoplasms', 'Hematopoietic Stem Cell Mobilization', 'Hematopoietic Stem Cell Transplantation', 'Hematopoietic Stem Cells', 'Humans', 'Male', 'Middle Aged'], 'reasoning_required_pred': ['n', 'o'], 'reasoning_free_pred': ['m', 'a', 'y', 'b', 'e']}, 'long_answer': 'The use of contaminated products with antibiotic prophylaxis may be safe in terms of the first day of fever, duration of fever, neutrophil, platelet engraftment and duration of hospitalization.', 'final_decision': 'maybe'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2025-12-21 12:00:44.700\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 22564465, 'question': 'Mammographic screening in Sami speaking municipalities and a control group. Are early outcome measures influenced by ethnicity?', 'context': {'contexts': ['Female citizens of Sami (the indigenous people of Norway) municipalities in northern Norway have a low risk of breast cancer. The objective of this study was to describe the attendance rate and outcome of the Norwegian Breast Cancer Screening Program (NBCSP) in the Sami-speaking municipalities and a control group.', 'A retrospective registry-based study.', 'The 8 municipalities included in the administration area of the Sami language law (Sami) were matched with a control group of 11 municipalities (non-Sami). Population data were accessed from Statistics Norway. Data regarding invitations and outcome in the NBCSP during the period 2001-2010 was derived from the Cancer Registry of Norway (CRN). The NBCSP targets women aged 50-69 years. Rates and percentages were compared using chi-square test with a p-value<0.05 as statistical significant.', 'The attendance rate in the NBCSP was 78% in the Sami and 75% in the non-Sami population (p<0.01). The recall rates were 2.4 and 3.3% in the Sami and non-Sami population, respectively (p<0.01). The rate of invasive screen detected cancer was not significantly lower in the Sami group (p=0.14). The percentage of all breast cancers detected in the NBCSP among the Sami (67%) was lower compared with the non-Sami population (86%, p=0.06).'], 'labels': ['OBJECTIVES', 'STUDY DESIGN', 'METHODS', 'RESULTS'], 'meshes': ['Aged', 'Breast Neoplasms', 'Cohort Studies', 'Ethnic Groups', 'Female', 'Humans', 'Mammography', 'Middle Aged', 'Norway', 'Population Groups', 'Registries', 'Retrospective Studies'], 'reasoning_required_pred': ['y', 'e', 's'], 'reasoning_free_pred': ['y', 'e', 's']}, 'long_answer': 'Despite a lower risk of breast cancer, the Sami attended the NBCSP more frequently than the control group. The recall and cancer detection rate was lower among the Sami compared with the non-Sami group.', 'final_decision': 'yes'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2025-12-21 12:00:44.701\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 25691513, 'question': 'Do communication disorders extend to musical messages?', 'context': {'contexts': ['Effective musical communication requires conveyance of the intended message in a manner perceptible to the receiver. Communication disorders that impair transmitting or decoding of structural features of music (e.g., pitch, timbre) and/or symbolic representation may result in atypical musical communication, which can have a negative impact on music therapy interventions.', 'This study compared recognition of symbolic representation of emotions or movements in music by two groups of children with different communicative characteristics: severe to profound hearing loss (using cochlear implants [CI]) and autism spectrum disorder (ASD). Their responses were compared to those of children with typical-development and normal hearing (TD-NH). Accuracy was examined as a function of communicative status, emotional or movement category, and individual characteristics.', 'Participants listened to recorded musical excerpts conveying emotions or movements and matched them with labels. Measures relevant to auditory and/or language function were also gathered.', 'There was no significant difference between the ASD and TD-NH groups in identification of musical emotions or movements. However, the CI group was significantly less accurate than the other two groups in identification of both emotions and movements. Mixed effects logistic regression revealed different patterns of accuracy for specific emotions as a function of group.'], 'labels': ['BACKGROUND', 'OBJECTIVE', 'METHODS', 'RESULTS'], 'meshes': ['Adolescent', 'Auditory Perception', 'Child', 'Child Development Disorders, Pervasive', 'Cochlear Implantation', 'Communication Disorders', 'Deafness', 'Emotions', 'Female', 'Humans', 'Male', 'Music'], 'reasoning_required_pred': ['y', 'e', 's'], 'reasoning_free_pred': ['m', 'a', 'y', 'b', 'e']}, 'long_answer': 'Conveyance of emotions or movements through music may be decoded differently by persons with different types of communication disorders. Because music is the primary therapeutic tool in music therapy sessions, clinicians should consider these differential abilities when selecting music for clinical interventions focusing on emotions or movement.', 'final_decision': 'yes'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 22%|██▏ | 112/500 [00:21<03:43, 1.74it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2025-12-21 12:00:44.714\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:00:44.738\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2025-12-21 12:00:44.739\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 17453263, 'question': 'Are physician estimates of asthma severity less accurate in black than in white patients?', 'context': {'contexts': ['Racial differences in asthma care are not fully explained by socioeconomic status, care access, and insurance status. Appropriate care requires accurate physician estimates of severity. It is unknown if accuracy of physician estimates differs between black and white patients, and how this relates to asthma care disparities.', 'We hypothesized that: 1) physician underestimation of asthma severity is more frequent among black patients; 2) among black patients, physician underestimation of severity is associated with poorer quality asthma care.', \"We conducted a cross-sectional survey among adult patients with asthma cared for in 15 managed care organizations in the United States. We collected physicians' estimates of their patients' asthma severity. Physicians' estimates of patients' asthma as being less severe than patient-reported symptoms were classified as underestimates of severity.\", 'Frequency of underestimation, asthma care, and communication.', 'Three thousand four hundred and ninety-four patients participated (13% were black). Blacks were significantly more likely than white patients to have their asthma severity underestimated (OR = 1.39, 95% CI 1.08-1.79). Among black patients, underestimation was associated with less use of daily inhaled corticosteroids (13% vs 20%, p<.05), less physician instruction on management of asthma flare-ups (33% vs 41%, p<.0001), and lower ratings of asthma care (p = .01) and physician communication (p = .04).'], 'labels': ['BACKGROUND', 'OBJECTIVE', 'DESIGN, SETTING AND PATIENTS', 'MEASUREMENTS', 'RESULTS'], 'meshes': ['Adult', 'African Americans', 'Asthma', 'Clinical Competence', 'Cohort Studies', 'Cross-Sectional Studies', 'European Continental Ancestry Group', 'Female', 'Humans', 'Male', 'Middle Aged', 'Odds Ratio', 'Physician-Patient Relations', 'Quality of Health Care', 'Severity of Illness Index'], 'reasoning_required_pred': ['n', 'o'], 'reasoning_free_pred': ['y', 'e', 's']}, 'long_answer': \"Biased estimates of asthma severity may contribute to racially disparate asthma care. Interventions to improve physicians' assessments of asthma severity and patient-physician communication may minimize racial disparities in asthma care.\", 'final_decision': 'yes'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:00:44.745\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:00:44.861\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 17562682, 'question': 'Is Panton-Valentine leucocidin associated with the pathogenesis of Staphylococcus aureus bacteraemia in the UK?', 'context': {'contexts': ['The morbidity and mortality associated with Panton-Valentine leucocidin (PVL)-positive Staphylococcus aureus suggest that this toxin is a key marker of disease severity. Nevertheless, the importance of PVL in the pathogenesis of primary bacteraemia caused by S. aureus is uncertain. We have determined the prevalence of PVL-encoding genes among isolates of S. aureus from bacteraemic patients.', 'Consecutive bacteraemia isolates of S. aureus (n=244) from patients hospitalized in 25 centres in the UK and Ireland during 2005 were screened for PVL and mecA genes. PVL-positive isolates were characterized by toxin gene profiling, PFGE, spa-typing and MIC determinations for a range of antimicrobials.', 'Four out of 244 isolates (1.6%) were PVL-positive and susceptible to oxacillin [methicillin-susceptible S. aureus (MSSA)]. Eighty-eight out of 244 (36%) were oxacillin-resistant (methicillin-resistant S. aureus), but none was PVL-positive. The four patients (two males: 30 and 33 years; two females: 62 and 80 years) had infection foci of: skin and soft tissue, unknown, indwelling line, and surgical site, and were located at one centre in Wales, one in England and two in Ireland. One of four PVL-positive isolates was resistant to penicillin and fusidic acid, the remainder were susceptible to all antibiotics tested. Genotypic analyses showed that the four isolates represented three distinct strains; the two isolates from Ireland were related.'], 'labels': ['BACKGROUND', 'METHODS', 'RESULTS'], 'meshes': ['Adolescent', 'Adult', 'Aged', 'Aged, 80 and over', 'Anti-Bacterial Agents', 'Bacteremia', 'Bacterial Proteins', 'Bacterial Toxins', 'Child', 'Child, Preschool', 'Exotoxins', 'Female', 'Fusidic Acid', 'Humans', 'Infant', 'Ireland', 'Leukocidins', 'Male', 'Methicillin Resistance', 'Microbial Sensitivity Tests', 'Middle Aged', 'Oxacillin', 'Penicillin Resistance', 'Penicillin-Binding Proteins', 'Staphylococcal Infections', 'Staphylococcus aureus', 'United Kingdom'], 'reasoning_required_pred': ['m', 'a', 'y', 'b', 'e'], 'reasoning_free_pred': ['n', 'o']}, 'long_answer': 'We found that 1.6% of S. aureus (all MSSA) from bacteraemic patients were PVL-positive. This low incidence suggests that PVL-positive S. aureus are of no particular significance as causative agents of S. aureus bacteraemia.', 'final_decision': 'maybe'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 24%|██▎ | 118/500 [00:21<01:51, 3.42it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 24%|██▍ | 120/500 [00:21<01:33, 4.07it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:00:45.137\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 16100194, 'question': 'Are physicians aware of the side effects of angiotensin-converting enzyme inhibitors?', 'context': {'contexts': ['Angiotensin-converting enzyme inhibitors (ACE-I) are considered safe, but they are associated with characteristic side effects, namely cough and angioedema, usually requiring discontinuation. We perceived that referrals for these side effects have become more and more frequent; therefore, we evaluated the degree of knowledge on the safety of ACE-I in different medical categories.', 'A questionnaire (13 questions) on side effects of ACE-I was posted to physicians.', 'Everyday clinical practice.', 'Cardiologists, allergists, and general practitioners (GPs) from the National Healthcare System.', 'Three hundred twelve physicians were contacted, and 154 returned questionnaires that could be analyzed. Of the 154 physicians (mean age, 45 years) 48 were cardiologists, 52 were GPs, and 54 were allergists. The percentage of correct answers was low: 31.9% for cardiologists, 40% for GPs, and 33% for allergists. Thus, GPs provided a significantly higher percentage of correct answers with respect to the remaining categories (p = 0.05). The lower rate of correct answers (0 to 15.9%) concerned the time of onset of cough and the action to take. Cardiologists seemed to be less aware of the fact that angiotensin receptor blockers (sartans) can cross-react with ACE-I.'], 'labels': ['STUDY OBJECTIVE', 'DESIGN', 'SETTING', 'PARTICIPANTS', 'MEASUREMENT AND RESULTS'], 'meshes': ['Adult', 'Allergy and Immunology', 'Angiotensin-Converting Enzyme Inhibitors', 'Cardiology', 'Family Practice', 'Health Knowledge, Attitudes, Practice', 'Humans', 'Middle Aged', \"Practice Patterns, Physicians'\", 'Surveys and Questionnaires'], 'reasoning_required_pred': ['n', 'o'], 'reasoning_free_pred': ['n', 'o']}, 'long_answer': 'Overall, there was a poor knowledge of the side effects of ACE-I. This may account for the increased referrals for chronic cough and angioedema.', 'final_decision': 'no'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2025-12-21 12:00:45.221\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 18322741, 'question': 'Does laparoscopic surgery decrease the risk of atrial fibrillation after foregut surgery?', 'context': {'contexts': ['Atrial fibrillation, which occurs in 12% of all major foregut surgeries, can prolong hospital stay and increase morbidity. Minimally invasive techniques in foregut surgery have been suggested to cause less tissue trauma. We examined the factors associated with new-onset atrial fibrillation after foregut surgery at our institution.', 'We retrospectively examined the records of 154 adult patients who underwent major foregut surgery which included esophagectomy, partial or total gastrectomy, redo Heller myotomy, redo or transthoracic fundoplications. Univariate and multivariate logistic regression analysis with standard modeling techniques were performed to determine risk factors for new-onset atrial fibrillation.', 'Of the 154 patients, 14 patients developed new-onset atrial fibrillation with a higher mean age of 67.1 years (+/-8.8 years) versus 56.4 years (+/-14.1 years) (p = 0.006). Laparoscopic (p = 0.004) and nonthoracic surgeries (p = 0.01) were associated with lower risk of atrial fibrillation. Patients with atrial fibrillation had received more fluid (6.5 +/- 2.8 liters versus 5.3 +/- 2.0 liters) and had longer operations (370 +/- 103 min versus 362 +/- 142 min), none of which were statistically significant. The average intensive care length of stay of patients was longer: 7.5 +/- 6.8 days versus 4.0 +/- 7.1 days (p = 0.004). Multivariate analysis revealed an association of atrial fibrillation with age (OR 1.08, 95% CI 1.02-1.14, p = 0.01), and laparoscopic surgery (OR 0.09, 95% CI 0.01-0.95, p = 0.04) after adjusting for surgery type.'], 'labels': ['BACKGROUND', 'METHODS', 'RESULTS'], 'meshes': ['Adult', 'Age Factors', 'Aged', 'Atrial Fibrillation', 'Cohort Studies', 'Digestive System Surgical Procedures', 'Female', 'Humans', 'Laparoscopy', 'Length of Stay', 'Male', 'Middle Aged', 'Retrospective Studies', 'Risk Factors'], 'reasoning_required_pred': ['y', 'e', 's'], 'reasoning_free_pred': ['y', 'e', 's']}, 'long_answer': 'Laparoscopic surgery is associated with lower risk of atrial fibrillation in foregut surgery. Development of atrial fibrillation is associated with increased length of intensive care stay. We recommend a prospective trial to confirm our findings.', 'final_decision': 'yes'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 24%|██▍ | 122/500 [00:21<01:22, 4.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2025-12-21 12:00:47.690\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 6 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:00:47.817\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 6 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 25%|██▍ | 124/500 [00:24<02:59, 2.10it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:00:47.856\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 27448572, 'question': 'Is duration of psychological treatment for depression related to return into treatment?', 'context': {'contexts': ['There is increasing pressure on mental health providers to reduce the duration of treatments, while retaining level of quality and effectiveness. The risk is that the population is underserved and therefore needs new treatment episodes. The primary aim of this study was to investigate whether duration of treatment and return into mental health care were related.', \"This study examined Dutch patients with an initial treatment episode in 2009 or 2010 in specialized mental health settings for depressive disorder (N\\xa0=\\xa085,754). Follow-up data about treatment episodes were available up until 2013. The data set included demographic (age, gender), and clinical factors (comorbidity with other DSM-IV Axis; scores on the 'Global Assessment of Functioning'). Cox regression analyses were used to assess whether duration of treatment and relapse into mental health care were related.\", 'The majority of patients did not return into mental health care (86\\xa0%). Patients with a shorter duration of treatment (5-250\\xa0min; 251-500\\xa0min and 751-1000\\xa0min) were slightly more likely to return (reference group:>1000\\xa0min) (HR 1.19 95\\xa0% CI 1.13-1.26; HR 1.11 95\\xa0% CI 1.06-1.17; HR 1.18 95\\xa0% CI 1.11-1.25), adjusted for demographic and clinical variables.'], 'labels': ['PURPOSE', 'METHODS', 'RESULTS'], 'meshes': ['Adult', 'Depression', 'Depressive Disorder', 'Diagnostic and Statistical Manual of Mental Disorders', 'Female', 'Humans', 'Male', 'Mental Health Services', 'Middle Aged', 'Psychotherapy', 'Recurrence', 'Time Factors'], 'reasoning_required_pred': ['y', 'e', 's'], 'reasoning_free_pred': ['y', 'e', 's']}, 'long_answer': 'The results suggest that a longer duration of treatment may prevent return into mental health care in some groups. However, because of the design of the study, no causal inference can be drawn. Further research, preferably in a RCT, is needed to determine whether the trend towards lower intensity treatments is associated with repeated mental health care use.', 'final_decision': 'yes'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 6 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2025-12-21 12:00:47.986\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 18540901, 'question': 'Transient tachypnea of the newborn (TTN): a role for polymorphisms in the beta-adrenergic receptor (ADRB) encoding genes?', 'context': {'contexts': ['DNA was collected for genotyping from 73 term newborns suffering from TTN and 55 healthy controls from a Caucasian cohort.', 'TTN infants were more likely to be male (70% vs. 49%; p<0.05), had a lower mean birthweight (3120 +/- 450 vs. 3396 +/- 504 g; p<0.001) and gestational age (GA) (38.4 +/- 1.2 vs. 39.4 +/- 1.3 weeks; p<0.001) and were more often delivered by caesarean section (CS) (71% vs. 26%; p<0.001). The beta1Ser49Gly polymorphism differed significantly between cases and controls. Multivariate analysis provided beta1Gly49 homozygotes with higher risk for TTN (OR 18.5; 95%CI 1.5-229; p = 0.023) than beta1Ser49 allele carrier. Further analysis showed significant association of T-47C, A46G, C79G and C491T (TACC) haplotype in ADRB2 gene with TTN (p = 0.048).'], 'labels': ['METHODS', 'RESULTS'], 'meshes': ['Apgar Score', 'Case-Control Studies', 'DNA', 'Female', 'Genotype', 'Gestational Age', 'Haplotypes', 'Humans', 'Infant, Newborn', 'Male', 'Polymorphism, Single Nucleotide', 'Receptors, Adrenergic, beta', 'Respiration Disorders', 'Time Factors'], 'reasoning_required_pred': ['y', 'e', 's'], 'reasoning_free_pred': ['y', 'e', 's']}, 'long_answer': 'We conclude that beta1Gly49 homozygosity and TACC haplotype of ADRB2 gene, both loss-of-function genetic variations, may predispose to TTN.', 'final_decision': 'yes'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 25%|██▌ | 126/500 [00:24<02:21, 2.65it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2025-12-21 12:00:48.255\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 21823940, 'question': 'Department of Transportation vs self-reported data on motor vehicle collisions and driving convictions for stroke survivors: do they agree?', 'context': {'contexts': [\"Research on stroke survivors' driving safety has typically used either self-reports or government records, but the extent to which the 2 may differ is not known. We compared government records and self-reports of motor vehicle collisions and driving convictions in a sample of stroke survivors.\", 'The 56 participants were originally recruited for a prospective study on driving and community re-integration post-stroke; the study population consisted of moderately impaired stroke survivors without severe communication disorders who had been referred for a driving assessment. The driving records of the 56 participants for the 5 years before study entry and the 1-year study period were acquired with written consent from the Ministry of Transportation of Ontario (MTO), Canada. Self-reports of collisions and convictions were acquired via a semistructured interview and then compared with the MTO records.', 'Forty-three participants completed the study. For 7 (13.5%) the MTO records did not match the self-reports regarding collision involvement, and for 9 (17.3%) the MTO records did not match self-reports regarding driving convictions. The kappa coefficient for the correlation between MTO records and self-reports was 0.52 for collisions and 0.47 for convictions (both in the moderate range of agreement). When both sources of data were consulted, up to 56 percent more accidents and up to 46 percent more convictions were identified in the study population in the 5 years before study entry compared to when either source was used alone.'], 'labels': ['OBJECTIVE', 'METHODS', 'RESULTS'], 'meshes': ['Accidents, Traffic', 'Aged', 'Automobile Driving', 'Female', 'Government Agencies', 'Humans', 'Male', 'Middle Aged', 'Ontario', 'Prospective Studies', 'Records as Topic', 'Reproducibility of Results', 'Safety', 'Self Report', 'Stroke', 'Survivors'], 'reasoning_required_pred': ['n', 'o'], 'reasoning_free_pred': ['n', 'o']}, 'long_answer': 'In our population of stroke survivors, self-reports of motor vehicle collisions and driving convictions differed from government records. In future studies, the use of both government and self-reported data would ensure a more accurate picture of driving safety post-stroke.', 'final_decision': 'no'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 26%|██▌ | 128/500 [00:24<02:06, 2.93it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2025-12-21 12:00:53.291\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 26%|██▌ | 129/500 [00:29<06:32, 1.06s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:00:53.355\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:00:53.739\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 26%|██▌ | 131/500 [00:30<04:53, 1.26it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:00:53.861\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 26%|██▋ | 132/500 [00:30<04:07, 1.48it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:00:53.861\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2025-12-21 12:00:53.869\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:00:53.886\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:00:53.903\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:00:53.903\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:00:53.998\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 28%|██▊ | 138/500 [00:30<01:39, 3.63it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:00:57.964\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:00:57.993\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 28%|██▊ | 140/500 [00:34<03:58, 1.51it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:00:58.120\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:00:58.391\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 17621202, 'question': 'Does shaving the incision site increase the infection rate after spinal surgery?', 'context': {'contexts': ['A prospective randomized clinical study.', 'To determine whether shaving the incision site before spinal surgery causes postsurgical infection.', 'Spine surgeons usually shave the skin of the incision site immediately before surgery is performed. However, evidence from some surgical series suggests that presurgical shaving may increase the postsurgical infection rate. To our knowledge, no previously published studies have addressed this issue.', 'A total of 789 patients scheduled to undergo spinal surgery were randomly allocated into 2 groups: those in whom the site of operation was shaved immediately before surgery (shaved group; 371 patients) and the patients in whom presurgical shaving was not performed (unshaved group; 418 patients). The mean duration of anesthesia and the infection rates in both groups were recorded and compared.', 'The duration of anesthesia did not differ in the 2 groups (P>0.05). A postoperative infection developed in 4 patients in the shaved group and in 1 patient in the nonshaved group (P<0.01).'], 'labels': ['STUDY DESIGN', 'OBJECTIVE', 'SUMMARY OF BACKGROUND DATA', 'METHODS', 'RESULTS'], 'meshes': ['Adult', 'Anti-Bacterial Agents', 'Dermatologic Surgical Procedures', 'Double-Blind Method', 'Female', 'Humans', 'Male', 'Middle Aged', 'Neurosurgical Procedures', 'Preoperative Care', 'Prospective Studies', 'Skin', 'Spinal Diseases', 'Surgical Wound Infection'], 'reasoning_required_pred': ['y', 'e', 's'], 'reasoning_free_pred': ['m', 'a', 'y', 'b', 'e']}, 'long_answer': 'The shaving of the incision site immediately before spinal surgery may increase the rate of postoperative infection.', 'final_decision': 'maybe'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2025-12-21 12:00:58.392\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 28%|██▊ | 142/500 [00:34<03:17, 1.81it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2025-12-21 12:00:58.402\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 12920330, 'question': 'Do somatic complaints predict subsequent symptoms of depression?', 'context': {'contexts': ['Evidence suggests substantial comorbidity between symptoms of somatization and depression in clinical as well as nonclinical populations. However, as most existing research has been retrospective or cross-sectional in design, very little is known about the specific nature of this relationship. In particular, it is unclear whether somatic complaints may heighten the risk for the subsequent development of depressive symptoms.', 'We report findings on the link between symptoms of somatization (assessed using the SCL-90-R) and depression 5 years later (assessed using the CES-D) in an initially healthy cohort of community adults, based on prospective data from the RENO Diet-Heart Study.', 'Gender-stratified multiple regression analyses revealed that baseline CES-D scores were the best predictors of subsequent depressive symptoms for men and women. Baseline scores on the SCL-90-R somatization subscale significantly predicted subsequent self-reported symptoms of depressed mood 5 years later, but only in women. However, somatic complaints were a somewhat less powerful predictor than income and age.'], 'labels': ['BACKGROUND', 'METHODS', 'RESULTS'], 'meshes': ['Adult', 'Aged', 'Comorbidity', 'Demography', 'Depression', 'Female', 'Humans', 'Male', 'Middle Aged', 'Primary Health Care', 'Prospective Studies', 'Somatoform Disorders'], 'reasoning_required_pred': ['y', 'e', 's'], 'reasoning_free_pred': ['m', 'a', 'y', 'b', 'e']}, 'long_answer': \"Our findings suggest that somatic complaints may represent one, but not necessarily the most important, risk factor for the subsequent development of depressive symptoms in women in nonclinical populations. The results also highlight the importance of including social variables in studies on women's depression as well as conducting additional research to further examine predictors of depressive symptoms in men.\", 'final_decision': 'maybe'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 29%|██▉ | 145/500 [00:34<02:12, 2.67it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:00:58.645\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:01.022\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 29%|██▉ | 147/500 [00:37<03:29, 1.69it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:01.435\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 12 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:01.486\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 12 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 30%|██▉ | 149/500 [00:37<02:54, 2.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:02.030\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 11 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2025-12-21 12:01:02.032\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 11 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:02.032\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 12913878, 'question': 'Locoregional opening of the rodent blood-brain barrier for paclitaxel using Nd:YAG laser-induced thermo therapy: a new concept of adjuvant glioma therapy?', 'context': {'contexts': ['Nd:YAG laser-induced thermo therapy (LITT) of rat brains is associated with blood-brain barrier (BBB) permeability changes. We address the question of whether LITT-induced locoregional disruption of the BBB could possibly allow a locoregional passage of chemotherapeutic agents into brain tissue to treat malignant glioma.STUDY DESIGN/', 'CD Fischer rats were subject to LITT of the left forebrain. Disruption of the BBB was analyzed using Evans blue and immunohistochemistry (IH). Animals were perfused with paclitaxel, and high-pressure liquid chromatography (HPLC) was employed to analyze the content of paclitaxel in brain and plasma samples.', 'LITT induces an opening of the BBB as demonstrated by locoregional extravasation of Evans blue, C3C, fibrinogen, and IgM. HPLC proved the passage of paclitaxel across the disrupted BBB.'], 'labels': ['BACKGROUND AND OBJECTIVES', 'MATERIALS AND METHODS', 'RESULTS'], 'meshes': ['Administration, Topical', 'Animals', 'Antineoplastic Agents, Phytogenic', 'Blood-Brain Barrier', 'Brain Neoplasms', 'Chemotherapy, Adjuvant', 'Glioma', 'Hyperthermia, Induced', 'Laser Therapy', 'Neurosurgical Procedures', 'Paclitaxel', 'Permeability', 'Rats', 'Rats, Inbred F344'], 'reasoning_required_pred': ['y', 'e', 's'], 'reasoning_free_pred': ['y', 'e', 's']}, 'long_answer': 'LITT induces a locoregional passage of chemotherapeutic agents into the brain tissue. This is of potential interest for the treatment of brain tumors.', 'final_decision': 'yes'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 11 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 30%|███ | 150/500 [00:38<02:57, 1.98it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2025-12-21 12:01:02.049\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 11 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:13.252\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 31%|███ | 154/500 [00:49<09:10, 1.59s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:13.477\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 28127977, 'question': 'Are serum leptin levels a prognostic factor in advanced lung cancer?', 'context': {'contexts': ['There are 71 previously untreated patients with cytological or histological evidence of primary lung cancer who were admitted to the oncology department between November 2013 and August 2014. Forty-five healthy individuals with age, sex and BMI matching the lung cancer patients, were recruited to take part in the study as a control group. Leptin levels were measured quantitatively by using a microELISA kit.', 'The serum leptin levels at diagnosis were significantly lower in lung cancer patients than those in control subjects (4.75±4.91 ng/ml, 9.67±8.02 ng/ml; p<0.001). We did not find any significant difference in leptin values related to clinicopathological parameters such as ECOG PS, weight loss, histological type, disease stage and TNM classification. Nevertheless, we demonstrated a significant correlation between serum leptin levels and BMI in lung cancer patients (correlation coefficient: 0.303; p>0.010). The analysis of serum leptin values did not show any association with the overall survival of the patients.'], 'labels': ['MATERIAL AND METHODS', 'RESULTS'], 'meshes': ['Aged', 'Biomarkers, Tumor', 'Body Mass Index', 'Female', 'Humans', 'Leptin', 'Lung Neoplasms', 'Male', 'Middle Aged', 'Neoplasm Staging', 'Prognosis', 'Reference Values', 'Statistics as Topic'], 'reasoning_required_pred': ['n', 'o'], 'reasoning_free_pred': ['n', 'o']}, 'long_answer': 'Our results showed that the serum leptin level has no prognostic indications in advanced lung cancer patients. Leptin is decreased in lung cancer, and there is lack of correlation with tumour‑related factors including prognosis. Therefore, leptin is not a useful clinical marker in lung cancer (Tab. 2, Fig. 2, Ref. 22).', 'final_decision': 'no'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 31%|███ | 155/500 [00:49<08:01, 1.39s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2025-12-21 12:01:13.535\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:13.561\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:13.609\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 10223070, 'question': 'Is perforation of the appendix a risk factor for tubal infertility and ectopic pregnancy?', 'context': {'contexts': ['To critically assess the evidence that appendiceal perforation is a risk factor for subsequent tubal infertility or ectopic pregnancy.', 'Epidemiologic studies investigating the relationship between appendectomy and infertility or ectopic pregnancy were identified by searching the MEDLINE database from 1966 to 1997. Appropriate citations were also extracted from a manual search of the bibliographies of selected papers.', 'Twenty-three articles were retrieved. Only 4 presented original data including comparisons to a nonexposed control group and they form the basis for this study.', 'Because the raw data or specific techniques of data analysis were not always explicitly described, indices of risk for exposure were extracted from the data as presented and were analysed without attempting to convert them to a common measure.', 'Articles were assessed according to the criteria of the Evidence-Based Medicine Working Group for evaluating articles on harm. Review of the literature yielded estimates of the risk of adverse fertility outcomes ranging from 1.6 (95% confidence interval [CI] 1.1 to 2.5) for ectopic pregnancy after an appendectomy to 4.8 (95% CI 1.5 to 14.9) for tubal infertility from perforation of the appendix. Recall bias, and poor adjustment for confounding variables in some reports, weakened the validity of the studies.'], 'labels': ['OBJECTIVE', 'DATA SOURCES', 'STUDY SELECTION', 'DATA EXTRACTION', 'DATA SYNTHESIS'], 'meshes': ['Appendicitis', 'Causality', 'Evidence-Based Medicine', 'Female', 'Humans', 'Infertility, Female', 'Intestinal Perforation', 'Pregnancy', 'Pregnancy, Ectopic', 'Research Design', 'Risk Factors'], 'reasoning_required_pred': ['y', 'e', 's'], 'reasoning_free_pred': ['n', 'o']}, 'long_answer': 'The methodologic weaknesses of the studies do not permit acceptance of increased risk of tubal pregnancy or infertility as a consequence of perforation of the appendix, so a causal relationship cannot be supported by the data currently available. Only a well-designed case-control study with unbiased ascertainment of exposure and adjustment for confounding variables will provide a definitive answer.', 'final_decision': 'maybe'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 32%|███▏ | 158/500 [00:50<05:00, 1.14it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2025-12-21 12:01:13.643\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:13.678\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:13.781\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 32%|███▏ | 162/500 [00:50<02:57, 1.90it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 33%|███▎ | 164/500 [00:51<03:18, 1.69it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 33%|███▎ | 167/500 [00:52<02:25, 2.29it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 34%|███▎ | 168/500 [00:52<02:14, 2.47it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 34%|███▍ | 169/500 [00:53<02:37, 2.11it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 34%|███▍ | 170/500 [00:53<02:35, 2.13it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 35%|███▍ | 174/500 [00:54<01:13, 4.46it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 35%|███▌ | 177/500 [00:54<00:50, 6.45it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 37%|███▋ | 183/500 [00:54<00:28, 10.94it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 37%|███▋ | 185/500 [00:55<00:44, 7.06it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 38%|███▊ | 189/500 [00:56<00:49, 6.26it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 38%|███▊ | 191/500 [00:56<00:58, 5.27it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:20.461\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 17008699, 'question': 'Do patients with localized prostate cancer treatment really want more aggressive treatment?', 'context': {'contexts': ['Examine whether patients with prostate cancer choose the more aggressive of two radiotherapeutic options, whether this choice is reasoned, and what the determinants of the choice are.', 'One hundred fifty patients with primary prostate cancer (T(1-3)N(0)M(0)) were informed by means of a decision aid of two treatment options: radiotherapy with 70 Gy versus 74 Gy. The latter treatment is associated with more cure and more toxicity. The patients were asked whether they wanted to choose, and if so which treatment they preferred. They also assigned importance weights to the probability of various outcomes, such as survival, cure and adverse effects. Patients who wanted to choose their own treatment (n = 119) are described here.', 'The majority of these patients (75%) chose the lower radiation dose. Their choice was highly consistent (P20%). All patients with short bowel (4) and 60% (6/10) of patients with a delay of enteral feeding more than 14 days postoperatively had cholestasis. Ten patients (36%) proceeded with in-depth evaluations for cholestasis, with 8 (28%) undergoing liver biopsy. No patient had biliary atresia. No deaths were related to isolated cholestasis/cirrhosis. Cholestasis resolved spontaneously in all the survivors.'], 'labels': ['PURPOSE', 'METHODS', 'RESULTS'], 'meshes': ['Cholestasis', 'Female', 'Gestational Age', 'Humans', 'Incidence', 'Infant, Newborn', 'Infant, Premature', 'Intestinal Atresia', 'Intestine, Small', 'Male', 'Parenteral Nutrition', 'Retrospective Studies', 'Risk Factors'], 'reasoning_required_pred': ['m', 'a', 'y', 'b', 'e'], 'reasoning_free_pred': ['n', 'o']}, 'long_answer': 'Small bowel atresia is frequently associated with postoperative cholestasis that will resolve with time. We recommend a more selective and expectant approach to SBA-associated cholestasis to minimize unnecessary investigations.', 'final_decision': 'no'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2025-12-21 12:01:23.386\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 43%|████▎ | 215/500 [00:59<00:40, 7.07it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 43%|████▎ | 217/500 [01:00<01:07, 4.16it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 44%|████▎ | 218/500 [01:01<01:07, 4.19it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 44%|████▍ | 220/500 [01:01<01:04, 4.36it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 44%|████▍ | 221/500 [01:01<00:58, 4.80it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 44%|████▍ | 222/500 [01:01<01:04, 4.33it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 45%|████▍ | 224/500 [01:02<00:54, 5.06it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 45%|████▌ | 226/500 [01:02<00:40, 6.69it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 46%|████▌ | 229/500 [01:02<00:40, 6.66it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 46%|████▋ | 232/500 [01:03<00:31, 8.51it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 47%|████▋ | 234/500 [01:03<00:32, 8.29it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 47%|████▋ | 236/500 [01:03<00:32, 8.17it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 48%|████▊ | 241/500 [01:03<00:19, 13.06it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 49%|████▊ | 243/500 [01:04<00:22, 11.61it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 49%|████▉ | 246/500 [01:04<00:39, 6.42it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 49%|████▉ | 247/500 [01:05<00:43, 5.81it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 50%|████▉ | 249/500 [01:05<00:51, 4.90it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 50%|█████ | 252/500 [01:05<00:33, 7.34it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 51%|█████ | 255/500 [01:06<00:33, 7.40it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 52%|█████▏ | 258/500 [01:06<00:24, 9.87it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 52%|█████▏ | 261/500 [01:06<00:18, 12.64it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 53%|█████▎ | 263/500 [01:07<00:30, 7.89it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 53%|█████▎ | 266/500 [01:07<00:26, 8.88it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 54%|█████▍ | 269/500 [01:07<00:26, 8.80it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 54%|█████▍ | 272/500 [01:08<00:27, 8.18it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 55%|█████▍ | 274/500 [01:08<00:27, 8.22it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 55%|█████▌ | 275/500 [01:08<00:29, 7.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 55%|█████▌ | 277/500 [01:08<00:27, 8.03it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 56%|█████▌ | 280/500 [01:09<00:26, 8.15it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 57%|█████▋ | 285/500 [01:09<00:19, 11.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 58%|█████▊ | 288/500 [01:09<00:18, 11.72it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 58%|█████▊ | 290/500 [01:10<00:21, 9.63it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 59%|█████▉ | 294/500 [01:11<00:27, 7.51it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 59%|█████▉ | 296/500 [01:11<00:28, 7.25it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 60%|█████▉ | 299/500 [01:11<00:22, 9.02it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 60%|██████ | 301/500 [01:11<00:23, 8.50it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 61%|██████ | 303/500 [01:12<00:26, 7.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 61%|██████ | 305/500 [01:12<00:22, 8.49it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 62%|██████▏ | 308/500 [01:12<00:24, 7.80it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 62%|██████▏ | 311/500 [01:13<00:24, 7.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 63%|██████▎ | 313/500 [01:13<00:29, 6.41it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 63%|██████▎ | 316/500 [01:14<00:26, 6.87it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 64%|██████▍ | 321/500 [01:14<00:16, 11.07it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 65%|██████▍ | 323/500 [01:14<00:17, 10.19it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 65%|██████▌ | 325/500 [01:14<00:18, 9.52it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 65%|██████▌ | 327/500 [01:17<01:12, 2.40it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2025-12-21 12:01:41.130\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 17704864, 'question': 'Is laparoscopic adrenalectomy safe and effective for adrenal masses larger than 7 cm?', 'context': {'contexts': ['Laparoscopic adrenalectomy (LA) has become the gold standard treatment for small (less than 6 cm) adrenal masses. However, the role of LA for large-volume (more than 6 cm) masses has not been well defined. Our aim was to evaluate, retrospectively, the outcome of LA for adrenal lesions larger than 7 cm.', '18 consecutive laparoscopic adrenalectomies were performed from 1996 to 2005 on patients with adrenal lesions larger than 7 cm.', 'The mean tumor size was 8.3 cm (range 7-13 cm), the mean operative time was 137 min, the mean blood loss was 182 mL (range 100-550 mL), the rate of intraoperative complications was 16%, and in three cases we switched from laparoscopic procedure to open surgery.'], 'labels': ['BACKGROUND', 'PATIENTS AND METHODS', 'RESULTS'], 'meshes': ['Adolescent', 'Adrenal Gland Neoplasms', 'Adrenalectomy', 'Adult', 'Aged', 'Aged, 80 and over', 'Female', 'Humans', 'Laparoscopy', 'Male', 'Middle Aged'], 'reasoning_required_pred': ['y', 'e', 's'], 'reasoning_free_pred': ['y', 'e', 's']}, 'long_answer': 'LA for adrenal masses larger than 7 cm is a safe and feasible technique, offering successful outcome in terms of intraoperative and postoperative morbidity, hospital stay and cosmesis for patients; it seems to replicate open surgical oncological principles demonstrating similar outcomes as survival rate and recurrence rate, when adrenal cortical carcinoma were treated. The main contraindication for this approach is the evidence, radiologically and intraoperatively, of local infiltration of periadrenal tissue.', 'final_decision': 'yes'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 66%|██████▌ | 328/500 [01:17<01:13, 2.34it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2025-12-21 12:01:44.880\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 26399179, 'question': 'Eyelid-parotid metastasis: do we screen for coexisting masses?', 'context': {'contexts': ['To report three cases illustrating that it is not unusual for a primary eyelid tumour to metastasise to the parotid gland and vice versa.', 'Two patients with malignant parotid tumours underwent radical parotidectomy and presented subsequently with eyelid lesions. Biopsy showed that both eyelid lesions were histologically similar to the primary parotid tumour. A third patient was noted to have ipsilateral upper eyelid and parotid gland tumours. Histology and immunocytochemistry were used to differentiate the primary tumour and the metastasis.'], 'labels': ['OBJECTIVE', 'CASE REPORTS'], 'meshes': ['Adult', 'Aged', 'Breast Neoplasms', 'Carcinoma, Squamous Cell', 'Chemoradiotherapy', 'Diagnosis, Differential', 'Eyelid Neoplasms', 'Female', 'Humans', 'Male', 'Middle Aged', 'Neoplasm Staging', 'Parotid Neoplasms', 'Reconstructive Surgical Procedures', 'Surgical Procedures, Operative'], 'reasoning_required_pred': ['y', 'e', 's'], 'reasoning_free_pred': ['y', 'e', 's']}, 'long_answer': 'These cases illustrate that tumours involving eyelids and parotid glands can present simultaneously or sequentially, and either of these structures could be the focus of primary or metastatic tumour. The important message for oculoplastic and parotid surgeons is to routinely assess both the periocular and parotid area when patients present with a mass in either structure.', 'final_decision': 'yes'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 66%|██████▌ | 329/500 [01:21<03:07, 1.10s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2025-12-21 12:01:44.919\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 9 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:44.933\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:44.942\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 14992556, 'question': 'Artefacts in 24-h pharyngeal and oesophageal pH monitoring: is simplification of pH data analysis feasible?', 'context': {'contexts': ['Ambulatory 24-h dual-channel pharyngeal and oesophageal pH monitoring is the standard test for measuring gastro-oesophageal and gastropharyngeal reflux. Artefacts caused by the intake of food may result in falsely positive gastropharyngeal reflux, which necessitates a manual review of 24-h pH data. The purpose of the study was to investigate the influence of meals and whether leaving out meals affected the reliability of the test.', 'Patients referred for otolaryngological complaints, suspected to have been caused by gastro-oesophageal reflux, underwent 24-h dual-channel pH monitoring. The raw unprocessed pH data were corrected by visual inspection of the 24-h tracings (corrected data), by leaving out meals or meals plus a 2-h postprandrial period.', 'The raw pH data were substantially influenced by artefacts of food intake and pseudoreflux. Data obtained by leaving out meals agreed best with manually corrected data. Many of the falsely positive reflux episodes could be removed, thereby inducing a 9%-18% chance of undetected reflux. When examining the fraction of time supine, manually corrected data and data leaving out meals were fully concordant and detected 79% of patients with gastropharyngeal reflux. However, leaving out meals plus a 2-h postprandrial period resulted in 21%-50% falsely negative tests.'], 'labels': ['BACKGROUND', 'METHODS', 'RESULTS'], 'meshes': ['Adult', 'Aged', 'Aged, 80 and over', 'Artifacts', 'Eating', 'Esophagus', 'Feasibility Studies', 'Female', 'Gastroesophageal Reflux', 'Humans', 'Hydrogen-Ion Concentration', 'Male', 'Middle Aged', 'Monitoring, Ambulatory', 'Pharynx', 'Postprandial Period', 'Prospective Studies', 'Reproducibility of Results', 'Statistics as Topic'], 'reasoning_required_pred': ['y', 'e', 's'], 'reasoning_free_pred': ['m', 'a', 'y', 'b', 'e']}, 'long_answer': 'Leaving out the period of intake of meals and beverages from the raw pH data might be the second best test after the time-consuming visual correction with a small chance of undetected gastropharyngeal reflux. For scientific purposes and when in doubt, it remains necessary to review the computer-generated data manually to discover every gastropharyngeal reflux event.', 'final_decision': 'maybe'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2025-12-21 12:01:45.020\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 67%|██████▋ | 334/500 [01:21<01:17, 2.15it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2025-12-21 12:01:45.262\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 8 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 67%|██████▋ | 336/500 [01:21<01:02, 2.64it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:45.307\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 15151701, 'question': 'Profiling quality of care: Is there a role for peer review?', 'context': {'contexts': ['We sought to develop a more reliable structured implicit chart review instrument for use in assessing the quality of care for chronic disease and to examine if ratings are more reliable for conditions in which the evidence base for practice is more developed.', 'We conducted a reliability study in a cohort with patient records including both outpatient and inpatient care as the objects of measurement. We developed a structured implicit review instrument to assess the quality of care over one year of treatment. 12 reviewers conducted a total of 496 reviews of 70 patient records selected from 26 VA clinical sites in two regions of the country. Each patient had between one and four conditions specified as having a highly developed evidence base (diabetes and hypertension) or a less developed evidence base (chronic obstructive pulmonary disease or a collection of acute conditions). Multilevel analysis that accounts for the nested and cross-classified structure of the data was used to estimate the signal and noise components of the measurement of quality and the reliability of implicit review.', 'For COPD and a collection of acute conditions the reliability of a single physician review was quite low (intra-class correlation = 0.16-0.26) but comparable to most previously published estimates for the use of this method in inpatient settings. However, for diabetes and hypertension the reliability is significantly higher at 0.46. The higher reliability is a result of the reviewers collectively being able to distinguish more differences in the quality of care between patients (p<0.007) and not due to less random noise or individual reviewer bias in the measurement. For these conditions the level of true quality (i.e. the rating of quality of care that would result from the full population of physician reviewers reviewing a record) varied from poor to good across patients.'], 'labels': ['BACKGROUND', 'METHODS', 'RESULTS'], 'meshes': ['Acute Disease', 'Chronic Disease', 'Cohort Studies', 'Continuity of Patient Care', 'Diabetes Mellitus', 'Disease Management', 'Evidence-Based Medicine', 'Health Services Misuse', 'Humans', 'Hypertension', 'Internal Medicine', 'Los Angeles', 'Medical Records', 'Michigan', 'Observer Variation', 'Outcome and Process Assessment (Health Care)', 'Peer Review, Health Care', 'Primary Health Care', 'Pulmonary Disease, Chronic Obstructive', 'Quality Assurance, Health Care', 'Veterans'], 'reasoning_required_pred': ['m', 'a', 'y', 'b', 'e'], 'reasoning_free_pred': ['y', 'e', 's']}, 'long_answer': 'For conditions with a well-developed quality of care evidence base, such as hypertension and diabetes, a single structured implicit review to assess the quality of care over a period of time is moderately reliable. This method could be a reasonable complement or alternative to explicit indicator approaches for assessing and comparing quality of care. Structured implicit review, like explicit quality measures, must be used more cautiously for illnesses for which the evidence base is less well developed, such as COPD and acute, short-course illnesses.', 'final_decision': 'yes'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2025-12-21 12:01:45.321\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 18378554, 'question': 'Are wandering and physically nonaggressive agitation equivalent?', 'context': {'contexts': ['The authors examined equivalence of wandering and physically nonaggressive agitation (PNA) as concepts.', 'A cross-sectional correlational design was used.', 'Participants were recruited from 22 nursing homes and 6 assisted living facilities in two states.', 'Ambulatory residents meeting DSM-IV criteria for dementia (N = 181) were studied.', 'Video-tapes for up to twelve 20-minute observations per participant were coded for wandering using an empirically derived taxonomy of ambulation patterns. Separate raters coded the same tapes for six PNA behaviors on the agitation behavior mapping instrument.', 'Most participants (73.5%) wandered; all showed PNA behaviors. Factor analyses yielded an one-factor solution for wandering (explained variance = 43.66%) and a two-factor solution for PNA (explained variance = 53.45%). Overall wandering correlated significantly with PNA Factor 1 (df =179, r = 0.68, p<0.001) and Factor 2, but at a lower value (df = 179, r = 0.26, p<0.01).'], 'labels': ['OBJECTIVE', 'DESIGN', 'SETTING', 'PARTICIPANTS', 'MEASUREMENTS', 'RESULTS'], 'meshes': ['Aged', 'Assisted Living Facilities', 'Cross-Sectional Studies', 'Homes for the Aged', 'Humans', 'Motor Activity', 'Nursing Homes', 'Patient Selection', 'Psychomotor Agitation', 'Videotape Recording', 'Walking'], 'reasoning_required_pred': ['n', 'o'], 'reasoning_free_pred': ['n', 'o']}, 'long_answer': 'Findings depict wandering and PNA as overlapping, but nonequivalent phenomena. Evidence supporting construct validity of wandering was more robust than that for PNA. Results have implications for accuracy in scientific and clinical detection and labeling of wandering and agitation.', 'final_decision': 'no'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2025-12-21 12:01:45.345\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 8 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:45.423\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 16296668, 'question': 'Can bedside assessment reliably exclude aspiration following acute stroke?', 'context': {'contexts': ['To investigate the ability of a bedside swallowing assessment to reliably exclude aspiration following acute stroke.', 'Consecutive patients admitted within 24 h of stroke onset to two hospitals.', 'A prospective study. Where possible, all patients had their ability to swallow assessed on the day of admission by both a doctor and a speech and language therapist using a standardized proforma. A videofluoroscopy examination was conducted within 3 days of admission.', '94 patients underwent videofluoroscopy; 20 (21%) were seen to be aspirating, although this was not detected at the bedside in 10. In 18 (22%) of the patients the speech and language therapist considered the swallow to be unsafe. In the medical assessment, 39 patients (41%) had an unsafe swallow. Bedside assessment by a speech and language therapist gave a sensitivity of 47%, a specificity of 86%, positive predictive value (PPV) of 50% and a negative predictive value (NPV) of 85% for the presence of aspiration. Multiple logistic regression was used to identify the optimum elements of the bedside assessments for predicting the presence of aspiration. A weak voluntary cough and any alteration in conscious level gave a sensitivity of 75%, specificity of 72%, PPV of 41% and NPV of 91% for aspiration.'], 'labels': ['OBJECTIVE', 'SUBJECTS', 'METHODS', 'RESULTS'], 'meshes': ['Acute Disease', 'Aged', 'Aged, 80 and over', 'Deglutition Disorders', 'Female', 'Fluoroscopy', 'Geriatric Assessment', 'Humans', 'Male', 'Middle Aged', 'Pneumonia, Aspiration', 'Predictive Value of Tests', 'Prospective Studies', 'Sensitivity and Specificity', 'Stroke', 'Videotape Recording'], 'reasoning_required_pred': ['y', 'e', 's'], 'reasoning_free_pred': ['n', 'o']}, 'long_answer': 'Bedside assessment of swallowing lacks the necessary sensitivity to be used as a screening instrument in acute stroke, but there are concerns about the use of videofluoroscopy as a gold standard. The relative importance of aspiration and bedside assessment in predicting complications and outcome needs to be studied.', 'final_decision': 'no'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 68%|██████▊ | 341/500 [01:21<00:34, 4.66it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2025-12-21 12:01:45.462\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 21889895, 'question': 'Will CT ordering practices change if we educate residents about the potential effects of radiation exposure?', 'context': {'contexts': ['The aim of this study was to determine if educating residents about the potential effects of radiation exposure from computed tomographic (CT) imaging alters ordering patterns. This study also explored whether referring physicians are interested in radiation education and was an initial effort to address their CT ordering behavior.', \"Two to four months after a radiologist's lecture on the potential effects of radiation exposure related to CT scans, urology and orthopedic residents were surveyed regarding the number and types of CT scans they ordered, the use of alternative imaging modalities, and whether they used the lecture information to educate patients.\", 'Twenty-one resident lecture attendants completed the survey. The number of CT scans ordered after the lecture stayed constant for 90% (19 of 21) and decreased for 10% (two of 21). The types of CT scans ordered changed after the lecture for 14% (three of 21). Thirty-three percent (seven of 21) reported increases in alternative imaging after the lecture, including 24% (five of 21) reporting increases in magnetic resonance imaging and 19% (four of 21) reporting increases in ultrasound. Patients directed questions about radiation exposure to 57% (12 of 21); 38% (eight of 21) used the lecture information to educate patients. Referring physicians were interested in the topic, and afterward, other physician groups requested radiation education lectures.'], 'labels': ['RATIONALE AND OBJECTIVES', 'MATERIALS AND METHODS', 'RESULTS'], 'meshes': ['Academic Medical Centers', 'Education, Medical, Graduate', 'Humans', 'Internship and Residency', 'Patient Education as Topic', \"Practice Patterns, Physicians'\", 'Radiation Dosage', 'Radiation Protection', 'Radiology', 'Risk', 'Tomography, X-Ray Computed'], 'reasoning_required_pred': ['y', 'e', 's'], 'reasoning_free_pred': ['n', 'o']}, 'long_answer': 'Most clinicians did not change their CT scan ordering after receiving education about radiation from a radiologist. Radiation education allowed clinicians to discuss CT benefits and risks with their patients and to choose appropriate CT protocols. Referring physician groups are interested in this topic, and radiologists should be encouraged to give radiation lectures to them.', 'final_decision': 'no'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2025-12-21 12:01:45.497\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:48.087\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 16872243, 'question': 'Can decisional algorithms replace global introspection in the individual causality assessment of spontaneously reported ADRs?', 'context': {'contexts': [\"In this study, an expert panel assessed causality of adverse reports by using the WHO global introspection (GI) method. The same reports were independently assessed using 15 published algorithms. The causality assessment level 'possible' was considered the lower limit for a report to be considered to be drug related. For a given algorithm, sensitivity was determined by the proportion of reports simultaneously classified as drug related by the algorithm and the GI method. Specificity was measured as the proportion of reports simultaneously considered non-drug related. The analysis was performed for the total sample and within serious or unexpected events.\", 'Five hundred adverse reports were studied. Algorithms presented high rates of sensitivity (average of 93%, positive predictive value of 89%) and low rates of specificity (average of 7%, negative predictive value of 31%).'], 'labels': ['METHOD', 'RESULTS'], 'meshes': ['Adverse Drug Reaction Reporting Systems', 'Algorithms', 'Decision Support Techniques', 'Drug-Related Side Effects and Adverse Reactions', 'Evaluation Studies as Topic', 'Gastrointestinal Diseases', 'Humans', 'Reproducibility of Results', 'Risk Factors', 'Skin Diseases', 'World Health Organization'], 'reasoning_required_pred': ['n', 'o'], 'reasoning_free_pred': ['n', 'o']}, 'long_answer': 'Decisional algorithms are sensitive methods for the detection of ADRs, but they present poor specificity. A reference method was not identified. Algorithms do not replace GI and are not definite alternatives in the individual causality assessment of suspected ADRs.', 'final_decision': 'no'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 69%|██████▉ | 344/500 [01:24<01:03, 2.44it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2025-12-21 12:01:48.115\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 15539888, 'question': 'Is the atopy patch test with house dust mites specific for atopic dermatitis?', 'context': {'contexts': ['The atopy patch test (APT), namely the patch test with aeroallergens, is regarded as specific for patients with atopic dermatitis (AD), but small numbers of positive APT were reported in the past also in atopic subjects without dermatitis and in healthy persons.', 'The aim of this study was to evaluate the response to the APT with house dust mites (HDM) in subjects nonaffected by AD and to compare the outcomes observed in these cases with those pointed out in AD patients, evaluating also the differences between two allergen extracts manufactured at different purifications and concentrations.', 'Forty-seven atopic subjects without eczema (AWE), 33 nonatopic (NA) subjects and 77 adult AD patients were patch tested with an extract of purified bodies of HDM at 20% and with another extract of whole bodies of HDM at 30%, the latter corresponding to 300 microg/g of Der p 1. The reproducibility of APT was also tested in 8 AD patients, in 37 AWE subjects and in 19 NA subjects.', 'Positive responses with extract at 20% were observed in 29 (37.7%) AD, in 5 (10.6%) AWE and in 4 (12.1%) NA subjects. The APT with HDM at 30% was positive in 32 (41.6%) AD, 9 (19.1%) AWE and 4 (12.1%) NA persons. The rates of positivity and the intensity scores of responses were significantly different between AD and non-AD subjects (p<0.01). The reproducibility of the APT in the three groups was satisfactory.'], 'labels': ['BACKGROUND', 'OBJECTIVE', 'METHODS', 'RESULTS'], 'meshes': ['Adolescent', 'Adult', 'Allergens', 'Animals', 'Case-Control Studies', 'Dermatitis, Atopic', 'Female', 'Humans', 'Male', 'Middle Aged', 'Mites', 'Patch Tests', 'Reference Values', 'Risk Assessment', 'Sensitivity and Specificity', 'Severity of Illness Index'], 'reasoning_required_pred': ['y', 'e', 's'], 'reasoning_free_pred': ['y', 'e', 's']}, 'long_answer': 'These observations lead to conclude that the APT with HDM is positive also in non-AD subjects but it is probably more specific for AD.', 'final_decision': 'yes'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2025-12-21 12:01:48.268\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 69%|██████▉ | 346/500 [01:24<00:52, 2.94it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:53.157\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 70%|██████▉ | 348/500 [01:29<02:07, 1.20it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:53.250\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 26134053, 'question': 'Outcome Feedback within Emergency Medicine Training Programs: An Opportunity to Apply the Theory of Deliberate Practice?', 'context': {'contexts': ['Outcome feedback is the process of learning patient outcomes after their care within the emergency department. We conducted a national survey of Canadian Royal College emergency medicine (EM) residents and program directors to determine the extent to which active outcome feedback and follow-up occurred. We also compared the perceived educational value of outcome feedback between residents and program directors.', \"We distributed surveys to all Royal College-accredited adult and pediatric EM training programs using a modified Dillman method. We analyzed the data using student's t-test for continuous variables and Fisher's exact test for categorical variables.\", 'We received 210 completed surveys from 260 eligible residents (80.8%) and 21 of 24 program directors (87.5%) (overall 81.3%). Mandatory active outcome feedback was not present in any EM training program for admitted or discharged patients (0/21). Follow-up was performed electively by 89.4% of residents for patients admitted to the hospital, and by 44.2% of residents for patients discharged home. A majority of residents (76.9%) believed that patient follow-up should be mandatory compared to 42.9% of program directors (p=0.002). The perceived educational value of outcome feedback was 5.8/7 for residents and 5.1/7 for program directors (difference 0.7; p=0.002) based on a seven-point Likert scale (1=not important; 7=very important).'], 'labels': ['OBJECTIVES', 'METHODS', 'RESULTS'], 'meshes': ['Adult', 'Canada', 'Child', 'Clinical Competence', 'Curriculum', 'Emergency Medicine', 'Female', 'Humans', 'Internship and Residency', 'Male', 'Models, Educational', 'Surveys and Questionnaires'], 'reasoning_required_pred': ['y', 'e', 's'], 'reasoning_free_pred': ['m', 'a', 'y', 'b', 'e']}, 'long_answer': 'While Canadian EM training programs do not mandate follow-up, it is performed electively by the majority of residents surveyed. Residents place a significantly greater educational value on outcome feedback than their program directors, and believe that follow-up should be a mandatory component of EM residencies.', 'final_decision': 'maybe'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2025-12-21 12:01:53.341\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 70%|███████ | 350/500 [01:29<01:37, 1.54it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:53.409\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:53.443\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 70%|███████ | 352/500 [01:29<01:12, 2.04it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:53.540\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2025-12-21 12:01:53.541\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 71%|███████ | 354/500 [01:29<00:54, 2.70it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:53.633\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:54.023\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 71%|███████ | 356/500 [01:30<00:47, 3.00it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:54.050\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:58.222\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 72%|███████▏ | 358/500 [01:34<01:59, 1.19it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:58.404\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 72%|███████▏ | 359/500 [01:34<01:42, 1.37it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:58.490\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:58.573\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 72%|███████▏ | 361/500 [01:35<01:11, 1.93it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:58.595\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:01:58.662\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:02:01.588\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 12 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 73%|███████▎ | 364/500 [01:38<01:37, 1.39it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:02:01.638\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 19468282, 'question': 'Is determination between complete and incomplete traumatic spinal cord injury clinically relevant?', 'context': {'contexts': ['Prospective multicenter longitudinal cohort study.', 'To validate the prognostic value of the acute phase sacral sparing measurements with regard to chronic phase-independent ambulation in patients with traumatic spinal cord injury (SCI).', 'European Multicenter Study of Human Spinal Cord Injury (EM-SCI).', 'In 432 patients, acute phase (0-15 days) American Spinal Injury Association (ASIA)/International Spinal Cord Society neurological standard scale (AIS) grades, ASIA sacral sparing measurements, which are S4-5 light touch (LT), S4-5 pin prick (PP), anal sensation and voluntary anal contraction; and chronic phase (6 or 12 months) indoor mobility Spinal Cord Independence Measure (SCIM) measurements were analyzed. Calculations of positive and negative predictive values (PPV/NPV) as well as univariate and multivariate logistic regressions were performed in all four sacral sparing criteria. The area under the receiver-operating characteristic curve (AUC) ratios of all regression equations was calculated.', 'To achieve independent ambulation 1-year post injury, a normal S4-5 PP score showed the best PPV (96.5%, P<0.001, 95% confidence interval (95% CI): 87.9-99.6). Best NPV was reported in the S4-5 LT score (91.7%, P<0.001, 95% CI: 81.6-97.2). The use of the combination of only voluntary anal contraction and the S4-5 LT and PP sensory scores (AUC: 0.906, P<0.001, 95% CI: 0.871-0.941) showed significantly better (P<0.001, 95% CI: 0.038-0.128) discriminating results in prognosticating 1-year independent ambulation than with the use of currently used distinction between complete and incomplete SCI (AUC: 0.823, P<0.001, 95% CI: 0.781-0.864).'], 'labels': ['STUDY DESIGN', 'OBJECTIVE', 'SETTING', 'METHODS', 'RESULTS'], 'meshes': ['Adolescent', 'Adult', 'Aged', 'Aged, 80 and over', 'Anal Canal', 'Cohort Studies', 'Diagnosis, Differential', 'Disability Evaluation', 'Female', 'Humans', 'Male', 'Middle Aged', 'Neurologic Examination', 'Paralysis', 'Predictive Value of Tests', 'Prospective Studies', 'Reproducibility of Results', 'Sacrum', 'Severity of Illness Index', 'Somatosensory Disorders', 'Spinal Cord', 'Spinal Cord Injuries', 'Young Adult'], 'reasoning_required_pred': ['n', 'o'], 'reasoning_free_pred': ['m', 'a', 'y', 'b', 'e']}, 'long_answer': 'Out of the four sacral sparing criteria, the acute phase anal sensory score measurements do not contribute significantly to the prognosis of independent ambulation. The combination of the acute phase voluntary anal contraction and the S4-5 LT and PP scores, predicts significantly better chronic phase-independent ambulation outcomes than the currently used distinction between complete and incomplete SCI.', 'final_decision': 'maybe'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 12 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2025-12-21 12:02:01.680\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 12 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:02:01.693\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 14 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 73%|███████▎ | 367/500 [01:38<01:01, 2.15it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:02:01.769\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 12 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 74%|███████▍ | 369/500 [01:38<00:47, 2.78it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:02:01.881\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 24901580, 'question': 'Is scintigraphy a guideline method in determining amputation levels in diabetic foot?', 'context': {'contexts': ['In this study, we aimed to evaluate the potential use of a 3-phase bone scintigraphy method to determine the level of amputation on treatment cost, morbidity and mortality, reamputation rates, and the duration of hospitalization in diabetic foot.', 'Thirty patients who were admitted to our clinic between September 2008 and July 2009, with diabetic foot were included. All patients were evaluated according to age, gender, diabetes duration, 3-phase bone scintigraphy, Doppler ultrasound, amputation/reamputation levels, and hospitalization periods. Patients underwent 3-phase bone scintigraphy using technetium-99m methylene diphosphonate, and the most distal site of the region displaying perfusion during the perfusion and early blood flow phase was marked as the amputation level. Amputation level was determined by 3-phase bone scintigraphy, Doppler ultrasound, and inspection of the infection-free clear region during surgery.', 'The amputation levels of the patients were as follows: finger in six (20%), ray amputation in five (16.6%), transmetatarsal in one (3.3%), Lisfranc in two (6.6%), Chopart in seven (23.3%), Syme in one (3.3%), below-the-knee in six (20%), above the knee in one (3.3%), knee disarticulation in one (3.3%), and two patients underwent amputation at other centers. After primary amputation, reamputation was performed on seven patients, and one patient was treated with debridement for wound site problems. No mortality was encountered during study.'], 'labels': ['BACKGROUND', 'METHODS', 'RESULTS'], 'meshes': ['Aged', 'Aged, 80 and over', 'Amputation', 'Cohort Studies', 'Decision Making', 'Diabetic Foot', 'Female', 'Follow-Up Studies', 'Humans', 'Imaging, Three-Dimensional', 'Male', 'Middle Aged', 'Practice Guidelines as Topic', 'Preoperative Care', 'Prospective Studies', 'Radionuclide Imaging', 'Risk Assessment', 'Sensitivity and Specificity', 'Severity of Illness Index', 'Technetium Tc 99m Sestamibi', 'Treatment Outcome', 'Wound Healing'], 'reasoning_required_pred': ['y', 'e', 's'], 'reasoning_free_pred': ['y', 'e', 's']}, 'long_answer': 'We conclude that 3-phase bone scintigraphy prior to surgery could be a useful method to determine the amputation level in a diabetic foot. We conclude that further, comparative, more comprehensive, long-term, and controlled studies are required.', 'final_decision': 'yes'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 12 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2025-12-21 12:02:13.628\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 74%|███████▍ | 372/500 [01:50<03:23, 1.59s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2025-12-21 12:02:13.811\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:02:13.815\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 28027677, 'question': 'Do prerecorded lecture VODcasts affect lecture attendance of first-yearpre-clinical Graduate Entry to Medicine students?', 'context': {'contexts': ['There is increasing concern amongst educators that the provision of recorded lectures may reduce student attendance of live lectures. We therefore sought to determine if the provision of prerecorded lecture video podcasts (VODcasts) to first-year Graduate Entry to Medicine (GEM) students, affected attendance at 21 Physiology lectures within three separate pre-clinical modules.', 'Data on lecture attendance, utilization of VODcasts, and whether VODcasts should replace live lectures were drawn from three surveys conducted in academic years 2014-2015 and 2015-2016 on all first-year GEM students in two first-year pre-clinical modules where prerecorded Physiology VODcasts were available for viewing or downloading prior to scheduled live lectures.', 'A total of 191/214 (89%) students responded to the three surveys, with 84.3% of students attending all 21 lectures in the study. Only 4% of students missed more than one lecture in each of the three lecture series, with 79% indicating that VODcasts should not replace lectures.'], 'labels': ['BACKGROUND', 'METHODS', 'RESULTS'], 'meshes': ['Adult', 'Education, Medical, Undergraduate', 'Female', 'Humans', 'Internet', 'Male', 'Surveys and Questionnaires', 'Teaching', 'Videotape Recording', 'Young Adult'], 'reasoning_required_pred': ['n', 'o'], 'reasoning_free_pred': ['n', 'o']}, 'long_answer': 'Therefore, we conclude that the attendance of pre-clinical GEM students at live lectures is not significantly impacted upon by the provision of lecture VODcasts, with most students viewing them as useful revision tools rather than as a replacement for live lectures.', 'final_decision': 'no'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2025-12-21 12:02:13.850\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 27216167, 'question': 'Gynecological cancer alarm symptoms: is contact with specialist care associated with lifestyle and socioeconomic status?', 'context': {'contexts': ['The aim of this study was to determine the proportion of patients who were referred to specialist care after reporting gynecological cancer alarm symptoms to their general practitioner. We sought to investigate whether contact with specialist care was associated with lifestyle factors or socioeconomic status.', 'Nationwide population-based prospective cohort study in Denmark, based on a random sample of 51 090 women aged 20 years or older from the general population. A web-based questionnaire regarding gynecological alarm symptoms and lifestyle was distributed to the invited individuals. Data about contact with specialist care were obtained from the National Patient Register and the National Health Insurance Service Registry, whereas information about socioeconomic status was collected from Statistics Denmark. Main outcome measures were percentages of patients having contact with specialist care and odds ratios (ORs) for associations between specialist care contact, lifestyle factors and socioeconomic status.', 'The study included 25 866 nonpregnant women; 2957 reported the onset of at least one gynecological cancer alarm symptom, and 683 of these (23.1%) reported symptoms to their general practitioner. The proportion of individuals having contact with specialist care ranged from 39.3% (pain during intercourse) to 47.8% (bleeding during intercourse). Individuals with higher educational level had significantly higher odds of contact with a specialist (OR 1.86, 95% CI 1.17-2.95).'], 'labels': ['INTRODUCTION', 'MATERIAL AND METHODS', 'RESULTS'], 'meshes': ['Adult', 'Cohort Studies', 'Coitus', 'Denmark', 'Educational Status', 'Female', 'General Practitioners', 'Genital Neoplasms, Female', 'Humans', 'Life Style', 'Middle Aged', 'Pain', 'Patient Acceptance of Health Care', 'Pelvic Pain', 'Postmenopause', 'Referral and Consultation', 'Social Class', 'Surveys and Questionnaires', 'Uterine Hemorrhage'], 'reasoning_required_pred': ['y', 'e', 's'], 'reasoning_free_pred': ['y', 'e', 's']}, 'long_answer': 'Educational level influences contact with specialist care among patients with gynecological cancer alarm symptoms. Future studies should investigate inequalities in access to the secondary healthcare system.', 'final_decision': 'yes'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2025-12-21 12:02:13.871\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 15879722, 'question': \"Is cytokeratin immunoreactivity useful in the diagnosis of short-segment Barrett's oesophagus in Korea?\", 'context': {'contexts': [\"Cytokeratin 7/20 staining has been reported to be helpful in diagnosing Barrett's oesophagus and gastric intestinal metaplasia. However, this is still a matter of some controversy.\", \"To determine the diagnostic usefulness of cytokeratin 7/20 immunostaining for short-segment Barrett's oesophagus in Korea.\", \"In patients with Barrett's oesophagus, diagnosed endoscopically, at least two biopsy specimens were taken from just below the squamocolumnar junction. If goblet cells were found histologically with alcian blue staining, cytokeratin 7/20 immunohistochemical stains were performed. Intestinal metaplasia at the cardia was diagnosed whenever biopsy specimens taken from within 2 cm below the oesophagogastric junction revealed intestinal metaplasia. Barrett's cytokeratin 7/20 pattern was defined as cytokeratin 20 positivity in only the superficial gland, combined with cytokeratin 7 positivity in both the superficial and deep glands.\", \"Barrett's cytokeratin 7/20 pattern was observed in 28 out of 36 cases (77.8%) with short-segment Barrett's oesophagus, 11 out of 28 cases (39.3%) with intestinal metaplasia at the cardia, and nine out of 61 cases (14.8%) with gastric intestinal metaplasia. The sensitivity and specificity of Barrett's cytokeratin 7/20 pattern were 77.8 and 77.5%, respectively.\"], 'labels': ['BACKGROUND', 'OBJECTIVE', 'METHODS', 'RESULTS'], 'meshes': ['Adult', 'Aged', 'Barrett Esophagus', 'Biomarkers', 'Biopsy', 'Cardia', 'Esophagoscopy', 'Female', 'Humans', 'Intermediate Filament Proteins', 'Keratin-20', 'Keratin-7', 'Keratins', 'Male', 'Metaplasia', 'Middle Aged', 'Sensitivity and Specificity', 'Stomach'], 'reasoning_required_pred': ['y', 'e', 's'], 'reasoning_free_pred': ['y', 'e', 's']}, 'long_answer': \"Barrett's cytokeratin 7/20 pattern can be a useful marker for the diagnosis of short-segment Barrett's oesophagus, although the false positive or false negative rate is approximately 25%.\", 'final_decision': 'yes'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2025-12-21 12:02:13.982\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 20684175, 'question': 'Vitamin D supplementation and regulatory T cells in apparently healthy subjects: vitamin D treatment for autoimmune diseases?', 'context': {'contexts': ['Epidemiological data show significant associations of vitamin D deficiency and autoimmune diseases. Vitamin D may prevent autoimmunity by stimulating naturally occurring regulatory T cells.', 'To elucidate whether vitamin D supplementation increases Tregs frequency (%Tregs) within circulating CD4+ T cells.', 'We performed an uncontrolled vitamin D supplementation trial among 50 apparently healthy subjects including supplementation of 140,000 IU at baseline and after 4 weeks (visit 1). The final follow-up visit was performed 8 weeks after the baseline examination (visit 2). Blood was drawn at each study visit to determine 25-hydroxyvitamin D levels and %Tregs. Tregs were characterized as CD4+CD25++ T cells with expression of the transcription factor forkhead box P3 and low or absent expression of CD127.', 'Forty-six study participants (65% females, mean age +/- SD 31 +/- 8 years) completed the trial. 25(OH)D levels increased from 23.9 +/- 12.9 ng/ml at baseline to 45.9 +/- 14.0 ng/ml at visit 1 and 58.0 +/- 15.1 ng/ml at visit 2. %Tregs at baseline were 4.8 +/- 1.4. Compared to baseline levels we noticed a significant increase of %Tregs at study visit 1 (5.9 +/- 1.7, P<0.001) and 2 (5.6 +/- 1.6, P<0.001).'], 'labels': ['BACKGROUND', 'OBJECTIVES', 'METHODS', 'RESULTS'], 'meshes': ['Adult', 'C-Reactive Protein', 'CD4 Lymphocyte Count', 'CD4-Positive T-Lymphocytes', 'Calcium', 'Dietary Supplements', 'Female', 'Follow-Up Studies', 'Forkhead Transcription Factors', 'Humans', 'Immunologic Factors', 'Interleukin-2 Receptor alpha Subunit', 'Interleukin-7 Receptor alpha Subunit', 'Male', 'Pilot Projects', 'T-Lymphocytes, Regulatory', 'Vitamin D', 'Vitamins'], 'reasoning_required_pred': ['y', 'e', 's'], 'reasoning_free_pred': ['y', 'e', 's']}, 'long_answer': 'Vitamin D supplementation was associated with significantly increased %Tregs in apparently healthy individuals. This immunomodulatory effect of vitamin D might underlie the associations of vitamin D deficiency and autoimmune diseases. Hence, our finding provides a rationale for further studies to investigate vitamin D effects on autoimmunological processes.', 'final_decision': 'yes'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 75%|███████▌ | 377/500 [01:50<01:32, 1.32it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2025-12-21 12:02:14.954\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:02:15.023\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 76%|███████▌ | 379/500 [01:51<01:24, 1.43it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 76%|███████▌ | 381/500 [01:52<01:15, 1.58it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2025-12-21 12:02:16.053\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 77%|███████▋ | 383/500 [01:52<00:57, 2.05it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 77%|███████▋ | 387/500 [01:53<00:44, 2.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 78%|███████▊ | 391/500 [01:54<00:29, 3.66it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 79%|███████▉ | 395/500 [01:54<00:17, 6.17it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 79%|███████▉ | 397/500 [01:54<00:15, 6.65it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 80%|████████ | 401/500 [01:55<00:11, 8.65it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 81%|████████ | 403/500 [01:56<00:20, 4.84it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 81%|████████▏ | 407/500 [01:56<00:15, 6.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 82%|████████▏ | 409/500 [01:56<00:13, 6.96it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 82%|████████▏ | 411/500 [01:57<00:17, 5.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 82%|████████▏ | 412/500 [01:57<00:19, 4.54it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 83%|████████▎ | 417/500 [01:58<00:10, 7.66it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2025-12-21 12:02:22.105\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 84%|████████▍ | 419/500 [01:58<00:12, 6.31it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 84%|████████▍ | 421/500 [01:58<00:13, 6.05it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 84%|████████▍ | 422/500 [01:59<00:12, 6.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2025-12-21 12:02:23.165\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 10749257, 'question': 'Has the mammography quality standards act affected the mammography quality in North Carolina?', 'context': {'contexts': ['The United States Food and Drug Administration implemented federal regulations governing mammography under the Mammography Quality Standards Act (MQSA) of 1992. During 1995, its first year in implementation, we examined the impact of the MQSA on the quality of mammography in North Carolina.', 'All mammography facilities were inspected during 1993-1994, and again in 1995. Both inspections evaluated mean glandular radiation dose, phantom image evaluation, darkroom fog, and developer temperature. Two mammography health specialists employed by the North Carolina Division of Radiation Protection performed all inspections and collected and codified data.', 'The percentage of facilities that met quality standards increased from the first inspection to the second inspection. Phantom scores passing rate was 31.6% versus 78.2%; darkroom fog passing rate was 74.3% versus 88.5%; and temperature difference passing rate was 62.4% versus 86.9%.'], 'labels': ['OBJECTIVE', 'MATERIALS AND METHODS', 'RESULTS'], 'meshes': ['Mammography', 'North Carolina', 'United States', 'United States Food and Drug Administration'], 'reasoning_required_pred': ['y', 'e', 's'], 'reasoning_free_pred': ['y', 'e', 's']}, 'long_answer': \"In 1995, the first year that the MQSA was in effect, there was a significant improvement in the quality of mammography in North Carolina. This improvement probably resulted from facilities' compliance with federal regulations.\", 'final_decision': 'yes'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 85%|████████▍ | 423/500 [01:59<00:19, 3.98it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2025-12-21 12:02:23.486\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 85%|████████▍ | 424/500 [01:59<00:20, 3.72it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:02:23.519\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 25501465, 'question': \"Evaluation of pediatric VCUG at an academic children's hospital: is the radiographic scout image necessary?\", 'context': {'contexts': ['There is heterogeneity in how pediatric voiding cystourethrography (VCUG) is performed. Some institutions, including our own, obtain a radiographic scout image prior to contrast agent instillation.', 'To demonstrate that the radiographic scout image does not augment VCUG interpretation or contribute management-changing information but nonetheless carries a non-negligible effective dose.', 'We evaluated 181 children who underwent VCUG in 2012, with an age breakdown of less than 1\\xa0year (56 children), 1-5 years (66 children), 6-10 years (43 children) and 11-18 years (16 children), with a mean age of 4.0\\xa0years. We investigated patient demographics, clinical indication for the examination, scout image findings and estimated effective radiation dose, as well as overall exam findings and impression.', 'No clinically significant or management-changing findings were present on scout images, and no radiopaque urinary tract calculi or concerning incidental finding was identified. Scout image estimated effective radiation dose averaged 0.09\\xa0mSv in children younger than 1\\xa0y, 0.09\\xa0mSv in children age 1-5, 0.13\\xa0mSv in children age 6-10 and 0.18\\xa0mSv in children age 11-18. Total fluoroscopy time per examination averaged 36.7\\xa0s (range 34.8-39.6\\xa0s for all age group averages). Evaluation of known or suspected vesicoureteral reflux (VUR) and urinary tract infection (UTI) were the most common clinical indications, stated in 40.9% and 37.0% of exams, respectively.'], 'labels': ['BACKGROUND', 'OBJECTIVE', 'MATERIALS AND METHODS', 'RESULTS'], 'meshes': ['Academic Medical Centers', 'Adolescent', 'Child', 'Child, Preschool', 'Female', 'Fluoroscopy', 'Hospitals, Pediatric', 'Humans', 'Infant', 'Male', 'Retrospective Studies', 'Vesico-Ureteral Reflux'], 'reasoning_required_pred': ['n', 'o'], 'reasoning_free_pred': ['n', 'o']}, 'long_answer': 'Although the estimated effective dose is low for VCUG radiographic scout images, this step did not augment VCUG interpretation or contribute management-changing information. This step should be omitted or substituted to further reduce dose in pediatric VCUG.', 'final_decision': 'no'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2025-12-21 12:02:23.703\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 19155657, 'question': 'Does accompanying metabolic syndrome contribute to heart dimensions in hypertensive patients?', 'context': {'contexts': ['Metabolic syndrome (MetS) is associated with increased risk for cardiovascular events. We evaluated heart dimensions in hypertensive patients with MetS.', 'The study included 75 hypertensive patients (34 males, 41 females; mean age 51+/-9 years) without coronary artery disease. Patients were evaluated in two groups depending on the presence or absence of MetS. Age- and gender-matched 20 healthy subjects (9 males, 11 females; mean age 50+/-5 years) comprised the control group. The diagnosis of MetS was based on the presence of at least three of five MetS criteria. Hypertension was defined as arterial blood pressure exceeding 140/85 mmHg on three consecutive measurements or the use of antihypertensive drugs. Echocardiographic measurements included interventricular septal thickness, left ventricular internal diameter, posterior wall thickness, aortic diameter, left atrial diameter, relative wall thickness, and left ventricular mass.', 'Metabolic syndrome was present in 32 hypertensive patients (42.7%; 18 males, 14 females). The mean number of MetS criteria was 2.6+/-1.0 in the hypertensive group. Compared to the control group, patients with or without MetS exhibited significantly increased interventricular septum and posterior wall thickness, left atrial diameter, relative wall thickness, and left ventricular mass (p<0.05). The only significant difference between the two patient groups was that MetS was associated with a greater left atrial diameter (p=0.019). Left atrial diameter was correlated with the number of MetS criteria (r=0.51; p<0.001).'], 'labels': ['OBJECTIVES', 'STUDY DESIGN', 'RESULTS'], 'meshes': ['Cardiovascular Diseases', 'Case-Control Studies', 'Echocardiography', 'Female', 'Humans', 'Hypertension', 'Hypertrophy, Left Ventricular', 'Male', 'Metabolic Syndrome', 'Middle Aged', 'Risk Factors', 'Ventricular Remodeling'], 'reasoning_required_pred': ['y', 'e', 's'], 'reasoning_free_pred': ['m', 'a', 'y', 'b', 'e']}, 'long_answer': 'Left ventricular dimensions are not influenced by MetS. Rather than MetS, hypertension is primarily responsible for changes in left ventricular dimensions. However, left atrial enlargement is more prominent in patients with MetS, suggesting that each MetS criterion contributes to left ventricular diastolic dysfunction.', 'final_decision': 'maybe'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 85%|████████▌ | 427/500 [02:00<00:12, 6.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2025-12-21 12:02:23.794\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 86%|████████▌ | 430/500 [02:00<00:13, 5.15it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 87%|████████▋ | 433/500 [02:01<00:09, 7.00it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2025-12-21 12:02:24.916\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 15222284, 'question': 'The effective orifice area/patient aortic annulus area ratio: a better way to compare different bioprostheses?', 'context': {'contexts': ['The aim of this prospective, randomized study was to compare the hemodynamic performance of the Medtronic Mosaic and Edwards Perimount bioprostheses in the aortic position, and to evaluate prosthesis-specific differences in valve sizing and valve-size labeling.', 'Between August 2000 and September 2002, 139 patients underwent isolated aortic valve replacement (AVR) with the Mosaic (n = 67) or Perimount (n = 72) bioprosthesis. Intraoperatively, the internal aortic annulus diameter was measured by insertion of a gauge (Hegar dilator), while prosthesis size was determined by using the original sizers. Transthoracic echocardiography was performed to determine hemodynamic and dimensional data. As the aim of AVR is to achieve a maximal effective orifice area (EOA) within a given aortic annulus, the ratio of EOA to patient aortic annulus area was calculated, the latter being based on annulus diameter measured intraoperatively.', \"Operative mortality was 2.2% (Mosaic 3.0%; Perimount 1.4%; p = NS). Upsizing (using a prosthesis larger in labeled valve size than the patient's measured internal aortic annulus diameter) was possible in 28.4% of Mosaic patients and 8.3% of Perimount patients. The postoperative mean systolic pressure gradient ranged from 10.5 to 22.2 mmHg in the Mosaic group, and from 9.4 to 12.6 mmHg in the Perimount group; it was significantly lower for 21 and 23 Perimount valves than for 21 and 23 Mosaic valves. The EOA ranged from 0.78 to 2.37 cm2 in Mosaic patients, and from 0.95 to 2.12 cm2 in Perimount patients. When indexing EOA by calculating the ratio of EOA to patient aortic annulus area to adjust for variables such as patient anatomy and valve dimensions, there was no significant difference between the two bioprostheses.\"], 'labels': ['BACKGROUND AND AIM OF THE STUDY', 'METHODS', 'RESULTS'], 'meshes': ['Aged', 'Aged, 80 and over', 'Aortic Valve', 'Aortic Valve Insufficiency', 'Aortic Valve Stenosis', 'Bioprosthesis', 'Blood Pressure', 'Female', 'Heart Valve Prosthesis', 'Heart Valve Prosthesis Implantation', 'Humans', 'Male', 'Middle Aged', 'Product Labeling', 'Prospective Studies', 'Prosthesis Design', 'Prosthesis Fitting'], 'reasoning_required_pred': ['n', 'o'], 'reasoning_free_pred': ['y', 'e', 's']}, 'long_answer': \"Comparisons of absolute EOA values grouped by the manufacturers' valve sizes are misleading because of specific differences in geometric dimensions. The EOA:patient aortic annulus area ratio provides a new hemodynamic index which may facilitate objective comparisons between different valve types.\", 'final_decision': 'yes'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 88%|████████▊ | 438/500 [02:01<00:06, 9.29it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 88%|████████▊ | 440/500 [02:02<00:12, 4.94it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 88%|████████▊ | 441/500 [02:02<00:11, 5.17it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 89%|████████▊ | 443/500 [02:02<00:11, 5.07it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 89%|████████▉ | 447/500 [02:03<00:07, 7.15it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 90%|████████▉ | 449/500 [02:03<00:08, 6.07it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 90%|█████████ | 451/500 [02:04<00:07, 6.91it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 91%|█████████ | 453/500 [02:04<00:06, 6.85it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 91%|█████████ | 455/500 [02:04<00:05, 7.63it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 92%|█████████▏| 460/500 [02:05<00:04, 8.15it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 92%|█████████▏| 462/500 [02:05<00:05, 6.91it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 93%|█████████▎| 465/500 [02:05<00:04, 8.28it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 93%|█████████▎| 467/500 [02:06<00:04, 7.50it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 94%|█████████▍| 470/500 [02:06<00:03, 8.88it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 94%|█████████▍| 472/500 [02:06<00:03, 8.95it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 95%|█████████▌| 476/500 [02:06<00:02, 10.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 96%|█████████▌| 478/500 [02:07<00:03, 6.72it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 96%|█████████▋| 482/500 [02:07<00:02, 8.50it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 97%|█████████▋| 484/500 [02:07<00:01, 10.23it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 97%|█████████▋| 486/500 [02:08<00:01, 8.08it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 98%|█████████▊| 488/500 [02:08<00:01, 7.58it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 98%|█████████▊| 491/500 [02:09<00:01, 7.69it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 99%|█████████▊| 493/500 [02:09<00:01, 6.17it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 99%|█████████▉| 497/500 [02:10<00:00, 7.39it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 100%|█████████▉| 499/500 [02:10<00:00, 8.52it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 100%|██████████| 500/500 [02:12<00:00, 3.76it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2025-12-21 12:02:36.508\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36m__main__\u001b[0m:\u001b[36m\u001b[0m:\u001b[36m13\u001b[0m - \u001b[1mEvaluation metrics (after optimization): {'f1': 0.0, 'em': 0.0, 'acc': 0.6052060737527115}\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\n" ] } ], "source": [ "logger.info(\"Evaluating workflow on test set...\")\n", "with suppress_logger_info():\n", " results = textgrad_optimizer.evaluate(dataset=benchmark, eval_mode=\"test\")\n", "logger.info(f\"Evaluation metrics (before optimization): {results}\")\n", "\n", "logger.info(\"Optimizing workflow...\")\n", "textgrad_optimizer.optimize(benchmark, seed=8)\n", "textgrad_optimizer.restore_best_graph()\n", "\n", "logger.info(\"Evaluating workflow on test set...\")\n", "with suppress_logger_info():\n", " results = textgrad_optimizer.evaluate(dataset=benchmark, eval_mode=\"test\")\n", "logger.info(f\"Evaluation metrics (after optimization): {results}\")" ] }, { "cell_type": "code", "execution_count": 9, "id": "4b6f274d", "metadata": { "scrolled": true }, "outputs": [ { "data": { "text/plain": [ "{'f1': 0.0, 'em': 0.0, 'acc': 0.6052060737527115}" ] }, "execution_count": 9, "metadata": {}, "output_type": "execute_result" } ], "source": [ "results" ] }, { "cell_type": "code", "execution_count": 11, "id": "56531fd3", "metadata": {}, "outputs": [], "source": [ "#textgrad-generated-workflow\n", "\n", "hotpotqa_graph_data = {\n", " \"goal\": \"Provide a direct answer to the question based on the context, without including explanations or reasoning.\",\n", " \"tasks\": [\n", " {\n", " \"name\": \"answer_generate\",\n", " \"description\": \"Generate a direct answer to the question based on the context.\",\n", " \"inputs\": [\n", " {\"name\": \"problem\", \"type\": \"str\", \"required\": True, \"description\": \"The question to answer directly.\"}\n", " ],\n", " \"outputs\": [\n", " {\"name\": \"answer\", \"type\": \"str\", \"required\": True, \"description\": \"The direct answer to the question.\"}\n", " ],\n", " \"prompt_template\": StringTemplate(instruction=\"Provide a concise and direct answer to the question.\"),\n", " \"parse_mode\": \"text\"\n", " }\n", " ] \n", "}" ] }, { "cell_type": "code", "execution_count": 12, "id": "fa82f0cf", "metadata": {}, "outputs": [ { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2025-12-21 12:56:09.728\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.benchmark.pubmedqa\u001b[0m:\u001b[36m_load_data_from_file\u001b[0m:\u001b[36m51\u001b[0m - \u001b[1mloading HotPotQA data from /gpfs/radev/pi/ying_rex/tl688/selfevolve/EvoAgentX/examples/pubmedqa/pubmedqa_train.json ...\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349655.794492449)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349655.91602788)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349657.528116504)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349656.595366654)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349656.720509929)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349657.279656617)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349657.069110437)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349657.850128811)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349717.170030702)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349716.911332504)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349717.367858409)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349717.394290517)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349717.115304642)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349717.796702131)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349716.743017815)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349716.795223625)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349718.27123203)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349717.307862742)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349717.215369296)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349716.866024508)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349717.670063238)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349716.662797149)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349717.781268261)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349717.38162481)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349717.249792906)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349717.070052424)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349718.339924365)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349721.114220095)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349720.511041476)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349720.426325599)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349721.787368023)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349722.20989129)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349722.260799649)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349780.781880301)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349780.502184599)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349780.644530383)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349781.584055148)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349780.760426617)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349780.732744673)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349781.388437646)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349781.59425938)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349780.747533459)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349782.784134746)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349780.847666129)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349781.22270922)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349780.693663633)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349780.200914306)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349783.829922054)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349785.099341043)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349790.173298285)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349789.513707126)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349789.759562665)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349785.640829715)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349786.281155627)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349785.819516318)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349786.18698921)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349790.033953066)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349788.520109584)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349789.919106368)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349790.228481696)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349790.869094954)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349788.601701352)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349788.629332454)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349788.525007637)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349791.154645834)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349788.406160213)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349790.351730719)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349790.417083266)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349790.459998194)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349789.9422543)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349792.793583557)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349793.802124876)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349793.757275403)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349793.447260883)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349794.846370846)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349792.540566988)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349792.74763959)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349792.055773662)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349792.141115052)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349794.47791502)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349797.566218658)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 349794.166243348)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349794.257405628)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349795.650618789)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349797.77956897)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349798.40912093)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349797.595529141)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349798.749999973)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349797.756265804)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349798.002070285)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349797.974106147)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349798.201617309)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349797.930895627)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349798.283719724)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349798.386270885)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349798.508120892)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349796.991887498)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349798.141031276)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349799.287130882)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349797.003828436)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349716.65261662)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349781.01680045)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 349780.450830657)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349779.563819391)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349781.086943779)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349789.159776595)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349789.895697618)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349789.067368289)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349796.99572547)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349799.740699646)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349656.221300623)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349655.141149416)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349656.48022287)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349656.119936049)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349718.252097926)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 349718.657782034)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349720.198651533)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349718.466671309)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349720.179574716)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349719.732705151)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349718.764685663)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349719.626206414)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349719.929260391)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349719.928047823)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349719.576268904)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349719.606142596)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349719.029803285)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349719.995665856)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349720.713630825)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349720.622092037)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349720.983917691)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349720.438317994)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349720.061497916)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349720.169668491)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349722.087090186)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349722.744021558)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349781.288305993)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349781.612078407)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349781.163223837)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349782.788928464)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349783.825804518)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349783.154693004)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349784.71128288)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349783.765779998)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349783.262180806)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349783.529747434)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349784.016758996)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349784.53978346)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349783.521381949)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349783.557435074)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349783.025174762)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349783.480433231)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349784.049259735)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349784.504515659)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349785.542974142)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349784.15572772)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349785.293362891)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 349785.180278629)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349784.431124826)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349786.977123489)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349785.963795787)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349789.487364213)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349789.948495853)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349788.605634385)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349789.864056088)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349790.297409878)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349791.638378764)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349791.819722156)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349791.727641521)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349792.248263299)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349792.565665254)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349794.454252978)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349792.374455672)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349792.573538671)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349791.344109941)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349791.649544289)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349792.636417705)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349792.957478272)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349793.280668719)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349795.942801162)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349794.053202412)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349793.681597953)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349795.247230281)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349794.284773335)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349795.019242992)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349795.798946558)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349795.408805341)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349795.244134178)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349794.947887281)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349795.316507181)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349796.021702696)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349796.169177557)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349796.179373826)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349794.81778685)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349796.014910627)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349720.490596251)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349722.193150197)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 349786.098221215)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349794.660553047)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349799.099928953)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349808.374944265)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349800.419834734)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349800.390360712)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349800.365218164)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349800.425399782)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349808.209346115)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349801.102050905)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349800.839909934)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349800.360046065)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349804.065515269)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349801.422253094)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349800.481494137)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349801.120178307)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349801.085126305)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349799.601630305)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349811.769199324)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349799.677836741)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349799.760469948)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349798.443578272)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349800.150944896)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349799.444703819)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349801.28775042)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349808.115292711)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349808.100346213)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349808.651425459)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349808.490120394)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 349808.246517966)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349808.735590054)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349804.072585931)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349808.244443518)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349811.500710753)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349808.428358496)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349804.152008807)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349808.207277883)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349811.988157832)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349808.221416393)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349808.255982624)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349811.369692884)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349817.368223321)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349817.366162871)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349821.899138154)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349811.203804005)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349811.33073012)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349821.912492686)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349817.253299869)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349817.377969622)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349817.413426855)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349817.512490592)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349825.537167812)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349816.805134196)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349816.868618451)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349837.123237536)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349821.506897333)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 349817.399395557)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349817.406653189)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349822.021864929)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349821.477953333)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349821.634652397)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349821.901138338)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349822.159596897)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349825.000063086)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349824.53613085)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349825.535180175)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349825.558366349)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349839.649499201)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349839.19959086)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349824.949089718)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349825.530195716)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349836.990860875)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349839.394288703)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349838.905959378)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349837.159995457)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349837.317321049)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349837.048508097)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349837.192222005)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349837.295657267)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349836.76649383)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349837.073132885)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349837.155940014)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349841.3931971)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349841.214758639)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349841.081091432)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 349841.415092689)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349839.000424306)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349842.220122174)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349840.853845819)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349841.69724469)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349841.300295438)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349841.445258339)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349843.630035939)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349844.574052247)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349841.16172834)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349842.282200838)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349841.592719544)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349841.304921179)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349844.533659141)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349841.673721281)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349840.401536777)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349841.094871058)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349841.606276395)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349845.342749326)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349842.905169927)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349845.413948243)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349844.213466942)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349843.618164309)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349843.433737682)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349844.272135453)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349844.152267069)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349842.970192508)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349844.041475991)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 349843.084177068)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349844.274790398)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349845.36578428)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349848.094157396)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349843.975669086)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349842.696540392)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349845.880070546)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349846.724825322)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349845.706868344)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349848.534221932)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349848.353056514)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349845.755988834)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349846.209046563)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349846.17598273)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349844.757895841)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349847.860606027)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349846.408680767)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349846.215560843)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349846.288594879)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349846.414415201)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349846.900686783)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349850.643631693)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349850.334732287)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349849.490626698)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349849.792528238)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349848.676818847)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349847.658899726)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 349848.973460052)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349850.158868536)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349850.84036771)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349849.202457123)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349850.105064002)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349850.963367097)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349849.340840281)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349849.313685952)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349849.813121043)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349850.262152726)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349854.052511917)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349850.766230361)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349850.874506729)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349851.252325706)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349851.241834875)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349849.988874871)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349850.514668637)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349852.049045525)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349852.285380426)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349854.576269178)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349851.571465721)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349852.835830245)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349853.126472148)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349852.90325363)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349853.672327365)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349853.589545048)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349853.756938043)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349850.972058727)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 349851.036176699)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349854.304427102)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349853.762385184)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349853.048606117)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349851.909475577)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349855.436691193)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349852.707675631)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349853.450350458)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349854.661849058)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349854.398443523)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349854.928968877)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349854.353469496)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349856.745405489)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349855.571187476)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349855.336196883)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349853.317535422)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349853.569716243)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349853.825359089)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349856.880419657)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349852.898085291)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349855.739585503)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349856.813879265)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349856.699431811)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349856.201461661)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349856.362366675)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349856.741278206)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349855.965179287)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349857.5736442)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349856.725412549)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 349857.28626653)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349856.272847043)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349856.809298525)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349855.205192671)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349858.251079384)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349854.640700897)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349854.922517904)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349855.746509734)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349858.534684358)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349857.969482378)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349858.00916313)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349859.409091965)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349857.349454826)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349858.606256885)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349859.270994578)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 349858.098481879)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349859.110955714)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 349857.037302578)])']\n", "connector: \n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2025-12-21 12:56:13.100\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.benchmark.pubmedqa\u001b[0m:\u001b[36m_load_data_from_file\u001b[0m:\u001b[36m51\u001b[0m - \u001b[1mloading HotPotQA data from /gpfs/radev/pi/ying_rex/tl688/selfevolve/EvoAgentX/examples/pubmedqa/pubmedqa_train.json ...\u001b[0m\n", "\u001b[32m2025-12-21 12:56:15.502\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.benchmark.pubmedqa\u001b[0m:\u001b[36m_load_data_from_file\u001b[0m:\u001b[36m51\u001b[0m - \u001b[1mloading HotPotQA data from /gpfs/radev/pi/ying_rex/tl688/selfevolve/EvoAgentX/examples/pubmedqa/pubmedqa_label.json ...\u001b[0m\n" ] }, { "ename": "ValidationError", "evalue": "1 validation error for CustomizeAgent\n Value error, 'text' is an invalid value for `parse_mode`. Available choices: ['str', 'json', 'xml', 'title', 'custom']. [type=value_error, input_value={'name': 'AnswerGenerateA..., response_format=None)}, input_type=dict]\n For further information visit https://errors.pydantic.dev/2.11/v/value_error", "output_type": "error", "traceback": [ "\u001b[31m---------------------------------------------------------------------------\u001b[39m", "\u001b[31mValidationError\u001b[39m Traceback (most recent call last)", "\u001b[36mCell\u001b[39m\u001b[36m \u001b[39m\u001b[32mIn[12]\u001b[39m\u001b[32m, line 4\u001b[39m\n\u001b[32m 2\u001b[39m workflow_graph = SequentialWorkFlowGraph.from_dict(hotpotqa_graph_data)\n\u001b[32m 3\u001b[39m agent_manager = AgentManager()\n\u001b[32m----> \u001b[39m\u001b[32m4\u001b[39m \u001b[43magent_manager\u001b[49m\u001b[43m.\u001b[49m\u001b[43madd_agents_from_workflow\u001b[49m\u001b[43m(\u001b[49m\u001b[43mworkflow_graph\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mexecutor_llm\u001b[49m\u001b[43m.\u001b[49m\u001b[43mconfig\u001b[49m\u001b[43m)\u001b[49m\n\u001b[32m 6\u001b[39m evaluator = Evaluator(\n\u001b[32m 7\u001b[39m llm=executor_llm, \n\u001b[32m 8\u001b[39m agent_manager=agent_manager, \n\u001b[32m (...)\u001b[39m\u001b[32m 11\u001b[39m verbose=\u001b[38;5;28;01mTrue\u001b[39;00m\n\u001b[32m 12\u001b[39m )\n\u001b[32m 14\u001b[39m textgrad_optimizer = TextGradOptimizer(\n\u001b[32m 15\u001b[39m graph=workflow_graph, \n\u001b[32m 16\u001b[39m optimize_mode=\u001b[33m\"\u001b[39m\u001b[33mall\u001b[39m\u001b[33m\"\u001b[39m,\n\u001b[32m (...)\u001b[39m\u001b[32m 27\u001b[39m constraints=[]\n\u001b[32m 28\u001b[39m )\n", "\u001b[36mFile \u001b[39m\u001b[32m/gpfs/radev/pi/ying_rex/tl688/selfevolve/EvoAgentX/evoagentx/agents/agent_manager.py:337\u001b[39m, in \u001b[36mAgentManager.add_agents_from_workflow\u001b[39m\u001b[34m(self, workflow_graph, llm_config, **kwargs)\u001b[39m\n\u001b[32m 335\u001b[39m \u001b[38;5;28;01mif\u001b[39;00m node.agents:\n\u001b[32m 336\u001b[39m \u001b[38;5;28;01mfor\u001b[39;00m agent \u001b[38;5;129;01min\u001b[39;00m node.agents:\n\u001b[32m--> \u001b[39m\u001b[32m337\u001b[39m \u001b[38;5;28;43mself\u001b[39;49m\u001b[43m.\u001b[49m\u001b[43madd_agent\u001b[49m\u001b[43m(\u001b[49m\u001b[43magent\u001b[49m\u001b[43m=\u001b[49m\u001b[43magent\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mllm_config\u001b[49m\u001b[43m=\u001b[49m\u001b[43mllm_config\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43m*\u001b[49m\u001b[43m*\u001b[49m\u001b[43mkwargs\u001b[49m\u001b[43m)\u001b[49m\n", "\u001b[36mFile \u001b[39m\u001b[32m/gpfs/radev/pi/ying_rex/tl688/selfevolve/EvoAgentX/evoagentx/core/decorators.py:32\u001b[39m, in \u001b[36matomic_method..wrapper\u001b[39m\u001b[34m(self, *args, **kwargs)\u001b[39m\n\u001b[32m 30\u001b[39m context = \u001b[38;5;28mgetattr\u001b[39m(\u001b[38;5;28mself\u001b[39m, \u001b[33m\"\u001b[39m\u001b[33m_lock\u001b[39m\u001b[33m\"\u001b[39m, nullcontext())\n\u001b[32m 31\u001b[39m \u001b[38;5;28;01mwith\u001b[39;00m context:\n\u001b[32m---> \u001b[39m\u001b[32m32\u001b[39m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[43mfunc\u001b[49m\u001b[43m(\u001b[49m\u001b[38;5;28;43mself\u001b[39;49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43m*\u001b[49m\u001b[43margs\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43m*\u001b[49m\u001b[43m*\u001b[49m\u001b[43mkwargs\u001b[49m\u001b[43m)\u001b[49m\n", "\u001b[36mFile \u001b[39m\u001b[32m/gpfs/radev/pi/ying_rex/tl688/selfevolve/EvoAgentX/evoagentx/agents/agent_manager.py:308\u001b[39m, in \u001b[36mAgentManager.add_agent\u001b[39m\u001b[34m(self, agent, llm_config, **kwargs)\u001b[39m\n\u001b[32m 306\u001b[39m \u001b[38;5;28;01mif\u001b[39;00m \u001b[38;5;28mself\u001b[39m.has_agent(agent_name=agent_name):\n\u001b[32m 307\u001b[39m \u001b[38;5;28;01mreturn\u001b[39;00m\n\u001b[32m--> \u001b[39m\u001b[32m308\u001b[39m agent_instance = \u001b[38;5;28;43mself\u001b[39;49m\u001b[43m.\u001b[49m\u001b[43mcreate_agent\u001b[49m\u001b[43m(\u001b[49m\u001b[43magent\u001b[49m\u001b[43m=\u001b[49m\u001b[43magent\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mllm_config\u001b[49m\u001b[43m=\u001b[49m\u001b[43mllm_config\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43m*\u001b[49m\u001b[43m*\u001b[49m\u001b[43mkwargs\u001b[49m\u001b[43m)\u001b[49m\n\u001b[32m 309\u001b[39m \u001b[38;5;28mself\u001b[39m.agents.append(agent_instance)\n\u001b[32m 310\u001b[39m \u001b[38;5;28mself\u001b[39m.agent_states[agent_instance.name] = AgentState.AVAILABLE\n", "\u001b[36mFile \u001b[39m\u001b[32m/gpfs/radev/pi/ying_rex/tl688/selfevolve/EvoAgentX/evoagentx/agents/agent_manager.py:276\u001b[39m, in \u001b[36mAgentManager.create_agent\u001b[39m\u001b[34m(self, agent, llm_config, **kwargs)\u001b[39m\n\u001b[32m 274\u001b[39m \u001b[38;5;28;01mif\u001b[39;00m \u001b[38;5;129;01mnot\u001b[39;00m agent.get(\u001b[33m\"\u001b[39m\u001b[33mis_human\u001b[39m\u001b[33m\"\u001b[39m, \u001b[38;5;28;01mFalse\u001b[39;00m) \u001b[38;5;129;01mand\u001b[39;00m (llm_config \u001b[38;5;129;01mis\u001b[39;00m \u001b[38;5;28;01mNone\u001b[39;00m \u001b[38;5;129;01mand\u001b[39;00m \u001b[33m\"\u001b[39m\u001b[33mllm_config\u001b[39m\u001b[33m\"\u001b[39m \u001b[38;5;129;01mnot\u001b[39;00m \u001b[38;5;129;01min\u001b[39;00m agent):\n\u001b[32m 275\u001b[39m \u001b[38;5;28;01mraise\u001b[39;00m \u001b[38;5;167;01mValueError\u001b[39;00m(\u001b[33m\"\u001b[39m\u001b[33mWhen providing an agent as a dictionary, you must either include \u001b[39m\u001b[33m'\u001b[39m\u001b[33mllm_config\u001b[39m\u001b[33m'\u001b[39m\u001b[33m in the dictionary or provide it as a parameter.\u001b[39m\u001b[33m\"\u001b[39m)\n\u001b[32m--> \u001b[39m\u001b[32m276\u001b[39m agent_instance = \u001b[38;5;28;43mself\u001b[39;49m\u001b[43m.\u001b[49m\u001b[43mcreate_customize_agent\u001b[49m\u001b[43m(\u001b[49m\u001b[43magent_data\u001b[49m\u001b[43m=\u001b[49m\u001b[43magent\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mllm_config\u001b[49m\u001b[43m=\u001b[49m\u001b[43mllm_config\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43m*\u001b[49m\u001b[43m*\u001b[49m\u001b[43mkwargs\u001b[49m\u001b[43m)\u001b[49m\n\u001b[32m 277\u001b[39m \u001b[38;5;28;01melif\u001b[39;00m \u001b[38;5;28misinstance\u001b[39m(agent, Agent):\n\u001b[32m 278\u001b[39m agent_instance = agent\n", "\u001b[36mFile \u001b[39m\u001b[32m/gpfs/radev/pi/ying_rex/tl688/selfevolve/EvoAgentX/evoagentx/agents/agent_manager.py:237\u001b[39m, in \u001b[36mAgentManager.create_customize_agent\u001b[39m\u001b[34m(self, agent_data, llm_config, **kwargs)\u001b[39m\n\u001b[32m 230\u001b[39m \u001b[38;5;66;03m# tool_mapping = {}\u001b[39;00m\n\u001b[32m 231\u001b[39m \u001b[38;5;66;03m# if self.tools is not None:\u001b[39;00m\n\u001b[32m 232\u001b[39m \u001b[38;5;66;03m# for tool in self.tools:\u001b[39;00m\n\u001b[32m 233\u001b[39m \u001b[38;5;66;03m# tool_mapping[tool.name] = tool\u001b[39;00m\n\u001b[32m 234\u001b[39m \u001b[38;5;66;03m# if agent_data.get(\"tool_names\", None):\u001b[39;00m\n\u001b[32m 235\u001b[39m \u001b[38;5;66;03m# agent_data[\"tools\"] = [tool_mapping[tool_name] for tool_name in agent_data[\"tool_names\"]]\u001b[39;00m\n\u001b[32m 236\u001b[39m \u001b[38;5;28mself\u001b[39m.update_tools(agent_data=agent_data) \u001b[38;5;66;03m# add `tools` field if needed \u001b[39;00m\n\u001b[32m--> \u001b[39m\u001b[32m237\u001b[39m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[43mCustomizeAgent\u001b[49m\u001b[43m.\u001b[49m\u001b[43mfrom_dict\u001b[49m\u001b[43m(\u001b[49m\u001b[43mdata\u001b[49m\u001b[43m=\u001b[49m\u001b[43magent_data\u001b[49m\u001b[43m)\u001b[49m\n", "\u001b[36mFile \u001b[39m\u001b[32m/gpfs/radev/pi/ying_rex/tl688/selfevolve/EvoAgentX/evoagentx/core/module.py:195\u001b[39m, in \u001b[36mBaseModule.from_dict\u001b[39m\u001b[34m(cls, data, **kwargs)\u001b[39m\n\u001b[32m 193\u001b[39m \u001b[38;5;28;01mif\u001b[39;00m class_name:\n\u001b[32m 194\u001b[39m \u001b[38;5;28mcls\u001b[39m = MODULE_REGISTRY.get_module(class_name)\n\u001b[32m--> \u001b[39m\u001b[32m195\u001b[39m module = \u001b[38;5;28;43mcls\u001b[39;49m\u001b[43m.\u001b[49m\u001b[43m_create_instance\u001b[49m\u001b[43m(\u001b[49m\u001b[43mdata\u001b[49m\u001b[43m)\u001b[49m\n\u001b[32m 196\u001b[39m \u001b[38;5;66;03m# module = cls.model_validate(data)\u001b[39;00m\n\u001b[32m 197\u001b[39m \u001b[38;5;28;01mif\u001b[39;00m \u001b[38;5;28mlen\u001b[39m(buffer.exceptions) > \u001b[32m0\u001b[39m:\n", "\u001b[36mFile \u001b[39m\u001b[32m/gpfs/radev/pi/ying_rex/tl688/selfevolve/EvoAgentX/evoagentx/core/module.py:150\u001b[39m, in \u001b[36mBaseModule._create_instance\u001b[39m\u001b[34m(cls, data)\u001b[39m\n\u001b[32m 148\u001b[39m processed_data = {k: \u001b[38;5;28mcls\u001b[39m._process_data(v) \u001b[38;5;28;01mfor\u001b[39;00m k, v \u001b[38;5;129;01min\u001b[39;00m data.items()}\n\u001b[32m 149\u001b[39m \u001b[38;5;66;03m# print(processed_data)\u001b[39;00m\n\u001b[32m--> \u001b[39m\u001b[32m150\u001b[39m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[38;5;28;43mcls\u001b[39;49m\u001b[43m.\u001b[49m\u001b[43mmodel_validate\u001b[49m\u001b[43m(\u001b[49m\u001b[43mprocessed_data\u001b[49m\u001b[43m)\u001b[49m\n", "\u001b[36mFile \u001b[39m\u001b[32m~/.conda/envs/evoagentx/lib/python3.11/site-packages/pydantic/main.py:705\u001b[39m, in \u001b[36mBaseModel.model_validate\u001b[39m\u001b[34m(cls, obj, strict, from_attributes, context, by_alias, by_name)\u001b[39m\n\u001b[32m 699\u001b[39m \u001b[38;5;28;01mif\u001b[39;00m by_alias \u001b[38;5;129;01mis\u001b[39;00m \u001b[38;5;28;01mFalse\u001b[39;00m \u001b[38;5;129;01mand\u001b[39;00m by_name \u001b[38;5;129;01mis\u001b[39;00m \u001b[38;5;129;01mnot\u001b[39;00m \u001b[38;5;28;01mTrue\u001b[39;00m:\n\u001b[32m 700\u001b[39m \u001b[38;5;28;01mraise\u001b[39;00m PydanticUserError(\n\u001b[32m 701\u001b[39m \u001b[33m'\u001b[39m\u001b[33mAt least one of `by_alias` or `by_name` must be set to True.\u001b[39m\u001b[33m'\u001b[39m,\n\u001b[32m 702\u001b[39m code=\u001b[33m'\u001b[39m\u001b[33mvalidate-by-alias-and-name-false\u001b[39m\u001b[33m'\u001b[39m,\n\u001b[32m 703\u001b[39m )\n\u001b[32m--> \u001b[39m\u001b[32m705\u001b[39m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[38;5;28;43mcls\u001b[39;49m\u001b[43m.\u001b[49m\u001b[43m__pydantic_validator__\u001b[49m\u001b[43m.\u001b[49m\u001b[43mvalidate_python\u001b[49m\u001b[43m(\u001b[49m\n\u001b[32m 706\u001b[39m \u001b[43m \u001b[49m\u001b[43mobj\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mstrict\u001b[49m\u001b[43m=\u001b[49m\u001b[43mstrict\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mfrom_attributes\u001b[49m\u001b[43m=\u001b[49m\u001b[43mfrom_attributes\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mcontext\u001b[49m\u001b[43m=\u001b[49m\u001b[43mcontext\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mby_alias\u001b[49m\u001b[43m=\u001b[49m\u001b[43mby_alias\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mby_name\u001b[49m\u001b[43m=\u001b[49m\u001b[43mby_name\u001b[49m\n\u001b[32m 707\u001b[39m \u001b[43m\u001b[49m\u001b[43m)\u001b[49m\n", "\u001b[31mValidationError\u001b[39m: 1 validation error for CustomizeAgent\n Value error, 'text' is an invalid value for `parse_mode`. Available choices: ['str', 'json', 'xml', 'title', 'custom']. [type=value_error, input_value={'name': 'AnswerGenerateA..., response_format=None)}, input_type=dict]\n For further information visit https://errors.pydantic.dev/2.11/v/value_error" ] } ], "source": [ "benchmark = PubMedQASplits()\n", "workflow_graph = SequentialWorkFlowGraph.from_dict(hotpotqa_graph_data)\n", "agent_manager = AgentManager()\n", "agent_manager.add_agents_from_workflow(workflow_graph, executor_llm.config)\n", "\n", "evaluator = Evaluator(\n", " llm=executor_llm, \n", " agent_manager=agent_manager, \n", " collate_func=collate_func, \n", " num_workers=20, \n", " verbose=True\n", ")\n", "\n", "textgrad_optimizer = TextGradOptimizer(\n", " graph=workflow_graph, \n", " optimize_mode=\"all\",\n", " executor_llm=executor_llm, \n", " optimizer_llm=optimizer_llm,\n", " batch_size=3,\n", " max_steps=20,\n", " evaluator=evaluator,\n", " eval_every_n_steps=1,\n", " eval_rounds=1,\n", " save_interval=None,\n", " save_path=\"./\",\n", " rollback=True,\n", " constraints=[]\n", ")\n" ] }, { "cell_type": "code", "execution_count": 13, "id": "8a6b52c8", "metadata": { "scrolled": true }, "outputs": [ { "data": { "text/plain": [ "500" ] }, "execution_count": 13, "metadata": {}, "output_type": "execute_result" } ], "source": [ "len(benchmark._test_data)" ] }, { "cell_type": "code", "execution_count": 14, "id": "1ac40ae0", "metadata": { "scrolled": true }, "outputs": [ { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2025-12-21 12:56:22.927\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36m__main__\u001b[0m:\u001b[36m\u001b[0m:\u001b[36m1\u001b[0m - \u001b[1mEvaluating workflow on test set...\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 0%| | 1/500 [00:01<14:05, 1.69s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 1%| | 4/500 [00:01<03:09, 2.62it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 1%| | 6/500 [00:02<02:19, 3.54it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 2%|▏ | 10/500 [00:02<01:32, 5.31it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 3%|▎ | 13/500 [00:02<01:05, 7.49it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 3%|▎ | 15/500 [00:03<00:55, 8.73it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 4%|▎ | 18/500 [00:03<01:11, 6.73it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 4%|▍ | 19/500 [00:03<01:09, 6.92it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 4%|▍ | 21/500 [00:04<01:29, 5.33it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 5%|▍ | 23/500 [00:04<01:27, 5.43it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 5%|▌ | 26/500 [00:05<01:16, 6.20it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 6%|▌ | 29/500 [00:05<00:54, 8.67it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 6%|▌ | 30/500 [00:05<01:08, 6.88it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 7%|▋ | 33/500 [00:05<01:05, 7.08it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 7%|▋ | 35/500 [00:06<00:51, 9.04it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 7%|▋ | 37/500 [00:06<01:08, 6.79it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 8%|▊ | 38/500 [00:07<01:55, 4.00it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 8%|▊ | 42/500 [00:07<01:12, 6.28it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 9%|▉ | 47/500 [00:07<00:43, 10.43it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 10%|█ | 50/500 [00:08<00:54, 8.29it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 11%|█ | 54/500 [00:08<00:48, 9.12it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 11%|█ | 56/500 [00:09<00:57, 7.69it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 12%|█▏ | 60/500 [00:09<01:00, 7.23it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 12%|█▏ | 62/500 [00:10<01:16, 5.70it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 13%|█▎ | 64/500 [00:10<01:02, 6.92it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 14%|█▍ | 69/500 [00:10<00:45, 9.42it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 15%|█▍ | 73/500 [00:11<00:52, 8.17it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 15%|█▌ | 77/500 [00:11<00:42, 9.98it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 16%|█▌ | 79/500 [00:12<00:55, 7.60it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 16%|█▌ | 81/500 [00:12<00:59, 7.03it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 17%|█▋ | 86/500 [00:13<00:42, 9.73it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 18%|█▊ | 89/500 [00:13<01:01, 6.69it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 18%|█▊ | 92/500 [00:14<01:02, 6.48it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 19%|█▉ | 94/500 [00:14<00:50, 8.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 19%|█▉ | 97/500 [00:14<00:44, 9.08it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 20%|█▉ | 99/500 [00:15<01:07, 5.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 20%|██ | 101/500 [00:15<01:19, 4.99it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 21%|██ | 106/500 [00:16<00:49, 8.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 22%|██▏ | 108/500 [00:16<01:11, 5.49it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 22%|██▏ | 111/500 [00:17<01:07, 5.76it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 23%|██▎ | 115/500 [00:17<00:51, 7.52it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 23%|██▎ | 116/500 [00:17<00:53, 7.18it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 24%|██▎ | 118/500 [00:18<01:03, 6.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 24%|██▍ | 121/500 [00:18<00:55, 6.81it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 25%|██▌ | 125/500 [00:18<00:35, 10.45it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 26%|██▌ | 129/500 [00:19<00:44, 8.38it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 26%|██▌ | 131/500 [00:19<00:38, 9.47it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 27%|██▋ | 135/500 [00:19<00:33, 10.86it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 27%|██▋ | 137/500 [00:20<00:40, 9.07it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 28%|██▊ | 139/500 [00:20<01:04, 5.60it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 28%|██▊ | 142/500 [00:21<00:51, 6.98it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 29%|██▉ | 144/500 [00:21<00:48, 7.41it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 30%|██▉ | 149/500 [00:21<00:30, 11.45it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 30%|███ | 152/500 [00:21<00:23, 14.60it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 31%|███ | 154/500 [00:22<00:48, 7.12it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 32%|███▏ | 158/500 [00:23<00:45, 7.44it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 32%|███▏ | 160/500 [00:23<00:38, 8.77it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 32%|███▏ | 162/500 [00:23<00:37, 9.10it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 33%|███▎ | 164/500 [00:23<00:36, 9.10it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 34%|███▍ | 170/500 [01:00<20:30, 3.73s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 35%|███▍ | 173/500 [01:01<12:15, 2.25s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 35%|███▌ | 176/500 [01:01<07:06, 1.32s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 36%|███▌ | 178/500 [01:01<05:21, 1.00it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 36%|███▌ | 180/500 [01:02<04:16, 1.25it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 37%|███▋ | 184/500 [01:02<02:11, 2.41it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 37%|███▋ | 186/500 [01:02<01:38, 3.19it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 38%|███▊ | 188/500 [01:03<01:28, 3.52it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 38%|███▊ | 191/500 [01:04<01:35, 3.23it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 39%|███▊ | 193/500 [01:04<01:16, 4.00it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 39%|███▉ | 195/500 [01:04<00:54, 5.58it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 40%|███▉ | 198/500 [01:05<00:50, 5.96it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 40%|███▉ | 199/500 [01:05<00:50, 5.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 40%|████ | 200/500 [01:05<01:10, 4.25it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 41%|████ | 203/500 [01:06<01:00, 4.92it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 41%|████ | 206/500 [01:06<00:37, 7.81it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 42%|████▏ | 208/500 [01:06<00:33, 8.75it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 42%|████▏ | 212/500 [01:07<00:35, 8.06it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 43%|████▎ | 214/500 [01:07<00:45, 6.25it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 43%|████▎ | 217/500 [01:08<00:39, 7.13it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 44%|████▍ | 220/500 [01:08<00:36, 7.57it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 44%|████▍ | 222/500 [01:08<00:29, 9.47it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 45%|████▍ | 224/500 [01:09<00:36, 7.63it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 45%|████▌ | 225/500 [01:09<00:49, 5.59it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 46%|████▌ | 228/500 [01:09<00:35, 7.61it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 46%|████▌ | 230/500 [01:10<00:49, 5.44it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 46%|████▋ | 232/500 [01:10<00:36, 7.27it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 47%|████▋ | 234/500 [01:10<00:30, 8.58it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 47%|████▋ | 237/500 [01:10<00:34, 7.58it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 48%|████▊ | 239/500 [01:11<00:56, 4.66it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 49%|████▉ | 244/500 [01:11<00:30, 8.48it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 49%|████▉ | 246/500 [01:12<00:31, 8.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 50%|█████ | 252/500 [01:12<00:22, 10.99it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 51%|█████ | 254/500 [01:12<00:25, 9.79it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 52%|█████▏ | 258/500 [01:13<00:27, 8.64it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 52%|█████▏ | 262/500 [01:13<00:25, 9.25it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 53%|█████▎ | 264/500 [01:14<00:27, 8.60it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 53%|█████▎ | 267/500 [01:14<00:24, 9.53it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 54%|█████▍ | 269/500 [01:14<00:20, 11.06it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 54%|█████▍ | 271/500 [01:15<00:33, 6.91it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 55%|█████▌ | 275/500 [01:15<00:26, 8.52it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 55%|█████▌ | 277/500 [01:16<00:37, 5.93it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 56%|█████▌ | 279/500 [01:16<00:39, 5.63it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 56%|█████▌ | 281/500 [01:16<00:38, 5.67it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 57%|█████▋ | 284/500 [01:17<00:29, 7.23it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 58%|█████▊ | 288/500 [01:17<00:19, 10.79it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 58%|█████▊ | 290/500 [01:17<00:24, 8.51it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 58%|█████▊ | 292/500 [01:17<00:20, 10.40it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 59%|█████▉ | 294/500 [01:18<00:28, 7.16it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 59%|█████▉ | 297/500 [01:18<00:28, 7.05it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 60%|█████▉ | 298/500 [01:19<00:36, 5.59it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 60%|██████ | 301/500 [01:19<00:30, 6.54it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 60%|██████ | 302/500 [01:19<00:28, 6.89it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 61%|██████ | 305/500 [01:20<00:27, 7.14it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 62%|██████▏ | 308/500 [01:20<00:19, 9.68it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 62%|██████▏ | 310/500 [01:20<00:18, 10.31it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 62%|██████▏ | 312/500 [01:20<00:18, 10.27it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 63%|██████▎ | 314/500 [01:21<00:31, 5.82it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 64%|██████▎ | 318/500 [01:21<00:20, 8.83it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 64%|██████▍ | 320/500 [01:21<00:23, 7.80it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 65%|██████▍ | 323/500 [01:22<00:30, 5.84it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 65%|██████▌ | 326/500 [01:22<00:21, 8.14it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 66%|██████▌ | 330/500 [01:23<00:18, 9.17it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 66%|██████▋ | 332/500 [01:23<00:22, 7.35it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 67%|██████▋ | 335/500 [01:24<00:24, 6.65it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 67%|██████▋ | 337/500 [01:24<00:23, 6.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 68%|██████▊ | 339/500 [01:24<00:19, 8.32it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 68%|██████▊ | 341/500 [01:25<00:24, 6.55it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 69%|██████▊ | 343/500 [02:00<14:55, 5.70s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 69%|██████▉ | 345/500 [02:01<09:05, 3.52s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 70%|███████ | 350/500 [02:01<03:08, 1.26s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 70%|███████ | 352/500 [02:01<02:16, 1.09it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:58:24.690\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 71%|███████ | 354/500 [02:01<01:38, 1.48it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:58:24.724\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'pubid': 25394614, 'question': 'Does timing of initial surfactant treatment make a difference in rates of chronic lung disease or mortality in premature infants?', 'context': {'contexts': ['To compare two treatment strategies in preterm infants with or at risk of respiratory distress syndrome: early surfactant administration (within one hour of birth) versus late surfactant administration, in a geographically defined population.', 'The primary outcome was chronic lung disease (CLD) and mortality before/at 36 weeks. Secondary outcomes included: duration of mechanical ventilation and continuous positive airway pressure (CPAP), post-natal steroids for CLD and major neonatal morbidities.', \"Premature infants born at 22-32 weeks' gestation between January 2006 and December 2009.\", 'Ten neonatal intensive care units (NICUs) in New South Wales (NSW) and Australian Capital Territory (ACT), Australia.', 'Retrospective analysis of prospectively collected data from the regional NICU database in NSW and ACT.', 'Of the 2170 infants who received surfactant, 1182 (54.5%) and 988 (45.5%) received early and late surfactant, respectively. The early surfactant group was less mature (27.1\\u2009±\\u20092.1 versus 29.4\\u2009±\\u20092.1 weeks) and had more CLD and mortality (40.2% versus 20.0%). The multivariable analysis showed early surfactant to be associated with less duration of ventilation, longer duration of CPAP and longer hospital stay but had little or no impact on CLD/mortality.'], 'labels': ['OBJECTIVE', 'OUTCOME', 'SUBJECTS', 'SETTING', 'DESIGN', 'RESULTS'], 'meshes': ['Australian Capital Territory', 'Chronic Disease', 'Female', 'Humans', 'Infant, Newborn', 'Male', 'New South Wales', 'Pulmonary Surfactants', 'Respiratory Distress Syndrome, Newborn', 'Treatment Outcome'], 'reasoning_required_pred': ['y', 'e', 's'], 'reasoning_free_pred': ['m', 'a', 'y', 'b', 'e']}, 'long_answer': 'Early surfactant administration is associated with shorter duration of ventilation but does not appear to be significantly protective against CLD/mortality among premature infants. This may support the growing evidence for consideration of CPAP as an alternative to routine intubation and early surfactant administration. Further investigation from large randomized clinical trials is warranted to confirm these results.', 'final_decision': 'maybe'}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:58:24.786\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 72%|███████▏ | 358/500 [02:02<01:11, 1.99it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 72%|███████▏ | 360/500 [02:03<00:58, 2.40it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 73%|███████▎ | 363/500 [02:03<00:49, 2.77it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 73%|███████▎ | 364/500 [02:04<00:46, 2.93it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 73%|███████▎ | 365/500 [02:04<00:46, 2.89it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 74%|███████▍ | 369/500 [02:04<00:29, 4.44it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 74%|███████▍ | 372/500 [02:05<00:18, 6.78it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 75%|███████▍ | 374/500 [02:05<00:19, 6.33it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 75%|███████▌ | 377/500 [02:05<00:20, 6.13it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 76%|███████▌ | 378/500 [02:06<00:21, 5.61it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 76%|███████▌ | 380/500 [02:06<00:18, 6.46it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 77%|███████▋ | 384/500 [02:06<00:12, 9.35it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 77%|███████▋ | 386/500 [02:06<00:10, 10.83it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 78%|███████▊ | 388/500 [02:07<00:21, 5.11it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 78%|███████▊ | 391/500 [02:08<00:19, 5.53it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 79%|███████▉ | 396/500 [02:08<00:14, 7.23it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 80%|███████▉ | 399/500 [02:09<00:16, 6.12it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 80%|████████ | 400/500 [02:09<00:16, 6.10it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 81%|████████ | 403/500 [02:09<00:15, 6.23it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 81%|████████ | 406/500 [02:10<00:12, 7.33it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 82%|████████▏ | 409/500 [02:10<00:12, 7.53it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 82%|████████▏ | 412/500 [02:11<00:13, 6.50it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 83%|████████▎ | 414/500 [02:11<00:14, 5.89it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 83%|████████▎ | 417/500 [02:11<00:10, 7.59it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 84%|████████▍ | 420/500 [02:12<00:09, 8.37it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 84%|████████▍ | 421/500 [02:12<00:13, 6.06it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 85%|████████▍ | 424/500 [02:13<00:12, 6.12it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 85%|████████▌ | 425/500 [02:13<00:12, 5.83it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 85%|████████▌ | 427/500 [02:13<00:13, 5.44it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 86%|████████▋ | 432/500 [02:13<00:05, 11.87it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 87%|████████▋ | 434/500 [02:14<00:05, 12.86it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 87%|████████▋ | 436/500 [02:14<00:07, 8.22it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 88%|████████▊ | 441/500 [02:14<00:05, 11.10it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 89%|████████▊ | 443/500 [02:15<00:05, 10.96it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 89%|████████▉ | 445/500 [02:15<00:07, 7.44it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 90%|████████▉ | 448/500 [02:16<00:07, 6.87it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 90%|█████████ | 450/500 [02:16<00:07, 6.76it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 91%|█████████ | 456/500 [02:16<00:04, 9.32it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 92%|█████████▏| 460/500 [02:17<00:05, 6.86it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 92%|█████████▏| 462/500 [02:17<00:04, 7.68it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 93%|█████████▎| 464/500 [02:18<00:04, 8.32it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 93%|█████████▎| 466/500 [02:18<00:04, 8.37it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 94%|█████████▍| 469/500 [02:18<00:03, 7.86it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 94%|█████████▍| 472/500 [02:19<00:05, 5.14it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 95%|█████████▍| 474/500 [02:19<00:03, 6.94it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 95%|█████████▌| 476/500 [02:19<00:03, 7.94it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 96%|█████████▌| 478/500 [02:20<00:02, 7.52it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 96%|█████████▌| 480/500 [02:20<00:02, 7.26it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 96%|█████████▋| 482/500 [02:21<00:02, 6.11it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 97%|█████████▋| 484/500 [02:21<00:02, 6.35it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 97%|█████████▋| 486/500 [02:21<00:02, 5.96it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 98%|█████████▊| 490/500 [02:21<00:01, 9.18it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 98%|█████████▊| 492/500 [02:22<00:01, 7.53it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 99%|█████████▉| 497/500 [02:22<00:00, 9.79it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 100%|█████████▉| 499/500 [02:24<00:00, 3.09it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 100%|██████████| 500/500 [02:24<00:00, 3.45it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 12:58:47.845\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36m__main__\u001b[0m:\u001b[36m\u001b[0m:\u001b[36m4\u001b[0m - \u001b[1mEvaluation metrics (before optimization): {'f1': 0.0, 'em': 0.0, 'acc': 0.6813627254509018}\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\n" ] } ], "source": [ "logger.info(\"Evaluating workflow on test set...\")\n", "with suppress_logger_info():\n", " results = textgrad_optimizer.evaluate(dataset=benchmark, eval_mode=\"test\")\n", "logger.info(f\"Evaluation metrics (before optimization): {results}\")\n", "\n", "# logger.info(\"Optimizing workflow...\")\n", "# textgrad_optimizer.optimize(benchmark, seed=8)\n", "# textgrad_optimizer.restore_best_graph()\n", "\n", "# logger.info(\"Evaluating workflow on test set...\")\n", "# with suppress_logger_info():\n", "# results = textgrad_optimizer.evaluate(dataset=benchmark, eval_mode=\"test\")\n", "# logger.info(f\"Evaluation metrics (after optimization): {results}\")" ] }, { "cell_type": "code", "execution_count": 15, "id": "809b5654", "metadata": {}, "outputs": [ { "data": { "text/plain": [ "{'f1': 0.0, 'em': 0.0, 'acc': 0.6813627254509018}" ] }, "execution_count": 15, "metadata": {}, "output_type": "execute_result" } ], "source": [ "results" ] }, { "cell_type": "code", "execution_count": 19, "id": "6ff35025", "metadata": {}, "outputs": [], "source": [ "#generated_workflow\n", "qa_workflow = {\n", " \"goal\": \"Provide a concise answer to the question using relevant context. The answer must be straightforward and avoid unnecessary explanations.\",\n", " \"tasks\": [\n", " {\n", " \"name\": \"generate_answer\",\n", " \"description\": \"Extract and formulate an answer from the given context.\",\n", " \"inputs\": [\n", " {\"name\": \"question\", \"type\": \"str\", \"required\": True, \"description\": \"The question that needs to be answered.\"},\n", " {\"name\": \"context\", \"type\": \"str\", \"required\": True, \"description\": \"The background information pertinent to the question.\"}\n", " ],\n", " \"outputs\": [\n", " {\"name\": \"answer\", \"type\": \"str\", \"required\": True, \"description\": \"The direct answer to the question.\"}\n", " ],\n", " \"prompt_template\": StringTemplate(instruction=\"Use the context to determine the best answer to the question. Provide your final answer in a clear format, without extra commentary or reasoning.\"),\n", " \"parse_mode\": \"xml\"\n", " }\n", " ]\n", "}" ] }, { "cell_type": "code", "execution_count": 25, "id": "a4af1fcb", "metadata": {}, "outputs": [], "source": [ "import nest_asyncio\n", "nest_asyncio.apply()\n", "\n", "class PubMedQASplits(PubMedQA):\n", "\n", " def _load_data(self):\n", " # load the original test data \n", " super()._load_data()\n", " # split the data into train, dev and test\n", " import numpy as np \n", " np.random.seed(42)\n", " permutation = np.random.permutation(len(self._dev_data))\n", " full_test_data = self._dev_data \n", " # randomly select 10 samples for train, 40 for dev, and 100 for test\n", " self._train_data = [full_test_data[idx] for idx in permutation[:50]]\n", " self._dev_data = [full_test_data[idx] for idx in permutation[:50]]\n", " self._test_data =self._test_data[0:500]\n", " self._fulldata = full_test_data\n", "\n", "\n", "def collate_func(example: dict) -> dict:\n", " context_list = []\n", " paragraphs = example[\"context\"][\"contexts\"]\n", " context = \"\\n\".join(paragraphs)\n", " problem = \"Question: {}\\n\\nAnswer:\".format(example[\"question\"])\n", " return {\"question\": problem, 'context':paragraphs}\n", "\n", "os.environ[\"AZURE_OPENAI_DEPLOYMENT_NAME\"] = \"gpt-4o-mini\"\n", "os.environ[\"AZURE_OPENAI_ENDPOINT\"] = \"https://tianyuliu-hua-raredisea-resource.cognitiveservices.azure.com/\"\n", "os.environ[\"AZURE_OPENAI_KEY\"] = \"2pa9h2ZIN1lQepFWwYADlXIKIansa9KPhxMoumeGbRQ08f2uDTXiJQQJ99BKACHYHv6XJ3w3AAAAACOGsQIt\"\n", "os.environ[\"AZURE_OPENAI_API_VERSION\"] = \"2025-01-01-preview\"\n", "llm_config = LiteLLMConfig(model=\"azure/\" + os.getenv(\"AZURE_OPENAI_DEPLOYMENT_NAME\"), # Azure model format\n", " azure_endpoint=os.getenv(\"AZURE_OPENAI_ENDPOINT\"),\n", " azure_key=os.getenv(\"AZURE_OPENAI_KEY\"),\n", " api_version=os.getenv(\"AZURE_OPENAI_API_VERSION\", \"2024-12-01-preview\"), top_p=0.85, temperature=0.2, frequency_penalty=0.0, presence_penalty=0.0)\n", "\n", "executor_llm = LiteLLM(config=llm_config)\n", "optimizer_llm = LiteLLM(config=llm_config)" ] }, { "cell_type": "code", "execution_count": 26, "id": "eab99c76", "metadata": {}, "outputs": [ { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2025-12-21 13:28:14.748\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.benchmark.pubmedqa\u001b[0m:\u001b[36m_load_data_from_file\u001b[0m:\u001b[36m51\u001b[0m - \u001b[1mloading HotPotQA data from /gpfs/radev/pi/ying_rex/tl688/selfevolve/EvoAgentX/examples/pubmedqa/pubmedqa_train.json ...\u001b[0m\n", "\u001b[32m2025-12-21 13:28:16.954\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.benchmark.pubmedqa\u001b[0m:\u001b[36m_load_data_from_file\u001b[0m:\u001b[36m51\u001b[0m - \u001b[1mloading HotPotQA data from /gpfs/radev/pi/ying_rex/tl688/selfevolve/EvoAgentX/examples/pubmedqa/pubmedqa_train.json ...\u001b[0m\n", "\u001b[32m2025-12-21 13:28:19.336\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.benchmark.pubmedqa\u001b[0m:\u001b[36m_load_data_from_file\u001b[0m:\u001b[36m51\u001b[0m - \u001b[1mloading HotPotQA data from /gpfs/radev/pi/ying_rex/tl688/selfevolve/EvoAgentX/examples/pubmedqa/pubmedqa_label.json ...\u001b[0m\n" ] } ], "source": [ "benchmark = PubMedQASplits()\n", "workflow_graph = SequentialWorkFlowGraph.from_dict(qa_workflow)\n", "agent_manager = AgentManager()\n", "agent_manager.add_agents_from_workflow(workflow_graph, executor_llm.config)\n", "\n", "evaluator = Evaluator(\n", " llm=executor_llm, \n", " agent_manager=agent_manager, \n", " collate_func=collate_func, \n", " num_workers=20, \n", " verbose=True\n", ")\n", "\n", "textgrad_optimizer = TextGradOptimizer(\n", " graph=workflow_graph, \n", " optimize_mode=\"all\",\n", " executor_llm=executor_llm, \n", " optimizer_llm=optimizer_llm,\n", " batch_size=3,\n", " max_steps=20,\n", " evaluator=evaluator,\n", " eval_every_n_steps=1,\n", " eval_rounds=1,\n", " save_interval=None,\n", " save_path=\"./\",\n", " rollback=True,\n", " constraints=[]\n", ")\n" ] }, { "cell_type": "code", "execution_count": 27, "id": "5f42406e", "metadata": { "scrolled": true }, "outputs": [ { "data": { "text/plain": [ "500" ] }, "execution_count": 27, "metadata": {}, "output_type": "execute_result" } ], "source": [ "len(benchmark._test_data)" ] }, { "cell_type": "code", "execution_count": 28, "id": "dcc6a821", "metadata": { "scrolled": true }, "outputs": [ { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2025-12-21 13:28:19.656\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36m__main__\u001b[0m:\u001b[36m\u001b[0m:\u001b[36m1\u001b[0m - \u001b[1mEvaluating workflow on test set...\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 1%| | 4/500 [00:01<01:45, 4.69it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 1%|▏ | 7/500 [00:01<00:58, 8.46it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 2%|▏ | 11/500 [00:01<00:40, 12.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 3%|▎ | 17/500 [00:01<00:33, 14.59it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 4%|▍ | 19/500 [00:01<00:37, 12.98it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 4%|▍ | 21/500 [00:02<00:46, 10.28it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 5%|▌ | 27/500 [00:02<00:34, 13.80it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 6%|▋ | 32/500 [00:02<00:29, 15.67it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 7%|▋ | 36/500 [00:03<00:34, 13.51it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 8%|▊ | 38/500 [00:03<00:33, 13.88it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 9%|▊ | 43/500 [00:03<00:32, 14.23it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 10%|▉ | 48/500 [00:04<00:33, 13.62it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 10%|█ | 52/500 [00:04<00:25, 17.38it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 11%|█ | 54/500 [00:04<00:30, 14.46it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 12%|█▏ | 59/500 [00:04<00:27, 15.87it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 12%|█▏ | 61/500 [00:04<00:26, 16.68it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 13%|█▎ | 65/500 [00:05<00:36, 11.99it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 13%|█▎ | 67/500 [00:05<00:32, 13.15it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 14%|█▍ | 70/500 [00:05<00:27, 15.55it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 15%|█▌ | 75/500 [00:05<00:29, 14.24it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 16%|█▌ | 78/500 [00:06<00:30, 14.07it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 16%|█▌ | 80/500 [00:06<00:32, 12.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 17%|█▋ | 84/500 [00:06<00:37, 11.10it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 17%|█▋ | 87/500 [00:06<00:30, 13.64it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 18%|█▊ | 92/500 [00:07<00:32, 12.72it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 20%|█▉ | 99/500 [00:07<00:23, 16.88it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 20%|██ | 101/500 [00:07<00:29, 13.41it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 21%|██ | 103/500 [00:08<00:32, 12.35it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 21%|██▏ | 107/500 [00:08<00:28, 13.74it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 22%|██▏ | 111/500 [00:08<00:32, 12.08it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 23%|██▎ | 117/500 [00:08<00:19, 19.90it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 24%|██▍ | 120/500 [00:09<00:27, 13.81it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 25%|██▍ | 123/500 [00:09<00:24, 15.35it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 25%|██▌ | 126/500 [00:09<00:29, 12.50it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 26%|██▌ | 128/500 [00:10<00:34, 10.64it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 27%|██▋ | 133/500 [00:10<00:28, 12.71it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 27%|██▋ | 137/500 [00:10<00:23, 15.61it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 29%|██▊ | 143/500 [00:10<00:19, 18.27it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 29%|██▉ | 146/500 [00:11<00:25, 13.87it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 30%|███ | 150/500 [00:11<00:28, 12.48it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 30%|███ | 152/500 [00:11<00:26, 13.30it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 31%|███ | 156/500 [00:12<00:25, 13.33it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 32%|███▏ | 161/500 [00:12<00:20, 16.94it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 33%|███▎ | 166/500 [00:12<00:20, 16.55it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 34%|███▎ | 168/500 [00:12<00:25, 12.94it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 35%|███▍ | 173/500 [00:13<00:24, 13.59it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 35%|███▌ | 176/500 [00:13<00:21, 15.19it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 36%|███▌ | 180/500 [00:13<00:20, 15.51it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 36%|███▋ | 182/500 [00:14<00:35, 9.03it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 37%|███▋ | 185/500 [01:00<24:07, 4.59s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 38%|███▊ | 189/500 [01:01<12:21, 2.38s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 39%|███▊ | 193/500 [01:01<06:16, 1.22s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 39%|███▉ | 197/500 [01:01<03:31, 1.43it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 40%|████ | 200/500 [01:02<02:28, 2.02it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 41%|████ | 204/500 [01:02<01:35, 3.11it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 41%|████ | 206/500 [01:02<01:17, 3.78it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 42%|████▏ | 208/500 [01:02<01:06, 4.42it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 42%|████▏ | 210/500 [01:02<00:56, 5.10it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 42%|████▏ | 212/500 [01:03<00:50, 5.74it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 44%|████▍ | 219/500 [01:03<00:30, 9.10it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 45%|████▌ | 226/500 [01:03<00:18, 14.78it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 46%|████▌ | 228/500 [01:03<00:20, 13.38it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 47%|████▋ | 233/500 [01:04<00:20, 12.79it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 48%|████▊ | 238/500 [01:04<00:19, 13.31it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 48%|████▊ | 240/500 [01:04<00:24, 10.82it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 49%|████▉ | 244/500 [01:05<00:27, 9.15it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 50%|█████ | 250/500 [01:05<00:17, 14.51it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 50%|█████ | 252/500 [01:05<00:18, 13.10it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 51%|█████ | 254/500 [01:06<00:20, 11.81it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 52%|█████▏ | 258/500 [01:06<00:23, 10.36it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 52%|█████▏ | 260/500 [01:06<00:21, 10.93it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 53%|█████▎ | 266/500 [01:07<00:15, 15.32it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 54%|█████▎ | 268/500 [01:07<00:20, 11.13it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 54%|█████▍ | 270/500 [01:07<00:18, 12.19it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 54%|█████▍ | 272/500 [01:07<00:19, 11.91it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 56%|█████▌ | 279/500 [01:08<00:15, 14.66it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 57%|█████▋ | 284/500 [01:08<00:14, 14.75it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 57%|█████▋ | 286/500 [01:08<00:18, 11.83it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 58%|█████▊ | 291/500 [01:09<00:16, 12.72it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 60%|█████▉ | 299/500 [01:09<00:09, 21.76it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 60%|██████ | 302/500 [01:09<00:10, 18.31it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 61%|██████ | 305/500 [01:10<00:14, 13.05it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 62%|██████▏ | 311/500 [01:10<00:13, 13.74it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 63%|██████▎ | 316/500 [01:10<00:12, 15.03it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 64%|██████▍ | 321/500 [01:10<00:09, 18.48it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 65%|██████▍ | 324/500 [01:11<00:10, 16.86it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 66%|██████▌ | 328/500 [01:11<00:09, 18.06it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 66%|██████▋ | 332/500 [01:11<00:11, 14.13it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 67%|██████▋ | 334/500 [01:11<00:11, 13.89it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 68%|██████▊ | 338/500 [01:12<00:13, 12.02it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 69%|██████▉ | 345/500 [01:12<00:08, 19.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 70%|██████▉ | 348/500 [01:12<00:08, 17.78it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 70%|███████ | 350/500 [01:13<00:11, 13.17it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 70%|███████ | 352/500 [01:13<00:12, 11.47it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 71%|███████ | 355/500 [01:13<00:13, 10.93it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 73%|███████▎ | 363/500 [01:13<00:09, 15.17it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 73%|███████▎ | 365/500 [01:14<00:11, 11.44it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 73%|███████▎ | 367/500 [01:14<00:14, 8.90it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 74%|███████▍ | 369/500 [02:01<12:34, 5.76s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 74%|███████▍ | 371/500 [02:01<08:50, 4.11s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 75%|███████▍ | 373/500 [02:01<05:30, 2.60s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 75%|███████▌ | 376/500 [02:01<02:47, 1.35s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 76%|███████▌ | 381/500 [02:02<01:13, 1.62it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 77%|███████▋ | 385/500 [02:02<00:41, 2.75it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 78%|███████▊ | 388/500 [02:02<00:27, 4.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 79%|███████▉ | 395/500 [02:03<00:12, 8.13it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 79%|███████▉ | 397/500 [02:03<00:11, 8.91it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 80%|████████ | 401/500 [02:03<00:10, 9.81it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 81%|████████ | 403/500 [02:04<00:09, 9.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 81%|████████▏ | 407/500 [02:04<00:09, 9.70it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 82%|████████▏ | 412/500 [02:04<00:06, 13.75it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 83%|████████▎ | 416/500 [02:05<00:06, 13.74it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 84%|████████▎ | 418/500 [02:05<00:09, 8.62it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 84%|████████▍ | 422/500 [02:05<00:08, 9.41it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 85%|████████▍ | 424/500 [02:06<00:08, 9.41it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 87%|████████▋ | 433/500 [02:06<00:03, 19.19it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 87%|████████▋ | 436/500 [02:06<00:04, 15.17it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 88%|████████▊ | 441/500 [02:07<00:05, 11.51it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 89%|████████▉ | 445/500 [02:07<00:03, 15.10it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 91%|█████████ | 453/500 [02:07<00:03, 14.94it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 91%|█████████▏| 457/500 [02:08<00:03, 14.14it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 92%|█████████▏| 459/500 [02:08<00:02, 14.60it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 93%|█████████▎| 463/500 [02:08<00:03, 12.30it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 93%|█████████▎| 467/500 [02:09<00:02, 12.07it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 94%|█████████▍| 469/500 [02:09<00:02, 12.26it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 95%|█████████▍| 474/500 [02:09<00:01, 17.93it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 96%|█████████▌| 479/500 [02:09<00:01, 13.27it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 97%|█████████▋| 483/500 [02:10<00:01, 11.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 97%|█████████▋| 485/500 [02:10<00:01, 13.08it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 98%|█████████▊| 488/500 [02:10<00:00, 13.44it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 98%|█████████▊| 492/500 [02:11<00:00, 12.19it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 99%|█████████▉| 497/500 [02:11<00:00, 13.41it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 100%|██████████| 500/500 [02:11<00:00, 3.80it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2025-12-21 13:30:31.423\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36m__main__\u001b[0m:\u001b[36m\u001b[0m:\u001b[36m4\u001b[0m - \u001b[1mEvaluation metrics (before optimization): {'f1': 0.0, 'em': 0.0, 'acc': 0.664}\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\n" ] } ], "source": [ "logger.info(\"Evaluating workflow on test set...\")\n", "with suppress_logger_info():\n", " results = textgrad_optimizer.evaluate(dataset=benchmark, eval_mode=\"test\")\n", "logger.info(f\"Evaluation metrics (before optimization): {results}\")\n", "\n", "# logger.info(\"Optimizing workflow...\")\n", "# textgrad_optimizer.optimize(benchmark, seed=8)\n", "# textgrad_optimizer.restore_best_graph()\n", "\n", "# logger.info(\"Evaluating workflow on test set...\")\n", "# with suppress_logger_info():\n", "# results = textgrad_optimizer.evaluate(dataset=benchmark, eval_mode=\"test\")\n", "# logger.info(f\"Evaluation metrics (after optimization): {results}\")" ] }, { "cell_type": "code", "execution_count": 29, "id": "c374857b", "metadata": {}, "outputs": [ { "data": { "text/plain": [ "{'f1': 0.0, 'em': 0.0, 'acc': 0.664}" ] }, "execution_count": 29, "metadata": {}, "output_type": "execute_result" } ], "source": [ "results" ] }, { "cell_type": "code", "execution_count": null, "id": "13d0c3d6", "metadata": {}, "outputs": [], "source": [] }, { "cell_type": "code", "execution_count": null, "id": "c11135bd", "metadata": {}, "outputs": [], "source": [] }, { "cell_type": "code", "execution_count": null, "id": "391e90be", "metadata": {}, "outputs": [], "source": [] } ], "metadata": { "kernelspec": { "display_name": "Python 3 (ipykernel)", "language": "python", "name": "python3" }, "language_info": { "codemirror_mode": { "name": "ipython", "version": 3 }, "file_extension": ".py", "mimetype": "text/x-python", "name": "python", "nbconvert_exporter": "python", "pygments_lexer": "ipython3", "version": "3.11.13" } }, "nbformat": 4, "nbformat_minor": 5 }