{ "cells": [ { "cell_type": "code", "execution_count": 1, "id": "4c2a6fa7", "metadata": {}, "outputs": [ { "name": "stderr", "output_type": "stream", "text": [ "/gpfs/radev/home/tl688/.conda/envs/evoagentx/lib/python3.11/site-packages/PyPDF2/__init__.py:21: DeprecationWarning: PyPDF2 is deprecated. Please move to the pypdf library instead.\n", " warnings.warn(\n" ] } ], "source": [ "import os\n", "\n", "from dotenv import load_dotenv\n", "\n", "from evoagentx.agents.agent_manager import AgentManager\n", "from evoagentx.benchmark import HotPotQA\n", "from evoagentx.core.callbacks import suppress_logger_info\n", "from evoagentx.core.logging import logger\n", "from evoagentx.evaluators import Evaluator\n", "from evoagentx.models import OpenAILLM, OpenAILLMConfig\n", "from evoagentx.optimizers import TextGradOptimizer\n", "from evoagentx.prompts import StringTemplate\n", "from evoagentx.workflow import SequentialWorkFlowGraph\n", "from dotenv import load_dotenv\n", "\n", "from evoagentx.agents.agent_manager import AgentManager\n", "from evoagentx.benchmark import MBPP\n", "from evoagentx.core.callbacks import suppress_logger_info\n", "from evoagentx.core.logging import logger\n", "from evoagentx.evaluators import Evaluator\n", "from evoagentx.models import OpenAILLM, OpenAILLMConfig\n", "from evoagentx.optimizers import TextGradOptimizer\n", "from evoagentx.prompts import StringTemplate\n", "from evoagentx.workflow import SequentialWorkFlowGraph\n", "\n", "from evoagentx.models import OpenAILLMConfig, OpenAILLM\n", "from evoagentx.workflow import SEWWorkFlowGraph, STRUCTUREWorkFlowGraph\n", "from evoagentx.agents import AgentManager\n", "from evoagentx.benchmark import HumanEval,AFlowMBPP\n", "from evoagentx.evaluators import Evaluator \n", "from evoagentx.optimizers import SEWOptimizer, STRUCTUREOptimizer\n", "from evoagentx.optimizers.structure_optimizer import STRUCTUREWorkFlowScheme\n", "from evoagentx.core.callbacks import suppress_logger_info\n", "\n", "from evoagentx.models import OpenAILLMConfig, OpenAILLM,AzureOpenAIConfig,LiteLLMConfig,LiteLLM\n", "from evoagentx.workflow import SEWWorkFlowGraph \n", "from evoagentx.agents import AgentManager\n", "from evoagentx.benchmark import MBPPPLUS, AFlowMBPPPLUS\n", "from evoagentx.evaluators import Evaluator \n", "from evoagentx.optimizers import SEWOptimizer \n", "from evoagentx.core.callbacks import suppress_logger_info\n", "from evoagentx.benchmark import HumanEvalPLUS\n", "from evoagentx.benchmark import SciCode\n", "from copy import deepcopy\n", "from evoagentx.benchmark import PertQA\n", "from copy import deepcopy\n", "from evoagentx.benchmark import MolQA\n", "import nest_asyncio\n", "nest_asyncio.apply()\n", "\n", "class HotPotQASplits(HotPotQA):\n", "\n", " def _load_data(self):\n", " # load the original test data \n", " super()._load_data()\n", " # split the data into train, dev and test\n", " import numpy as np \n", " np.random.seed(42)\n", " permutation = np.random.permutation(len(self._dev_data))\n", " full_test_data = self._dev_data \n", " # randomly select 10 samples for train, 40 for dev, and 100 for test\n", " self._train_data = [full_test_data[idx] for idx in permutation[:50]]\n", " self._dev_data = [full_test_data[idx] for idx in permutation[:50]]\n", " self._test_data = [full_test_data[idx] for idx in permutation[50:550]]\n", " self._fulldata = full_test_data\n", "\n", "\n", "def collate_func(example: dict) -> dict:\n", " problem = \"Question: {}\\n\\nAnswer:\".format(example[\"question\"])\n", " return {\"problem\": problem}\n", "\n", "\n", "hotpotqa_graph_data = {\n", " \"goal\": \"Answer the question based on the context. The answer should be a direct response to the question, without including explanations or reasoning.\",\n", " \"tasks\": [\n", " {\n", " \"name\": \"answer_generate\",\n", " \"description\": \"Answer the question based on the context.\",\n", " \"inputs\": [\n", " {\"name\": \"problem\", \"type\": \"str\", \"required\": True, \"description\": \"The problem to solve.\"}\n", " ],\n", " \"outputs\": [\n", " {\"name\": \"answer\", \"type\": \"str\", \"required\": True, \"description\": \"The answer to the problem.\"}\n", " ],\n", " \"prompt_template\": StringTemplate(instruction=\"Think step by step to answer the question. You should explain your thinking process in the 'thought' field, and provide the final answer in the 'answer' field.\\nFormat your output in xml format, such as xxx and xxx.\"),\n", " \"parse_mode\": \"xml\"\n", " }\n", " ] \n", "}\n", "os.environ[\"AZURE_OPENAI_DEPLOYMENT_NAME\"] = \"gpt-4o-mini\"\n", "os.environ[\"AZURE_OPENAI_ENDPOINT\"] = \"https://tianyuliu-hua-raredisea-resource.cognitiveservices.azure.com/\"\n", "os.environ[\"AZURE_OPENAI_KEY\"] = \"2pa9h2ZIN1lQepFWwYADlXIKIansa9KPhxMoumeGbRQ08f2uDTXiJQQJ99BKACHYHv6XJ3w3AAAAACOGsQIt\"\n", "os.environ[\"AZURE_OPENAI_API_VERSION\"] = \"2025-01-01-preview\"\n", "llm_config = LiteLLMConfig(model=\"azure/\" + os.getenv(\"AZURE_OPENAI_DEPLOYMENT_NAME\"), # Azure model format\n", " azure_endpoint=os.getenv(\"AZURE_OPENAI_ENDPOINT\"),\n", " azure_key=os.getenv(\"AZURE_OPENAI_KEY\"),\n", " api_version=os.getenv(\"AZURE_OPENAI_API_VERSION\", \"2024-12-01-preview\"), top_p=0.85, temperature=0.2, frequency_penalty=0.0, presence_penalty=0.0)\n", "\n", "executor_llm = LiteLLM(config=llm_config)\n", "optimizer_llm = LiteLLM(config=llm_config)\n", "llm = executor_llm" ] }, { "cell_type": "code", "execution_count": 2, "id": "ad0efa03", "metadata": {}, "outputs": [ { "data": { "text/plain": [ "evoagentx.optimizers.sew_optimizer.SEWOptimizer" ] }, "execution_count": 2, "metadata": {}, "output_type": "execute_result" } ], "source": [ "SEWOptimizer " ] }, { "cell_type": "code", "execution_count": 3, "id": "ad4b2024", "metadata": {}, "outputs": [], "source": [ "# difficult easy " ] }, { "cell_type": "code", "execution_count": 4, "id": "c95059f0", "metadata": {}, "outputs": [], "source": [ "from evoagentx.benchmark import HotPotQA" ] }, { "cell_type": "code", "execution_count": 5, "id": "84efabfa", "metadata": {}, "outputs": [ { "name": "stderr", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:32:09.571\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.benchmark.hitqa\u001b[0m:\u001b[36m_load_data_from_file\u001b[0m:\u001b[36m51\u001b[0m - \u001b[1mloading HotPotQA data from /home/tl688/pitl688/selfevolve/EvoAgentX/examples/antibiotic_pred/ec_train.json ...\u001b[0m\n", "\u001b[32m2026-01-10 17:32:09.662\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.benchmark.hitqa\u001b[0m:\u001b[36m_load_data_from_file\u001b[0m:\u001b[36m51\u001b[0m - \u001b[1mloading HotPotQA data from /home/tl688/pitl688/selfevolve/EvoAgentX/examples/antibiotic_pred/ec_train.json ...\u001b[0m\n", "\u001b[32m2026-01-10 17:32:09.751\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.benchmark.hitqa\u001b[0m:\u001b[36m_load_data_from_file\u001b[0m:\u001b[36m51\u001b[0m - \u001b[1mloading HotPotQA data from /home/tl688/pitl688/selfevolve/EvoAgentX/examples/antibiotic_pred/ec_test.json ...\u001b[0m\n" ] } ], "source": [ "# llm_config = OpenAILLMConfig(model=\"gpt-4o-mini-2024-07-18\", openai_key=OPENAI_API_KEY, top_p=0.85, temperature=0.2, frequency_penalty=0.0, presence_penalty=0.0)\n", "# llm = OpenAILLM(config=llm_config)\n", "llm = executor_llm\n", "\n", "# obtain SEW workflow \n", "sew_graph = SEWWorkFlowGraph.from_dict(hotpotqa_graph_data)\n", "agent_manager = AgentManager()\n", "agent_manager.add_agents_from_workflow(sew_graph, executor_llm.config)\n", "\n", "benchmark = MolQA()\n", "\n", "# obtain Evaluator\n", "evaluator = Evaluator(llm=llm, agent_manager=agent_manager, collate_func=collate_func, num_workers=20, verbose=True)" ] }, { "cell_type": "code", "execution_count": 6, "id": "d2bba683", "metadata": {}, "outputs": [], "source": [ "# import json\n", "# # with open(\"../../MaAS/maas/ext/maas/data/humaneval_train.jsonl\", 'w') as f:\n", "# # json.dump(humaneval._dev_data, f, indent=2) # indent=4 makes the JSON output more readable\n", "\n", "\n", "# # with open(\"../../MaAS/maas/ext/maas/data/humaneval_test.jsonl\", 'w') as f:\n", "# # json.dump(humaneval._test_data, f, indent=2) # indent=4 makes the JSON output more readable\n", "\n", "# with open(\"../../MaAS/maas/ext/maas/data/humaneval_train.jsonl\", 'w') as f:\n", "# for obj in humaneval._dev_data:\n", "# json_line = json.dumps(obj)\n", "# f.write(json_line + '\\n')\n", " \n", "# with open(\"../../MaAS/maas/ext/maas/data/humaneval_test.jsonl\", 'w') as f:\n", "# for obj in humaneval._test_data:\n", "# json_line = json.dumps(obj)\n", "# f.write(json_line + '\\n')\n" ] }, { "cell_type": "code", "execution_count": 7, "id": "8598151b", "metadata": { "scrolled": false }, "outputs": [ { "data": { "text/plain": [ "1" ] }, "execution_count": 7, "metadata": {}, "output_type": "execute_result" } ], "source": [ "len(sew_graph.to_dict()['nodes'])" ] }, { "cell_type": "code", "execution_count": 8, "id": "b1f7fc18", "metadata": {}, "outputs": [ { "data": { "text/plain": [ "0" ] }, "execution_count": 8, "metadata": {}, "output_type": "execute_result" } ], "source": [ "len(sew_graph.edges)" ] }, { "cell_type": "code", "execution_count": 9, "id": "33859fa8", "metadata": {}, "outputs": [ { "data": { "text/plain": [ "[]" ] }, "execution_count": 9, "metadata": {}, "output_type": "execute_result" } ], "source": [ "sew_graph.edges" ] }, { "cell_type": "code", "execution_count": 10, "id": "3c048529", "metadata": { "scrolled": true }, "outputs": [], "source": [ "# obtain SEWOptimizer after having more roles, default\n", "optimizer = SEWOptimizer(\n", " graph=sew_graph, \n", " evaluator=evaluator, \n", " llm=llm, \n", " max_steps=20,\n", " eval_rounds=3, \n", " repr_scheme=\"python\", \n", " optimize_mode=\"all\", \n", " order=\"zero-order\",\n", " max_rounds=20,\n", ")\n", "\n", "# with suppress_logger_info():\n", "# metrics = optimizer.evaluate(dataset=humaneval, eval_mode=\"test\")\n", "# print(\"Evaluation metrics: \", metrics)\n" ] }, { "cell_type": "code", "execution_count": 11, "id": "9cbdd001", "metadata": {}, "outputs": [], "source": [ "import numpy as np\n", "np.random.seed(2024)\n", "out = np.random.choice(benchmark._train_data, size=50, replace=False)\n", "benchmark._train_data = out\n", "benchmark._dev_data = out" ] }, { "cell_type": "code", "execution_count": 12, "id": "8b05058e", "metadata": { "scrolled": true }, "outputs": [ { "name": "stderr", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:32:09.823\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m678\u001b[0m - \u001b[1mOptimizing the SEWWorkFlowGraph workflow with python representation.\u001b[0m\n", "\u001b[32m2026-01-10 17:32:09.823\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m682\u001b[0m - \u001b[1mRun initial evaluation on the original workflow ...\u001b[0m\n", "Evaluating workflow: 2%|▏ | 1/50 [00:02<01:41, 2.06s/it]Task exception was never retrieved\n", "future: exception=RuntimeError('Event loop is closed')>\n", "Traceback (most recent call last):\n", " File \"/gpfs/radev/home/tl688/.conda/envs/evoagentx/lib/python3.11/asyncio/tasks.py\", line 277, in __step\n", " result = coro.send(None)\n", " ^^^^^^^^^^^^^^^\n", " File \"/gpfs/radev/home/tl688/.conda/envs/evoagentx/lib/python3.11/site-packages/litellm/utils.py\", line 873, in _client_async_logging_helper\n", " GLOBAL_LOGGING_WORKER.ensure_initialized_and_enqueue(\n", " File \"/gpfs/radev/home/tl688/.conda/envs/evoagentx/lib/python3.11/site-packages/litellm/litellm_core_utils/logging_worker.py\", line 322, in ensure_initialized_and_enqueue\n", " self.enqueue(async_coroutine)\n", " File \"/gpfs/radev/home/tl688/.conda/envs/evoagentx/lib/python3.11/site-packages/litellm/litellm_core_utils/logging_worker.py\", line 131, in enqueue\n", " self._queue.put_nowait(task)\n", " File \"/gpfs/radev/home/tl688/.conda/envs/evoagentx/lib/python3.11/asyncio/queues.py\", line 147, in put_nowait\n", " self._wakeup_next(self._getters)\n", " File \"/gpfs/radev/home/tl688/.conda/envs/evoagentx/lib/python3.11/asyncio/queues.py\", line 63, in _wakeup_next\n", " waiter.set_result(None)\n", " File \"/gpfs/radev/home/tl688/.conda/envs/evoagentx/lib/python3.11/asyncio/futures.py\", line 263, in set_result\n", " self.__schedule_callbacks()\n", " File \"/gpfs/radev/home/tl688/.conda/envs/evoagentx/lib/python3.11/asyncio/futures.py\", line 173, in __schedule_callbacks\n", " self._loop.call_soon(callback, self, context=ctx)\n", " File \"/gpfs/radev/home/tl688/.conda/envs/evoagentx/lib/python3.11/asyncio/base_events.py\", line 762, in call_soon\n", " self._check_closed()\n", " File \"/gpfs/radev/home/tl688/.conda/envs/evoagentx/lib/python3.11/asyncio/base_events.py\", line 520, in _check_closed\n", " raise RuntimeError('Event loop is closed')\n", "RuntimeError: Event loop is closed\n", "Evaluating workflow: 4%|▍ | 2/50 [00:02<00:45, 1.05it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 10%|█ | 5/50 [00:02<00:14, 3.16it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 22%|██▏ | 11/50 [00:02<00:04, 9.30it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 36%|███▌ | 18/50 [00:03<00:02, 14.59it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 42%|████▏ | 21/50 [00:04<00:05, 4.89it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 50%|█████ | 25/50 [00:05<00:04, 6.23it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 64%|██████▍ | 32/50 [00:05<00:01, 10.70it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 74%|███████▍ | 37/50 [00:05<00:00, 14.07it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 78%|███████▊ | 39/50 [00:06<00:01, 7.80it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 82%|████████▏ | 41/50 [00:06<00:01, 5.61it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 90%|█████████ | 45/50 [00:07<00:00, 6.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 94%|█████████▍| 47/50 [00:07<00:00, 7.36it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 98%|█████████▊| 49/50 [00:08<00:00, 6.57it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 100%|██████████| 50/50 [00:08<00:00, 6.14it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2026-01-10 17:32:18.025\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m685\u001b[0m - \u001b[1mInitial metrics: {'f1': 0.0, 'em': 0.0, 'acc': 0.56}\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:32:18.657\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.014 | Total tokens: 44852 | Current cost: $0.000 | Current tokens: 104\u001b[0m\n", "\u001b[32m2026-01-10 17:32:19.282\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.014 | Total tokens: 44989 | Current cost: $0.000 | Current tokens: 137\u001b[0m\n", "\u001b[32m2026-01-10 17:32:19.284\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m700\u001b[0m - \u001b[33m\u001b[1mError in step 0: can only concatenate str (not \"NoneType\") to str. Skip this step.\u001b[0m\n", "\u001b[32m2026-01-10 17:32:21.018\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.014 | Total tokens: 45241 | Current cost: $0.000 | Current tokens: 252\u001b[0m\n", "\u001b[32m2026-01-10 17:32:21.820\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.014 | Total tokens: 45586 | Current cost: $0.000 | Current tokens: 345\u001b[0m\n", "\u001b[32m2026-01-10 17:32:21.821\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36mparse_workflow_python_repr\u001b[0m:\u001b[36m403\u001b[0m - \u001b[33m\u001b[1mFailed to parse workflow string: 'llm_config'. Return the original workflow.\u001b[0m\n", "\u001b[32m2026-01-10 17:32:21.821\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m700\u001b[0m - \u001b[33m\u001b[1mError in step 1: can only concatenate str (not \"NoneType\") to str. Skip this step.\u001b[0m\n", "\u001b[32m2026-01-10 17:32:22.665\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.014 | Total tokens: 45702 | Current cost: $0.000 | Current tokens: 116\u001b[0m\n", "\u001b[32m2026-01-10 17:32:23.315\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.014 | Total tokens: 45882 | Current cost: $0.000 | Current tokens: 180\u001b[0m\n", "\u001b[32m2026-01-10 17:32:23.317\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m700\u001b[0m - \u001b[33m\u001b[1mError in step 2: can only concatenate str (not \"NoneType\") to str. Skip this step.\u001b[0m\n", "\u001b[32m2026-01-10 17:32:23.929\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.014 | Total tokens: 45975 | Current cost: $0.000 | Current tokens: 93\u001b[0m\n", "\u001b[32m2026-01-10 17:32:24.587\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.014 | Total tokens: 46113 | Current cost: $0.000 | Current tokens: 138\u001b[0m\n", "\u001b[32m2026-01-10 17:32:24.588\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m700\u001b[0m - \u001b[33m\u001b[1mError in step 3: can only concatenate str (not \"NoneType\") to str. Skip this step.\u001b[0m\n", "\u001b[32m2026-01-10 17:32:25.185\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.014 | Total tokens: 46227 | Current cost: $0.000 | Current tokens: 114\u001b[0m\n", "\u001b[32m2026-01-10 17:32:25.782\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.014 | Total tokens: 46365 | Current cost: $0.000 | Current tokens: 138\u001b[0m\n", "\u001b[32m2026-01-10 17:32:25.783\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m700\u001b[0m - \u001b[33m\u001b[1mError in step 4: can only concatenate str (not \"NoneType\") to str. Skip this step.\u001b[0m\n", "\u001b[32m2026-01-10 17:32:26.398\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.014 | Total tokens: 46444 | Current cost: $0.000 | Current tokens: 79\u001b[0m\n", "\u001b[32m2026-01-10 17:32:27.046\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.014 | Total tokens: 46590 | Current cost: $0.000 | Current tokens: 146\u001b[0m\n", "\u001b[32m2026-01-10 17:32:27.047\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36mparse_workflow_python_repr\u001b[0m:\u001b[36m403\u001b[0m - \u001b[33m\u001b[1mFailed to parse workflow string: 'llm_config'. Return the original workflow.\u001b[0m\n", "\u001b[32m2026-01-10 17:32:27.047\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m700\u001b[0m - \u001b[33m\u001b[1mError in step 5: can only concatenate str (not \"NoneType\") to str. Skip this step.\u001b[0m\n", "\u001b[32m2026-01-10 17:32:28.261\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.014 | Total tokens: 46668 | Current cost: $0.000 | Current tokens: 78\u001b[0m\n", "\u001b[32m2026-01-10 17:32:28.900\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.014 | Total tokens: 46809 | Current cost: $0.000 | Current tokens: 141\u001b[0m\n", "\u001b[32m2026-01-10 17:32:28.901\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m700\u001b[0m - \u001b[33m\u001b[1mError in step 6: can only concatenate str (not \"NoneType\") to str. Skip this step.\u001b[0m\n", "\u001b[32m2026-01-10 17:32:29.546\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.014 | Total tokens: 46891 | Current cost: $0.000 | Current tokens: 82\u001b[0m\n", "\u001b[32m2026-01-10 17:32:30.199\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.014 | Total tokens: 47032 | Current cost: $0.000 | Current tokens: 141\u001b[0m\n", "\u001b[32m2026-01-10 17:32:30.200\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m700\u001b[0m - \u001b[33m\u001b[1mError in step 7: can only concatenate str (not \"NoneType\") to str. Skip this step.\u001b[0m\n", "\u001b[32m2026-01-10 17:32:30.906\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.014 | Total tokens: 47129 | Current cost: $0.000 | Current tokens: 97\u001b[0m\n", "\u001b[32m2026-01-10 17:32:31.945\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.014 | Total tokens: 47345 | Current cost: $0.000 | Current tokens: 216\u001b[0m\n", "\u001b[32m2026-01-10 17:32:31.946\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36mparse_workflow_python_repr\u001b[0m:\u001b[36m403\u001b[0m - \u001b[33m\u001b[1mFailed to parse workflow string: 'llm_config'. Return the original workflow.\u001b[0m\n", "\u001b[32m2026-01-10 17:32:31.946\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m700\u001b[0m - \u001b[33m\u001b[1mError in step 8: can only concatenate str (not \"NoneType\") to str. Skip this step.\u001b[0m\n", "\u001b[32m2026-01-10 17:32:32.769\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.015 | Total tokens: 47442 | Current cost: $0.000 | Current tokens: 97\u001b[0m\n", "\u001b[32m2026-01-10 17:32:33.417\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.015 | Total tokens: 47603 | Current cost: $0.000 | Current tokens: 161\u001b[0m\n", "\u001b[32m2026-01-10 17:32:33.418\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m700\u001b[0m - \u001b[33m\u001b[1mError in step 9: can only concatenate str (not \"NoneType\") to str. Skip this step.\u001b[0m\n", "\u001b[32m2026-01-10 17:32:34.022\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.015 | Total tokens: 47670 | Current cost: $0.000 | Current tokens: 67\u001b[0m\n", "\u001b[32m2026-01-10 17:32:34.719\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.015 | Total tokens: 47808 | Current cost: $0.000 | Current tokens: 138\u001b[0m\n", "\u001b[32m2026-01-10 17:32:34.720\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m700\u001b[0m - \u001b[33m\u001b[1mError in step 10: can only concatenate str (not \"NoneType\") to str. Skip this step.\u001b[0m\n", "\u001b[32m2026-01-10 17:32:35.311\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.015 | Total tokens: 47887 | Current cost: $0.000 | Current tokens: 79\u001b[0m\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:32:36.582\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.015 | Total tokens: 48028 | Current cost: $0.000 | Current tokens: 141\u001b[0m\n", "\u001b[32m2026-01-10 17:32:36.584\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m700\u001b[0m - \u001b[33m\u001b[1mError in step 11: can only concatenate str (not \"NoneType\") to str. Skip this step.\u001b[0m\n", "\u001b[32m2026-01-10 17:32:38.113\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.015 | Total tokens: 48229 | Current cost: $0.000 | Current tokens: 201\u001b[0m\n", "\u001b[32m2026-01-10 17:32:38.758\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.015 | Total tokens: 48498 | Current cost: $0.000 | Current tokens: 269\u001b[0m\n", "\u001b[32m2026-01-10 17:32:38.759\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m700\u001b[0m - \u001b[33m\u001b[1mError in step 12: can only concatenate str (not \"NoneType\") to str. Skip this step.\u001b[0m\n", "\u001b[32m2026-01-10 17:32:39.357\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.015 | Total tokens: 48587 | Current cost: $0.000 | Current tokens: 89\u001b[0m\n", "\u001b[32m2026-01-10 17:32:40.103\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.015 | Total tokens: 48758 | Current cost: $0.000 | Current tokens: 171\u001b[0m\n", "\u001b[32m2026-01-10 17:32:40.104\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36mparse_workflow_python_repr\u001b[0m:\u001b[36m403\u001b[0m - \u001b[33m\u001b[1mFailed to parse workflow string: 'llm_config'. Return the original workflow.\u001b[0m\n", "\u001b[32m2026-01-10 17:32:40.105\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m700\u001b[0m - \u001b[33m\u001b[1mError in step 13: can only concatenate str (not \"NoneType\") to str. Skip this step.\u001b[0m\n", "\u001b[32m2026-01-10 17:32:40.917\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.015 | Total tokens: 48880 | Current cost: $0.000 | Current tokens: 122\u001b[0m\n", "\u001b[32m2026-01-10 17:32:41.664\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.015 | Total tokens: 49065 | Current cost: $0.000 | Current tokens: 185\u001b[0m\n", "\u001b[32m2026-01-10 17:32:41.665\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36mparse_workflow_python_repr\u001b[0m:\u001b[36m403\u001b[0m - \u001b[33m\u001b[1mFailed to parse workflow string: 'llm_config'. Return the original workflow.\u001b[0m\n", "\u001b[32m2026-01-10 17:32:41.665\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m700\u001b[0m - \u001b[33m\u001b[1mError in step 14: can only concatenate str (not \"NoneType\") to str. Skip this step.\u001b[0m\n", "\u001b[32m2026-01-10 17:32:43.439\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.015 | Total tokens: 49322 | Current cost: $0.000 | Current tokens: 257\u001b[0m\n", "\u001b[32m2026-01-10 17:32:43.892\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.015 | Total tokens: 49633 | Current cost: $0.000 | Current tokens: 311\u001b[0m\n", "\u001b[32m2026-01-10 17:32:43.894\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m700\u001b[0m - \u001b[33m\u001b[1mError in step 15: can only concatenate str (not \"NoneType\") to str. Skip this step.\u001b[0m\n", "\u001b[32m2026-01-10 17:32:44.466\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.015 | Total tokens: 49744 | Current cost: $0.000 | Current tokens: 111\u001b[0m\n", "\u001b[32m2026-01-10 17:32:45.222\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.015 | Total tokens: 49882 | Current cost: $0.000 | Current tokens: 138\u001b[0m\n", "\u001b[32m2026-01-10 17:32:45.223\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m700\u001b[0m - \u001b[33m\u001b[1mError in step 16: can only concatenate str (not \"NoneType\") to str. Skip this step.\u001b[0m\n", "\u001b[32m2026-01-10 17:32:45.655\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.015 | Total tokens: 49954 | Current cost: $0.000 | Current tokens: 72\u001b[0m\n", "\u001b[32m2026-01-10 17:32:46.290\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.015 | Total tokens: 50081 | Current cost: $0.000 | Current tokens: 127\u001b[0m\n", "\u001b[32m2026-01-10 17:32:46.292\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m700\u001b[0m - \u001b[33m\u001b[1mError in step 17: can only concatenate str (not \"NoneType\") to str. Skip this step.\u001b[0m\n", "\u001b[32m2026-01-10 17:32:46.742\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.015 | Total tokens: 50149 | Current cost: $0.000 | Current tokens: 68\u001b[0m\n", "\u001b[32m2026-01-10 17:32:47.433\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.015 | Total tokens: 50276 | Current cost: $0.000 | Current tokens: 127\u001b[0m\n", "\u001b[32m2026-01-10 17:32:47.434\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m700\u001b[0m - \u001b[33m\u001b[1mError in step 18: can only concatenate str (not \"NoneType\") to str. Skip this step.\u001b[0m\n", "\u001b[32m2026-01-10 17:32:48.020\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.015 | Total tokens: 50353 | Current cost: $0.000 | Current tokens: 77\u001b[0m\n", "\u001b[32m2026-01-10 17:32:48.616\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.models.model_utils\u001b[0m:\u001b[36mupdate_cost\u001b[0m:\u001b[36m87\u001b[0m - \u001b[1mTotal cost: $0.015 | Total tokens: 50492 | Current cost: $0.000 | Current tokens: 139\u001b[0m\n", "\u001b[32m2026-01-10 17:32:48.617\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m700\u001b[0m - \u001b[33m\u001b[1mError in step 19: can only concatenate str (not \"NoneType\") to str. Skip this step.\u001b[0m\n", "\u001b[32m2026-01-10 17:32:48.617\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m707\u001b[0m - \u001b[1mReach the maximum number of steps 20. Stop the optimization.\u001b[0m\n", "\u001b[32m2026-01-10 17:32:48.618\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36moptimize\u001b[0m:\u001b[36m710\u001b[0m - \u001b[1mRestore the best graph from the snapshot ...\u001b[0m\n", "\u001b[32m2026-01-10 17:32:48.618\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.optimizers.sew_optimizer\u001b[0m:\u001b[36mrestore_best_graph\u001b[0m:\u001b[36m814\u001b[0m - \u001b[1mRestore the best graph from snapshot with metrics {'f1': 0.0, 'em': 0.0, 'acc': 0.56} ...\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 0%| | 1/3000 [00:01<1:27:45, 1.76s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 0%| | 5/3000 [00:02<15:39, 3.19it/s] " ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 0%| | 8/3000 [00:02<08:58, 5.55it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 0%| | 12/3000 [00:02<06:20, 7.85it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 0%| | 14/3000 [00:02<05:15, 9.46it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 1%| | 16/3000 [00:03<05:45, 8.65it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 1%| | 18/3000 [00:03<06:14, 7.96it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 1%| | 20/3000 [00:03<08:19, 5.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 1%| | 22/3000 [00:04<07:33, 6.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 1%| | 23/3000 [00:04<08:34, 5.79it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 1%| | 29/3000 [00:04<04:56, 10.02it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 1%| | 33/3000 [00:05<05:43, 8.65it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 1%| | 37/3000 [00:05<05:30, 8.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 1%|▏ | 39/3000 [00:06<06:24, 7.69it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 1%|▏ | 41/3000 [00:06<06:41, 7.38it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 1%|▏ | 43/3000 [00:06<07:30, 6.57it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 2%|▏ | 46/3000 [00:07<05:37, 8.75it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 2%|▏ | 48/3000 [00:07<05:07, 9.61it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 2%|▏ | 50/3000 [00:07<05:13, 9.41it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 2%|▏ | 53/3000 [00:07<04:33, 10.79it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 2%|▏ | 56/3000 [00:08<05:55, 8.27it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 2%|▏ | 58/3000 [00:08<09:01, 5.43it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 2%|▏ | 62/3000 [00:09<06:35, 7.44it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 2%|▏ | 64/3000 [00:09<05:53, 8.29it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 2%|▏ | 66/3000 [00:09<05:13, 9.35it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 2%|▏ | 70/3000 [00:10<04:28, 10.92it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 2%|▏ | 74/3000 [00:10<06:24, 7.60it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 3%|▎ | 79/3000 [00:11<04:47, 10.15it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 3%|▎ | 81/3000 [00:11<07:47, 6.25it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 3%|▎ | 83/3000 [00:12<07:02, 6.90it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 3%|▎ | 87/3000 [00:12<06:39, 7.29it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 3%|▎ | 91/3000 [00:12<04:51, 9.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 3%|▎ | 93/3000 [00:13<06:13, 7.78it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 3%|▎ | 95/3000 [00:13<06:54, 7.00it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 3%|▎ | 97/3000 [00:13<06:40, 7.24it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 3%|▎ | 99/3000 [00:14<07:18, 6.61it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 3%|▎ | 102/3000 [00:14<07:24, 6.53it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 4%|▎ | 105/3000 [00:14<05:02, 9.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 4%|▎ | 107/3000 [00:15<04:46, 10.09it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 4%|▎ | 111/3000 [00:15<06:13, 7.74it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 4%|▍ | 113/3000 [00:16<06:07, 7.85it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 4%|▍ | 115/3000 [00:16<06:06, 7.88it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 4%|▍ | 119/3000 [00:16<05:09, 9.31it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 4%|▍ | 121/3000 [00:16<05:13, 9.18it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 4%|▍ | 125/3000 [00:17<04:07, 11.61it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 4%|▍ | 127/3000 [00:17<06:35, 7.26it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 4%|▍ | 129/3000 [00:18<07:31, 6.36it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 4%|▍ | 130/3000 [00:18<07:55, 6.04it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 4%|▍ | 132/3000 [00:18<07:57, 6.00it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097694.871948225)])']\n", "connector: \n", "Evaluating workflow: 4%|▍ | 133/3000 [00:18<09:40, 4.94it/s]Unclosed connector\n", "connections: ['deque([(, 2097695.159673983)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097694.905455873)])']\n", "connector: \n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 5%|▍ | 142/3000 [00:19<03:58, 11.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 5%|▍ | 144/3000 [00:19<04:49, 9.88it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 5%|▍ | 146/3000 [00:20<05:45, 8.27it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 5%|▍ | 148/3000 [00:20<05:35, 8.50it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 5%|▌ | 151/3000 [00:20<06:06, 7.78it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 5%|▌ | 156/3000 [00:21<04:55, 9.61it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 5%|▌ | 158/3000 [00:21<06:15, 7.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 5%|▌ | 160/3000 [00:22<07:25, 6.38it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 5%|▌ | 164/3000 [00:22<05:26, 8.68it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 6%|▌ | 169/3000 [00:22<04:14, 11.10it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 6%|▌ | 172/3000 [00:22<04:09, 11.32it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 6%|▌ | 174/3000 [00:23<04:14, 11.12it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 6%|▌ | 176/3000 [00:23<06:46, 6.94it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 6%|▌ | 179/3000 [00:24<08:15, 5.69it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 6%|▌ | 184/3000 [00:24<04:28, 10.50it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 6%|▌ | 186/3000 [00:24<04:50, 9.69it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 6%|▋ | 188/3000 [00:25<06:49, 6.86it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 6%|▋ | 190/3000 [00:25<07:21, 6.36it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 6%|▋ | 194/3000 [00:26<05:59, 7.81it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 7%|▋ | 196/3000 [00:26<06:07, 7.62it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 7%|▋ | 197/3000 [00:26<09:01, 5.18it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 7%|▋ | 198/3000 [00:27<09:10, 5.09it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 7%|▋ | 203/3000 [00:27<04:59, 9.34it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 7%|▋ | 208/3000 [00:27<04:14, 10.99it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metricsmetrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", " {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 7%|▋ | 211/3000 [00:28<06:11, 7.50it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 7%|▋ | 213/3000 [00:28<07:40, 6.05it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 7%|▋ | 215/3000 [00:29<06:30, 7.14it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 7%|▋ | 221/3000 [00:29<04:33, 10.17it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 7%|▋ | 223/3000 [00:29<05:08, 8.99it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 8%|▊ | 227/3000 [00:30<04:45, 9.70it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 8%|▊ | 231/3000 [00:30<05:09, 8.96it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 8%|▊ | 233/3000 [00:31<08:39, 5.32it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 8%|▊ | 238/3000 [00:31<05:58, 7.71it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 8%|▊ | 243/3000 [00:32<04:01, 11.44it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 8%|▊ | 248/3000 [00:32<04:35, 9.99it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 8%|▊ | 252/3000 [00:33<05:39, 8.09it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 9%|▊ | 257/3000 [00:34<05:55, 7.71it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 9%|▊ | 259/3000 [00:34<05:09, 8.86it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 9%|▊ | 262/3000 [00:34<05:32, 8.23it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 9%|▉ | 264/3000 [00:35<06:46, 6.73it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:27.830\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0148671', 'INHIB_AVE': 11.18, 'SMILES': 'c1c(C(CC(=O)N2)c(snc3c4cccc(F)c4)c23)cc(c5c1)OCO5', 'HIT': 0.0, 'question': 'Is c1c(C(CC(=O)N2)c(snc3c4cccc(F)c4)c23)cc(c5c1)OCO5 capable of inhibiting the growth or survival of E. coli?', 'answer': 'No', '_id': 14135}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 9%|▉ | 266/3000 [00:39<27:55, 1.63it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:33:31.060\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0277736', 'INHIB_AVE': -5.91, 'SMILES': 'c12c(c3c(cccc3)OC1=O)nc(c(cc4)ccc4C(C)C)nc2NC(C)=O', 'HIT': 0.0, 'question': 'Can c12c(c3c(cccc3)OC1=O)nc(c(cc4)ccc4C(C)C)nc2NC(C)=O effectively kill or suppress Escherichia coli cells?', 'answer': 'No', '_id': 73552}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 9%|▉ | 267/3000 [00:42<46:44, 1.03s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:33:31.283\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0225640', 'INHIB_AVE': 2.95, 'SMILES': 'C1(=N2)N(C=CN=C1Oc(cc3)ccc3C)C(N2CC(=O)Nc4ccc(cc4F)Br)=O', 'HIT': 0.0, 'question': 'Does the proposed molecule C1(=N2)N(C=CN=C1Oc(cc3)ccc3C)C(N2CC(=O)Nc4ccc(cc4F)Br)=O exhibit antibacterial activity against Escherichia coli?', 'answer': 'No', '_id': 44869}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 9%|▉ | 268/3000 [00:42<39:58, 1.14it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:33:31.400\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0240505', 'INHIB_AVE': 22.96, 'SMILES': 'C(C#N)(C1c2scc(Br)c2)=C(N)N(C(CCCC3=O)=C13)c4ccccc4', 'HIT': 0.0, 'question': 'Is C(C#N)(C1c2scc(Br)c2)=C(N)N(C(CCCC3=O)=C13)c4ccccc4 capable of inhibiting the growth or survival of E. coli?', 'answer': 'No', '_id': 1627}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 9%|▉ | 269/3000 [00:42<32:45, 1.39it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:33:31.418\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0331221', 'INHIB_AVE': 0.72, 'SMILES': 'n1(CC=C)c(SCC(=O)Nc(cc2)ccc2C(=O)OC(C)C)nnc1c3ccco3', 'HIT': 0.0, 'question': 'Can exposure to n1(CC=C)c(SCC(=O)Nc(cc2)ccc2C(=O)OC(C)C)nnc1c3ccco3 lead to reduced viability of E. coli?', 'answer': 'No', '_id': 54634}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:33:31.429\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0217226', 'INHIB_AVE': 8.1, 'SMILES': 'C(F)(F)(COc1ccccc1NC(=O)c2ccc(cc2)C)C(F)F', 'HIT': 0.0, 'question': 'Can C(F)(F)(COc1ccccc1NC(=O)c2ccc(cc2)C)C(F)F effectively kill or suppress Escherichia coli cells?', 'answer': 'No', '_id': 22842}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:33:31.484\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0234228', 'INHIB_AVE': 10.17, 'SMILES': 'c1(c(O)c(\\\\C=N\\\\c2ccc(cc2Cl)Cl)c(c(Cl)c1C)C)[N+](=O)[O-]', 'HIT': 0.0, 'question': 'Does c1(c(O)c(\\\\C=N\\\\c2ccc(cc2Cl)Cl)c(c(Cl)c1C)C)[N+](=O)[O-] show antimicrobial efficacy specifically against Escherichia coli strains?', 'answer': 'No', '_id': 16564}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:33:31.643\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0326254', 'INHIB_AVE': 5.32, 'SMILES': 'N1(c2ccc(cc2)Cl)C(=O)CC(N(C3CCCC3)C(C)=O)C1=O', 'HIT': 0.0, 'question': 'Is N1(c2ccc(cc2)Cl)C(=O)CC(N(C3CCCC3)C(C)=O)C1=O effective in eliminating E. coli under experimental conditions?', 'answer': 'No', '_id': 34235}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 9%|▉ | 273/3000 [00:42<16:26, 2.76it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:33:31.683\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0161257', 'INHIB_AVE': -7.52, 'SMILES': 'C1(=O)N(\\\\N=C\\\\c(cccc2[N+](=O)[O-])c2)C(C)=CC(C)=C1C(N)=O', 'HIT': 0.0, 'question': 'Is C1(=O)N(\\\\N=C\\\\c(cccc2[N+](=O)[O-])c2)C(C)=CC(C)=C1C(N)=O capable of inhibiting the growth or survival of E. coli?', 'answer': 'No', '_id': 76097}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:33:31.771\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0303166', 'INHIB_AVE': 0.06, 'SMILES': 'N1(c2ccc(cc2)OC)C(=O)NC(\\\\C(=C/c3cc(Br)ccc3OCCOc(cc(CC)cc4C)c4)\\\\C1=O)=O', 'HIT': 0.0, 'question': 'Does N1(c2ccc(cc2)OC)C(=O)NC(\\\\C(=C/c3cc(Br)ccc3OCCOc(cc(CC)cc4C)c4)\\\\C1=O)=O show antimicrobial efficacy specifically against Escherichia coli strains?', 'answer': 'No', '_id': 57221}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 9%|▉ | 275/3000 [00:43<12:38, 3.59it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:33:31.781\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0277496', 'INHIB_AVE': -1.48, 'SMILES': 'C1(C(OCCOC)=O)=C(C)NC(NC1c(cc2)cc(c2OC(F)F)OC)=O', 'HIT': 0.0, 'question': 'Can exposure to C1(C(OCCOC)=O)=C(C)NC(NC1c(cc2)cc(c2OC(F)F)OC)=O lead to reduced viability of E. coli?', 'answer': 'No', '_id': 62675}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:33:31.929\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0162954', 'INHIB_AVE': -7.52, 'SMILES': 'S(=O)(=O)(CCC(C)=O)c1ccc(cc1)Cc(cc2)ccc2S(=O)(=O)CCC(C)=O', 'HIT': 0.0, 'question': 'Is Escherichia coli susceptible to treatment with the compound S(=O)(=O)(CCC(C)=O)c1ccc(cc1)Cc(cc2)ccc2S(=O)(=O)CCC(C)=O?', 'answer': 'No', '_id': 76090}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 9%|▉ | 277/3000 [00:43<10:03, 4.51it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:33:31.980\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0222713', 'INHIB_AVE': 16.99, 'SMILES': 'N1(CNc2ccccc2C(OC)=O)C(=O)\\\\C(\\\\SC1=O)=C/c(cc3OC)cc(c3OC)OC', 'HIT': 0.0, 'question': 'Does N1(CNc2ccccc2C(OC)=O)C(=O)\\\\C(\\\\SC1=O)=C/c(cc3OC)cc(c3OC)OC show antimicrobial efficacy specifically against Escherichia coli strains?', 'answer': 'No', '_id': 5300}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:33:32.084\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0326010', 'INHIB_AVE': -1.13, 'SMILES': 'S(=O)(=O)(Nc(cc1)ccc1OCC)c2ccc(c(C)c2Cl)Cl', 'HIT': 0.0, 'question': 'Is Escherichia coli susceptible to treatment with the compound S(=O)(=O)(Nc(cc1)ccc1OCC)c2ccc(c(C)c2Cl)Cl?', 'answer': 'No', '_id': 61539}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 9%|▉ | 279/3000 [00:43<08:09, 5.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:33:32.149\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0156154', 'INHIB_AVE': 11.42, 'SMILES': 'O=C(C(C=C1)=CC=C1Cl)N2C3=C(SC4=C2C=CC=C4)C=CC(Cl)=C3', 'HIT': 0.0, 'question': 'Is O=C(C(C=C1)=CC=C1Cl)N2C3=C(SC4=C2C=CC=C4)C=CC(Cl)=C3 capable of inhibiting the growth or survival of E. coli?', 'answer': 'No', '_id': 13619}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:33:33.003\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0236947', 'INHIB_AVE': 3.12, 'SMILES': 'O=C(C1=CC=C(OC)C=C1CC/2)C2=CC3=CC=CN=C3Br', 'HIT': 0.0, 'question': 'Is O=C(C1=CC=C(OC)C=C1CC/2)C2=CC3=CC=CN=C3Br capable of inhibiting the growth or survival of E. coli?', 'answer': 'No', '_id': 44066}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 7 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 9%|▉ | 281/3000 [00:44<11:51, 3.82it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:33:33.069\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0233044', 'INHIB_AVE': -1.93, 'SMILES': 'N1(c2ccccc2c3ccccc3)C(=O)c(c4C1=O)ccc(c4)C(=O)Nc5ccc(cc5C(O)=O)Br', 'HIT': 0.0, 'question': 'Is Escherichia coli susceptible to treatment with the compound N1(c2ccccc2c3ccccc3)C(=O)c(c4C1=O)ccc(c4)C(=O)Nc5ccc(cc5C(O)=O)Br?', 'answer': 'No', '_id': 64166}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 7 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:33:33.180\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0287710', 'INHIB_AVE': -0.2, 'SMILES': 'N1(\\\\N=C(/CCCCC2)\\\\C2)C(=O)c3c(NC1c4ccccc4Cl)cccc3', 'HIT': 0.0, 'question': 'Does treatment with N1(\\\\N=C(/CCCCC2)\\\\C2)C(=O)c3c(NC1c4ccccc4Cl)cccc3 result in bacteriostatic or bactericidal effects on E. coli?', 'answer': 'No', '_id': 58255}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 7 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 9%|▉ | 283/3000 [00:44<09:32, 4.75it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:33:33.348\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0166323', 'INHIB_AVE': 14.81, 'SMILES': 'C(OC)(COc(cc1)ccc1C=O)=O', 'HIT': 0.0, 'question': 'Is Escherichia coli susceptible to treatment with the compound C(OC)(COc(cc1)ccc1C=O)=O?', 'answer': 'No', '_id': 7841}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 7 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 9%|▉ | 284/3000 [00:44<09:12, 4.92it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:33:35.088\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0296119', 'INHIB_AVE': 0.49, 'SMILES': 'c1(cc(F)c(cc1Cl)F)C(=O)N(CC2)CCN2c(cc3)c4c(c3[N+](=O)[O-])cccn4', 'HIT': 0.0, 'question': 'Does c1(cc(F)c(cc1Cl)F)C(=O)N(CC2)CCN2c(cc3)c4c(c3[N+](=O)[O-])cccn4 demonstrate bactericidal effects against E. coli in vitro?', 'answer': 'No', '_id': 55543}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 10%|▉ | 285/3000 [00:46<23:02, 1.96it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:33:35.503\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0161314', 'INHIB_AVE': 2.25, 'SMILES': 'C(C#N)(C1c(ccc(c2[N+](=O)[O-])Cl)c2)=C(N)SC(N)=C1C#N', 'HIT': 0.0, 'question': 'Can C(C#N)(C1c(ccc(c2[N+](=O)[O-])Cl)c2)=C(N)SC(N)=C1C#N effectively kill or suppress Escherichia coli cells?', 'answer': 'No', '_id': 47986}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 10%|▉ | 286/3000 [00:46<22:05, 2.05it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:33:38.281\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 10%|▉ | 287/3000 [00:49<46:57, 1.04s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:38.535\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 10%|▉ | 288/3000 [00:49<37:52, 1.19it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:38.555\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:33:38.560\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:38.621\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:38.779\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 10%|▉ | 292/3000 [00:50<17:00, 2.65it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:39.060\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 10%|▉ | 293/3000 [00:50<16:15, 2.78it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:39.113\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:39.225\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 10%|▉ | 295/3000 [00:50<12:04, 3.74it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:39.319\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:39.479\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 10%|▉ | 297/3000 [00:50<10:00, 4.50it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:39.644\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 10%|▉ | 298/3000 [00:50<09:31, 4.73it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:40.417\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 10%|▉ | 299/3000 [00:51<14:51, 3.03it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:40.665\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 10%|█ | 300/3000 [00:51<13:59, 3.22it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:40.698\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:40.702\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:40.751\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:41.138\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 10%|█ | 304/3000 [00:52<09:02, 4.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:41.884\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0233758', 'INHIB_AVE': 8.35, 'SMILES': 'N1(CCC[C@@H](C1)c(ccn2)o2)C(=O)[C@@H]3CCCN3.Cl', 'HIT': 0.0, 'question': 'Can N1(CCC[C@@H](C1)c(ccn2)o2)C(=O)[C@@H]3CCCN3.Cl effectively kill or suppress Escherichia coli cells?', 'answer': 'No', '_id': 21942}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 10%|█ | 305/3000 [00:53<13:10, 3.41it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:33:41.959\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0256464', 'INHIB_AVE': -9.4, 'SMILES': 'c1(cnc(C)n1CC(NN)=O)[N+](=O)[O-]', 'HIT': 0.0, 'question': 'Can c1(cnc(C)n1CC(NN)=O)[N+](=O)[O-] act as an antibacterial agent targeting Escherichia coli?', 'answer': 'No', '_id': 78235}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:33:42.107\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0215412', 'INHIB_AVE': 2.96, 'SMILES': 'c(C#N)(c1c2ccccc2)c(SCc(ccc(c3Cl)Cl)c3)nc(N)c1C#N', 'HIT': 0.0, 'question': 'Does c(C#N)(c1c2ccccc2)c(SCc(ccc(c3Cl)Cl)c3)nc(N)c1C#N show antimicrobial efficacy specifically against Escherichia coli strains?', 'answer': 'No', '_id': 44825}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 10%|█ | 307/3000 [00:53<10:30, 4.27it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:33:42.247\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0232115', 'INHIB_AVE': 5.87, 'SMILES': 'C(=O)(c1ccc(cc1)Cl)NCC(=O)N\\\\N=C\\\\c(ccc(c23)OCO2)c3', 'HIT': 0.0, 'question': 'Is Escherichia coli susceptible to treatment with the compound C(=O)(c1ccc(cc1)Cl)NCC(=O)N\\\\N=C\\\\c(ccc(c23)OCO2)c3?', 'answer': 'No', '_id': 31806}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 10%|█ | 308/3000 [00:53<09:42, 4.62it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:33:43.613\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 10%|█ | 309/3000 [00:54<20:39, 2.17it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:43.774\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 10%|█ | 310/3000 [00:55<17:31, 2.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:44.362\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 10%|█ | 311/3000 [00:55<19:44, 2.27it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:44.450\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:44.842\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 10%|█ | 313/3000 [00:56<15:59, 2.80it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:44.849\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:44.880\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:44.892\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:45.022\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 11%|█ | 317/3000 [00:56<08:23, 5.33it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:45.122\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:45.248\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 11%|█ | 319/3000 [00:56<07:26, 6.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:45.271\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:45.382\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:33:45.383\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0278342', 'INHIB_AVE': 0.69, 'SMILES': 'N1(c2cccc(Cl)c2Cl)C(=O)\\\\C(=C\\\\c3ccc(C)o3)\\\\C(NC1=S)=O', 'HIT': 0.0, 'question': 'Does treatment with N1(c2cccc(Cl)c2Cl)C(=O)\\\\C(=C\\\\c3ccc(C)o3)\\\\C(NC1=S)=O result in bacteriostatic or bactericidal effects on E. coli?', 'answer': 'No', '_id': 54760}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 11%|█ | 321/3000 [00:56<06:09, 7.24it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:33:46.501\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 11%|█ | 323/3000 [00:57<11:36, 3.84it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:46.638\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 11%|█ | 324/3000 [00:57<10:40, 4.18it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:47.021\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 11%|█ | 325/3000 [00:58<11:55, 3.74it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:47.153\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 11%|█ | 326/3000 [00:58<10:35, 4.21it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:47.259\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 11%|█ | 327/3000 [00:58<09:11, 4.85it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:47.382\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 11%|█ | 328/3000 [00:58<08:13, 5.42it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:47.401\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:47.515\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 11%|█ | 330/3000 [00:58<06:00, 7.41it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:47.522\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:33:47.622\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0155675', 'INHIB_AVE': -3.55, 'SMILES': '[O-][N+](C1=CC=CC=C1OCC(N/N=C/C2=CC=C(C3=CC=CC([N+]([O-])=O)=C3)O2)=O)=O', 'HIT': 0.0, 'question': 'Does [O-][N+](C1=CC=CC=C1OCC(N/N=C/C2=CC=C(C3=CC=CC([N+]([O-])=O)=C3)O2)=O)=O show antimicrobial efficacy specifically against Escherichia coli strains?', 'answer': 'No', '_id': 68660}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 11%|█ | 332/3000 [00:58<04:38, 9.57it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:33:47.838\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0232143', 'INHIB_AVE': 5.88, 'SMILES': 'C(=O)(NCC(=O)N\\\\N=C(/C)\\\\c1ccc(cc1)C)c2ccco2', 'HIT': 0.0, 'question': 'Is C(=O)(NCC(=O)N\\\\N=C(/C)\\\\c1ccc(cc1)C)c2ccco2 effective in eliminating E. coli under experimental conditions?', 'answer': 'No', '_id': 31731}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 11%|█ | 336/3000 [01:00<12:10, 3.65it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2026-01-10 17:33:49.636\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 11%|█▏ | 338/3000 [01:00<09:08, 4.85it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2026-01-10 17:33:49.765\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 11%|█▏ | 340/3000 [01:01<07:09, 6.19it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 11%|█▏ | 342/3000 [01:02<13:09, 3.37it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 11%|█▏ | 343/3000 [01:02<12:51, 3.45it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 12%|█▏ | 347/3000 [01:03<09:11, 4.81it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 12%|█▏ | 352/3000 [01:04<08:07, 5.44it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 12%|█▏ | 356/3000 [01:04<05:32, 7.96it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 12%|█▏ | 358/3000 [01:04<05:36, 7.84it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 12%|█▏ | 362/3000 [01:04<05:02, 8.73it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 12%|█▏ | 364/3000 [01:05<04:29, 9.77it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 12%|█▏ | 368/3000 [01:05<06:40, 6.58it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 12%|█▏ | 370/3000 [01:06<06:23, 6.86it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 12%|█▏ | 372/3000 [01:06<07:17, 6.00it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 12%|█▎ | 375/3000 [01:06<05:43, 7.64it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 13%|█▎ | 378/3000 [01:07<05:07, 8.53it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 13%|█▎ | 380/3000 [01:07<05:22, 8.12it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 13%|█▎ | 384/3000 [01:07<04:59, 8.75it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 13%|█▎ | 385/3000 [01:08<05:15, 8.29it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 13%|█▎ | 388/3000 [01:08<05:40, 7.67it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 13%|█▎ | 390/3000 [01:08<05:46, 7.54it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 13%|█▎ | 393/3000 [01:09<06:11, 7.02it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 13%|█▎ | 398/3000 [01:09<03:26, 12.58it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 13%|█▎ | 400/3000 [01:09<03:58, 10.91it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 13%|█▎ | 402/3000 [01:09<04:20, 9.96it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 14%|█▎ | 405/3000 [01:10<08:04, 5.35it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 14%|█▎ | 411/3000 [01:11<03:49, 11.28it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 14%|█▍ | 413/3000 [01:11<03:57, 10.90it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 14%|█▍ | 415/3000 [01:11<05:19, 8.09it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 14%|█▍ | 418/3000 [01:11<04:02, 10.64it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 14%|█▍ | 420/3000 [01:12<03:42, 11.59it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 14%|█▍ | 422/3000 [01:12<06:32, 6.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 14%|█▍ | 424/3000 [01:13<07:39, 5.61it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 14%|█▍ | 428/3000 [01:13<05:53, 7.27it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 14%|█▍ | 433/3000 [01:13<03:55, 10.92it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 15%|█▍ | 436/3000 [01:14<04:09, 10.29it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 15%|█▍ | 438/3000 [01:14<06:03, 7.05it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 15%|█▍ | 440/3000 [01:15<06:57, 6.13it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 15%|█▍ | 442/3000 [01:15<06:15, 6.80it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 15%|█▍ | 443/3000 [01:15<06:06, 6.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 15%|█▍ | 447/3000 [01:15<05:05, 8.35it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 15%|█▍ | 449/3000 [01:16<04:39, 9.13it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 15%|█▌ | 451/3000 [01:16<04:32, 9.37it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 15%|█▌ | 453/3000 [01:16<05:11, 8.18it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 15%|█▌ | 456/3000 [01:17<06:52, 6.16it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 15%|█▌ | 461/3000 [01:17<05:28, 7.73it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 15%|█▌ | 464/3000 [01:18<04:27, 9.46it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 16%|█▌ | 467/3000 [01:18<03:47, 11.12it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 16%|█▌ | 469/3000 [01:18<04:01, 10.48it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 16%|█▌ | 472/3000 [01:18<04:45, 8.86it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 16%|█▌ | 474/3000 [01:19<05:51, 7.18it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 16%|█▌ | 477/3000 [01:19<05:34, 7.55it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 16%|█▌ | 479/3000 [01:19<04:35, 9.15it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 16%|█▌ | 482/3000 [01:20<07:11, 5.84it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 16%|█▌ | 484/3000 [01:20<05:22, 7.80it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 16%|█▋ | 488/3000 [01:21<04:02, 10.35it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 16%|█▋ | 490/3000 [01:21<04:06, 10.17it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 16%|█▋ | 492/3000 [01:21<06:27, 6.47it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 16%|█▋ | 493/3000 [01:22<10:12, 4.09it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 17%|█▋ | 497/3000 [01:23<06:55, 6.02it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 17%|█▋ | 500/3000 [01:23<06:38, 6.28it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 17%|█▋ | 506/3000 [01:23<03:03, 13.57it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 17%|█▋ | 509/3000 [01:24<03:51, 10.77it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 17%|█▋ | 511/3000 [01:24<03:36, 11.51it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 17%|█▋ | 513/3000 [01:24<06:15, 6.63it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 17%|█▋ | 517/3000 [01:25<06:12, 6.67it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 17%|█▋ | 520/3000 [01:26<06:31, 6.33it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 17%|█▋ | 524/3000 [01:26<04:37, 8.92it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 18%|█▊ | 526/3000 [01:26<04:00, 10.27it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 18%|█▊ | 528/3000 [01:26<04:10, 9.86it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 18%|█▊ | 532/3000 [01:27<05:39, 7.26it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 18%|█▊ | 534/3000 [01:27<05:28, 7.51it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 18%|█▊ | 539/3000 [01:28<04:53, 8.37it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 18%|█▊ | 544/3000 [01:28<04:13, 9.69it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 18%|█▊ | 546/3000 [01:29<04:23, 9.31it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 18%|█▊ | 548/3000 [01:29<04:46, 8.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 18%|█▊ | 550/3000 [01:29<04:41, 8.69it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 18%|█▊ | 552/3000 [01:30<06:13, 6.55it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 18%|█▊ | 553/3000 [01:30<07:26, 5.48it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 18%|█▊ | 555/3000 [01:30<06:47, 6.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 19%|█▊ | 557/3000 [01:31<07:04, 5.76it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 19%|█▊ | 561/3000 [01:31<04:28, 9.08it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 19%|█▉ | 563/3000 [01:31<04:17, 9.46it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 19%|█▉ | 568/3000 [01:31<03:27, 11.70it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 19%|█▉ | 570/3000 [01:32<05:51, 6.91it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 19%|█▉ | 573/3000 [01:33<07:12, 5.61it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 19%|█▉ | 578/3000 [01:33<04:02, 9.99it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 19%|█▉ | 583/3000 [01:33<03:36, 11.15it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 20%|█▉ | 586/3000 [01:33<02:52, 14.00it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 20%|█▉ | 588/3000 [01:34<04:58, 8.09it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 20%|█▉ | 593/3000 [01:35<05:53, 6.82it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 20%|█▉ | 595/3000 [01:35<05:44, 6.99it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 20%|█▉ | 598/3000 [01:36<05:27, 7.34it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Evaluating workflow: 20%|██ | 601/3000 [01:36<05:35, 7.16it/s]Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097695.199769279)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097695.473596559)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097695.352257943)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097695.751721669)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097695.058871246)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097694.457078643)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097695.391940769)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097695.076186155)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097695.091972919)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097695.036603625)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097695.030669327)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097695.43318479)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097695.096205965)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097695.445775692)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097695.282213716)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097695.839455878)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097697.484573281)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097698.006550075)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097697.064096366)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097697.508680813)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097699.652659826)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097699.895076367)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097699.800058514)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097699.825371565)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097700.288237764)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097700.533241261)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097699.277893031)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097700.398903814)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097700.013124506)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097699.707131608)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097733.945262679)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097733.490579246)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097733.080230055)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097734.642944896)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097733.276366821)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097734.346525543)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097733.979476669)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097735.404317318)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097733.304579214)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097733.954019606)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097734.071394677)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097734.411640277)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097734.327106703)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097733.507360054)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097733.615553136)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097733.808525755)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097736.002233049)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097736.116433316)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097736.08346601)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097736.605640692)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097735.707893423)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097735.928419455)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097736.003416848)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097738.121596471)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097738.251584906)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097738.54080827)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097741.147427051)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097737.717460166)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097739.167082527)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097740.591460884)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097740.929533925)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097744.054695319)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097743.305958045)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097746.138833164)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097743.784974065)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097746.303999855)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097746.364009174)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097747.750131061)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097747.356587002)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097746.064856508)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097745.846524025)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097748.369081888)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097748.255387545)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097750.27266823)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097750.178740437)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097748.208971438)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097748.106459671)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097748.001517441)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097750.181111645)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097750.503915919)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097750.555761064)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097751.45864045)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097750.989127303)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097750.878137267)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097751.241963473)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097751.902758353)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097751.460825997)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097751.349015151)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097751.468810292)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097750.460273226)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097750.177719457)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097694.638522362)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097734.649952486)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097750.17629479)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097750.183074033)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097750.419207368)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097753.030467953)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097697.361545777)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097697.489786393)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097697.872586616)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097697.943495553)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097697.767786294)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097697.842698304)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097697.576722448)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097697.40194085)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097697.164908596)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097697.476756607)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097697.645078353)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097698.127868362)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097697.745834785)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097697.877772556)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097698.855252702)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097698.627521987)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097733.608734889)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097733.431555891)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097732.9839041)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097735.693790129)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097735.242833606)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097735.168906396)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097735.671334142)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097736.056387847)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097736.788694347)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097736.107765964)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097736.895780858)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097736.971384064)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097737.05244723)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097737.343212626)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097738.001551927)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097737.41699843)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097738.371160171)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097737.459833638)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097738.78167479)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097738.51714945)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097738.286901523)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097738.740124749)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097738.793815238)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097738.983360256)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097738.762175483)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097739.456702805)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097739.2628008)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097740.454174037)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097740.118402736)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097740.007958521)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097740.77448332)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097742.279689438)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097741.258589133)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097740.653932679)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097740.9866516)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097741.932776286)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097741.18960442)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097741.962934038)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097740.412761242)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097740.408892016)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097740.872741355)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097741.282085025)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097742.354012003)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097742.445025897)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097742.538528307)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097742.310470589)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097743.098341146)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097743.516059217)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097742.402166126)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097743.797633039)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097743.888674058)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097744.017630755)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097744.616698959)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097743.093544726)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097743.938355286)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097744.316809536)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097744.887434239)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097744.912838508)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097745.929582621)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097746.028621127)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097746.309488597)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097745.142073842)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097746.090927381)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097745.137920091)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097745.506503688)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097744.534729499)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097747.766308486)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097747.256765466)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097746.837052712)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097746.904015421)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097747.010532349)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097747.06199613)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097747.523862895)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097747.505441322)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097747.881580681)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097748.390229068)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097749.521464388)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097748.344501209)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097748.885922814)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097749.040636136)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097749.496475309)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097749.833243313)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097749.291283161)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097750.566236274)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097742.035252128)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097744.143716742)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097751.958133859)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097752.525632367)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097752.363841568)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097752.320420584)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097752.328961096)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097753.970355562)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097752.250704894)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097753.865193719)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097753.850052961)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097753.851223436)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097753.805848823)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097753.247593784)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097753.530896437)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097753.564315747)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097753.467761071)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097752.796547706)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097754.225530799)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097753.582645255)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097754.029834566)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097754.367733473)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097754.946262476)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097755.723130185)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097755.007741125)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097754.414723202)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097755.713973045)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097755.516110179)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097757.026487172)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097755.77412453)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097755.78210428)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097757.393229661)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097754.000372481)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097756.648023319)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097757.287247745)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097756.782833348)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097757.116186197)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097757.664005117)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097756.239060417)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097757.516642282)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097757.125475272)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097758.075334777)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097758.826628624)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097755.717891796)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097755.843275861)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097755.987922492)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097755.564627435)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097756.103354139)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097758.554232214)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097758.327223939)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097758.364861296)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097758.341801001)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097758.281773065)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097758.649156396)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097758.301370315)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097759.051448777)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097758.827803778)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097760.155120963)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097758.886611748)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097759.029655606)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097759.621087029)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097758.976646019)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097759.979688093)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097760.149369069)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097760.313605353)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097760.285593252)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097760.751963426)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097760.630849963)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097760.597416239)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097761.34060961)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097760.632113803)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097760.643923883)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097761.878010576)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097761.214697024)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097761.057565722)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097761.366428769)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097761.459245617)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097760.798146625)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097761.792954529)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097761.495529679)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097761.970462392)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097763.403505452)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097763.147138153)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097763.022885761)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097763.498352133)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097763.48209731)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097762.309168336)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097762.738672965)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097763.34409163)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097763.515780884)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097762.724623193)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097762.816875468)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097763.045549068)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097763.456371042)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097763.428990821)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097763.268012017)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097764.594137735)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097765.363812452)])']\n", "connector: \n", "Evaluating workflow: 20%|██ | 603/3000 [01:36<04:54, 8.13it/s]Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097765.518637901)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097764.619587751)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097764.72991682)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097765.415163169)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097765.337130974)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097765.32650264)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097764.000501925)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097764.646668568)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097766.502040243)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097765.632646074)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097763.322900797)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097765.508415151)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097774.488665325)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097765.682418657)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097766.267089844)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097766.571902277)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097766.045230488)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097774.28008692)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097773.568860665)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097773.938006908)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097770.280230107)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097774.290213497)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097773.909172128)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097773.927133389)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097773.993291686)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097765.68640315)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097773.791585869)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097775.512182595)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097774.657835625)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097774.151749773)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097774.191865236)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097778.012261701)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097775.85700308)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097774.437677088)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097775.688722743)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097775.576994746)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097774.59327113)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097777.596698243)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097787.887817573)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097784.468170282)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097781.129528368)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097781.287730822)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097781.043312807)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097781.062227153)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097781.067746656)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097784.616482648)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097781.569143407)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097780.790486944)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097790.34756317)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097781.827298523)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097784.392763901)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097784.756159259)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097781.987666192)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097781.621865259)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097783.64700824)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097792.22815981)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097781.734580055)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097782.152934544)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097782.926350085)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097783.203988404)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097783.259677253)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097792.024509439)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097783.173676007)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097783.20818133)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097792.360671728)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097790.131499741)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097786.870861946)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097786.95858195)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097786.28337463)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097793.827550062)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097786.121655695)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097787.389262964)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097787.349246231)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097787.353166059)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097787.398379419)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097787.780234302)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097792.064505511)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097789.146885665)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097793.475068055)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097791.940212767)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097787.531177499)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097787.756722941)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097793.734555305)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097789.010231119)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097787.631034537)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097787.885975076)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097791.844956273)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097789.530525892)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097789.662222028)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097789.767653249)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097789.909974648)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097790.02149784)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097790.025627901)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097792.144667295)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097789.89015601)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097792.273778998)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097793.802183764)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097795.209859408)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097795.302559683)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097793.804102281)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097794.248632132)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097795.71876919)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097795.16818472)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097795.544922746)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097795.318235871)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097795.2547554)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097795.366851653)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097796.124235679)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097796.219991046)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097795.215651581)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097795.962488062)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097796.38124251)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097796.262659354)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097797.083237887)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097795.796589455)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097797.105373634)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097796.053184056)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097795.297037469)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097795.451320356)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097797.963972771)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097798.148549612)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097798.715399023)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097797.629498636)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097798.400333403)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097797.919856125)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097798.681617095)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097798.122830451)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097798.524070926)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097799.307062063)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097798.176885194)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097799.457476212)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097799.232864318)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097798.945950685)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097797.90187815)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097798.739438434)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097797.207155928)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097797.464765181)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097797.391868386)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097799.640885514)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097799.659955608)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097800.615414741)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097800.323823592)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097799.914385618)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097800.554069769)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097800.573128489)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097800.672348367)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097800.262569551)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097800.918462764)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097800.702772148)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097800.430802096)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097799.080839688)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097801.163098993)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097800.64992962)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097802.291168593)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097802.015145865)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097802.340808941)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097802.320869362)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097802.301209742)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097801.739411248)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097803.970479602)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097803.086990304)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097802.871769453)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097802.283099098)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097802.626409669)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097802.164809467)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097802.424324129)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097803.07672536)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097802.496877819)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097801.03422385)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097803.176008374)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097803.595121425)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097803.051445005)])']\n", "connector: \n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2026-01-10 17:34:29.487\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:34:29.488\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 20%|██ | 605/3000 [01:40<24:29, 1.63it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:34:31.698\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 20%|██ | 607/3000 [01:42<29:43, 1.34it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:34:31.909\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 20%|██ | 608/3000 [01:43<26:17, 1.52it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:34:33.269\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0306866', 'INHIB_AVE': -1.53, 'SMILES': 'C12(c3c(N(CC(OC)=O)C1=O)cccc3)C(C#N)=C(N)Oc([nH]nc4c5ccc(cc5)OCCCC)c24', 'HIT': 0.0, 'question': 'Is C12(c3c(N(CC(OC)=O)C1=O)cccc3)C(C#N)=C(N)Oc([nH]nc4c5ccc(cc5)OCCCC)c24 effective in eliminating E. coli under experimental conditions?', 'answer': 'No', '_id': 62821}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 7 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 20%|██ | 609/3000 [01:44<31:29, 1.27it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:34:33.293\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0234077', 'INHIB_AVE': 10.18, 'SMILES': 'c1(NC(=O)Nc2c(F)c(F)c(c(F)c2F)F)c(F)c(F)c(c(F)c1F)F', 'HIT': 0.0, 'question': 'Does c1(NC(=O)Nc2c(F)c(F)c(c(F)c2F)F)c(F)c(F)c(c(F)c1F)F show antimicrobial efficacy specifically against Escherichia coli strains?', 'answer': 'No', '_id': 16543}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 7 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:34:33.350\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0253426', 'INHIB_AVE': 1.27, 'SMILES': 'n1c(c2ccccc2)c[nH]c1SCC(=O)Nc3cccc(C)c3C', 'HIT': 0.0, 'question': 'Is n1c(c2ccccc2)c[nH]c1SCC(=O)Nc3cccc(C)c3C effective in eliminating E. coli under experimental conditions?', 'answer': 'No', '_id': 52298}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 7 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:34:33.361\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0283756', 'INHIB_AVE': 4.62, 'SMILES': 'c12c(nc3c(cccc3)n1)c(nn2c4ccc(cc4)C)c5ccccc5', 'HIT': 0.0, 'question': 'Does c12c(nc3c(cccc3)n1)c(nn2c4ccc(cc4)C)c5ccccc5 show antimicrobial efficacy specifically against Escherichia coli strains?', 'answer': 'No', '_id': 37355}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 7 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:34:33.379\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0295406', 'INHIB_AVE': -10.47, 'SMILES': 'c1(OC(N)=C(C2c3ccc(c(OC)c3)OC)C#N)c2c(n[nH]1)CC', 'HIT': 0.0, 'question': 'Is c1(OC(N)=C(C2c3ccc(c(OC)c3)OC)C#N)c2c(n[nH]1)CC effective in eliminating E. coli under experimental conditions?', 'answer': 'No', '_id': 79064}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 7 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:34:33.380\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0255796', 'INHIB_AVE': 7.65, 'SMILES': 'C1(C(OCCOc2ccccc2)=O)=C(C)NC(=O)NC1c3ccc(cc3)O', 'HIT': 0.0, 'question': 'Is C1(C(OCCOc2ccccc2)=O)=C(C)NC(=O)NC1c3ccc(cc3)O effective in eliminating E. coli under experimental conditions?', 'answer': 'No', '_id': 24469}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 7 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 20%|██ | 613/3000 [01:44<15:45, 2.52it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:34:33.415\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0312189', 'INHIB_AVE': -1.6, 'SMILES': 'n(c(c(cc1)ccc1[N+](=O)[O-])nn2CC(=O)Nc3ccc(cc3C)C)c2SCC(=O)Nc4ccc(cc4C)C', 'HIT': 0.0, 'question': 'Does n(c(c(cc1)ccc1[N+](=O)[O-])nn2CC(=O)Nc3ccc(cc3C)C)c2SCC(=O)Nc4ccc(cc4C)C demonstrate bactericidal effects against E. coli in vitro?', 'answer': 'No', '_id': 63068}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 7 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:34:33.436\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0245049', 'INHIB_AVE': 10.78, 'SMILES': 'c1(c(Br)cc(c(c1)NC(=O)c2ccccc2F)Br)C(F)(F)F', 'HIT': 0.0, 'question': 'Can c1(c(Br)cc(c(c1)NC(=O)c2ccccc2F)Br)C(F)(F)F effectively kill or suppress Escherichia coli cells?', 'answer': 'No', '_id': 15040}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 7 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:34:38.724\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0307612', 'INHIB_AVE': -11.84, 'SMILES': 'C(=O)(N(CC1)CCN1c2ccc(cc2F)C(CCC)=O)c3ccccc3Br', 'HIT': 0.0, 'question': 'Can exposure to C(=O)(N(CC1)CCN1c2ccc(cc2F)C(CCC)=O)c3ccccc3Br lead to reduced viability of E. coli?', 'answer': 'No', '_id': 79934}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 21%|██ | 617/3000 [01:50<31:35, 1.26it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:34:38.744\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0156221', 'INHIB_AVE': 11.5, 'SMILES': 'N#CC1=CC(C2=C3C=CC=C2)=C(C=C3)OC1=N', 'HIT': 0.0, 'question': 'Does the proposed molecule N#CC1=CC(C2=C3C=CC=C2)=C(C=C3)OC1=N exhibit antibacterial activity against Escherichia coli?', 'answer': 'No', '_id': 13449}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:34:38.883\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 21%|██ | 619/3000 [01:50<24:56, 1.59it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:34:39.101\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0299561', 'INHIB_AVE': -0.36, 'SMILES': 'C(=O)(CN(CC1)CCO1)N\\\\N=C\\\\c2ccccc2O', 'HIT': 0.0, 'question': 'Does the proposed molecule C(=O)(CN(CC1)CCO1)N\\\\N=C\\\\c2ccccc2O exhibit antibacterial activity against Escherichia coli?', 'answer': 'No', '_id': 58849}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:34:39.111\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0223787', 'INHIB_AVE': -0.83, 'SMILES': 'C(=O)(Nc1ccccc1C)c2ccccc2OC', 'HIT': 0.0, 'question': 'Does C(=O)(Nc1ccccc1C)c2ccccc2OC demonstrate bactericidal effects against E. coli in vitro?', 'answer': 'No', '_id': 60530}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 21%|██ | 621/3000 [01:50<19:50, 2.00it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:34:39.487\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:34:39.497\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 21%|██ | 623/3000 [01:50<16:37, 2.38it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:34:40.519\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 21%|██ | 624/3000 [01:51<20:22, 1.94it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:34:40.525\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:34:40.567\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:34:40.601\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:34:40.619\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0293892', 'INHIB_AVE': 9.43, 'SMILES': 'N1(c2ccc(cc2)OCC)CC(CC1=O)C(=O)NC(CC3(C)C)CC(N3)(C)C', 'HIT': 0.0, 'question': 'Is N1(c2ccc(cc2)OCC)CC(CC1=O)C(=O)NC(CC3(C)C)CC(N3)(C)C capable of inhibiting the growth or survival of E. coli?', 'answer': 'No', '_id': 18649}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:34:40.620\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0246017', 'INHIB_AVE': 10.84, 'SMILES': 'N1(C(=O)c2ccccc2F)C(O)(c3ccc(cc3)Cl)CC(C(F)(F)F)=N1', 'HIT': 0.0, 'question': 'Is Escherichia coli susceptible to treatment with the compound N1(C(=O)c2ccccc2F)C(O)(c3ccc(cc3)Cl)CC(C(F)(F)F)=N1?', 'answer': 'No', '_id': 14900}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 21%|██ | 630/3000 [01:52<10:26, 3.78it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:34:41.193\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0256618', 'INHIB_AVE': 2.66, 'SMILES': 'C(C(OC)=O)(=C1C)\\\\C(=C\\\\c2ccc(cc2Cl)Cl)\\\\C(N1CCOC)=O', 'HIT': 0.0, 'question': 'Is C(C(OC)=O)(=C1C)\\\\C(=C\\\\c2ccc(cc2Cl)Cl)\\\\C(N1CCOC)=O capable of inhibiting the growth or survival of E. coli?', 'answer': 'No', '_id': 46175}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:34:43.632\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 21%|██ | 632/3000 [01:54<21:14, 1.86it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:34:43.864\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 21%|██ | 633/3000 [01:55<19:18, 2.04it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:34:43.875\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:34:44.017\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 21%|██ | 635/3000 [01:55<14:11, 2.78it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:34:44.239\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 21%|██ | 636/3000 [01:55<13:11, 2.99it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:34:44.250\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:34:44.804\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0320126', 'INHIB_AVE': -6.38, 'SMILES': 'n1(CC(=O)N(CC)C2(CCCCC2)C(=O)Nc(cc3)ccc3C)nnnc1N', 'HIT': 0.0, 'question': 'Is Escherichia coli susceptible to treatment with the compound n1(CC(=O)N(CC)C2(CCCCC2)C(=O)Nc(cc3)ccc3C)nnnc1N?', 'answer': 'No', '_id': 74376}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 21%|██▏ | 638/3000 [01:56<12:28, 3.16it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:34:44.863\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0255273', 'INHIB_AVE': 4.7, 'SMILES': 'C(C(OCCOc1ccccc1)=O)(=C(C)NC(=C23)CC(CC2=O)(C)C)C3c4ccccc4F', 'HIT': 0.0, 'question': 'Can exposure to C(C(OCCOc1ccccc1)=O)(=C(C)NC(=C23)CC(CC2=O)(C)C)C3c4ccccc4F lead to reduced viability of E. coli?', 'answer': 'No', '_id': 36992}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:34:44.889\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0259605', 'INHIB_AVE': -5.9, 'SMILES': 'c1(SCc2ccccc2)sc3c(ccc(c3)NC(=O)c4ccccc4OC)n1', 'HIT': 0.0, 'question': 'Does treatment with c1(SCc2ccccc2)sc3c(ccc(c3)NC(=O)c4ccccc4OC)n1 result in bacteriostatic or bactericidal effects on E. coli?', 'answer': 'No', '_id': 73539}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:34:44.901\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0293976', 'INHIB_AVE': -10.01, 'SMILES': 'c1(NC(=O)c2ccccc2F)n(c3ccccc3)ncc1C(=O)N4CCC(CC4)C(N)=O', 'HIT': 0.0, 'question': 'Can c1(NC(=O)c2ccccc2F)n(c3ccccc3)ncc1C(=O)N4CCC(CC4)C(N)=O act as an antibacterial agent targeting Escherichia coli?', 'answer': 'No', '_id': 78709}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:34:46.703\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 21%|██▏ | 642/3000 [01:57<15:31, 2.53it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:34:46.739\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:34:47.329\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 21%|██▏ | 644/3000 [01:58<14:40, 2.68it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:34:47.756\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 22%|██▏ | 645/3000 [01:59<15:00, 2.62it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:34:47.769\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:34:47.791\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:34:47.838\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0286502', 'INHIB_AVE': -2.92, 'SMILES': 'C(/c1c(N2Cc3c(cccc3Cl)Cl)cccc1)(=N/N(C)C(NC)=S)\\\\C2=O', 'HIT': 0.0, 'question': 'Is Escherichia coli susceptible to treatment with the compound C(/c1c(N2Cc3c(cccc3Cl)Cl)cccc1)(=N/N(C)C(NC)=S)\\\\C2=O?', 'answer': 'No', '_id': 67027}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:34:48.030\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 22%|██▏ | 649/3000 [01:59<09:08, 4.28it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:34:49.000\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 22%|██▏ | 650/3000 [02:00<13:19, 2.94it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 22%|██▏ | 651/3000 [02:00<14:55, 2.62it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:34:49.730\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0230253', 'INHIB_AVE': 10.67, 'SMILES': 'C(OC)(=O)c1ccc(cc1)OCc2ccccc2', 'HIT': 0.0, 'question': 'Does C(OC)(=O)c1ccc(cc1)OCc2ccccc2 show antimicrobial efficacy specifically against Escherichia coli strains?', 'answer': 'No', '_id': 15285}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 22%|██▏ | 653/3000 [02:01<10:58, 3.57it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:34:49.842\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 22%|██▏ | 654/3000 [02:01<09:42, 4.03it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:34:49.875\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 22%|██▏ | 657/3000 [02:01<07:38, 5.11it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 22%|██▏ | 658/3000 [02:02<12:20, 3.16it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 22%|██▏ | 660/3000 [02:02<09:32, 4.09it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 22%|██▏ | 663/3000 [02:02<06:51, 5.68it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 22%|██▏ | 664/3000 [02:03<07:59, 4.87it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 22%|██▏ | 665/3000 [02:03<09:12, 4.22it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 22%|██▏ | 669/3000 [02:04<06:51, 5.66it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 22%|██▏ | 671/3000 [02:04<06:47, 5.72it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 22%|██▎ | 675/3000 [02:04<03:37, 10.69it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 23%|██▎ | 677/3000 [02:04<04:11, 9.23it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 23%|██▎ | 679/3000 [02:05<03:41, 10.46it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 23%|██▎ | 681/3000 [02:05<05:50, 6.61it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 23%|██▎ | 683/3000 [02:05<05:00, 7.70it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 23%|██▎ | 685/3000 [02:06<04:54, 7.85it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 23%|██▎ | 688/3000 [02:06<06:22, 6.04it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 23%|██▎ | 692/3000 [02:07<04:15, 9.02it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 23%|██▎ | 694/3000 [02:07<04:06, 9.36it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 23%|██▎ | 696/3000 [02:07<04:58, 7.72it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 23%|██▎ | 697/3000 [02:07<06:05, 6.31it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 23%|██▎ | 698/3000 [02:08<06:32, 5.86it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 23%|██▎ | 700/3000 [02:08<07:05, 5.41it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 23%|██▎ | 702/3000 [02:08<05:57, 6.43it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 24%|██▎ | 705/3000 [02:08<04:10, 9.16it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 24%|██▎ | 709/3000 [02:09<03:14, 11.81it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 24%|██▎ | 711/3000 [02:09<03:49, 9.96it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 24%|██▍ | 715/3000 [02:10<04:58, 7.65it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 24%|██▍ | 717/3000 [02:10<06:25, 5.92it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 24%|██▍ | 719/3000 [02:10<05:02, 7.54it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 24%|██▍ | 721/3000 [02:11<05:24, 7.02it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 24%|██▍ | 724/3000 [02:11<04:16, 8.87it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 24%|██▍ | 726/3000 [02:11<04:58, 7.61it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 24%|██▍ | 730/3000 [02:12<04:14, 8.93it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 24%|██▍ | 734/3000 [02:12<03:56, 9.57it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 25%|██▍ | 736/3000 [02:12<05:25, 6.96it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 25%|██▍ | 737/3000 [02:13<05:56, 6.35it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 25%|██▍ | 740/3000 [02:13<04:32, 8.30it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 25%|██▍ | 741/3000 [02:13<06:13, 6.04it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 25%|██▍ | 744/3000 [02:14<06:07, 6.13it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 25%|██▍ | 747/3000 [02:14<03:57, 9.50it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metricsmetrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", " {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 25%|██▌ | 750/3000 [02:14<04:09, 9.03it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 25%|██▌ | 752/3000 [02:14<04:03, 9.21it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 25%|██▌ | 755/3000 [02:15<04:53, 7.64it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 25%|██▌ | 756/3000 [02:15<04:58, 7.51it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 25%|██▌ | 757/3000 [02:15<06:00, 6.22it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 25%|██▌ | 760/3000 [02:16<05:06, 7.31it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 25%|██▌ | 763/3000 [02:16<04:15, 8.76it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 26%|██▌ | 768/3000 [02:16<03:26, 10.80it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 26%|██▌ | 772/3000 [02:17<03:29, 10.62it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 26%|██▌ | 774/3000 [02:17<05:36, 6.62it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 26%|██▌ | 778/3000 [02:18<03:51, 9.58it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 26%|██▌ | 780/3000 [02:18<05:28, 6.76it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 26%|██▌ | 782/3000 [02:18<05:36, 6.59it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 26%|██▌ | 786/3000 [02:19<03:52, 9.53it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 26%|██▋ | 789/3000 [02:19<03:27, 10.64it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 26%|██▋ | 791/3000 [02:19<05:23, 6.82it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 26%|██▋ | 793/3000 [02:20<05:24, 6.80it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 27%|██▋ | 796/3000 [02:20<05:10, 7.10it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 27%|██▋ | 799/3000 [02:20<03:39, 10.05it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 27%|██▋ | 801/3000 [02:21<03:26, 10.65it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 27%|██▋ | 806/3000 [02:21<03:55, 9.33it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 27%|██▋ | 808/3000 [02:21<03:21, 10.86it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 27%|██▋ | 810/3000 [02:22<05:47, 6.30it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 27%|██▋ | 812/3000 [02:22<04:50, 7.53it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 27%|██▋ | 815/3000 [02:23<06:30, 5.60it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 27%|██▋ | 821/3000 [02:24<05:04, 7.15it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 28%|██▊ | 825/3000 [02:24<03:17, 10.99it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 28%|██▊ | 827/3000 [02:24<03:46, 9.59it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 28%|██▊ | 829/3000 [02:24<03:46, 9.60it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 28%|██▊ | 833/3000 [02:25<05:09, 7.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 28%|██▊ | 835/3000 [02:25<05:00, 7.20it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 28%|██▊ | 839/3000 [02:26<05:02, 7.13it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 28%|██▊ | 840/3000 [02:26<05:49, 6.17it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 28%|██▊ | 844/3000 [02:27<04:35, 7.82it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 28%|██▊ | 846/3000 [02:27<03:46, 9.51it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 28%|██▊ | 849/3000 [02:27<04:40, 7.66it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 28%|██▊ | 852/3000 [02:28<05:40, 6.31it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 28%|██▊ | 854/3000 [02:28<04:41, 7.62it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 29%|██▊ | 856/3000 [02:28<04:42, 7.59it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 29%|██▊ | 860/3000 [02:29<04:16, 8.33it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 29%|██▉ | 867/3000 [02:29<02:21, 15.04it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 29%|██▉ | 871/3000 [02:30<04:33, 7.78it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 29%|██▉ | 876/3000 [02:31<05:17, 6.69it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 29%|██▉ | 881/3000 [02:31<04:12, 8.38it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 29%|██▉ | 884/3000 [02:32<03:42, 9.49it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 30%|██▉ | 888/3000 [02:32<04:29, 7.84it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 30%|██▉ | 890/3000 [02:33<05:12, 6.75it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 30%|██▉ | 893/3000 [02:33<04:41, 7.47it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 30%|██▉ | 895/3000 [02:33<03:52, 9.07it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 30%|██▉ | 897/3000 [02:33<03:45, 9.33it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 30%|██▉ | 899/3000 [02:34<05:23, 6.50it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 30%|███ | 900/3000 [02:34<05:15, 6.67it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 30%|███ | 902/3000 [02:34<04:52, 7.17it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 30%|███ | 908/3000 [02:35<04:28, 7.78it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 30%|███ | 911/3000 [02:35<04:41, 7.43it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 30%|███ | 915/3000 [02:36<03:16, 10.63it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 31%|███ | 917/3000 [02:36<04:02, 8.59it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 31%|███ | 919/3000 [02:36<04:01, 8.63it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2026-01-10 17:35:33.924\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0328930', 'INHIB_AVE': 6.66, 'SMILES': 'n1(N)c(CC)nnc1N\\\\N=C\\\\c2c[nH]nc2c3ccc(cc3)F.Cl.Cl', 'HIT': 0.0, 'question': 'Is n1(N)c(CC)nnc1N\\\\N=C\\\\c2c[nH]nc2c3ccc(cc3)F.Cl.Cl capable of inhibiting the growth or survival of E. coli?', 'answer': 'No', '_id': 28469}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 31%|███ | 921/3000 [02:45<46:20, 1.34s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:35:33.934\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0160279', 'INHIB_AVE': 13.72, 'SMILES': 'ClC1=C(OC)C=CC(/C=N/C2=CC=C(C=C2)C(O3)=NC4=C3C=C(C)C=C4)=C1', 'HIT': 0.0, 'question': 'Is ClC1=C(OC)C=CC(/C=N/C2=CC=C(C=C2)C(O3)=NC4=C3C=C(C)C=C4)=C1 effective in eliminating E. coli under experimental conditions?', 'answer': 'No', '_id': 9448}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:35:33.946\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0231210', 'INHIB_AVE': -4.68, 'SMILES': 'c1(O)c(cccc1\\\\C=N\\\\NC(CCCc(ccc(c2C)OC)c2)=O)OC', 'HIT': 0.0, 'question': 'Can exposure to c1(O)c(cccc1\\\\C=N\\\\NC(CCCc(ccc(c2C)OC)c2)=O)OC lead to reduced viability of E. coli?', 'answer': 'No', '_id': 71201}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:35:34.076\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0301930', 'INHIB_AVE': 9.07, 'SMILES': 'c1(cc(ccc1N2CCCCC2)[N+](=O)[O-])C(=O)Nc(cc3)ccc3Oc4ccccc4', 'HIT': 0.0, 'question': 'Does c1(cc(ccc1N2CCCCC2)[N+](=O)[O-])C(=O)Nc(cc3)ccc3Oc4ccccc4 show antimicrobial efficacy specifically against Escherichia coli strains?', 'answer': 'No', '_id': 19659}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 7 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 31%|███ | 924/3000 [02:45<28:55, 1.20it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:35:34.162\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0254720', 'INHIB_AVE': 2.64, 'SMILES': 'N1(C2c3ccc(c(OC)c3)OC)C(S\\\\C(=C\\\\c(ccc4c5ccccc5C(O)=O)o4)\\\\C1=O)=NC(C)=C2C(OCC)=O', 'HIT': 0.0, 'question': 'Is Escherichia coli susceptible to treatment with the compound N1(C2c3ccc(c(OC)c3)OC)C(S\\\\C(=C\\\\c(ccc4c5ccccc5C(O)=O)o4)\\\\C1=O)=NC(C)=C2C(OCC)=O?', 'answer': 'No', '_id': 46222}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 7 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:35:38.617\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0299557', 'INHIB_AVE': 1.84, 'SMILES': 'c1(cc(C#N)c(cc1Nc(cc2)ccc2Oc(cc3)ccc3Cl)C#N)[N+](=O)[O-]', 'HIT': 0.0, 'question': 'Is c1(cc(C#N)c(cc1Nc(cc2)ccc2Oc(cc3)ccc3Cl)C#N)[N+](=O)[O-] effective in eliminating E. coli under experimental conditions?', 'answer': 'No', '_id': 49807}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 31%|███ | 926/3000 [02:49<42:19, 1.22s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:35:38.647\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0285220', 'INHIB_AVE': -0.19, 'SMILES': 'N1(c2cccc(Cl)c2)C(=O)C(CC1=O)SCC(N)C(OCC)=O.Cl', 'HIT': 0.0, 'question': 'Does the proposed molecule N1(c2cccc(Cl)c2)C(=O)C(CC1=O)SCC(N)C(OCC)=O.Cl exhibit antibacterial activity against Escherichia coli?', 'answer': 'No', '_id': 58194}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:35:38.648\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0218193', 'INHIB_AVE': 0.39, 'SMILES': 'n1c(ccc(cccc2)c12)\\\\C=C\\\\c(cc3)ccc3\\\\C=C\\\\c4ccc(c5n4)cccc5', 'HIT': 0.0, 'question': 'Can exposure to n1c(ccc(cccc2)c12)\\\\C=C\\\\c(cc3)ccc3\\\\C=C\\\\c4ccc(c5n4)cccc5 lead to reduced viability of E. coli?', 'answer': 'No', '_id': 55886}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:35:38.671\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0296007', 'INHIB_AVE': -6.34, 'SMILES': 'C(=CC(\\\\C=C\\\\c(cc1)ccc1N2CCCC2)=NC3=O)(C(F)(F)F)N3', 'HIT': 0.0, 'question': 'Does the proposed molecule C(=CC(\\\\C=C\\\\c(cc1)ccc1N2CCCC2)=NC3=O)(C(F)(F)F)N3 exhibit antibacterial activity against Escherichia coli?', 'answer': 'No', '_id': 74308}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:35:38.686\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0137181', 'INHIB_AVE': -12.43, 'SMILES': 'c1(O)cc(OC(=O)c(c2c(o3)cc(O)cc2)c34)c4cc1', 'HIT': 0.0, 'question': 'Can c1(O)cc(OC(=O)c(c2c(o3)cc(O)cc2)c34)c4cc1 act as an antibacterial agent targeting Escherichia coli?', 'answer': 'No', '_id': 80264}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:35:38.694\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0254449', 'INHIB_AVE': -33.21, 'SMILES': 'C12(C=Cc(c(c(OC)cc3[N+](=O)[O-])O1)c3)c4c(cccc4)c(c5N2C)cccc5', 'HIT': 0.0, 'question': 'Is Escherichia coli susceptible to treatment with the compound C12(C=Cc(c(c(OC)cc3[N+](=O)[O-])O1)c3)c4c(cccc4)c(c5N2C)cccc5?', 'answer': 'No', '_id': 82628}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:35:38.720\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0155451', 'INHIB_AVE': -2.26, 'SMILES': 'O=C1NC(C(Br)=CC=C2)=C2/C1=N\\\\NC(COC3=C(Br)C=C(C)C=C3Br)=O', 'HIT': 0.0, 'question': 'Is Escherichia coli susceptible to treatment with the compound O=C1NC(C(Br)=CC=C2)=C2/C1=N\\\\NC(COC3=C(Br)C=C(C)C=C3Br)=O?', 'answer': 'No', '_id': 65128}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 31%|███ | 932/3000 [02:50<19:51, 1.74it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:35:39.168\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0284914', 'INHIB_AVE': -5.97, 'SMILES': 'n1c(SCC)c2c(c(C)n1)cccc2', 'HIT': 0.0, 'question': 'Is n1c(SCC)c2c(c(C)n1)cccc2 capable of inhibiting the growth or survival of E. coli?', 'answer': 'No', '_id': 73681}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:35:39.220\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0329742', 'INHIB_AVE': -3.53, 'SMILES': 'C1(C(F)(F)F)(NC(=O)c2ccccc2F)C(=O)NC3=C1C(=O)NC(N3c4ccccc4)=O', 'HIT': 0.0, 'question': 'Does C1(C(F)(F)F)(NC(=O)c2ccccc2F)C(=O)NC3=C1C(=O)NC(N3c4ccccc4)=O show antimicrobial efficacy specifically against Escherichia coli strains?', 'answer': 'No', '_id': 68595}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:35:39.272\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0255911', 'INHIB_AVE': -9.52, 'SMILES': 'S(=O)(=O)(c1ccc(cc1)C)N(Cc2ccccc2)c3ccccc3C(NCCSc4sc(c5n4)cccc5)=O', 'HIT': 0.0, 'question': 'Does the proposed molecule S(=O)(=O)(c1ccc(cc1)C)N(Cc2ccccc2)c3ccccc3C(NCCSc4sc(c5n4)cccc5)=O exhibit antibacterial activity against Escherichia coli?', 'answer': 'No', '_id': 78339}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 31%|███ | 935/3000 [02:50<16:05, 2.14it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:35:39.371\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0320735', 'INHIB_AVE': 18.19, 'SMILES': 'N1(c2cccc(Cl)c2)C(c3ccc(c4c3)OCO4)C(C1=O)Oc5ccccc5', 'HIT': 0.0, 'question': 'Does treatment with N1(c2cccc(Cl)c2)C(c3ccc(c4c3)OCO4)C(C1=O)Oc5ccccc5 result in bacteriostatic or bactericidal effects on E. coli?', 'answer': 'No', '_id': 4146}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:35:39.387\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0226016', 'INHIB_AVE': 13.62, 'SMILES': 'C1(=N2)N(C=CN=C1N3CCCC(C)C3)C(N2CC(=O)Nc4ccc(cc4F)F)=O', 'HIT': 0.0, 'question': 'Does C1(=N2)N(C=CN=C1N3CCCC(C)C3)C(N2CC(=O)Nc4ccc(cc4F)F)=O demonstrate bactericidal effects against E. coli in vitro?', 'answer': 'No', '_id': 9593}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 31%|███ | 937/3000 [02:50<13:07, 2.62it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:35:43.821\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 31%|███▏ | 939/3000 [02:55<27:36, 1.24it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:35:43.832\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:35:43.852\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 4 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:35:44.230\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 31%|███▏ | 942/3000 [02:55<20:00, 1.71it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:35:44.294\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:35:44.304\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:35:44.352\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 32%|███▏ | 945/3000 [02:55<14:01, 2.44it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:35:44.416\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:35:47.668\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0241385', 'INHIB_AVE': 7.91, 'SMILES': 'c1(C(=O)OCC(=O)c2ccc(cc2)Cl)c(cccc1[N+](=O)[O-])C(=O)OCC(=O)c3ccc(cc3)Cl', 'HIT': 0.0, 'question': 'Does c1(C(=O)OCC(=O)c2ccc(cc2)Cl)c(cccc1[N+](=O)[O-])C(=O)OCC(=O)c3ccc(cc3)Cl demonstrate bactericidal effects against E. coli in vitro?', 'answer': 'No', '_id': 23537}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 32%|███▏ | 947/3000 [02:58<23:59, 1.43it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:35:47.749\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0300665', 'INHIB_AVE': 8.65, 'SMILES': 'C1(=COc(c2C1=O)ccc(Cl)c2)C3Nc4c(NC(CC(c5ccc(c(OC)c5)OC)CC6=O)=C36)cccc4', 'HIT': 0.0, 'question': 'Does the proposed molecule C1(=COc(c2C1=O)ccc(Cl)c2)C3Nc4c(NC(CC(c5ccc(c(OC)c5)OC)CC6=O)=C36)cccc4 exhibit antibacterial activity against Escherichia coli?', 'answer': 'No', '_id': 21017}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:35:47.860\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0255002', 'INHIB_AVE': 9.83, 'SMILES': 'C(C#N)(C1c(cc2)ccc2N(CC)CC)=C(N)N(C(CC(C)(C)CC3=O)=C13)c4ccc(cc4)C', 'HIT': 0.0, 'question': 'Does C(C#N)(C1c(cc2)ccc2N(CC)CC)=C(N)N(C(CC(C)(C)CC3=O)=C13)c4ccc(cc4)C show antimicrobial efficacy specifically against Escherichia coli strains?', 'answer': 'No', '_id': 17499}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 32%|███▏ | 949/3000 [02:59<18:47, 1.82it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:35:47.955\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0243783', 'INHIB_AVE': 8.58, 'SMILES': 'c(c(SCC(=CC1=O)OC(C(c2cc3c(OCO3)c(OC)c2)CC(=O)OC)=C1O)c(F)c4)cc4F', 'HIT': 0.0, 'question': 'Does treatment with c(c(SCC(=CC1=O)OC(C(c2cc3c(OCO3)c(OC)c2)CC(=O)OC)=C1O)c(F)c4)cc4F result in bacteriostatic or bactericidal effects on E. coli?', 'answer': 'No', '_id': 21224}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 2 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:35:48.193\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0320218', 'INHIB_AVE': -6.68, 'SMILES': 'C12=C(CCCC1=O)Nc(c3C2c4cccc(Cl)c4Cl)ccc5c3ccc(n5)C', 'HIT': 0.0, 'question': 'Is C12=C(CCCC1=O)Nc(c3C2c4cccc(Cl)c4Cl)ccc5c3ccc(n5)C capable of inhibiting the growth or survival of E. coli?', 'answer': 'No', '_id': 74858}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 32%|███▏ | 952/3000 [03:00<20:10, 1.69it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 32%|███▏ | 954/3000 [03:01<15:27, 2.21it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:35:49.957\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0234442', 'INHIB_AVE': 1.43, 'SMILES': 'c1(C#N)c2c(sc1\\\\N=C\\\\c3ccc(I)s3)CC(C)CC2', 'HIT': 0.0, 'question': 'Does c1(C#N)c2c(sc1\\\\N=C\\\\c3ccc(I)s3)CC(C)CC2 demonstrate bactericidal effects against E. coli in vitro?', 'answer': 'No', '_id': 51612}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 32%|███▏ | 957/3000 [03:01<10:45, 3.16it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:35:50.435\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 32%|███▏ | 958/3000 [03:01<10:11, 3.34it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:35:50.483\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 32%|███▏ | 960/3000 [03:02<13:46, 2.47it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 32%|███▏ | 963/3000 [03:03<09:04, 3.74it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 32%|███▏ | 965/3000 [03:03<08:19, 4.07it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 32%|███▏ | 966/3000 [03:03<08:01, 4.23it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 32%|███▏ | 968/3000 [03:04<08:13, 4.12it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 32%|███▏ | 971/3000 [03:04<06:14, 5.42it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 33%|███▎ | 976/3000 [03:05<03:56, 8.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 33%|███▎ | 978/3000 [03:05<03:56, 8.55it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 33%|███▎ | 982/3000 [03:05<02:59, 11.27it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 33%|███▎ | 984/3000 [03:05<03:21, 10.02it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 33%|███▎ | 986/3000 [03:06<05:53, 5.70it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 33%|███▎ | 988/3000 [03:06<04:38, 7.24it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 33%|███▎ | 990/3000 [03:07<05:40, 5.91it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 33%|███▎ | 995/3000 [03:07<03:57, 8.46it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 33%|███▎ | 1000/3000 [03:07<02:52, 11.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 33%|███▎ | 1002/3000 [03:08<05:29, 6.07it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 34%|███▎ | 1006/3000 [03:09<04:46, 6.96it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 34%|███▎ | 1008/3000 [03:09<04:45, 6.98it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 34%|███▍ | 1013/3000 [03:10<03:48, 8.71it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 34%|███▍ | 1015/3000 [03:10<03:52, 8.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 34%|███▍ | 1018/3000 [03:11<05:55, 5.58it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 34%|███▍ | 1019/3000 [03:11<06:16, 5.26it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 34%|███▍ | 1023/3000 [03:11<04:33, 7.23it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 34%|███▍ | 1028/3000 [03:12<02:38, 12.46it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 34%|███▍ | 1031/3000 [03:12<02:19, 14.08it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 34%|███▍ | 1033/3000 [03:12<03:02, 10.79it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 35%|███▍ | 1037/3000 [03:13<03:41, 8.87it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 35%|███▍ | 1039/3000 [03:13<04:24, 7.41it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 35%|███▍ | 1043/3000 [03:14<04:33, 7.16it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 35%|███▍ | 1046/3000 [03:14<03:42, 8.80it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 35%|███▍ | 1048/3000 [03:14<04:00, 8.11it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 35%|███▌ | 1051/3000 [03:14<03:22, 9.62it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 35%|███▌ | 1054/3000 [03:15<04:32, 7.15it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 35%|███▌ | 1056/3000 [03:15<03:40, 8.81it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 35%|███▌ | 1062/3000 [03:16<04:09, 7.77it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 35%|███▌ | 1064/3000 [03:16<03:48, 8.49it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 36%|███▌ | 1066/3000 [03:17<04:31, 7.13it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 36%|███▌ | 1070/3000 [03:17<03:07, 10.28it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 36%|███▌ | 1072/3000 [03:17<04:21, 7.37it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 36%|███▌ | 1074/3000 [03:18<04:30, 7.12it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 36%|███▌ | 1077/3000 [03:18<03:39, 8.76it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 36%|███▌ | 1079/3000 [03:18<04:05, 7.82it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Evaluating workflow: 36%|███▌ | 1081/3000 [03:19<05:15, 6.08it/s]Unclosed connector\n", "connections: ['deque([(, 2097804.72845879)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097804.782479729)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097806.272878961)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097806.502743109)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097805.183298568)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097804.86514611)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097805.122965769)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097808.430080132)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097806.854035307)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097807.100781122)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097807.201417192)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097808.619547143)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097807.374626808)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097807.409021884)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097809.704639351)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097809.402212511)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097809.18223911)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097811.122795448)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097812.091757648)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097814.13307571)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097815.249524193)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097811.033222113)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097813.776584449)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097814.829623055)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097813.074253799)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097816.741614439)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097817.951768916)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097815.063624101)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097817.259600319)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097817.477198415)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097817.672440816)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097817.729774431)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097817.801985565)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097817.109263545)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097819.475259072)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097820.347305488)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097819.592308574)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097820.027849883)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097819.47934268)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097822.526043142)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097822.920856605)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097822.796667857)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097821.866334728)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097824.443503917)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097824.958537478)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097824.522655965)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097825.132883364)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097826.302182796)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097835.882511673)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097827.045731593)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097835.884470838)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097841.253533446)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097835.869870848)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097835.855556674)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097841.232590548)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097835.80134697)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097835.944148925)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097841.619014217)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097843.701270045)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097835.777464728)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097841.609924655)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097835.923913238)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097827.392232538)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097826.62771993)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097827.94506094)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097827.794545643)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097827.796060292)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097828.284314986)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097843.124461764)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097843.122545105)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097803.304013069)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097804.911064863)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097804.992677521)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097805.986158685)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097805.0921768)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097804.502253866)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097804.462735501)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097803.271711265)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097804.410475068)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097804.924850897)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097805.144779355)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097806.427968715)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097805.623992203)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097806.637561241)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097808.676302269)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097807.575490519)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097805.444169927)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097806.763276983)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097807.115112262)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097807.890677117)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097807.354413951)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097807.623760356)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097808.678034908)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097809.121803277)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097809.327512699)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097808.454252025)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097808.971809909)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097808.679132127)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097809.194342444)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097810.87743713)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097809.337162245)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097809.846216115)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097810.059764719)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097809.512844057)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097810.568718796)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097810.819745974)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097809.735699315)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097810.171220411)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097810.985738365)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097811.160984751)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097812.086926802)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097810.437053868)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097811.896550014)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097811.773975646)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097812.702752446)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097812.079902175)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097812.272545343)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097811.900370682)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097812.262521434)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097812.46831433)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097811.999529876)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097814.261796011)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097814.164069515)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097814.814139429)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097814.355499574)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097814.735649353)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097815.211095853)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097814.769065762)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097814.613490782)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097814.802699565)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097815.595657552)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097814.126705427)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097814.727283164)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097814.767076016)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097815.242222195)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097816.078258225)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097816.217198053)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097817.567816093)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097816.640238373)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097816.74555946)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097818.007330175)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097817.994969918)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097816.601303584)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097817.602598487)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097818.038958886)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097817.507067388)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097819.495016613)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097818.687151142)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097819.596969482)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097818.671172154)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097819.435025554)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097818.909506619)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097820.559164244)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097819.574707502)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097820.865884027)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097819.905272106)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097820.552805003)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097820.773591033)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097820.267883592)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097820.193538631)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097821.270195333)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097822.12357256)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097821.599696225)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097821.602105517)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097819.973506562)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097822.260687189)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097823.096550659)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097822.633463564)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097822.292208118)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097822.952360677)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097823.305640167)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097822.45050986)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097822.386937982)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097822.948382195)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097823.689780687)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097824.615487587)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097822.720019142)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097823.693068491)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097824.883852724)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097824.237069243)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097824.655206731)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097824.503808217)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097824.429040906)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097824.399868453)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097824.939740734)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097825.16938406)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097825.144551355)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097825.054823715)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097826.994128152)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097825.435091171)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097826.729367963)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097826.618912183)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097826.67495655)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097827.011484781)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097827.371434619)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097826.910321996)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097827.364868039)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097815.374072989)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097825.725779997)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097847.371892304)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097831.992655231)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097831.990673059)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097834.206904834)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097834.418532873)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097841.391726563)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097847.397866328)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097850.346971713)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097847.31282985)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097847.409852284)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097843.555252533)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097843.029372551)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097843.075093602)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097846.384068583)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097846.526339264)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097846.748060509)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097846.141381678)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097852.386289246)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097841.9965671)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097842.005107735)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097843.025416807)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097843.109463861)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097846.759347946)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097846.372838866)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097850.299611289)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097850.26549267)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097853.834509722)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097849.83796387)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097852.23943339)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097855.442644808)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097852.748042987)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097849.211525232)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097849.247871725)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097850.276150185)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097853.657519596)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097852.096084442)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097854.428631407)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097850.539027278)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097852.350698698)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097852.382352644)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097852.111456317)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097854.79760621)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097853.629799038)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097851.509510029)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097855.283194821)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097855.808667804)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097853.836730346)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097854.075780679)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097855.363805775)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097855.630047068)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097857.102082277)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097853.909706147)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097855.890653441)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097856.10382463)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097856.186810645)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097855.345957343)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097855.791168777)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097856.999757397)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097856.314303509)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097855.868575213)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097856.228649749)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097855.889560086)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097857.045686942)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097856.818040638)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097856.888872269)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097857.302836703)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097859.084565545)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097857.288534215)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097858.256534111)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097858.099031607)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097857.996719193)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097858.66617556)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097857.980043389)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097859.819751575)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097858.795130084)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097858.239461296)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097859.730293498)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097860.204555718)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097861.225722959)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097860.020551022)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097859.95826388)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097859.301772618)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097857.846771972)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097858.110055653)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097859.570689938)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097859.998577226)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097858.435285744)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097860.209946694)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097860.620639623)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097860.357121175)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097861.9301507)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097860.743772687)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097860.663964751)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097861.672690192)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097861.346995303)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097862.314451428)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097861.862204424)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097860.127960598)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097860.650864182)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097861.997780359)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097862.15785438)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097860.144172297)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097863.309652461)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097862.560299053)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097862.939730186)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097862.90303589)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097863.128209301)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097862.604931206)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097863.129563341)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097863.292358394)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097863.58056572)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097864.614512611)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097863.613313539)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097864.178265661)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097863.710142228)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097862.321544093)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097864.140230047)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097862.36956793)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097864.494040277)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097864.399788021)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097865.51054235)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097865.466493838)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097865.92737622)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097865.275567845)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097865.596885756)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097865.51272637)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097864.999174243)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097865.642803609)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097865.054893836)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097866.178985922)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097865.56930824)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097864.423784623)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097867.019864838)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097866.07632669)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097866.428825801)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097866.132346928)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097866.758935165)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097867.624160658)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097866.616463076)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097867.281974226)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097867.293262127)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097867.654579302)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097868.078868812)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097867.784502369)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097868.120400267)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097867.771922164)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097867.947410315)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097867.413675058)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097868.540194821)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097869.352440384)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097869.000887383)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097868.434679525)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097868.422616867)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097869.107384786)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097868.46320146)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097870.411385041)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097869.121019023)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097869.278797005)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097870.062733117)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097870.172425537)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097870.440700805)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097869.249871968)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097869.557350148)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097867.72994388)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097870.300542813)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097870.346226573)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097871.38802437)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097871.950665145)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097871.182420742)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097870.469462982)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097870.645632554)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097871.7664542)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097871.221233068)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097871.517351977)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097871.991139406)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097872.06998248)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097872.096046674)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097870.228500541)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097871.828599079)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097869.85193098)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097872.794316756)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097873.000634778)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097873.008110175)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097872.764034967)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097872.858759834)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097872.907224625)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097872.323781666)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097873.788005623)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097875.300230364)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097873.676702828)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097875.378574087)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097873.565295704)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097874.628851057)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097872.228908388)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097872.036532252)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097874.572608482)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097875.273336632)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097874.391041793)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097874.671099855)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097874.612291153)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097875.162004201)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097875.412473782)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097875.672308803)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097875.267705096)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097875.695475729)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097874.571381408)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097875.315188769)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097873.810748198)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097876.551613175)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097875.880025551)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097873.976133848)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097876.987918886)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097876.729286779)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097877.963467198)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097876.687785984)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097877.19458723)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097876.776342171)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097878.251358245)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097877.519493958)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097878.237836255)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097878.352248108)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097876.595880651)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097878.107752016)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097878.353400647)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097877.421553358)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097877.952663979)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097878.359994924)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097879.679549268)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097880.357349736)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097878.854213666)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097879.40249699)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097879.601399188)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097879.369133447)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097879.944833959)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097879.5373167)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097880.616513322)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097880.681496077)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097877.59905177)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097877.866081629)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097879.971558991)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097880.487265755)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097880.50647224)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097880.313189148)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097881.197458318)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097880.427121525)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097880.564537974)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097880.580596236)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097881.420564548)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097881.61489926)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097881.531922796)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097883.054112423)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097882.447401525)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097881.799443459)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097880.463084735)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097883.277219187)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097879.913791377)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097883.035694951)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097883.974782652)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097883.0740836)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097882.600567242)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097882.439591576)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097882.599444962)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097882.629286171)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097882.521792368)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097884.325074267)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097883.041012093)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097882.843040924)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097883.941240661)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097884.607161414)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097885.069547613)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097883.822098823)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097885.595136721)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097883.84175177)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097886.0085586)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097884.382032469)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097884.875508539)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097885.785287055)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097885.075864121)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097884.832624758)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097884.640816679)])']\n", "connector: \n", "Evaluating workflow: 36%|███▌ | 1084/3000 [03:19<05:02, 6.34it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 36%|███▌ | 1087/3000 [03:20<04:27, 7.14it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 36%|███▋ | 1091/3000 [03:20<02:57, 10.73it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 36%|███▋ | 1093/3000 [03:20<03:02, 10.44it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 37%|███▋ | 1097/3000 [03:20<03:02, 10.41it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 37%|███▋ | 1099/3000 [03:21<03:33, 8.90it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 37%|███▋ | 1101/3000 [03:21<04:52, 6.50it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 37%|███▋ | 1103/3000 [03:21<04:26, 7.11it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 37%|███▋ | 1106/3000 [03:22<04:03, 7.78it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 37%|███▋ | 1108/3000 [03:22<03:23, 9.30it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 37%|███▋ | 1112/3000 [03:22<02:57, 10.65it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 37%|███▋ | 1114/3000 [03:23<03:47, 8.28it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 37%|███▋ | 1116/3000 [03:23<04:35, 6.83it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 37%|███▋ | 1118/3000 [03:23<05:36, 5.59it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 37%|███▋ | 1120/3000 [03:24<04:18, 7.28it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 37%|███▋ | 1122/3000 [03:24<04:30, 6.93it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 37%|███▋ | 1124/3000 [03:24<04:31, 6.91it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 38%|███▊ | 1126/3000 [03:24<03:39, 8.52it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 38%|███▊ | 1128/3000 [03:25<03:47, 8.22it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 38%|███▊ | 1134/3000 [03:26<04:15, 7.29it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 38%|███▊ | 1136/3000 [03:26<04:08, 7.50it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 38%|███▊ | 1138/3000 [03:26<03:49, 8.10it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 38%|███▊ | 1140/3000 [03:26<03:51, 8.05it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 38%|███▊ | 1142/3000 [03:27<03:42, 8.35it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 38%|███▊ | 1145/3000 [03:27<02:50, 10.89it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 38%|███▊ | 1147/3000 [03:27<03:10, 9.70it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 38%|███▊ | 1149/3000 [03:27<03:20, 9.23it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 38%|███▊ | 1151/3000 [03:28<05:11, 5.94it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 39%|███▊ | 1157/3000 [03:28<04:08, 7.43it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 39%|███▊ | 1161/3000 [03:29<03:11, 9.61it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 39%|███▉ | 1163/3000 [03:29<03:35, 8.51it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 39%|███▉ | 1167/3000 [03:30<03:26, 8.89it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 39%|███▉ | 1169/3000 [03:30<04:12, 7.26it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 39%|███▉ | 1170/3000 [03:30<05:01, 6.07it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 39%|███▉ | 1172/3000 [03:31<05:22, 5.67it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 39%|███▉ | 1176/3000 [03:31<03:46, 8.06it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 39%|███▉ | 1180/3000 [03:31<03:08, 9.66it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 39%|███▉ | 1182/3000 [03:32<03:30, 8.63it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 40%|███▉ | 1185/3000 [03:32<03:00, 10.06it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 40%|███▉ | 1189/3000 [03:33<05:03, 5.98it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 40%|███▉ | 1191/3000 [03:33<05:30, 5.48it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 40%|███▉ | 1193/3000 [03:34<04:54, 6.14it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 40%|███▉ | 1197/3000 [03:34<03:10, 9.49it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 40%|███▉ | 1199/3000 [03:34<03:10, 9.44it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 40%|████ | 1203/3000 [03:34<02:53, 10.35it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 40%|████ | 1205/3000 [03:35<03:12, 9.33it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 40%|████ | 1207/3000 [03:35<05:40, 5.27it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 40%|████ | 1210/3000 [03:36<04:37, 6.46it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 40%|████ | 1212/3000 [03:36<04:37, 6.45it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 40%|████ | 1213/3000 [03:36<04:53, 6.08it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 41%|████ | 1218/3000 [03:37<03:24, 8.70it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 41%|████ | 1220/3000 [03:37<02:57, 10.03it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 41%|████ | 1223/3000 [03:37<03:10, 9.33it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:36:33.157\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:36:33.220\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 41%|████ | 1225/3000 [03:44<29:36, 1.00s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:36:33.260\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:36:34.382\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0292965', 'INHIB_AVE': 14.04, 'SMILES': 'N1(O)C(C)(C)CC(CC1(C)C)NC(c2ccccc2)=O.Cl', 'HIT': 0.0, 'question': 'Is Escherichia coli susceptible to treatment with the compound N1(O)C(C)(C)CC(CC1(C)C)NC(c2ccccc2)=O.Cl?', 'answer': 'No', '_id': 8978}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 7 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 41%|████ | 1227/3000 [03:45<26:06, 1.13it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:36:34.392\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0218951', 'INHIB_AVE': 4.97, 'SMILES': 'C(C(c1ccc(cc1)Cl)Nc2cccc(Br)c2)C(=O)c3ccc(cc3)C', 'HIT': 0.0, 'question': 'Can C(C(c1ccc(cc1)Cl)Nc2cccc(Br)c2)C(=O)c3ccc(cc3)C act as an antibacterial agent targeting Escherichia coli?', 'answer': 'No', '_id': 35761}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 7 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:36:34.478\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0155122', 'INHIB_AVE': 7.02, 'SMILES': 'O=C(C(C=C1)=CC=C1F)C2=C(N)C(C=C3C)=C(N=C3CC)S2', 'HIT': 0.0, 'question': 'Does O=C(C(C=C1)=CC=C1F)C2=C(N)C(C=C3C)=C(N=C3CC)S2 demonstrate bactericidal effects against E. coli in vitro?', 'answer': 'No', '_id': 26914}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 7 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:36:34.554\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0300046', 'INHIB_AVE': -10.32, 'SMILES': 'C(C#N)(C(CC(N1)=O)c(ccc(c2OCc3ccccc3)OC)c2)=C1SCC', 'HIT': 0.0, 'question': 'Does C(C#N)(C(CC(N1)=O)c(ccc(c2OCc3ccccc3)OC)c2)=C1SCC show antimicrobial efficacy specifically against Escherichia coli strains?', 'answer': 'No', '_id': 78930}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 6 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 41%|████ | 1230/3000 [03:45<16:57, 1.74it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:36:34.651\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0166775', 'INHIB_AVE': 14.79, 'SMILES': 'N1(c2ccc(cc2)Br)C(=O)C(CC1=O)SC(N\\\\N=C\\\\c(ccc(c3O)O)c3)=N', 'HIT': 0.0, 'question': 'Can N1(c2ccc(cc2)Br)C(=O)C(CC1=O)SC(N\\\\N=C\\\\c(ccc(c3O)O)c3)=N effectively kill or suppress Escherichia coli cells?', 'answer': 'No', '_id': 7868}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 6 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:36:34.721\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0231851', 'INHIB_AVE': 5.02, 'SMILES': 'c1(c2ccccc2C)oc(c3n1)ccc(c3)\\\\N=C\\\\c(ccc(c4Cl)OC)c4', 'HIT': 0.0, 'question': 'Does c1(c2ccccc2C)oc(c3n1)ccc(c3)\\\\N=C\\\\c(ccc(c4Cl)OC)c4 show antimicrobial efficacy specifically against Escherichia coli strains?', 'answer': 'No', '_id': 35540}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 6 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 41%|████ | 1232/3000 [03:46<13:08, 2.24it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:36:34.745\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0225305', 'INHIB_AVE': 10.45, 'SMILES': 'N12C(C=CC(SCc(cc3)ccc3Cl)=N1)=NN(CC(=O)Nc4cccc(OC)c4)C2=O', 'HIT': 0.0, 'question': 'Can N12C(C=CC(SCc(cc3)ccc3Cl)=N1)=NN(CC(=O)Nc4cccc(OC)c4)C2=O act as an antibacterial agent targeting Escherichia coli?', 'answer': 'No', '_id': 15855}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 6 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:36:34.866\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 6 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:36:34.872\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 6 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 41%|████ | 1234/3000 [03:46<10:08, 2.90it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:36:34.899\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 6 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:36:34.934\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0221600', 'INHIB_AVE': 14.62, 'SMILES': 'N1(O)C(=O)CNC1c(cccc2[N+](=O)[O-])c2', 'HIT': 0.0, 'question': 'Is N1(O)C(=O)CNC1c(cccc2[N+](=O)[O-])c2 effective in eliminating E. coli under experimental conditions?', 'answer': 'No', '_id': 8096}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 6 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:36:35.083\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0304632', 'INHIB_AVE': 3.29, 'SMILES': 'c1(NCc(cc2)ccc2C)nnnn1CCC', 'HIT': 0.0, 'question': 'Can exposure to c1(NCc(cc2)ccc2C)nnnn1CCC lead to reduced viability of E. coli?', 'answer': 'No', '_id': 43308}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 6 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 41%|████▏ | 1239/3000 [03:46<05:44, 5.11it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:36:35.094\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0285338', 'INHIB_AVE': -0.17, 'SMILES': 'c12c(n(c(SCC(OCC)=O)n1)C)C(N(C)C(=O)N2C)=O', 'HIT': 0.0, 'question': 'Does treatment with c12c(n(c(SCC(OCC)=O)n1)C)C(N(C)C(=O)N2C)=O result in bacteriostatic or bactericidal effects on E. coli?', 'answer': 'No', '_id': 58150}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 6 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:36:40.920\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0311738', 'INHIB_AVE': 2.23, 'SMILES': 'c1(NC(=O)c2ccccc2[N+](=O)[O-])nnnn1\\\\N=C\\\\c(cc3)ccc3F', 'HIT': 0.0, 'question': 'Can exposure to c1(NC(=O)c2ccccc2[N+](=O)[O-])nnnn1\\\\N=C\\\\c(cc3)ccc3F lead to reduced viability of E. coli?', 'answer': 'No', '_id': 48082}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 41%|████▏ | 1241/3000 [03:52<23:17, 1.26it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:36:41.103\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0231843', 'INHIB_AVE': 10.09, 'SMILES': 'c1(c2ccccc2C)oc(c3n1)ccc(c3)\\\\N=C\\\\c4ccccc4O', 'HIT': 0.0, 'question': 'Can c1(c2ccccc2C)oc(c3n1)ccc(c3)\\\\N=C\\\\c4ccccc4O effectively kill or suppress Escherichia coli cells?', 'answer': 'No', '_id': 16769}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:36:41.390\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 3 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 41%|████▏ | 1243/3000 [03:52<19:21, 1.51it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:36:44.080\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:36:44.084\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 6 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 41%|████▏ | 1244/3000 [03:55<28:01, 1.04it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:36:44.212\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0249897', 'INHIB_AVE': 4.78, 'SMILES': 'c1(nnc(COc2ccccc2)s1)NC(=O)c(cc3)ccc3C(C)(C)C', 'HIT': 0.0, 'question': 'Can c1(nnc(COc2ccccc2)s1)NC(=O)c(cc3)ccc3C(C)(C)C act as an antibacterial agent targeting Escherichia coli?', 'answer': 'No', '_id': 36654}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 6 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 42%|████▏ | 1246/3000 [03:55<20:18, 1.44it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:36:44.223\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:36:44.257\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:36:44.288\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0245359', 'INHIB_AVE': -0.99, 'SMILES': 'N1(c2ccccc2)C(=O)c(c3C1=O)ccc(c3)N4Cc5c(cccc5)C4', 'HIT': 0.0, 'question': 'Can exposure to N1(c2ccccc2)C(=O)c(c3C1=O)ccc(c3)N4Cc5c(cccc5)C4 lead to reduced viability of E. coli?', 'answer': 'No', '_id': 61083}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:36:44.522\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 42%|████▏ | 1250/3000 [03:55<12:02, 2.42it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:36:44.532\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0215608', 'INHIB_AVE': 8.11, 'SMILES': 'S(c1ccccc1)(=O)(=O)N(c2ccc(c(Cl)c2)Cl)CC(=O)N\\\\N=C\\\\c(c(OC)ccc3[N+](=O)[O-])c3', 'HIT': 0.0, 'question': 'Can exposure to S(c1ccccc1)(=O)(=O)N(c2ccc(c(Cl)c2)Cl)CC(=O)N\\\\N=C\\\\c(c(OC)ccc3[N+](=O)[O-])c3 lead to reduced viability of E. coli?', 'answer': 'No', '_id': 22818}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:36:44.537\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:36:49.530\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 42%|████▏ | 1253/3000 [04:00<24:02, 1.21it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 42%|████▏ | 1256/3000 [04:01<16:16, 1.79it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2026-01-10 17:36:50.117\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0235287', 'INHIB_AVE': 4.8, 'SMILES': 'C(=O)(c1ccc(cc1)Br)Oc2cc(ccc2OC(=O)c3ccc(cc3)Br)\\\\C=N\\\\NC(c4ccccc4)=O', 'HIT': 0.0, 'question': 'Is C(=O)(c1ccc(cc1)Br)Oc2cc(ccc2OC(=O)c3ccc(cc3)Br)\\\\C=N\\\\NC(c4ccccc4)=O effective in eliminating E. coli under experimental conditions?', 'answer': 'No', '_id': 36527}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 42%|████▏ | 1257/3000 [04:01<14:28, 2.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2026-01-10 17:36:50.226\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:36:50.227\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0157623', 'INHIB_AVE': 14.59, 'SMILES': 'O=C(C1=CC=CO1)NC2=C(C(OCC)=O)SC3=C2C(C4=CC=CC=C4)=CC(C5=CC=CC=C5)=N3', 'HIT': 0.0, 'question': 'Does O=C(C1=CC=CO1)NC2=C(C(OCC)=O)SC3=C2C(C4=CC=CC=C4)=CC(C5=CC=CC=C5)=N3 show antimicrobial efficacy specifically against Escherichia coli strains?', 'answer': 'No', '_id': 8131}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 42%|████▏ | 1259/3000 [04:01<10:18, 2.81it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2026-01-10 17:36:50.506\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0221456', 'INHIB_AVE': 10.02, 'SMILES': 'S(=O)(=O)(Nc(cc1)ccc1OCCCC)c2cccc3c2cccc3S(=O)(=O)Nc(cc4)ccc4OCCCC', 'HIT': 0.0, 'question': 'Does treatment with S(=O)(=O)(Nc(cc1)ccc1OCCCC)c2cccc3c2cccc3S(=O)(=O)Nc(cc4)ccc4OCCCC result in bacteriostatic or bactericidal effects on E. coli?', 'answer': 'No', '_id': 16954}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 42%|████▏ | 1262/3000 [04:01<07:11, 4.03it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:36:50.670\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:36:50.679\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 42%|████▏ | 1264/3000 [04:01<05:52, 4.92it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2026-01-10 17:36:50.723\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 42%|████▏ | 1268/3000 [04:02<05:57, 4.84it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2026-01-10 17:36:51.914\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 42%|████▏ | 1269/3000 [04:03<06:43, 4.29it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 42%|████▏ | 1272/3000 [04:03<05:20, 5.39it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 42%|████▏ | 1274/3000 [04:04<07:41, 3.74it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 43%|████▎ | 1277/3000 [04:04<05:55, 4.85it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 43%|████▎ | 1280/3000 [04:05<03:36, 7.94it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 43%|████▎ | 1282/3000 [04:05<03:31, 8.12it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 43%|████▎ | 1284/3000 [04:05<03:26, 8.31it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 43%|████▎ | 1286/3000 [04:05<03:30, 8.14it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 43%|████▎ | 1288/3000 [04:06<03:45, 7.61it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 43%|████▎ | 1290/3000 [04:06<03:33, 8.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 43%|████▎ | 1294/3000 [04:06<03:23, 8.38it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 43%|████▎ | 1298/3000 [04:07<03:37, 7.81it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 43%|████▎ | 1300/3000 [04:07<03:57, 7.14it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 43%|████▎ | 1302/3000 [04:07<03:20, 8.46it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 44%|████▎ | 1305/3000 [04:08<04:07, 6.85it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 44%|████▎ | 1309/3000 [04:09<03:50, 7.34it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 44%|████▎ | 1311/3000 [04:09<03:09, 8.93it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 44%|████▍ | 1313/3000 [04:09<03:20, 8.40it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 44%|████▍ | 1317/3000 [04:10<04:07, 6.79it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 44%|████▍ | 1321/3000 [04:10<02:48, 9.98it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 44%|████▍ | 1323/3000 [04:10<03:30, 7.98it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 44%|████▍ | 1326/3000 [04:11<05:23, 5.17it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 44%|████▍ | 1328/3000 [04:11<04:46, 5.84it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 44%|████▍ | 1332/3000 [04:12<03:46, 7.36it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 45%|████▍ | 1336/3000 [04:12<03:43, 7.43it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 45%|████▍ | 1340/3000 [04:13<02:45, 10.05it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 45%|████▍ | 1342/3000 [04:13<03:49, 7.24it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 45%|████▍ | 1346/3000 [04:14<03:38, 7.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 45%|████▍ | 1348/3000 [04:14<03:12, 8.57it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 45%|████▌ | 1352/3000 [04:14<02:48, 9.80it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 45%|████▌ | 1355/3000 [04:14<02:11, 12.48it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 45%|████▌ | 1357/3000 [04:15<02:39, 10.33it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 45%|████▌ | 1359/3000 [04:15<04:10, 6.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 45%|████▌ | 1363/3000 [04:16<03:04, 8.87it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 46%|████▌ | 1369/3000 [04:16<02:49, 9.64it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 46%|████▌ | 1371/3000 [04:17<03:10, 8.55it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 46%|████▌ | 1375/3000 [04:17<03:08, 8.61it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 46%|████▌ | 1377/3000 [04:17<03:16, 8.28it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 46%|████▌ | 1379/3000 [04:18<03:08, 8.61it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 46%|████▌ | 1380/3000 [04:18<03:34, 7.55it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 46%|████▌ | 1385/3000 [04:18<03:17, 8.18it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 46%|████▌ | 1387/3000 [04:19<02:52, 9.36it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 46%|████▋ | 1392/3000 [04:19<02:26, 10.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 46%|████▋ | 1394/3000 [04:20<03:55, 6.81it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 47%|████▋ | 1396/3000 [04:20<03:48, 7.02it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 47%|████▋ | 1398/3000 [04:20<03:36, 7.40it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 47%|████▋ | 1402/3000 [04:21<03:39, 7.27it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 47%|████▋ | 1405/3000 [04:21<02:55, 9.06it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 47%|████▋ | 1409/3000 [04:21<03:04, 8.65it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 47%|████▋ | 1411/3000 [04:22<03:38, 7.27it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 47%|████▋ | 1413/3000 [04:22<04:06, 6.43it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 47%|████▋ | 1414/3000 [04:23<04:54, 5.39it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 47%|████▋ | 1419/3000 [04:23<02:58, 8.86it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 47%|████▋ | 1421/3000 [04:23<03:05, 8.53it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 48%|████▊ | 1425/3000 [04:24<02:31, 10.39it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 48%|████▊ | 1427/3000 [04:24<03:05, 8.50it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 48%|████▊ | 1431/3000 [04:24<02:51, 9.16it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 48%|████▊ | 1433/3000 [04:25<03:36, 7.24it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 48%|████▊ | 1435/3000 [04:25<03:02, 8.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 48%|████▊ | 1437/3000 [04:25<03:59, 6.54it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 48%|████▊ | 1439/3000 [04:26<03:46, 6.88it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 48%|████▊ | 1443/3000 [04:26<02:58, 8.73it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 48%|████▊ | 1445/3000 [04:26<03:39, 7.09it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 48%|████▊ | 1447/3000 [04:26<03:10, 8.16it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 48%|████▊ | 1451/3000 [04:27<02:23, 10.79it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 48%|████▊ | 1453/3000 [04:27<04:06, 6.27it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 49%|████▊ | 1457/3000 [04:28<04:08, 6.20it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 49%|████▊ | 1461/3000 [04:28<03:22, 7.59it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 49%|████▉ | 1463/3000 [04:29<03:38, 7.04it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 49%|████▉ | 1469/3000 [04:29<02:14, 11.39it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 49%|████▉ | 1471/3000 [04:29<02:08, 11.90it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 49%|████▉ | 1475/3000 [04:30<02:42, 9.36it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 49%|████▉ | 1477/3000 [04:30<03:30, 7.24it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 49%|████▉ | 1478/3000 [04:31<04:02, 6.28it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 49%|████▉ | 1484/3000 [04:31<03:22, 7.48it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 50%|████▉ | 1487/3000 [04:32<02:46, 9.08it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 50%|████▉ | 1490/3000 [04:32<03:26, 7.30it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 50%|████▉ | 1492/3000 [04:33<03:56, 6.37it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 50%|████▉ | 1495/3000 [04:33<02:53, 8.68it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 50%|████▉ | 1497/3000 [04:33<03:59, 6.29it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 50%|█████ | 1500/3000 [04:34<03:30, 7.13it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 50%|█████ | 1503/3000 [04:34<02:24, 10.35it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 50%|█████ | 1506/3000 [04:34<03:15, 7.66it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 50%|█████ | 1510/3000 [04:35<02:59, 8.31it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 50%|█████ | 1513/3000 [04:35<02:52, 8.64it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 51%|█████ | 1516/3000 [04:36<04:15, 5.82it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 51%|█████ | 1518/3000 [04:36<03:26, 7.19it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 51%|█████ | 1521/3000 [04:37<03:38, 6.76it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 51%|█████ | 1524/3000 [04:37<02:51, 8.63it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 51%|█████ | 1527/3000 [04:37<03:09, 7.77it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2026-01-10 17:37:34.656\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 10 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:37:34.660\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 10 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 51%|█████ | 1530/3000 [04:45<25:30, 1.04s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 51%|█████ | 1534/3000 [04:46<16:19, 1.50it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 51%|█████ | 1535/3000 [04:46<14:56, 1.63it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:37:44.269\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0295457', 'INHIB_AVE': -8.28, 'SMILES': 'c1(Sc2cc(ccc2[N+](=O)[O-])N(CC3)CCO3)nncn1C', 'HIT': 0.0, 'question': 'Can c1(Sc2cc(ccc2[N+](=O)[O-])N(CC3)CCO3)nncn1C effectively kill or suppress Escherichia coli cells?', 'answer': 'No', '_id': 77048}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 6 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 51%|█████ | 1536/3000 [04:55<45:46, 1.88s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:37:44.422\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0301377', 'INHIB_AVE': -0.06, 'SMILES': 'C1(C2CC(C=C2)N1C(=O)c3ccc(cc3)Cl)(C(F)(F)F)C(F)(F)F', 'HIT': 0.0, 'question': 'Can exposure to C1(C2CC(C=C2)N1C(=O)c3ccc(cc3)Cl)(C(F)(F)F)C(F)(F)F lead to reduced viability of E. coli?', 'answer': 'No', '_id': 57702}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 51%|█████ | 1537/3000 [04:55<38:17, 1.57s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:37:44.436\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0149061', 'INHIB_AVE': -3.79, 'SMILES': 'c1(c2c([nH]c1)cccc2)C(=O)C(=O)Nc3cccc(F)c3', 'HIT': 0.0, 'question': 'Is c1(c2c([nH]c1)cccc2)C(=O)C(=O)Nc3cccc(F)c3 effective in eliminating E. coli under experimental conditions?', 'answer': 'No', '_id': 69264}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:37:44.436\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0331833', 'INHIB_AVE': 0.81, 'SMILES': 'C(C#N)(C1=O)=C(c2ccccc2)N=C(SCC(=O)Nc3ccccc3)N1', 'HIT': 0.0, 'question': 'Does the proposed molecule C(C#N)(C1=O)=C(c2ccccc2)N=C(SCC(=O)Nc3ccccc3)N1 exhibit antibacterial activity against Escherichia coli?', 'answer': 'No', '_id': 54247}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:37:44.501\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0164434', 'INHIB_AVE': 98.17, 'SMILES': 'CCN1CCN(C(=O)N[C@@H](C(=O)N[C@H]2[C@H]3SC(C)(C)[C@@H](N3C2=O)C([O-])=O)c2ccccc2)C(=O)C1=O.[Na+]', 'HIT': 1.0, 'question': 'Is Escherichia coli susceptible to treatment with the compound CCN1CCN(C(=O)N[C@@H](C(=O)N[C@H]2[C@H]3SC(C)(C)[C@@H](N3C2=O)C([O-])=O)c2ccccc2)C(=O)C1=O.[Na+]?', 'answer': 'Yes', '_id': 185}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:37:44.514\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0291078', 'INHIB_AVE': 2.41, 'SMILES': 'N1(c2ccc(cc2)Cl)C3(CCC(CC3)C(C)(C)C)Nc(c4C1=O)cccc4', 'HIT': 0.0, 'question': 'Can exposure to N1(c2ccc(cc2)Cl)C3(CCC(CC3)C(C)(C)C)Nc(c4C1=O)cccc4 lead to reduced viability of E. coli?', 'answer': 'No', '_id': 47258}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:37:44.644\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0321291', 'INHIB_AVE': -11.47, 'SMILES': 'C1(C2(C)C)=C(C(SS1)=S)c(c3N2C(=O)Nc(cc4)ccc4Cl)cc(cc3)C', 'HIT': 0.0, 'question': 'Does C1(C2(C)C)=C(C(SS1)=S)c(c3N2C(=O)Nc(cc4)ccc4Cl)cc(cc3)C show antimicrobial efficacy specifically against Escherichia coli strains?', 'answer': 'No', '_id': 79699}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 51%|█████▏ | 1542/3000 [04:55<17:24, 1.40it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:37:44.669\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0299818', 'INHIB_AVE': -10.77, 'SMILES': 'C(=C/c1cccs1)(/C(=O)N2C)\\\\C(=O)N(C2=S)C', 'HIT': 0.0, 'question': 'Can C(=C/c1cccs1)(/C(=O)N2C)\\\\C(=O)N(C2=S)C effectively kill or suppress Escherichia coli cells?', 'answer': 'No', '_id': 79255}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:37:44.834\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0332375', 'INHIB_AVE': 4.23, 'SMILES': 's1c(C)nnc1SCCC(=O)Nc(cc2)ccc2OC', 'HIT': 0.0, 'question': 'Does s1c(C)nnc1SCCC(=O)Nc(cc2)ccc2OC demonstrate bactericidal effects against E. coli in vitro?', 'answer': 'No', '_id': 39123}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 52%|█████▏ | 1546/3000 [04:56<11:01, 2.20it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:37:44.845\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0162369', 'INHIB_AVE': 0.21, 'SMILES': 'n1c2c(oc1c3ccc(cc3)I)ccc(c2)\\\\N=C\\\\c4cccc(I)c4', 'HIT': 0.0, 'question': 'Can n1c2c(oc1c3ccc(cc3)I)ccc(c2)\\\\N=C\\\\c4cccc(I)c4 act as an antibacterial agent targeting Escherichia coli?', 'answer': 'No', '_id': 56616}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:37:44.857\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0223652', 'INHIB_AVE': -1.15, 'SMILES': 'C(COc(cc1)ccc1CCCC)(=O)N\\\\N=C\\\\c2ccccc2Cl', 'HIT': 0.0, 'question': 'Is Escherichia coli susceptible to treatment with the compound C(COc(cc1)ccc1CCCC)(=O)N\\\\N=C\\\\c2ccccc2Cl?', 'answer': 'No', '_id': 61592}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 52%|█████▏ | 1549/3000 [05:01<20:46, 1.16it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:37:50.980\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 52%|█████▏ | 1550/3000 [05:02<20:13, 1.19it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:37:50.994\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:37:51.006\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 52%|█████▏ | 1555/3000 [05:03<12:09, 1.98it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 52%|█████▏ | 1556/3000 [05:03<11:35, 2.08it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 52%|█████▏ | 1557/3000 [05:03<10:40, 2.25it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 52%|█████▏ | 1560/3000 [05:03<07:01, 3.42it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 52%|█████▏ | 1564/3000 [05:05<06:14, 3.83it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 52%|█████▏ | 1567/3000 [05:05<04:14, 5.62it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 52%|█████▏ | 1571/3000 [05:05<02:33, 9.30it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 52%|█████▏ | 1573/3000 [05:05<03:38, 6.55it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 53%|█████▎ | 1577/3000 [05:06<03:32, 6.69it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 53%|█████▎ | 1580/3000 [05:06<02:50, 8.35it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 53%|█████▎ | 1584/3000 [05:07<02:43, 8.68it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 53%|█████▎ | 1586/3000 [05:07<02:25, 9.70it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097885.982132635)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097885.965633144)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097885.968017556)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097887.330713244)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097887.35195843)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097887.454721123)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097888.005601727)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097888.205939224)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097910.702390986)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097901.151592522)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097903.616946819)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097910.257923551)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097901.125170255)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097906.802592688)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097906.338611018)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097906.738889126)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097912.44836105)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097912.713638154)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097912.351807224)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097906.329744068)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097906.861283053)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097906.92483773)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097912.135011059)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097906.361361284)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097906.811768763)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097912.052359391)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097915.644258307)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097916.90440157)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097914.119738869)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097912.944422878)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097912.465803241)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097914.883433302)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097916.743290587)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097914.2134882)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097914.362802664)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097916.305609978)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097914.467502305)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097914.486916192)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097915.591245788)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097916.830436916)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097915.955411791)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097915.842587399)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097916.281896748)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097918.810044459)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097916.795582598)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097917.978860451)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097919.030156828)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097918.45266898)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097918.808862497)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097921.229440926)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097920.336805354)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097921.541097528)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097921.29244771)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097923.342343141)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097920.477432793)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097923.004735861)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097922.775168543)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097923.215259248)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097924.39120034)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097922.285409853)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097922.467202649)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097922.701163881)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097925.224087231)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097925.906805702)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097925.655431277)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097924.286067933)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097926.784541185)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097926.664644915)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097926.899810675)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097926.923923117)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097927.930993198)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097929.883382012)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097928.916993311)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097931.382081659)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097928.157311907)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097931.155421976)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097931.242666755)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097931.091230554)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097931.863705553)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097931.230640172)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097931.307084044)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097931.702083892)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097931.999634188)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097931.589741205)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097931.295269097)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097931.545012318)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097931.925058019)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097930.408786384)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097929.961193364)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097930.413025041)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097930.407330898)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097932.308031294)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097884.759030961)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097885.450548516)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097886.477249421)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097885.727891316)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097887.046095731)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097887.776312526)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097887.165267822)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097887.437809882)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097887.341618988)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097887.143061544)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097887.846444089)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097901.728283597)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097896.443224941)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097886.665625067)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097901.676442539)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097901.895696683)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097901.19246637)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097901.78135321)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097901.149573279)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097896.671098832)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097896.431947277)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097896.584808196)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097901.198119991)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097901.228529553)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097901.179626901)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097896.454774608)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097901.87983398)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097910.463762244)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097910.177392879)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097910.369083376)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097915.168375871)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097912.992008499)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097915.084418856)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097916.247066334)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097916.42433365)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097916.658432404)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097916.54656019)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097917.90941575)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097917.866540499)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097917.156520759)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097917.967319667)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097916.928401364)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097918.891779694)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097918.551695424)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097919.194251591)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097918.742516524)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097919.433958978)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097919.041274448)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097919.05488869)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097919.013179329)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097920.771719359)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097920.12882263)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097919.971555702)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097920.295170015)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097920.652924102)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097921.376461775)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097921.620420092)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097921.28833632)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097921.281477782)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097921.757203818)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097923.208481773)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097922.977490806)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097923.201998672)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097923.125050182)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097923.134441056)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097923.327645982)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097923.7683266)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097923.585551209)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097923.968439191)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097924.294349345)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097923.286586792)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097923.440475834)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097924.598698735)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097925.305686651)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097924.776012184)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097925.228385331)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097925.996588297)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097926.191836411)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097925.406311627)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097925.958965428)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097925.557369876)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097925.942558184)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097925.510380005)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097926.169882186)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097926.90512344)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097926.980645596)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097926.84021526)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097927.632460585)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097927.615766717)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097927.954682212)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097928.383696014)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097928.352560413)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097927.779158157)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097928.357233502)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097928.524264009)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097929.039599585)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097929.215691104)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097928.358424427)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097929.546269223)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097929.348781588)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097929.353473871)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097929.4049502)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097929.738525396)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097930.853522817)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097932.200352602)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097885.976063636)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097916.266674311)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097917.754851616)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097934.014739909)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097933.317590419)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097933.284062174)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097933.633339942)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097933.509068168)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097933.419127143)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097934.255700417)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097933.893180086)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097934.664522184)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097933.780108676)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097934.443733867)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097935.298465739)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097936.313362997)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097935.508326175)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097935.229790079)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097933.792364219)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097936.05397)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097932.342103311)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097932.824080205)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097932.841743725)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097933.040242732)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097933.544915233)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097935.167303798)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097935.625456298)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097935.916153297)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097936.312178632)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097937.245148695)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097937.191455661)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097935.934483927)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097936.088712782)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097936.365150583)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097937.398789212)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097937.637193337)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097937.264912785)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097938.322442277)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097935.718095504)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097934.974407577)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097937.321437029)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097938.443176248)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097938.103667175)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097938.73021915)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097938.369634177)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097937.867095774)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097938.230931217)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097938.013736692)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097938.236133368)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097938.882399206)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097939.297037003)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097938.63795691)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097940.061748087)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097940.468604819)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097937.566989704)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097939.549364032)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097939.524191566)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097937.761086059)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097940.126149475)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097940.834388864)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097940.333356197)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097940.249387382)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097941.645740551)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097941.111794781)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097941.177168251)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097942.5463971)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097940.539810497)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097940.479716763)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097940.211514475)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097940.785093394)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097941.335890936)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097939.619639425)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097939.576414615)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097941.940315305)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097942.690409011)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097941.956704892)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097942.340063493)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097941.247659115)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097942.668741634)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097943.625584565)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097942.421338638)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097942.912694487)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097943.541046484)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097943.367932137)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097942.615742659)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097943.306821206)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097943.030980178)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097943.318683293)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097944.519799519)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097945.32771177)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097944.656835495)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097945.356232778)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097945.493770247)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097945.706373528)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097945.097475749)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097944.596430975)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097945.377024605)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097945.167373956)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097944.941611184)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097945.686770595)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097945.341081636)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097942.939004472)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097943.355062778)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097946.057429074)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097945.955462939)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097946.32648595)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097946.157864556)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097947.160262715)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097946.807130546)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097947.414249345)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097947.120589236)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097948.411434888)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097947.94096982)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097947.740456142)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097945.95073701)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097948.599919675)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097947.19871896)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097945.472065318)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097948.412607977)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097948.251390799)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097947.619644667)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097948.896607255)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097948.268835417)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097957.160472872)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097948.224777761)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097948.539786203)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097948.591595741)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097956.890557685)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097956.986978731)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097957.442636774)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097956.90051111)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097957.229348621)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097947.953331072)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097957.253792527)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097957.602908637)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097957.592453743)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097957.062872958)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097957.376542229)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097957.407408263)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097966.79710569)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097963.428658634)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097955.768604642)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097955.729178867)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097963.611411924)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097957.46107736)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097957.372717822)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097955.665554682)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097966.72127761)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097967.032478277)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097963.899151141)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097972.62612153)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097972.296836076)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097967.028643478)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097972.731318888)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097967.04101575)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097972.639823811)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097966.587239235)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097966.765644867)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097972.813933129)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097972.356312397)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097973.014865039)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097973.205093949)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097972.442023661)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097966.592724293)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097976.150139881)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097966.732380552)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097972.039413218)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097974.810189688)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097975.559513348)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097973.177344845)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097974.071435663)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097973.94707732)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097974.525692001)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097973.232167839)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097973.182756507)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097972.729394058)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097975.697772381)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097974.658886744)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097976.217228199)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097976.360590599)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097976.190081747)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097975.982029715)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097978.028064003)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097976.740014424)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097974.422732595)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097975.761757443)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097976.277439374)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097976.524984739)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097976.984053641)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097976.997425107)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097978.078654306)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097976.512232651)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097977.068724752)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097977.518152407)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097978.4612615)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097977.974691204)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097977.299339554)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097977.304869081)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097978.913967983)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097979.052639848)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097979.128903598)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097979.719476318)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097978.995354623)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097981.669337593)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097980.077182475)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097979.195363263)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097979.705200211)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097980.282693339)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097980.253115587)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097980.115574755)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097978.65874486)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097978.111628174)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097981.576824707)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097978.069674046)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097979.576679026)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097978.615230842)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097981.287195188)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097982.086122304)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097982.020834505)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097981.639560539)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097981.301105385)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097981.689655177)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097980.66602306)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097982.126055338)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097983.733086387)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097980.371410442)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097980.46093569)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097983.646259834)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097982.968744196)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097981.423873713)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097980.640939864)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097984.06423745)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097983.463641365)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097984.285502628)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097983.190337826)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097983.038167933)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097984.183878443)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097983.743963061)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097984.897479726)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097983.559874043)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097984.895032085)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097984.437725572)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097984.460775847)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097985.452299381)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097984.217591438)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097982.062041532)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097983.547468805)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097984.306732495)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097985.606807067)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097985.446207524)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097985.296823109)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097985.662299754)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097986.37379335)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097985.860927521)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097985.995202655)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097986.12025332)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097985.982104259)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097986.403278616)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097985.893980762)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097987.168308898)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097985.323455562)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097986.889525069)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097986.995336397)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097987.275406224)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097987.086648759)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097987.337860889)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097987.116725089)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097987.884248642)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097987.800539868)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097989.044273621)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097988.306384648)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097988.278235112)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097988.640552435)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097987.84871395)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097987.963117368)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097987.869436802)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097988.850335126)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097988.655059233)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097985.989060515)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097990.15542492)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097989.211534192)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097990.025874294)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097990.134338595)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097989.465341231)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097989.999397046)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097989.25329239)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097989.530925278)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097990.269068972)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097990.697500307)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097990.720911073)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097990.067130522)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097990.618949834)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097988.778222482)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097988.241163059)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097990.635379309)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097990.539070642)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097991.816359456)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097991.325623635)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097991.321712948)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097991.058423619)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097991.728053542)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097992.867716575)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097992.451132773)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097991.583997628)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097992.283876818)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097992.390213219)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097992.342903615)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097993.192876875)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097992.650833856)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097992.534705262)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097993.579076276)])']\n", "connector: \n", "Evaluating workflow: 53%|█████▎ | 1592/3000 [05:08<03:54, 6.00it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 53%|█████▎ | 1599/3000 [05:09<01:58, 11.79it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 53%|█████▎ | 1602/3000 [05:09<02:55, 7.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 53%|█████▎ | 1604/3000 [05:10<04:39, 5.00it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 54%|█████▎ | 1608/3000 [05:11<03:38, 6.38it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 54%|█████▍ | 1613/3000 [05:11<02:28, 9.34it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 54%|█████▍ | 1615/3000 [05:11<02:25, 9.55it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 54%|█████▍ | 1617/3000 [05:12<03:02, 7.58it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 54%|█████▍ | 1619/3000 [05:12<03:51, 5.96it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 54%|█████▍ | 1621/3000 [05:12<03:32, 6.50it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 54%|█████▍ | 1625/3000 [05:13<02:24, 9.51it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 54%|█████▍ | 1627/3000 [05:13<03:46, 6.06it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 54%|█████▍ | 1630/3000 [05:14<03:10, 7.20it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 54%|█████▍ | 1631/3000 [05:14<03:18, 6.90it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 54%|█████▍ | 1632/3000 [05:14<03:40, 6.21it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 55%|█████▍ | 1636/3000 [05:14<02:45, 8.23it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 55%|█████▍ | 1638/3000 [05:15<02:42, 8.40it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 55%|█████▍ | 1641/3000 [05:15<03:02, 7.46it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 55%|█████▍ | 1646/3000 [05:15<01:53, 11.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 55%|█████▍ | 1649/3000 [05:16<02:31, 8.91it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 55%|█████▌ | 1652/3000 [05:16<03:23, 6.64it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 55%|█████▌ | 1657/3000 [05:17<02:28, 9.04it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 55%|█████▌ | 1659/3000 [05:17<02:10, 10.24it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 55%|█████▌ | 1663/3000 [05:17<02:12, 10.07it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 56%|█████▌ | 1666/3000 [05:18<01:59, 11.20it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 56%|█████▌ | 1668/3000 [05:18<01:47, 12.38it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 56%|█████▌ | 1670/3000 [05:18<03:34, 6.21it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 56%|█████▌ | 1674/3000 [05:19<02:52, 7.71it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 56%|█████▌ | 1676/3000 [05:19<02:22, 9.31it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 56%|█████▌ | 1681/3000 [05:19<02:00, 10.93it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 56%|█████▌ | 1683/3000 [05:20<01:57, 11.17it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 56%|█████▌ | 1687/3000 [05:21<03:46, 5.79it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 56%|█████▋ | 1691/3000 [05:21<02:43, 8.02it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 57%|█████▋ | 1696/3000 [05:22<02:14, 9.72it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 57%|█████▋ | 1699/3000 [05:22<01:43, 12.63it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 57%|█████▋ | 1701/3000 [05:22<03:11, 6.79it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 57%|█████▋ | 1705/3000 [05:23<03:09, 6.84it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 57%|█████▋ | 1709/3000 [05:23<02:49, 7.64it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 57%|█████▋ | 1713/3000 [05:24<02:16, 9.41it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 57%|█████▋ | 1715/3000 [05:24<02:09, 9.89it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 57%|█████▋ | 1717/3000 [05:25<03:48, 5.63it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 57%|█████▋ | 1719/3000 [05:25<03:50, 5.55it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 57%|█████▋ | 1722/3000 [05:25<03:02, 7.00it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 57%|█████▊ | 1725/3000 [05:25<02:01, 10.47it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 58%|█████▊ | 1727/3000 [05:26<02:10, 9.77it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 58%|█████▊ | 1729/3000 [05:26<02:09, 9.80it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 58%|█████▊ | 1731/3000 [05:26<02:56, 7.21it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 58%|█████▊ | 1732/3000 [05:27<03:19, 6.35it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 58%|█████▊ | 1733/3000 [05:27<03:32, 5.98it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 58%|█████▊ | 1737/3000 [05:27<03:01, 6.95it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 58%|█████▊ | 1740/3000 [05:28<03:02, 6.90it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 58%|█████▊ | 1744/3000 [05:28<02:16, 9.20it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 58%|█████▊ | 1747/3000 [05:28<02:09, 9.66it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 58%|█████▊ | 1749/3000 [05:29<02:09, 9.68it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 58%|█████▊ | 1751/3000 [05:29<02:53, 7.22it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 58%|█████▊ | 1753/3000 [05:30<03:36, 5.76it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 58%|█████▊ | 1755/3000 [05:30<02:54, 7.15it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 59%|█████▊ | 1757/3000 [05:30<02:35, 7.98it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 59%|█████▊ | 1759/3000 [05:30<03:13, 6.43it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 59%|█████▉ | 1763/3000 [05:31<02:07, 9.70it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 59%|█████▉ | 1765/3000 [05:31<02:44, 7.53it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 59%|█████▉ | 1768/3000 [05:31<02:17, 8.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 59%|█████▉ | 1772/3000 [05:32<02:01, 10.14it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 59%|█████▉ | 1774/3000 [05:32<02:27, 8.32it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 59%|█████▉ | 1775/3000 [05:32<03:01, 6.74it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 59%|█████▉ | 1776/3000 [05:33<03:34, 5.71it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 59%|█████▉ | 1777/3000 [05:33<04:02, 5.05it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 59%|█████▉ | 1781/3000 [05:33<02:31, 8.07it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 59%|█████▉ | 1783/3000 [05:33<02:10, 9.32it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 60%|█████▉ | 1785/3000 [05:34<02:16, 8.89it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 60%|█████▉ | 1787/3000 [05:34<02:21, 8.57it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 60%|█████▉ | 1790/3000 [05:34<02:03, 9.77it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 60%|█████▉ | 1792/3000 [05:35<03:12, 6.27it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 60%|█████▉ | 1795/3000 [05:35<03:15, 6.16it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 60%|█████▉ | 1798/3000 [05:35<02:24, 8.29it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 60%|██████ | 1800/3000 [05:36<02:18, 8.64it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 60%|██████ | 1802/3000 [05:36<02:15, 8.82it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 60%|██████ | 1807/3000 [05:36<01:51, 10.65it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 60%|██████ | 1809/3000 [05:37<01:59, 9.98it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 60%|██████ | 1811/3000 [05:37<02:46, 7.13it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 60%|██████ | 1812/3000 [05:37<02:57, 6.70it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 60%|██████ | 1813/3000 [05:38<04:17, 4.61it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 61%|██████ | 1816/3000 [05:38<03:03, 6.46it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2026-01-10 17:38:34.707\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0295442', 'INHIB_AVE': 5.64, 'SMILES': 'c1(cc(ccc1Cl)NC(Cc(cc2)ccc2F)=O)C(F)(F)F', 'HIT': 0.0, 'question': 'Is Escherichia coli susceptible to treatment with the compound c1(cc(ccc1Cl)NC(Cc(cc2)ccc2F)=O)C(F)(F)F?', 'answer': 'No', '_id': 32735}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 10 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 61%|██████ | 1817/3000 [05:45<29:16, 1.48s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:38:34.757\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0304080', 'INHIB_AVE': 39.07, 'SMILES': 'c1(CC(NCCO)=O)n(C)c2c(cc(cc2)C)c1Sc3ccccc3', 'HIT': 0.0, 'question': 'Can c1(CC(NCCO)=O)n(C)c2c(cc(cc2)C)c1Sc3ccccc3 effectively kill or suppress Escherichia coli cells?', 'answer': 'No', '_id': 560}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 10 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:38:34.912\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0324735', 'INHIB_AVE': -10.94, 'SMILES': 'C12=C(CC(C)(C)CC1=O)Nc(c3NC2c4ccccc4Cl)cc(c(C)c3)C', 'HIT': 0.0, 'question': 'Does the proposed molecule C12=C(CC(C)(C)CC1=O)Nc(c3NC2c4ccccc4Cl)cc(c(C)c3)C exhibit antibacterial activity against Escherichia coli?', 'answer': 'No', '_id': 79390}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 10 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 61%|██████ | 1819/3000 [05:46<20:01, 1.02s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:38:35.092\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 10 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 61%|██████ | 1820/3000 [05:46<16:48, 1.17it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 61%|██████ | 1823/3000 [05:46<09:26, 2.08it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2026-01-10 17:38:35.253\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 9 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:38:44.722\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0258117', 'INHIB_AVE': -0.28, 'SMILES': 'C(=N/c1cccc(C)c1)(/S\\\\C(\\\\C2=O)=C\\\\c(ccc(c3OC)OCc4cccc(C)c4)c3)\\\\N2', 'HIT': 0.0, 'question': 'Does treatment with C(=N/c1cccc(C)c1)(/S\\\\C(\\\\C2=O)=C\\\\c(ccc(c3OC)OCc4cccc(C)c4)c3)\\\\N2 result in bacteriostatic or bactericidal effects on E. coli?', 'answer': 'No', '_id': 58541}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 61%|██████ | 1826/3000 [05:56<30:10, 1.54s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:38:44.822\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0218560', 'INHIB_AVE': 8.37, 'SMILES': 'C(OCCCCCCCCC)(NCCCCNC(OCCCCCCCCC)=O)=O', 'HIT': 0.0, 'question': 'Does C(OCCCCCCCCC)(NCCCCNC(OCCCCCCCCC)=O)=O demonstrate bactericidal effects against E. coli in vitro?', 'answer': 'No', '_id': 21892}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 61%|██████ | 1827/3000 [05:56<25:42, 1.31s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:38:44.908\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0241431', 'INHIB_AVE': -5.6, 'SMILES': 'c1(N(CC2)CCO2)nc(N\\\\N=C\\\\c3ccc(cc3OC)OC)nc(Nc(cc4)ccc4F)n1.Cl', 'HIT': 0.0, 'question': 'Does treatment with c1(N(CC2)CCO2)nc(N\\\\N=C\\\\c3ccc(cc3OC)OC)nc(Nc(cc4)ccc4F)n1.Cl result in bacteriostatic or bactericidal effects on E. coli?', 'answer': 'No', '_id': 73007}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:38:44.927\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0292202', 'INHIB_AVE': -0.42, 'SMILES': 'n1c(c(cccc2[N+](=O)[O-])c2)[nH]c(c3ccccc3)c1c(cc4)ccc4c(cc5)ccc5[N+](=O)[O-]', 'HIT': 0.0, 'question': 'Can n1c(c(cccc2[N+](=O)[O-])c2)[nH]c(c3ccccc3)c1c(cc4)ccc4c(cc5)ccc5[N+](=O)[O-] effectively kill or suppress Escherichia coli cells?', 'answer': 'No', '_id': 59079}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 61%|██████ | 1829/3000 [05:56<17:59, 1.08it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:38:44.943\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0219307', 'INHIB_AVE': 7.45, 'SMILES': 'c1(ccccc1NC(=O)c2ccc(cc2)F)C(=O)N\\\\N=C\\\\C=C\\\\c3ccccc3', 'HIT': 0.0, 'question': 'Does treatment with c1(ccccc1NC(=O)c2ccc(cc2)F)C(=O)N\\\\N=C\\\\C=C\\\\c3ccccc3 result in bacteriostatic or bactericidal effects on E. coli?', 'answer': 'No', '_id': 25247}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:38:45.036\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0303801', 'INHIB_AVE': -1.56, 'SMILES': '[N+](=O)([O-])c1ccc(cc1NCc2ccccc2)N(CC3)CCN3C(=O)c4ccccc4Br', 'HIT': 0.0, 'question': 'Does the proposed molecule [N+](=O)([O-])c1ccc(cc1NCc2ccccc2)N(CC3)CCN3C(=O)c4ccccc4Br exhibit antibacterial activity against Escherichia coli?', 'answer': 'No', '_id': 62948}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 61%|██████ | 1831/3000 [05:56<12:45, 1.53it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:38:45.122\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0254450', 'INHIB_AVE': 8.62, 'SMILES': 'C1(c2ccccc2)(c3ccccc3)OC(CCc4ccc(C)o4)Nc(cccc5)c15', 'HIT': 0.0, 'question': 'Can C1(c2ccccc2)(c3ccccc3)OC(CCc4ccc(C)o4)Nc(cccc5)c15 effectively kill or suppress Escherichia coli cells?', 'answer': 'No', '_id': 21099}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2026-01-10 17:38:45.174\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0163065', 'INHIB_AVE': 6.94, 'SMILES': 'C(=O)(N\\\\N=C\\\\c(ccc1I)o1)c2ccco2', 'HIT': 0.0, 'question': 'Does treatment with C(=O)(N\\\\N=C\\\\c(ccc1I)o1)c2ccco2 result in bacteriostatic or bactericidal effects on E. coli?', 'answer': 'No', '_id': 27236}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 61%|██████ | 1835/3000 [05:56<07:05, 2.74it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:38:45.227\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:38:45.257\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0286775', 'INHIB_AVE': 13.16, 'SMILES': 'C(C#N)(C1c2ccc(cc2)F)=C(N)N(C(CC(C)(C)CC3=O)=C13)c4cccc(Br)c4', 'HIT': 0.0, 'question': 'Can C(C#N)(C1c2ccc(cc2)F)=C(N)N(C(CC(C)(C)CC3=O)=C13)c4cccc(Br)c4 effectively kill or suppress Escherichia coli cells?', 'answer': 'No', '_id': 10317}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 61%|██████▏ | 1838/3000 [06:01<15:35, 1.24it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 61%|██████▏ | 1840/3000 [06:02<13:54, 1.39it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2026-01-10 17:38:51.157\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0304152', 'INHIB_AVE': 3.17, 'SMILES': 'C(C(OC)=O)(=CN(Cc(cc1)ccc1F)C=C2C(OC)=O)C2c3ccc(cc3)Cl', 'HIT': 0.0, 'question': 'Does the proposed molecule C(C(OC)=O)(=CN(Cc(cc1)ccc1F)C=C2C(OC)=O)C2c3ccc(cc3)Cl exhibit antibacterial activity against Escherichia coli?', 'answer': 'No', '_id': 43852}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:38:51.270\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 61%|██████▏ | 1843/3000 [06:02<09:28, 2.04it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:38:51.277\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:38:51.399\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 62%|██████▏ | 1845/3000 [06:02<07:26, 2.59it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 62%|██████▏ | 1847/3000 [06:04<08:55, 2.15it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 62%|██████▏ | 1850/3000 [06:04<06:29, 2.95it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 62%|██████▏ | 1855/3000 [06:05<04:52, 3.91it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 62%|██████▏ | 1857/3000 [06:05<03:45, 5.07it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 62%|██████▏ | 1862/3000 [06:06<02:27, 7.69it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 62%|██████▏ | 1864/3000 [06:06<02:18, 8.21it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 62%|██████▏ | 1867/3000 [06:06<03:02, 6.22it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 62%|██████▏ | 1868/3000 [06:07<03:37, 5.20it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 62%|██████▏ | 1872/3000 [06:07<02:53, 6.50it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 62%|██████▎ | 1875/3000 [06:08<02:16, 8.25it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 63%|██████▎ | 1876/3000 [06:08<02:42, 6.91it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 63%|██████▎ | 1878/3000 [06:08<02:50, 6.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 63%|██████▎ | 1879/3000 [06:08<03:12, 5.82it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 63%|██████▎ | 1882/3000 [06:09<02:40, 6.98it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 63%|██████▎ | 1886/3000 [06:09<01:53, 9.85it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 63%|██████▎ | 1888/3000 [06:09<02:00, 9.21it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 63%|██████▎ | 1890/3000 [06:09<01:57, 9.47it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 63%|██████▎ | 1892/3000 [06:10<03:44, 4.93it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 63%|██████▎ | 1896/3000 [06:11<02:25, 7.59it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 63%|██████▎ | 1898/3000 [06:11<02:29, 7.36it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 63%|██████▎ | 1901/3000 [06:11<02:17, 8.00it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 63%|██████▎ | 1904/3000 [06:12<02:07, 8.59it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 64%|██████▎ | 1907/3000 [06:12<02:59, 6.08it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 64%|██████▎ | 1911/3000 [06:13<02:36, 6.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 64%|██████▍ | 1913/3000 [06:13<02:20, 7.75it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 64%|██████▍ | 1916/3000 [06:13<01:58, 9.12it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 64%|██████▍ | 1921/3000 [06:14<01:59, 9.03it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 64%|██████▍ | 1923/3000 [06:14<02:09, 8.32it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 64%|██████▍ | 1925/3000 [06:14<02:05, 8.55it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 64%|██████▍ | 1928/3000 [06:15<02:43, 6.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 64%|██████▍ | 1929/3000 [06:15<02:32, 7.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 64%|██████▍ | 1932/3000 [06:16<02:40, 6.66it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 64%|██████▍ | 1934/3000 [06:16<02:21, 7.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 65%|██████▍ | 1938/3000 [06:16<01:41, 10.51it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 65%|██████▍ | 1940/3000 [06:16<01:42, 10.32it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 65%|██████▍ | 1942/3000 [06:17<02:04, 8.51it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 65%|██████▍ | 1944/3000 [06:17<01:50, 9.55it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 65%|██████▍ | 1946/3000 [06:17<01:52, 9.38it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 65%|██████▍ | 1948/3000 [06:17<02:14, 7.79it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 65%|██████▌ | 1950/3000 [06:18<02:10, 8.05it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 65%|██████▌ | 1953/3000 [06:18<02:35, 6.75it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 65%|██████▌ | 1956/3000 [06:18<01:55, 9.03it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 65%|██████▌ | 1958/3000 [06:19<01:56, 8.98it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 65%|██████▌ | 1961/3000 [06:19<02:25, 7.16it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 65%|██████▌ | 1964/3000 [06:20<02:12, 7.81it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 66%|██████▌ | 1970/3000 [06:20<01:37, 10.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 66%|██████▌ | 1972/3000 [06:20<01:28, 11.59it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 66%|██████▌ | 1976/3000 [06:20<01:24, 12.18it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 66%|██████▌ | 1979/3000 [06:21<02:51, 5.96it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 66%|██████▌ | 1982/3000 [06:22<02:32, 6.67it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 66%|██████▌ | 1987/3000 [06:22<01:40, 10.03it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 66%|██████▋ | 1989/3000 [06:22<01:42, 9.87it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 66%|██████▋ | 1991/3000 [06:23<02:21, 7.11it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 66%|██████▋ | 1993/3000 [06:23<02:33, 6.54it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 66%|██████▋ | 1995/3000 [06:24<02:20, 7.15it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 67%|██████▋ | 1996/3000 [06:24<02:32, 6.57it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 67%|██████▋ | 1999/3000 [06:24<02:37, 6.35it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 67%|██████▋ | 2003/3000 [06:24<01:44, 9.50it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 67%|██████▋ | 2005/3000 [06:25<01:49, 9.11it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 67%|██████▋ | 2009/3000 [06:25<01:31, 10.87it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 67%|██████▋ | 2011/3000 [06:26<02:30, 6.59it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 67%|██████▋ | 2013/3000 [06:26<02:30, 6.54it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 67%|██████▋ | 2015/3000 [06:26<02:52, 5.71it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 67%|██████▋ | 2017/3000 [06:27<02:32, 6.46it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 67%|██████▋ | 2019/3000 [06:27<01:56, 8.41it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 67%|██████▋ | 2023/3000 [06:27<01:32, 10.55it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 68%|██████▊ | 2025/3000 [06:27<01:36, 10.11it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 68%|██████▊ | 2029/3000 [06:28<01:24, 11.44it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 68%|██████▊ | 2031/3000 [06:28<02:09, 7.49it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 68%|██████▊ | 2033/3000 [06:28<02:24, 6.67it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 68%|██████▊ | 2035/3000 [06:29<02:24, 6.69it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 68%|██████▊ | 2038/3000 [06:29<02:13, 7.20it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 68%|██████▊ | 2040/3000 [06:29<02:16, 7.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 68%|██████▊ | 2043/3000 [06:30<02:12, 7.20it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 68%|██████▊ | 2046/3000 [06:30<02:01, 7.88it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 68%|██████▊ | 2049/3000 [06:30<01:37, 9.79it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 68%|██████▊ | 2052/3000 [06:31<01:53, 8.32it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 68%|██████▊ | 2054/3000 [06:31<01:35, 9.87it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 69%|██████▊ | 2057/3000 [06:32<02:47, 5.63it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 69%|██████▊ | 2062/3000 [06:32<01:57, 7.95it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 69%|██████▉ | 2065/3000 [06:33<01:58, 7.86it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 69%|██████▉ | 2067/3000 [06:33<02:19, 6.67it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 69%|██████▉ | 2068/3000 [06:33<02:40, 5.82it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 69%|██████▉ | 2074/3000 [06:34<01:19, 11.62it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 69%|██████▉ | 2076/3000 [06:34<01:20, 11.47it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 69%|██████▉ | 2078/3000 [06:34<01:41, 9.04it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 69%|██████▉ | 2082/3000 [06:35<02:14, 6.82it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 69%|██████▉ | 2083/3000 [06:35<02:07, 7.21it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 70%|██████▉ | 2087/3000 [06:36<02:07, 7.17it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 70%|██████▉ | 2092/3000 [06:36<01:17, 11.65it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 70%|██████▉ | 2094/3000 [06:36<01:16, 11.85it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 70%|██████▉ | 2097/3000 [06:36<01:09, 12.95it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 70%|███████ | 2102/3000 [06:37<02:10, 6.87it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 70%|███████ | 2104/3000 [06:38<01:55, 7.73it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 70%|███████ | 2106/3000 [06:38<02:19, 6.40it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 70%|███████ | 2108/3000 [06:44<14:54, 1.00s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 70%|███████ | 2109/3000 [06:46<15:20, 1.03s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:39:34.945\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0235450', 'INHIB_AVE': -1.19, 'SMILES': 's1c(c2ccccc2)nnc1NC(=O)c3ccccc3Br', 'HIT': 0.0, 'question': 'Does treatment with s1c(c2ccccc2)nnc1NC(=O)c3ccccc3Br result in bacteriostatic or bactericidal effects on E. coli?', 'answer': 'No', '_id': 61701}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 10 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:39:34.973\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0295120', 'INHIB_AVE': -20.23, 'SMILES': 'c1(Oc2cccc(c2)NC(COc3ccc(cc3Cl)Cl)=O)nc(C)cc(C)c1C#N', 'HIT': 0.0, 'question': 'Does treatment with c1(Oc2cccc(c2)NC(COc3ccc(cc3Cl)Cl)=O)nc(C)cc(C)c1C#N result in bacteriostatic or bactericidal effects on E. coli?', 'answer': 'No', '_id': 82188}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 10 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 70%|███████ | 2112/3000 [06:46<09:38, 1.54it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2026-01-10 17:39:35.466\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 9 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 70%|███████ | 2113/3000 [06:46<08:37, 1.72it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:39:35.608\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0252493', 'INHIB_AVE': 4.7, 'SMILES': 'C(/NC(c1ccccc1)=O)(\\\\C(=O)N\\\\N=C\\\\c2c(cccc2Cl)Cl)=C/c3cccs3', 'HIT': 0.0, 'question': 'Is Escherichia coli susceptible to treatment with the compound C(/NC(c1ccccc1)=O)(\\\\C(=O)N\\\\N=C\\\\c2c(cccc2Cl)Cl)=C/c3cccs3?', 'answer': 'No', '_id': 37004}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 9 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 70%|███████ | 2114/3000 [06:46<07:21, 2.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:39:44.403\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0163993', 'INHIB_AVE': 13.88, 'SMILES': 'O=C(N1N(S(=O)(C2=CC=C([N+]([O-])=O)C=C2)=O)C3C(CCC3)C1O)OC(C)(C)C', 'HIT': 0.0, 'question': 'Can O=C(N1N(S(=O)(C2=CC=C([N+]([O-])=O)C=C2)=O)C3C(CCC3)C1O)OC(C)(C)C effectively kill or suppress Escherichia coli cells?', 'answer': 'No', '_id': 9230}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 70%|███████ | 2115/3000 [06:55<33:36, 2.28s/it]Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097993.217209857)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097992.515879588)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097995.037735843)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097994.581430171)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097995.172525656)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097995.135217403)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097997.537831381)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097998.641360771)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097996.423252204)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2097999.098943959)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098000.531200871)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098000.936019865)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098001.611423902)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098000.986046292)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098001.022102034)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098002.36591194)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098002.327093094)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098000.769575027)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098000.503198796)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098000.841546723)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098002.927651291)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098003.118876441)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098005.441532737)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098005.008917249)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098005.363733066)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098006.796275519)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098006.470155077)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098005.437216704)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098004.95123119)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098007.753323407)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098008.352377548)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098007.622437057)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098006.471321252)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098006.805658923)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098006.105254791)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098027.205288483)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:39:45.142\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0279089', 'INHIB_AVE': 7.36, 'SMILES': 'c1(C(=O)Nc2sc(c3c2C(OCC)=O)CCCC3)cn(nc1C45CC6CC(CC(C6)C4)C5)c7ccccc7', 'HIT': 0.0, 'question': 'Is Escherichia coli susceptible to treatment with the compound c1(C(=O)Nc2sc(c3c2C(OCC)=O)CCCC3)cn(nc1C45CC6CC(CC(C6)C4)C5)c7ccccc7?', 'answer': 'No', '_id': 25632}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2098026.934710099)])']\n", "connector: \n", "Evaluating workflow: 71%|███████ | 2116/3000 [06:56<28:18, 1.92s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:39:45.168\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0321350', 'INHIB_AVE': 0.04, 'SMILES': 'n1(c2c(n3)cccc2)c3S\\\\C(=C/c4cccc(c4)OCc5ccccc5F)\\\\C1=O', 'HIT': 0.0, 'question': 'Can n1(c2c(n3)cccc2)c3S\\\\C(=C/c4cccc(c4)OCc5ccccc5F)\\\\C1=O effectively kill or suppress Escherichia coli cells?', 'answer': 'No', '_id': 57294}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2098026.92916599)])']\n", "connector: \n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:39:45.177\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0285792', 'INHIB_AVE': -6.34, 'SMILES': 'C1(C(CC(NC1(C)C)(C)C)NC(=O)C(F)(F)F)NC(=O)C(F)(F)F', 'HIT': 0.0, 'question': 'Does treatment with C1(C(CC(NC1(C)C)(C)C)NC(=O)C(F)(F)F)NC(=O)C(F)(F)F result in bacteriostatic or bactericidal effects on E. coli?', 'answer': 'No', '_id': 74305}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:39:45.190\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0217589', 'INHIB_AVE': -0.71, 'SMILES': 'c1(OC)c(OC)ccc(\\\\C=N\\\\NC(=O)c(cc2OC)cc(c2OC)OC)c1OC', 'HIT': 0.0, 'question': 'Is Escherichia coli susceptible to treatment with the compound c1(OC)c(OC)ccc(\\\\C=N\\\\NC(=O)c(cc2OC)cc(c2OC)OC)c1OC?', 'answer': 'No', '_id': 60107}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2098017.163014604)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098017.604969329)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098027.36624143)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098032.820830408)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098017.167054859)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098036.339581815)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098034.105182356)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098036.6032045)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098033.494721149)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098034.277994202)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098035.224480434)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098035.202858367)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098036.405494354)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098036.477816083)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098034.980616749)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098036.469708739)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098036.495092321)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098034.922737159)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098036.232080574)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098036.032769777)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098033.496328995)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:39:45.192\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0161726', 'INHIB_AVE': 11.49, 'SMILES': 'c1(oc2c(ccc(C)c2)n1)c3cc(ccc3O)\\\\N=C\\\\c(cc(Cl)cc4I)c4O', 'HIT': 0.0, 'question': 'Does treatment with c1(oc2c(ccc(C)c2)n1)c3cc(ccc3O)\\\\N=C\\\\c(cc(Cl)cc4I)c4O result in bacteriostatic or bactericidal effects on E. coli?', 'answer': 'No', '_id': 13464}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2098033.509156936)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098034.907923966)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098037.892530625)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098037.868063667)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098033.488861762)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098036.511714485)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098038.659902452)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098038.194766298)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098040.282672413)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098038.38457773)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098037.16903343)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098037.180575507)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098040.310751272)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098039.95261084)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098040.164227089)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098039.961590539)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098040.219090177)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098040.938752542)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098040.267022737)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098040.34180763)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098041.394248032)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098042.473523962)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098040.623448349)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098041.02096349)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098038.504480506)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098042.396997425)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098040.010398942)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:39:45.194\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098042.878736946)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098042.560377007)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097993.07891324)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097993.087408537)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097993.978584751)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097994.569511345)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097994.3400417)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097994.676471942)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097994.310736883)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097994.574236402)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097993.824059896)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097994.853814973)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097994.932563783)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097995.586086462)])']\n", "connector: \n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:39:45.285\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0163421', 'INHIB_AVE': -7.54, 'SMILES': 'c1(c2ccc(cc2)I)ccc(\\\\C=N/NC(COc3cccc(C)c3)=O)o1', 'HIT': 0.0, 'question': 'Does c1(c2ccc(cc2)I)ccc(\\\\C=N/NC(COc3cccc(C)c3)=O)o1 demonstrate bactericidal effects against E. coli in vitro?', 'answer': 'No', '_id': 76127}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097995.811556959)])']\n", "connector: \n", "Evaluating workflow: 71%|███████ | 2125/3000 [06:56<07:07, 2.04it/s]Unclosed connector\n", "connections: ['deque([(, 2097995.319273409)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097995.858299004)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097996.473587063)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097995.957094813)])']\n", "connector: \n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:39:45.297\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0243288', 'INHIB_AVE': 18.36, 'SMILES': 'c(CC(C1)ON=C1C(N)=O)(c(OC)c2OC)c(Br)c(c23)OCO3', 'HIT': 0.0, 'question': 'Is c(CC(C1)ON=C1C(N)=O)(c(OC)c2OC)c(Br)c(c23)OCO3 capable of inhibiting the growth or survival of E. coli?', 'answer': 'No', '_id': 4014}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2097996.321741269)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097995.247711418)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097996.01049996)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097996.556928818)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097997.598874025)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097997.066597565)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097998.393201909)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097998.100934304)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097997.650955883)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097998.468772313)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097997.032126958)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097996.940346839)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097998.215754816)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097998.056577253)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097997.868049161)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097997.287260585)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097997.356032439)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097998.335263697)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097998.312491136)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098000.163225677)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097999.142707529)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097999.114646738)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098000.351584091)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097999.160360582)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097999.824597653)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2097999.752371384)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098000.062677407)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098000.045045066)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098001.550549762)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098000.768352367)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098000.703417319)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098000.796853405)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098001.414144652)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098001.570345588)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098002.303539036)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098002.276878218)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098002.009621747)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098002.89709858)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098003.125141689)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098003.814837385)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098004.230844595)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098003.231143109)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2098004.411907889)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098003.027927115)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098004.346482222)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098004.476643263)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098003.709691264)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098004.098392489)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098004.935478928)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098005.176278528)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098005.312078482)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098005.507612055)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098006.457555672)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098005.985884576)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098006.369052558)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098006.534843594)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098005.304004408)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098009.041777098)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098007.620039169)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098008.13378193)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098008.315403057)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098007.433740568)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098008.135176784)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098008.961224763)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098008.295176318)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098008.511674799)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098008.546907176)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098008.586929229)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098027.010071416)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098026.944274117)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098017.823809105)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098009.034821399)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098027.173663675)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098027.153214143)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098026.777952938)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098017.187057045)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098027.178323925)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098027.343193088)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098017.208050152)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098027.022629153)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098027.352436382)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098034.63485449)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098037.70075355)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098036.111710379)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2098037.064137557)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098037.823671112)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098038.033856186)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098038.554742252)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098037.757041792)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098038.648383451)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098040.111612734)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098038.356458884)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098038.733570998)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098040.269892929)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098003.672927909)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098044.205129022)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098042.531084057)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098043.287805582)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098044.06181009)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098042.989887545)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098042.38041098)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098042.455547848)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098042.682569061)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098043.668249546)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098043.806505723)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098042.70426464)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098044.251127451)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098043.94749581)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098041.966210417)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098042.117108915)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098042.286039337)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098044.364972228)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098044.95656699)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098045.121013543)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098045.641737685)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2098044.991633756)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098046.057362649)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098045.257449232)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098044.857164257)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098045.981578004)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098046.811923287)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098045.987584738)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098046.336035301)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098045.426053592)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098044.189229808)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098046.698976389)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098046.940946085)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098046.854831619)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098047.5651435)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098047.052078655)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098047.059534086)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098047.115463508)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098047.587388003)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098047.129302364)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098048.553155124)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098047.531031377)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098048.274675576)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098046.843938792)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098046.396600516)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098046.207153555)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098048.940249797)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098048.50521515)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098046.107901253)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2098048.776069002)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098048.191094352)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098048.612275255)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098049.123955561)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098049.722025336)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098048.673543161)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098048.976636719)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098049.315163752)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098050.77718387)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098049.097839793)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098049.168443703)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098049.423793917)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098049.428075811)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098050.706834049)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098048.270580104)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098048.434056845)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098050.601245929)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098050.179182682)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098050.489706024)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098052.479090496)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098050.733104048)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098052.771690545)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098050.648124416)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098051.119497927)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098051.097247139)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098051.290129863)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098051.01722943)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098051.155102933)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098050.329379572)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098051.475599444)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098053.265480088)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098050.562209366)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098052.312323121)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098052.58865226)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098053.536961651)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098053.327757044)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098052.368488166)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098053.374791137)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098053.149041077)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098053.164021241)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098052.747872194)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098055.203678458)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098053.288513205)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098055.149872062)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098052.477426798)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098053.075826374)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098053.225873308)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098054.499184834)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098054.070425494)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098054.276414543)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098054.692603373)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098054.644262072)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098055.281704219)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098054.857975651)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098055.702645768)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098055.452225182)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2098055.094475132)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098055.524842966)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098055.349624783)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098055.542246946)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098057.221447278)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098056.321471607)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098057.142646597)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098056.420761258)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098056.899891525)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098056.65936908)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098058.348749016)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098057.658924645)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098058.104490367)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098057.147027618)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098056.806500174)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098057.095089824)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098057.463477953)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098056.931667084)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098058.614804522)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098057.456451548)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098058.574288594)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098057.257952574)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098059.048756837)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098059.350258841)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098059.805731178)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098058.640393644)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098059.496452497)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098059.679916955)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098059.091514293)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098059.490986367)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098060.406282819)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098059.665575228)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098059.852083344)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098057.707445208)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098058.553991843)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098061.266183379)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098060.223852726)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098061.076706634)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098061.483876568)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098061.308709707)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098061.422761891)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098062.288177877)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098062.072918431)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098062.308644081)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098061.959908862)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098060.753647412)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098059.81027215)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098060.298511146)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098061.61819554)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098060.092784995)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098063.114396972)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098062.671068445)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098063.634425084)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098062.208942842)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098062.766505466)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098063.05184665)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2098063.392188107)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098062.957905569)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098064.306420363)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098063.421761158)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098063.060584911)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098063.755543042)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098063.266841319)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098064.029384305)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098064.612912007)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098062.377127008)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098065.062745516)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098064.902789543)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098064.809137888)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098064.581827278)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098064.940546805)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098064.912817214)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098065.538935651)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098065.296793061)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098065.309039956)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098066.142544014)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098065.587546638)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098065.589441034)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098065.804424805)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098067.386275783)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098066.418679314)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098065.560767805)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098066.507903463)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098067.080197136)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2098067.17409917)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098067.60374447)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098066.807675194)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098066.93474308)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098067.649780684)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098067.686806201)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098067.983717964)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098068.231933094)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098068.000276535)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098067.839088706)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098067.049341203)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098067.221674237)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098068.899978883)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098067.56444331)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098068.694595085)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098069.473922658)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098068.218796566)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098068.708145898)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098069.394101938)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098069.636697641)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098077.266045944)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098087.631158935)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098077.615578543)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098087.623936097)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098077.728935752)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098069.417373117)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098077.215652541)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098087.331511408)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098087.544771613)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098087.76649529)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098087.682487991)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098087.452040434)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098077.690938079)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098077.601086986)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098077.761612088)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098087.23115339)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098087.416665321)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098077.711974501)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098087.43634822)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098087.647972258)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098077.420963493)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098097.188866387)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098093.64980749)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098092.879235158)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098093.665985813)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098093.636037143)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098092.736767345)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098094.317546546)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098095.729891189)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098087.736517369)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098097.097735693)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098095.503782777)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098095.69205086)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098095.286817095)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098095.78438447)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098096.719314408)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2098093.780460023)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098093.776579797)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098097.811841907)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098095.326437453)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098097.122398766)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098093.908187019)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098097.384732713)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098098.190427368)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098096.865478654)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098096.773065484)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098096.876584614)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098096.639341511)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098098.516692847)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098097.483162589)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098097.283915137)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098099.518692399)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098098.797495317)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098098.597911179)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098098.599260137)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098099.101487143)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098098.979488934)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098100.695434024)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098101.142883876)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098099.81781186)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098099.486715979)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098100.650730158)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098097.034660705)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098098.079474613)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2098099.137333515)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098100.062873707)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098100.479572027)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098100.33731338)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098101.207734317)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098100.761227916)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098101.009018624)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098102.041220048)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098102.598880708)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098102.072314192)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098100.854442513)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098101.881252554)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098102.186600443)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098099.252334915)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098100.465165559)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098100.3831558)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098102.927639659)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098102.650999378)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098103.012318584)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098104.642405301)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098103.229403101)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098102.925816618)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098103.248437925)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098102.600107323)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098103.263763122)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098104.571706976)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098104.435181089)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098104.525052664)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098104.833958772)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098104.974584488)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098102.310271854)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098102.246270658)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098102.552454927)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098104.003259669)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098104.80424238)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098105.624674183)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098105.544185853)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098105.92094388)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098105.841144871)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098105.56112137)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098105.171220848)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098106.36390059)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098106.137310659)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098104.749054104)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098104.445582225)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098105.500744668)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098105.961509188)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098106.911566878)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098106.7951572)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098106.673125738)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098106.89591475)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098107.616710662)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098107.362740685)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098107.575749565)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098107.852430591)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098107.469242123)])']\n", "connector: \n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:39:45.558\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 71%|███████ | 2128/3000 [07:01<12:07, 1.20it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 71%|███████ | 2130/3000 [07:02<10:08, 1.43it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2026-01-10 17:39:51.599\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:39:51.693\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 71%|███████ | 2132/3000 [07:02<08:58, 1.61it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 71%|███████ | 2134/3000 [07:03<07:23, 1.95it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 71%|███████ | 2135/3000 [07:04<08:14, 1.75it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2026-01-10 17:39:53.093\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 71%|███████ | 2136/3000 [07:04<07:09, 2.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:39:53.141\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 71%|███████▏ | 2139/3000 [07:04<04:19, 3.32it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 71%|███████▏ | 2141/3000 [07:04<03:18, 4.32it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 71%|███████▏ | 2143/3000 [07:04<02:53, 4.94it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 72%|███████▏ | 2145/3000 [07:05<04:12, 3.38it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 72%|███████▏ | 2150/3000 [07:06<02:21, 6.00it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 72%|███████▏ | 2153/3000 [07:06<02:17, 6.14it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 72%|███████▏ | 2155/3000 [07:06<02:12, 6.38it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 72%|███████▏ | 2157/3000 [07:07<02:37, 5.34it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 72%|███████▏ | 2160/3000 [07:08<02:33, 5.48it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 72%|███████▏ | 2162/3000 [07:08<01:59, 7.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 72%|███████▏ | 2164/3000 [07:08<02:11, 6.34it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 72%|███████▏ | 2168/3000 [07:08<01:43, 8.04it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 72%|███████▏ | 2169/3000 [07:09<02:04, 6.67it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 72%|███████▏ | 2171/3000 [07:09<02:12, 6.25it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 72%|███████▎ | 2175/3000 [07:09<01:45, 7.82it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 73%|███████▎ | 2180/3000 [07:10<01:13, 11.14it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 73%|███████▎ | 2182/3000 [07:10<01:34, 8.69it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 73%|███████▎ | 2184/3000 [07:11<01:56, 7.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 73%|███████▎ | 2186/3000 [07:11<02:24, 5.61it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 73%|███████▎ | 2188/3000 [07:11<01:52, 7.23it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 73%|███████▎ | 2190/3000 [07:11<01:36, 8.35it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 73%|███████▎ | 2195/3000 [07:12<01:14, 10.85it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 73%|███████▎ | 2197/3000 [07:13<02:50, 4.71it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 73%|███████▎ | 2200/3000 [07:13<02:23, 5.57it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 73%|███████▎ | 2203/3000 [07:13<01:40, 7.92it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 74%|███████▎ | 2210/3000 [07:14<01:00, 12.99it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 74%|███████▎ | 2212/3000 [07:14<01:16, 10.34it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 74%|███████▍ | 2215/3000 [07:15<02:32, 5.16it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 74%|███████▍ | 2216/3000 [07:15<02:19, 5.60it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 74%|███████▍ | 2222/3000 [07:16<01:30, 8.59it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 74%|███████▍ | 2224/3000 [07:16<01:42, 7.60it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 74%|███████▍ | 2229/3000 [07:17<01:28, 8.73it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 74%|███████▍ | 2231/3000 [07:17<01:27, 8.79it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 74%|███████▍ | 2233/3000 [07:17<01:14, 10.23it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 75%|███████▍ | 2237/3000 [07:18<02:05, 6.08it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 75%|███████▍ | 2241/3000 [07:18<01:21, 9.36it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 75%|███████▍ | 2244/3000 [07:19<01:12, 10.47it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 75%|███████▍ | 2246/3000 [07:19<01:37, 7.73it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 75%|███████▍ | 2248/3000 [07:19<01:29, 8.37it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 75%|███████▌ | 2250/3000 [07:19<01:27, 8.53it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 75%|███████▌ | 2252/3000 [07:20<02:01, 6.18it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 75%|███████▌ | 2254/3000 [07:20<01:49, 6.84it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 75%|███████▌ | 2259/3000 [07:21<01:22, 8.98it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 75%|███████▌ | 2262/3000 [07:21<01:08, 10.75it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 76%|███████▌ | 2269/3000 [07:22<01:04, 11.28it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 76%|███████▌ | 2271/3000 [07:22<01:35, 7.66it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 76%|███████▌ | 2273/3000 [07:22<01:44, 6.96it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 76%|███████▌ | 2275/3000 [07:23<01:44, 6.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 76%|███████▌ | 2277/3000 [07:23<01:21, 8.84it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 76%|███████▌ | 2282/3000 [07:23<00:59, 12.11it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 76%|███████▌ | 2284/3000 [07:23<01:01, 11.71it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 76%|███████▌ | 2286/3000 [07:24<02:21, 5.05it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 76%|███████▋ | 2289/3000 [07:25<02:05, 5.68it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 76%|███████▋ | 2294/3000 [07:25<01:11, 9.92it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 77%|███████▋ | 2296/3000 [07:25<01:32, 7.61it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 77%|███████▋ | 2300/3000 [07:26<01:39, 7.03it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 77%|███████▋ | 2302/3000 [07:26<01:23, 8.35it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 77%|███████▋ | 2304/3000 [07:26<01:17, 8.98it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 77%|███████▋ | 2309/3000 [07:27<01:11, 9.63it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 77%|███████▋ | 2313/3000 [07:27<01:04, 10.72it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 77%|███████▋ | 2315/3000 [07:28<01:19, 8.65it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 77%|███████▋ | 2317/3000 [07:28<01:33, 7.30it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 77%|███████▋ | 2319/3000 [07:28<01:21, 8.36it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 77%|███████▋ | 2320/3000 [07:28<01:44, 6.51it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 77%|███████▋ | 2322/3000 [07:29<01:36, 7.00it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 78%|███████▊ | 2327/3000 [07:29<01:16, 8.85it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 78%|███████▊ | 2329/3000 [07:29<01:14, 9.03it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 78%|███████▊ | 2331/3000 [07:30<01:44, 6.40it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 78%|███████▊ | 2335/3000 [07:30<01:25, 7.74it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 78%|███████▊ | 2337/3000 [07:31<01:29, 7.43it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 78%|███████▊ | 2339/3000 [07:31<01:12, 9.06it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 78%|███████▊ | 2343/3000 [07:31<01:20, 8.15it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 78%|███████▊ | 2345/3000 [07:32<01:21, 8.00it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 78%|███████▊ | 2347/3000 [07:32<01:43, 6.32it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 78%|███████▊ | 2351/3000 [07:32<01:09, 9.37it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 78%|███████▊ | 2353/3000 [07:33<00:57, 11.23it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 78%|███████▊ | 2355/3000 [07:33<01:04, 10.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 79%|███████▊ | 2359/3000 [07:33<01:22, 7.77it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 79%|███████▉ | 2365/3000 [07:34<01:20, 7.92it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 79%|███████▉ | 2367/3000 [07:35<01:20, 7.90it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 79%|███████▉ | 2369/3000 [07:35<01:40, 6.30it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 79%|███████▉ | 2371/3000 [07:36<01:57, 5.35it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 79%|███████▉ | 2375/3000 [07:36<01:35, 6.55it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 79%|███████▉ | 2379/3000 [07:36<01:11, 8.66it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 79%|███████▉ | 2382/3000 [07:37<00:52, 11.86it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 79%|███████▉ | 2384/3000 [07:37<01:13, 8.37it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 80%|███████▉ | 2388/3000 [07:38<01:27, 6.98it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 80%|███████▉ | 2390/3000 [07:38<01:29, 6.85it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 80%|███████▉ | 2392/3000 [07:38<01:25, 7.15it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 80%|███████▉ | 2394/3000 [07:39<01:34, 6.43it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 80%|███████▉ | 2398/3000 [07:44<07:20, 1.37it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 80%|███████▉ | 2399/3000 [07:45<06:51, 1.46it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:40:34.784\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0304156', 'INHIB_AVE': 29.06, 'SMILES': 'C(C(OC)=O)(=CN(Cc1cccs1)C=C2C(OC)=O)C2c3ccc(cc3)Cl', 'HIT': 0.0, 'question': 'Does the proposed molecule C(C(OC)=O)(=CN(Cc1cccs1)C=C2C(OC)=O)C2c3ccc(cc3)Cl exhibit antibacterial activity against Escherichia coli?', 'answer': 'No', '_id': 813}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 10 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 80%|████████ | 2400/3000 [07:46<06:59, 1.43it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:40:34.829\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:40:34.852\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0166060', 'INHIB_AVE': 85.55, 'SMILES': 'c1(c(C)cccc1C)OCC(=O)N\\\\N=C\\\\c(cc2)ccc2C(O)=O', 'HIT': 0.0, 'question': 'Is Escherichia coli susceptible to treatment with the compound c1(c(C)cccc1C)OCC(=O)N\\\\N=C\\\\c(cc2)ccc2C(O)=O?', 'answer': 'No', '_id': 394}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 10 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:40:35.033\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0230568', 'INHIB_AVE': -0.86, 'SMILES': 'C(CSc1ccccc1)(=O)N\\\\N=C\\\\c(cc2)ccc2N(C)c3ccccc3', 'HIT': 0.0, 'question': 'Does the proposed molecule C(CSc1ccccc1)(=O)N\\\\N=C\\\\c(cc2)ccc2N(C)c3ccccc3 exhibit antibacterial activity against Escherichia coli?', 'answer': 'No', '_id': 60637}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 10 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 80%|████████ | 2403/3000 [07:46<04:24, 2.26it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:40:35.075\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:40:35.079\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:40:35.132\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 80%|████████ | 2406/3000 [07:46<02:52, 3.45it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:40:35.172\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:40:35.263\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0291294', 'INHIB_AVE': 4.5, 'SMILES': 'C1(Nc(cccc2C(F)(F)F)c2)(C(OC)=O)N(c3ccc(cc3)OCC)C(C(OC)=C1C(c4ccccc4)=O)=O', 'HIT': 0.0, 'question': 'Does treatment with C1(Nc(cccc2C(F)(F)F)c2)(C(OC)=O)N(c3ccc(cc3)OCC)C(C(OC)=C1C(c4ccccc4)=O)=O result in bacteriostatic or bactericidal effects on E. coli?', 'answer': 'No', '_id': 37921}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 10 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 80%|████████ | 2408/3000 [07:46<02:16, 4.33it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:40:35.469\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0249209', 'INHIB_AVE': -9.08, 'SMILES': 'S(=O)(=O)(c1ccc(cc1)C)Oc(cc2)ccc2\\\\C=C(\\\\C3=O)/N=C(O3)c(cc4)ccc4C(C)(C)C', 'HIT': 0.0, 'question': 'Can S(=O)(=O)(c1ccc(cc1)C)Oc(cc2)ccc2\\\\C=C(\\\\C3=O)/N=C(O3)c(cc4)ccc4C(C)(C)C effectively kill or suppress Escherichia coli cells?', 'answer': 'No', '_id': 77936}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 10 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 80%|████████ | 2410/3000 [07:46<01:55, 5.10it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:40:35.492\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0293861', 'INHIB_AVE': 9.48, 'SMILES': 'c(c(C)c(s1)C)(C2=O)c1N=C(SCC(=O)NC(C)(C)C)N2c3ccccc3', 'HIT': 0.0, 'question': 'Does c(c(C)c(s1)C)(C2=O)c1N=C(SCC(=O)NC(C)(C)C)N2c3ccccc3 show antimicrobial efficacy specifically against Escherichia coli strains?', 'answer': 'No', '_id': 18485}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 10 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:40:35.539\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0166308', 'INHIB_AVE': 7.41, 'SMILES': 'C(=O)(C(NCCc1ccccc1F)=O)Nc2cccc(Cl)c2Cl', 'HIT': 0.0, 'question': 'Is Escherichia coli susceptible to treatment with the compound C(=O)(C(NCCc1ccccc1F)=O)Nc2cccc(Cl)c2Cl?', 'answer': 'No', '_id': 25443}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 9 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:40:35.578\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0294801', 'INHIB_AVE': -6.31, 'SMILES': 'n(c1c(n2C)ccc(c1)NCc3n(C)c4c(cccc4)n3)c2C', 'HIT': 0.0, 'question': 'Can n(c1c(n2C)ccc(c1)NCc3n(C)c4c(cccc4)n3)c2C act as an antibacterial agent targeting Escherichia coli?', 'answer': 'No', '_id': 74250}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 9 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 80%|████████ | 2413/3000 [07:46<01:20, 7.29it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:40:35.625\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0291088', 'INHIB_AVE': 2.5, 'SMILES': 'n1(Cc2ccccc2)c(S)nnc1c3ccc(c(OC)c3)OC', 'HIT': 0.0, 'question': 'Can n1(Cc2ccccc2)c(S)nnc1c3ccc(c(OC)c3)OC act as an antibacterial agent targeting Escherichia coli?', 'answer': 'No', '_id': 46885}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 9 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:40:35.757\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 9 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 80%|████████ | 2415/3000 [07:47<01:13, 8.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:40:36.214\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 9 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 81%|████████ | 2417/3000 [07:47<01:29, 6.54it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:40:45.173\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 6 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:40:45.222\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 81%|████████ | 2419/3000 [07:56<13:14, 1.37s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:40:45.225\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:40:45.266\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0250734', 'INHIB_AVE': -2.33, 'SMILES': 'S(=O)(=O)(N(CC1)CCO1)c2c(Br)ccc(C(OC)=O)c2', 'HIT': 0.0, 'question': 'Is Escherichia coli susceptible to treatment with the compound S(=O)(=O)(N(CC1)CCO1)c2c(Br)ccc(C(OC)=O)c2?', 'answer': 'No', '_id': 65368}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:40:45.285\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:40:45.302\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:40:45.411\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 81%|████████ | 2424/3000 [07:56<06:45, 1.42it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:40:45.471\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 81%|████████ | 2426/3000 [07:56<05:27, 1.75it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 81%|████████ | 2428/3000 [08:02<10:38, 1.12s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2026-01-10 17:40:51.871\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0244855', 'INHIB_AVE': 4.93, 'SMILES': 'C(N(CC)C(=O)N1CC)(C1N2)N=C3N2C(=O)\\\\C(=C\\\\c(c(Cl)cc(Cl)c4)c4)\\\\S3', 'HIT': 0.0, 'question': 'Is C(N(CC)C(=O)N1CC)(C1N2)N=C3N2C(=O)\\\\C(=C\\\\c(c(Cl)cc(Cl)c4)c4)\\\\S3 capable of inhibiting the growth or survival of E. coli?', 'answer': 'No', '_id': 35960}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 81%|████████ | 2430/3000 [08:03<08:35, 1.11it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:40:51.932\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0295302', 'INHIB_AVE': 1.78, 'SMILES': 'C(=C\\\\NCCCCCC)(/C1=O)\\\\C(=O)NC(=O)N1', 'HIT': 0.0, 'question': 'Can C(=C\\\\NCCCCCC)(/C1=O)\\\\C(=O)NC(=O)N1 act as an antibacterial agent targeting Escherichia coli?', 'answer': 'No', '_id': 50088}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:40:51.942\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0217586', 'INHIB_AVE': 0.3, 'SMILES': 's1c(\\\\C=N\\\\NC(COc(ccc(c2C)C)c2)=O)ccc1Br', 'HIT': 0.0, 'question': 'Does treatment with s1c(\\\\C=N\\\\NC(COc(ccc(c2C)C)c2)=O)ccc1Br result in bacteriostatic or bactericidal effects on E. coli?', 'answer': 'No', '_id': 56228}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:40:52.010\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0328467', 'INHIB_AVE': 4.33, 'SMILES': 'N1(CC(C)=C(CC1(C(OC)=O)C(F)(F)F)C)S(c2ccccc2)(=O)=O', 'HIT': 0.0, 'question': 'Is Escherichia coli susceptible to treatment with the compound N1(CC(C)=C(CC1(C(OC)=O)C(F)(F)F)C)S(c2ccccc2)(=O)=O?', 'answer': 'No', '_id': 38680}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 81%|████████ | 2435/3000 [08:03<04:21, 2.16it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 81%|████████ | 2437/3000 [08:04<04:37, 2.03it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 81%|████████▏ | 2439/3000 [08:04<03:36, 2.59it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 81%|████████▏ | 2442/3000 [08:05<03:23, 2.75it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 81%|████████▏ | 2444/3000 [08:06<02:59, 3.10it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 82%|████████▏ | 2446/3000 [08:06<02:09, 4.29it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 82%|████████▏ | 2450/3000 [08:06<01:23, 6.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 82%|████████▏ | 2452/3000 [08:07<01:30, 6.04it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 82%|████████▏ | 2454/3000 [08:07<01:24, 6.49it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 82%|████████▏ | 2456/3000 [08:07<01:31, 5.96it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 82%|████████▏ | 2460/3000 [08:08<01:11, 7.52it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 82%|████████▏ | 2462/3000 [08:08<01:16, 7.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 82%|████████▏ | 2465/3000 [08:08<01:02, 8.54it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 82%|████████▏ | 2467/3000 [08:08<00:50, 10.61it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 82%|████████▏ | 2470/3000 [08:09<01:21, 6.50it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 82%|████████▏ | 2472/3000 [08:10<01:41, 5.18it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 83%|████████▎ | 2476/3000 [08:10<01:17, 6.78it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 83%|████████▎ | 2478/3000 [08:10<01:05, 8.03it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 83%|████████▎ | 2480/3000 [08:11<01:08, 7.60it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 83%|████████▎ | 2483/3000 [08:11<01:12, 7.13it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 83%|████████▎ | 2484/3000 [08:12<01:34, 5.47it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 83%|████████▎ | 2490/3000 [08:12<00:57, 8.85it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 83%|████████▎ | 2495/3000 [08:13<01:00, 8.35it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 83%|████████▎ | 2497/3000 [08:13<01:12, 6.90it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 83%|████████▎ | 2498/3000 [08:13<01:21, 6.14it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 83%|████████▎ | 2501/3000 [08:14<01:14, 6.67it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 83%|████████▎ | 2504/3000 [08:14<00:58, 8.48it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 84%|████████▎ | 2510/3000 [08:14<00:41, 11.70it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 84%|████████▍ | 2514/3000 [08:15<01:16, 6.38it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 84%|████████▍ | 2517/3000 [08:16<01:09, 6.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 84%|████████▍ | 2522/3000 [08:16<00:55, 8.67it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 84%|████████▍ | 2526/3000 [08:17<00:48, 9.80it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 84%|████████▍ | 2530/3000 [08:17<00:35, 13.20it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 84%|████████▍ | 2532/3000 [08:18<01:15, 6.22it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 84%|████████▍ | 2534/3000 [08:18<01:20, 5.78it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 85%|████████▍ | 2538/3000 [08:19<00:59, 7.70it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 85%|████████▍ | 2540/3000 [08:19<00:55, 8.22it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 85%|████████▍ | 2543/3000 [08:19<00:47, 9.53it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 85%|████████▍ | 2546/3000 [08:19<00:53, 8.49it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 85%|████████▍ | 2548/3000 [08:20<01:02, 7.27it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 85%|████████▍ | 2549/3000 [08:20<01:18, 5.74it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 85%|████████▌ | 2551/3000 [08:20<01:14, 6.07it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 85%|████████▌ | 2555/3000 [08:21<01:10, 6.27it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 85%|████████▌ | 2560/3000 [08:22<00:49, 8.85it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 85%|████████▌ | 2562/3000 [08:22<00:45, 9.67it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 86%|████████▌ | 2565/3000 [08:22<00:40, 10.71it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 86%|████████▌ | 2567/3000 [08:22<00:44, 9.83it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 86%|████████▌ | 2569/3000 [08:23<00:52, 8.23it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 86%|████████▌ | 2570/3000 [08:23<01:00, 7.16it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 86%|████████▌ | 2573/3000 [08:23<00:52, 8.14it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 86%|████████▌ | 2575/3000 [08:23<00:57, 7.38it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 86%|████████▌ | 2579/3000 [08:24<00:49, 8.56it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 86%|████████▌ | 2582/3000 [08:24<00:36, 11.61it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 86%|████████▌ | 2584/3000 [08:24<00:36, 11.29it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 86%|████████▌ | 2586/3000 [08:25<01:03, 6.50it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 86%|████████▌ | 2587/3000 [08:25<01:14, 5.55it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 86%|████████▋ | 2589/3000 [08:25<01:03, 6.49it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 86%|████████▋ | 2594/3000 [08:26<00:38, 10.43it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 87%|████████▋ | 2599/3000 [08:26<00:30, 13.22it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 87%|████████▋ | 2601/3000 [08:26<00:35, 11.10it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 87%|████████▋ | 2603/3000 [08:27<00:48, 8.21it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 87%|████████▋ | 2606/3000 [08:27<01:05, 6.04it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 87%|████████▋ | 2608/3000 [08:28<01:12, 5.42it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 87%|████████▋ | 2610/3000 [08:28<00:56, 6.93it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 87%|████████▋ | 2613/3000 [08:28<00:54, 7.06it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 87%|████████▋ | 2617/3000 [08:28<00:32, 11.96it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 87%|████████▋ | 2619/3000 [08:29<00:52, 7.29it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 87%|████████▋ | 2624/3000 [08:29<00:40, 9.37it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 88%|████████▊ | 2626/3000 [08:30<00:46, 8.07it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 88%|████████▊ | 2629/3000 [08:30<00:36, 10.19it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 88%|████████▊ | 2631/3000 [08:30<00:40, 9.13it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 88%|████████▊ | 2636/3000 [08:31<00:50, 7.27it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 88%|████████▊ | 2638/3000 [08:31<00:44, 8.21it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 88%|████████▊ | 2642/3000 [08:32<00:38, 9.33it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 88%|████████▊ | 2644/3000 [08:32<00:35, 9.89it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 88%|████████▊ | 2646/3000 [08:32<00:35, 9.90it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 88%|████████▊ | 2648/3000 [08:32<00:36, 9.73it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 88%|████████▊ | 2652/3000 [08:33<00:42, 8.14it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 89%|████████▊ | 2656/3000 [08:34<00:44, 7.80it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 89%|████████▊ | 2657/3000 [08:34<00:54, 6.29it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 89%|████████▊ | 2662/3000 [08:35<00:44, 7.61it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 89%|████████▉ | 2664/3000 [08:35<00:40, 8.38it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 89%|████████▉ | 2668/3000 [08:35<00:44, 7.54it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 89%|████████▉ | 2670/3000 [08:36<00:48, 6.86it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 89%|████████▉ | 2671/3000 [08:36<00:58, 5.60it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098109.104576336)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098110.94024912)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098110.021616306)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Evaluating workflow: 89%|████████▉ | 2673/3000 [08:37<01:20, 4.05it/s]Unclosed connector\n", "connections: ['deque([(, 2098108.611182403)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098108.206397917)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098108.462960852)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098111.316786756)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098111.315327314)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098111.857567793)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098110.392869282)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098109.97254053)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098110.035778308)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098111.853564183)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098113.786843997)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098112.097752498)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098116.00711064)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098115.03295858)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098114.175018545)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098113.577617787)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098116.05006713)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098116.598952533)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098116.470530476)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098118.774994638)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098118.476259675)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098115.918939489)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098116.04891635)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098118.646016483)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098119.104617529)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098119.335066346)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098118.186296623)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098118.445104048)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098121.789197038)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098123.911014312)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098122.455535933)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098122.181495056)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098123.048545598)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098123.43422142)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098122.208080147)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098125.327555889)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098125.544978684)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098125.423804878)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098126.579291782)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098124.436350185)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098125.953515657)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098128.118741838)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098127.836859504)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098147.650970831)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098129.137285872)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098137.45342093)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098138.117452906)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098146.911447775)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098147.660761308)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098147.80300844)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098129.393166719)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098147.671596976)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098147.793591969)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2098147.664710948)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098153.196849441)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098148.066618099)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098137.975337939)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098147.759311502)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098153.51830271)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098155.842645036)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098155.721523681)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098147.674339935)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098154.550152519)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098157.111807916)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098153.45165629)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098154.443715035)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098107.181861736)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098107.851282819)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098108.528397008)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2098107.682037416)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098107.99177641)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098108.056078444)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098108.375311364)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098108.751581636)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098109.198870798)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098109.33497851)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098108.928737144)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098109.946642571)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098109.871522726)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098111.248834927)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098110.150759189)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098110.194156382)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098110.376130203)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098110.956642871)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098111.516964738)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098110.722351378)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098111.29838616)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098110.9125485)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098111.657567509)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098111.786448783)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098111.787626584)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098111.796645978)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098111.667991467)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098113.060114124)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098112.841480191)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098113.630326232)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098113.186466244)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098113.816696947)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098114.682771857)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098113.524381453)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098114.056825341)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098113.674893002)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098113.964138747)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098113.784653281)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098115.020211909)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098114.653706095)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098114.599106168)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098116.226305845)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098117.665774972)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098115.236939347)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098115.844978683)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2098116.534678309)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098116.024431945)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098116.343508813)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098116.756062607)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098117.509065591)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098118.362527841)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098117.355739764)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098116.73303494)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098117.229130615)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098117.977314607)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098118.700436158)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098118.112221479)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098119.232822251)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098118.801210458)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098118.929097707)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098119.016714181)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098119.205571915)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098120.036213915)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098119.811149006)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098120.185165816)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098119.487635803)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098120.276975028)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098120.884072565)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098120.482009576)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098120.72647362)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098121.600618099)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098120.885328072)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098120.693335665)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098121.22784502)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098121.857428134)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098121.491454137)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098121.984906497)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098121.187377618)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098121.967972265)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098122.04764629)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098122.612012656)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098122.739371234)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098122.671709481)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098123.639338489)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098123.901659493)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098123.629027821)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098124.069140518)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2098124.101703554)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098125.29317033)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098123.598524533)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098124.258646126)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098125.315497543)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098125.387232123)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098124.826777164)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098125.090594204)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098126.043548199)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098125.309894484)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098125.605099999)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098127.85390151)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098126.022878335)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098125.804089461)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098127.521073518)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098126.859116575)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098127.659595446)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098126.726764252)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098127.310784108)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098126.828134353)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098127.423109462)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098127.223038429)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098127.565437996)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098127.680372649)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098128.137082779)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098127.841095441)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098127.602707078)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098128.037372912)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098137.48086735)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098129.864747705)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098129.226708574)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098129.379397907)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098147.657678778)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098129.11579377)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098147.685558556)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098129.696739406)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098129.850341063)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098137.393530223)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098147.634539987)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098136.208461943)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098137.741436363)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098115.447729757)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098124.472842586)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098127.909268414)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098157.187366976)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098155.793528981)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098155.470761302)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098157.480885987)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098157.149987243)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098155.670782915)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098154.107654493)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098157.145816788)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098157.395257142)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098155.950789437)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098154.201437093)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098158.210153508)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098155.649604093)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098156.108865257)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098157.357136734)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098157.521728621)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098158.227645326)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098155.602089613)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098158.831696302)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098157.460896119)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098159.206891741)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098161.405255194)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098160.543839614)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098159.806037367)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098159.425335577)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098159.885888301)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098160.154965199)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Evaluating workflow: 89%|████████▉ | 2674/3000 [08:37<01:13, 4.43it/s]Unclosed connector\n", "connections: ['deque([(, 2098160.456515157)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098159.308144068)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098159.819315306)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098159.395272469)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098159.784580738)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098159.871079564)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098160.825123169)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098158.750326081)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098158.1049178)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098157.931411302)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098161.57438148)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098161.083106683)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098161.483282342)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098162.011305738)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098161.732779669)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098161.406483453)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098161.474891075)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098162.378797432)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098161.946237466)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098163.419414502)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098162.972039319)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098163.887557145)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098163.202515145)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098160.827616855)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098161.063959858)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098163.252456837)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2098161.223349306)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098163.27605466)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098163.525389523)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098163.564850548)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098164.964892038)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098163.083046347)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098165.16929698)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098164.790988259)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098165.243398326)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098165.434490764)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098165.267533576)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098164.580595426)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098162.708459912)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098162.907910239)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098163.037140212)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098164.624601439)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098165.906425175)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098165.5505104)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098165.857479482)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098167.101062546)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098165.211702164)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098165.729858624)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098165.422617668)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098167.592894253)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098166.851701868)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098166.988879394)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098165.090151865)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2098165.2514549)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098167.112961143)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098165.191940682)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098167.173186327)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098168.80088353)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098167.934748848)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098167.869549725)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098167.45986365)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098168.72096344)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098168.010214865)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098168.497534369)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098168.389786169)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098168.330011741)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098168.577584428)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098168.831528595)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098168.96390826)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098168.34897085)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098167.152548769)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098170.1477666)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098167.418082123)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098169.809852778)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098169.969888443)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098169.922377142)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098169.908825718)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098169.931056788)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098171.709287126)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098169.762616608)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098170.055034309)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098170.845248116)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098170.619771594)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098171.150328517)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098170.060487065)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098170.935119056)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098171.443144011)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098170.270943646)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098170.9283121)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098172.388472104)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098171.921628331)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098171.7458594)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098172.615839945)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098172.560111457)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098172.280638829)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098172.414147911)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098172.003835048)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098172.265287092)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098172.226967546)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098172.634537024)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098173.518614782)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098172.38264898)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098173.080595793)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098173.126219905)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098170.748400004)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098173.235274736)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098173.067342438)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098173.157866946)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098174.169935266)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098174.305764432)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2098173.797717287)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098174.033436823)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098174.556370904)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098174.805553088)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098174.529775865)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098174.769467278)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098174.454079486)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098176.162007619)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098174.702822003)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098175.392805812)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098174.892859148)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098176.266832725)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098176.502918091)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098175.078235701)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098176.528483272)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098176.687573977)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098176.82292156)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098176.648766403)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098176.040197061)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098177.668583174)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098176.454676246)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098177.85734937)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098174.777667835)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098174.953849802)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098177.291528398)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098176.570015233)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098177.09920776)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098177.595008337)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2098177.901580486)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098178.133309513)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098177.768013636)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098178.632798856)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098178.607707306)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098178.550161229)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098178.084663473)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098178.422433057)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098178.905252807)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098178.061605239)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098179.185159865)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098178.608877738)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098178.141535458)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098179.572533096)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098179.755733617)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098179.279947273)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098179.68884163)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098181.084177496)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098180.222145762)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098179.857036248)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098182.017279218)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098180.1556675)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098180.429506507)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098180.401672903)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098180.453643562)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098180.708561621)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098180.907551457)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098178.587060246)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098180.805764742)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098181.500910794)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098181.627317749)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098182.103318928)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098182.431078728)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098182.296697132)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098183.833812756)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098181.699695076)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098181.973020154)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098182.405961376)])']\n", "connector: \n" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098182.549135784)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098183.012700987)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098182.727863805)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098182.981257403)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098183.386875673)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098180.871317536)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098183.135457976)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098183.645792724)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098184.054763177)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098183.12708377)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098184.097229984)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098184.175698596)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098184.113955841)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098185.995828195)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098184.56213583)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098184.229643754)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098184.615399138)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098184.607999452)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098184.526950148)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098184.279853759)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098184.347807459)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098185.170066826)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098186.027662312)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098186.000093118)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098185.937942591)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098185.968839549)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098186.306249692)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098186.08071476)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098186.63403781)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098186.336516461)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098186.849261454)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098186.971015728)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098188.221727638)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098187.381105814)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098188.052071636)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098187.886606058)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098188.205599671)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098187.426964111)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098188.231033378)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098188.176140014)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098187.793415731)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098188.034533863)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098188.738876359)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098188.307665008)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098188.316390248)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098188.875187546)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098190.209135575)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098189.558059946)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098189.411302871)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098189.419132522)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098189.514827948)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098187.895876583)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098190.070867923)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098189.820114019)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098190.456961086)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098198.086939122)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098190.512704845)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098198.047773364)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098190.463112038)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098190.553630739)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098189.899807232)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098198.000657156)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098197.846792326)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098197.361103266)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098198.33158583)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098196.126767623)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098197.772592212)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098208.138291513)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Evaluating workflow: 89%|████████▉ | 2675/3000 [08:37<01:05, 4.94it/s]Unclosed connector\n", "connections: ['deque([(, 2098197.679929783)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098197.541697337)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098196.524513216)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098198.134568092)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098197.97797717)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098197.581948876)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098197.640955764)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098197.292827532)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098207.774932757)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098197.337224512)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098197.587676283)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098198.266281103)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098198.722810621)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098215.228903495)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098207.727245856)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098207.729194952)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098214.632098885)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098214.440387393)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098207.681696987)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098207.794133123)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098207.811070087)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098213.819975186)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098213.34992079)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098214.379508124)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098214.587389665)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098214.450071594)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098214.51887836)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098217.925725613)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098207.920017338)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098213.80519511)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098216.069808346)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098207.98027027)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098217.45872413)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098215.782986066)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098216.078326856)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098217.811885494)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098219.4302656)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098215.987463159)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098215.846543943)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098217.502634664)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098218.478932303)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098216.97267545)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098218.197479559)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098218.968885803)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed connector\n", "connections: ['deque([(, 2098217.32509098)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098218.670108198)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098218.59018761)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098219.58824351)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098217.81981533)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098218.315075717)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098217.708061737)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098217.568235975)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098219.054783905)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098219.74892719)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098219.979083964)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098220.108486654)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098221.793341497)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098221.341132516)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098220.180471015)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098220.264535321)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098219.965461144)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098220.072049052)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098220.870551148)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098222.384179315)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098224.243933303)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098222.158300894)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098219.2994752)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098219.394587379)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098221.966281831)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098220.209766368)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098221.735255196)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098221.745183887)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098222.82143341)])']\n", "connector: \n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098222.062252783)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098221.926498149)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098223.481971615)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098222.35932476)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098222.450334292)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098223.316074476)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098223.272806562)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098221.445045145)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098223.237287582)])']\n", "connector: \n", "Unclosed connector\n", "connections: ['deque([(, 2098224.393475381)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098223.665740301)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098224.331696492)])']\n", "connector: \n", "Unclosed client session\n", "client_session: \n", "Unclosed connector\n", "connections: ['deque([(, 2098223.524224535)])']\n", "connector: \n", "Evaluating workflow: 89%|████████▉ | 2676/3000 [08:37<00:58, 5.52it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 89%|████████▉ | 2681/3000 [08:38<00:31, 10.06it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 89%|████████▉ | 2684/3000 [08:38<00:29, 10.82it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 90%|████████▉ | 2686/3000 [08:38<00:25, 12.15it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 90%|████████▉ | 2688/3000 [08:38<00:31, 10.02it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 90%|████████▉ | 2690/3000 [08:38<00:31, 9.75it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 90%|████████▉ | 2694/3000 [08:39<00:50, 6.01it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 90%|████████▉ | 2696/3000 [08:45<04:29, 1.13it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 90%|████████▉ | 2699/3000 [08:45<03:07, 1.60it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:41:44.854\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0256631', 'INHIB_AVE': 2.62, 'SMILES': 'c1([nH]c(c2n1)ccc(c2)\\\\N=C\\\\c3cccc(Br)c3)c4ccc(cc4Cl)Cl', 'HIT': 0.0, 'question': 'Does c1([nH]c(c2n1)ccc(c2)\\\\N=C\\\\c3cccc(Br)c3)c4ccc(cc4Cl)Cl demonstrate bactericidal effects against E. coli in vitro?', 'answer': 'No', '_id': 46324}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 90%|█████████ | 2700/3000 [08:56<10:39, 2.13s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:41:44.863\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0288843', 'INHIB_AVE': 0.72, 'SMILES': 'C(N(C(=O)OC(C)(C)C)CC1)C(C1)CC(=O)NCC#C', 'HIT': 0.0, 'question': 'Can exposure to C(N(C(=O)OC(C)(C)C)CC1)C(C1)CC(=O)NCC#C lead to reduced viability of E. coli?', 'answer': 'No', '_id': 54625}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:41:45.000\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0291484', 'INHIB_AVE': 4.55, 'SMILES': 'n1(nc(c2C(=O)Nc(cccc3C(F)(F)F)c3)C)c2nc(c4ccc(cc4)Cl)cc1C(F)(F)F', 'HIT': 0.0, 'question': 'Is n1(nc(c2C(=O)Nc(cccc3C(F)(F)F)c3)C)c2nc(c4ccc(cc4)Cl)cc1C(F)(F)F capable of inhibiting the growth or survival of E. coli?', 'answer': 'No', '_id': 37664}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 90%|█████████ | 2702/3000 [08:56<07:26, 1.50s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:41:45.159\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0255999', 'INHIB_AVE': 4.77, 'SMILES': 'C(C(OC)=O)(C1c(cccc2[N+](=O)[O-])c2)=C(N)N(C(CCCC3=O)=C13)c4ccc(cc4)Cl', 'HIT': 0.0, 'question': 'Does C(C(OC)=O)(C1c(cccc2[N+](=O)[O-])c2)=C(N)N(C(CCCC3=O)=C13)c4ccc(cc4)Cl demonstrate bactericidal effects against E. coli in vitro?', 'answer': 'No', '_id': 36688}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 6 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 90%|█████████ | 2703/3000 [08:56<06:13, 1.26s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:41:45.178\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0287254', 'INHIB_AVE': 3.49, 'SMILES': 'C(C(OCC)=O)(Nc1ccccn1)(C(F)(F)F)NC(OCC)=O', 'HIT': 0.0, 'question': 'Can exposure to C(C(OCC)=O)(Nc1ccccn1)(C(F)(F)F)NC(OCC)=O lead to reduced viability of E. coli?', 'answer': 'No', '_id': 42415}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 6 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:41:45.292\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0223606', 'INHIB_AVE': 1.37, 'SMILES': 'c1(CC)c(\\\\C=N\\\\O)c(CC)cc(CC)c1\\\\C=N\\\\O', 'HIT': 0.0, 'question': 'Is c1(CC)c(\\\\C=N\\\\O)c(CC)cc(CC)c1\\\\C=N\\\\O capable of inhibiting the growth or survival of E. coli?', 'answer': 'No', '_id': 51852}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 6 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 90%|█████████ | 2705/3000 [08:56<04:11, 1.17it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:41:45.336\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0319201', 'INHIB_AVE': 2.06, 'SMILES': 'c1(NC(=O)c(ccc(c2Cl)Cl)c2)sc(c(C)c1C(OCC)=O)C(C)=O', 'HIT': 0.0, 'question': 'Can c1(NC(=O)c(ccc(c2Cl)Cl)c2)sc(c(C)c1C(OCC)=O)C(C)=O act as an antibacterial agent targeting Escherichia coli?', 'answer': 'No', '_id': 48810}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2026-01-10 17:41:45.375\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0161298', 'INHIB_AVE': -2.38, 'SMILES': 'c(C(=O)Nc(cccc1C(F)(F)F)c1)(c2N)sc3c2cc(c4n3)CCCCC4', 'HIT': 0.0, 'question': 'Does c(C(=O)Nc(cccc1C(F)(F)F)c1)(c2N)sc3c2cc(c4n3)CCCCC4 show antimicrobial efficacy specifically against Escherichia coli strains?', 'answer': 'No', '_id': 65494}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:41:45.630\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0283411', 'INHIB_AVE': -2.74, 'SMILES': 'N1(c2ccccc2)C(=O)c3c(N=C1SCC(=O)N4CCCCC4)cccc3', 'HIT': 0.0, 'question': 'Can exposure to N1(c2ccccc2)C(=O)c3c(N=C1SCC(=O)N4CCCCC4)cccc3 lead to reduced viability of E. coli?', 'answer': 'No', '_id': 66481}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 90%|█████████ | 2710/3000 [08:56<02:02, 2.37it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:41:45.637\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0232361', 'INHIB_AVE': -0.8, 'SMILES': 'C1(C(=O)Nc(cccc2[N+](=O)[O-])c2)=Cc(cccc3CC=C)c3OC1=O', 'HIT': 0.0, 'question': 'Can C1(C(=O)Nc(cccc2[N+](=O)[O-])c2)=Cc(cccc3CC=C)c3OC1=O effectively kill or suppress Escherichia coli cells?', 'answer': 'No', '_id': 60433}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:41:45.666\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:41:45.677\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0259883', 'INHIB_AVE': 2.56, 'SMILES': 'C1(C)(C(OCC)=O)C=C(C(=O)N1c2ccccc2)Nc(cc3)ccc3C(O)=O', 'HIT': 0.0, 'question': 'Can exposure to C1(C)(C(OCC)=O)C=C(C(=O)N1c2ccccc2)Nc(cc3)ccc3C(O)=O lead to reduced viability of E. coli?', 'answer': 'No', '_id': 46604}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "\u001b[32m2026-01-10 17:41:45.717\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0252662', 'INHIB_AVE': 5.79, 'SMILES': 'C1(C(Oc(c23)cc(c(C)c2)NCC)=CC(=[N+]([H])CC)C(C)=C1)=C3c4ccccc4C(OCC)=O.C(C(C)C)(=O)[O-]', 'HIT': 0.0, 'question': 'Does C1(C(Oc(c23)cc(c(C)c2)NCC)=CC(=[N+]([H])CC)C(C)=C1)=C3c4ccccc4C(OCC)=O.C(C(C)C)(=O)[O-] demonstrate bactericidal effects against E. coli in vitro?', 'answer': 'No', '_id': 32171}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 90%|█████████ | 2715/3000 [08:57<01:15, 3.78it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:41:51.049\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 91%|█████████ | 2717/3000 [09:02<03:29, 1.35it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 91%|█████████ | 2718/3000 [09:02<03:11, 1.47it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "\u001b[32m2026-01-10 17:41:51.848\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0232691', 'INHIB_AVE': 4.8, 'SMILES': 'C1(C(=O)OC2CCCC(C)C2)=C(C)NC(=O)NC1c3ccc(cc3)OCCC', 'HIT': 0.0, 'question': 'Does C1(C(=O)OC2CCCC(C)C2)=C(C)NC(=O)NC1c3ccc(cc3)OCCC demonstrate bactericidal effects against E. coli in vitro?', 'answer': 'No', '_id': 36531}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 91%|█████████ | 2721/3000 [09:03<02:17, 2.03it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:41:52.095\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:41:52.301\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 91%|█████████ | 2723/3000 [09:03<01:45, 2.63it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:41:52.312\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 91%|█████████ | 2725/3000 [09:03<01:21, 3.38it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 91%|█████████ | 2726/3000 [09:04<01:51, 2.45it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:41:53.617\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 91%|█████████ | 2727/3000 [09:04<01:38, 2.76it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:41:53.637\u001b[0m | \u001b[31m\u001b[1mERROR \u001b[0m | \u001b[36mevoagentx.workflow.workflow\u001b[0m:\u001b[36masync_execute\u001b[0m:\u001b[36m104\u001b[0m - \u001b[31m\u001b[1mAn Error occurs when executing the workflow: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 1 second. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 91%|█████████ | 2731/3000 [09:05<00:49, 5.46it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 91%|█████████ | 2734/3000 [09:05<00:45, 5.84it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 91%|█████████▏| 2738/3000 [09:06<00:54, 4.77it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 91%|█████████▏| 2741/3000 [09:07<00:52, 4.95it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 92%|█████████▏| 2745/3000 [09:07<00:37, 6.88it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 92%|█████████▏| 2749/3000 [09:08<00:38, 6.47it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 92%|█████████▏| 2751/3000 [09:08<00:33, 7.50it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 92%|█████████▏| 2755/3000 [09:08<00:23, 10.22it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 92%|█████████▏| 2759/3000 [09:09<00:21, 11.26it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 92%|█████████▏| 2761/3000 [09:09<00:34, 6.88it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 92%|█████████▏| 2763/3000 [09:10<00:46, 5.04it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 92%|█████████▏| 2767/3000 [09:10<00:35, 6.58it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 92%|█████████▏| 2771/3000 [09:11<00:26, 8.70it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 92%|█████████▏| 2773/3000 [09:11<00:24, 9.27it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 93%|█████████▎| 2778/3000 [09:12<00:35, 6.27it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 93%|█████████▎| 2781/3000 [09:12<00:25, 8.45it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 93%|█████████▎| 2785/3000 [09:13<00:25, 8.33it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 93%|█████████▎| 2790/3000 [09:14<00:30, 6.96it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 93%|█████████▎| 2796/3000 [09:14<00:25, 8.10it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 93%|█████████▎| 2798/3000 [09:15<00:27, 7.23it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 93%|█████████▎| 2802/3000 [09:15<00:21, 9.33it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 94%|█████████▎| 2806/3000 [09:16<00:20, 9.38it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 94%|█████████▎| 2808/3000 [09:16<00:19, 9.62it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 94%|█████████▎| 2810/3000 [09:17<00:43, 4.40it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 94%|█████████▍| 2814/3000 [09:17<00:27, 6.83it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 94%|█████████▍| 2818/3000 [09:17<00:19, 9.20it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 94%|█████████▍| 2820/3000 [09:17<00:19, 9.23it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 94%|█████████▍| 2823/3000 [09:18<00:15, 11.13it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 94%|█████████▍| 2825/3000 [09:18<00:19, 8.75it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 94%|█████████▍| 2827/3000 [09:19<00:28, 6.15it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 94%|█████████▍| 2830/3000 [09:19<00:25, 6.80it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 94%|█████████▍| 2832/3000 [09:19<00:20, 8.03it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 94%|█████████▍| 2834/3000 [09:19<00:18, 8.94it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 95%|█████████▍| 2838/3000 [09:20<00:22, 7.21it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 95%|█████████▍| 2840/3000 [09:20<00:19, 8.13it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 95%|█████████▍| 2844/3000 [09:21<00:21, 7.37it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 95%|█████████▍| 2847/3000 [09:21<00:25, 6.05it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 95%|█████████▌| 2850/3000 [09:22<00:18, 7.93it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 95%|█████████▌| 2853/3000 [09:22<00:12, 11.36it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 95%|█████████▌| 2855/3000 [09:22<00:14, 9.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 95%|█████████▌| 2857/3000 [09:23<00:19, 7.29it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 95%|█████████▌| 2859/3000 [09:23<00:24, 5.86it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 95%|█████████▌| 2860/3000 [09:23<00:29, 4.77it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 95%|█████████▌| 2861/3000 [09:24<00:29, 4.77it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 95%|█████████▌| 2864/3000 [09:24<00:23, 5.90it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 96%|█████████▌| 2869/3000 [09:24<00:13, 9.66it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 96%|█████████▌| 2872/3000 [09:25<00:10, 12.09it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 96%|█████████▌| 2877/3000 [09:25<00:09, 12.95it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 96%|█████████▌| 2879/3000 [09:25<00:13, 8.98it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 96%|█████████▌| 2882/3000 [09:26<00:18, 6.32it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 96%|█████████▌| 2885/3000 [09:26<00:14, 7.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 96%|█████████▌| 2887/3000 [09:27<00:15, 7.27it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 96%|█████████▋| 2893/3000 [09:27<00:10, 10.11it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 96%|█████████▋| 2895/3000 [09:27<00:08, 11.72it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 97%|█████████▋| 2899/3000 [09:28<00:14, 6.97it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 97%|█████████▋| 2905/3000 [09:29<00:12, 7.82it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 97%|█████████▋| 2909/3000 [09:29<00:08, 10.17it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 97%|█████████▋| 2912/3000 [09:29<00:08, 10.39it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 97%|█████████▋| 2914/3000 [09:30<00:11, 7.55it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 97%|█████████▋| 2916/3000 [09:30<00:13, 6.35it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 97%|█████████▋| 2921/3000 [09:31<00:08, 9.74it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 97%|█████████▋| 2923/3000 [09:31<00:08, 8.66it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 98%|█████████▊| 2926/3000 [09:31<00:07, 9.60it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 98%|█████████▊| 2930/3000 [09:32<00:07, 9.60it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 98%|█████████▊| 2933/3000 [09:32<00:09, 7.13it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 98%|█████████▊| 2938/3000 [09:33<00:05, 10.85it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 98%|█████████▊| 2941/3000 [09:33<00:06, 8.43it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 98%|█████████▊| 2943/3000 [09:33<00:06, 8.21it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 98%|█████████▊| 2946/3000 [09:34<00:07, 7.24it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 98%|█████████▊| 2947/3000 [09:34<00:07, 7.29it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 98%|█████████▊| 2949/3000 [09:34<00:08, 5.79it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 98%|█████████▊| 2952/3000 [09:35<00:06, 7.06it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 98%|█████████▊| 2954/3000 [09:35<00:07, 6.49it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 99%|█████████▊| 2959/3000 [09:35<00:03, 11.62it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 99%|█████████▉| 2964/3000 [09:36<00:02, 14.55it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 99%|█████████▉| 2966/3000 [09:37<00:08, 4.09it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 99%|█████████▉| 2968/3000 [09:37<00:06, 4.92it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 99%|█████████▉| 2970/3000 [09:38<00:06, 4.54it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 99%|█████████▉| 2974/3000 [09:38<00:03, 6.84it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 99%|█████████▉| 2976/3000 [09:38<00:03, 7.44it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 99%|█████████▉| 2980/3000 [09:39<00:02, 9.20it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 99%|█████████▉| 2983/3000 [09:40<00:03, 4.67it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 100%|█████████▉| 2985/3000 [09:40<00:02, 5.39it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 100%|█████████▉| 2988/3000 [09:45<00:08, 1.34it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 100%|█████████▉| 2990/3000 [09:45<00:05, 1.76it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:42:45.436\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0229799', 'INHIB_AVE': 1.44, 'SMILES': 'c(C(OCC)=O)(s1)c(NC(C)=O)c2c1\\\\C(=N/O)\\\\C(CC2)Sc3ccccc3', 'HIT': 0.0, 'question': 'Is c(C(OCC)=O)(s1)c(NC(C)=O)c2c1\\\\C(=N/O)\\\\C(CC2)Sc3ccccc3 capable of inhibiting the growth or survival of E. coli?', 'answer': 'No', '_id': 51532}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 6 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 100%|█████████▉| 2994/3000 [09:56<00:08, 1.44s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n", "\u001b[32m2026-01-10 17:42:45.760\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0219498', 'INHIB_AVE': 0.61, 'SMILES': 'c1(c(C)ccc(Cl)c1)\\\\N=C\\\\C=C(/O)\\\\c2ccc(cc2)OC', 'HIT': 0.0, 'question': 'Can exposure to c1(c(C)ccc(Cl)c1)\\\\N=C\\\\C=C(/O)\\\\c2ccc(cc2)OC lead to reduced viability of E. coli?', 'answer': 'No', '_id': 55046}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 100%|█████████▉| 2996/3000 [09:57<00:04, 1.05s/it]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "\u001b[32m2026-01-10 17:42:45.779\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[36mevoagentx.evaluators.evaluator\u001b[0m:\u001b[36m_evaluate_single_example\u001b[0m:\u001b[36m205\u001b[0m - \u001b[33m\u001b[1mError evaluating example and set the metrics to None:\n", "Example: {'COADD_ID': 'CO-ADD:0235054', 'INHIB_AVE': -5.63, 'SMILES': 'c1(oc(c2n1)ccc(c2)\\\\N=C\\\\c(cc(Br)cc3[N+](=O)[O-])c3O)c4c(O)cc(c5c4)cccc5', 'HIT': 0.0, 'question': 'Is c1(oc(c2n1)ccc(c2)\\\\N=C\\\\c(cc(Br)cc3[N+](=O)[O-])c3O)c4c(O)cc(c5c4)cccc5 capable of inhibiting the growth or survival of E. coli?', 'answer': 'No', '_id': 73057}\n", "Error: Error during single_generate_async: litellm.RateLimitError: AzureException RateLimitError - Your requests to gpt-4o-mini for gpt-4o-mini in East US 2 have exceeded the token rate limit for your current AIServices S0 pricing tier. This request was for ChatCompletions_Create under Azure OpenAI API version 2025-01-01-preview. Please retry after 5 seconds. To increase your default rate limit, visit: https://aka.ms/oai/quotaincrease.\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\r", "Evaluating workflow: 100%|█████████▉| 2998/3000 [09:57<00:01, 1.32it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "metrics {'f1': 0, 'em': 0.0, 'acc': 0.0}\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "Evaluating workflow: 100%|██████████| 3000/3000 [09:57<00:00, 5.02it/s]" ] }, { "name": "stdout", "output_type": "stream", "text": [ "metrics {'f1': 0, 'em': 0.0, 'acc': 1.0}\n", "Evaluation metrics: {'f1': 0.0, 'em': 0.0, 'acc': 0.442530755711775}\n", "\u001b[32m2026-01-10 17:42:46.239\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.workflow.workflow_graph\u001b[0m:\u001b[36msave_module\u001b[0m:\u001b[36m1204\u001b[0m - \u001b[1mSaving SEWWorkFlowGraph to debug/optimized_sew_workflow_update_molqa.json\u001b[0m\n", "\u001b[32m2026-01-10 17:42:46.240\u001b[0m | \u001b[1mINFO \u001b[0m | \u001b[36mevoagentx.utils.utils\u001b[0m:\u001b[36mmake_parent_folder\u001b[0m:\u001b[36m19\u001b[0m - \u001b[1mcreating folder debug ...\u001b[0m\n" ] }, { "name": "stderr", "output_type": "stream", "text": [ "\n" ] } ], "source": [ "# obtain SEWOptimizer after having more roles\n", "optimizer = SEWOptimizer(\n", " graph=sew_graph, \n", " evaluator=evaluator, \n", " llm=llm, \n", " max_steps=20,\n", " eval_rounds=1, \n", " repr_scheme=\"python\", \n", " optimize_mode=\"all\", \n", " order=\"zero-order\",\n", " max_rounds=20,\n", ")\n", "\n", "# with suppress_logger_info():\n", "# metrics = optimizer.evaluate(dataset=humaneval, eval_mode=\"test\")\n", "# print(\"Evaluation metrics: \", metrics)\n", "\n", "# optimize the SEW workflow\n", "optimizer.optimize(dataset=benchmark)\n", "\n", "# evaluate the optimized SEW workflow\n", "with suppress_logger_info():\n", " metrics = optimizer.evaluate(dataset=benchmark, eval_mode=\"test\")\n", "print(\"Evaluation metrics: \", metrics)\n", "\n", "# save the optimized SEW workflow\n", "optimizer.save(\"debug/optimized_sew_workflow_update_molqa.json\")" ] }, { "cell_type": "code", "execution_count": 14, "id": "c5f272e3", "metadata": {}, "outputs": [ { "data": { "text/plain": [ "{'f1': 0.0, 'em': 0.0, 'acc': 0.442530755711775}" ] }, "execution_count": 14, "metadata": {}, "output_type": "execute_result" } ], "source": [ "metrics" ] }, { "cell_type": "code", "execution_count": null, "id": "360eeccc", "metadata": {}, "outputs": [], "source": [] }, { "cell_type": "code", "execution_count": null, "id": "42106cd4", "metadata": {}, "outputs": [], "source": [] }, { "cell_type": "code", "execution_count": null, "id": "33a7a554", "metadata": {}, "outputs": [], "source": [] } ], "metadata": { "kernelspec": { "display_name": "Python 3 (ipykernel)", "language": "python", "name": "python3" }, "language_info": { "codemirror_mode": { "name": "ipython", "version": 3 }, "file_extension": ".py", "mimetype": "text/x-python", "name": "python", "nbconvert_exporter": "python", "pygments_lexer": "ipython3", "version": "3.11.13" } }, "nbformat": 4, "nbformat_minor": 5 }