diff --git a/agentgraph/__pycache__/__init__.cpython-311.pyc b/agentgraph/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index 0805b6b430aacf3d2ba628fe8f0f36e80bf66acd..0000000000000000000000000000000000000000 Binary files a/agentgraph/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/__pycache__/__init__.cpython-312.pyc b/agentgraph/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index 721c14109c5a2b3cf1ee99143190017bb0e1641e..0000000000000000000000000000000000000000 Binary files a/agentgraph/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/__pycache__/__init__.cpython-313.pyc b/agentgraph/__pycache__/__init__.cpython-313.pyc deleted file mode 100644 index 4cad4a7f2935c9d7489714233e914970c267815d..0000000000000000000000000000000000000000 Binary files a/agentgraph/__pycache__/__init__.cpython-313.pyc and /dev/null differ diff --git a/agentgraph/__pycache__/pipeline.cpython-311.pyc b/agentgraph/__pycache__/pipeline.cpython-311.pyc deleted file mode 100644 index 3502ddc519d1d4833f0c6a14fbacb114ba3fc1bd..0000000000000000000000000000000000000000 Binary files a/agentgraph/__pycache__/pipeline.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/__pycache__/pipeline.cpython-312.pyc b/agentgraph/__pycache__/pipeline.cpython-312.pyc deleted file mode 100644 index 991a626032096cc645c32bffd033d099bcc249e3..0000000000000000000000000000000000000000 Binary files a/agentgraph/__pycache__/pipeline.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/__pycache__/sdk.cpython-312.pyc b/agentgraph/__pycache__/sdk.cpython-312.pyc deleted file mode 100644 index 24f9e1cf42b94efa9831db168194cc740f874e9e..0000000000000000000000000000000000000000 Binary files a/agentgraph/__pycache__/sdk.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/causal/__pycache__/__init__.cpython-311.pyc b/agentgraph/causal/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index 077980b008141750729958da0e4480548da067a1..0000000000000000000000000000000000000000 Binary files a/agentgraph/causal/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/causal/__pycache__/__init__.cpython-312.pyc b/agentgraph/causal/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index 15087bd5b770efaf36d9b69765350d10304ea497..0000000000000000000000000000000000000000 Binary files a/agentgraph/causal/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/causal/__pycache__/causal_interface.cpython-311.pyc b/agentgraph/causal/__pycache__/causal_interface.cpython-311.pyc deleted file mode 100644 index 1b25f8b5a82675013e69d5e30df3b7f50fcdd9a2..0000000000000000000000000000000000000000 Binary files a/agentgraph/causal/__pycache__/causal_interface.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/causal/__pycache__/causal_interface.cpython-312.pyc b/agentgraph/causal/__pycache__/causal_interface.cpython-312.pyc deleted file mode 100644 index 721b7f98df299183a3ec95bdd4b736ded610a490..0000000000000000000000000000000000000000 Binary files a/agentgraph/causal/__pycache__/causal_interface.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/causal/__pycache__/component_analysis.cpython-311.pyc b/agentgraph/causal/__pycache__/component_analysis.cpython-311.pyc deleted file mode 100644 index c331775c19421996893ca02f14a8d4ed8b562b5d..0000000000000000000000000000000000000000 Binary files a/agentgraph/causal/__pycache__/component_analysis.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/causal/__pycache__/component_analysis.cpython-312.pyc b/agentgraph/causal/__pycache__/component_analysis.cpython-312.pyc deleted file mode 100644 index a7cac1f7f1dfd50564e4c4342a3c520c19bbb5d4..0000000000000000000000000000000000000000 Binary files a/agentgraph/causal/__pycache__/component_analysis.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/causal/__pycache__/dowhy_analysis.cpython-311.pyc b/agentgraph/causal/__pycache__/dowhy_analysis.cpython-311.pyc deleted file mode 100644 index 1de468364e69b2f37e0fce3551070a9b0bbc44ec..0000000000000000000000000000000000000000 Binary files a/agentgraph/causal/__pycache__/dowhy_analysis.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/causal/__pycache__/dowhy_analysis.cpython-312.pyc b/agentgraph/causal/__pycache__/dowhy_analysis.cpython-312.pyc deleted file mode 100644 index ab19af9fd41e5ef791a6507d0db4db8fe1421f77..0000000000000000000000000000000000000000 Binary files a/agentgraph/causal/__pycache__/dowhy_analysis.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/causal/__pycache__/graph_analysis.cpython-311.pyc b/agentgraph/causal/__pycache__/graph_analysis.cpython-311.pyc deleted file mode 100644 index f95bec52b8891c9c6a63f98abecda801fdba4b8b..0000000000000000000000000000000000000000 Binary files a/agentgraph/causal/__pycache__/graph_analysis.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/causal/__pycache__/graph_analysis.cpython-312.pyc b/agentgraph/causal/__pycache__/graph_analysis.cpython-312.pyc deleted file mode 100644 index 268bf0852143dbdbaf13f8179af094d2f2461201..0000000000000000000000000000000000000000 Binary files a/agentgraph/causal/__pycache__/graph_analysis.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/causal/__pycache__/influence_analysis.cpython-311.pyc b/agentgraph/causal/__pycache__/influence_analysis.cpython-311.pyc deleted file mode 100644 index 67686682ef393eb653a441d6c8b9f9f2f58dc3a9..0000000000000000000000000000000000000000 Binary files a/agentgraph/causal/__pycache__/influence_analysis.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/causal/__pycache__/influence_analysis.cpython-312.pyc b/agentgraph/causal/__pycache__/influence_analysis.cpython-312.pyc deleted file mode 100644 index be871e65a464fc14f683de5f79da35bd3afa2e81..0000000000000000000000000000000000000000 Binary files a/agentgraph/causal/__pycache__/influence_analysis.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/causal/confounders/__pycache__/__init__.cpython-311.pyc b/agentgraph/causal/confounders/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index 21b054d4dc26d30a3a00db26f8d073c3e5fedf49..0000000000000000000000000000000000000000 Binary files a/agentgraph/causal/confounders/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/causal/confounders/__pycache__/__init__.cpython-312.pyc b/agentgraph/causal/confounders/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index 5a01a11555b43c655dee1e90df1cc5d78583305a..0000000000000000000000000000000000000000 Binary files a/agentgraph/causal/confounders/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/causal/confounders/__pycache__/basic_detection.cpython-311.pyc b/agentgraph/causal/confounders/__pycache__/basic_detection.cpython-311.pyc deleted file mode 100644 index d6f29f2a0e671d575d68987e7f3d0845d5200aae..0000000000000000000000000000000000000000 Binary files a/agentgraph/causal/confounders/__pycache__/basic_detection.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/causal/confounders/__pycache__/basic_detection.cpython-312.pyc b/agentgraph/causal/confounders/__pycache__/basic_detection.cpython-312.pyc deleted file mode 100644 index e56437bdff8c6df23afbad77a17b449303cc2fd6..0000000000000000000000000000000000000000 Binary files a/agentgraph/causal/confounders/__pycache__/basic_detection.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/causal/confounders/__pycache__/multi_signal_detection.cpython-311.pyc b/agentgraph/causal/confounders/__pycache__/multi_signal_detection.cpython-311.pyc deleted file mode 100644 index ca2c5396d2f82e8d8212545361c2b2bf038c2563..0000000000000000000000000000000000000000 Binary files a/agentgraph/causal/confounders/__pycache__/multi_signal_detection.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/causal/confounders/__pycache__/multi_signal_detection.cpython-312.pyc b/agentgraph/causal/confounders/__pycache__/multi_signal_detection.cpython-312.pyc deleted file mode 100644 index f2422e2896b07c04b7ab05966126f3265859c845..0000000000000000000000000000000000000000 Binary files a/agentgraph/causal/confounders/__pycache__/multi_signal_detection.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/causal/utils/__pycache__/__init__.cpython-311.pyc b/agentgraph/causal/utils/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index 07a15974a6daa0b471957ac44ab511d032af99d0..0000000000000000000000000000000000000000 Binary files a/agentgraph/causal/utils/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/causal/utils/__pycache__/__init__.cpython-312.pyc b/agentgraph/causal/utils/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index d7cacbe0c94a64c83f450f2189704fe12461b057..0000000000000000000000000000000000000000 Binary files a/agentgraph/causal/utils/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/causal/utils/__pycache__/dataframe_builder.cpython-311.pyc b/agentgraph/causal/utils/__pycache__/dataframe_builder.cpython-311.pyc deleted file mode 100644 index 7d3c73ed786f58e448a3255e6bde7391a5861c89..0000000000000000000000000000000000000000 Binary files a/agentgraph/causal/utils/__pycache__/dataframe_builder.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/causal/utils/__pycache__/dataframe_builder.cpython-312.pyc b/agentgraph/causal/utils/__pycache__/dataframe_builder.cpython-312.pyc deleted file mode 100644 index 8fff38c442beb807ff15fb6e1394721748d286a2..0000000000000000000000000000000000000000 Binary files a/agentgraph/causal/utils/__pycache__/dataframe_builder.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/causal/utils/__pycache__/shared_utils.cpython-311.pyc b/agentgraph/causal/utils/__pycache__/shared_utils.cpython-311.pyc deleted file mode 100644 index f24d4b2962269d1233536b9a222ab7ee90b4fc6d..0000000000000000000000000000000000000000 Binary files a/agentgraph/causal/utils/__pycache__/shared_utils.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/causal/utils/__pycache__/shared_utils.cpython-312.pyc b/agentgraph/causal/utils/__pycache__/shared_utils.cpython-312.pyc deleted file mode 100644 index 51a4b1fb0c3c2419eedfdbe0ee979df58ef1275c..0000000000000000000000000000000000000000 Binary files a/agentgraph/causal/utils/__pycache__/shared_utils.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/extraction/__pycache__/__init__.cpython-311.pyc b/agentgraph/extraction/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index 1309766c503770998a5d946c73deeede44654260..0000000000000000000000000000000000000000 Binary files a/agentgraph/extraction/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/extraction/__pycache__/__init__.cpython-312.pyc b/agentgraph/extraction/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index 72912a616475eb08add6469c669a1cdab2cfd65e..0000000000000000000000000000000000000000 Binary files a/agentgraph/extraction/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/extraction/graph_processing/__pycache__/__init__.cpython-311.pyc b/agentgraph/extraction/graph_processing/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index c4430c54d4a83924f72290f25f079d2e766f9e26..0000000000000000000000000000000000000000 Binary files a/agentgraph/extraction/graph_processing/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/extraction/graph_processing/__pycache__/__init__.cpython-312.pyc b/agentgraph/extraction/graph_processing/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index cb113faeed44fcf6390579fca1d802e2b47a6b13..0000000000000000000000000000000000000000 Binary files a/agentgraph/extraction/graph_processing/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/extraction/graph_processing/__pycache__/knowledge_graph_processor.cpython-311.pyc b/agentgraph/extraction/graph_processing/__pycache__/knowledge_graph_processor.cpython-311.pyc deleted file mode 100644 index 7e340b050b54e6573a32e042bfb256b3edb1779a..0000000000000000000000000000000000000000 Binary files a/agentgraph/extraction/graph_processing/__pycache__/knowledge_graph_processor.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/extraction/graph_processing/__pycache__/knowledge_graph_processor.cpython-312.pyc b/agentgraph/extraction/graph_processing/__pycache__/knowledge_graph_processor.cpython-312.pyc deleted file mode 100644 index 98eb46b8b0088a740910b6d853080d8fbac831ee..0000000000000000000000000000000000000000 Binary files a/agentgraph/extraction/graph_processing/__pycache__/knowledge_graph_processor.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/extraction/graph_utilities/__pycache__/__init__.cpython-311.pyc b/agentgraph/extraction/graph_utilities/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index 40d2a663b3ac13610b1aeb8a273f89b8fbcbd915..0000000000000000000000000000000000000000 Binary files a/agentgraph/extraction/graph_utilities/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/extraction/graph_utilities/__pycache__/__init__.cpython-312.pyc b/agentgraph/extraction/graph_utilities/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index d744a124f80f717b3639eed18b8006642da44ced..0000000000000000000000000000000000000000 Binary files a/agentgraph/extraction/graph_utilities/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/extraction/graph_utilities/__pycache__/graph_comparator.cpython-311.pyc b/agentgraph/extraction/graph_utilities/__pycache__/graph_comparator.cpython-311.pyc deleted file mode 100644 index 0987bd3202e949da58232eac1340d1024b200150..0000000000000000000000000000000000000000 Binary files a/agentgraph/extraction/graph_utilities/__pycache__/graph_comparator.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/extraction/graph_utilities/__pycache__/graph_comparator.cpython-312.pyc b/agentgraph/extraction/graph_utilities/__pycache__/graph_comparator.cpython-312.pyc deleted file mode 100644 index 0f36eb0367e100c431df411bbfb80527e4acfa90..0000000000000000000000000000000000000000 Binary files a/agentgraph/extraction/graph_utilities/__pycache__/graph_comparator.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/extraction/graph_utilities/__pycache__/knowledge_graph_merger.cpython-311.pyc b/agentgraph/extraction/graph_utilities/__pycache__/knowledge_graph_merger.cpython-311.pyc deleted file mode 100644 index de4d5c093041f9e5d2aebfe54641784829431384..0000000000000000000000000000000000000000 Binary files a/agentgraph/extraction/graph_utilities/__pycache__/knowledge_graph_merger.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/extraction/graph_utilities/__pycache__/knowledge_graph_merger.cpython-312.pyc b/agentgraph/extraction/graph_utilities/__pycache__/knowledge_graph_merger.cpython-312.pyc deleted file mode 100644 index dbd17e691e3b4a3bb918d265e731d12333744952..0000000000000000000000000000000000000000 Binary files a/agentgraph/extraction/graph_utilities/__pycache__/knowledge_graph_merger.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/extraction/knowledge_extraction/__pycache__/__init__.cpython-311.pyc b/agentgraph/extraction/knowledge_extraction/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index 30344d5126093c42d098d1a4681e4ca190f12695..0000000000000000000000000000000000000000 Binary files a/agentgraph/extraction/knowledge_extraction/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/extraction/knowledge_extraction/__pycache__/__init__.cpython-312.pyc b/agentgraph/extraction/knowledge_extraction/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index 4d930d6faa1caf3296640e3953717ca7fc7edfc1..0000000000000000000000000000000000000000 Binary files a/agentgraph/extraction/knowledge_extraction/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/extraction/knowledge_extraction/__pycache__/multi_agent_knowledge_extractor.cpython-312.pyc b/agentgraph/extraction/knowledge_extraction/__pycache__/multi_agent_knowledge_extractor.cpython-312.pyc deleted file mode 100644 index fbc6c2fe6cac762af2bd44187b263f81fe5113ab..0000000000000000000000000000000000000000 Binary files a/agentgraph/extraction/knowledge_extraction/__pycache__/multi_agent_knowledge_extractor.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/input/__pycache__/__init__.cpython-311.pyc b/agentgraph/input/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index f019271abefec20446f036a73a202f5d6ed027b4..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/input/__pycache__/__init__.cpython-312.pyc b/agentgraph/input/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index ff78224b414b92c955e219817e7c84729d3ab073..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/input/__pycache__/__init__.cpython-313.pyc b/agentgraph/input/__pycache__/__init__.cpython-313.pyc deleted file mode 100644 index d2513806bd7a9c6d359f3ef9f17c7fcbd63784bc..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/__pycache__/__init__.cpython-313.pyc and /dev/null differ diff --git a/agentgraph/input/content_analysis/__pycache__/__init__.cpython-311.pyc b/agentgraph/input/content_analysis/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index be43d59c81dfdd21552c3eb3c4f908effcb04b95..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/content_analysis/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/input/content_analysis/__pycache__/__init__.cpython-312.pyc b/agentgraph/input/content_analysis/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index 8887783a835037262be01c8c1759bb4c5c1ac555..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/content_analysis/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/input/content_analysis/__pycache__/__init__.cpython-313.pyc b/agentgraph/input/content_analysis/__pycache__/__init__.cpython-313.pyc deleted file mode 100644 index 45ab4f878f574ea70f748a22c178e4e2fd82a5eb..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/content_analysis/__pycache__/__init__.cpython-313.pyc and /dev/null differ diff --git a/agentgraph/input/content_analysis/__pycache__/boundary_detector.cpython-311.pyc b/agentgraph/input/content_analysis/__pycache__/boundary_detector.cpython-311.pyc deleted file mode 100644 index a71ee957875fe87985e8246e3674469dae23f8d9..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/content_analysis/__pycache__/boundary_detector.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/input/content_analysis/__pycache__/boundary_detector.cpython-312.pyc b/agentgraph/input/content_analysis/__pycache__/boundary_detector.cpython-312.pyc deleted file mode 100644 index 2e987bb5440c88e0f036c6244f6b7f66945d605f..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/content_analysis/__pycache__/boundary_detector.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/input/content_analysis/__pycache__/boundary_detector.cpython-313.pyc b/agentgraph/input/content_analysis/__pycache__/boundary_detector.cpython-313.pyc deleted file mode 100644 index dac1521c298c03a4818da4939c904fc8d2e116b7..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/content_analysis/__pycache__/boundary_detector.cpython-313.pyc and /dev/null differ diff --git a/agentgraph/input/content_analysis/__pycache__/log_type_detector.cpython-311.pyc b/agentgraph/input/content_analysis/__pycache__/log_type_detector.cpython-311.pyc deleted file mode 100644 index d4850c6f8bafef9f2120b902eeabfcfc417d3f21..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/content_analysis/__pycache__/log_type_detector.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/input/content_analysis/__pycache__/log_type_detector.cpython-312.pyc b/agentgraph/input/content_analysis/__pycache__/log_type_detector.cpython-312.pyc deleted file mode 100644 index fa61e451af14513b319515bbaf648a9fab1775af..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/content_analysis/__pycache__/log_type_detector.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/input/content_analysis/__pycache__/log_type_detector.cpython-313.pyc b/agentgraph/input/content_analysis/__pycache__/log_type_detector.cpython-313.pyc deleted file mode 100644 index 4ec4270f6fa10e931a74abda39a43ad6daf1b95e..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/content_analysis/__pycache__/log_type_detector.cpython-313.pyc and /dev/null differ diff --git a/agentgraph/input/content_analysis/__pycache__/semantic_analyzer.cpython-311.pyc b/agentgraph/input/content_analysis/__pycache__/semantic_analyzer.cpython-311.pyc deleted file mode 100644 index 8b996c16e55223b875fe9863412adbe1d33160dd..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/content_analysis/__pycache__/semantic_analyzer.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/input/content_analysis/__pycache__/semantic_analyzer.cpython-312.pyc b/agentgraph/input/content_analysis/__pycache__/semantic_analyzer.cpython-312.pyc deleted file mode 100644 index 01a57b6984e67a76594f5b5bd080412a59682098..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/content_analysis/__pycache__/semantic_analyzer.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/input/content_analysis/__pycache__/semantic_analyzer.cpython-313.pyc b/agentgraph/input/content_analysis/__pycache__/semantic_analyzer.cpython-313.pyc deleted file mode 100644 index b635afbb432044bf25e0bdafff9fbec9628120f4..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/content_analysis/__pycache__/semantic_analyzer.cpython-313.pyc and /dev/null differ diff --git a/agentgraph/input/parsers/__pycache__/__init__.cpython-311.pyc b/agentgraph/input/parsers/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index 703051a8eb34da16e38031850c73cacf1e46dc08..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/parsers/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/input/parsers/__pycache__/__init__.cpython-312.pyc b/agentgraph/input/parsers/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index e782a194a240ec1b418c4501cdefd54d7e2a7e9c..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/parsers/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/input/parsers/__pycache__/base_parser.cpython-311.pyc b/agentgraph/input/parsers/__pycache__/base_parser.cpython-311.pyc deleted file mode 100644 index 2a9e1a600e820c7a759ce26cf885957ca5478605..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/parsers/__pycache__/base_parser.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/input/parsers/__pycache__/base_parser.cpython-312.pyc b/agentgraph/input/parsers/__pycache__/base_parser.cpython-312.pyc deleted file mode 100644 index 4779322373bb9ee99af20526d8977b592bbfb6d1..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/parsers/__pycache__/base_parser.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/input/parsers/__pycache__/langsmith_parser.cpython-311.pyc b/agentgraph/input/parsers/__pycache__/langsmith_parser.cpython-311.pyc deleted file mode 100644 index 3605456fffe9ddf432da308f2223bb53245da129..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/parsers/__pycache__/langsmith_parser.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/input/parsers/__pycache__/langsmith_parser.cpython-312.pyc b/agentgraph/input/parsers/__pycache__/langsmith_parser.cpython-312.pyc deleted file mode 100644 index 398640380c60a3e51a8f97fb5e644adfbbe479bb..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/parsers/__pycache__/langsmith_parser.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/input/parsers/__pycache__/parser_factory.cpython-311.pyc b/agentgraph/input/parsers/__pycache__/parser_factory.cpython-311.pyc deleted file mode 100644 index c398322acbd0f0e18ecb800a177c282553656327..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/parsers/__pycache__/parser_factory.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/input/parsers/__pycache__/parser_factory.cpython-312.pyc b/agentgraph/input/parsers/__pycache__/parser_factory.cpython-312.pyc deleted file mode 100644 index 0198f92b0d4e15c6847e01b067f1f7a2af0d06c0..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/parsers/__pycache__/parser_factory.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/input/parsers/__pycache__/universal_parser.cpython-311.pyc b/agentgraph/input/parsers/__pycache__/universal_parser.cpython-311.pyc deleted file mode 100644 index 5115ae88f6d79ae1576c25a4ed82d2d20647235e..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/parsers/__pycache__/universal_parser.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/input/parsers/__pycache__/universal_parser.cpython-312.pyc b/agentgraph/input/parsers/__pycache__/universal_parser.cpython-312.pyc deleted file mode 100644 index 04e4044fdf77a045523dd0df1730605ca1469fa3..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/parsers/__pycache__/universal_parser.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/input/text_processing/__pycache__/__init__.cpython-311.pyc b/agentgraph/input/text_processing/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index 751eb269910f0c5a7336bc477d8f1eb4155ed614..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/text_processing/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/input/text_processing/__pycache__/__init__.cpython-312.pyc b/agentgraph/input/text_processing/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index ff78bfac5d411d63512799bf8fcb960f2bb7f2b8..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/text_processing/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/input/text_processing/__pycache__/chunking_service.cpython-311.pyc b/agentgraph/input/text_processing/__pycache__/chunking_service.cpython-311.pyc deleted file mode 100644 index 4997c4d84ddb3ba3b9a008c95f7abf7047def14a..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/text_processing/__pycache__/chunking_service.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/input/text_processing/__pycache__/chunking_service.cpython-312.pyc b/agentgraph/input/text_processing/__pycache__/chunking_service.cpython-312.pyc deleted file mode 100644 index 43eb6c2ad00f1f8533afc43a4914f0a987cfde9c..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/text_processing/__pycache__/chunking_service.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/input/text_processing/__pycache__/text_chunking_strategies.cpython-311.pyc b/agentgraph/input/text_processing/__pycache__/text_chunking_strategies.cpython-311.pyc deleted file mode 100644 index 8c84bbcc0dc0275b5e73424885905efa17076d03..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/text_processing/__pycache__/text_chunking_strategies.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/input/text_processing/__pycache__/text_chunking_strategies.cpython-312.pyc b/agentgraph/input/text_processing/__pycache__/text_chunking_strategies.cpython-312.pyc deleted file mode 100644 index b0ec4195d783cdc8f7ec4aa80d732cfeb7d67f27..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/text_processing/__pycache__/text_chunking_strategies.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/input/text_processing/__pycache__/trace_line_processor.cpython-311.pyc b/agentgraph/input/text_processing/__pycache__/trace_line_processor.cpython-311.pyc deleted file mode 100644 index f52ded48182a69c41b78195cf1b76b5e2ccc0bc6..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/text_processing/__pycache__/trace_line_processor.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/input/text_processing/__pycache__/trace_line_processor.cpython-312.pyc b/agentgraph/input/text_processing/__pycache__/trace_line_processor.cpython-312.pyc deleted file mode 100644 index 17f241e37b15d40e12e3401bab3390c02e45b8f0..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/text_processing/__pycache__/trace_line_processor.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/input/text_processing/__pycache__/trace_preprocessor.cpython-311.pyc b/agentgraph/input/text_processing/__pycache__/trace_preprocessor.cpython-311.pyc deleted file mode 100644 index 2d27b1f7136993f91bb1a6c12e5f6f2b5172ae22..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/text_processing/__pycache__/trace_preprocessor.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/input/text_processing/__pycache__/trace_preprocessor.cpython-312.pyc b/agentgraph/input/text_processing/__pycache__/trace_preprocessor.cpython-312.pyc deleted file mode 100644 index 793fcbbb6c44b1e24d6888464f4ab0dcd3cee261..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/text_processing/__pycache__/trace_preprocessor.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/input/trace_management/__pycache__/__init__.cpython-311.pyc b/agentgraph/input/trace_management/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index 57760983be671efe7abbdb36717109f14be26e7b..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/trace_management/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/input/trace_management/__pycache__/__init__.cpython-312.pyc b/agentgraph/input/trace_management/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index de2e0108a6fa29cc6283586f4e8bb47058d86414..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/trace_management/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/input/trace_management/__pycache__/__init__.cpython-313.pyc b/agentgraph/input/trace_management/__pycache__/__init__.cpython-313.pyc deleted file mode 100644 index acf0380b095344d07bb36e7ad325520f349eb238..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/trace_management/__pycache__/__init__.cpython-313.pyc and /dev/null differ diff --git a/agentgraph/input/trace_management/__pycache__/langsmith_metadata_parser.cpython-312.pyc b/agentgraph/input/trace_management/__pycache__/langsmith_metadata_parser.cpython-312.pyc deleted file mode 100644 index d5d089808b3f3c0e7bc2aebc9a2f5d4139673a7f..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/trace_management/__pycache__/langsmith_metadata_parser.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/input/trace_management/__pycache__/trace_analysis.cpython-311.pyc b/agentgraph/input/trace_management/__pycache__/trace_analysis.cpython-311.pyc deleted file mode 100644 index 96b1ad5f7a7a128e9359e397ab076f338ed787b8..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/trace_management/__pycache__/trace_analysis.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/input/trace_management/__pycache__/trace_analysis.cpython-312.pyc b/agentgraph/input/trace_management/__pycache__/trace_analysis.cpython-312.pyc deleted file mode 100644 index b87358b0ef653054c53a2dbefbeea4f5e49ce82a..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/trace_management/__pycache__/trace_analysis.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/input/trace_management/__pycache__/trace_analysis.cpython-313.pyc b/agentgraph/input/trace_management/__pycache__/trace_analysis.cpython-313.pyc deleted file mode 100644 index 0273a385f2c1102550fb653ba6dccc7316477c02..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/trace_management/__pycache__/trace_analysis.cpython-313.pyc and /dev/null differ diff --git a/agentgraph/input/trace_management/__pycache__/trace_loader_service.cpython-311.pyc b/agentgraph/input/trace_management/__pycache__/trace_loader_service.cpython-311.pyc deleted file mode 100644 index a3f52e5115a9a0fdac44cc529d164b37c9b758fc..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/trace_management/__pycache__/trace_loader_service.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/input/trace_management/__pycache__/trace_loader_service.cpython-312.pyc b/agentgraph/input/trace_management/__pycache__/trace_loader_service.cpython-312.pyc deleted file mode 100644 index 9189fdaf4d4b4ed31c175150ee6dbabaf9ececc4..0000000000000000000000000000000000000000 Binary files a/agentgraph/input/trace_management/__pycache__/trace_loader_service.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/methods/__pycache__/__init__.cpython-311.pyc b/agentgraph/methods/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index 7bb8c405112692b5136ed859bb79f99977341d24..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/methods/__pycache__/__init__.cpython-312.pyc b/agentgraph/methods/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index bd6059363068d3c57569d2427ae92cdc76c5cddc..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/methods/baseline/__pycache__/__init__.cpython-311.pyc b/agentgraph/methods/baseline/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index 0795d4156ff2de404ae0381ba3fda460b25de3ee..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/baseline/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/methods/baseline/__pycache__/__init__.cpython-312.pyc b/agentgraph/methods/baseline/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index 0893a2c8cf794fc66b37a286e8de957a6477d378..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/baseline/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/methods/baseline/__pycache__/base_method.cpython-311.pyc b/agentgraph/methods/baseline/__pycache__/base_method.cpython-311.pyc deleted file mode 100644 index 3316a9a0fc2d3d72798151f27f97bd2e7007d42b..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/baseline/__pycache__/base_method.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/methods/baseline/__pycache__/base_method.cpython-312.pyc b/agentgraph/methods/baseline/__pycache__/base_method.cpython-312.pyc deleted file mode 100644 index 5d5e2b9cfd74a299e42450f9efd76d09b3eb7715..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/baseline/__pycache__/base_method.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/methods/baseline/__pycache__/clustering_method.cpython-311.pyc b/agentgraph/methods/baseline/__pycache__/clustering_method.cpython-311.pyc deleted file mode 100644 index 58891aa4f0db20887cdfa59b1adcd36c2801df42..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/baseline/__pycache__/clustering_method.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/methods/baseline/__pycache__/clustering_method.cpython-312.pyc b/agentgraph/methods/baseline/__pycache__/clustering_method.cpython-312.pyc deleted file mode 100644 index 811d369939aa7f69bf6daf58333783fbd3b0e431..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/baseline/__pycache__/clustering_method.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/methods/baseline/__pycache__/direct_llm_method.cpython-311.pyc b/agentgraph/methods/baseline/__pycache__/direct_llm_method.cpython-311.pyc deleted file mode 100644 index 3ee25eddaef2efc178b73dfe435be1bb92cd4dc2..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/baseline/__pycache__/direct_llm_method.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/methods/baseline/__pycache__/direct_llm_method.cpython-312.pyc b/agentgraph/methods/baseline/__pycache__/direct_llm_method.cpython-312.pyc deleted file mode 100644 index 5aea433556e008bf109b6f2e49f77ea5a28cd1ff..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/baseline/__pycache__/direct_llm_method.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/methods/baseline/__pycache__/dumb_method.cpython-312.pyc b/agentgraph/methods/baseline/__pycache__/dumb_method.cpython-312.pyc deleted file mode 100644 index 336d7aaed9a4b96089470cdc53d4f8b486ed5f88..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/baseline/__pycache__/dumb_method.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/methods/baseline/__pycache__/hybrid_method.cpython-311.pyc b/agentgraph/methods/baseline/__pycache__/hybrid_method.cpython-311.pyc deleted file mode 100644 index 60c7c98ee7fad9c388fda25a70297f143d6c400c..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/baseline/__pycache__/hybrid_method.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/methods/baseline/__pycache__/hybrid_method.cpython-312.pyc b/agentgraph/methods/baseline/__pycache__/hybrid_method.cpython-312.pyc deleted file mode 100644 index ca4a29902f68e2ce2a289f8431477283c92a82f7..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/baseline/__pycache__/hybrid_method.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/methods/baseline/__pycache__/openai_agent.cpython-311.pyc b/agentgraph/methods/baseline/__pycache__/openai_agent.cpython-311.pyc deleted file mode 100644 index 6f8b8a47aa58944eb1d4ae7c12d778c4c475046a..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/baseline/__pycache__/openai_agent.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/methods/baseline/__pycache__/original_method.cpython-311.pyc b/agentgraph/methods/baseline/__pycache__/original_method.cpython-311.pyc deleted file mode 100644 index 6001e65fd6bb35b5a1fae62ba437c380030b6a4b..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/baseline/__pycache__/original_method.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/methods/baseline/__pycache__/original_method.cpython-312.pyc b/agentgraph/methods/baseline/__pycache__/original_method.cpython-312.pyc deleted file mode 100644 index 3cb8c02a5ce9e2bd3c7019fbad49563f2df8fb1a..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/baseline/__pycache__/original_method.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/methods/baseline/__pycache__/perfect_method.cpython-312.pyc b/agentgraph/methods/baseline/__pycache__/perfect_method.cpython-312.pyc deleted file mode 100644 index 8d4e7559ca1f27dea0e564a412ce0b938f15e2ed..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/baseline/__pycache__/perfect_method.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/methods/baseline/__pycache__/pydantic_method.cpython-311.pyc b/agentgraph/methods/baseline/__pycache__/pydantic_method.cpython-311.pyc deleted file mode 100644 index 406c143f8ac2fbc51454211beeef0350973cda6d..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/baseline/__pycache__/pydantic_method.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/methods/baseline/__pycache__/pydantic_method.cpython-312.pyc b/agentgraph/methods/baseline/__pycache__/pydantic_method.cpython-312.pyc deleted file mode 100644 index 9d7b5edf07953c0b193c4c1f7a3bfd3d510c2225..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/baseline/__pycache__/pydantic_method.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/methods/baseline/__pycache__/rule_based_method.cpython-312.pyc b/agentgraph/methods/baseline/__pycache__/rule_based_method.cpython-312.pyc deleted file mode 100644 index a7bcfdcb02fb9a407e013bfc7f1267892bc952b7..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/baseline/__pycache__/rule_based_method.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/methods/baseline/__pycache__/unified_method.cpython-311.pyc b/agentgraph/methods/baseline/__pycache__/unified_method.cpython-311.pyc deleted file mode 100644 index 5e4a3e805ccd51b938020c2ee71767f9b7939ad0..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/baseline/__pycache__/unified_method.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/methods/baseline/__pycache__/unified_method.cpython-312.pyc b/agentgraph/methods/baseline/__pycache__/unified_method.cpython-312.pyc deleted file mode 100644 index 4e5605c1bc46cfb7adcae1f562373e7f6423adb4..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/baseline/__pycache__/unified_method.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/methods/experimental/__pycache__/__init__.cpython-312.pyc b/agentgraph/methods/experimental/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index a3c193ebc5282fe0fe58876def185abe1cd72a88..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/experimental/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/methods/experimental/__pycache__/adaptive_knowledge_extractor.cpython-312.pyc b/agentgraph/methods/experimental/__pycache__/adaptive_knowledge_extractor.cpython-312.pyc deleted file mode 100644 index b0acd2239ac9adaf414168640e7879f4eda8fec7..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/experimental/__pycache__/adaptive_knowledge_extractor.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/methods/experimental/__pycache__/enhanced_adaptive_extractor.cpython-312.pyc b/agentgraph/methods/experimental/__pycache__/enhanced_adaptive_extractor.cpython-312.pyc deleted file mode 100644 index 06140d940b3da839a94f227cfb610603e6e40569..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/experimental/__pycache__/enhanced_adaptive_extractor.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/methods/experimental/__pycache__/optimized_adaptive_extractor.cpython-312.pyc b/agentgraph/methods/experimental/__pycache__/optimized_adaptive_extractor.cpython-312.pyc deleted file mode 100644 index 094ada3319eda833828ebaccb21e93004c6be71b..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/experimental/__pycache__/optimized_adaptive_extractor.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/methods/experimental/__pycache__/streamlined_agent_extractor.cpython-312.pyc b/agentgraph/methods/experimental/__pycache__/streamlined_agent_extractor.cpython-312.pyc deleted file mode 100644 index 103e25b18193bb1ceae15a06a0174abbd5df9628..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/experimental/__pycache__/streamlined_agent_extractor.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/methods/logs/agent_monitoring.log b/agentgraph/methods/logs/agent_monitoring.log deleted file mode 100644 index b5971050871ab3005100347c5cd3d295503cb23c..0000000000000000000000000000000000000000 --- a/agentgraph/methods/logs/agent_monitoring.log +++ /dev/null @@ -1,7550 +0,0 @@ -2025-07-24 13:01:19,557 - openlit - INFO - Starting openLIT initialization... -2025-07-24 13:01:19,573 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 13:01:20,192 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 13:01:20,244 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 13:01:20,244 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 13:01:20,244 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 13:01:20,614 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 13:01:20,723 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 13:01:20,723 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 13:01:21,306 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 13:01:21,307 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 13:01:22,751 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 13:01:22,754 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 13:01:22,754 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 13:01:22,755 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 13:01:22,755 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 13:01:22,755 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 13:01:22,755 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 13:01:22,755 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 13:01:22,755 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 13:01:22,755 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 13:01:22,755 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 13:01:22,755 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 13:01:22,755 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 13:01:22,756 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 13:01:22,756 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 13:01:22,756 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 13:01:22,756 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 13:01:22,756 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 13:01:22,756 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 14:29:57,970 - openlit - INFO - Starting openLIT initialization... -2025-07-24 14:29:57,991 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 14:29:58,795 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 14:29:58,861 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 14:29:58,862 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 14:29:58,862 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 14:29:59,324 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 14:29:59,465 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 14:29:59,465 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 14:30:00,256 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 14:30:00,258 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 14:30:02,427 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 14:30:02,430 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 14:30:02,430 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 14:30:02,431 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 14:30:02,431 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 14:30:02,431 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 14:30:02,431 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 14:30:02,431 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 14:30:02,431 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 14:30:02,431 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 14:30:02,431 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 14:30:02,431 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 14:30:02,432 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 14:30:02,432 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 14:30:02,432 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 14:30:02,432 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 14:30:02,432 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 14:30:02,432 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 14:30:02,432 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 14:30:03,703 - __main__ - INFO - Loaded 3 texts from scripts/example_texts.json -2025-07-24 14:30:03,703 - agentgraph.input.text_processing.chunking_service - INFO - ChunkingService initialized with batch_size=3, model=gpt-4o-mini -2025-07-24 14:30:03,703 - __main__ - INFO - Initialized BatchKGExtractor with model: gpt-4o-mini, method: production -2025-07-24 14:30:03,703 - __main__ - INFO - Processing batch of 3 texts -2025-07-24 14:30:03,703 - __main__ - INFO - Processing text 1/3: text_0 -2025-07-24 14:30:03,703 - __main__ - INFO - Processing text text_0 -2025-07-24 14:30:03,703 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 14:30:03,703 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 360 characters -2025-07-24 14:30:03,703 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 14:30:03,703 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 14:30:03,711 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 14:30:03,711 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 14:30:03,715 - agentgraph.input.content_analysis.semantic_analyzer - INFO - Detected agent trace type: unknown (confidence: 0.12) -2025-07-24 14:30:03,715 - agentgraph.input.text_processing.chunking_service - INFO - Applied rule-based line splitting to 1 chunks (max_line_length=800) -2025-07-24 14:30:03,716 - agentgraph.input.text_processing.chunking_service - INFO - Assigning global line numbers to 1 chunks -2025-07-24 14:30:03,716 - agentgraph.input.text_processing.trace_line_processor - INFO - Added line numbers to 6 lines, starting from line 1 -2025-07-24 14:30:03,716 - agentgraph.input.text_processing.chunking_service - DEBUG - Chunk 0: chars 0-360 → lines 1-6 -2025-07-24 14:30:03,716 - agentgraph.input.text_processing.chunking_service - INFO - Successfully assigned global line numbers to all chunks -2025-07-24 14:30:03,716 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using agent_semantic splitter -2025-07-24 14:30:03,716 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=350000, overlap_size=17500 -2025-07-24 14:30:03,717 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context enhancement complete: 0 total documents -2025-07-24 14:30:03,717 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - SlidingWindowMonitor initialized with model: gpt-4o-mini, method: production -2025-07-24 14:30:03,717 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context documents: 0 provided -2025-07-24 14:30:03,717 - agentgraph.extraction.graph_utilities.knowledge_graph_merger - INFO - KnowledgeGraphMerger initialized with model: gpt-4o-mini -2025-07-24 14:30:03,719 - __main__ - INFO - Extracting knowledge graph from 1 chunks -2025-07-24 14:30:03,719 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks (splitter: agent_semantic, window_size=350000, overlap=0) -2025-07-24 14:30:03,719 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided source trace ID: text_0 -2025-07-24 14:30:03,719 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided processing run ID: batch_text_0 -2025-07-24 14:30:03,719 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks -2025-07-24 14:30:03,719 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using parallel processing with batch size 3 (1 batches) -2025-07-24 14:30:03,719 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 windows in 1 batches -2025-07-24 14:30:03,732 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks in 1 sub-batches (max 1 concurrent per sub-batch) -2025-07-24 14:30:03,732 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Starting sub-batch 1/1 with 1 chunks -2025-07-24 14:30:03,732 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing window 0 -2025-07-24 14:30:03,732 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - 🤖 Starting async_crew extraction for window 0 with method: production using model: gpt-4o-mini -2025-07-24 14:30:03,732 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - No context documents available for window 0, passing empty context string -2025-07-24 14:30:03,732 - agentgraph.methods.production.multi_agent_knowledge_extractor - INFO - Creating agent monitoring crew with model: gpt-4o-mini -2025-07-24 14:30:50,165 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Result Generated (production method) -2025-07-24 14:30:50,166 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 6 lines -2025-07-24 14:30:50,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:30:50,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 14:30:50,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L2-L2 -2025-07-24 14:30:50,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L2-L2 -2025-07-24 14:30:50,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 1, end_idx: 2 -2025-07-24 14:30:50,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 14:30:50,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:30:50,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Role: You are a research assistant AI th... -2025-07-24 14:30:50,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Role: You are a research assistant AI th... -2025-07-24 14:30:50,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Role: You are a research assistant AI th... -2025-07-24 14:30:50,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:30:50,166 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 1 resolution debug: -2025-07-24 14:30:50,166 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 14:30:50,166 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L2-L2 -2025-07-24 14:30:50,166 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 14:30:50,166 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Role: You are a research assistant AI th... -2025-07-24 14:30:50,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 1: 96 characters -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L3-L3 -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L3-L3 -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 2, end_idx: 3 -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Task: Search for information about clima... -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Task: Search for information about clima... -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Task: Search for information about clima... -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:30:50,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 2 resolution debug: -2025-07-24 14:30:50,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 14:30:50,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L3-L3 -2025-07-24 14:30:50,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 14:30:50,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Task: Search for information about clima... -2025-07-24 14:30:50,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 2: 59 characters -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L4-L4 -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L4-L4 -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 3, end_idx: 4 -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Tool: web_search... -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Tool: web_search... -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Tool: web_search... -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:30:50,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 3 resolution debug: -2025-07-24 14:30:50,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 14:30:50,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L4-L4 -2025-07-24 14:30:50,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 14:30:50,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Tool: web_search... -2025-07-24 14:30:50,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 3: 26 characters -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L5-L5 -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L5-L5 -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 4, end_idx: 5 -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Input: climate change impacts 2024... -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Input: climate change impacts 2024... -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Input: climate change impacts 2024... -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:30:50,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 4 resolution debug: -2025-07-24 14:30:50,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 14:30:50,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L5-L5 -2025-07-24 14:30:50,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 14:30:50,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Input: climate change impacts 2024... -2025-07-24 14:30:50,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 4: 44 characters -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L6-L6 -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L6-L6 -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 5, end_idx: 6 -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Output: Climate change continues to show... -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Output: Climate change continues to show... -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Output: Climate change continues to show... -2025-07-24 14:30:50,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:30:50,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 5 resolution debug: -2025-07-24 14:30:50,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 14:30:50,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L6-L6 -2025-07-24 14:30:50,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 14:30:50,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Output: Climate change continues to show... -2025-07-24 14:30:50,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 5: 150 characters -2025-07-24 14:30:50,167 - agentgraph.reconstruction.content_reference_resolver - INFO - Entity prompt resolution stats: {'total_entities': 5, 'entities_with_refs': 5, 'successful_resolutions': 5, 'failed_resolutions': 0} -2025-07-24 14:30:50,168 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 6 lines -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L5-L5 -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L5-L5 -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 4, end_idx: 5 -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Input: climate change impacts 2024... -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Input: climate change impacts 2024... -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Input: climate change impacts 2024... -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:30:50,168 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation 1: 44 characters -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L4-L4 -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L4-L4 -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 3, end_idx: 4 -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Tool: web_search... -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Tool: web_search... -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Tool: web_search... -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:30:50,168 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation 3: 26 characters -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L3-L3 -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L3-L3 -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 2, end_idx: 3 -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Task: Search for information about clima... -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Task: Search for information about clima... -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Task: Search for information about clima... -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:30:50,168 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation 4: 59 characters -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L6-L6 -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L6-L6 -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 5, end_idx: 6 -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Output: Climate change continues to show... -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Output: Climate change continues to show... -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Output: Climate change continues to show... -2025-07-24 14:30:50,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:30:50,168 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation 6: 150 characters -2025-07-24 14:30:50,168 - agentgraph.reconstruction.content_reference_resolver - INFO - Relation prompt resolution stats: {'total_relations': 6, 'relations_with_refs': 4, 'successful_resolutions': 4, 'failed_resolutions': 0} -2025-07-24 14:30:50,168 - agentgraph.reconstruction.content_reference_resolver - INFO - Resolved content references for knowledge graph with 5 entities and 6 relations -2025-07-24 14:30:50,168 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Resolved content references for window 0 -2025-07-24 14:30:50,169 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Completed sub-batch 1/1 -2025-07-24 14:30:50,169 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Preparing window knowledge graphs with metadata -2025-07-24 14:30:50,173 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Merging knowledge graphs... -2025-07-24 14:30:50,173 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Only one knowledge graph generated, skipping merge process -2025-07-24 14:30:50,173 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing complete. Knowledge graph generated with 5 entities and 6 relations -2025-07-24 14:30:50,176 - __main__ - INFO - Processing text 2/3: text_1 -2025-07-24 14:30:50,176 - __main__ - INFO - Processing text text_1 -2025-07-24 14:30:50,176 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 14:30:50,176 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 445 characters -2025-07-24 14:30:50,176 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 14:30:50,177 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 14:30:50,186 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 14:30:50,186 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 14:30:50,187 - agentgraph.input.content_analysis.semantic_analyzer - INFO - Detected agent trace type: unknown (confidence: 0.12) -2025-07-24 14:30:50,187 - agentgraph.input.text_processing.chunking_service - INFO - Applied rule-based line splitting to 1 chunks (max_line_length=800) -2025-07-24 14:30:50,187 - agentgraph.input.text_processing.chunking_service - INFO - Assigning global line numbers to 1 chunks -2025-07-24 14:30:50,187 - agentgraph.input.text_processing.trace_line_processor - INFO - Added line numbers to 8 lines, starting from line 1 -2025-07-24 14:30:50,187 - agentgraph.input.text_processing.chunking_service - DEBUG - Chunk 0: chars 0-445 → lines 1-8 -2025-07-24 14:30:50,187 - agentgraph.input.text_processing.chunking_service - INFO - Successfully assigned global line numbers to all chunks -2025-07-24 14:30:50,187 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using agent_semantic splitter -2025-07-24 14:30:50,187 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=350000, overlap_size=17500 -2025-07-24 14:30:50,188 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context enhancement complete: 0 total documents -2025-07-24 14:30:50,188 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - SlidingWindowMonitor initialized with model: gpt-4o-mini, method: production -2025-07-24 14:30:50,188 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context documents: 0 provided -2025-07-24 14:30:50,188 - agentgraph.extraction.graph_utilities.knowledge_graph_merger - INFO - KnowledgeGraphMerger initialized with model: gpt-4o-mini -2025-07-24 14:30:50,190 - __main__ - INFO - Extracting knowledge graph from 1 chunks -2025-07-24 14:30:50,190 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks (splitter: agent_semantic, window_size=350000, overlap=0) -2025-07-24 14:30:50,190 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided source trace ID: text_1 -2025-07-24 14:30:50,190 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided processing run ID: batch_text_1 -2025-07-24 14:30:50,190 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks -2025-07-24 14:30:50,190 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using parallel processing with batch size 3 (1 batches) -2025-07-24 14:30:50,190 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 windows in 1 batches -2025-07-24 14:30:50,190 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks in 1 sub-batches (max 1 concurrent per sub-batch) -2025-07-24 14:30:50,190 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Starting sub-batch 1/1 with 1 chunks -2025-07-24 14:30:50,191 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing window 0 -2025-07-24 14:30:50,191 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - 🤖 Starting async_crew extraction for window 0 with method: production using model: gpt-4o-mini -2025-07-24 14:30:50,191 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - No context documents available for window 0, passing empty context string -2025-07-24 14:30:50,191 - agentgraph.methods.production.multi_agent_knowledge_extractor - INFO - Creating agent monitoring crew with model: gpt-4o-mini -2025-07-24 14:31:42,003 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Result Generated (production method) -2025-07-24 14:31:42,003 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 8 lines -2025-07-24 14:31:42,003 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:31:42,003 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 8 -2025-07-24 14:31:42,003 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L2-L2 -2025-07-24 14:31:42,003 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L2-L2 -2025-07-24 14:31:42,003 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 1, end_idx: 2 -2025-07-24 14:31:42,003 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 8 -2025-07-24 14:31:42,003 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:31:42,003 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Agent 1: Data Analyst - responsible for ... -2025-07-24 14:31:42,003 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Agent 1: Data Analyst - responsible for ... -2025-07-24 14:31:42,003 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Agent 1: Data Analyst - responsible for ... -2025-07-24 14:31:42,003 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:31:42,003 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity agent_1 resolution debug: -2025-07-24 14:31:42,003 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 14:31:42,003 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L2-L2 -2025-07-24 14:31:42,003 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 14:31:42,003 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Agent 1: Data Analyst - responsible for ... -2025-07-24 14:31:42,003 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity agent_1: 73 characters -2025-07-24 14:31:42,003 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:31:42,003 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 8 -2025-07-24 14:31:42,003 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L3-L3 -2025-07-24 14:31:42,003 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L3-L3 -2025-07-24 14:31:42,003 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 2, end_idx: 3 -2025-07-24 14:31:42,003 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 8 -2025-07-24 14:31:42,003 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:31:42,003 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Agent 2: Report Generator - creates comp... -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Agent 2: Report Generator - creates comp... -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Agent 2: Report Generator - creates comp... -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity agent_2 resolution debug: -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L3-L3 -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Agent 2: Report Generator - creates comp... -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity agent_2: 67 characters -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 8 -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L4-L4 -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L4-L4 -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 3, end_idx: 4 -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 8 -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Task 1: Analyze sales data from Q4 2023... -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Task 1: Analyze sales data from Q4 2023... -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Task 1: Analyze sales data from Q4 2023... -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity task_1 resolution debug: -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L4-L4 -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Task 1: Analyze sales data from Q4 2023... -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity task_1: 49 characters -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 8 -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L5-L5 -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L5-L5 -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 4, end_idx: 5 -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 8 -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Task 2: Generate executive summary repor... -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Task 2: Generate executive summary repor... -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Task 2: Generate executive summary repor... -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity task_2 resolution debug: -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L5-L5 -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Task 2: Generate executive summary repor... -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity task_2: 51 characters -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 8 -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L6-L6 -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L6-L6 -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 5, end_idx: 6 -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 8 -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Tool: pandas_analyzer - analyzes CSV dat... -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Tool: pandas_analyzer - analyzes CSV dat... -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Tool: pandas_analyzer - analyzes CSV dat... -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity tool_1 resolution debug: -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L6-L6 -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Tool: pandas_analyzer - analyzes CSV dat... -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity tool_1: 51 characters -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 8 -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L7-L7 -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L7-L7 -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 6, end_idx: 7 -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 8 -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Tool: report_generator - creates PDF rep... -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Tool: report_generator - creates PDF rep... -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Tool: report_generator - creates PDF rep... -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity tool_2 resolution debug: -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L7-L7 -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Tool: report_generator - creates PDF rep... -2025-07-24 14:31:42,004 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity tool_2: 54 characters -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 8 -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L8-L8 -2025-07-24 14:31:42,004 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L8-L8 -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 7, end_idx: 8 -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 8 -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Flow: Data Analyst performs Task 1 using... -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Flow: Data Analyst performs Task 1 using... -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Flow: Data Analyst performs Task 1 using... -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:31:42,005 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity flow_1 resolution debug: -2025-07-24 14:31:42,005 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 14:31:42,005 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L8-L8 -2025-07-24 14:31:42,005 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 14:31:42,005 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Flow: Data Analyst performs Task 1 using... -2025-07-24 14:31:42,005 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity flow_1: 128 characters -2025-07-24 14:31:42,005 - agentgraph.reconstruction.content_reference_resolver - INFO - Entity prompt resolution stats: {'total_entities': 7, 'entities_with_refs': 7, 'successful_resolutions': 7, 'failed_resolutions': 0} -2025-07-24 14:31:42,005 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 8 lines -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 8 -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L8-L8 -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L8-L8 -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 7, end_idx: 8 -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 8 -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Flow: Data Analyst performs Task 1 using... -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Flow: Data Analyst performs Task 1 using... -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Flow: Data Analyst performs Task 1 using... -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:31:42,005 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation rel_1: 128 characters -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 8 -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L8-L8 -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L8-L8 -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 7, end_idx: 8 -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 8 -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Flow: Data Analyst performs Task 1 using... -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Flow: Data Analyst performs Task 1 using... -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Flow: Data Analyst performs Task 1 using... -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:31:42,005 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation rel_3: 128 characters -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 8 -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L8-L8 -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L8-L8 -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 7, end_idx: 8 -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 8 -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Flow: Data Analyst performs Task 1 using... -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Flow: Data Analyst performs Task 1 using... -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Flow: Data Analyst performs Task 1 using... -2025-07-24 14:31:42,005 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:31:42,005 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation rel_5: 128 characters -2025-07-24 14:31:42,005 - agentgraph.reconstruction.content_reference_resolver - INFO - Relation prompt resolution stats: {'total_relations': 6, 'relations_with_refs': 3, 'successful_resolutions': 3, 'failed_resolutions': 0} -2025-07-24 14:31:42,005 - agentgraph.reconstruction.content_reference_resolver - INFO - Resolved content references for knowledge graph with 7 entities and 6 relations -2025-07-24 14:31:42,005 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Resolved content references for window 0 -2025-07-24 14:31:42,006 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Completed sub-batch 1/1 -2025-07-24 14:31:42,006 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Preparing window knowledge graphs with metadata -2025-07-24 14:31:42,006 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Merging knowledge graphs... -2025-07-24 14:31:42,006 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Only one knowledge graph generated, skipping merge process -2025-07-24 14:31:42,006 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing complete. Knowledge graph generated with 7 entities and 6 relations -2025-07-24 14:31:42,006 - __main__ - INFO - Processing text 3/3: text_2 -2025-07-24 14:31:42,006 - __main__ - INFO - Processing text text_2 -2025-07-24 14:31:42,006 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 14:31:42,006 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 434 characters -2025-07-24 14:31:42,006 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 14:31:42,007 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 14:31:42,016 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 14:31:42,016 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 14:31:42,016 - agentgraph.input.content_analysis.semantic_analyzer - INFO - Detected agent trace type: unknown (confidence: 0.12) -2025-07-24 14:31:42,017 - agentgraph.input.text_processing.chunking_service - INFO - Applied rule-based line splitting to 1 chunks (max_line_length=800) -2025-07-24 14:31:42,017 - agentgraph.input.text_processing.chunking_service - INFO - Assigning global line numbers to 1 chunks -2025-07-24 14:31:42,017 - agentgraph.input.text_processing.trace_line_processor - INFO - Added line numbers to 7 lines, starting from line 1 -2025-07-24 14:31:42,017 - agentgraph.input.text_processing.chunking_service - DEBUG - Chunk 0: chars 0-434 → lines 1-7 -2025-07-24 14:31:42,017 - agentgraph.input.text_processing.chunking_service - INFO - Successfully assigned global line numbers to all chunks -2025-07-24 14:31:42,017 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using agent_semantic splitter -2025-07-24 14:31:42,017 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=350000, overlap_size=17500 -2025-07-24 14:31:42,018 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context enhancement complete: 0 total documents -2025-07-24 14:31:42,018 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - SlidingWindowMonitor initialized with model: gpt-4o-mini, method: production -2025-07-24 14:31:42,018 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context documents: 0 provided -2025-07-24 14:31:42,018 - agentgraph.extraction.graph_utilities.knowledge_graph_merger - INFO - KnowledgeGraphMerger initialized with model: gpt-4o-mini -2025-07-24 14:31:42,020 - __main__ - INFO - Extracting knowledge graph from 1 chunks -2025-07-24 14:31:42,020 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks (splitter: agent_semantic, window_size=350000, overlap=0) -2025-07-24 14:31:42,020 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided source trace ID: text_2 -2025-07-24 14:31:42,020 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided processing run ID: batch_text_2 -2025-07-24 14:31:42,020 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks -2025-07-24 14:31:42,020 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using parallel processing with batch size 3 (1 batches) -2025-07-24 14:31:42,020 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 windows in 1 batches -2025-07-24 14:31:42,020 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks in 1 sub-batches (max 1 concurrent per sub-batch) -2025-07-24 14:31:42,020 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Starting sub-batch 1/1 with 1 chunks -2025-07-24 14:31:42,020 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing window 0 -2025-07-24 14:31:42,020 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - 🤖 Starting async_crew extraction for window 0 with method: production using model: gpt-4o-mini -2025-07-24 14:31:42,020 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - No context documents available for window 0, passing empty context string -2025-07-24 14:31:42,020 - agentgraph.methods.production.multi_agent_knowledge_extractor - INFO - Creating agent monitoring crew with model: gpt-4o-mini -2025-07-24 14:32:26,250 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Result Generated (production method) -2025-07-24 14:32:26,251 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 7 lines -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L3-L3 -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L3-L3 -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 2, end_idx: 3 -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Agent: Customer Support AI... -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Agent: Customer Support AI... -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Agent: Customer Support AI... -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:32:26,251 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 1 resolution debug: -2025-07-24 14:32:26,251 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 14:32:26,251 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L3-L3 -2025-07-24 14:32:26,251 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 14:32:26,251 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Agent: Customer Support AI... -2025-07-24 14:32:26,251 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 1: 36 characters -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L2-L2 -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L2-L2 -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 1, end_idx: 2 -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: User Input: I need help with my order... -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: User Input: I need help with my order... -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: User Input: I need help with my order... -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:32:26,251 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 2 resolution debug: -2025-07-24 14:32:26,251 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 14:32:26,251 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L2-L2 -2025-07-24 14:32:26,251 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 14:32:26,251 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: User Input: I need help with my order... -2025-07-24 14:32:26,251 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 2: 47 characters -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L5-L5 -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L5-L5 -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 4, end_idx: 5 -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Tool: order_lookup - searches order data... -2025-07-24 14:32:26,251 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Tool: order_lookup - searches order data... -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Tool: order_lookup - searches order data... -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:32:26,252 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 3 resolution debug: -2025-07-24 14:32:26,252 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 14:32:26,252 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L5-L5 -2025-07-24 14:32:26,252 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 14:32:26,252 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Tool: order_lookup - searches order data... -2025-07-24 14:32:26,252 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 3: 54 characters -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L6-L6 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L6-L6 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 5, end_idx: 6 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Tool: email_sender - sends emails to cus... -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Tool: email_sender - sends emails to cus... -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Tool: email_sender - sends emails to cus... -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:32:26,252 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 4 resolution debug: -2025-07-24 14:32:26,252 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 14:32:26,252 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L6-L6 -2025-07-24 14:32:26,252 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 14:32:26,252 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Tool: email_sender - sends emails to cus... -2025-07-24 14:32:26,252 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 4: 56 characters -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L7-L7 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L7-L7 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 6, end_idx: 7 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Process: Agent uses order_lookup to find... -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Process: Agent uses order_lookup to find... -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Process: Agent uses order_lookup to find... -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:32:26,252 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 5 resolution debug: -2025-07-24 14:32:26,252 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 14:32:26,252 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L7-L7 -2025-07-24 14:32:26,252 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 14:32:26,252 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Process: Agent uses order_lookup to find... -2025-07-24 14:32:26,252 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 5: 136 characters -2025-07-24 14:32:26,252 - agentgraph.reconstruction.content_reference_resolver - INFO - Entity prompt resolution stats: {'total_entities': 5, 'entities_with_refs': 5, 'successful_resolutions': 5, 'failed_resolutions': 0} -2025-07-24 14:32:26,252 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 7 lines -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L2-L2 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L2-L2 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 1, end_idx: 2 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: User Input: I need help with my order... -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: User Input: I need help with my order... -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: User Input: I need help with my order... -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:32:26,252 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation relation_1: 47 characters -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L7-L7 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L7-L7 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 6, end_idx: 7 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Process: Agent uses order_lookup to find... -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Process: Agent uses order_lookup to find... -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Process: Agent uses order_lookup to find... -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:32:26,252 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation relation_4: 136 characters -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L7-L7 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L7-L7 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 6, end_idx: 7 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Process: Agent uses order_lookup to find... -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Process: Agent uses order_lookup to find... -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Process: Agent uses order_lookup to find... -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:32:26,252 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation relation_5: 136 characters -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 14:32:26,252 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 14:32:26,253 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L7-L7 -2025-07-24 14:32:26,253 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L7-L7 -2025-07-24 14:32:26,253 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 6, end_idx: 7 -2025-07-24 14:32:26,253 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 14:32:26,253 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 14:32:26,253 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Process: Agent uses order_lookup to find... -2025-07-24 14:32:26,253 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Process: Agent uses order_lookup to find... -2025-07-24 14:32:26,253 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Process: Agent uses order_lookup to find... -2025-07-24 14:32:26,253 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 14:32:26,253 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation relation_6: 136 characters -2025-07-24 14:32:26,253 - agentgraph.reconstruction.content_reference_resolver - INFO - Relation prompt resolution stats: {'total_relations': 6, 'relations_with_refs': 4, 'successful_resolutions': 4, 'failed_resolutions': 0} -2025-07-24 14:32:26,253 - agentgraph.reconstruction.content_reference_resolver - INFO - Resolved content references for knowledge graph with 5 entities and 6 relations -2025-07-24 14:32:26,253 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Resolved content references for window 0 -2025-07-24 14:32:26,253 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Completed sub-batch 1/1 -2025-07-24 14:32:26,253 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Preparing window knowledge graphs with metadata -2025-07-24 14:32:26,254 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Merging knowledge graphs... -2025-07-24 14:32:26,254 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Only one knowledge graph generated, skipping merge process -2025-07-24 14:32:26,254 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing complete. Knowledge graph generated with 5 entities and 6 relations -2025-07-24 15:29:14,811 - openlit - INFO - Starting openLIT initialization... -2025-07-24 15:29:14,830 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 15:29:15,545 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 15:29:15,597 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 15:29:15,597 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 15:29:15,597 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 15:29:15,962 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 15:29:16,076 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 15:29:16,077 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 15:29:16,678 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 15:29:16,679 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 15:29:18,146 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 15:29:18,149 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 15:29:18,149 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 15:29:18,150 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 15:29:18,150 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 15:29:18,150 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 15:29:18,150 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 15:29:18,150 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 15:29:18,150 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 15:29:18,150 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 15:29:18,151 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 15:29:18,151 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 15:29:18,151 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 15:29:18,151 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 15:29:18,151 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 15:29:18,151 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 15:29:18,151 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 15:29:18,151 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 15:29:18,151 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 15:29:49,092 - openlit - INFO - Starting openLIT initialization... -2025-07-24 15:29:49,111 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 15:29:49,700 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 15:29:49,749 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 15:29:49,749 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 15:29:49,749 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 15:29:50,179 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 15:29:50,297 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 15:29:50,297 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 15:29:50,866 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 15:29:50,867 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 15:29:52,154 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 15:29:52,156 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 15:29:52,156 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 15:29:52,157 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 15:29:52,157 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 15:29:52,157 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 15:29:52,157 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 15:29:52,157 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 15:29:52,157 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 15:29:52,157 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 15:29:52,157 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 15:29:52,157 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 15:29:52,157 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 15:29:52,158 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 15:29:52,158 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 15:29:52,158 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 15:29:52,158 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 15:29:52,158 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 15:29:52,158 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 15:30:04,021 - openlit - INFO - Starting openLIT initialization... -2025-07-24 15:30:04,038 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 15:30:04,603 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 15:30:04,650 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 15:30:04,650 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 15:30:04,650 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 15:30:04,902 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 15:30:05,090 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 15:30:05,090 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 15:30:05,388 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 15:30:05,388 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 15:30:06,562 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 15:30:06,564 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 15:30:06,565 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 15:30:06,565 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 15:30:06,565 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 15:30:06,565 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 15:30:06,565 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 15:30:06,565 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 15:30:06,565 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 15:30:06,565 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 15:30:06,565 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 15:30:06,566 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 15:30:06,566 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 15:30:06,566 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 15:30:06,566 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 15:30:06,566 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 15:30:06,566 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 15:30:06,566 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 15:30:06,566 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 15:55:12,888 - openlit - INFO - Starting openLIT initialization... -2025-07-24 15:55:12,904 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 15:55:13,529 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 15:55:13,580 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 15:55:13,581 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 15:55:13,581 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 15:55:13,962 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 15:55:14,081 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 15:55:14,081 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 15:55:14,679 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 15:55:14,680 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 15:55:16,131 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 15:55:16,133 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 15:55:16,133 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 15:55:16,133 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 15:55:16,134 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 15:55:16,134 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 15:55:16,134 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 15:55:16,134 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 15:55:16,134 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 15:55:16,134 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 15:55:16,134 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 15:55:16,134 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 15:55:16,134 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 15:55:16,134 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 15:55:16,134 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 15:55:16,134 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 15:55:16,135 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 15:55:16,135 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 15:55:16,135 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 15:55:17,084 - agentgraph.input.text_processing.chunking_service - INFO - ChunkingService initialized with batch_size=3, model=gpt-4o-mini -2025-07-24 15:55:17,084 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with json splitter -2025-07-24 15:55:17,084 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 225 characters -2025-07-24 15:55:17,084 - agentgraph.input.text_processing.chunking_service - INFO - Preprocessed content length: 248 characters -2025-07-24 15:55:17,084 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=200, overlap_size=50 -2025-07-24 15:55:17,084 - agentgraph.input.text_processing.text_chunking_strategies - INFO - JSONSplitter initialized with max_chunk_size=200 -2025-07-24 15:55:17,084 - agentgraph.input.text_processing.chunking_service - INFO - Created JSONSplitter with max_chunk_size=200 -2025-07-24 15:55:17,084 - agentgraph.input.text_processing.text_chunking_strategies - INFO - Splitting content into JSON-based chunks -2025-07-24 15:55:17,084 - agentgraph.input.text_processing.text_chunking_strategies - INFO - Split content into 1 JSON-based chunks -2025-07-24 15:55:17,085 - agentgraph.input.text_processing.chunking_service - INFO - Assigning global line numbers to 1 chunks -2025-07-24 15:55:17,085 - agentgraph.input.text_processing.trace_line_processor - INFO - Added line numbers to 2 lines, starting from line 1 -2025-07-24 15:55:17,085 - agentgraph.input.text_processing.chunking_service - INFO - Successfully assigned global line numbers to all chunks -2025-07-24 15:55:17,085 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using json splitter -2025-07-24 15:55:17,085 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=200, overlap_size=50 -2025-07-24 15:55:34,254 - openlit - INFO - Starting openLIT initialization... -2025-07-24 15:55:34,270 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 15:55:34,852 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 15:55:34,898 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 15:55:34,898 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 15:55:34,898 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 15:55:35,238 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 15:55:35,338 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 15:55:35,339 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 15:55:35,890 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 15:55:35,891 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 15:55:37,263 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 15:55:37,265 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 15:55:37,265 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 15:55:37,265 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 15:55:37,265 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 15:55:37,265 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 15:55:37,266 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 15:55:37,266 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 15:55:37,266 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 15:55:37,266 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 15:55:37,266 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 15:55:37,266 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 15:55:37,266 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 15:55:37,266 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 15:55:37,266 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 15:55:37,266 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 15:55:37,266 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 15:55:37,266 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 15:55:37,266 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 15:55:49,066 - openlit - INFO - Starting openLIT initialization... -2025-07-24 15:55:49,082 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 15:55:49,549 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 15:55:49,587 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 15:55:49,587 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 15:55:49,587 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 15:55:49,881 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 15:55:49,965 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 15:55:49,965 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 15:55:50,422 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 15:55:50,423 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 15:55:51,495 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 15:55:51,497 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 15:55:51,497 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 15:55:51,498 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 15:55:51,498 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 15:55:51,498 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 15:55:51,498 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 15:55:51,498 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 15:55:51,498 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 15:55:51,498 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 15:55:51,498 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 15:55:51,498 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 15:55:51,498 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 15:55:51,499 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 15:55:51,499 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 15:55:51,499 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 15:55:51,499 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 15:55:51,499 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 15:55:51,499 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 15:55:52,278 - __main__ - INFO - Processing single text input -2025-07-24 15:55:52,278 - agentgraph.input.text_processing.chunking_service - INFO - ChunkingService initialized with batch_size=3, model=gpt-4o-mini -2025-07-24 15:55:52,278 - __main__ - INFO - Initialized BatchKGExtractor with model: gpt-4o-mini, method: production -2025-07-24 15:55:52,278 - __main__ - INFO - Processing batch of 1 texts -2025-07-24 15:55:52,278 - __main__ - INFO - Processing text 1/1: text_0 -2025-07-24 15:55:52,278 - __main__ - INFO - Processing text text_0 -2025-07-24 15:55:52,278 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 15:55:52,278 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 135 characters -2025-07-24 15:55:52,278 - agentgraph.input.text_processing.chunking_service - INFO - Preprocessed content length: 161 characters -2025-07-24 15:55:52,278 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 15:55:52,278 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 15:55:52,284 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 15:55:52,284 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 15:55:52,287 - agentgraph.input.content_analysis.semantic_analyzer - INFO - Detected agent trace type: unknown (confidence: 0.14) -2025-07-24 15:55:52,287 - agentgraph.input.text_processing.chunking_service - INFO - Applied rule-based line splitting to 1 chunks (max_line_length=800) -2025-07-24 15:55:52,288 - agentgraph.input.text_processing.chunking_service - INFO - Assigning global line numbers to 1 chunks -2025-07-24 15:55:52,288 - agentgraph.input.text_processing.trace_line_processor - INFO - Added line numbers to 10 lines, starting from line 1 -2025-07-24 15:55:52,288 - agentgraph.input.text_processing.chunking_service - DEBUG - Chunk 0: chars 0-161 → lines 1-10 -2025-07-24 15:55:52,288 - agentgraph.input.text_processing.chunking_service - INFO - Successfully assigned global line numbers to all chunks -2025-07-24 15:55:52,288 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using agent_semantic splitter -2025-07-24 15:55:52,288 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=350000, overlap_size=17500 -2025-07-24 15:55:52,288 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context enhancement complete: 0 total documents -2025-07-24 15:55:52,288 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - SlidingWindowMonitor initialized with model: gpt-4o-mini, method: production -2025-07-24 15:55:52,288 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context documents: 0 provided -2025-07-24 15:55:52,288 - agentgraph.extraction.graph_utilities.knowledge_graph_merger - INFO - KnowledgeGraphMerger initialized with model: gpt-4o-mini -2025-07-24 15:55:52,290 - __main__ - INFO - Extracting knowledge graph from 1 chunks -2025-07-24 15:55:52,290 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks (splitter: agent_semantic, window_size=350000, overlap=0) -2025-07-24 15:55:52,290 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided source trace ID: text_0 -2025-07-24 15:55:52,290 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided processing run ID: batch_text_0 -2025-07-24 15:55:52,290 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks -2025-07-24 15:55:52,290 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using parallel processing with batch size 3 (1 batches) -2025-07-24 15:55:52,290 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 windows in 1 batches -2025-07-24 15:55:52,301 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks in 1 sub-batches (max 1 concurrent per sub-batch) -2025-07-24 15:55:52,301 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Starting sub-batch 1/1 with 1 chunks -2025-07-24 15:55:52,301 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing window 0 -2025-07-24 15:55:52,301 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - 🤖 Starting async_crew extraction for window 0 with method: production using model: gpt-4o-mini -2025-07-24 15:55:52,301 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - No context documents available for window 0, passing empty context string -2025-07-24 15:55:52,301 - agentgraph.methods.production.multi_agent_knowledge_extractor - INFO - Creating agent monitoring crew with model: gpt-4o-mini -2025-07-24 15:55:53,514 - asyncio - ERROR - Task was destroyed but it is pending! -task: wait_for= cb=[gather.._done_callback() at /Users/zekunwu/anaconda3/lib/python3.11/asyncio/tasks.py:764]> -2025-07-24 15:55:53,514 - asyncio - ERROR - Task exception was never retrieved -future: exception=SystemExit(1)> -Traceback (most recent call last): - File "/Users/zekunwu/anaconda3/lib/python3.11/asyncio/tasks.py", line 476, in wait_for - await waiter -asyncio.exceptions.CancelledError - -During handling of the above exception, another exception occurred: - -Traceback (most recent call last): - File "/Users/zekunwu/anaconda3/lib/python3.11/asyncio/runners.py", line 189, in run - with Runner(debug=debug) as runner: - File "/Users/zekunwu/anaconda3/lib/python3.11/asyncio/runners.py", line 63, in __exit__ - self.close() - File "/Users/zekunwu/anaconda3/lib/python3.11/asyncio/runners.py", line 71, in close - _cancel_all_tasks(loop) - File "/Users/zekunwu/anaconda3/lib/python3.11/asyncio/runners.py", line 201, in _cancel_all_tasks - loop.run_until_complete(tasks.gather(*to_cancel, return_exceptions=True)) - File "/Users/zekunwu/anaconda3/lib/python3.11/asyncio/base_events.py", line 641, in run_until_complete - self.run_forever() - File "/Users/zekunwu/anaconda3/lib/python3.11/asyncio/base_events.py", line 608, in run_forever - self._run_once() - File "/Users/zekunwu/anaconda3/lib/python3.11/asyncio/base_events.py", line 1936, in _run_once - handle._run() - File "/Users/zekunwu/anaconda3/lib/python3.11/asyncio/events.py", line 84, in _run - self._context.run(self._callback, *self._args) - File "/Users/zekunwu/Desktop/agent_monitoring/agentgraph/extraction/graph_processing/knowledge_graph_processor.py", line 238, in process_window - result = await asyncio.wait_for( - ^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/asyncio/tasks.py", line 479, in wait_for - return fut.result() - ^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/crewai/crew.py", line 693, in kickoff_async - return await asyncio.to_thread(self.kickoff, inputs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/asyncio/threads.py", line 25, in to_thread - return await loop.run_in_executor(None, func_call) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/concurrent/futures/thread.py", line 58, in run - result = self.fn(*self.args, **self.kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/crewai/crew.py", line 646, in kickoff - result = self._run_sequential_process() - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/crewai/crew.py", line 758, in _run_sequential_process - return self._execute_tasks(self.tasks) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/crewai/crew.py", line 861, in _execute_tasks - task_output = task.execute_sync( - ^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/crewai/task.py", line 328, in execute_sync - return self._execute_core(agent, context, tools) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/openlit/instrumentation/crewai/crewai.py", line 67, in wrapper - response = wrapped(*args, **kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/crewai/task.py", line 392, in _execute_core - result = agent.execute_task( - ^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/openlit/instrumentation/crewai/crewai.py", line 67, in wrapper - response = wrapped(*args, **kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/crewai/agent.py", line 250, in execute_task - result = self.agent_executor.invoke( - ^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/crewai/agents/crew_agent_executor.py", line 112, in invoke - formatted_answer = self._invoke_loop() - ^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/crewai/agents/crew_agent_executor.py", line 155, in _invoke_loop - answer = get_llm_response( - ^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/crewai/utilities/agent_utils.py", line 148, in get_llm_response - answer = llm.call( - ^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/crewai/llm.py", line 756, in call - crewai_event_bus.emit( - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/crewai/utilities/events/crewai_event_bus.py", line 73, in emit - handler(source, event) - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/crewai/utilities/events/event_listener.py", line 288, in on_llm_call_started - self.formatter.handle_llm_call_started( - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/crewai/utilities/events/utils/console_formatter.py", line 511, in handle_llm_call_started - self.print(tree_to_use) - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/crewai/utilities/events/utils/console_formatter.py", line 72, in print - self.console.print(*args, **kwargs) - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/rich/console.py", line 1678, in print - with self: - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/rich/console.py", line 864, in __exit__ - self._exit_buffer() - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/rich/console.py", line 822, in _exit_buffer - self._check_buffer() - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/rich/console.py", line 2021, in _check_buffer - self.on_broken_pipe() - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/rich/console.py", line 2006, in on_broken_pipe - raise SystemExit(1) -SystemExit: 1 -2025-07-24 16:21:52,413 - openlit - INFO - Starting openLIT initialization... -2025-07-24 16:21:52,429 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 16:21:53,054 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 16:21:53,106 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 16:21:53,106 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 16:21:53,106 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 16:21:53,467 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 16:21:53,581 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 16:21:53,581 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 16:21:54,185 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 16:21:54,187 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 16:21:55,617 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 16:21:55,619 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 16:21:55,619 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 16:21:55,620 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 16:21:55,620 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 16:21:55,620 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 16:21:55,620 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 16:21:55,620 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 16:21:55,620 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 16:21:55,620 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 16:21:55,621 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 16:21:55,621 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 16:21:55,621 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 16:21:55,621 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 16:21:55,621 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 16:21:55,621 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 16:21:55,621 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 16:21:55,621 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 16:21:55,621 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 16:22:08,469 - openlit - INFO - Starting openLIT initialization... -2025-07-24 16:22:08,485 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 16:22:09,085 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 16:22:09,132 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 16:22:09,133 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 16:22:09,133 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 16:22:09,497 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 16:22:09,605 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 16:22:09,605 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 16:22:10,171 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 16:22:10,172 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 16:22:11,447 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 16:22:11,450 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 16:22:11,450 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 16:22:11,450 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 16:22:11,451 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 16:22:11,451 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 16:22:11,451 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 16:22:11,451 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 16:22:11,451 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 16:22:11,451 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 16:22:11,451 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 16:22:11,451 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 16:22:11,451 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 16:22:11,452 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 16:22:11,452 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 16:22:11,452 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 16:22:11,452 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 16:22:11,452 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 16:22:11,452 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 16:38:26,975 - openlit - INFO - Starting openLIT initialization... -2025-07-24 16:38:26,991 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 16:38:27,679 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 16:38:27,737 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 16:38:27,737 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 16:38:27,737 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 16:38:28,035 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 16:38:28,138 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 16:38:28,138 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 16:38:28,630 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 16:38:28,631 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 16:38:30,112 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 16:38:30,115 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 16:38:30,115 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 16:38:30,116 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 16:38:30,116 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 16:38:30,116 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 16:38:30,116 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 16:38:30,116 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 16:38:30,116 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 16:38:30,116 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 16:38:30,117 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 16:38:30,117 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 16:38:30,117 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 16:38:30,117 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 16:38:30,117 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 16:38:30,117 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 16:38:30,117 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 16:38:30,117 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 16:38:30,117 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 16:55:34,677 - openlit - INFO - Starting openLIT initialization... -2025-07-24 16:55:34,694 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 16:55:35,310 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 16:55:35,361 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 16:55:35,362 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 16:55:35,362 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 16:55:35,643 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 16:55:35,736 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 16:55:35,736 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 16:55:36,179 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 16:55:36,180 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 16:55:37,442 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 16:55:37,445 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 16:55:37,445 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 16:55:37,445 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 16:55:37,445 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 16:55:37,445 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 16:55:37,445 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 16:55:37,445 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 16:55:37,445 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 16:55:37,446 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 16:55:37,446 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 16:55:37,446 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 16:55:37,446 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 16:55:37,446 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 16:55:37,446 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 16:55:37,446 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 16:55:37,446 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 16:55:37,446 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 16:55:37,446 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 17:06:54,583 - openlit - INFO - Starting openLIT initialization... -2025-07-24 17:06:54,600 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 17:06:55,235 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 17:06:55,290 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 17:06:55,290 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 17:06:55,290 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 17:06:55,668 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 17:06:55,867 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 17:06:55,867 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 17:06:56,163 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 17:06:56,164 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 17:06:57,396 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 17:06:57,399 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 17:06:57,399 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 17:06:57,399 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 17:06:57,399 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 17:06:57,399 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 17:06:57,399 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 17:06:57,400 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 17:06:57,400 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 17:06:57,400 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 17:06:57,400 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 17:06:57,400 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 17:06:57,400 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 17:06:57,400 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 17:06:57,400 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 17:06:57,400 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 17:06:57,400 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 17:06:57,400 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 17:06:57,401 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 17:16:06,173 - openlit - INFO - Starting openLIT initialization... -2025-07-24 17:16:06,191 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 17:16:06,839 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 17:16:06,893 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 17:16:06,893 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 17:16:06,893 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 17:16:07,274 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 17:16:07,393 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 17:16:07,393 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 17:16:08,027 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 17:16:08,028 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 17:16:09,522 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 17:16:09,525 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 17:16:09,525 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 17:16:09,526 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 17:16:09,526 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 17:16:09,526 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 17:16:09,526 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 17:16:09,526 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 17:16:09,526 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 17:16:09,526 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 17:16:09,526 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 17:16:09,526 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 17:16:09,526 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 17:16:09,527 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 17:16:09,527 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 17:16:09,527 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 17:16:09,527 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 17:16:09,527 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 17:16:09,527 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 17:16:30,166 - openlit - INFO - Starting openLIT initialization... -2025-07-24 17:16:30,183 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 17:16:30,794 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 17:16:30,844 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 17:16:30,845 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 17:16:30,845 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 17:16:31,211 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 17:16:31,326 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 17:16:31,326 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 17:16:31,927 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 17:16:31,928 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 17:16:34,471 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 17:16:34,474 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 17:16:34,474 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 17:16:34,474 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 17:16:34,474 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 17:16:34,474 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 17:16:34,474 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 17:16:34,474 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 17:16:34,474 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 17:16:34,474 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 17:16:34,474 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 17:16:34,475 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 17:16:34,475 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 17:16:34,475 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 17:16:34,475 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 17:16:34,475 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 17:16:34,475 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 17:16:34,475 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 17:16:34,475 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 17:18:15,628 - openlit - INFO - Starting openLIT initialization... -2025-07-24 17:18:15,646 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 17:18:16,206 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 17:18:16,253 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 17:18:16,254 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 17:18:16,254 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 17:18:16,597 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 17:18:16,701 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 17:18:16,701 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 17:18:17,262 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 17:18:17,263 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 17:18:18,523 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 17:18:18,525 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 17:18:18,525 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 17:18:18,525 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 17:18:18,525 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 17:18:18,525 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 17:18:18,525 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 17:18:18,526 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 17:18:18,526 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 17:18:18,526 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 17:18:18,526 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 17:18:18,526 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 17:18:18,526 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 17:18:18,526 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 17:18:18,526 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 17:18:18,526 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 17:18:18,526 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 17:18:18,526 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 17:18:18,526 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 17:18:19,656 - __main__ - INFO - Loaded 3 texts from scripts/example_texts.json -2025-07-24 17:18:19,656 - agentgraph.input.text_processing.chunking_service - INFO - ChunkingService initialized with batch_size=3, model=gpt-4o-mini -2025-07-24 17:18:19,656 - __main__ - INFO - Initialized BatchKGExtractor: -2025-07-24 17:18:19,656 - __main__ - INFO - - Model: gpt-4o-mini, Method: production -2025-07-24 17:18:19,656 - __main__ - INFO - - Preprocessing: False -2025-07-24 17:18:19,656 - __main__ - INFO - - Line numbers: False -2025-07-24 17:18:19,656 - __main__ - INFO - Processing batch of 3 texts -2025-07-24 17:18:19,656 - __main__ - INFO - Processing text 1/3: text_0 -2025-07-24 17:18:19,656 - __main__ - INFO - Processing text text_0 (format: auto) -2025-07-24 17:18:19,656 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 17:18:19,656 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 360 characters -2025-07-24 17:18:19,656 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 17:18:19,656 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 17:18:19,662 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 17:18:19,662 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 17:18:19,665 - agentgraph.input.content_analysis.semantic_analyzer - INFO - Detected agent trace type: unknown (confidence: 0.12) -2025-07-24 17:18:19,665 - agentgraph.input.text_processing.chunking_service - INFO - Applied rule-based line splitting to 1 chunks (max_line_length=800) -2025-07-24 17:18:19,666 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using agent_semantic splitter -2025-07-24 17:18:19,666 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=350000, overlap_size=17500 -2025-07-24 17:18:19,667 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context enhancement complete: 0 total documents -2025-07-24 17:18:19,667 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - SlidingWindowMonitor initialized with model: gpt-4o-mini, method: production -2025-07-24 17:18:19,667 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context documents: 0 provided -2025-07-24 17:18:19,667 - agentgraph.extraction.graph_utilities.knowledge_graph_merger - INFO - KnowledgeGraphMerger initialized with model: gpt-4o-mini -2025-07-24 17:18:19,668 - __main__ - INFO - Extracting knowledge graph from 1 chunks -2025-07-24 17:18:19,668 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks (splitter: agent_semantic, window_size=350000, overlap=0) -2025-07-24 17:18:19,668 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided source trace ID: text_0 -2025-07-24 17:18:19,668 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided processing run ID: batch_text_0 -2025-07-24 17:18:19,668 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks -2025-07-24 17:18:19,668 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using parallel processing with batch size 3 (1 batches) -2025-07-24 17:18:19,668 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 windows in 1 batches -2025-07-24 17:18:19,678 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks in 1 sub-batches (max 1 concurrent per sub-batch) -2025-07-24 17:18:19,678 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Starting sub-batch 1/1 with 1 chunks -2025-07-24 17:18:19,679 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing window 0 -2025-07-24 17:18:19,679 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - 🤖 Starting async_crew extraction for window 0 with method: production using model: gpt-4o-mini -2025-07-24 17:18:19,679 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - No context documents available for window 0, passing empty context string -2025-07-24 17:18:19,679 - agentgraph.methods.production.multi_agent_knowledge_extractor - INFO - Creating agent monitoring crew with model: gpt-4o-mini -2025-07-24 17:19:05,054 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Result Generated (production method) -2025-07-24 17:19:05,054 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 6 lines -2025-07-24 17:19:05,055 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:19:05,055 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 6 -2025-07-24 17:19:05,055 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L2-L2 -2025-07-24 17:19:05,055 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L2-L2 -2025-07-24 17:19:05,055 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 1, end_idx: 2 -2025-07-24 17:19:05,055 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 6 -2025-07-24 17:19:05,055 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:19:05,055 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Role: You are a research assistant AI that he... -2025-07-24 17:19:05,055 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Role: You are a research assistant AI that he... -2025-07-24 17:19:05,055 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Role: You are a research assistant AI that he... -2025-07-24 17:19:05,055 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:19:05,055 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 1 resolution debug: -2025-07-24 17:19:05,055 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:19:05,055 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L2-L2 -2025-07-24 17:19:05,055 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:19:05,055 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Role: You are a research assistant AI that he... -2025-07-24 17:19:05,055 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 1: 91 characters -2025-07-24 17:19:05,055 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:19:05,055 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 6 -2025-07-24 17:19:05,055 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L4-L4 -2025-07-24 17:19:05,055 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L4-L4 -2025-07-24 17:19:05,055 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 3, end_idx: 4 -2025-07-24 17:19:05,055 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 6 -2025-07-24 17:19:05,055 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:19:05,055 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Tool: web_search... -2025-07-24 17:19:05,055 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Tool: web_search... -2025-07-24 17:19:05,055 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Tool: web_search... -2025-07-24 17:19:05,055 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:19:05,055 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 2 resolution debug: -2025-07-24 17:19:05,055 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:19:05,055 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L4-L4 -2025-07-24 17:19:05,055 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:19:05,055 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Tool: web_search... -2025-07-24 17:19:05,055 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 2: 21 characters -2025-07-24 17:19:05,055 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:19:05,055 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 6 -2025-07-24 17:19:05,055 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L5-L5 -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L5-L5 -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 4, end_idx: 5 -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 6 -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Input: climate change impacts 2024... -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Input: climate change impacts 2024... -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Input: climate change impacts 2024... -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:19:05,056 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 3 resolution debug: -2025-07-24 17:19:05,056 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:19:05,056 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L5-L5 -2025-07-24 17:19:05,056 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:19:05,056 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Input: climate change impacts 2024... -2025-07-24 17:19:05,056 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 3: 39 characters -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 6 -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L6-L6 -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L6-L6 -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 5, end_idx: 6 -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 6 -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Output: Climate change continues to show sign... -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Output: Climate change continues to show sign... -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Output: Climate change continues to show sign... -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:19:05,056 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 4 resolution debug: -2025-07-24 17:19:05,056 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:19:05,056 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L6-L6 -2025-07-24 17:19:05,056 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:19:05,056 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Output: Climate change continues to show sign... -2025-07-24 17:19:05,056 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 4: 150 characters -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 6 -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L7-L7 -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L7-L7 -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 6, end_idx: 7 -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 6 -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - WARNING - Line range in ContentReference is out of bounds. -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - bounds check failed: start_idx=6, end_idx=7, lines_len=6 -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: EMPTY... -2025-07-24 17:19:05,056 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:19:05,056 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 5 resolution debug: -2025-07-24 17:19:05,056 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:19:05,056 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L7-L7 -2025-07-24 17:19:05,056 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:19:05,056 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: EMPTY... -2025-07-24 17:19:05,056 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 5: 0 characters -2025-07-24 17:19:05,056 - agentgraph.reconstruction.content_reference_resolver - INFO - Entity prompt resolution stats: {'total_entities': 5, 'entities_with_refs': 5, 'successful_resolutions': 5, 'failed_resolutions': 0} -2025-07-24 17:19:05,057 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 6 lines -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 6 -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L6-L6 -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L6-L6 -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 5, end_idx: 6 -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 6 -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Output: Climate change continues to show sign... -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Output: Climate change continues to show sign... -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Output: Climate change continues to show sign... -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:19:05,057 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation relation_1: 150 characters -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 6 -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L5-L5 -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L5-L5 -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 4, end_idx: 5 -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 6 -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Input: climate change impacts 2024... -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Input: climate change impacts 2024... -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Input: climate change impacts 2024... -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:19:05,057 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation relation_3: 39 characters -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 6 -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L5-L5 -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L5-L5 -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 4, end_idx: 5 -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 6 -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Input: climate change impacts 2024... -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Input: climate change impacts 2024... -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Input: climate change impacts 2024... -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:19:05,057 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation relation_4: 39 characters -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 6 -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L7-L7 -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L7-L7 -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 6, end_idx: 7 -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 6 -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - WARNING - Line range in ContentReference is out of bounds. -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - bounds check failed: start_idx=6, end_idx=7, lines_len=6 -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: EMPTY... -2025-07-24 17:19:05,057 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:19:05,057 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation relation_6: 0 characters -2025-07-24 17:19:05,057 - agentgraph.reconstruction.content_reference_resolver - INFO - Relation prompt resolution stats: {'total_relations': 6, 'relations_with_refs': 4, 'successful_resolutions': 4, 'failed_resolutions': 0} -2025-07-24 17:19:05,058 - agentgraph.reconstruction.content_reference_resolver - INFO - Resolved content references for knowledge graph with 5 entities and 6 relations -2025-07-24 17:19:05,058 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Resolved content references for window 0 -2025-07-24 17:19:05,058 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Completed sub-batch 1/1 -2025-07-24 17:19:05,058 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Preparing window knowledge graphs with metadata -2025-07-24 17:19:05,059 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Merging knowledge graphs... -2025-07-24 17:19:05,059 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Only one knowledge graph generated, skipping merge process -2025-07-24 17:19:05,059 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing complete. Knowledge graph generated with 5 entities and 6 relations -2025-07-24 17:19:05,059 - __main__ - INFO - Processing text 2/3: text_1 -2025-07-24 17:19:05,059 - __main__ - INFO - Processing text text_1 (format: auto) -2025-07-24 17:19:05,059 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 17:19:05,059 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 445 characters -2025-07-24 17:19:05,059 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 17:19:05,059 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 17:19:05,070 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 17:19:05,070 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 17:19:05,070 - agentgraph.input.content_analysis.semantic_analyzer - INFO - Detected agent trace type: unknown (confidence: 0.12) -2025-07-24 17:19:05,070 - agentgraph.input.text_processing.chunking_service - INFO - Applied rule-based line splitting to 1 chunks (max_line_length=800) -2025-07-24 17:19:05,071 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using agent_semantic splitter -2025-07-24 17:19:05,071 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=350000, overlap_size=17500 -2025-07-24 17:19:05,071 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context enhancement complete: 0 total documents -2025-07-24 17:19:05,071 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - SlidingWindowMonitor initialized with model: gpt-4o-mini, method: production -2025-07-24 17:19:05,071 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context documents: 0 provided -2025-07-24 17:19:05,071 - agentgraph.extraction.graph_utilities.knowledge_graph_merger - INFO - KnowledgeGraphMerger initialized with model: gpt-4o-mini -2025-07-24 17:19:05,074 - __main__ - INFO - Extracting knowledge graph from 1 chunks -2025-07-24 17:19:05,074 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks (splitter: agent_semantic, window_size=350000, overlap=0) -2025-07-24 17:19:05,074 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided source trace ID: text_1 -2025-07-24 17:19:05,074 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided processing run ID: batch_text_1 -2025-07-24 17:19:05,074 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks -2025-07-24 17:19:05,074 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using parallel processing with batch size 3 (1 batches) -2025-07-24 17:19:05,074 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 windows in 1 batches -2025-07-24 17:19:05,074 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks in 1 sub-batches (max 1 concurrent per sub-batch) -2025-07-24 17:19:05,074 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Starting sub-batch 1/1 with 1 chunks -2025-07-24 17:19:05,074 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing window 0 -2025-07-24 17:19:05,074 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - 🤖 Starting async_crew extraction for window 0 with method: production using model: gpt-4o-mini -2025-07-24 17:19:05,075 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - No context documents available for window 0, passing empty context string -2025-07-24 17:19:05,075 - agentgraph.methods.production.multi_agent_knowledge_extractor - INFO - Creating agent monitoring crew with model: gpt-4o-mini -2025-07-24 17:19:15,429 - instructor - DEBUG - Patching `client.chat.completions.create` with mode= -2025-07-24 17:19:15,434 - instructor - DEBUG - Instructor Request: mode.value='tool_call', response_model=, new_kwargs={'messages': [{'role': 'user', 'content': '{\n "entities": [\n {\n "id": "agent_1",\n "type": "Agent",\n "name": "Data Analyst",\n "importance": "HIGH",\n "raw_prompt": "responsible for analyzing customer data",\n "raw_prompt_ref": [\n {\n "line_start": 4,\n "line_end": 4,\n "confidence": 1.0\n }\n ]\n },\n {\n "id": "agent_2",\n "type": "Agent",\n "name": "Report Generator",\n "importance": "HIGH",\n "raw_prompt": "creates comprehensive reports",\n "raw_prompt_ref": [\n {\n "line_start": 5,\n "line_end": 5,\n "confidence": 1.0\n }\n ]\n },\n {\n "id": "task_1",\n "type": "Task",\n "name": "Analyze sales data from Q4 2023",\n "importance": "HIGH",\n "raw_prompt": "Analyze sales data from Q4 2023",\n "raw_prompt_ref": [\n {\n "line_start": 6,\n "line_end": 6,\n "confidence": 1.0\n }\n ]\n },\n {\n "id": "task_2",\n "type": "Task",\n "name": "Generate executive summary report",\n "importance": "HIGH",\n "raw_prompt": "Generate executive summary report",\n "raw_prompt_ref": [\n {\n "line_start": 7,\n "line_end": 7,\n "confidence": 1.0\n }\n ]\n },\n {\n "id": "tool_1",\n "type": "Tool",\n "name": "pandas_analyzer",\n "importance": "HIGH",\n "raw_prompt": "analyzes CSV data",\n "raw_prompt_ref": [\n {\n "line_start": 8,\n "line_end": 8,\n "confidence": 1.0\n }\n ]\n },\n {\n "id": "tool_2",\n "type": "Tool",\n "name": "report_generator",\n "importance": "HIGH",\n "raw_prompt": "creates PDF reports",\n "raw_prompt_ref": [\n {\n "line_start": 9,\n "line_end": 9,\n "confidence": 1.0\n }\n ]\n },\n {\n "id": "flow_1",\n "type": "Flow",\n "name": "Data Analyst performs Task 1 using pandas_analyzer, then Report Generator performs Task 2 using report_generator",\n "importance": "HIGH",\n "raw_prompt": "Data Analyst performs Task 1 using pandas_analyzer, then Report Generator performs Task 2 using report_generator",\n "raw_prompt_ref": [\n {\n "line_start": 10,\n "line_end": 10,\n "confidence": 1.0\n }\n ]\n }\n ]\n}'}], 'model': 'gpt-4o-mini', 'tools': [{'type': 'function', 'function': {'name': 'EntityExtractionList', 'description': 'Correctly extracted `EntityExtractionList` with all the required parameters with correct types', 'parameters': {'$defs': {'ContentReference': {'description': 'Reference to content location in the original trace using line numbers and character positions.\nThis allows AI agents to provide position metadata instead of full content, enabling \nefficient mapping back to the original trace while reducing hallucination risks.\n\nCRITICAL FOR LLMs: Line counting accuracy is essential for proper content resolution.\nUse systematic counting methods and verify your line numbers before submission.', 'properties': {'line_start': {'description': 'Starting line number where the content begins (1-based indexing from , ... markers).\n \n ACCURACY REQUIREMENTS FOR LLMs:\n - Count markers systematically from the beginning of the input\n - Use anchor points: find distinctive text first, then count nearby lines\n - Double-check by counting backwards from a known reference point\n - For multi-line content, this should be the FIRST line containing the content\n - In key-value pairs (e.g. "content": "..."), reference the line where the VALUE starts, not the key\n \n COMMON ERRORS TO AVOID:\n - Miscounting due to skipping indented continuation lines\n - Confusing line numbers when content spans multiple markers\n - Using approximate counting instead of precise marker identification\n \n VERIFICATION: Before submitting, locate your chosen line number and confirm it contains the expected content start.', 'title': 'Line Start', 'type': 'integer'}, 'line_end': {'description': 'Ending line number where content ends (1-based indexing from , ... markers).\n \n ACCURACY REQUIREMENTS FOR LLMs:\n - Must be >= line_start (validation will fail otherwise)\n - For single-line content, line_end should equal line_start\n - For multi-line content, find the LAST line containing the content\n - Include indented continuation lines that are part of the same logical content block\n \n VERIFICATION STRATEGY:\n - Count from line_start to ensure proper range\n - Confirm the line_end marker contains the actual end of the content\n - Check that no content continues beyond your specified line_end', 'title': 'Line End', 'type': 'integer'}, 'confidence': {'default': None, 'description': 'Confidence score for the location reference accuracy (0.0 - 1.0).\n \n CONFIDENCE SCORING GUIDE FOR LLMs:\n - 1.0: Verified by counting twice with consistent results, clear content boundaries\n - 0.9: High confidence with single verification, unambiguous content location \n - 0.8: Good confidence but content boundaries somewhat ambiguous\n - 0.7: Moderate confidence, some uncertainty in exact line boundaries\n - 0.6: Lower confidence due to complex content structure or counting difficulty\n - 0.5 or below: Uncertain about accuracy, recommend manual verification\n \n FACTORS AFFECTING CONFIDENCE:\n - Clarity of content boundaries (higher = more confident)\n - Complexity of surrounding text (simpler = more confident) \n - Verification method used (double-counting = more confident)\n - Presence of clear anchor points (more anchors = more confident)', 'maximum': 1.0, 'minimum': 0.0, 'title': 'Confidence', 'type': 'number'}}, 'required': ['line_start', 'line_end'], 'title': 'ContentReference', 'type': 'object'}, 'Entity': {'properties': {'id': {'description': 'Unique identifier for the entity', 'title': 'Id', 'type': 'string'}, 'type': {'description': 'Type of entity defined by prompt type: Agent (system prompt), Task (instruction prompt), Tool (description prompt), Input (input format prompt), Output (output format prompt), Human (optional prompt). The raw_prompt field is the primary distinguishing factor for entity uniqueness and classification.', 'enum': ['Agent', 'Task', 'Tool', 'Input', 'Output', 'Human'], 'title': 'Type', 'type': 'string'}, 'name': {'description': "Name of the entity derived from the prompt content. Names should reflect the specific prompt or specification that defines this entity. For composite entities, use descriptive names that capture the prompt's scope (e.g., 'SQL Query Generation System Prompt', 'Data Analysis Instruction Set').", 'title': 'Name', 'type': 'string'}, 'importance': {'description': 'Importance level of this entity in the system. HIGH: Core agents, critical tasks, essential tools that are central to system function. MEDIUM: Supporting agents, standard tasks, commonly used tools. LOW: Auxiliary entities, simple tasks, rarely used components.', 'enum': ['HIGH', 'MEDIUM', 'LOW'], 'title': 'Importance', 'type': 'string'}, 'raw_prompt': {'default': '', 'description': 'PRIMARY DISTINGUISHING CONTENT: The actual prompt, specification, or instruction that defines this entity. This is the core content that makes each entity unique and should contain: For Agents (system prompts defining role/capabilities), For Tasks (instruction prompts defining objectives), For Tools (description prompts defining functionality), For Inputs (format specifications), For Outputs (format specifications), For Humans (interaction patterns). This field is more important than the name for entity distinction and relationship mapping.', 'title': 'Raw Prompt', 'type': 'string'}, 'raw_prompt_ref': {'description': 'A list of references to the locations of the raw prompt content in the original trace. When provided, this allows mapping back to all exact positions in the trace where this prompt was found.', 'items': {'$ref': '#/$defs/ContentReference'}, 'title': 'Raw Prompt Ref', 'type': 'array'}}, 'required': ['id', 'type', 'name', 'importance'], 'title': 'Entity', 'type': 'object'}}, 'properties': {'entities': {'default': [], 'items': {'$ref': '#/$defs/Entity'}, 'title': 'Entities', 'type': 'array'}}, 'type': 'object', 'required': []}}}], 'tool_choice': {'type': 'function', 'function': {'name': 'EntityExtractionList'}}} -2025-07-24 17:19:15,435 - instructor - DEBUG - max_retries: 3 -2025-07-24 17:19:15,435 - instructor - DEBUG - Retrying, attempt: 1 -2025-07-24 17:19:23,424 - instructor - DEBUG - Instructor Raw Response: ModelResponse(id='chatcmpl-Bwsx1zPUXi70jSDZiM1LR34XydZID', created=1753373955, model='gpt-4o-mini-2024-07-18', object='chat.completion', system_fingerprint=None, choices=[Choices(finish_reason='stop', index=0, message=Message(content=None, role='assistant', tool_calls=[ChatCompletionMessageToolCall(function=Function(arguments='{"entities":[{"id":"agent_1","type":"Agent","name":"Data Analyst","importance":"HIGH","raw_prompt":"responsible for analyzing customer data","raw_prompt_ref":[{"line_start":4,"line_end":4,"confidence":1.0}]},{"id":"agent_2","type":"Agent","name":"Report Generator","importance":"HIGH","raw_prompt":"creates comprehensive reports","raw_prompt_ref":[{"line_start":5,"line_end":5,"confidence":1.0}]},{"id":"task_1","type":"Task","name":"Analyze sales data from Q4 2023","importance":"HIGH","raw_prompt":"Analyze sales data from Q4 2023","raw_prompt_ref":[{"line_start":6,"line_end":6,"confidence":1.0}]},{"id":"task_2","type":"Task","name":"Generate executive summary report","importance":"HIGH","raw_prompt":"Generate executive summary report","raw_prompt_ref":[{"line_start":7,"line_end":7,"confidence":1.0}]},{"id":"tool_1","type":"Tool","name":"pandas_analyzer","importance":"HIGH","raw_prompt":"analyzes CSV data","raw_prompt_ref":[{"line_start":8,"line_end":8,"confidence":1.0}]},{"id":"tool_2","type":"Tool","name":"report_generator","importance":"HIGH","raw_prompt":"creates PDF reports","raw_prompt_ref":[{"line_start":9,"line_end":9,"confidence":1.0}]},{"id":"flow_1","type":"Flow","name":"Data Analyst performs Task 1 using pandas_analyzer, then Report Generator performs Task 2 using report_generator","importance":"HIGH","raw_prompt":"Data Analyst performs Task 1 using pandas_analyzer, then Report Generator performs Task 2 using report_generator","raw_prompt_ref":[{"line_start":10,"line_end":10,"confidence":1.0}]}]}', name='EntityExtractionList'), id='call_UGXmGnmv2SmudEEKl7UiIdjL', type='function')], function_call=None, provider_specific_fields={'refusal': None, 'annotations': []}, refusal=None, annotations=[]))], usage=CompletionUsage(completion_tokens=401, prompt_tokens=1753, total_tokens=2154, completion_tokens_details=CompletionTokensDetails(accepted_prediction_tokens=None, audio_tokens=0, reasoning_tokens=0, rejected_prediction_tokens=None), prompt_tokens_details=PromptTokensDetails(audio_tokens=0, cached_tokens=0)), service_tier='default') -2025-07-24 17:19:23,429 - instructor - DEBUG - Parse error: 1 validation error for EntityExtractionList -entities.6.type - Input should be 'Agent', 'Task', 'Tool', 'Input', 'Output' or 'Human' [type=literal_error, input_value='Flow', input_type=str] - For further information visit https://errors.pydantic.dev/2.11/v/literal_error -2025-07-24 17:19:23,429 - instructor - DEBUG - Retrying, attempt: 2 -2025-07-24 17:19:28,688 - instructor - DEBUG - Instructor Raw Response: ModelResponse(id='chatcmpl-Bwsx9Xqy9PJF5HS4HwANNoFefK8Zw', created=1753373963, model='gpt-4o-mini-2024-07-18', object='chat.completion', system_fingerprint=None, choices=[Choices(finish_reason='stop', index=0, message=Message(content=None, role='assistant', tool_calls=[ChatCompletionMessageToolCall(function=Function(arguments='{"entities":[{"id":"agent_1","type":"Agent","name":"Data Analyst","importance":"HIGH","raw_prompt":"responsible for analyzing customer data","raw_prompt_ref":[{"line_start":4,"line_end":4,"confidence":1.0}]},{"id":"agent_2","type":"Agent","name":"Report Generator","importance":"HIGH","raw_prompt":"creates comprehensive reports","raw_prompt_ref":[{"line_start":5,"line_end":5,"confidence":1.0}]},{"id":"task_1","type":"Task","name":"Analyze sales data from Q4 2023","importance":"HIGH","raw_prompt":"Analyze sales data from Q4 2023","raw_prompt_ref":[{"line_start":6,"line_end":6,"confidence":1.0}]},{"id":"task_2","type":"Task","name":"Generate executive summary report","importance":"HIGH","raw_prompt":"Generate executive summary report","raw_prompt_ref":[{"line_start":7,"line_end":7,"confidence":1.0}]},{"id":"tool_1","type":"Tool","name":"pandas_analyzer","importance":"HIGH","raw_prompt":"analyzes CSV data","raw_prompt_ref":[{"line_start":8,"line_end":8,"confidence":1.0}]},{"id":"tool_2","type":"Tool","name":"report_generator","importance":"HIGH","raw_prompt":"creates PDF reports","raw_prompt_ref":[{"line_start":9,"line_end":9,"confidence":1.0}]}]}', name='EntityExtractionList'), id='call_kDTgRpJxnHx8uZC1f5tP0U9y', type='function')], function_call=None, provider_specific_fields={'refusal': None, 'annotations': []}, refusal=None, annotations=[]))], usage=CompletionUsage(completion_tokens=717, prompt_tokens=4010, total_tokens=4727, completion_tokens_details=CompletionTokensDetails(accepted_prediction_tokens=None, audio_tokens=0, reasoning_tokens=0, rejected_prediction_tokens=None), prompt_tokens_details=PromptTokensDetails(audio_tokens=0, cached_tokens=2048)), service_tier='default') -2025-07-24 17:20:20,674 - instructor - DEBUG - Patching `client.chat.completions.create` with mode= -2025-07-24 17:20:20,698 - instructor - DEBUG - Instructor Request: mode.value='tool_call', response_model=, new_kwargs={'messages': [{'role': 'user', 'content': '{\n "entities": [\n {\n "id": "agent_1",\n "type": "Agent",\n "name": "Data Analyst",\n "importance": "HIGH",\n "raw_prompt": "responsible for analyzing customer data",\n "raw_prompt_ref": [\n {\n "line_start": 4,\n "line_end": 4,\n "confidence": 1.0\n }\n ]\n },\n {\n "id": "agent_2",\n "type": "Agent",\n "name": "Report Generator",\n "importance": "HIGH",\n "raw_prompt": "creates comprehensive reports",\n "raw_prompt_ref": [\n {\n "line_start": 5,\n "line_end": 5,\n "confidence": 1.0\n }\n ]\n },\n {\n "id": "task_1",\n "type": "Task",\n "name": "Analyze sales data from Q4 2023",\n "importance": "HIGH",\n "raw_prompt": "Analyze sales data from Q4 2023",\n "raw_prompt_ref": [\n {\n "line_start": 6,\n "line_end": 6,\n "confidence": 1.0\n }\n ]\n },\n {\n "id": "task_2",\n "type": "Task",\n "name": "Generate executive summary report",\n "importance": "HIGH",\n "raw_prompt": "Generate executive summary report",\n "raw_prompt_ref": [\n {\n "line_start": 7,\n "line_end": 7,\n "confidence": 1.0\n }\n ]\n },\n {\n "id": "tool_1",\n "type": "Tool",\n "name": "pandas_analyzer",\n "importance": "HIGH",\n "raw_prompt": "analyzes CSV data",\n "raw_prompt_ref": [\n {\n "line_start": 8,\n "line_end": 8,\n "confidence": 1.0\n }\n ]\n },\n {\n "id": "tool_2",\n "type": "Tool",\n "name": "report_generator",\n "importance": "HIGH",\n "raw_prompt": "creates PDF reports",\n "raw_prompt_ref": [\n {\n "line_start": 9,\n "line_end": 9,\n "confidence": 1.0\n }\n ]\n },\n {\n "id": "flow_1",\n "type": "Flow",\n "name": "Data Analyst performs Task 1 using pandas_analyzer, then Report Generator performs Task 2 using report_generator",\n "importance": "HIGH",\n "raw_prompt": "Data Analyst performs Task 1 using pandas_analyzer, then Report Generator performs Task 2 using report_generator",\n "raw_prompt_ref": [\n {\n "line_start": 10,\n "line_end": 10,\n "confidence": 1.0\n }\n ]\n }\n ],\n "relations": [\n {\n "id": "relation_1",\n "source": "task_1",\n "target": "agent_1",\n "type": "ASSIGNED_TO",\n "importance": "HIGH",\n "interaction_prompt": "",\n "interaction_prompt_ref": []\n },\n {\n "id": "relation_2",\n "source": "agent_1",\n "target": "task_1",\n "type": "PERFORMS",\n "importance": "HIGH",\n "interaction_prompt": "",\n "interaction_prompt_ref": []\n },\n {\n "id": "relation_3",\n "source": "task_1",\n "target": "tool_1",\n "type": "REQUIRED_BY",\n "importance": "HIGH",\n "interaction_prompt": "analyzes CSV data",\n "interaction_prompt_ref": [\n {\n "line_start": 8,\n "line_end": 8,\n "confidence": 1.0\n }\n ]\n },\n {\n "id": "relation_4",\n "source": "agent_1",\n "target": "tool_1",\n "type": "USES",\n "importance": "HIGH",\n "interaction_prompt": "analyzes CSV data",\n "interaction_prompt_ref": [\n {\n "line_start": 8,\n "line_end": 8,\n "confidence": 1.0\n }\n ]\n },\n {\n "id": "relation_5",\n "source": "agent_2",\n "target": "task_2",\n "type": "ASSIGNED_TO",\n "importance": "HIGH",\n "interaction_prompt": "",\n "interaction_prompt_ref": []\n },\n {\n "id": "relation_6",\n "source": "agent_2",\n "target": "task_2",\n "type": "PERFORMS",\n "importance": "HIGH",\n "interaction_prompt": "",\n "interaction_prompt_ref": []\n },\n {\n "id": "relation_7",\n "source": "task_2",\n "target": "tool_2",\n "type": "REQUIRED_BY",\n "importance": "HIGH",\n "interaction_prompt": "creates PDF reports",\n "interaction_prompt_ref": [\n {\n "line_start": 9,\n "line_end": 9,\n "confidence": 1.0\n }\n ]\n },\n {\n "id": "relation_8",\n "source": "agent_2",\n "target": "tool_2",\n "type": "USES",\n "importance": "HIGH",\n "interaction_prompt": "creates PDF reports",\n "interaction_prompt_ref": [\n {\n "line_start": 9,\n "line_end": 9,\n "confidence": 1.0\n }\n ]\n },\n {\n "id": "relation_9",\n "source": "task_1",\n "target": "task_2",\n "type": "NEXT",\n "importance": "HIGH",\n "interaction_prompt": "",\n "interaction_prompt_ref": []\n },\n {\n "id": "relation_10",\n "source": "task_1",\n "target": "output_1",\n "type": "PRODUCES",\n "importance": "HIGH",\n "interaction_prompt": "",\n "interaction_prompt_ref": []\n },\n {\n "id": "relation_11",\n "source": "output_1",\n "target": "human_1",\n "type": "DELIVERS_TO",\n "importance": "HIGH",\n "interaction_prompt": "Output format specification",\n "interaction_prompt_ref": [\n {\n "line_start": 10,\n "line_end": 10,\n "confidence": 1.0\n }\n ]\n }\n ],\n "failures": [\n {\n "id": "failure_1",\n "risk_type": "AGENT_ERROR",\n "description": "Data Analyst might not perform task due to unavailability.",\n "raw_text": "assumed presence of Data Analyst for all tasks.",\n "raw_text_ref": [\n {\n "line_start": 4,\n "line_end": 4,\n "confidence": 1.0\n }\n ],\n "affected_id": "agent_1"\n },\n {\n "id": "failure_2",\n "risk_type": "PLANNING_ERROR",\n "description": "Lack of a buffer for delays can cause schedule overlaps.",\n "raw_text": "tasks dependent on completion.",\n "raw_text_ref": [\n {\n "line_start": 6,\n "line_end": 6,\n "confidence": 1.0\n }\n ],\n "affected_id": "task_1"\n }\n ],\n "system_name": "Sales Data Analysis and Reporting System",\n "system_summary": "This system analyzes sales data from Q4 2023 through the Data Analyst\'s performance of the task using the pandas_analyzer tool. Once the analysis is complete, a summary report is generated by the Report Generator using the report_generator tool, culminating in execution of key data processes in a structured flow."\n}'}], 'model': 'gpt-4o-mini', 'tools': [{'type': 'function', 'function': {'name': 'KnowledgeGraph', 'description': 'Correctly extracted `KnowledgeGraph` with all the required parameters with correct types', 'parameters': {'$defs': {'ContentReference': {'description': 'Reference to content location in the original trace using line numbers and character positions.\nThis allows AI agents to provide position metadata instead of full content, enabling \nefficient mapping back to the original trace while reducing hallucination risks.\n\nCRITICAL FOR LLMs: Line counting accuracy is essential for proper content resolution.\nUse systematic counting methods and verify your line numbers before submission.', 'properties': {'line_start': {'description': 'Starting line number where the content begins (1-based indexing from , ... markers).\n \n ACCURACY REQUIREMENTS FOR LLMs:\n - Count markers systematically from the beginning of the input\n - Use anchor points: find distinctive text first, then count nearby lines\n - Double-check by counting backwards from a known reference point\n - For multi-line content, this should be the FIRST line containing the content\n - In key-value pairs (e.g. "content": "..."), reference the line where the VALUE starts, not the key\n \n COMMON ERRORS TO AVOID:\n - Miscounting due to skipping indented continuation lines\n - Confusing line numbers when content spans multiple markers\n - Using approximate counting instead of precise marker identification\n \n VERIFICATION: Before submitting, locate your chosen line number and confirm it contains the expected content start.', 'title': 'Line Start', 'type': 'integer'}, 'line_end': {'description': 'Ending line number where content ends (1-based indexing from , ... markers).\n \n ACCURACY REQUIREMENTS FOR LLMs:\n - Must be >= line_start (validation will fail otherwise)\n - For single-line content, line_end should equal line_start\n - For multi-line content, find the LAST line containing the content\n - Include indented continuation lines that are part of the same logical content block\n \n VERIFICATION STRATEGY:\n - Count from line_start to ensure proper range\n - Confirm the line_end marker contains the actual end of the content\n - Check that no content continues beyond your specified line_end', 'title': 'Line End', 'type': 'integer'}, 'confidence': {'default': None, 'description': 'Confidence score for the location reference accuracy (0.0 - 1.0).\n \n CONFIDENCE SCORING GUIDE FOR LLMs:\n - 1.0: Verified by counting twice with consistent results, clear content boundaries\n - 0.9: High confidence with single verification, unambiguous content location \n - 0.8: Good confidence but content boundaries somewhat ambiguous\n - 0.7: Moderate confidence, some uncertainty in exact line boundaries\n - 0.6: Lower confidence due to complex content structure or counting difficulty\n - 0.5 or below: Uncertain about accuracy, recommend manual verification\n \n FACTORS AFFECTING CONFIDENCE:\n - Clarity of content boundaries (higher = more confident)\n - Complexity of surrounding text (simpler = more confident) \n - Verification method used (double-counting = more confident)\n - Presence of clear anchor points (more anchors = more confident)', 'maximum': 1.0, 'minimum': 0.0, 'title': 'Confidence', 'type': 'number'}}, 'required': ['line_start', 'line_end'], 'title': 'ContentReference', 'type': 'object'}, 'Entity': {'properties': {'id': {'description': 'Unique identifier for the entity', 'title': 'Id', 'type': 'string'}, 'type': {'description': 'Type of entity defined by prompt type: Agent (system prompt), Task (instruction prompt), Tool (description prompt), Input (input format prompt), Output (output format prompt), Human (optional prompt). The raw_prompt field is the primary distinguishing factor for entity uniqueness and classification.', 'enum': ['Agent', 'Task', 'Tool', 'Input', 'Output', 'Human'], 'title': 'Type', 'type': 'string'}, 'name': {'description': "Name of the entity derived from the prompt content. Names should reflect the specific prompt or specification that defines this entity. For composite entities, use descriptive names that capture the prompt's scope (e.g., 'SQL Query Generation System Prompt', 'Data Analysis Instruction Set').", 'title': 'Name', 'type': 'string'}, 'importance': {'description': 'Importance level of this entity in the system. HIGH: Core agents, critical tasks, essential tools that are central to system function. MEDIUM: Supporting agents, standard tasks, commonly used tools. LOW: Auxiliary entities, simple tasks, rarely used components.', 'enum': ['HIGH', 'MEDIUM', 'LOW'], 'title': 'Importance', 'type': 'string'}, 'raw_prompt': {'default': '', 'description': 'PRIMARY DISTINGUISHING CONTENT: The actual prompt, specification, or instruction that defines this entity. This is the core content that makes each entity unique and should contain: For Agents (system prompts defining role/capabilities), For Tasks (instruction prompts defining objectives), For Tools (description prompts defining functionality), For Inputs (format specifications), For Outputs (format specifications), For Humans (interaction patterns). This field is more important than the name for entity distinction and relationship mapping.', 'title': 'Raw Prompt', 'type': 'string'}, 'raw_prompt_ref': {'description': 'A list of references to the locations of the raw prompt content in the original trace. When provided, this allows mapping back to all exact positions in the trace where this prompt was found.', 'items': {'$ref': '#/$defs/ContentReference'}, 'title': 'Raw Prompt Ref', 'type': 'array'}}, 'required': ['id', 'type', 'name', 'importance'], 'title': 'Entity', 'type': 'object'}, 'Failure': {'description': 'Represents a failure / risk event located via ContentReference.', 'properties': {'id': {'description': 'Unique identifier for the failure event', 'title': 'Id', 'type': 'string'}, 'risk_type': {'description': 'Categorised failure type (predefined list)', 'enum': ['AGENT_ERROR', 'PLANNING_ERROR', 'EXECUTION_ERROR', 'RETRIEVAL_ERROR', 'HALLUCINATION'], 'title': 'Risk Type', 'type': 'string'}, 'description': {'description': 'One-sentence explanation of the failure', 'title': 'Description', 'type': 'string'}, 'raw_text': {'default': '', 'description': 'Exact snippet of trace text that evidences the failure (can be left blank and recovered via raw_text_ref)', 'title': 'Raw Text', 'type': 'string'}, 'raw_text_ref': {'description': 'List of references to every occurrence of the failure evidence in the trace', 'items': {'$ref': '#/$defs/ContentReference'}, 'title': 'Raw Text Ref', 'type': 'array'}, 'affected_id': {'anyOf': [{'type': 'string'}, {'type': 'null'}], 'default': None, 'description': 'ID of related Entity or Relation responsible for or impacted by the failure', 'title': 'Affected Id'}}, 'required': ['risk_type', 'description', 'raw_text_ref'], 'title': 'Failure', 'type': 'object'}, 'Relation': {'properties': {'id': {'description': 'Unique identifier for the relation', 'title': 'Id', 'type': 'string'}, 'source': {'description': 'ID of the source entity', 'title': 'Source', 'type': 'string'}, 'target': {'description': 'ID of the target entity', 'title': 'Target', 'type': 'string'}, 'type': {'description': 'Type of relation (only predefined types are allowed)', 'enum': ['CONSUMED_BY', 'PERFORMS', 'ASSIGNED_TO', 'USES', 'REQUIRED_BY', 'SUBTASK_OF', 'NEXT', 'PRODUCES', 'DELIVERS_TO', 'INTERVENES'], 'title': 'Type', 'type': 'string'}, 'importance': {'description': 'Importance level of this relationship in the system. HIGH: Critical data flows, core agent-task assignments, essential tool usage. MEDIUM: Standard workflows, common interactions, regular data processing. LOW: Auxiliary connections, optional steps, rarely activated relationships.', 'enum': ['HIGH', 'MEDIUM', 'LOW'], 'title': 'Importance', 'type': 'string'}, 'interaction_prompt': {'default': '', 'description': "Actual runtime interaction message/log that shows this relationship occurring during execution. Contains the exact text from the trace where this interaction happened (e.g., 'Agent started task X', 'Calling tool Y with parameters Z', 'User provided feedback: ABC'). This is NOT the static prompt definition but the dynamic interaction evidence.", 'title': 'Interaction Prompt', 'type': 'string'}, 'interaction_prompt_ref': {'description': 'List of references to the locations of interaction prompt content in the original trace. Enables mapping back to all occurrences of the interaction prompt.', 'items': {'$ref': '#/$defs/ContentReference'}, 'title': 'Interaction Prompt Ref', 'type': 'array'}}, 'required': ['source', 'target', 'type', 'importance'], 'title': 'Relation', 'type': 'object'}}, 'properties': {'entities': {'description': 'List of entities in the knowledge graph', 'items': {'$ref': '#/$defs/Entity'}, 'title': 'Entities', 'type': 'array'}, 'relations': {'description': 'List of relations in the knowledge graph', 'items': {'$ref': '#/$defs/Relation'}, 'title': 'Relations', 'type': 'array'}, 'failures': {'description': 'List of detected risk or failure events across the trace', 'items': {'$ref': '#/$defs/Failure'}, 'title': 'Failures', 'type': 'array'}, 'system_name': {'default': '', 'description': 'A concise, descriptive name for the agent system', 'title': 'System Name', 'type': 'string'}, 'system_summary': {'default': '', 'description': "A short 2-3 sentence summary of the agent system's purpose and structure", 'title': 'System Summary', 'type': 'string'}}, 'type': 'object', 'required': ['entities', 'failures', 'relations']}}}], 'tool_choice': {'type': 'function', 'function': {'name': 'KnowledgeGraph'}}} -2025-07-24 17:20:20,699 - instructor - DEBUG - max_retries: 3 -2025-07-24 17:20:20,699 - instructor - DEBUG - Retrying, attempt: 1 -2025-07-24 17:20:25,164 - instructor - DEBUG - Instructor Raw Response: ModelResponse(id='chatcmpl-Bwsy4ouss1y22gkMaiAyxZcBUrSxX', created=1753374020, model='gpt-4o-mini-2024-07-18', object='chat.completion', system_fingerprint=None, choices=[Choices(finish_reason='stop', index=0, message=Message(content=None, role='assistant', tool_calls=[ChatCompletionMessageToolCall(function=Function(arguments='{"entities":[{"id":"agent_1","type":"Agent","name":"Data Analyst","importance":"HIGH","raw_prompt":"responsible for analyzing customer data","raw_prompt_ref":[{"line_start":4,"line_end":4,"confidence":1}]}],"relations":[{"source":"task_1","target":"agent_1","type":"ASSIGNED_TO","importance":"HIGH"},{"source":"agent_1","target":"task_1","type":"PERFORMS","importance":"HIGH"}],"failures":[{"risk_type":"AGENT_ERROR","description":"Data Analyst might not perform task due to unavailability.","raw_text":"assumed presence of Data Analyst for all tasks.","raw_text_ref":[{"line_start":4,"line_end":4,"confidence":1}],"affected_id":"agent_1"}],"system_name":"Sales Data Analysis and Reporting System","system_summary":"This system analyzes sales data from Q4 2023 through the Data Analyst\'s performance of the task using the pandas_analyzer tool. Once the analysis is complete, a summary report is generated by the Report Generator using the report_generator tool, culminating in execution of key data processes in a structured flow."}', name='KnowledgeGraph'), id='call_46UwZ1t6PnunkPLzfiCT7IIf', type='function')], function_call=None, provider_specific_fields={'refusal': None, 'annotations': []}, refusal=None, annotations=[]))], usage=CompletionUsage(completion_tokens=241, prompt_tokens=4751, total_tokens=4992, completion_tokens_details=CompletionTokensDetails(accepted_prediction_tokens=None, audio_tokens=0, reasoning_tokens=0, rejected_prediction_tokens=None), prompt_tokens_details=PromptTokensDetails(audio_tokens=0, cached_tokens=0)), service_tier='default') -2025-07-24 17:20:25,173 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Result Generated (production method) -2025-07-24 17:20:25,173 - agentgraph.extraction.graph_processing.knowledge_graph_processor - WARNING - Failed to resolve content references for window 0: 1 validation error for Entity -type - Input should be 'Agent', 'Task', 'Tool', 'Input', 'Output' or 'Human' [type=literal_error, input_value='Flow', input_type=str] - For further information visit https://errors.pydantic.dev/2.11/v/literal_error -2025-07-24 17:20:25,173 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Completed sub-batch 1/1 -2025-07-24 17:20:25,174 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Preparing window knowledge graphs with metadata -2025-07-24 17:20:25,174 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Merging knowledge graphs... -2025-07-24 17:20:25,174 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Only one knowledge graph generated, skipping merge process -2025-07-24 17:20:25,174 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing complete. Knowledge graph generated with 7 entities and 11 relations -2025-07-24 17:20:25,174 - __main__ - INFO - Processing text 3/3: text_2 -2025-07-24 17:20:25,174 - __main__ - INFO - Processing text text_2 (format: auto) -2025-07-24 17:20:25,174 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 17:20:25,174 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 434 characters -2025-07-24 17:20:25,174 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 17:20:25,174 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 17:20:25,183 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 17:20:25,183 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 17:20:25,183 - agentgraph.input.content_analysis.semantic_analyzer - INFO - Detected agent trace type: unknown (confidence: 0.12) -2025-07-24 17:20:25,184 - agentgraph.input.text_processing.chunking_service - INFO - Applied rule-based line splitting to 1 chunks (max_line_length=800) -2025-07-24 17:20:25,184 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using agent_semantic splitter -2025-07-24 17:20:25,184 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=350000, overlap_size=17500 -2025-07-24 17:20:25,184 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context enhancement complete: 0 total documents -2025-07-24 17:20:25,184 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - SlidingWindowMonitor initialized with model: gpt-4o-mini, method: production -2025-07-24 17:20:25,184 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context documents: 0 provided -2025-07-24 17:20:25,184 - agentgraph.extraction.graph_utilities.knowledge_graph_merger - INFO - KnowledgeGraphMerger initialized with model: gpt-4o-mini -2025-07-24 17:20:25,187 - __main__ - INFO - Extracting knowledge graph from 1 chunks -2025-07-24 17:20:25,187 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks (splitter: agent_semantic, window_size=350000, overlap=0) -2025-07-24 17:20:25,187 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided source trace ID: text_2 -2025-07-24 17:20:25,187 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided processing run ID: batch_text_2 -2025-07-24 17:20:25,187 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks -2025-07-24 17:20:25,187 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using parallel processing with batch size 3 (1 batches) -2025-07-24 17:20:25,187 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 windows in 1 batches -2025-07-24 17:20:25,187 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks in 1 sub-batches (max 1 concurrent per sub-batch) -2025-07-24 17:20:25,187 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Starting sub-batch 1/1 with 1 chunks -2025-07-24 17:20:25,187 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing window 0 -2025-07-24 17:20:25,187 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - 🤖 Starting async_crew extraction for window 0 with method: production using model: gpt-4o-mini -2025-07-24 17:20:25,187 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - No context documents available for window 0, passing empty context string -2025-07-24 17:20:25,187 - agentgraph.methods.production.multi_agent_knowledge_extractor - INFO - Creating agent monitoring crew with model: gpt-4o-mini -2025-07-24 17:21:01,068 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Result Generated (production method) -2025-07-24 17:21:01,068 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 7 lines -2025-07-24 17:21:01,068 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:21:01,068 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:21:01,068 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L6-L6 -2025-07-24 17:21:01,068 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L6-L6 -2025-07-24 17:21:01,068 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 5, end_idx: 6 -2025-07-24 17:21:01,068 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:21:01,068 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:21:01,068 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Tool: email_sender - sends emails to customer... -2025-07-24 17:21:01,068 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Tool: email_sender - sends emails to customer... -2025-07-24 17:21:01,068 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Tool: email_sender - sends emails to customer... -2025-07-24 17:21:01,068 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:21:01,069 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 1 resolution debug: -2025-07-24 17:21:01,069 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:21:01,069 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L6-L6 -2025-07-24 17:21:01,069 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:21:01,069 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Tool: email_sender - sends emails to customer... -2025-07-24 17:21:01,069 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 1: 51 characters -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L3-L3 -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L3-L3 -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 2, end_idx: 3 -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Agent: Customer Support AI... -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Agent: Customer Support AI... -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Agent: Customer Support AI... -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:21:01,069 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 2 resolution debug: -2025-07-24 17:21:01,069 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:21:01,069 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L3-L3 -2025-07-24 17:21:01,069 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:21:01,069 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Agent: Customer Support AI... -2025-07-24 17:21:01,069 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 2: 31 characters -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L7-L7 -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L7-L7 -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 6, end_idx: 7 -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Process: Agent uses order_lookup to find cust... -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Process: Agent uses order_lookup to find cust... -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Process: Agent uses order_lookup to find cust... -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:21:01,069 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 3 resolution debug: -2025-07-24 17:21:01,069 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:21:01,069 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L7-L7 -2025-07-24 17:21:01,069 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:21:01,069 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Process: Agent uses order_lookup to find cust... -2025-07-24 17:21:01,069 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 3: 131 characters -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L8-L8 -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L8-L8 -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 7, end_idx: 8 -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - WARNING - Line range in ContentReference is out of bounds. -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - bounds check failed: start_idx=7, end_idx=8, lines_len=7 -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: EMPTY... -2025-07-24 17:21:01,069 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:21:01,069 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 4 resolution debug: -2025-07-24 17:21:01,069 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:21:01,069 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L8-L8 -2025-07-24 17:21:01,069 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:21:01,069 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: EMPTY... -2025-07-24 17:21:01,070 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 4: 0 characters -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L9-L9 -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L9-L9 -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 8, end_idx: 9 -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - WARNING - Line range in ContentReference is out of bounds. -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - bounds check failed: start_idx=8, end_idx=9, lines_len=7 -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: EMPTY... -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:21:01,070 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 5 resolution debug: -2025-07-24 17:21:01,070 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:21:01,070 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L9-L9 -2025-07-24 17:21:01,070 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:21:01,070 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: EMPTY... -2025-07-24 17:21:01,070 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 5: 0 characters -2025-07-24 17:21:01,070 - agentgraph.reconstruction.content_reference_resolver - INFO - Entity prompt resolution stats: {'total_entities': 5, 'entities_with_refs': 5, 'successful_resolutions': 5, 'failed_resolutions': 0} -2025-07-24 17:21:01,070 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 7 lines -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L3-L3 -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L3-L3 -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 2, end_idx: 3 -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Agent: Customer Support AI... -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Agent: Customer Support AI... -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Agent: Customer Support AI... -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:21:01,070 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation rel1: 31 characters -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L9-L9 -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L9-L9 -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 8, end_idx: 9 -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - WARNING - Line range in ContentReference is out of bounds. -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - bounds check failed: start_idx=8, end_idx=9, lines_len=7 -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: EMPTY... -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:21:01,070 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation rel3: 0 characters -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L9-L9 -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L9-L9 -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 8, end_idx: 9 -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:21:01,070 - agentgraph.input.text_processing.trace_line_processor - WARNING - Line range in ContentReference is out of bounds. -2025-07-24 17:21:01,071 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - bounds check failed: start_idx=8, end_idx=9, lines_len=7 -2025-07-24 17:21:01,071 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: EMPTY... -2025-07-24 17:21:01,071 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:21:01,071 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation rel4: 0 characters -2025-07-24 17:21:01,071 - agentgraph.reconstruction.content_reference_resolver - INFO - Relation prompt resolution stats: {'total_relations': 4, 'relations_with_refs': 3, 'successful_resolutions': 3, 'failed_resolutions': 0} -2025-07-24 17:21:01,071 - agentgraph.reconstruction.content_reference_resolver - INFO - Resolved content references for knowledge graph with 5 entities and 4 relations -2025-07-24 17:21:01,071 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Resolved content references for window 0 -2025-07-24 17:21:01,071 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Completed sub-batch 1/1 -2025-07-24 17:21:01,071 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Preparing window knowledge graphs with metadata -2025-07-24 17:21:01,072 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Merging knowledge graphs... -2025-07-24 17:21:01,072 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Only one knowledge graph generated, skipping merge process -2025-07-24 17:21:01,072 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing complete. Knowledge graph generated with 5 entities and 4 relations -2025-07-24 17:21:01,074 - __main__ - INFO - Results saved to /tmp/test_results.json -2025-07-24 17:22:17,872 - openlit - INFO - Starting openLIT initialization... -2025-07-24 17:22:17,888 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 17:22:18,462 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 17:22:18,508 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 17:22:18,508 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 17:22:18,508 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 17:22:18,842 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 17:22:18,941 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 17:22:18,941 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 17:22:19,495 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 17:22:19,496 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 17:22:20,824 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 17:22:20,826 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 17:22:20,826 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 17:22:20,826 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 17:22:20,826 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 17:22:20,826 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 17:22:20,826 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 17:22:20,826 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 17:22:20,827 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 17:22:20,827 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 17:22:20,827 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 17:22:20,827 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 17:22:20,827 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 17:22:20,827 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 17:22:20,827 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 17:22:20,827 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 17:22:20,827 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 17:22:20,827 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 17:22:20,827 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 17:22:21,931 - __main__ - INFO - Treating JSON object as single trace/text input -2025-07-24 17:22:21,931 - __main__ - INFO - Loaded 1 texts from /tmp/test_langsmith.json -2025-07-24 17:22:21,931 - agentgraph.input.text_processing.chunking_service - INFO - ChunkingService initialized with batch_size=3, model=gpt-4o-mini -2025-07-24 17:22:21,931 - __main__ - INFO - Initialized BatchKGExtractor: -2025-07-24 17:22:21,931 - __main__ - INFO - - Model: gpt-4o-mini, Method: production -2025-07-24 17:22:21,931 - __main__ - INFO - - Preprocessing: False -2025-07-24 17:22:21,931 - __main__ - INFO - - Line numbers: False -2025-07-24 17:22:21,931 - __main__ - INFO - Processing batch of 1 texts -2025-07-24 17:22:21,931 - __main__ - INFO - Processing text 1/1: text_0 -2025-07-24 17:22:21,931 - __main__ - INFO - Processing text text_0 (format: auto) -2025-07-24 17:22:21,931 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 17:22:21,931 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 238 characters -2025-07-24 17:22:21,931 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 17:22:21,931 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 17:22:21,937 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 17:22:21,937 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 17:22:21,939 - agentgraph.input.content_analysis.semantic_analyzer - INFO - Detected agent trace type: unknown (confidence: 0.14) -2025-07-24 17:22:21,939 - agentgraph.input.text_processing.chunking_service - INFO - Applied rule-based line splitting to 1 chunks (max_line_length=800) -2025-07-24 17:22:21,940 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using agent_semantic splitter -2025-07-24 17:22:21,940 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=350000, overlap_size=17500 -2025-07-24 17:22:21,941 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context enhancement complete: 0 total documents -2025-07-24 17:22:21,941 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - SlidingWindowMonitor initialized with model: gpt-4o-mini, method: production -2025-07-24 17:22:21,941 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context documents: 0 provided -2025-07-24 17:22:21,941 - agentgraph.extraction.graph_utilities.knowledge_graph_merger - INFO - KnowledgeGraphMerger initialized with model: gpt-4o-mini -2025-07-24 17:22:21,942 - __main__ - INFO - Extracting knowledge graph from 1 chunks -2025-07-24 17:22:21,942 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks (splitter: agent_semantic, window_size=350000, overlap=0) -2025-07-24 17:22:21,942 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided source trace ID: text_0 -2025-07-24 17:22:21,942 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided processing run ID: batch_text_0 -2025-07-24 17:22:21,943 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks -2025-07-24 17:22:21,943 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using parallel processing with batch size 3 (1 batches) -2025-07-24 17:22:21,943 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 windows in 1 batches -2025-07-24 17:22:21,952 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks in 1 sub-batches (max 1 concurrent per sub-batch) -2025-07-24 17:22:21,952 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Starting sub-batch 1/1 with 1 chunks -2025-07-24 17:22:21,953 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing window 0 -2025-07-24 17:22:21,953 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - 🤖 Starting async_crew extraction for window 0 with method: production using model: gpt-4o-mini -2025-07-24 17:22:21,953 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - No context documents available for window 0, passing empty context string -2025-07-24 17:22:21,953 - agentgraph.methods.production.multi_agent_knowledge_extractor - INFO - Creating agent monitoring crew with model: gpt-4o-mini -2025-07-24 17:22:49,928 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Result Generated (production method) -2025-07-24 17:22:49,928 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 16 lines -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 16 -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L4-L4 -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L4-L4 -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 3, end_idx: 4 -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 16 -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: {... -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: {... -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: {... -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:22:49,929 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity agent_1 resolution debug: -2025-07-24 17:22:49,929 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:22:49,929 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L4-L4 -2025-07-24 17:22:49,929 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:22:49,929 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: {... -2025-07-24 17:22:49,929 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity agent_1: 10 characters -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 16 -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L5-L5 -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L5-L5 -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 4, end_idx: 5 -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 16 -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: "id": "run1",... -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: "id": "run1",... -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: "id": "run1",... -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:22:49,929 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity task_1 resolution debug: -2025-07-24 17:22:49,929 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:22:49,929 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L5-L5 -2025-07-24 17:22:49,929 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:22:49,929 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: "id": "run1",... -2025-07-24 17:22:49,929 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity task_1: 24 characters -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 16 -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L7-L7 -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L7-L7 -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 6, end_idx: 7 -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 16 -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:22:49,929 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: "run_type": "llm",... -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: "run_type": "llm",... -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: "run_type": "llm",... -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:22:49,930 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity input_1 resolution debug: -2025-07-24 17:22:49,930 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:22:49,930 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L7-L7 -2025-07-24 17:22:49,930 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:22:49,930 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: "run_type": "llm",... -2025-07-24 17:22:49,930 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity input_1: 29 characters -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 16 -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L10-L10 -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L10-L10 -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 9, end_idx: 10 -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 16 -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: },... -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: },... -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: },... -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:22:49,930 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity output_1 resolution debug: -2025-07-24 17:22:49,930 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:22:49,930 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L10-L10 -2025-07-24 17:22:49,930 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:22:49,930 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: },... -2025-07-24 17:22:49,930 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity output_1: 14 characters -2025-07-24 17:22:49,930 - agentgraph.reconstruction.content_reference_resolver - INFO - Entity prompt resolution stats: {'total_entities': 4, 'entities_with_refs': 4, 'successful_resolutions': 4, 'failed_resolutions': 0} -2025-07-24 17:22:49,930 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 16 lines -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 16 -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L7-L7 -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L7-L7 -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 6, end_idx: 7 -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 16 -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: "run_type": "llm",... -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: "run_type": "llm",... -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: "run_type": "llm",... -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:22:49,930 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation relation_1: 29 characters -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 16 -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L10-L10 -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L10-L10 -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 9, end_idx: 10 -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 16 -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: },... -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: },... -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: },... -2025-07-24 17:22:49,930 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:22:49,930 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation relation_4: 14 characters -2025-07-24 17:22:49,930 - agentgraph.reconstruction.content_reference_resolver - INFO - Relation prompt resolution stats: {'total_relations': 4, 'relations_with_refs': 2, 'successful_resolutions': 2, 'failed_resolutions': 0} -2025-07-24 17:22:49,930 - agentgraph.reconstruction.content_reference_resolver - INFO - Resolved content references for knowledge graph with 4 entities and 4 relations -2025-07-24 17:22:49,930 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Resolved content references for window 0 -2025-07-24 17:22:49,931 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Completed sub-batch 1/1 -2025-07-24 17:22:49,931 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Preparing window knowledge graphs with metadata -2025-07-24 17:22:49,932 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Merging knowledge graphs... -2025-07-24 17:22:49,932 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Only one knowledge graph generated, skipping merge process -2025-07-24 17:22:49,932 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing complete. Knowledge graph generated with 4 entities and 4 relations -2025-07-24 17:22:49,933 - __main__ - INFO - Results saved to /tmp/test_output.json -2025-07-24 17:23:10,280 - openlit - INFO - Starting openLIT initialization... -2025-07-24 17:23:10,295 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 17:23:10,856 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 17:23:10,901 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 17:23:10,902 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 17:23:10,902 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 17:23:11,227 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 17:23:11,339 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 17:23:11,339 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 17:23:11,879 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 17:23:11,880 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 17:23:13,118 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 17:23:13,121 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 17:23:13,121 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 17:23:13,121 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 17:23:13,121 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 17:23:13,122 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 17:23:13,122 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 17:23:13,122 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 17:23:13,122 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 17:23:13,122 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 17:23:13,122 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 17:23:13,122 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 17:23:13,122 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 17:23:13,122 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 17:23:13,122 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 17:23:13,122 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 17:23:13,123 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 17:23:13,123 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 17:23:13,123 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 17:23:14,040 - __main__ - INFO - Treating JSON object as single trace/text input -2025-07-24 17:23:14,040 - __main__ - INFO - Loaded 1 texts from /tmp/test_langsmith.json -2025-07-24 17:23:14,040 - agentgraph.input.text_processing.chunking_service - INFO - ChunkingService initialized with batch_size=3, model=gpt-4o-mini -2025-07-24 17:23:14,040 - __main__ - INFO - Initialized BatchKGExtractor: -2025-07-24 17:23:14,040 - __main__ - INFO - - Model: gpt-4o-mini, Method: production -2025-07-24 17:23:14,040 - __main__ - INFO - - Preprocessing: True -2025-07-24 17:23:14,040 - __main__ - INFO - - Line numbers: True -2025-07-24 17:23:14,040 - __main__ - INFO - Processing batch of 1 texts -2025-07-24 17:23:14,040 - __main__ - INFO - Processing text 1/1: text_0 -2025-07-24 17:23:14,040 - __main__ - WARNING - Trace preprocessing failed: 31 validation errors for LangSmithTrace -trace_name - Field required [type=missing, input_value={'trace_id': 'test_id', '...sponse': 'Hi there!'}}]}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -project_name - Field required [type=missing, input_value={'trace_id': 'test_id', '...sponse': 'Hi there!'}}]}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -export_time - Field required [type=missing, input_value={'trace_id': 'test_id', '...sponse': 'Hi there!'}}]}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -total_runs - Field required [type=missing, input_value={'trace_id': 'test_id', '...sponse': 'Hi there!'}}]}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs.0.start_time - Field required [type=missing, input_value={'id': 'run1', 'name': 't...response': 'Hi there!'}}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs.0.end_time - Field required [type=missing, input_value={'id': 'run1', 'name': 't...response': 'Hi there!'}}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs.0.extra - Field required [type=missing, input_value={'id': 'run1', 'name': 't...response': 'Hi there!'}}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs.0.error - Field required [type=missing, input_value={'id': 'run1', 'name': 't...response': 'Hi there!'}}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs.0.serialized - Field required [type=missing, input_value={'id': 'run1', 'name': 't...response': 'Hi there!'}}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs.0.events - Field required [type=missing, input_value={'id': 'run1', 'name': 't...response': 'Hi there!'}}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs.0.parent_run_id - Field required [type=missing, input_value={'id': 'run1', 'name': 't...response': 'Hi there!'}}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs.0.tags - Field required [type=missing, input_value={'id': 'run1', 'name': 't...response': 'Hi there!'}}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs.0.attachments - Field required [type=missing, input_value={'id': 'run1', 'name': 't...response': 'Hi there!'}}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs.0.session_id - Field required [type=missing, input_value={'id': 'run1', 'name': 't...response': 'Hi there!'}}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs.0.child_run_ids - Field required [type=missing, input_value={'id': 'run1', 'name': 't...response': 'Hi there!'}}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs.0.child_runs - Field required [type=missing, input_value={'id': 'run1', 'name': 't...response': 'Hi there!'}}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs.0.feedback_stats - Field required [type=missing, input_value={'id': 'run1', 'name': 't...response': 'Hi there!'}}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs.0.app_path - Field required [type=missing, input_value={'id': 'run1', 'name': 't...response': 'Hi there!'}}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs.0.manifest_id - Field required [type=missing, input_value={'id': 'run1', 'name': 't...response': 'Hi there!'}}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs.0.status - Field required [type=missing, input_value={'id': 'run1', 'name': 't...response': 'Hi there!'}}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs.0.prompt_tokens - Field required [type=missing, input_value={'id': 'run1', 'name': 't...response': 'Hi there!'}}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs.0.completion_tokens - Field required [type=missing, input_value={'id': 'run1', 'name': 't...response': 'Hi there!'}}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs.0.total_tokens - Field required [type=missing, input_value={'id': 'run1', 'name': 't...response': 'Hi there!'}}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs.0.first_token_time - Field required [type=missing, input_value={'id': 'run1', 'name': 't...response': 'Hi there!'}}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs.0.total_cost - Field required [type=missing, input_value={'id': 'run1', 'name': 't...response': 'Hi there!'}}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs.0.prompt_cost - Field required [type=missing, input_value={'id': 'run1', 'name': 't...response': 'Hi there!'}}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs.0.completion_cost - Field required [type=missing, input_value={'id': 'run1', 'name': 't...response': 'Hi there!'}}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs.0.parent_run_ids - Field required [type=missing, input_value={'id': 'run1', 'name': 't...response': 'Hi there!'}}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs.0.trace_id - Field required [type=missing, input_value={'id': 'run1', 'name': 't...response': 'Hi there!'}}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs.0.dotted_order - Field required [type=missing, input_value={'id': 'run1', 'name': 't...response': 'Hi there!'}}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs.0.in_dataset - Field required [type=missing, input_value={'id': 'run1', 'name': 't...response': 'Hi there!'}}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing, using original text -2025-07-24 17:23:14,041 - agentgraph.input.text_processing.trace_line_processor - INFO - Added line numbers to 16 lines, starting from line 1 -2025-07-24 17:23:14,041 - __main__ - INFO - Added line numbers: 16 lines -2025-07-24 17:23:14,041 - __main__ - INFO - Processing text text_0 (format: langsmith) -2025-07-24 17:23:14,041 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 17:23:14,041 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 325 characters -2025-07-24 17:23:14,041 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 17:23:14,041 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 17:23:14,046 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 17:23:14,046 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 17:23:14,049 - agentgraph.input.content_analysis.semantic_analyzer - INFO - Detected agent trace type: unknown (confidence: 0.00) -2025-07-24 17:23:14,049 - agentgraph.input.text_processing.chunking_service - INFO - Applied rule-based line splitting to 1 chunks (max_line_length=800) -2025-07-24 17:23:14,050 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using agent_semantic splitter -2025-07-24 17:23:14,050 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=350000, overlap_size=17500 -2025-07-24 17:23:14,050 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context enhancement complete: 0 total documents -2025-07-24 17:23:14,050 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - SlidingWindowMonitor initialized with model: gpt-4o-mini, method: production -2025-07-24 17:23:14,050 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context documents: 0 provided -2025-07-24 17:23:14,050 - agentgraph.extraction.graph_utilities.knowledge_graph_merger - INFO - KnowledgeGraphMerger initialized with model: gpt-4o-mini -2025-07-24 17:23:14,052 - __main__ - INFO - Extracting knowledge graph from 1 chunks -2025-07-24 17:23:14,052 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks (splitter: agent_semantic, window_size=350000, overlap=0) -2025-07-24 17:23:14,052 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided source trace ID: text_0 -2025-07-24 17:23:14,052 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided processing run ID: batch_text_0 -2025-07-24 17:23:14,052 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks -2025-07-24 17:23:14,052 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using parallel processing with batch size 3 (1 batches) -2025-07-24 17:23:14,052 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 windows in 1 batches -2025-07-24 17:23:14,063 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks in 1 sub-batches (max 1 concurrent per sub-batch) -2025-07-24 17:23:14,063 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Starting sub-batch 1/1 with 1 chunks -2025-07-24 17:23:14,063 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing window 0 -2025-07-24 17:23:14,063 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - 🤖 Starting async_crew extraction for window 0 with method: production using model: gpt-4o-mini -2025-07-24 17:23:14,063 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - No context documents available for window 0, passing empty context string -2025-07-24 17:23:14,063 - agentgraph.methods.production.multi_agent_knowledge_extractor - INFO - Creating agent monitoring crew with model: gpt-4o-mini -2025-07-24 17:23:39,556 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Result Generated (production method) -2025-07-24 17:23:39,556 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 16 lines -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 16 -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L7-L7 -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L7-L7 -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 6, end_idx: 7 -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 16 -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: "run_type": "llm",... -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: "run_type": "llm",... -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: "run_type": "llm",... -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:23:39,557 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 1 resolution debug: -2025-07-24 17:23:39,557 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:23:39,557 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L7-L7 -2025-07-24 17:23:39,557 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:23:39,557 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: "run_type": "llm",... -2025-07-24 17:23:39,557 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 1: 34 characters -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 16 -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L6-L6 -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L6-L6 -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 5, end_idx: 6 -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 16 -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: "name": "test_run",... -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: "name": "test_run",... -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: "name": "test_run",... -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:23:39,557 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 2 resolution debug: -2025-07-24 17:23:39,557 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:23:39,557 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L6-L6 -2025-07-24 17:23:39,557 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:23:39,557 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: "name": "test_run",... -2025-07-24 17:23:39,557 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 2: 35 characters -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 16 -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L9-L9 -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L9-L9 -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 8, end_idx: 9 -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 16 -2025-07-24 17:23:39,557 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:23:39,558 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: "query": "Hello"... -2025-07-24 17:23:39,558 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: "query": "Hello"... -2025-07-24 17:23:39,558 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: "query": "Hello"... -2025-07-24 17:23:39,558 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:23:39,558 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 3 resolution debug: -2025-07-24 17:23:39,558 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:23:39,558 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L9-L9 -2025-07-24 17:23:39,558 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:23:39,558 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: "query": "Hello"... -2025-07-24 17:23:39,558 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 3: 34 characters -2025-07-24 17:23:39,558 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:23:39,558 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 16 -2025-07-24 17:23:39,558 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L12-L12 -2025-07-24 17:23:39,558 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L12-L12 -2025-07-24 17:23:39,558 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 11, end_idx: 12 -2025-07-24 17:23:39,558 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 16 -2025-07-24 17:23:39,558 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:23:39,558 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: "response": "Hi there!"... -2025-07-24 17:23:39,558 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: "response": "Hi there!"... -2025-07-24 17:23:39,558 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: "response": "Hi there!"... -2025-07-24 17:23:39,558 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:23:39,558 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 4 resolution debug: -2025-07-24 17:23:39,558 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:23:39,558 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L12-L12 -2025-07-24 17:23:39,558 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:23:39,558 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: "response": "Hi there!"... -2025-07-24 17:23:39,558 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 4: 43 characters -2025-07-24 17:23:39,558 - agentgraph.reconstruction.content_reference_resolver - INFO - Entity prompt resolution stats: {'total_entities': 4, 'entities_with_refs': 4, 'successful_resolutions': 4, 'failed_resolutions': 0} -2025-07-24 17:23:39,558 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 16 lines -2025-07-24 17:23:39,558 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:23:39,558 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 16 -2025-07-24 17:23:39,558 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L9-L9 -2025-07-24 17:23:39,558 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L9-L9 -2025-07-24 17:23:39,558 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 8, end_idx: 9 -2025-07-24 17:23:39,558 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 16 -2025-07-24 17:23:39,558 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:23:39,558 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: "query": "Hello"... -2025-07-24 17:23:39,558 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: "query": "Hello"... -2025-07-24 17:23:39,558 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: "query": "Hello"... -2025-07-24 17:23:39,558 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:23:39,558 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation relation_1: 34 characters -2025-07-24 17:23:39,558 - agentgraph.reconstruction.content_reference_resolver - INFO - Relation prompt resolution stats: {'total_relations': 3, 'relations_with_refs': 1, 'successful_resolutions': 1, 'failed_resolutions': 0} -2025-07-24 17:23:39,558 - agentgraph.reconstruction.content_reference_resolver - INFO - Resolved content references for knowledge graph with 4 entities and 3 relations -2025-07-24 17:23:39,559 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Resolved content references for window 0 -2025-07-24 17:23:39,559 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Completed sub-batch 1/1 -2025-07-24 17:23:39,559 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Preparing window knowledge graphs with metadata -2025-07-24 17:23:39,560 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Merging knowledge graphs... -2025-07-24 17:23:39,560 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Only one knowledge graph generated, skipping merge process -2025-07-24 17:23:39,560 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing complete. Knowledge graph generated with 4 entities and 3 relations -2025-07-24 17:23:39,561 - __main__ - INFO - Results saved to /tmp/test_preprocessed.json -2025-07-24 17:23:58,773 - openlit - INFO - Starting openLIT initialization... -2025-07-24 17:23:58,787 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 17:23:59,317 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 17:23:59,363 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 17:23:59,363 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 17:23:59,364 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 17:23:59,683 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 17:23:59,784 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 17:23:59,784 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 17:24:00,311 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 17:24:00,311 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 17:24:01,529 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 17:24:01,531 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 17:24:01,531 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 17:24:01,532 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 17:24:01,532 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 17:24:01,532 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 17:24:01,532 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 17:24:01,532 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 17:24:01,532 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 17:24:01,532 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 17:24:01,532 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 17:24:01,532 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 17:24:01,532 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 17:24:01,532 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 17:24:01,533 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 17:24:01,533 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 17:24:01,533 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 17:24:01,533 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 17:24:01,533 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 17:24:02,434 - __main__ - INFO - Treating file as plain text input -2025-07-24 17:24:02,434 - __main__ - INFO - Loaded 1 texts from /tmp/plain_text.txt -2025-07-24 17:24:02,434 - agentgraph.input.text_processing.chunking_service - INFO - ChunkingService initialized with batch_size=3, model=gpt-4o-mini -2025-07-24 17:24:02,434 - __main__ - INFO - Initialized BatchKGExtractor: -2025-07-24 17:24:02,434 - __main__ - INFO - - Model: gpt-4o-mini, Method: production -2025-07-24 17:24:02,434 - __main__ - INFO - - Preprocessing: False -2025-07-24 17:24:02,434 - __main__ - INFO - - Line numbers: False -2025-07-24 17:24:02,434 - __main__ - INFO - Processing batch of 1 texts -2025-07-24 17:24:02,434 - __main__ - INFO - Processing text 1/1: text_0 -2025-07-24 17:24:02,434 - __main__ - INFO - Processing text text_0 (format: auto) -2025-07-24 17:24:02,434 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 17:24:02,434 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 109 characters -2025-07-24 17:24:02,434 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 17:24:02,434 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 17:24:02,439 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 17:24:02,440 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 17:24:02,442 - agentgraph.input.content_analysis.semantic_analyzer - INFO - Detected agent trace type: unknown (confidence: 0.00) -2025-07-24 17:24:02,442 - agentgraph.input.text_processing.chunking_service - INFO - Applied rule-based line splitting to 1 chunks (max_line_length=800) -2025-07-24 17:24:02,443 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using agent_semantic splitter -2025-07-24 17:24:02,443 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=350000, overlap_size=17500 -2025-07-24 17:24:02,443 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context enhancement complete: 0 total documents -2025-07-24 17:24:02,443 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - SlidingWindowMonitor initialized with model: gpt-4o-mini, method: production -2025-07-24 17:24:02,443 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context documents: 0 provided -2025-07-24 17:24:02,443 - agentgraph.extraction.graph_utilities.knowledge_graph_merger - INFO - KnowledgeGraphMerger initialized with model: gpt-4o-mini -2025-07-24 17:24:02,445 - __main__ - INFO - Extracting knowledge graph from 1 chunks -2025-07-24 17:24:02,445 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks (splitter: agent_semantic_fallback, window_size=350000, overlap=0) -2025-07-24 17:24:02,445 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided source trace ID: text_0 -2025-07-24 17:24:02,445 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided processing run ID: batch_text_0 -2025-07-24 17:24:02,445 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks -2025-07-24 17:24:02,445 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using parallel processing with batch size 3 (1 batches) -2025-07-24 17:24:02,445 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 windows in 1 batches -2025-07-24 17:24:02,457 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks in 1 sub-batches (max 1 concurrent per sub-batch) -2025-07-24 17:24:02,457 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Starting sub-batch 1/1 with 1 chunks -2025-07-24 17:24:02,457 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing window 0 -2025-07-24 17:24:02,457 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - 🤖 Starting async_crew extraction for window 0 with method: production using model: gpt-4o-mini -2025-07-24 17:24:02,457 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - No context documents available for window 0, passing empty context string -2025-07-24 17:24:02,457 - agentgraph.methods.production.multi_agent_knowledge_extractor - INFO - Creating agent monitoring crew with model: gpt-4o-mini -2025-07-24 17:24:50,142 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Result Generated (production method) -2025-07-24 17:24:50,143 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 2 lines -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 2 -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L1-L1 -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L1-L1 -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 0, end_idx: 1 -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 2 -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:24:50,143 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 1 resolution debug: -2025-07-24 17:24:50,143 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:24:50,143 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L1-L1 -2025-07-24 17:24:50,143 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:24:50,143 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,143 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 1: 113 characters -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 2 -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L1-L1 -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L1-L1 -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 0, end_idx: 1 -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 2 -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:24:50,143 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 2 resolution debug: -2025-07-24 17:24:50,143 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:24:50,143 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L1-L1 -2025-07-24 17:24:50,143 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:24:50,143 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,143 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 2: 113 characters -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 2 -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L1-L1 -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L1-L1 -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 0, end_idx: 1 -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 2 -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,143 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:24:50,144 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 3 resolution debug: -2025-07-24 17:24:50,144 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:24:50,144 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L1-L1 -2025-07-24 17:24:50,144 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:24:50,144 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,144 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 3: 113 characters -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 2 -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L1-L1 -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L1-L1 -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 0, end_idx: 1 -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 2 -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:24:50,144 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 4 resolution debug: -2025-07-24 17:24:50,144 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:24:50,144 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L1-L1 -2025-07-24 17:24:50,144 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:24:50,144 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,144 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 4: 113 characters -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 2 -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L1-L1 -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L1-L1 -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 0, end_idx: 1 -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 2 -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:24:50,144 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity 5 resolution debug: -2025-07-24 17:24:50,144 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:24:50,144 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L1-L1 -2025-07-24 17:24:50,144 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:24:50,144 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,144 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity 5: 113 characters -2025-07-24 17:24:50,144 - agentgraph.reconstruction.content_reference_resolver - INFO - Entity prompt resolution stats: {'total_entities': 5, 'entities_with_refs': 5, 'successful_resolutions': 5, 'failed_resolutions': 0} -2025-07-24 17:24:50,144 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 2 lines -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 2 -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L1-L1 -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L1-L1 -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 0, end_idx: 1 -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 2 -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:24:50,144 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation relation_1: 113 characters -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 2 -2025-07-24 17:24:50,144 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L1-L1 -2025-07-24 17:24:50,145 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L1-L1 -2025-07-24 17:24:50,145 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 0, end_idx: 1 -2025-07-24 17:24:50,145 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 2 -2025-07-24 17:24:50,145 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:24:50,145 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,145 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,145 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,145 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:24:50,145 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation relation_3: 113 characters -2025-07-24 17:24:50,145 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:24:50,145 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 2 -2025-07-24 17:24:50,145 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L1-L1 -2025-07-24 17:24:50,145 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L1-L1 -2025-07-24 17:24:50,145 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 0, end_idx: 1 -2025-07-24 17:24:50,145 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 2 -2025-07-24 17:24:50,145 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:24:50,145 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,145 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,145 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: This is a simple agent trace: User asks about... -2025-07-24 17:24:50,145 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:24:50,145 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation relation_6: 113 characters -2025-07-24 17:24:50,145 - agentgraph.reconstruction.content_reference_resolver - INFO - Relation prompt resolution stats: {'total_relations': 6, 'relations_with_refs': 3, 'successful_resolutions': 3, 'failed_resolutions': 0} -2025-07-24 17:24:50,145 - agentgraph.reconstruction.content_reference_resolver - INFO - Resolved content references for knowledge graph with 5 entities and 6 relations -2025-07-24 17:24:50,145 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Resolved content references for window 0 -2025-07-24 17:24:50,145 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Completed sub-batch 1/1 -2025-07-24 17:24:50,146 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Preparing window knowledge graphs with metadata -2025-07-24 17:24:50,147 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Merging knowledge graphs... -2025-07-24 17:24:50,147 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Only one knowledge graph generated, skipping merge process -2025-07-24 17:24:50,147 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing complete. Knowledge graph generated with 5 entities and 6 relations -2025-07-24 17:24:50,148 - __main__ - INFO - Results saved to /tmp/test_plaintext.json -2025-07-24 17:26:43,677 - openlit - INFO - Starting openLIT initialization... -2025-07-24 17:26:43,694 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 17:26:44,340 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 17:26:44,386 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 17:26:44,386 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 17:26:44,387 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 17:26:44,720 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 17:26:44,827 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 17:26:44,827 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 17:26:45,385 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 17:26:45,386 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 17:26:46,643 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 17:26:46,645 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 17:26:46,645 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 17:26:46,645 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 17:26:46,645 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 17:26:46,645 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 17:26:46,645 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 17:26:46,646 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 17:26:46,646 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 17:26:46,646 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 17:26:46,646 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 17:26:46,646 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 17:26:46,646 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 17:26:46,646 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 17:26:46,646 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 17:26:46,646 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 17:26:46,646 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 17:26:46,646 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 17:26:46,646 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 17:26:47,607 - __main__ - INFO - Treating JSON object as single trace/text input -2025-07-24 17:26:47,607 - __main__ - INFO - Loaded 1 texts from /tmp/test_langsmith.json -2025-07-24 17:26:47,607 - __main__ - INFO - Auto-enabling line numbers for trace preprocessing -2025-07-24 17:26:47,607 - agentgraph.input.text_processing.chunking_service - INFO - ChunkingService initialized with batch_size=3, model=gpt-4o-mini -2025-07-24 17:26:47,607 - __main__ - INFO - Initialized BatchKGExtractor: -2025-07-24 17:26:47,607 - __main__ - INFO - - Model: gpt-4o-mini, Method: production -2025-07-24 17:26:47,607 - __main__ - INFO - - Preprocessing: True -2025-07-24 17:26:47,607 - __main__ - INFO - - Line numbers: True -2025-07-24 17:26:47,607 - __main__ - INFO - Processing batch of 1 texts -2025-07-24 17:26:47,607 - __main__ - INFO - Processing text 1/1: text_0 -2025-07-24 17:26:47,607 - __main__ - INFO - Trace preprocessing: 238 → 238 chars -2025-07-24 17:26:47,607 - agentgraph.input.text_processing.trace_line_processor - INFO - Added line numbers to 16 lines, starting from line 1 -2025-07-24 17:26:47,607 - __main__ - INFO - Added line numbers: 16 lines -2025-07-24 17:26:47,607 - __main__ - INFO - Processing text text_0 (format: auto) -2025-07-24 17:26:47,607 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 17:26:47,607 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 325 characters -2025-07-24 17:26:47,607 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 17:26:47,607 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 17:26:47,613 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 17:26:47,613 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 17:26:47,615 - agentgraph.input.content_analysis.semantic_analyzer - INFO - Detected agent trace type: unknown (confidence: 0.00) -2025-07-24 17:26:47,616 - agentgraph.input.text_processing.chunking_service - INFO - Applied rule-based line splitting to 1 chunks (max_line_length=800) -2025-07-24 17:26:47,617 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using agent_semantic splitter -2025-07-24 17:26:47,617 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=350000, overlap_size=17500 -2025-07-24 17:26:47,617 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context enhancement complete: 0 total documents -2025-07-24 17:26:47,617 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - SlidingWindowMonitor initialized with model: gpt-4o-mini, method: production -2025-07-24 17:26:47,618 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context documents: 0 provided -2025-07-24 17:26:47,618 - agentgraph.extraction.graph_utilities.knowledge_graph_merger - INFO - KnowledgeGraphMerger initialized with model: gpt-4o-mini -2025-07-24 17:26:47,619 - __main__ - INFO - Extracting knowledge graph from 1 chunks -2025-07-24 17:26:47,619 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks (splitter: agent_semantic, window_size=350000, overlap=0) -2025-07-24 17:26:47,619 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided source trace ID: text_0 -2025-07-24 17:26:47,619 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided processing run ID: batch_text_0 -2025-07-24 17:26:47,619 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks -2025-07-24 17:26:47,619 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using parallel processing with batch size 3 (1 batches) -2025-07-24 17:26:47,619 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 windows in 1 batches -2025-07-24 17:26:47,629 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks in 1 sub-batches (max 1 concurrent per sub-batch) -2025-07-24 17:26:47,630 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Starting sub-batch 1/1 with 1 chunks -2025-07-24 17:26:47,630 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing window 0 -2025-07-24 17:26:47,630 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - 🤖 Starting async_crew extraction for window 0 with method: production using model: gpt-4o-mini -2025-07-24 17:26:47,630 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - No context documents available for window 0, passing empty context string -2025-07-24 17:26:47,630 - agentgraph.methods.production.multi_agent_knowledge_extractor - INFO - Creating agent monitoring crew with model: gpt-4o-mini -2025-07-24 17:27:11,408 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Result Generated (production method) -2025-07-24 17:27:11,409 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 16 lines -2025-07-24 17:27:11,409 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:27:11,409 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 16 -2025-07-24 17:27:11,409 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L6-L7 -2025-07-24 17:27:11,409 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L6-L7 -2025-07-24 17:27:11,409 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 5, end_idx: 7 -2025-07-24 17:27:11,409 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 16 -2025-07-24 17:27:11,409 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 2 -2025-07-24 17:27:11,409 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: "name": "test_run",... -2025-07-24 17:27:11,409 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[1]: "run_type": "llm",... -2025-07-24 17:27:11,409 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: "name": "test_run",\n ... -2025-07-24 17:27:11,409 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: "name": "test_run",\n ... -2025-07-24 17:27:11,409 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:27:11,409 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity agent_run1 resolution debug: -2025-07-24 17:27:11,409 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:27:11,409 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L6-L7 -2025-07-24 17:27:11,409 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:27:11,409 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: "name": "test_run",\n ... -2025-07-24 17:27:11,409 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity agent_run1: 70 characters -2025-07-24 17:27:11,409 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:27:11,409 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 16 -2025-07-24 17:27:11,409 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L6-L7 -2025-07-24 17:27:11,409 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L6-L7 -2025-07-24 17:27:11,409 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 5, end_idx: 7 -2025-07-24 17:27:11,409 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 16 -2025-07-24 17:27:11,409 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 2 -2025-07-24 17:27:11,409 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: "name": "test_run",... -2025-07-24 17:27:11,409 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[1]: "run_type": "llm",... -2025-07-24 17:27:11,409 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: "name": "test_run",\n ... -2025-07-24 17:27:11,409 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: "name": "test_run",\n ... -2025-07-24 17:27:11,409 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:27:11,409 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity task_run1 resolution debug: -2025-07-24 17:27:11,409 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:27:11,409 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L6-L7 -2025-07-24 17:27:11,409 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:27:11,410 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: "name": "test_run",\n ... -2025-07-24 17:27:11,410 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity task_run1: 70 characters -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 16 -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L9-L9 -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L9-L9 -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 8, end_idx: 9 -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 16 -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: "query": "Hello"... -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: "query": "Hello"... -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: "query": "Hello"... -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:27:11,410 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity input_run1 resolution debug: -2025-07-24 17:27:11,410 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:27:11,410 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L9-L9 -2025-07-24 17:27:11,410 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:27:11,410 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: "query": "Hello"... -2025-07-24 17:27:11,410 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity input_run1: 34 characters -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 16 -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L12-L12 -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L12-L12 -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 11, end_idx: 12 -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 16 -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: "response": "Hi there!"... -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: "response": "Hi there!"... -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: "response": "Hi there!"... -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:27:11,410 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity output_run1 resolution debug: -2025-07-24 17:27:11,410 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:27:11,410 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L12-L12 -2025-07-24 17:27:11,410 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:27:11,410 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: "response": "Hi there!"... -2025-07-24 17:27:11,410 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity output_run1: 43 characters -2025-07-24 17:27:11,410 - agentgraph.reconstruction.content_reference_resolver - INFO - Entity prompt resolution stats: {'total_entities': 4, 'entities_with_refs': 4, 'successful_resolutions': 4, 'failed_resolutions': 0} -2025-07-24 17:27:11,410 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 16 lines -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 16 -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L9-L9 -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L9-L9 -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 8, end_idx: 9 -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 16 -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: "query": "Hello"... -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: "query": "Hello"... -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: "query": "Hello"... -2025-07-24 17:27:11,410 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:27:11,410 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation relation_1: 34 characters -2025-07-24 17:27:11,410 - agentgraph.reconstruction.content_reference_resolver - INFO - Relation prompt resolution stats: {'total_relations': 3, 'relations_with_refs': 1, 'successful_resolutions': 1, 'failed_resolutions': 0} -2025-07-24 17:27:11,410 - agentgraph.reconstruction.content_reference_resolver - INFO - Resolved content references for knowledge graph with 4 entities and 3 relations -2025-07-24 17:27:11,410 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Resolved content references for window 0 -2025-07-24 17:27:11,410 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Completed sub-batch 1/1 -2025-07-24 17:27:11,411 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Preparing window knowledge graphs with metadata -2025-07-24 17:27:11,411 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Merging knowledge graphs... -2025-07-24 17:27:11,411 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Only one knowledge graph generated, skipping merge process -2025-07-24 17:27:11,411 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing complete. Knowledge graph generated with 4 entities and 3 relations -2025-07-24 17:27:11,412 - __main__ - INFO - Results saved to /tmp/test_auto_lines.json -2025-07-24 17:43:45,726 - openlit - INFO - Starting openLIT initialization... -2025-07-24 17:43:45,742 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 17:43:46,296 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 17:43:46,342 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 17:43:46,342 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 17:43:46,342 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 17:43:46,686 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 17:43:46,788 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 17:43:46,788 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 17:43:47,342 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 17:43:47,343 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 17:43:48,613 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 17:43:48,615 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 17:43:48,616 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 17:43:48,616 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 17:43:48,616 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 17:43:48,616 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 17:43:48,616 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 17:43:48,616 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 17:43:48,616 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 17:43:48,616 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 17:43:48,616 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 17:43:48,616 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 17:43:48,617 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 17:43:48,617 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 17:43:48,617 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 17:43:48,617 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 17:43:48,617 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 17:43:48,617 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 17:43:48,617 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 17:43:49,532 - __main__ - INFO - Loaded 3 texts from scripts/example_texts.json -2025-07-24 17:43:49,532 - agentgraph.input.text_processing.chunking_service - INFO - ChunkingService initialized with batch_size=3, model=gpt-4o-mini -2025-07-24 17:43:49,532 - __main__ - INFO - Initialized BatchKGExtractor: -2025-07-24 17:43:49,532 - __main__ - INFO - - Model: gpt-4o-mini, Method: production -2025-07-24 17:43:49,532 - __main__ - INFO - - Preprocessing: False -2025-07-24 17:43:49,532 - __main__ - INFO - - Line numbers: True -2025-07-24 17:43:49,532 - __main__ - INFO - Processing batch of 3 texts -2025-07-24 17:43:49,532 - __main__ - INFO - Processing text 1/3: text_0 -2025-07-24 17:43:49,532 - agentgraph.input.text_processing.trace_line_processor - INFO - Added line numbers to 6 lines, starting from line 1 -2025-07-24 17:43:49,532 - __main__ - INFO - Added line numbers: 6 lines -2025-07-24 17:43:49,532 - __main__ - INFO - Processing text text_0 (format: auto) -2025-07-24 17:43:49,532 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 17:43:49,532 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 390 characters -2025-07-24 17:43:49,532 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 17:43:49,532 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 17:43:49,538 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 17:43:49,538 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 17:43:49,540 - agentgraph.input.content_analysis.semantic_analyzer - INFO - Detected agent trace type: unknown (confidence: 0.00) -2025-07-24 17:43:49,541 - agentgraph.input.text_processing.chunking_service - INFO - Applied rule-based line splitting to 1 chunks (max_line_length=800) -2025-07-24 17:43:49,541 - agentgraph.input.text_processing.chunking_service - INFO - Assigning global line numbers to 1 chunks -2025-07-24 17:43:49,541 - agentgraph.input.text_processing.trace_line_processor - INFO - Added line numbers to 6 lines, starting from line 1 -2025-07-24 17:43:49,541 - agentgraph.input.text_processing.chunking_service - DEBUG - Chunk 0: chars 0-390 → lines 1-6 -2025-07-24 17:43:49,541 - agentgraph.input.text_processing.chunking_service - INFO - Successfully assigned global line numbers to all chunks -2025-07-24 17:43:49,541 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using agent_semantic splitter -2025-07-24 17:43:49,541 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=350000, overlap_size=17500 -2025-07-24 17:43:49,542 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context enhancement complete: 0 total documents -2025-07-24 17:43:49,542 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - SlidingWindowMonitor initialized with model: gpt-4o-mini, method: production -2025-07-24 17:43:49,542 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context documents: 0 provided -2025-07-24 17:43:49,542 - agentgraph.extraction.graph_utilities.knowledge_graph_merger - INFO - KnowledgeGraphMerger initialized with model: gpt-4o-mini -2025-07-24 17:43:49,543 - __main__ - INFO - Extracting knowledge graph from 1 chunks -2025-07-24 17:43:49,543 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks (splitter: agent_semantic, window_size=350000, overlap=0) -2025-07-24 17:43:49,543 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided source trace ID: text_0 -2025-07-24 17:43:49,543 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided processing run ID: batch_text_0 -2025-07-24 17:43:49,543 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks -2025-07-24 17:43:49,543 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using parallel processing with batch size 3 (1 batches) -2025-07-24 17:43:49,543 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 windows in 1 batches -2025-07-24 17:43:49,555 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks in 1 sub-batches (max 1 concurrent per sub-batch) -2025-07-24 17:43:49,556 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Starting sub-batch 1/1 with 1 chunks -2025-07-24 17:43:49,556 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing window 0 -2025-07-24 17:43:49,556 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - 🤖 Starting async_crew extraction for window 0 with method: production using model: gpt-4o-mini -2025-07-24 17:43:49,556 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - No context documents available for window 0, passing empty context string -2025-07-24 17:43:49,556 - agentgraph.methods.production.multi_agent_knowledge_extractor - INFO - Creating agent monitoring crew with model: gpt-4o-mini -2025-07-24 17:44:38,347 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Result Generated (production method) -2025-07-24 17:44:38,348 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 6 lines -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L2-L2 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L2-L2 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 1, end_idx: 2 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Role: You are a research assistant ... -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Role: You are a research assistant ... -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Role: You are a research assistant ... -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:44:38,348 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity agent_001 resolution debug: -2025-07-24 17:44:38,348 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:44:38,348 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L2-L2 -2025-07-24 17:44:38,348 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:44:38,348 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Role: You are a research assistant ... -2025-07-24 17:44:38,348 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity agent_001: 101 characters -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L3-L3 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L3-L3 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 2, end_idx: 3 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Task: Search for information about ... -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Task: Search for information about ... -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Task: Search for information about ... -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:44:38,348 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity task_001 resolution debug: -2025-07-24 17:44:38,348 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:44:38,348 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L3-L3 -2025-07-24 17:44:38,348 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:44:38,348 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Task: Search for information about ... -2025-07-24 17:44:38,348 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity task_001: 64 characters -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L4-L4 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L4-L4 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 3, end_idx: 4 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Tool: web_search... -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Tool: web_search... -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Tool: web_search... -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:44:38,348 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity tool_001 resolution debug: -2025-07-24 17:44:38,348 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:44:38,348 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L4-L4 -2025-07-24 17:44:38,348 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:44:38,348 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Tool: web_search... -2025-07-24 17:44:38,348 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity tool_001: 31 characters -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L5-L5 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L5-L5 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 4, end_idx: 5 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Input: climate change impacts 2024... -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Input: climate change impacts 2024... -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Input: climate change impacts 2024... -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:44:38,348 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity input_001 resolution debug: -2025-07-24 17:44:38,348 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:44:38,348 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L5-L5 -2025-07-24 17:44:38,348 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:44:38,348 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Input: climate change impacts 2024... -2025-07-24 17:44:38,348 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity input_001: 49 characters -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L6-L6 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L6-L6 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 5, end_idx: 6 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:44:38,348 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Output: Climate change continues to... -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Output: Climate change continues to... -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Output: Climate change continues to... -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:44:38,349 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity output_001 resolution debug: -2025-07-24 17:44:38,349 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:44:38,349 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L6-L6 -2025-07-24 17:44:38,349 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:44:38,349 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Output: Climate change continues to... -2025-07-24 17:44:38,349 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity output_001: 150 characters -2025-07-24 17:44:38,349 - agentgraph.reconstruction.content_reference_resolver - INFO - Entity prompt resolution stats: {'total_entities': 5, 'entities_with_refs': 5, 'successful_resolutions': 5, 'failed_resolutions': 0} -2025-07-24 17:44:38,349 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 6 lines -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L5-L5 -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L5-L5 -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 4, end_idx: 5 -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Input: climate change impacts 2024... -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Input: climate change impacts 2024... -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Input: climate change impacts 2024... -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:44:38,349 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation rel_001: 49 characters -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L4-L4 -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L4-L4 -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 3, end_idx: 4 -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Tool: web_search... -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Tool: web_search... -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Tool: web_search... -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:44:38,349 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation rel_004: 31 characters -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L6-L6 -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L6-L6 -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 5, end_idx: 6 -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Output: Climate change continues to... -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Output: Climate change continues to... -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Output: Climate change continues to... -2025-07-24 17:44:38,349 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:44:38,349 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation rel_007: 150 characters -2025-07-24 17:44:38,349 - agentgraph.reconstruction.content_reference_resolver - INFO - Relation prompt resolution stats: {'total_relations': 7, 'relations_with_refs': 3, 'successful_resolutions': 3, 'failed_resolutions': 0} -2025-07-24 17:44:38,349 - agentgraph.reconstruction.content_reference_resolver - INFO - Resolved content references for knowledge graph with 5 entities and 7 relations -2025-07-24 17:44:38,349 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Resolved content references for window 0 -2025-07-24 17:44:38,349 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Completed sub-batch 1/1 -2025-07-24 17:44:38,350 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Preparing window knowledge graphs with metadata -2025-07-24 17:44:38,350 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Merging knowledge graphs... -2025-07-24 17:44:38,350 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Only one knowledge graph generated, skipping merge process -2025-07-24 17:44:38,350 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing complete. Knowledge graph generated with 5 entities and 7 relations -2025-07-24 17:44:38,350 - __main__ - INFO - Processing text 2/3: text_1 -2025-07-24 17:44:38,350 - agentgraph.input.text_processing.trace_line_processor - INFO - Added line numbers to 8 lines, starting from line 1 -2025-07-24 17:44:38,350 - __main__ - INFO - Added line numbers: 8 lines -2025-07-24 17:44:38,350 - __main__ - INFO - Processing text text_1 (format: auto) -2025-07-24 17:44:38,350 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 17:44:38,350 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 485 characters -2025-07-24 17:44:38,350 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 17:44:38,350 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 17:44:38,358 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 17:44:38,358 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 17:44:38,358 - agentgraph.input.content_analysis.semantic_analyzer - INFO - Detected agent trace type: unknown (confidence: 0.00) -2025-07-24 17:44:38,359 - agentgraph.input.text_processing.chunking_service - INFO - Applied rule-based line splitting to 1 chunks (max_line_length=800) -2025-07-24 17:44:38,359 - agentgraph.input.text_processing.chunking_service - INFO - Assigning global line numbers to 1 chunks -2025-07-24 17:44:38,359 - agentgraph.input.text_processing.trace_line_processor - INFO - Added line numbers to 8 lines, starting from line 1 -2025-07-24 17:44:38,359 - agentgraph.input.text_processing.chunking_service - DEBUG - Chunk 0: chars 0-485 → lines 1-8 -2025-07-24 17:44:38,359 - agentgraph.input.text_processing.chunking_service - INFO - Successfully assigned global line numbers to all chunks -2025-07-24 17:44:38,359 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using agent_semantic splitter -2025-07-24 17:44:38,359 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=350000, overlap_size=17500 -2025-07-24 17:44:38,360 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context enhancement complete: 0 total documents -2025-07-24 17:44:38,360 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - SlidingWindowMonitor initialized with model: gpt-4o-mini, method: production -2025-07-24 17:44:38,360 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context documents: 0 provided -2025-07-24 17:44:38,360 - agentgraph.extraction.graph_utilities.knowledge_graph_merger - INFO - KnowledgeGraphMerger initialized with model: gpt-4o-mini -2025-07-24 17:44:38,363 - __main__ - INFO - Extracting knowledge graph from 1 chunks -2025-07-24 17:44:38,363 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks (splitter: agent_semantic, window_size=350000, overlap=0) -2025-07-24 17:44:38,363 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided source trace ID: text_1 -2025-07-24 17:44:38,363 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided processing run ID: batch_text_1 -2025-07-24 17:44:38,363 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks -2025-07-24 17:44:38,363 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using parallel processing with batch size 3 (1 batches) -2025-07-24 17:44:38,363 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 windows in 1 batches -2025-07-24 17:44:38,363 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks in 1 sub-batches (max 1 concurrent per sub-batch) -2025-07-24 17:44:38,363 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Starting sub-batch 1/1 with 1 chunks -2025-07-24 17:44:38,363 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing window 0 -2025-07-24 17:44:38,363 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - 🤖 Starting async_crew extraction for window 0 with method: production using model: gpt-4o-mini -2025-07-24 17:44:38,364 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - No context documents available for window 0, passing empty context string -2025-07-24 17:44:38,364 - agentgraph.methods.production.multi_agent_knowledge_extractor - INFO - Creating agent monitoring crew with model: gpt-4o-mini -2025-07-24 17:45:22,224 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Result Generated (production method) -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 8 lines -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 8 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L2-L2 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L2-L2 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 1, end_idx: 2 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 8 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Agent 1: Data Analyst - responsible... -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Agent 1: Data Analyst - responsible... -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Agent 1: Data Analyst - responsible... -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity agent_001 resolution debug: -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L2-L2 -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Agent 1: Data Analyst - responsible... -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity agent_001: 78 characters -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 8 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L3-L3 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L3-L3 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 2, end_idx: 3 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 8 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Agent 2: Report Generator - creates... -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Agent 2: Report Generator - creates... -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Agent 2: Report Generator - creates... -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity agent_002 resolution debug: -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L3-L3 -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Agent 2: Report Generator - creates... -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity agent_002: 72 characters -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 8 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L4-L4 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L4-L4 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 3, end_idx: 4 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 8 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Task 1: Analyze sales data from Q4 ... -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Task 1: Analyze sales data from Q4 ... -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Task 1: Analyze sales data from Q4 ... -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity task_001 resolution debug: -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L4-L4 -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Task 1: Analyze sales data from Q4 ... -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity task_001: 54 characters -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 8 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L5-L5 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L5-L5 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 4, end_idx: 5 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 8 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Task 2: Generate executive summary ... -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Task 2: Generate executive summary ... -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Task 2: Generate executive summary ... -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity task_002 resolution debug: -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L5-L5 -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Task 2: Generate executive summary ... -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity task_002: 56 characters -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 8 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L6-L6 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L6-L6 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 5, end_idx: 6 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 8 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Tool: pandas_analyzer - analyzes CS... -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Tool: pandas_analyzer - analyzes CS... -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Tool: pandas_analyzer - analyzes CS... -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity tool_001 resolution debug: -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L6-L6 -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Tool: pandas_analyzer - analyzes CS... -2025-07-24 17:45:22,225 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity tool_001: 56 characters -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 8 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L7-L7 -2025-07-24 17:45:22,225 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L7-L7 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 6, end_idx: 7 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 8 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Tool: report_generator - creates PD... -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Tool: report_generator - creates PD... -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Tool: report_generator - creates PD... -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:45:22,226 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity tool_002 resolution debug: -2025-07-24 17:45:22,226 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:45:22,226 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L7-L7 -2025-07-24 17:45:22,226 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:45:22,226 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Tool: report_generator - creates PD... -2025-07-24 17:45:22,226 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity tool_002: 59 characters -2025-07-24 17:45:22,226 - agentgraph.reconstruction.content_reference_resolver - INFO - Entity prompt resolution stats: {'total_entities': 6, 'entities_with_refs': 6, 'successful_resolutions': 6, 'failed_resolutions': 0} -2025-07-24 17:45:22,226 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 8 lines -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 8 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L8-L8 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L8-L8 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 7, end_idx: 8 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 8 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Flow: Data Analyst performs Task 1 ... -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Flow: Data Analyst performs Task 1 ... -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Flow: Data Analyst performs Task 1 ... -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:45:22,226 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation relation_001: 133 characters -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 8 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L8-L8 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L8-L8 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 7, end_idx: 8 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 8 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Flow: Data Analyst performs Task 1 ... -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Flow: Data Analyst performs Task 1 ... -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Flow: Data Analyst performs Task 1 ... -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:45:22,226 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation relation_002: 133 characters -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 8 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L8-L8 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L8-L8 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 7, end_idx: 8 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 8 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Flow: Data Analyst performs Task 1 ... -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Flow: Data Analyst performs Task 1 ... -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Flow: Data Analyst performs Task 1 ... -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:45:22,226 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation relation_003: 133 characters -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 8 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L8-L8 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L8-L8 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 7, end_idx: 8 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 8 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Flow: Data Analyst performs Task 1 ... -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Flow: Data Analyst performs Task 1 ... -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Flow: Data Analyst performs Task 1 ... -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:45:22,226 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation relation_004: 133 characters -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 8 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L8-L8 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L8-L8 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 7, end_idx: 8 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 8 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Flow: Data Analyst performs Task 1 ... -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Flow: Data Analyst performs Task 1 ... -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Flow: Data Analyst performs Task 1 ... -2025-07-24 17:45:22,226 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:45:22,226 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation relation_005: 133 characters -2025-07-24 17:45:22,226 - agentgraph.reconstruction.content_reference_resolver - INFO - Relation prompt resolution stats: {'total_relations': 5, 'relations_with_refs': 5, 'successful_resolutions': 5, 'failed_resolutions': 0} -2025-07-24 17:45:22,226 - agentgraph.reconstruction.content_reference_resolver - INFO - Resolved content references for knowledge graph with 6 entities and 5 relations -2025-07-24 17:45:22,226 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Resolved content references for window 0 -2025-07-24 17:45:22,226 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Completed sub-batch 1/1 -2025-07-24 17:45:22,227 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Preparing window knowledge graphs with metadata -2025-07-24 17:45:22,227 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Merging knowledge graphs... -2025-07-24 17:45:22,227 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Only one knowledge graph generated, skipping merge process -2025-07-24 17:45:22,227 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing complete. Knowledge graph generated with 6 entities and 5 relations -2025-07-24 17:45:22,227 - __main__ - INFO - Processing text 3/3: text_2 -2025-07-24 17:45:22,227 - agentgraph.input.text_processing.trace_line_processor - INFO - Added line numbers to 7 lines, starting from line 1 -2025-07-24 17:45:22,227 - __main__ - INFO - Added line numbers: 7 lines -2025-07-24 17:45:22,227 - __main__ - INFO - Processing text text_2 (format: auto) -2025-07-24 17:45:22,227 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 17:45:22,227 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 469 characters -2025-07-24 17:45:22,227 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 17:45:22,227 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 17:45:22,234 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 17:45:22,234 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 17:45:22,234 - agentgraph.input.content_analysis.semantic_analyzer - INFO - Detected agent trace type: unknown (confidence: 0.00) -2025-07-24 17:45:22,234 - agentgraph.input.text_processing.chunking_service - INFO - Applied rule-based line splitting to 1 chunks (max_line_length=800) -2025-07-24 17:45:22,234 - agentgraph.input.text_processing.chunking_service - INFO - Assigning global line numbers to 1 chunks -2025-07-24 17:45:22,234 - agentgraph.input.text_processing.trace_line_processor - INFO - Added line numbers to 7 lines, starting from line 1 -2025-07-24 17:45:22,234 - agentgraph.input.text_processing.chunking_service - DEBUG - Chunk 0: chars 0-469 → lines 1-7 -2025-07-24 17:45:22,234 - agentgraph.input.text_processing.chunking_service - INFO - Successfully assigned global line numbers to all chunks -2025-07-24 17:45:22,234 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using agent_semantic splitter -2025-07-24 17:45:22,234 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=350000, overlap_size=17500 -2025-07-24 17:45:22,235 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context enhancement complete: 0 total documents -2025-07-24 17:45:22,235 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - SlidingWindowMonitor initialized with model: gpt-4o-mini, method: production -2025-07-24 17:45:22,235 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context documents: 0 provided -2025-07-24 17:45:22,235 - agentgraph.extraction.graph_utilities.knowledge_graph_merger - INFO - KnowledgeGraphMerger initialized with model: gpt-4o-mini -2025-07-24 17:45:22,237 - __main__ - INFO - Extracting knowledge graph from 1 chunks -2025-07-24 17:45:22,237 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks (splitter: agent_semantic, window_size=350000, overlap=0) -2025-07-24 17:45:22,237 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided source trace ID: text_2 -2025-07-24 17:45:22,237 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided processing run ID: batch_text_2 -2025-07-24 17:45:22,237 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks -2025-07-24 17:45:22,237 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using parallel processing with batch size 3 (1 batches) -2025-07-24 17:45:22,237 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 windows in 1 batches -2025-07-24 17:45:22,237 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks in 1 sub-batches (max 1 concurrent per sub-batch) -2025-07-24 17:45:22,237 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Starting sub-batch 1/1 with 1 chunks -2025-07-24 17:45:22,237 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing window 0 -2025-07-24 17:45:22,237 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - 🤖 Starting async_crew extraction for window 0 with method: production using model: gpt-4o-mini -2025-07-24 17:45:22,237 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - No context documents available for window 0, passing empty context string -2025-07-24 17:45:22,237 - agentgraph.methods.production.multi_agent_knowledge_extractor - INFO - Creating agent monitoring crew with model: gpt-4o-mini -2025-07-24 17:46:06,616 - asyncio - ERROR - _GatheringFuture exception was never retrieved -future: <_GatheringFuture finished exception=CancelledError()> -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/agentgraph/extraction/graph_processing/knowledge_graph_processor.py", line 238, in process_window - result = await asyncio.wait_for( - ^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/asyncio/tasks.py", line 476, in wait_for - await waiter -asyncio.exceptions.CancelledError -2025-07-24 17:47:30,978 - openlit - INFO - Starting openLIT initialization... -2025-07-24 17:47:30,996 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 17:47:31,559 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 17:47:31,605 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 17:47:31,606 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 17:47:31,606 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 17:47:31,946 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 17:47:32,050 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 17:47:32,050 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 17:47:32,595 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 17:47:32,596 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 17:47:33,862 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 17:47:33,864 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 17:47:33,864 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 17:47:33,864 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 17:47:33,864 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 17:47:33,864 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 17:47:33,865 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 17:47:33,865 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 17:47:33,865 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 17:47:33,865 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 17:47:33,865 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 17:47:33,865 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 17:47:33,865 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 17:47:33,865 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 17:47:33,865 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 17:47:33,865 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 17:47:33,865 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 17:47:33,865 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 17:47:33,866 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 17:47:35,190 - __main__ - INFO - Loaded 3 texts from scripts/example_texts.json -2025-07-24 17:47:35,190 - agentgraph.input.text_processing.chunking_service - INFO - ChunkingService initialized with batch_size=3, model=gpt-4o-mini -2025-07-24 17:47:35,190 - __main__ - INFO - Initialized BatchKGExtractor: -2025-07-24 17:47:35,190 - __main__ - INFO - - Model: gpt-4o-mini, Method: production -2025-07-24 17:47:35,190 - __main__ - INFO - - Preprocessing: False -2025-07-24 17:47:35,190 - __main__ - INFO - - Line numbers: True -2025-07-24 17:47:35,190 - __main__ - INFO - Processing batch of 3 texts -2025-07-24 17:47:35,190 - __main__ - INFO - Processing text 1/3: text_0 -2025-07-24 17:47:35,190 - agentgraph.input.text_processing.trace_line_processor - INFO - Added line numbers to 6 lines, starting from line 1 -2025-07-24 17:47:35,190 - __main__ - INFO - Added line numbers: 6 lines -2025-07-24 17:47:35,190 - __main__ - INFO - Processing text text_0 (format: auto) -2025-07-24 17:47:35,190 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 17:47:35,190 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 390 characters -2025-07-24 17:47:35,190 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 17:47:35,190 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 17:47:35,196 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 17:47:35,196 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 17:47:35,199 - agentgraph.input.content_analysis.semantic_analyzer - INFO - Detected agent trace type: unknown (confidence: 0.00) -2025-07-24 17:47:35,199 - agentgraph.input.text_processing.chunking_service - INFO - Applied rule-based line splitting to 1 chunks (max_line_length=800) -2025-07-24 17:47:35,200 - agentgraph.input.text_processing.chunking_service - INFO - Content already has line numbers, skipping line number assignment -2025-07-24 17:47:35,200 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using agent_semantic splitter -2025-07-24 17:47:35,200 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=350000, overlap_size=17500 -2025-07-24 17:47:35,201 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context enhancement complete: 0 total documents -2025-07-24 17:47:35,201 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - SlidingWindowMonitor initialized with model: gpt-4o-mini, method: production -2025-07-24 17:47:35,201 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context documents: 0 provided -2025-07-24 17:47:35,201 - agentgraph.extraction.graph_utilities.knowledge_graph_merger - INFO - KnowledgeGraphMerger initialized with model: gpt-4o-mini -2025-07-24 17:47:35,203 - __main__ - INFO - Extracting knowledge graph from 1 chunks -2025-07-24 17:47:35,203 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks (splitter: agent_semantic, window_size=350000, overlap=0) -2025-07-24 17:47:35,203 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided source trace ID: text_0 -2025-07-24 17:47:35,203 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided processing run ID: batch_text_0 -2025-07-24 17:47:35,203 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks -2025-07-24 17:47:35,203 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using parallel processing with batch size 3 (1 batches) -2025-07-24 17:47:35,203 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 windows in 1 batches -2025-07-24 17:47:35,216 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks in 1 sub-batches (max 1 concurrent per sub-batch) -2025-07-24 17:47:35,217 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Starting sub-batch 1/1 with 1 chunks -2025-07-24 17:47:35,217 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing window 0 -2025-07-24 17:47:35,217 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - 🤖 Starting async_crew extraction for window 0 with method: production using model: gpt-4o-mini -2025-07-24 17:47:35,217 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - No context documents available for window 0, passing empty context string -2025-07-24 17:47:35,217 - agentgraph.methods.production.multi_agent_knowledge_extractor - INFO - Creating agent monitoring crew with model: gpt-4o-mini -2025-07-24 17:48:03,564 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Result Generated (production method) -2025-07-24 17:48:03,565 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 6 lines -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L2-L2 -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L2-L2 -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 1, end_idx: 2 -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Role: You are a research assistant AI th... -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Role: You are a research assistant AI th... -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Role: You are a research assistant AI th... -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:48:03,565 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity agent_001 resolution debug: -2025-07-24 17:48:03,565 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:48:03,565 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L2-L2 -2025-07-24 17:48:03,565 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:48:03,565 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Role: You are a research assistant AI th... -2025-07-24 17:48:03,565 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity agent_001: 96 characters -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L3-L3 -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L3-L3 -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 2, end_idx: 3 -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Task: Search for information about clima... -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Task: Search for information about clima... -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Task: Search for information about clima... -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:48:03,565 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity task_001 resolution debug: -2025-07-24 17:48:03,565 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:48:03,565 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L3-L3 -2025-07-24 17:48:03,565 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:48:03,565 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Task: Search for information about clima... -2025-07-24 17:48:03,565 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity task_001: 59 characters -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L4-L4 -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L4-L4 -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 3, end_idx: 4 -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Tool: web_search... -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Tool: web_search... -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Tool: web_search... -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:48:03,565 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity tool_001 resolution debug: -2025-07-24 17:48:03,565 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:48:03,565 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L4-L4 -2025-07-24 17:48:03,565 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:48:03,565 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Tool: web_search... -2025-07-24 17:48:03,565 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity tool_001: 26 characters -2025-07-24 17:48:03,565 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L5-L5 -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L5-L5 -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 4, end_idx: 5 -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Input: climate change impacts 2024... -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Input: climate change impacts 2024... -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Input: climate change impacts 2024... -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:48:03,566 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity input_001 resolution debug: -2025-07-24 17:48:03,566 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:48:03,566 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L5-L5 -2025-07-24 17:48:03,566 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:48:03,566 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Input: climate change impacts 2024... -2025-07-24 17:48:03,566 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity input_001: 44 characters -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L6-L6 -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L6-L6 -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 5, end_idx: 6 -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Output: Climate change continues to show... -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Output: Climate change continues to show... -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Output: Climate change continues to show... -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:48:03,566 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity output_001 resolution debug: -2025-07-24 17:48:03,566 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:48:03,566 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L6-L6 -2025-07-24 17:48:03,566 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:48:03,566 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Output: Climate change continues to show... -2025-07-24 17:48:03,566 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity output_001: 150 characters -2025-07-24 17:48:03,566 - agentgraph.reconstruction.content_reference_resolver - INFO - Entity prompt resolution stats: {'total_entities': 5, 'entities_with_refs': 5, 'successful_resolutions': 5, 'failed_resolutions': 0} -2025-07-24 17:48:03,566 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 6 lines -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L5-L5 -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L5-L5 -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 4, end_idx: 5 -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Input: climate change impacts 2024... -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Input: climate change impacts 2024... -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Input: climate change impacts 2024... -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:48:03,566 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation relation_001: 44 characters -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L2-L2 -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L2-L2 -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 1, end_idx: 2 -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Role: You are a research assistant AI th... -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Role: You are a research assistant AI th... -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Role: You are a research assistant AI th... -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:48:03,566 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation relation_002: 96 characters -2025-07-24 17:48:03,566 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:48:03,567 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:48:03,567 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L4-L4 -2025-07-24 17:48:03,567 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L4-L4 -2025-07-24 17:48:03,567 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 3, end_idx: 4 -2025-07-24 17:48:03,567 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:48:03,567 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:48:03,567 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Tool: web_search... -2025-07-24 17:48:03,567 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Tool: web_search... -2025-07-24 17:48:03,567 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Tool: web_search... -2025-07-24 17:48:03,567 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:48:03,567 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation relation_003: 26 characters -2025-07-24 17:48:03,567 - agentgraph.reconstruction.content_reference_resolver - INFO - Relation prompt resolution stats: {'total_relations': 4, 'relations_with_refs': 3, 'successful_resolutions': 3, 'failed_resolutions': 0} -2025-07-24 17:48:03,567 - agentgraph.reconstruction.content_reference_resolver - INFO - Resolved content references for knowledge graph with 5 entities and 4 relations -2025-07-24 17:48:03,567 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Resolved content references for window 0 -2025-07-24 17:48:03,567 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Completed sub-batch 1/1 -2025-07-24 17:48:03,567 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Preparing window knowledge graphs with metadata -2025-07-24 17:48:03,568 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Merging knowledge graphs... -2025-07-24 17:48:03,568 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Only one knowledge graph generated, skipping merge process -2025-07-24 17:48:03,568 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing complete. Knowledge graph generated with 5 entities and 4 relations -2025-07-24 17:48:03,568 - __main__ - INFO - Processing text 2/3: text_1 -2025-07-24 17:48:03,568 - agentgraph.input.text_processing.trace_line_processor - INFO - Added line numbers to 8 lines, starting from line 1 -2025-07-24 17:48:03,568 - __main__ - INFO - Added line numbers: 8 lines -2025-07-24 17:48:03,568 - __main__ - INFO - Processing text text_1 (format: auto) -2025-07-24 17:48:03,568 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 17:48:03,568 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 485 characters -2025-07-24 17:48:03,568 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 17:48:03,568 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 17:48:03,577 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 17:48:03,577 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 17:48:03,578 - agentgraph.input.content_analysis.semantic_analyzer - INFO - Detected agent trace type: unknown (confidence: 0.00) -2025-07-24 17:48:03,578 - agentgraph.input.text_processing.chunking_service - INFO - Applied rule-based line splitting to 1 chunks (max_line_length=800) -2025-07-24 17:48:03,578 - agentgraph.input.text_processing.chunking_service - INFO - Content already has line numbers, skipping line number assignment -2025-07-24 17:48:03,578 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using agent_semantic splitter -2025-07-24 17:48:03,578 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=350000, overlap_size=17500 -2025-07-24 17:48:03,579 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context enhancement complete: 0 total documents -2025-07-24 17:48:03,579 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - SlidingWindowMonitor initialized with model: gpt-4o-mini, method: production -2025-07-24 17:48:03,579 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context documents: 0 provided -2025-07-24 17:48:03,579 - agentgraph.extraction.graph_utilities.knowledge_graph_merger - INFO - KnowledgeGraphMerger initialized with model: gpt-4o-mini -2025-07-24 17:48:03,582 - __main__ - INFO - Extracting knowledge graph from 1 chunks -2025-07-24 17:48:03,582 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks (splitter: agent_semantic, window_size=350000, overlap=0) -2025-07-24 17:48:03,582 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided source trace ID: text_1 -2025-07-24 17:48:03,582 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided processing run ID: batch_text_1 -2025-07-24 17:48:03,582 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks -2025-07-24 17:48:03,582 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using parallel processing with batch size 3 (1 batches) -2025-07-24 17:48:03,582 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 windows in 1 batches -2025-07-24 17:48:03,582 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks in 1 sub-batches (max 1 concurrent per sub-batch) -2025-07-24 17:48:03,582 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Starting sub-batch 1/1 with 1 chunks -2025-07-24 17:48:03,583 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing window 0 -2025-07-24 17:48:03,583 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - 🤖 Starting async_crew extraction for window 0 with method: production using model: gpt-4o-mini -2025-07-24 17:48:03,583 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - No context documents available for window 0, passing empty context string -2025-07-24 17:48:03,583 - agentgraph.methods.production.multi_agent_knowledge_extractor - INFO - Creating agent monitoring crew with model: gpt-4o-mini -2025-07-24 17:48:14,078 - instructor - DEBUG - Patching `client.chat.completions.create` with mode= -2025-07-24 17:48:14,084 - instructor - DEBUG - Instructor Request: mode.value='tool_call', response_model=, new_kwargs={'messages': [{'role': 'user', 'content': '{\n "entities": [\n {\n "id": "agent_001",\n "type": "Agent",\n "name": "Data Analyst",\n "importance": "HIGH",\n "raw_prompt": "",\n "raw_prompt_ref": [\n {"line_start": 2, "line_end": 2, "confidence": 1.0}\n ]\n },\n {\n "id": "agent_002",\n "type": "Agent",\n "name": "Report Generator",\n "importance": "HIGH",\n "raw_prompt": "",\n "raw_prompt_ref": [\n {"line_start": 3, "line_end": 3, "confidence": 1.0}\n ]\n },\n {\n "id": "task_001",\n "type": "Task",\n "name": "Analyze sales data from Q4 2023",\n "importance": "HIGH",\n "raw_prompt": "",\n "raw_prompt_ref": [\n {"line_start": 4, "line_end": 4, "confidence": 1.0}\n ]\n },\n {\n "id": "task_002",\n "type": "Task",\n "name": "Generate executive summary report",\n "importance": "HIGH",\n "raw_prompt": "",\n "raw_prompt_ref": [\n {"line_start": 5, "line_end": 5, "confidence": 1.0}\n ]\n },\n {\n "id": "tool_001",\n "type": "Tool",\n "name": "pandas_analyzer",\n "importance": "MEDIUM",\n "raw_prompt": "",\n "raw_prompt_ref": [\n {"line_start": 6, "line_end": 6, "confidence": 1.0}\n ]\n },\n {\n "id": "tool_002",\n "type": "Tool",\n "name": "report_generator",\n "importance": "MEDIUM",\n "raw_prompt": "",\n "raw_prompt_ref": [\n {"line_start": 7, "line_end": 7, "confidence": 1.0}\n ]\n },\n {\n "id": "flow_001",\n "type": "Flow",\n "name": "Data Analysis Workflow",\n "importance": "HIGH",\n "raw_prompt": "",\n "raw_prompt_ref": [\n {"line_start": 8, "line_end": 8, "confidence": 1.0}\n ]\n }\n ]\n}'}], 'model': 'gpt-4o-mini', 'tools': [{'type': 'function', 'function': {'name': 'EntityExtractionList', 'description': 'Correctly extracted `EntityExtractionList` with all the required parameters with correct types', 'parameters': {'$defs': {'ContentReference': {'description': 'Reference to content location in the original trace using line numbers and character positions.\nThis allows AI agents to provide position metadata instead of full content, enabling \nefficient mapping back to the original trace while reducing hallucination risks.\n\nCRITICAL FOR LLMs: Line counting accuracy is essential for proper content resolution.\nUse systematic counting methods and verify your line numbers before submission.', 'properties': {'line_start': {'description': 'Starting line number where the content begins (1-based indexing from , ... markers).\n \n ACCURACY REQUIREMENTS FOR LLMs:\n - Count markers systematically from the beginning of the input\n - Use anchor points: find distinctive text first, then count nearby lines\n - Double-check by counting backwards from a known reference point\n - For multi-line content, this should be the FIRST line containing the content\n - In key-value pairs (e.g. "content": "..."), reference the line where the VALUE starts, not the key\n \n COMMON ERRORS TO AVOID:\n - Miscounting due to skipping indented continuation lines\n - Confusing line numbers when content spans multiple markers\n - Using approximate counting instead of precise marker identification\n \n VERIFICATION: Before submitting, locate your chosen line number and confirm it contains the expected content start.', 'title': 'Line Start', 'type': 'integer'}, 'line_end': {'description': 'Ending line number where content ends (1-based indexing from , ... markers).\n \n ACCURACY REQUIREMENTS FOR LLMs:\n - Must be >= line_start (validation will fail otherwise)\n - For single-line content, line_end should equal line_start\n - For multi-line content, find the LAST line containing the content\n - Include indented continuation lines that are part of the same logical content block\n \n VERIFICATION STRATEGY:\n - Count from line_start to ensure proper range\n - Confirm the line_end marker contains the actual end of the content\n - Check that no content continues beyond your specified line_end', 'title': 'Line End', 'type': 'integer'}, 'confidence': {'default': None, 'description': 'Confidence score for the location reference accuracy (0.0 - 1.0).\n \n CONFIDENCE SCORING GUIDE FOR LLMs:\n - 1.0: Verified by counting twice with consistent results, clear content boundaries\n - 0.9: High confidence with single verification, unambiguous content location \n - 0.8: Good confidence but content boundaries somewhat ambiguous\n - 0.7: Moderate confidence, some uncertainty in exact line boundaries\n - 0.6: Lower confidence due to complex content structure or counting difficulty\n - 0.5 or below: Uncertain about accuracy, recommend manual verification\n \n FACTORS AFFECTING CONFIDENCE:\n - Clarity of content boundaries (higher = more confident)\n - Complexity of surrounding text (simpler = more confident) \n - Verification method used (double-counting = more confident)\n - Presence of clear anchor points (more anchors = more confident)', 'maximum': 1.0, 'minimum': 0.0, 'title': 'Confidence', 'type': 'number'}}, 'required': ['line_start', 'line_end'], 'title': 'ContentReference', 'type': 'object'}, 'Entity': {'properties': {'id': {'description': 'Unique identifier for the entity', 'title': 'Id', 'type': 'string'}, 'type': {'description': 'Type of entity defined by prompt type: Agent (system prompt), Task (instruction prompt), Tool (description prompt), Input (input format prompt), Output (output format prompt), Human (optional prompt). The raw_prompt field is the primary distinguishing factor for entity uniqueness and classification.', 'enum': ['Agent', 'Task', 'Tool', 'Input', 'Output', 'Human'], 'title': 'Type', 'type': 'string'}, 'name': {'description': "Name of the entity derived from the prompt content. Names should reflect the specific prompt or specification that defines this entity. For composite entities, use descriptive names that capture the prompt's scope (e.g., 'SQL Query Generation System Prompt', 'Data Analysis Instruction Set').", 'title': 'Name', 'type': 'string'}, 'importance': {'description': 'Importance level of this entity in the system. HIGH: Core agents, critical tasks, essential tools that are central to system function. MEDIUM: Supporting agents, standard tasks, commonly used tools. LOW: Auxiliary entities, simple tasks, rarely used components.', 'enum': ['HIGH', 'MEDIUM', 'LOW'], 'title': 'Importance', 'type': 'string'}, 'raw_prompt': {'default': '', 'description': 'PRIMARY DISTINGUISHING CONTENT: The actual prompt, specification, or instruction that defines this entity. This is the core content that makes each entity unique and should contain: For Agents (system prompts defining role/capabilities), For Tasks (instruction prompts defining objectives), For Tools (description prompts defining functionality), For Inputs (format specifications), For Outputs (format specifications), For Humans (interaction patterns). This field is more important than the name for entity distinction and relationship mapping.', 'title': 'Raw Prompt', 'type': 'string'}, 'raw_prompt_ref': {'description': 'A list of references to the locations of the raw prompt content in the original trace. When provided, this allows mapping back to all exact positions in the trace where this prompt was found.', 'items': {'$ref': '#/$defs/ContentReference'}, 'title': 'Raw Prompt Ref', 'type': 'array'}}, 'required': ['id', 'type', 'name', 'importance'], 'title': 'Entity', 'type': 'object'}}, 'properties': {'entities': {'default': [], 'items': {'$ref': '#/$defs/Entity'}, 'title': 'Entities', 'type': 'array'}}, 'type': 'object', 'required': []}}}], 'tool_choice': {'type': 'function', 'function': {'name': 'EntityExtractionList'}}} -2025-07-24 17:48:14,084 - instructor - DEBUG - max_retries: 3 -2025-07-24 17:48:14,085 - instructor - DEBUG - Retrying, attempt: 1 -2025-07-24 17:48:19,775 - instructor - DEBUG - Instructor Raw Response: ModelResponse(id='chatcmpl-BwtP4mSLELyJH2dGnPQLZuZpXAzia', created=1753375694, model='gpt-4o-mini-2024-07-18', object='chat.completion', system_fingerprint=None, choices=[Choices(finish_reason='stop', index=0, message=Message(content=None, role='assistant', tool_calls=[ChatCompletionMessageToolCall(function=Function(arguments='{"entities":[{"id":"agent_001","type":"Agent","name":"Data Analyst","importance":"HIGH","raw_prompt":"","raw_prompt_ref":[{"line_start":2,"line_end":2,"confidence":1.0}]},{"id":"agent_002","type":"Agent","name":"Report Generator","importance":"HIGH","raw_prompt":"","raw_prompt_ref":[{"line_start":3,"line_end":3,"confidence":1.0}]},{"id":"task_001","type":"Task","name":"Analyze sales data from Q4 2023","importance":"HIGH","raw_prompt":"","raw_prompt_ref":[{"line_start":4,"line_end":4,"confidence":1.0}]},{"id":"task_002","type":"Task","name":"Generate executive summary report","importance":"HIGH","raw_prompt":"","raw_prompt_ref":[{"line_start":5,"line_end":5,"confidence":1.0}]},{"id":"tool_001","type":"Tool","name":"pandas_analyzer","importance":"MEDIUM","raw_prompt":"","raw_prompt_ref":[{"line_start":6,"line_end":6,"confidence":1.0}]},{"id":"tool_002","type":"Tool","name":"report_generator","importance":"MEDIUM","raw_prompt":"","raw_prompt_ref":[{"line_start":7,"line_end":7,"confidence":1.0}]},{"id":"flow_001","type":"Flow","name":"Data Analysis Workflow","importance":"HIGH","raw_prompt":"","raw_prompt_ref":[{"line_start":8,"line_end":8,"confidence":1.0}]}]}', name='EntityExtractionList'), id='call_gQBWjcRANh1wMndIJxzFQrhE', type='function')], function_call=None, provider_specific_fields={'refusal': None, 'annotations': []}, refusal=None, annotations=[]))], usage=CompletionUsage(completion_tokens=327, prompt_tokens=1637, total_tokens=1964, completion_tokens_details=CompletionTokensDetails(accepted_prediction_tokens=None, audio_tokens=0, reasoning_tokens=0, rejected_prediction_tokens=None), prompt_tokens_details=PromptTokensDetails(audio_tokens=0, cached_tokens=0)), service_tier='default') -2025-07-24 17:48:19,778 - instructor - DEBUG - Parse error: 1 validation error for EntityExtractionList -entities.6.type - Input should be 'Agent', 'Task', 'Tool', 'Input', 'Output' or 'Human' [type=literal_error, input_value='Flow', input_type=str] - For further information visit https://errors.pydantic.dev/2.11/v/literal_error -2025-07-24 17:48:19,778 - instructor - DEBUG - Retrying, attempt: 2 -2025-07-24 17:48:24,738 - instructor - DEBUG - Instructor Raw Response: ModelResponse(id='chatcmpl-BwtP9yXkWM7H4vShUKhIshsWEEyp3', created=1753375699, model='gpt-4o-mini-2024-07-18', object='chat.completion', system_fingerprint=None, choices=[Choices(finish_reason='stop', index=0, message=Message(content=None, role='assistant', tool_calls=[ChatCompletionMessageToolCall(function=Function(arguments='{"entities":[{"id":"agent_001","type":"Agent","name":"Data Analyst","importance":"HIGH","raw_prompt":"","raw_prompt_ref":[{"line_start":2,"line_end":2,"confidence":1.0}]},{"id":"agent_002","type":"Agent","name":"Report Generator","importance":"HIGH","raw_prompt":"","raw_prompt_ref":[{"line_start":3,"line_end":3,"confidence":1.0}]},{"id":"task_001","type":"Task","name":"Analyze sales data from Q4 2023","importance":"HIGH","raw_prompt":"","raw_prompt_ref":[{"line_start":4,"line_end":4,"confidence":1.0}]},{"id":"task_002","type":"Task","name":"Generate executive summary report","importance":"HIGH","raw_prompt":"","raw_prompt_ref":[{"line_start":5,"line_end":5,"confidence":1.0}]},{"id":"tool_001","type":"Tool","name":"pandas_analyzer","importance":"MEDIUM","raw_prompt":"","raw_prompt_ref":[{"line_start":6,"line_end":6,"confidence":1.0}]},{"id":"tool_002","type":"Tool","name":"report_generator","importance":"MEDIUM","raw_prompt":"","raw_prompt_ref":[{"line_start":7,"line_end":7,"confidence":1.0}]}]}', name='EntityExtractionList'), id='call_m3K4MX35QA2x82MKECDvQljk', type='function')], function_call=None, provider_specific_fields={'refusal': None, 'annotations': []}, refusal=None, annotations=[]))], usage=CompletionUsage(completion_tokens=609, prompt_tokens=3704, total_tokens=4313, completion_tokens_details=CompletionTokensDetails(accepted_prediction_tokens=None, audio_tokens=0, reasoning_tokens=0, rejected_prediction_tokens=None), prompt_tokens_details=PromptTokensDetails(audio_tokens=0, cached_tokens=1920)), service_tier='default') -2025-07-24 17:49:00,344 - instructor - DEBUG - Patching `client.chat.completions.create` with mode= -2025-07-24 17:49:00,357 - instructor - DEBUG - Instructor Request: mode.value='tool_call', response_model=, new_kwargs={'messages': [{'role': 'user', 'content': '{\n "entities": [\n {\n "id": "agent_001",\n "type": "Agent",\n "name": "Data Analyst",\n "importance": "HIGH",\n "raw_prompt": "",\n "raw_prompt_ref": [\n {\n "line_start": 2,\n "line_end": 2,\n "confidence": 1.0\n }\n ]\n },\n {\n "id": "agent_002",\n "type": "Agent",\n "name": "Report Generator",\n "importance": "HIGH",\n "raw_prompt": "",\n "raw_prompt_ref": [\n {\n "line_start": 3,\n "line_end": 3,\n "confidence": 1.0\n }\n ]\n },\n {\n "id": "task_001",\n "type": "Task",\n "name": "Analyze sales data from Q4 2023",\n "importance": "HIGH",\n "raw_prompt": "",\n "raw_prompt_ref": [\n {\n "line_start": 4,\n "line_end": 4,\n "confidence": 1.0\n }\n ]\n },\n {\n "id": "task_002",\n "type": "Task",\n "name": "Generate executive summary report",\n "importance": "HIGH",\n "raw_prompt": "",\n "raw_prompt_ref": [\n {\n "line_start": 5,\n "line_end": 5,\n "confidence": 1.0\n }\n ]\n },\n {\n "id": "tool_001",\n "type": "Tool",\n "name": "pandas_analyzer",\n "importance": "MEDIUM",\n "raw_prompt": "",\n "raw_prompt_ref": [\n {\n "line_start": 6,\n "line_end": 6,\n "confidence": 1.0\n }\n ]\n },\n {\n "id": "tool_002",\n "type": "Tool",\n "name": "report_generator",\n "importance": "MEDIUM",\n "raw_prompt": "",\n "raw_prompt_ref": [\n {\n "line_start": 7,\n "line_end": 7,\n "confidence": 1.0\n }\n ]\n },\n {\n "id": "flow_001",\n "type": "Flow",\n "name": "Data Analysis Workflow",\n "importance": "HIGH",\n "raw_prompt": "",\n "raw_prompt_ref": [\n {\n "line_start": 8,\n "line_end": 8,\n "confidence": 1.0\n }\n ]\n }\n ],\n "relations": [\n {\n "id": "relation_001",\n "source": "agent_001",\n "target": "task_001",\n "type": "PERFORMS",\n "importance": "HIGH",\n "interaction_prompt": "",\n "interaction_prompt_ref": [\n {\n "line_start": 8,\n "line_end": 8,\n "confidence": 1.0\n }\n ]\n },\n {\n "id": "relation_002",\n "source": "agent_001",\n "target": "tool_001",\n "type": "USES",\n "importance": "MEDIUM",\n "interaction_prompt": "",\n "interaction_prompt_ref": [\n {\n "line_start": 8,\n "line_end": 8,\n "confidence": 1.0\n }\n ]\n },\n {\n "id": "relation_003",\n "source": "agent_002",\n "target": "task_002",\n "type": "PERFORMS",\n "importance": "HIGH",\n "interaction_prompt": "",\n "interaction_prompt_ref": [\n {\n "line_start": 8,\n "line_end": 8,\n "confidence": 1.0\n }\n ]\n },\n {\n "id": "relation_004",\n "source": "agent_002",\n "target": "tool_002",\n "type": "USES",\n "importance": "MEDIUM",\n "interaction_prompt": "",\n "interaction_prompt_ref": [\n {\n "line_start": 8,\n "line_end": 8,\n "confidence": 1.0\n }\n ]\n },\n {\n "id": "relation_005",\n "source": "task_001",\n "target": "agent_001",\n "type": "ASSIGNED_TO",\n "importance": "HIGH",\n "interaction_prompt": "",\n "interaction_prompt_ref": []\n },\n {\n "id": "relation_006",\n "source": "task_002",\n "target": "agent_002",\n "type": "ASSIGNED_TO",\n "importance": "HIGH",\n "interaction_prompt": "",\n "interaction_prompt_ref": []\n }\n ],\n "failures": [\n {\n "id": "failure_001",\n "risk_type": "EXECUTION_ERROR",\n "description": "Data Analyst failed to properly analyze the sales data leading to incomplete reports.",\n "raw_text": "",\n "raw_text_ref": [\n {\n "line_start": 4,\n "line_end": 4,\n "confidence": 1.0\n }\n ],\n "affected_id": "task_001"\n }\n ],\n "system_name": "CrewAI Multi-Agent Workflow System",\n "system_summary": "The CrewAI Multi-Agent Workflow System coordinates a Data Analyst and Report Generator to analyze customer sales data and generate executive reports. The system improves efficiency by automating data analysis and report generation, with clear interaction pathways between agents and tasks."\n}'}], 'model': 'gpt-4o-mini', 'tools': [{'type': 'function', 'function': {'name': 'KnowledgeGraph', 'description': 'Correctly extracted `KnowledgeGraph` with all the required parameters with correct types', 'parameters': {'$defs': {'ContentReference': {'description': 'Reference to content location in the original trace using line numbers and character positions.\nThis allows AI agents to provide position metadata instead of full content, enabling \nefficient mapping back to the original trace while reducing hallucination risks.\n\nCRITICAL FOR LLMs: Line counting accuracy is essential for proper content resolution.\nUse systematic counting methods and verify your line numbers before submission.', 'properties': {'line_start': {'description': 'Starting line number where the content begins (1-based indexing from , ... markers).\n \n ACCURACY REQUIREMENTS FOR LLMs:\n - Count markers systematically from the beginning of the input\n - Use anchor points: find distinctive text first, then count nearby lines\n - Double-check by counting backwards from a known reference point\n - For multi-line content, this should be the FIRST line containing the content\n - In key-value pairs (e.g. "content": "..."), reference the line where the VALUE starts, not the key\n \n COMMON ERRORS TO AVOID:\n - Miscounting due to skipping indented continuation lines\n - Confusing line numbers when content spans multiple markers\n - Using approximate counting instead of precise marker identification\n \n VERIFICATION: Before submitting, locate your chosen line number and confirm it contains the expected content start.', 'title': 'Line Start', 'type': 'integer'}, 'line_end': {'description': 'Ending line number where content ends (1-based indexing from , ... markers).\n \n ACCURACY REQUIREMENTS FOR LLMs:\n - Must be >= line_start (validation will fail otherwise)\n - For single-line content, line_end should equal line_start\n - For multi-line content, find the LAST line containing the content\n - Include indented continuation lines that are part of the same logical content block\n \n VERIFICATION STRATEGY:\n - Count from line_start to ensure proper range\n - Confirm the line_end marker contains the actual end of the content\n - Check that no content continues beyond your specified line_end', 'title': 'Line End', 'type': 'integer'}, 'confidence': {'default': None, 'description': 'Confidence score for the location reference accuracy (0.0 - 1.0).\n \n CONFIDENCE SCORING GUIDE FOR LLMs:\n - 1.0: Verified by counting twice with consistent results, clear content boundaries\n - 0.9: High confidence with single verification, unambiguous content location \n - 0.8: Good confidence but content boundaries somewhat ambiguous\n - 0.7: Moderate confidence, some uncertainty in exact line boundaries\n - 0.6: Lower confidence due to complex content structure or counting difficulty\n - 0.5 or below: Uncertain about accuracy, recommend manual verification\n \n FACTORS AFFECTING CONFIDENCE:\n - Clarity of content boundaries (higher = more confident)\n - Complexity of surrounding text (simpler = more confident) \n - Verification method used (double-counting = more confident)\n - Presence of clear anchor points (more anchors = more confident)', 'maximum': 1.0, 'minimum': 0.0, 'title': 'Confidence', 'type': 'number'}}, 'required': ['line_start', 'line_end'], 'title': 'ContentReference', 'type': 'object'}, 'Entity': {'properties': {'id': {'description': 'Unique identifier for the entity', 'title': 'Id', 'type': 'string'}, 'type': {'description': 'Type of entity defined by prompt type: Agent (system prompt), Task (instruction prompt), Tool (description prompt), Input (input format prompt), Output (output format prompt), Human (optional prompt). The raw_prompt field is the primary distinguishing factor for entity uniqueness and classification.', 'enum': ['Agent', 'Task', 'Tool', 'Input', 'Output', 'Human'], 'title': 'Type', 'type': 'string'}, 'name': {'description': "Name of the entity derived from the prompt content. Names should reflect the specific prompt or specification that defines this entity. For composite entities, use descriptive names that capture the prompt's scope (e.g., 'SQL Query Generation System Prompt', 'Data Analysis Instruction Set').", 'title': 'Name', 'type': 'string'}, 'importance': {'description': 'Importance level of this entity in the system. HIGH: Core agents, critical tasks, essential tools that are central to system function. MEDIUM: Supporting agents, standard tasks, commonly used tools. LOW: Auxiliary entities, simple tasks, rarely used components.', 'enum': ['HIGH', 'MEDIUM', 'LOW'], 'title': 'Importance', 'type': 'string'}, 'raw_prompt': {'default': '', 'description': 'PRIMARY DISTINGUISHING CONTENT: The actual prompt, specification, or instruction that defines this entity. This is the core content that makes each entity unique and should contain: For Agents (system prompts defining role/capabilities), For Tasks (instruction prompts defining objectives), For Tools (description prompts defining functionality), For Inputs (format specifications), For Outputs (format specifications), For Humans (interaction patterns). This field is more important than the name for entity distinction and relationship mapping.', 'title': 'Raw Prompt', 'type': 'string'}, 'raw_prompt_ref': {'description': 'A list of references to the locations of the raw prompt content in the original trace. When provided, this allows mapping back to all exact positions in the trace where this prompt was found.', 'items': {'$ref': '#/$defs/ContentReference'}, 'title': 'Raw Prompt Ref', 'type': 'array'}}, 'required': ['id', 'type', 'name', 'importance'], 'title': 'Entity', 'type': 'object'}, 'Failure': {'description': 'Represents a failure / risk event located via ContentReference.', 'properties': {'id': {'description': 'Unique identifier for the failure event', 'title': 'Id', 'type': 'string'}, 'risk_type': {'description': 'Categorised failure type (predefined list)', 'enum': ['AGENT_ERROR', 'PLANNING_ERROR', 'EXECUTION_ERROR', 'RETRIEVAL_ERROR', 'HALLUCINATION'], 'title': 'Risk Type', 'type': 'string'}, 'description': {'description': 'One-sentence explanation of the failure', 'title': 'Description', 'type': 'string'}, 'raw_text': {'default': '', 'description': 'Exact snippet of trace text that evidences the failure (can be left blank and recovered via raw_text_ref)', 'title': 'Raw Text', 'type': 'string'}, 'raw_text_ref': {'description': 'List of references to every occurrence of the failure evidence in the trace', 'items': {'$ref': '#/$defs/ContentReference'}, 'title': 'Raw Text Ref', 'type': 'array'}, 'affected_id': {'anyOf': [{'type': 'string'}, {'type': 'null'}], 'default': None, 'description': 'ID of related Entity or Relation responsible for or impacted by the failure', 'title': 'Affected Id'}}, 'required': ['risk_type', 'description', 'raw_text_ref'], 'title': 'Failure', 'type': 'object'}, 'Relation': {'properties': {'id': {'description': 'Unique identifier for the relation', 'title': 'Id', 'type': 'string'}, 'source': {'description': 'ID of the source entity', 'title': 'Source', 'type': 'string'}, 'target': {'description': 'ID of the target entity', 'title': 'Target', 'type': 'string'}, 'type': {'description': 'Type of relation (only predefined types are allowed)', 'enum': ['CONSUMED_BY', 'PERFORMS', 'ASSIGNED_TO', 'USES', 'REQUIRED_BY', 'SUBTASK_OF', 'NEXT', 'PRODUCES', 'DELIVERS_TO', 'INTERVENES'], 'title': 'Type', 'type': 'string'}, 'importance': {'description': 'Importance level of this relationship in the system. HIGH: Critical data flows, core agent-task assignments, essential tool usage. MEDIUM: Standard workflows, common interactions, regular data processing. LOW: Auxiliary connections, optional steps, rarely activated relationships.', 'enum': ['HIGH', 'MEDIUM', 'LOW'], 'title': 'Importance', 'type': 'string'}, 'interaction_prompt': {'default': '', 'description': "Actual runtime interaction message/log that shows this relationship occurring during execution. Contains the exact text from the trace where this interaction happened (e.g., 'Agent started task X', 'Calling tool Y with parameters Z', 'User provided feedback: ABC'). This is NOT the static prompt definition but the dynamic interaction evidence.", 'title': 'Interaction Prompt', 'type': 'string'}, 'interaction_prompt_ref': {'description': 'List of references to the locations of interaction prompt content in the original trace. Enables mapping back to all occurrences of the interaction prompt.', 'items': {'$ref': '#/$defs/ContentReference'}, 'title': 'Interaction Prompt Ref', 'type': 'array'}}, 'required': ['source', 'target', 'type', 'importance'], 'title': 'Relation', 'type': 'object'}}, 'properties': {'entities': {'description': 'List of entities in the knowledge graph', 'items': {'$ref': '#/$defs/Entity'}, 'title': 'Entities', 'type': 'array'}, 'relations': {'description': 'List of relations in the knowledge graph', 'items': {'$ref': '#/$defs/Relation'}, 'title': 'Relations', 'type': 'array'}, 'failures': {'description': 'List of detected risk or failure events across the trace', 'items': {'$ref': '#/$defs/Failure'}, 'title': 'Failures', 'type': 'array'}, 'system_name': {'default': '', 'description': 'A concise, descriptive name for the agent system', 'title': 'System Name', 'type': 'string'}, 'system_summary': {'default': '', 'description': "A short 2-3 sentence summary of the agent system's purpose and structure", 'title': 'System Summary', 'type': 'string'}}, 'type': 'object', 'required': ['entities', 'failures', 'relations']}}}], 'tool_choice': {'type': 'function', 'function': {'name': 'KnowledgeGraph'}}} -2025-07-24 17:49:00,358 - instructor - DEBUG - max_retries: 3 -2025-07-24 17:49:00,358 - instructor - DEBUG - Retrying, attempt: 1 -2025-07-24 17:49:02,134 - instructor - DEBUG - Instructor Raw Response: ModelResponse(id='chatcmpl-BwtPoKBEsF7XmmJfhn9hyiIzOYd8b', created=1753375740, model='gpt-4o-mini-2024-07-18', object='chat.completion', system_fingerprint=None, choices=[Choices(finish_reason='stop', index=0, message=Message(content=None, role='assistant', tool_calls=[ChatCompletionMessageToolCall(function=Function(arguments='{"entities":[{"id":"agent_001","type":"Agent","name":"Data Analyst","importance":"HIGH","raw_prompt":"","raw_prompt_ref":[{"line_start":2,"line_end":2,"confidence":1}]}]}', name='KnowledgeGraph'), id='call_ljKJ51SkEByHQTQ9itUnMgsv', type='function')], function_call=None, provider_specific_fields={'refusal': None, 'annotations': []}, refusal=None, annotations=[]))], usage=CompletionUsage(completion_tokens=47, prompt_tokens=4205, total_tokens=4252, completion_tokens_details=CompletionTokensDetails(accepted_prediction_tokens=None, audio_tokens=0, reasoning_tokens=0, rejected_prediction_tokens=None), prompt_tokens_details=PromptTokensDetails(audio_tokens=0, cached_tokens=0)), service_tier='default') -2025-07-24 17:49:02,145 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Result Generated (production method) -2025-07-24 17:49:02,145 - agentgraph.extraction.graph_processing.knowledge_graph_processor - WARNING - Failed to resolve content references for window 0: 1 validation error for Entity -type - Input should be 'Agent', 'Task', 'Tool', 'Input', 'Output' or 'Human' [type=literal_error, input_value='Flow', input_type=str] - For further information visit https://errors.pydantic.dev/2.11/v/literal_error -2025-07-24 17:49:02,146 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Completed sub-batch 1/1 -2025-07-24 17:49:02,146 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Preparing window knowledge graphs with metadata -2025-07-24 17:49:02,146 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Merging knowledge graphs... -2025-07-24 17:49:02,146 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Only one knowledge graph generated, skipping merge process -2025-07-24 17:49:02,146 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing complete. Knowledge graph generated with 7 entities and 6 relations -2025-07-24 17:49:02,147 - __main__ - INFO - Processing text 3/3: text_2 -2025-07-24 17:49:02,147 - agentgraph.input.text_processing.trace_line_processor - INFO - Added line numbers to 7 lines, starting from line 1 -2025-07-24 17:49:02,147 - __main__ - INFO - Added line numbers: 7 lines -2025-07-24 17:49:02,147 - __main__ - INFO - Processing text text_2 (format: auto) -2025-07-24 17:49:02,147 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 17:49:02,147 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 469 characters -2025-07-24 17:49:02,147 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 17:49:02,147 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 17:49:02,155 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 17:49:02,155 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 17:49:02,156 - agentgraph.input.content_analysis.semantic_analyzer - INFO - Detected agent trace type: unknown (confidence: 0.00) -2025-07-24 17:49:02,156 - agentgraph.input.text_processing.chunking_service - INFO - Applied rule-based line splitting to 1 chunks (max_line_length=800) -2025-07-24 17:49:02,156 - agentgraph.input.text_processing.chunking_service - INFO - Content already has line numbers, skipping line number assignment -2025-07-24 17:49:02,156 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using agent_semantic splitter -2025-07-24 17:49:02,156 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=350000, overlap_size=17500 -2025-07-24 17:49:02,157 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context enhancement complete: 0 total documents -2025-07-24 17:49:02,157 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - SlidingWindowMonitor initialized with model: gpt-4o-mini, method: production -2025-07-24 17:49:02,157 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context documents: 0 provided -2025-07-24 17:49:02,157 - agentgraph.extraction.graph_utilities.knowledge_graph_merger - INFO - KnowledgeGraphMerger initialized with model: gpt-4o-mini -2025-07-24 17:49:02,159 - __main__ - INFO - Extracting knowledge graph from 1 chunks -2025-07-24 17:49:02,159 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks (splitter: agent_semantic, window_size=350000, overlap=0) -2025-07-24 17:49:02,159 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided source trace ID: text_2 -2025-07-24 17:49:02,159 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided processing run ID: batch_text_2 -2025-07-24 17:49:02,159 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks -2025-07-24 17:49:02,159 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using parallel processing with batch size 3 (1 batches) -2025-07-24 17:49:02,159 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 windows in 1 batches -2025-07-24 17:49:02,159 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks in 1 sub-batches (max 1 concurrent per sub-batch) -2025-07-24 17:49:02,159 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Starting sub-batch 1/1 with 1 chunks -2025-07-24 17:49:02,160 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing window 0 -2025-07-24 17:49:02,160 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - 🤖 Starting async_crew extraction for window 0 with method: production using model: gpt-4o-mini -2025-07-24 17:49:02,160 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - No context documents available for window 0, passing empty context string -2025-07-24 17:49:02,160 - agentgraph.methods.production.multi_agent_knowledge_extractor - INFO - Creating agent monitoring crew with model: gpt-4o-mini -2025-07-24 17:49:38,165 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Result Generated (production method) -2025-07-24 17:49:38,166 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 7 lines -2025-07-24 17:49:38,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:49:38,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:49:38,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L4-L4 -2025-07-24 17:49:38,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L4-L4 -2025-07-24 17:49:38,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 3, end_idx: 4 -2025-07-24 17:49:38,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:49:38,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:49:38,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: System Prompt: You are a helpful custome... -2025-07-24 17:49:38,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: System Prompt: You are a helpful custome... -2025-07-24 17:49:38,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: System Prompt: You are a helpful custome... -2025-07-24 17:49:38,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:49:38,166 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity agent_001 resolution debug: -2025-07-24 17:49:38,166 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:49:38,166 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L4-L4 -2025-07-24 17:49:38,166 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:49:38,166 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: System Prompt: You are a helpful custome... -2025-07-24 17:49:38,166 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity agent_001: 119 characters -2025-07-24 17:49:38,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:49:38,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:49:38,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L5-L5 -2025-07-24 17:49:38,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L5-L5 -2025-07-24 17:49:38,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 4, end_idx: 5 -2025-07-24 17:49:38,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:49:38,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:49:38,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Tool: order_lookup - searches order data... -2025-07-24 17:49:38,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Tool: order_lookup - searches order data... -2025-07-24 17:49:38,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Tool: order_lookup - searches order data... -2025-07-24 17:49:38,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:49:38,166 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity tool_001 resolution debug: -2025-07-24 17:49:38,166 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:49:38,166 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L5-L5 -2025-07-24 17:49:38,166 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:49:38,166 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Tool: order_lookup - searches order data... -2025-07-24 17:49:38,166 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity tool_001: 54 characters -2025-07-24 17:49:38,166 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L6-L6 -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L6-L6 -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 5, end_idx: 6 -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Tool: email_sender - sends emails to cus... -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Tool: email_sender - sends emails to cus... -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Tool: email_sender - sends emails to cus... -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:49:38,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity tool_002 resolution debug: -2025-07-24 17:49:38,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:49:38,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L6-L6 -2025-07-24 17:49:38,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:49:38,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Tool: email_sender - sends emails to cus... -2025-07-24 17:49:38,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity tool_002: 56 characters -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L7-L7 -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L7-L7 -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 6, end_idx: 7 -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Process: Agent uses order_lookup to find... -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Process: Agent uses order_lookup to find... -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Process: Agent uses order_lookup to find... -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:49:38,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity task_001 resolution debug: -2025-07-24 17:49:38,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:49:38,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L7-L7 -2025-07-24 17:49:38,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:49:38,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: Process: Agent uses order_lookup to find... -2025-07-24 17:49:38,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity task_001: 136 characters -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L2-L2 -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L2-L2 -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 1, end_idx: 2 -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: User Input: I need help with my order... -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: User Input: I need help with my order... -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: User Input: I need help with my order... -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:49:38,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity input_001 resolution debug: -2025-07-24 17:49:38,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 17:49:38,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L2-L2 -2025-07-24 17:49:38,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 17:49:38,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: User Input: I need help with my order... -2025-07-24 17:49:38,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity input_001: 47 characters -2025-07-24 17:49:38,167 - agentgraph.reconstruction.content_reference_resolver - INFO - Entity prompt resolution stats: {'total_entities': 5, 'entities_with_refs': 5, 'successful_resolutions': 5, 'failed_resolutions': 0} -2025-07-24 17:49:38,167 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 7 lines -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L2-L2 -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L2-L2 -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 1, end_idx: 2 -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: User Input: I need help with my order... -2025-07-24 17:49:38,167 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: User Input: I need help with my order... -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: User Input: I need help with my order... -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:49:38,168 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation relation_001: 47 characters -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L7-L7 -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L7-L7 -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 6, end_idx: 7 -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Process: Agent uses order_lookup to find... -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Process: Agent uses order_lookup to find... -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Process: Agent uses order_lookup to find... -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:49:38,168 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation relation_002: 136 characters -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L5-L5 -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L5-L5 -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 4, end_idx: 5 -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Tool: order_lookup - searches order data... -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Tool: order_lookup - searches order data... -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Tool: order_lookup - searches order data... -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:49:38,168 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation relation_003: 54 characters -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 7 -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L6-L6 -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L6-L6 -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 5, end_idx: 6 -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 7 -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: Tool: email_sender - sends emails to cus... -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: Tool: email_sender - sends emails to cus... -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: Tool: email_sender - sends emails to cus... -2025-07-24 17:49:38,168 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 17:49:38,168 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation relation_004: 56 characters -2025-07-24 17:49:38,168 - agentgraph.reconstruction.content_reference_resolver - INFO - Relation prompt resolution stats: {'total_relations': 4, 'relations_with_refs': 4, 'successful_resolutions': 4, 'failed_resolutions': 0} -2025-07-24 17:49:38,168 - agentgraph.reconstruction.content_reference_resolver - INFO - Resolved content references for knowledge graph with 5 entities and 4 relations -2025-07-24 17:49:38,168 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Resolved content references for window 0 -2025-07-24 17:49:38,168 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Completed sub-batch 1/1 -2025-07-24 17:49:38,169 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Preparing window knowledge graphs with metadata -2025-07-24 17:49:38,169 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Merging knowledge graphs... -2025-07-24 17:49:38,169 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Only one knowledge graph generated, skipping merge process -2025-07-24 17:49:38,169 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing complete. Knowledge graph generated with 5 entities and 4 relations -2025-07-24 17:49:38,171 - __main__ - INFO - Results saved to test_results_fixed.json -2025-07-24 18:06:41,465 - openlit - INFO - Starting openLIT initialization... -2025-07-24 18:06:41,470 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 18:06:42,052 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 18:06:42,085 - openlit - INFO - Library for ollama (ollama) not found. Skipping instrumentation -2025-07-24 18:06:42,085 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 18:06:42,085 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 18:06:42,085 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 18:06:42,381 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 18:06:42,555 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 18:06:42,555 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 18:06:43,138 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 18:06:43,139 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 18:06:43,139 - openlit - INFO - Library for transformers (transformers) not found. Skipping instrumentation -2025-07-24 18:06:43,142 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 18:06:43,142 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 18:06:43,142 - openlit - INFO - Library for autogen (autogen) not found. Skipping instrumentation -2025-07-24 18:06:43,142 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 18:06:43,143 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 18:06:43,143 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 18:06:43,143 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 18:06:43,143 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 18:06:43,143 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 18:06:43,143 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 18:06:43,143 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 18:06:43,143 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 18:06:43,143 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 18:06:43,143 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 18:06:43,143 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 18:06:43,143 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 18:06:43,143 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 18:06:43,143 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 18:06:44,277 - __main__ - INFO - Loaded 10 texts from scripts/example.json -2025-07-24 18:06:44,277 - agentgraph.input.text_processing.chunking_service - INFO - ChunkingService initialized with batch_size=3, model=gpt-4o-mini -2025-07-24 18:06:44,277 - __main__ - INFO - Initialized BatchKGExtractor: -2025-07-24 18:06:44,277 - __main__ - INFO - - Model: gpt-4o-mini, Method: production -2025-07-24 18:06:44,277 - __main__ - INFO - - Preprocessing: False -2025-07-24 18:06:44,277 - __main__ - INFO - - Line numbers: True -2025-07-24 18:06:44,277 - __main__ - INFO - Processing batch of 10 texts -2025-07-24 18:06:44,277 - __main__ - INFO - Processing text 1/10: text_0 -2025-07-24 18:06:44,277 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:06:44,277 - __main__ - INFO - Processing text text_0 (format: auto) -2025-07-24 18:06:44,277 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:06:44,277 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:06:44,277 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:06:44,277 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:06:44,283 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:06:44,283 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:06:44,283 - __main__ - ERROR - Error processing text text_0: 'dict' object has no attribute 'strip' -2025-07-24 18:06:44,283 - __main__ - INFO - Processing text 2/10: text_1 -2025-07-24 18:06:44,283 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:06:44,283 - __main__ - INFO - Processing text text_1 (format: auto) -2025-07-24 18:06:44,283 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:06:44,283 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:06:44,283 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:06:44,283 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:06:44,288 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:06:44,288 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:06:44,288 - __main__ - ERROR - Error processing text text_1: 'dict' object has no attribute 'strip' -2025-07-24 18:06:44,288 - __main__ - INFO - Processing text 3/10: text_2 -2025-07-24 18:06:44,288 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:06:44,288 - __main__ - INFO - Processing text text_2 (format: auto) -2025-07-24 18:06:44,288 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:06:44,288 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:06:44,288 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:06:44,288 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:06:44,293 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:06:44,293 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:06:44,293 - __main__ - ERROR - Error processing text text_2: 'dict' object has no attribute 'strip' -2025-07-24 18:06:44,294 - __main__ - INFO - Processing text 4/10: text_3 -2025-07-24 18:06:44,294 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:06:44,294 - __main__ - INFO - Processing text text_3 (format: auto) -2025-07-24 18:06:44,294 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:06:44,294 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:06:44,294 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:06:44,294 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:06:44,301 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:06:44,301 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:06:44,301 - __main__ - ERROR - Error processing text text_3: 'dict' object has no attribute 'strip' -2025-07-24 18:06:44,301 - __main__ - INFO - Processing text 5/10: text_4 -2025-07-24 18:06:44,301 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:06:44,301 - __main__ - INFO - Processing text text_4 (format: auto) -2025-07-24 18:06:44,301 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:06:44,301 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:06:44,301 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:06:44,301 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:06:44,306 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:06:44,306 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:06:44,306 - __main__ - ERROR - Error processing text text_4: 'dict' object has no attribute 'strip' -2025-07-24 18:06:44,306 - __main__ - INFO - Processing text 6/10: text_5 -2025-07-24 18:06:44,306 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:06:44,306 - __main__ - INFO - Processing text text_5 (format: auto) -2025-07-24 18:06:44,306 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:06:44,306 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:06:44,306 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:06:44,306 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:06:44,443 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:06:44,443 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:06:44,443 - __main__ - ERROR - Error processing text text_5: 'dict' object has no attribute 'strip' -2025-07-24 18:06:44,443 - __main__ - INFO - Processing text 7/10: text_6 -2025-07-24 18:06:44,443 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:06:44,443 - __main__ - INFO - Processing text text_6 (format: auto) -2025-07-24 18:06:44,443 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:06:44,443 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:06:44,443 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:06:44,443 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:06:44,449 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:06:44,449 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:06:44,449 - __main__ - ERROR - Error processing text text_6: 'dict' object has no attribute 'strip' -2025-07-24 18:06:44,449 - __main__ - INFO - Processing text 8/10: text_7 -2025-07-24 18:06:44,449 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:06:44,449 - __main__ - INFO - Processing text text_7 (format: auto) -2025-07-24 18:06:44,449 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:06:44,449 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:06:44,449 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:06:44,449 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:06:44,454 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:06:44,454 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:06:44,454 - __main__ - ERROR - Error processing text text_7: 'dict' object has no attribute 'strip' -2025-07-24 18:06:44,454 - __main__ - INFO - Processing text 9/10: text_8 -2025-07-24 18:06:44,454 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:06:44,454 - __main__ - INFO - Processing text text_8 (format: auto) -2025-07-24 18:06:44,454 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:06:44,454 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:06:44,454 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:06:44,454 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:06:44,459 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:06:44,459 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:06:44,459 - __main__ - ERROR - Error processing text text_8: 'dict' object has no attribute 'strip' -2025-07-24 18:06:44,459 - __main__ - INFO - Processing text 10/10: text_9 -2025-07-24 18:06:44,459 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:06:44,459 - __main__ - INFO - Processing text text_9 (format: auto) -2025-07-24 18:06:44,459 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:06:44,459 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:06:44,459 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:06:44,459 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:06:44,464 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:06:44,464 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:06:44,464 - __main__ - ERROR - Error processing text text_9: 'dict' object has no attribute 'strip' -2025-07-24 18:06:44,465 - __main__ - INFO - Results saved to test_results_fixed.json -2025-07-24 18:08:07,996 - openlit - INFO - Starting openLIT initialization... -2025-07-24 18:08:08,011 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 18:08:08,600 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 18:08:08,650 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 18:08:08,650 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 18:08:08,651 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 18:08:08,995 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 18:08:09,104 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 18:08:09,104 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 18:08:09,703 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 18:08:09,704 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 18:08:11,134 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 18:08:11,136 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 18:08:11,136 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 18:08:11,137 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 18:08:11,137 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 18:08:11,137 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 18:08:11,137 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 18:08:11,137 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 18:08:11,137 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 18:08:11,137 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 18:08:11,137 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 18:08:11,137 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 18:08:11,137 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 18:08:11,137 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 18:08:11,137 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 18:08:11,138 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 18:08:11,138 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 18:08:11,138 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 18:08:11,138 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 18:08:12,187 - __main__ - INFO - Loaded 10 texts from scripts/example.json -2025-07-24 18:08:12,187 - agentgraph.input.text_processing.chunking_service - INFO - ChunkingService initialized with batch_size=3, model=gpt-4o-mini -2025-07-24 18:08:12,187 - __main__ - INFO - Initialized BatchKGExtractor: -2025-07-24 18:08:12,187 - __main__ - INFO - - Model: gpt-4o-mini, Method: production -2025-07-24 18:08:12,187 - __main__ - INFO - - Preprocessing: False -2025-07-24 18:08:12,187 - __main__ - INFO - - Line numbers: True -2025-07-24 18:08:12,187 - __main__ - INFO - Processing batch of 10 texts -2025-07-24 18:08:12,187 - __main__ - INFO - Processing text 1/10: text_0 -2025-07-24 18:08:12,187 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:08:12,187 - __main__ - INFO - Processing text text_0 (format: auto) -2025-07-24 18:08:12,187 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:08:12,187 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:08:12,187 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:08:12,188 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:08:12,193 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:08:12,193 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:08:12,193 - __main__ - ERROR - Error processing text text_0: 'dict' object has no attribute 'strip' -2025-07-24 18:08:12,194 - __main__ - INFO - Processing text 2/10: text_1 -2025-07-24 18:08:12,194 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:08:12,194 - __main__ - INFO - Processing text text_1 (format: auto) -2025-07-24 18:08:12,194 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:08:12,194 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:08:12,194 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:08:12,194 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:08:12,199 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:08:12,199 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:08:12,199 - __main__ - ERROR - Error processing text text_1: 'dict' object has no attribute 'strip' -2025-07-24 18:08:12,200 - __main__ - INFO - Processing text 3/10: text_2 -2025-07-24 18:08:12,200 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:08:12,200 - __main__ - INFO - Processing text text_2 (format: auto) -2025-07-24 18:08:12,200 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:08:12,200 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:08:12,200 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:08:12,200 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:08:12,205 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:08:12,205 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:08:12,205 - __main__ - ERROR - Error processing text text_2: 'dict' object has no attribute 'strip' -2025-07-24 18:08:12,206 - __main__ - INFO - Processing text 4/10: text_3 -2025-07-24 18:08:12,206 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:08:12,206 - __main__ - INFO - Processing text text_3 (format: auto) -2025-07-24 18:08:12,206 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:08:12,206 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:08:12,206 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:08:12,206 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:08:12,211 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:08:12,211 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:08:12,211 - __main__ - ERROR - Error processing text text_3: 'dict' object has no attribute 'strip' -2025-07-24 18:08:12,211 - __main__ - INFO - Processing text 5/10: text_4 -2025-07-24 18:08:12,211 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:08:12,211 - __main__ - INFO - Processing text text_4 (format: auto) -2025-07-24 18:08:12,212 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:08:12,212 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:08:12,212 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:08:12,212 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:08:12,216 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:08:12,216 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:08:12,217 - __main__ - ERROR - Error processing text text_4: 'dict' object has no attribute 'strip' -2025-07-24 18:08:12,217 - __main__ - INFO - Processing text 6/10: text_5 -2025-07-24 18:08:12,217 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:08:12,217 - __main__ - INFO - Processing text text_5 (format: auto) -2025-07-24 18:08:12,217 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:08:12,217 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:08:12,217 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:08:12,217 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:08:12,222 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:08:12,222 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:08:12,222 - __main__ - ERROR - Error processing text text_5: 'dict' object has no attribute 'strip' -2025-07-24 18:08:12,223 - __main__ - INFO - Processing text 7/10: text_6 -2025-07-24 18:08:12,223 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:08:12,223 - __main__ - INFO - Processing text text_6 (format: auto) -2025-07-24 18:08:12,223 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:08:12,223 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:08:12,223 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:08:12,223 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:08:12,228 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:08:12,228 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:08:12,228 - __main__ - ERROR - Error processing text text_6: 'dict' object has no attribute 'strip' -2025-07-24 18:08:12,228 - __main__ - INFO - Processing text 8/10: text_7 -2025-07-24 18:08:12,228 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:08:12,228 - __main__ - INFO - Processing text text_7 (format: auto) -2025-07-24 18:08:12,228 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:08:12,228 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:08:12,228 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:08:12,228 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:08:12,233 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:08:12,233 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:08:12,233 - __main__ - ERROR - Error processing text text_7: 'dict' object has no attribute 'strip' -2025-07-24 18:08:12,234 - __main__ - INFO - Processing text 9/10: text_8 -2025-07-24 18:08:12,234 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:08:12,234 - __main__ - INFO - Processing text text_8 (format: auto) -2025-07-24 18:08:12,234 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:08:12,234 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:08:12,234 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:08:12,234 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:08:12,238 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:08:12,239 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:08:12,239 - __main__ - ERROR - Error processing text text_8: 'dict' object has no attribute 'strip' -2025-07-24 18:08:12,239 - __main__ - INFO - Processing text 10/10: text_9 -2025-07-24 18:08:12,239 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:08:12,239 - __main__ - INFO - Processing text text_9 (format: auto) -2025-07-24 18:08:12,239 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:08:12,239 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:08:12,239 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:08:12,239 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:08:12,244 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:08:12,244 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:08:12,244 - __main__ - ERROR - Error processing text text_9: 'dict' object has no attribute 'strip' -2025-07-24 18:08:12,245 - __main__ - INFO - Results saved to test_debug.json -2025-07-24 18:08:38,423 - openlit - INFO - Starting openLIT initialization... -2025-07-24 18:08:38,439 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 18:08:39,018 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 18:08:39,065 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 18:08:39,065 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 18:08:39,065 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 18:08:39,410 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 18:08:39,514 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 18:08:39,515 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 18:08:40,068 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 18:08:40,069 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 18:08:41,338 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 18:08:41,340 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 18:08:41,341 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 18:08:41,341 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 18:08:41,341 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 18:08:41,341 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 18:08:41,341 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 18:08:41,341 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 18:08:41,341 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 18:08:41,341 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 18:08:41,341 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 18:08:41,341 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 18:08:41,342 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 18:08:41,342 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 18:08:41,342 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 18:08:41,342 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 18:08:41,342 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 18:08:41,342 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 18:08:41,342 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 18:08:42,255 - __main__ - INFO - Loaded 10 texts from scripts/example.json -2025-07-24 18:08:42,255 - agentgraph.input.text_processing.chunking_service - INFO - ChunkingService initialized with batch_size=3, model=gpt-4o-mini -2025-07-24 18:08:42,255 - __main__ - INFO - Initialized BatchKGExtractor: -2025-07-24 18:08:42,255 - __main__ - INFO - - Model: gpt-4o-mini, Method: production -2025-07-24 18:08:42,255 - __main__ - INFO - - Preprocessing: False -2025-07-24 18:08:42,255 - __main__ - INFO - - Line numbers: True -2025-07-24 18:08:42,255 - __main__ - INFO - Processing batch of 10 texts -2025-07-24 18:08:42,255 - __main__ - INFO - Processing text 1/10: text_0 -2025-07-24 18:08:42,255 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:08:42,255 - __main__ - INFO - Processing text text_0 (format: auto) -2025-07-24 18:08:42,255 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:08:42,255 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:08:42,255 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:08:42,255 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:08:42,261 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:08:42,261 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:08:42,261 - __main__ - ERROR - Error processing text text_0: 'dict' object has no attribute 'strip' -2025-07-24 18:08:42,261 - __main__ - INFO - Processing text 2/10: text_1 -2025-07-24 18:08:42,261 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:08:42,261 - __main__ - INFO - Processing text text_1 (format: auto) -2025-07-24 18:08:42,261 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:08:42,261 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:08:42,261 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:08:42,261 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:08:42,267 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:08:42,267 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:08:42,267 - __main__ - ERROR - Error processing text text_1: 'dict' object has no attribute 'strip' -2025-07-24 18:08:42,267 - __main__ - INFO - Processing text 3/10: text_2 -2025-07-24 18:08:42,267 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:08:42,268 - __main__ - INFO - Processing text text_2 (format: auto) -2025-07-24 18:08:42,268 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:08:42,268 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:08:42,268 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:08:42,268 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:08:42,272 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:08:42,272 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:08:42,272 - __main__ - ERROR - Error processing text text_2: 'dict' object has no attribute 'strip' -2025-07-24 18:08:42,273 - __main__ - INFO - Processing text 4/10: text_3 -2025-07-24 18:08:42,273 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:08:42,273 - __main__ - INFO - Processing text text_3 (format: auto) -2025-07-24 18:08:42,273 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:08:42,273 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:08:42,273 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:08:42,273 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:08:42,278 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:08:42,278 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:08:42,278 - __main__ - ERROR - Error processing text text_3: 'dict' object has no attribute 'strip' -2025-07-24 18:08:42,279 - __main__ - INFO - Processing text 5/10: text_4 -2025-07-24 18:08:42,279 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:08:42,279 - __main__ - INFO - Processing text text_4 (format: auto) -2025-07-24 18:08:42,279 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:08:42,279 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:08:42,279 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:08:42,279 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:08:42,284 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:08:42,284 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:08:42,284 - __main__ - ERROR - Error processing text text_4: 'dict' object has no attribute 'strip' -2025-07-24 18:08:42,284 - __main__ - INFO - Processing text 6/10: text_5 -2025-07-24 18:08:42,285 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:08:42,285 - __main__ - INFO - Processing text text_5 (format: auto) -2025-07-24 18:08:42,285 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:08:42,285 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:08:42,285 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:08:42,285 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:08:42,289 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:08:42,289 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:08:42,289 - __main__ - ERROR - Error processing text text_5: 'dict' object has no attribute 'strip' -2025-07-24 18:08:42,290 - __main__ - INFO - Processing text 7/10: text_6 -2025-07-24 18:08:42,290 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:08:42,290 - __main__ - INFO - Processing text text_6 (format: auto) -2025-07-24 18:08:42,290 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:08:42,290 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:08:42,290 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:08:42,290 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:08:42,295 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:08:42,295 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:08:42,295 - __main__ - ERROR - Error processing text text_6: 'dict' object has no attribute 'strip' -2025-07-24 18:08:42,296 - __main__ - INFO - Processing text 8/10: text_7 -2025-07-24 18:08:42,296 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:08:42,296 - __main__ - INFO - Processing text text_7 (format: auto) -2025-07-24 18:08:42,296 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:08:42,296 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:08:42,296 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:08:42,296 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:08:42,301 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:08:42,301 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:08:42,301 - __main__ - ERROR - Error processing text text_7: 'dict' object has no attribute 'strip' -2025-07-24 18:08:42,301 - __main__ - INFO - Processing text 9/10: text_8 -2025-07-24 18:08:42,301 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:08:42,301 - __main__ - INFO - Processing text text_8 (format: auto) -2025-07-24 18:08:42,301 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:08:42,301 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:08:42,301 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:08:42,301 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:08:42,306 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:08:42,306 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:08:42,306 - __main__ - ERROR - Error processing text text_8: 'dict' object has no attribute 'strip' -2025-07-24 18:08:42,307 - __main__ - INFO - Processing text 10/10: text_9 -2025-07-24 18:08:42,307 - __main__ - WARNING - Line number processing failed: 'dict' object has no attribute 'split', using original text -2025-07-24 18:08:42,307 - __main__ - INFO - Processing text text_9 (format: auto) -2025-07-24 18:08:42,307 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:08:42,307 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 13 characters -2025-07-24 18:08:42,307 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:08:42,307 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:08:42,312 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:08:42,312 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:08:42,312 - __main__ - ERROR - Error processing text text_9: 'dict' object has no attribute 'strip' -2025-07-24 18:08:52,697 - openlit - INFO - Starting openLIT initialization... -2025-07-24 18:08:52,713 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 18:08:53,230 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 18:08:53,269 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 18:08:53,269 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 18:08:53,269 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 18:08:53,580 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 18:08:53,672 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 18:08:53,672 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 18:08:54,154 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 18:08:54,154 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 18:08:55,313 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 18:08:55,315 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 18:08:55,315 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 18:08:55,315 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 18:08:55,315 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 18:08:55,315 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 18:08:55,315 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 18:08:55,315 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 18:08:55,316 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 18:08:55,316 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 18:08:55,316 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 18:08:55,316 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 18:08:55,316 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 18:08:55,316 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 18:08:55,316 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 18:08:55,316 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 18:08:55,316 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 18:08:55,316 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 18:08:55,317 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 18:09:32,835 - openlit - INFO - Starting openLIT initialization... -2025-07-24 18:09:32,853 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 18:09:33,361 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 18:09:33,403 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 18:09:33,403 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 18:09:33,403 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 18:09:33,757 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 18:09:33,851 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 18:09:33,851 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 18:09:34,416 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 18:09:34,417 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 18:09:35,721 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 18:09:35,723 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 18:09:35,723 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 18:09:35,724 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 18:09:35,724 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 18:09:35,724 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 18:09:35,724 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 18:09:35,724 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 18:09:35,724 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 18:09:35,724 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 18:09:35,724 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 18:09:35,724 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 18:09:35,724 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 18:09:35,725 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 18:09:35,725 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 18:09:35,725 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 18:09:35,725 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 18:09:35,725 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 18:09:35,725 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 18:09:36,667 - scripts.batch_kg_extraction - INFO - Loaded JSON array with 10 items -2025-07-24 18:09:46,781 - openlit - INFO - Starting openLIT initialization... -2025-07-24 18:09:46,798 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 18:09:47,280 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 18:09:47,320 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 18:09:47,320 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 18:09:47,320 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 18:09:47,636 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 18:09:47,727 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 18:09:47,727 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 18:09:48,277 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 18:09:48,277 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 18:09:49,552 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 18:09:49,554 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 18:09:49,555 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 18:09:49,555 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 18:09:49,555 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 18:09:49,555 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 18:09:49,555 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 18:09:49,555 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 18:09:49,555 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 18:09:49,555 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 18:09:49,555 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 18:09:49,556 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 18:09:49,556 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 18:09:49,556 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 18:09:49,556 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 18:09:49,556 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 18:09:49,556 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 18:09:49,556 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 18:09:49,556 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 18:09:50,514 - __main__ - INFO - Loaded JSON array with 10 items -2025-07-24 18:09:50,514 - __main__ - INFO - Loaded 10 texts from scripts/example.json -2025-07-24 18:09:50,514 - agentgraph.input.text_processing.chunking_service - INFO - ChunkingService initialized with batch_size=3, model=gpt-4o-mini -2025-07-24 18:09:50,514 - __main__ - INFO - Initialized BatchKGExtractor: -2025-07-24 18:09:50,514 - __main__ - INFO - - Model: gpt-4o-mini, Method: production -2025-07-24 18:09:50,514 - __main__ - INFO - - Preprocessing: True -2025-07-24 18:09:50,514 - __main__ - INFO - - Line numbers: True -2025-07-24 18:09:50,514 - __main__ - INFO - Processing batch of 10 texts -2025-07-24 18:09:50,514 - __main__ - INFO - Processing text 1/10: text_0 -2025-07-24 18:09:50,514 - __main__ - WARNING - Trace preprocessing failed: 6 validation errors for LangSmithTrace -trace_id - Field required [type=missing, input_value={'name': 'ReAct Agent', '... 'feedback_stats': None}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -trace_name - Field required [type=missing, input_value={'name': 'ReAct Agent', '... 'feedback_stats': None}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -project_name - Field required [type=missing, input_value={'name': 'ReAct Agent', '... 'feedback_stats': None}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -export_time - Field required [type=missing, input_value={'name': 'ReAct Agent', '... 'feedback_stats': None}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -total_runs - Field required [type=missing, input_value={'name': 'ReAct Agent', '... 'feedback_stats': None}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing -runs - Field required [type=missing, input_value={'name': 'ReAct Agent', '... 'feedback_stats': None}, input_type=dict] - For further information visit https://errors.pydantic.dev/2.11/v/missing, using original text -2025-07-24 18:09:50,515 - agentgraph.input.text_processing.trace_line_processor - INFO - Added line numbers to 239 lines, starting from line 1 -2025-07-24 18:09:50,515 - __main__ - INFO - Added line numbers: 239 lines -2025-07-24 18:09:50,515 - __main__ - INFO - Processing text text_0 (format: langsmith) -2025-07-24 18:09:50,515 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:09:50,515 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 16247 characters -2025-07-24 18:09:50,515 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:09:50,515 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:09:50,521 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:09:50,521 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:09:50,526 - agentgraph.input.content_analysis.semantic_analyzer - INFO - Detected agent trace type: unknown (confidence: 0.00) -2025-07-24 18:09:51,448 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Generated embeddings for 10 sentences using text-embedding-3-small with smart batching -2025-07-24 18:09:51,454 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Detected agent trace type: unknown (confidence: 0.00) -2025-07-24 18:09:51,454 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Unknown format detected - using data-driven threshold selection -2025-07-24 18:09:51,454 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Highly coherent content detected, using gradient threshold -2025-07-24 18:09:51,454 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Using adaptive threshold: 0.722 for 9 similarities -2025-07-24 18:09:51,456 - agentgraph.input.text_processing.chunking_service - INFO - Applied rule-based line splitting to 1 chunks (max_line_length=800) -2025-07-24 18:09:51,457 - agentgraph.input.text_processing.chunking_service - INFO - Content already has line numbers, skipping line number assignment -2025-07-24 18:09:51,457 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using agent_semantic splitter -2025-07-24 18:09:51,457 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=350000, overlap_size=17500 -2025-07-24 18:09:51,457 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context enhancement complete: 0 total documents -2025-07-24 18:09:51,457 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - SlidingWindowMonitor initialized with model: gpt-4o-mini, method: production -2025-07-24 18:09:51,457 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context documents: 0 provided -2025-07-24 18:09:51,457 - agentgraph.extraction.graph_utilities.knowledge_graph_merger - INFO - KnowledgeGraphMerger initialized with model: gpt-4o-mini -2025-07-24 18:09:51,459 - __main__ - INFO - Extracting knowledge graph from 1 chunks -2025-07-24 18:09:51,459 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks (splitter: agent_semantic, window_size=350000, overlap=0) -2025-07-24 18:09:51,459 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided source trace ID: text_0 -2025-07-24 18:09:51,459 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided processing run ID: batch_text_0 -2025-07-24 18:09:51,459 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks -2025-07-24 18:09:51,459 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using parallel processing with batch size 3 (1 batches) -2025-07-24 18:09:51,459 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 windows in 1 batches -2025-07-24 18:09:51,472 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks in 1 sub-batches (max 1 concurrent per sub-batch) -2025-07-24 18:09:51,472 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Starting sub-batch 1/1 with 1 chunks -2025-07-24 18:09:51,473 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing window 0 -2025-07-24 18:09:51,473 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - 🤖 Starting async_crew extraction for window 0 with method: production using model: gpt-4o-mini -2025-07-24 18:09:51,473 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - No context documents available for window 0, passing empty context string -2025-07-24 18:09:51,473 - agentgraph.methods.production.multi_agent_knowledge_extractor - INFO - Creating agent monitoring crew with model: gpt-4o-mini -2025-07-24 18:11:19,749 - openlit - INFO - Starting openLIT initialization... -2025-07-24 18:11:19,764 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 18:11:20,338 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 18:11:20,385 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 18:11:20,385 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 18:11:20,386 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 18:11:20,728 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 18:11:20,838 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 18:11:20,839 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 18:11:21,383 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 18:11:21,384 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 18:11:22,642 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 18:11:22,644 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 18:11:22,644 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 18:11:22,644 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 18:11:22,644 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 18:11:22,645 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 18:11:22,645 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 18:11:22,645 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 18:11:22,645 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 18:11:22,645 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 18:11:22,645 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 18:11:22,645 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 18:11:22,645 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 18:11:22,645 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 18:11:22,645 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 18:11:22,646 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 18:11:22,646 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 18:11:22,646 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 18:11:22,646 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 18:11:23,662 - scripts.batch_kg_extraction - INFO - Combined 3 texts from 'texts' key into single text -2025-07-24 18:11:23,663 - scripts.batch_kg_extraction - INFO - Combined 10 array items into single text -2025-07-24 18:11:35,400 - openlit - INFO - Starting openLIT initialization... -2025-07-24 18:11:35,417 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 18:11:35,997 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 18:11:36,046 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 18:11:36,046 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 18:11:36,046 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 18:11:36,406 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 18:11:36,514 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 18:11:36,514 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 18:11:37,091 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 18:11:37,092 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 18:11:38,504 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 18:11:38,509 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 18:11:38,509 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 18:11:38,509 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 18:11:38,510 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 18:11:38,510 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 18:11:38,510 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 18:11:38,510 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 18:11:38,510 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 18:11:38,510 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 18:11:38,510 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 18:11:38,510 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 18:11:38,510 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 18:11:38,510 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 18:11:38,510 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 18:11:38,510 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 18:11:38,511 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 18:11:38,511 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 18:11:38,511 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 18:11:39,537 - __main__ - INFO - Combined 3 texts from 'texts' key into single text -2025-07-24 18:11:39,537 - __main__ - INFO - Loaded 1 texts from scripts/example_texts.json -2025-07-24 18:11:39,537 - agentgraph.input.text_processing.chunking_service - INFO - ChunkingService initialized with batch_size=3, model=gpt-4o-mini -2025-07-24 18:11:39,537 - __main__ - INFO - Initialized BatchKGExtractor: -2025-07-24 18:11:39,537 - __main__ - INFO - - Model: gpt-4o-mini, Method: production -2025-07-24 18:11:39,537 - __main__ - INFO - - Preprocessing: False -2025-07-24 18:11:39,537 - __main__ - INFO - - Line numbers: True -2025-07-24 18:11:39,537 - __main__ - INFO - Processing batch of 1 texts -2025-07-24 18:11:39,537 - __main__ - INFO - Processing text 1/1: text_0 -2025-07-24 18:11:39,538 - agentgraph.input.text_processing.trace_line_processor - INFO - Added line numbers to 26 lines, starting from line 1 -2025-07-24 18:11:39,538 - __main__ - INFO - Added line numbers: 26 lines -2025-07-24 18:11:39,538 - __main__ - INFO - Processing text text_0 (format: auto) -2025-07-24 18:11:39,538 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:11:39,538 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 1435 characters -2025-07-24 18:11:39,538 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:11:39,538 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:11:39,544 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:11:39,544 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:11:39,546 - agentgraph.input.content_analysis.semantic_analyzer - INFO - Detected agent trace type: unknown (confidence: 0.00) -2025-07-24 18:11:39,547 - agentgraph.input.text_processing.chunking_service - INFO - Applied rule-based line splitting to 1 chunks (max_line_length=800) -2025-07-24 18:11:39,548 - agentgraph.input.text_processing.chunking_service - INFO - Content already has line numbers, skipping line number assignment -2025-07-24 18:11:39,548 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using agent_semantic splitter -2025-07-24 18:11:39,548 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=350000, overlap_size=17500 -2025-07-24 18:11:39,548 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context enhancement complete: 0 total documents -2025-07-24 18:11:39,548 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - SlidingWindowMonitor initialized with model: gpt-4o-mini, method: production -2025-07-24 18:11:39,548 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context documents: 0 provided -2025-07-24 18:11:39,548 - agentgraph.extraction.graph_utilities.knowledge_graph_merger - INFO - KnowledgeGraphMerger initialized with model: gpt-4o-mini -2025-07-24 18:11:39,550 - __main__ - INFO - Extracting knowledge graph from 1 chunks -2025-07-24 18:11:39,550 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks (splitter: agent_semantic, window_size=350000, overlap=0) -2025-07-24 18:11:39,550 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided source trace ID: text_0 -2025-07-24 18:11:39,550 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided processing run ID: batch_text_0 -2025-07-24 18:11:39,550 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks -2025-07-24 18:11:39,550 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using parallel processing with batch size 3 (1 batches) -2025-07-24 18:11:39,550 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 windows in 1 batches -2025-07-24 18:11:39,562 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks in 1 sub-batches (max 1 concurrent per sub-batch) -2025-07-24 18:11:39,563 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Starting sub-batch 1/1 with 1 chunks -2025-07-24 18:11:39,563 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing window 0 -2025-07-24 18:11:39,564 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - 🤖 Starting async_crew extraction for window 0 with method: production using model: gpt-4o-mini -2025-07-24 18:11:39,564 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - No context documents available for window 0, passing empty context string -2025-07-24 18:11:39,564 - agentgraph.methods.production.multi_agent_knowledge_extractor - INFO - Creating agent monitoring crew with model: gpt-4o-mini -2025-07-24 18:12:41,267 - asyncio - ERROR - _GatheringFuture exception was never retrieved -future: <_GatheringFuture finished exception=CancelledError()> -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/agentgraph/extraction/graph_processing/knowledge_graph_processor.py", line 238, in process_window - result = await asyncio.wait_for( - ^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/asyncio/tasks.py", line 476, in wait_for - await waiter -asyncio.exceptions.CancelledError -2025-07-24 18:13:23,105 - openlit - INFO - Starting openLIT initialization... -2025-07-24 18:13:23,112 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 18:13:23,519 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 18:13:23,548 - openlit - INFO - Library for ollama (ollama) not found. Skipping instrumentation -2025-07-24 18:13:23,548 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 18:13:23,548 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 18:13:23,548 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 18:13:23,806 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 18:13:23,990 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 18:13:23,990 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 18:13:24,521 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 18:13:24,522 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 18:13:24,522 - openlit - INFO - Library for transformers (transformers) not found. Skipping instrumentation -2025-07-24 18:13:24,525 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 18:13:24,525 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 18:13:24,525 - openlit - INFO - Library for autogen (autogen) not found. Skipping instrumentation -2025-07-24 18:13:24,525 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 18:13:24,525 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 18:13:24,525 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 18:13:24,525 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 18:13:24,525 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 18:13:24,525 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 18:13:24,526 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 18:13:24,526 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 18:13:24,526 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 18:13:24,526 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 18:13:24,526 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 18:13:24,526 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 18:13:24,526 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 18:13:24,526 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 18:13:24,526 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 18:13:25,650 - __main__ - INFO - Loaded 1 texts from scripts/example_texts.json -2025-07-24 18:13:25,650 - agentgraph.input.text_processing.chunking_service - INFO - ChunkingService initialized with batch_size=3, model=gpt-4o-mini -2025-07-24 18:13:25,650 - __main__ - INFO - Initialized BatchKGExtractor: -2025-07-24 18:13:25,650 - __main__ - INFO - - Model: gpt-4o-mini, Method: production -2025-07-24 18:13:25,650 - __main__ - INFO - - Preprocessing: False -2025-07-24 18:13:25,650 - __main__ - INFO - - Line numbers: True -2025-07-24 18:13:25,650 - __main__ - INFO - Processing batch of 1 texts -2025-07-24 18:13:25,650 - __main__ - INFO - Processing text 1/1: text_0 -2025-07-24 18:13:25,650 - agentgraph.input.text_processing.trace_line_processor - INFO - Added line numbers to 17 lines, starting from line 1 -2025-07-24 18:13:25,650 - __main__ - INFO - Added line numbers: 17 lines -2025-07-24 18:13:25,650 - __main__ - INFO - Processing text text_0 (format: auto) -2025-07-24 18:13:25,650 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:13:25,650 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 1417 characters -2025-07-24 18:13:25,650 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:13:25,650 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:13:25,657 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:13:25,657 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:13:25,659 - agentgraph.input.content_analysis.semantic_analyzer - INFO - Detected agent trace type: unknown (confidence: 0.00) -2025-07-24 18:13:25,660 - agentgraph.input.text_processing.chunking_service - INFO - Applied rule-based line splitting to 1 chunks (max_line_length=800) -2025-07-24 18:13:25,661 - agentgraph.input.text_processing.chunking_service - INFO - Content already has line numbers, skipping line number assignment -2025-07-24 18:13:25,661 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using agent_semantic splitter -2025-07-24 18:13:25,661 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=350000, overlap_size=17500 -2025-07-24 18:13:25,661 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context enhancement complete: 0 total documents -2025-07-24 18:13:25,661 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - SlidingWindowMonitor initialized with model: gpt-4o-mini, method: production -2025-07-24 18:13:25,661 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context documents: 0 provided -2025-07-24 18:13:25,661 - agentgraph.extraction.graph_utilities.knowledge_graph_merger - INFO - KnowledgeGraphMerger initialized with model: gpt-4o-mini -2025-07-24 18:13:25,662 - __main__ - INFO - Extracting knowledge graph from 1 chunks -2025-07-24 18:13:25,662 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks (splitter: agent_semantic, window_size=350000, overlap=0) -2025-07-24 18:13:25,662 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided source trace ID: text_0 -2025-07-24 18:13:25,662 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided processing run ID: batch_text_0 -2025-07-24 18:13:25,662 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks -2025-07-24 18:13:25,662 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using parallel processing with batch size 3 (1 batches) -2025-07-24 18:13:25,662 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 windows in 1 batches -2025-07-24 18:13:25,679 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks in 1 sub-batches (max 1 concurrent per sub-batch) -2025-07-24 18:13:25,679 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Starting sub-batch 1/1 with 1 chunks -2025-07-24 18:13:25,679 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing window 0 -2025-07-24 18:13:25,680 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - 🤖 Starting async_crew extraction for window 0 with method: production using model: gpt-4o-mini -2025-07-24 18:13:25,680 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - No context documents available for window 0, passing empty context string -2025-07-24 18:13:25,680 - agentgraph.methods.production.multi_agent_knowledge_extractor - INFO - Creating agent monitoring crew with model: gpt-4o-mini -2025-07-24 18:14:12,944 - openlit - INFO - Starting openLIT initialization... -2025-07-24 18:14:12,950 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 18:14:13,375 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 18:14:13,404 - openlit - INFO - Library for ollama (ollama) not found. Skipping instrumentation -2025-07-24 18:14:13,404 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 18:14:13,404 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 18:14:13,404 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 18:14:13,679 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 18:14:13,863 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 18:14:13,864 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 18:14:14,481 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 18:14:14,482 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 18:14:14,482 - openlit - INFO - Library for transformers (transformers) not found. Skipping instrumentation -2025-07-24 18:14:14,487 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 18:14:14,487 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 18:14:14,487 - openlit - INFO - Library for autogen (autogen) not found. Skipping instrumentation -2025-07-24 18:14:14,487 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 18:14:14,487 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 18:14:14,487 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 18:14:14,487 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 18:14:14,487 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 18:14:14,487 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 18:14:14,487 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 18:14:14,487 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 18:14:14,487 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 18:14:14,487 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 18:14:14,488 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 18:14:14,488 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 18:14:14,488 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 18:14:14,488 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 18:14:14,488 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 18:14:15,613 - __main__ - INFO - Loaded 1 texts from scripts/example.json -2025-07-24 18:14:15,613 - agentgraph.input.text_processing.chunking_service - INFO - ChunkingService initialized with batch_size=3, model=gpt-4o-mini -2025-07-24 18:14:15,613 - __main__ - INFO - Initialized BatchKGExtractor: -2025-07-24 18:14:15,613 - __main__ - INFO - - Model: gpt-4o-mini, Method: production -2025-07-24 18:14:15,613 - __main__ - INFO - - Preprocessing: False -2025-07-24 18:14:15,613 - __main__ - INFO - - Line numbers: True -2025-07-24 18:14:15,613 - __main__ - INFO - Processing batch of 1 texts -2025-07-24 18:14:15,613 - __main__ - INFO - Processing text 1/1: text_0 -2025-07-24 18:14:15,614 - agentgraph.input.text_processing.trace_line_processor - INFO - Added line numbers to 595 lines, starting from line 1 -2025-07-24 18:14:15,614 - __main__ - INFO - Added line numbers: 595 lines -2025-07-24 18:14:15,614 - __main__ - INFO - Processing text text_0 (format: auto) -2025-07-24 18:14:15,614 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-24 18:14:15,614 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 94970 characters -2025-07-24 18:14:15,614 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-24 18:14:15,614 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized tiktoken for accurate token counting -2025-07-24 18:14:15,620 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Initialized OpenAI client for text-embedding-3-small -2025-07-24 18:14:15,620 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-24 18:14:15,668 - agentgraph.input.content_analysis.semantic_analyzer - INFO - Detected agent trace type: unknown (confidence: 0.00) -2025-07-24 18:14:20,067 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Generated embeddings for 96 sentences using text-embedding-3-small with smart batching -2025-07-24 18:14:20,099 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Detected agent trace type: unknown (confidence: 0.00) -2025-07-24 18:14:20,099 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Unknown format detected - using data-driven threshold selection -2025-07-24 18:14:20,099 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Using standard deviation threshold -2025-07-24 18:14:20,099 - agentgraph.input.content_analysis.semantic_analyzer - DEBUG - Using adaptive threshold: 0.578 for 95 similarities -2025-07-24 18:14:20,111 - agentgraph.input.text_processing.chunking_service - INFO - Applied rule-based line splitting to 1 chunks (max_line_length=800) -2025-07-24 18:14:20,113 - agentgraph.input.text_processing.chunking_service - INFO - Content already has line numbers, skipping line number assignment -2025-07-24 18:14:20,113 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using agent_semantic splitter -2025-07-24 18:14:20,113 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=350000, overlap_size=17500 -2025-07-24 18:14:20,113 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context enhancement complete: 0 total documents -2025-07-24 18:14:20,113 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - SlidingWindowMonitor initialized with model: gpt-4o-mini, method: production -2025-07-24 18:14:20,113 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context documents: 0 provided -2025-07-24 18:14:20,113 - agentgraph.extraction.graph_utilities.knowledge_graph_merger - INFO - KnowledgeGraphMerger initialized with model: gpt-4o-mini -2025-07-24 18:14:20,114 - __main__ - INFO - Extracting knowledge graph from 1 chunks -2025-07-24 18:14:20,114 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks (splitter: agent_semantic, window_size=350000, overlap=2355) -2025-07-24 18:14:20,114 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided source trace ID: text_0 -2025-07-24 18:14:20,114 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided processing run ID: batch_text_0 -2025-07-24 18:14:20,114 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks -2025-07-24 18:14:20,114 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using parallel processing with batch size 3 (1 batches) -2025-07-24 18:14:20,114 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 windows in 1 batches -2025-07-24 18:14:20,130 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks in 1 sub-batches (max 1 concurrent per sub-batch) -2025-07-24 18:14:20,130 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Starting sub-batch 1/1 with 1 chunks -2025-07-24 18:14:20,130 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing window 0 -2025-07-24 18:14:20,131 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - 🤖 Starting async_crew extraction for window 0 with method: production using model: gpt-4o-mini -2025-07-24 18:14:20,131 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - No context documents available for window 0, passing empty context string -2025-07-24 18:14:20,131 - agentgraph.methods.production.multi_agent_knowledge_extractor - INFO - Creating agent monitoring crew with model: gpt-4o-mini -2025-07-24 18:15:15,514 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Result Generated (production method) -2025-07-24 18:15:15,521 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 610 lines -2025-07-24 18:15:15,522 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 18:15:15,522 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 1218 -2025-07-24 18:15:15,522 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L1-L1 -2025-07-24 18:15:15,522 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L1-L1 -2025-07-24 18:15:15,522 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 0, end_idx: 1 -2025-07-24 18:15:15,522 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 1218 -2025-07-24 18:15:15,522 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 18:15:15,522 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: [{"name": "ReAct Agent", "start_time": "... -2025-07-24 18:15:15,522 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: [{"name": "ReAct Agent", "start_time": "... -2025-07-24 18:15:15,522 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: [{"name": "ReAct Agent", "start_time": "... -2025-07-24 18:15:15,522 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 18:15:15,522 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity agent_001 resolution debug: -2025-07-24 18:15:15,522 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 18:15:15,522 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L1-L1 -2025-07-24 18:15:15,522 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 18:15:15,522 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: [{"name": "ReAct Agent", "start_time": "... -2025-07-24 18:15:15,522 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity agent_001: 150 characters -2025-07-24 18:15:15,522 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 18:15:15,522 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 1218 -2025-07-24 18:15:15,522 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L2-L2 -2025-07-24 18:15:15,522 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L2-L2 -2025-07-24 18:15:15,522 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 1, end_idx: 2 -2025-07-24 18:15:15,522 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 1218 -2025-07-24 18:15:15,522 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 18:15:15,522 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: : {"messag... -2025-07-24 18:15:15,522 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: : {"messag... -2025-07-24 18:15:15,522 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: : {"messag... -2025-07-24 18:15:15,522 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 18:15:15,522 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity task_001 resolution debug: -2025-07-24 18:15:15,522 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 18:15:15,522 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L2-L2 -2025-07-24 18:15:15,522 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 18:15:15,522 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: : {"messag... -2025-07-24 18:15:15,522 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity task_001: 12 characters -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 7 references -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 1218 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L3-L4 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L3-L4 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 2, end_idx: 4 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 1218 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 2 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: es": [{"role": "human", "content": "Wh... -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[1]: : {}, "respo... -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: es": [{"role": "human", "content": "Wh... -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: es": [{"role": "human", "content": "Wh... -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[1]: L69-L69 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L69-L69 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 68, end_idx: 69 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 1218 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: pecial LangChain-built library that ... -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: pecial LangChain-built library that ... -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[1]: pecial LangChain-built library that ... -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[2]: L135-L135 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L135-L135 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 134, end_idx: 135 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 1218 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: peat a cycle but its really clear.\"... -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: peat a cycle but its really clear.\"... -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[2]: peat a cycle but its really clear.\"... -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[3]: L201-L201 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L201-L201 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 200, end_idx: 201 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 1218 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: : "2025-07-18T14:02:40.631212", "r... -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: : "2025-07-18T14:02:40.631212", "r... -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[3]: : "2025-07-18T14:02:40.631212", "r... -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[4]: L205-L205 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L205-L205 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 204, end_idx: 205 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 1218 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: : 1, "type": "constructor", "id": ... -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: : 1, "type": "constructor", "id": ... -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[4]: : 1, "type": "constructor", "id": ... -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[5]: L267-L269 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L267-L269 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 266, end_idx: 269 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 1218 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 3 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: ", "additional_kwargs": {}, "respo... -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[1]: {"content": "",... -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[2]: "additional_kwargs": {"tool_calls... -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: ", "additional_kwargs": {}, "respo... -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[5]: ", "additional_kwargs": {}, "respo... -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[6]: L328-L328 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L328-L328 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 327, end_idx: 328 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 1218 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: , {\"url\": \"ht... -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: , {\"url\": \"ht... -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[6]: , {\"url\": \"ht... -2025-07-24 18:15:15,523 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 7 references. -2025-07-24 18:15:15,523 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity tool_001 resolution debug: -2025-07-24 18:15:15,523 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 7 -2025-07-24 18:15:15,523 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L3-L4 -2025-07-24 18:15:15,523 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[1]: L69-L69 -2025-07-24 18:15:15,523 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[2]: L135-L135 -2025-07-24 18:15:15,523 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[3]: L201-L201 -2025-07-24 18:15:15,523 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[4]: L205-L205 -2025-07-24 18:15:15,523 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[5]: L267-L269 -2025-07-24 18:15:15,523 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[6]: L328-L328 -2025-07-24 18:15:15,523 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 7 -2025-07-24 18:15:15,523 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: es": [{"role": "human", "content": "Wh... -2025-07-24 18:15:15,523 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[1]: pecial LangChain-built library that ... -2025-07-24 18:15:15,523 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[2]: peat a cycle but its really clear.\"... -2025-07-24 18:15:15,523 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[3]: : "2025-07-18T14:02:40.631212", "r... -2025-07-24 18:15:15,523 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[4]: : 1, "type": "constructor", "id": ... -2025-07-24 18:15:15,523 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[5]: ", "additional_kwargs": {}, "respo... -2025-07-24 18:15:15,523 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[6]: , {\"url\": \"ht... -2025-07-24 18:15:15,523 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity tool_001: 1104 characters -2025-07-24 18:15:15,523 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - joined with delimiter, split count will be: 7 -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 1218 -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L267-L267 -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L267-L267 -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 266, end_idx: 267 -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 1218 -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: ", "additional_kwargs": {}, "respo... -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: ", "additional_kwargs": {}, "respo... -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: ", "additional_kwargs": {}, "respo... -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 18:15:15,524 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity tool_002 resolution debug: -2025-07-24 18:15:15,524 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 18:15:15,524 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L267-L267 -2025-07-24 18:15:15,524 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 18:15:15,524 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: ", "additional_kwargs": {}, "respo... -2025-07-24 18:15:15,524 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity tool_002: 150 characters -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 1218 -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L205-L205 -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L205-L205 -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 204, end_idx: 205 -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 1218 -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: : 1, "type": "constructor", "id": ... -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: : 1, "type": "constructor", "id": ... -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: : 1, "type": "constructor", "id": ... -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 18:15:15,524 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity task_002 resolution debug: -2025-07-24 18:15:15,524 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 18:15:15,524 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L205-L205 -2025-07-24 18:15:15,524 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 18:15:15,524 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: : 1, "type": "constructor", "id": ... -2025-07-24 18:15:15,524 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity task_002: 150 characters -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 1218 -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L263-L263 -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L263-L263 -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 262, end_idx: 263 -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 1218 -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: ": ["1f063dfd-c8d4-64e7-84a7-6e7ca... -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: ": ["1f063dfd-c8d4-64e7-84a7-6e7ca... -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: ": ["1f063dfd-c8d4-64e7-84a7-6e7ca... -2025-07-24 18:15:15,524 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 18:15:15,524 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Entity output_001 resolution debug: -2025-07-24 18:15:15,524 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - raw_prompt_ref count: 1 -2025-07-24 18:15:15,524 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - ref[0]: L263-L263 -2025-07-24 18:15:15,524 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - extracted snippets count: 1 -2025-07-24 18:15:15,525 - agentgraph.reconstruction.content_reference_resolver - DEBUG - - snippet[0]: ": ["1f063dfd-c8d4-64e7-84a7-6e7ca... -2025-07-24 18:15:15,525 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved prompt for entity output_001: 150 characters -2025-07-24 18:15:15,525 - agentgraph.reconstruction.content_reference_resolver - INFO - Entity prompt resolution stats: {'total_entities': 6, 'entities_with_refs': 6, 'successful_resolutions': 6, 'failed_resolutions': 0} -2025-07-24 18:15:15,532 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Created extraction-compatible numbering for 610 lines -2025-07-24 18:15:15,533 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 18:15:15,533 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 1218 -2025-07-24 18:15:15,533 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L2-L2 -2025-07-24 18:15:15,533 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L2-L2 -2025-07-24 18:15:15,533 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 1, end_idx: 2 -2025-07-24 18:15:15,533 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 1218 -2025-07-24 18:15:15,533 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 18:15:15,533 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: : {"messag... -2025-07-24 18:15:15,533 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: : {"messag... -2025-07-24 18:15:15,533 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: : {"messag... -2025-07-24 18:15:15,533 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 18:15:15,533 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation rel_001: 12 characters -2025-07-24 18:15:15,533 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 18:15:15,533 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 1218 -2025-07-24 18:15:15,533 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L14-L14 -2025-07-24 18:15:15,533 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L14-L14 -2025-07-24 18:15:15,533 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 13, end_idx: 14 -2025-07-24 18:15:15,533 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 1218 -2025-07-24 18:15:15,533 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 18:15:15,533 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: v4Lr8", "ser... -2025-07-24 18:15:15,533 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: v4Lr8", "ser... -2025-07-24 18:15:15,533 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: v4Lr8", "ser... -2025-07-24 18:15:15,533 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 18:15:15,533 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation rel_002: 14 characters -2025-07-24 18:15:15,533 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 2 references -2025-07-24 18:15:15,533 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 1218 -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L3-L4 -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L3-L4 -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 2, end_idx: 4 -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 1218 -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 2 -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: es": [{"role": "human", "content": "Wh... -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[1]: : {}, "respo... -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: es": [{"role": "human", "content": "Wh... -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: es": [{"role": "human", "content": "Wh... -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[1]: L201-L201 -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L201-L201 -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 200, end_idx: 201 -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 1218 -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: : "2025-07-18T14:02:40.631212", "r... -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: : "2025-07-18T14:02:40.631212", "r... -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[1]: : "2025-07-18T14:02:40.631212", "r... -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 2 references. -2025-07-24 18:15:15,534 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation rel_003: 314 characters -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 1218 -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L67-L67 -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L67-L67 -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 66, end_idx: 67 -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 1218 -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: than LangChain - YouTube\", \"conten... -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: than LangChain - YouTube\", \"conten... -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: than LangChain - YouTube\", \"conten... -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 18:15:15,534 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation rel_004: 150 characters -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 1218 -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L263-L263 -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L263-L263 -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 262, end_idx: 263 -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 1218 -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: ": ["1f063dfd-c8d4-64e7-84a7-6e7ca... -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: ": ["1f063dfd-c8d4-64e7-84a7-6e7ca... -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: ": ["1f063dfd-c8d4-64e7-84a7-6e7ca... -2025-07-24 18:15:15,534 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 18:15:15,534 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation rel_005: 150 characters -2025-07-24 18:15:15,535 - agentgraph.input.text_processing.trace_line_processor - DEBUG - extract_content_by_reference: Processing 1 references -2025-07-24 18:15:15,535 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines count: 1218 -2025-07-24 18:15:15,535 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processing ref[0]: L263-L263 -2025-07-24 18:15:15,535 - agentgraph.input.text_processing.trace_line_processor - DEBUG - _extract_single_reference: L263-L263 -2025-07-24 18:15:15,535 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - start_idx: 262, end_idx: 263 -2025-07-24 18:15:15,535 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - processed_lines length: 1218 -2025-07-24 18:15:15,535 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines count: 1 -2025-07-24 18:15:15,535 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - slice_lines[0]: ": ["1f063dfd-c8d4-64e7-84a7-6e7ca... -2025-07-24 18:15:15,535 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - reconstructed: ": ["1f063dfd-c8d4-64e7-84a7-6e7ca... -2025-07-24 18:15:15,535 - agentgraph.input.text_processing.trace_line_processor - DEBUG - - extracted content[0]: ": ["1f063dfd-c8d4-64e7-84a7-6e7ca... -2025-07-24 18:15:15,535 - agentgraph.input.text_processing.trace_line_processor - DEBUG - Successfully extracted content for 1 references. -2025-07-24 18:15:15,535 - agentgraph.reconstruction.content_reference_resolver - DEBUG - Resolved interaction prompt for relation rel_006: 150 characters -2025-07-24 18:15:15,535 - agentgraph.reconstruction.content_reference_resolver - INFO - Relation prompt resolution stats: {'total_relations': 6, 'relations_with_refs': 6, 'successful_resolutions': 6, 'failed_resolutions': 0} -2025-07-24 18:15:15,535 - agentgraph.reconstruction.content_reference_resolver - INFO - Resolved content references for knowledge graph with 6 entities and 6 relations -2025-07-24 18:15:15,535 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Resolved content references for window 0 -2025-07-24 18:15:15,535 - agentgraph.extraction.graph_processing.knowledge_graph_processor - DEBUG - Completed sub-batch 1/1 -2025-07-24 18:15:15,536 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Preparing window knowledge graphs with metadata -2025-07-24 18:15:15,536 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Merging knowledge graphs... -2025-07-24 18:15:15,536 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Only one knowledge graph generated, skipping merge process -2025-07-24 18:15:15,536 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing complete. Knowledge graph generated with 6 entities and 6 relations -2025-07-24 18:15:15,537 - __main__ - INFO - Results saved to test_results_fixed.json -2025-07-25 11:57:48,965 - openlit - INFO - Starting openLIT initialization... -2025-07-25 11:57:48,982 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-25 11:57:49,467 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-25 11:57:49,507 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-25 11:57:49,507 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-25 11:57:49,507 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-25 11:57:49,819 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-25 11:57:49,907 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-25 11:57:49,907 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-25 11:57:50,390 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-25 11:57:50,391 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-25 11:57:51,288 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-25 11:57:51,290 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-25 11:57:51,290 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-25 11:57:51,290 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-25 11:57:51,290 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-25 11:57:51,290 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-25 11:57:51,290 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-25 11:57:51,290 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-25 11:57:51,290 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-25 11:57:51,290 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-25 11:57:51,290 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-25 11:57:51,290 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-25 11:57:51,290 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-25 11:57:51,291 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-25 11:57:51,291 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-25 11:57:51,291 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-25 11:57:51,291 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-25 11:57:51,291 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-25 11:57:51,291 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-25 11:58:41,271 - openlit - INFO - Starting openLIT initialization... -2025-07-25 11:58:41,288 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-25 11:58:41,771 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-25 11:58:41,812 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-25 11:58:41,812 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-25 11:58:41,812 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-25 11:58:42,117 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-25 11:58:42,205 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-25 11:58:42,205 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-25 11:58:42,676 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-25 11:58:42,677 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-25 11:58:43,571 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-25 11:58:43,573 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-25 11:58:43,573 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-25 11:58:43,573 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-25 11:58:43,573 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-25 11:58:43,573 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-25 11:58:43,573 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-25 11:58:43,573 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-25 11:58:43,573 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-25 11:58:43,574 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-25 11:58:43,574 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-25 11:58:43,574 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-25 11:58:43,574 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-25 11:58:43,574 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-25 11:58:43,574 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-25 11:58:43,574 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-25 11:58:43,574 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-25 11:58:43,574 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-25 11:58:43,574 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-25 11:59:13,180 - openlit - INFO - Starting openLIT initialization... -2025-07-25 11:59:13,197 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-25 11:59:13,777 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-25 11:59:13,828 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-25 11:59:13,828 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-25 11:59:13,828 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-25 11:59:14,236 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-25 11:59:14,336 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-25 11:59:14,336 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-25 11:59:14,891 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-25 11:59:14,891 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-25 11:59:16,117 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-25 11:59:16,119 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-25 11:59:16,119 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-25 11:59:16,119 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-25 11:59:16,120 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-25 11:59:16,120 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-25 11:59:16,120 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-25 11:59:16,120 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-25 11:59:16,120 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-25 11:59:16,120 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-25 11:59:16,120 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-25 11:59:16,120 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-25 11:59:16,120 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-25 11:59:16,120 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-25 11:59:16,120 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-25 11:59:16,121 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-25 11:59:16,121 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-25 11:59:16,121 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-25 11:59:16,121 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-25 11:59:16,941 - agentgraph.input.text_processing.chunking_service - INFO - ChunkingService initialized with batch_size=3, model=gpt-4o-mini -2025-07-25 11:59:16,941 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-25 11:59:16,941 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 131981 characters -2025-07-25 11:59:16,941 - agentgraph.input.text_processing.chunking_service - INFO - Analyzing trace to determine optimal parameters... -2025-07-25 11:59:16,942 - agentgraph.input.trace_management.trace_analysis - INFO - Analyzing trace characteristics... -2025-07-25 11:59:16,942 - agentgraph.input.trace_management.trace_analysis - INFO - Applying content optimization for cost savings... -2025-07-25 11:59:16,951 - agentgraph.input.trace_management.trace_analysis - INFO - Content optimization complete: 3,700 characters removed (2.8% reduction) -2025-07-25 11:59:16,963 - agentgraph.input.text_processing.chunking_service - INFO - Using recommended parameters from trace analysis: -2025-07-25 11:59:16,963 - agentgraph.input.text_processing.chunking_service - INFO - - Window size: 400,000 characters -2025-07-25 11:59:16,963 - agentgraph.input.text_processing.chunking_service - INFO - - Overlap size: 12,000 characters -2025-07-25 11:59:16,963 - agentgraph.input.text_processing.chunking_service - INFO - - Estimated windows: 1 -2025-07-25 11:59:16,971 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=400000, overlap_ratio=0.03 -2025-07-25 11:59:16,995 - agentgraph.input.content_analysis.semantic_analyzer - INFO - Detected agent trace type: unknown (confidence: 0.14) -2025-07-25 11:59:25,856 - agentgraph.input.text_processing.chunking_service - INFO - Applied rule-based line splitting to 1 chunks (max_line_length=800) -2025-07-25 11:59:25,857 - agentgraph.input.text_processing.chunking_service - INFO - Assigning global line numbers to 1 chunks -2025-07-25 11:59:25,865 - agentgraph.input.text_processing.trace_line_processor - INFO - Added line numbers to 991 lines, starting from line 1 -2025-07-25 11:59:25,866 - agentgraph.input.text_processing.chunking_service - INFO - Successfully assigned global line numbers to all chunks -2025-07-25 11:59:25,866 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using agent_semantic splitter -2025-07-25 11:59:25,866 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=400000, overlap_size=12000 -2025-07-25 11:59:25,866 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context enhancement complete: 0 total documents -2025-07-25 11:59:25,866 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - SlidingWindowMonitor initialized with model: gpt-4.1-mini, method: production -2025-07-25 11:59:25,866 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context documents: 0 provided -2025-07-25 11:59:25,866 - agentgraph.extraction.graph_utilities.knowledge_graph_merger - INFO - KnowledgeGraphMerger initialized with model: gpt-4.1-mini -2025-07-25 11:59:25,868 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks (splitter: agent_semantic, window_size=400000, overlap=164) -2025-07-25 11:59:25,868 - agentgraph.extraction.graph_processing.knowledge_graph_processor - WARNING - No source trace ID provided, generated new trace ID: d70d0214-a08e-4d69-a967-97e2937ebc33 -2025-07-25 11:59:25,868 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Generated processing run ID: 54dd2b16 -2025-07-25 11:59:25,868 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks -2025-07-25 11:59:25,868 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using sequential processing for 1 windows -2025-07-25 11:59:25,880 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing window 0 -2025-07-25 11:59:25,880 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - 🤖 Starting async_crew extraction for window 0 with method: production using model: gpt-4.1-mini -2025-07-25 11:59:25,880 - agentgraph.methods.production.multi_agent_knowledge_extractor - INFO - Creating agent monitoring crew with model: gpt-4.1-mini -2025-07-25 12:01:44,886 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Result Generated (production method) -2025-07-25 12:01:44,976 - agentgraph.reconstruction.content_reference_resolver - INFO - Entity prompt resolution stats: {'total_entities': 11, 'entities_with_refs': 11, 'successful_resolutions': 11, 'failed_resolutions': 0} -2025-07-25 12:01:44,992 - agentgraph.reconstruction.content_reference_resolver - INFO - Relation prompt resolution stats: {'total_relations': 16, 'relations_with_refs': 16, 'successful_resolutions': 16, 'failed_resolutions': 0} -2025-07-25 12:01:44,992 - agentgraph.reconstruction.content_reference_resolver - INFO - Resolved content references for knowledge graph with 11 entities and 16 relations -2025-07-25 12:01:44,995 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Preparing window knowledge graphs with metadata -2025-07-25 12:01:44,995 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Merging knowledge graphs... -2025-07-25 12:01:44,995 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Only one knowledge graph generated, skipping merge process -2025-07-25 12:01:44,995 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing complete. Knowledge graph generated with 11 entities and 16 relations -2025-07-25 12:01:44,995 - agentgraph.reconstruction.prompt_reconstructor - INFO - Successfully initialized PromptReconstructor with 11 entities and 16 relations -2025-07-25 12:01:44,996 - agentgraph.testing.perturbation_types.base - INFO - Prepared testing data with 14 relations with reconstructed prompts -2025-07-25 12:01:45,011 - agentgraph.testing.perturbation_types.jailbreak - INFO - Successfully loaded 185 jailbreak techniques from CSV file -2025-07-25 12:01:45,011 - agentgraph.testing.perturbation_types.jailbreak - INFO - Running jailbreak tests on 5 relations using 185 techniques -2025-07-25 12:03:45,859 - agentgraph.testing.perturbation_types.jailbreak - INFO - Jailbreak testing completed: 5/5 successful -2025-07-28 17:16:44,396 - openlit - INFO - Starting openLIT initialization... -2025-07-28 17:16:44,402 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-28 17:16:45,068 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-28 17:16:45,130 - openlit - INFO - Library for ollama (ollama) not found. Skipping instrumentation -2025-07-28 17:16:45,131 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-28 17:16:45,131 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-28 17:16:45,131 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-28 17:16:45,564 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-28 17:16:45,816 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-28 17:16:45,816 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-28 17:16:46,513 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-28 17:16:46,514 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-28 17:16:46,514 - openlit - INFO - Library for transformers (transformers) not found. Skipping instrumentation -2025-07-28 17:16:46,517 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-28 17:16:46,517 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-28 17:16:46,517 - openlit - INFO - Library for autogen (autogen) not found. Skipping instrumentation -2025-07-28 17:16:46,517 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-28 17:16:46,517 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-28 17:16:46,518 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-28 17:16:46,518 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-28 17:16:46,518 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-28 17:16:46,518 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-28 17:16:46,518 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-28 17:16:46,518 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-28 17:16:46,518 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-28 17:16:46,518 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-28 17:16:46,518 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-28 17:16:46,518 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-28 17:16:46,518 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-28 17:16:46,518 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-28 17:16:46,518 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-29 14:22:38,272 - openlit - INFO - Starting openLIT initialization... -2025-07-29 14:22:38,288 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-29 14:22:38,925 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-29 14:22:38,977 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-29 14:22:38,977 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-29 14:22:38,977 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-29 14:22:39,343 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-29 14:22:39,456 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-29 14:22:39,456 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-29 14:22:40,051 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-29 14:22:40,052 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-29 14:22:41,609 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-29 14:22:41,612 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-29 14:22:41,612 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-29 14:22:41,612 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-29 14:22:41,612 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-29 14:22:41,612 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-29 14:22:41,613 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-29 14:22:41,613 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-29 14:22:41,613 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-29 14:22:41,613 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-29 14:22:41,613 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-29 14:22:41,613 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-29 14:22:41,613 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-29 14:22:41,613 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-29 14:22:41,613 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-29 14:22:41,613 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-29 14:22:41,613 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-29 14:22:41,614 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-29 14:22:41,614 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-29 14:24:43,475 - openlit - INFO - Starting openLIT initialization... -2025-07-29 14:24:43,492 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-29 14:24:44,096 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-29 14:24:44,144 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-29 14:24:44,144 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-29 14:24:44,144 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-29 14:24:44,484 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-29 14:24:44,590 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-29 14:24:44,591 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-29 14:24:45,138 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-29 14:24:45,139 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-29 14:24:46,375 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-29 14:24:46,377 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-29 14:24:46,377 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-29 14:24:46,377 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-29 14:24:46,377 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-29 14:24:46,377 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-29 14:24:46,377 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-29 14:24:46,378 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-29 14:24:46,378 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-29 14:24:46,378 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-29 14:24:46,378 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-29 14:24:46,378 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-29 14:24:46,378 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-29 14:24:46,378 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-29 14:24:46,378 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-29 14:24:46,378 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-29 14:24:46,378 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-29 14:24:46,378 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-29 14:24:46,378 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-29 14:28:17,225 - openlit - INFO - Starting openLIT initialization... -2025-07-29 14:28:17,239 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-29 14:28:17,810 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-29 14:28:17,854 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-29 14:28:17,854 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-29 14:28:17,854 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-29 14:28:18,182 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-29 14:28:18,282 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-29 14:28:18,282 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-29 14:28:18,777 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-29 14:28:18,778 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-29 14:28:19,929 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-29 14:28:19,932 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-29 14:28:19,932 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-29 14:28:19,932 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-29 14:28:19,932 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-29 14:28:19,932 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-29 14:28:19,932 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-29 14:28:19,932 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-29 14:28:19,932 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-29 14:28:19,932 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-29 14:28:19,932 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-29 14:28:19,933 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-29 14:28:19,933 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-29 14:28:19,933 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-29 14:28:19,933 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-29 14:28:19,933 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-29 14:28:19,933 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-29 14:28:19,933 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-29 14:28:19,933 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-29 14:28:26,062 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.seed' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:28:26,062 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.frequency_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:28:26,062 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.presence_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:28:26,062 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.stop_sequences' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:28:26,062 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.top_p' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:28:26,062 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.service_tier' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:28:26,062 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.user' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:29:03,915 - openlit - INFO - Starting openLIT initialization... -2025-07-29 14:29:03,928 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-29 14:29:04,478 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-29 14:29:04,522 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-29 14:29:04,522 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-29 14:29:04,522 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-29 14:29:04,843 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-29 14:29:04,940 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-29 14:29:04,940 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-29 14:29:05,443 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-29 14:29:05,444 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-29 14:29:06,623 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-29 14:29:06,625 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-29 14:29:06,625 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-29 14:29:06,625 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-29 14:29:06,625 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-29 14:29:06,626 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-29 14:29:06,626 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-29 14:29:06,626 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-29 14:29:06,626 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-29 14:29:06,626 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-29 14:29:06,626 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-29 14:29:06,626 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-29 14:29:06,626 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-29 14:29:06,626 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-29 14:29:06,626 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-29 14:29:06,626 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-29 14:29:06,626 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-29 14:29:06,627 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-29 14:29:06,627 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-29 14:29:11,993 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.seed' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:29:11,993 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.frequency_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:29:11,993 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.presence_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:29:11,993 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.stop_sequences' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:29:11,993 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.top_p' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:29:11,993 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.service_tier' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:29:11,993 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.user' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:29:16,169 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.seed' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:29:16,169 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.frequency_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:29:16,169 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.presence_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:29:16,169 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.stop_sequences' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:29:16,169 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.top_p' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:29:16,170 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.service_tier' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:29:16,170 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.user' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:29:23,605 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.seed' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:29:23,605 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.frequency_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:29:23,605 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.presence_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:29:23,605 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.stop_sequences' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:29:23,605 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.top_p' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:29:23,605 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.service_tier' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:29:23,605 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.user' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:30:57,471 - openlit - INFO - Starting openLIT initialization... -2025-07-29 14:30:57,487 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-29 14:30:58,075 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-29 14:30:58,122 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-29 14:30:58,122 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-29 14:30:58,122 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-29 14:30:58,463 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-29 14:30:58,568 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-29 14:30:58,569 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-29 14:30:59,119 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-29 14:30:59,119 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-29 14:31:00,356 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-29 14:31:00,358 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-29 14:31:00,358 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-29 14:31:00,359 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-29 14:31:00,359 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-29 14:31:00,359 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-29 14:31:00,359 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-29 14:31:00,359 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-29 14:31:00,359 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-29 14:31:00,359 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-29 14:31:00,359 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-29 14:31:00,359 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-29 14:31:00,359 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-29 14:31:00,359 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-29 14:31:00,360 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-29 14:31:00,360 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-29 14:31:00,360 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-29 14:31:00,360 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-29 14:31:00,360 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-29 14:35:12,059 - openlit - INFO - Starting openLIT initialization... -2025-07-29 14:35:12,064 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-29 14:35:12,507 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-29 14:35:12,540 - openlit - INFO - Library for ollama (ollama) not found. Skipping instrumentation -2025-07-29 14:35:12,540 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-29 14:35:12,540 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-29 14:35:12,540 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-29 14:35:12,815 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-29 14:35:13,004 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-29 14:35:13,004 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-29 14:35:13,566 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-29 14:35:13,567 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-29 14:35:13,567 - openlit - INFO - Library for transformers (transformers) not found. Skipping instrumentation -2025-07-29 14:35:13,570 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-29 14:35:13,570 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-29 14:35:13,570 - openlit - INFO - Library for autogen (autogen) not found. Skipping instrumentation -2025-07-29 14:35:13,570 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-29 14:35:13,570 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-29 14:35:13,570 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-29 14:35:13,571 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-29 14:35:13,571 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-29 14:35:13,571 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-29 14:35:13,571 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-29 14:35:13,571 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-29 14:35:13,571 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-29 14:35:13,571 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-29 14:35:13,571 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-29 14:35:13,571 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-29 14:35:13,571 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-29 14:35:13,571 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-29 14:35:13,571 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-29 14:36:08,723 - openlit - INFO - Starting openLIT initialization... -2025-07-29 14:36:08,738 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-29 14:36:09,345 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-29 14:36:09,396 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-29 14:36:09,397 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-29 14:36:09,397 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-29 14:36:09,755 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-29 14:36:09,873 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-29 14:36:09,873 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-29 14:36:10,468 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-29 14:36:10,469 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-29 14:36:11,873 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-29 14:36:11,875 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-29 14:36:11,875 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-29 14:36:11,876 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-29 14:36:11,876 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-29 14:36:11,876 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-29 14:36:11,876 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-29 14:36:11,876 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-29 14:36:11,876 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-29 14:36:11,876 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-29 14:36:11,876 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-29 14:36:11,877 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-29 14:36:11,877 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-29 14:36:11,877 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-29 14:36:11,877 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-29 14:36:11,877 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-29 14:36:11,877 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-29 14:36:11,877 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-29 14:36:11,877 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-29 14:36:34,454 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.seed' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:36:34,455 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.frequency_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:36:34,455 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.presence_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:36:34,455 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.stop_sequences' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:36:34,455 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.top_p' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:36:34,455 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.service_tier' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:36:34,455 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.user' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:36:47,364 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.seed' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:36:47,365 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.frequency_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:36:47,365 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.presence_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:36:47,365 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.stop_sequences' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:36:47,365 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.top_p' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:36:47,365 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.service_tier' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:36:47,365 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.user' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:37:16,145 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.seed' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:37:16,146 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.frequency_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:37:16,146 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.presence_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:37:16,146 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.stop_sequences' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:37:16,146 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.top_p' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:37:16,146 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.service_tier' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:37:16,146 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.user' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:38:34,340 - openlit - INFO - Starting openLIT initialization... -2025-07-29 14:38:34,345 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-29 14:38:34,750 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-29 14:38:34,779 - openlit - INFO - Library for ollama (ollama) not found. Skipping instrumentation -2025-07-29 14:38:34,779 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-29 14:38:34,779 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-29 14:38:34,779 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-29 14:38:35,040 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-29 14:38:35,211 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-29 14:38:35,212 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-29 14:38:35,748 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-29 14:38:35,749 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-29 14:38:35,749 - openlit - INFO - Library for transformers (transformers) not found. Skipping instrumentation -2025-07-29 14:38:35,752 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-29 14:38:35,752 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-29 14:38:35,752 - openlit - INFO - Library for autogen (autogen) not found. Skipping instrumentation -2025-07-29 14:38:35,752 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-29 14:38:35,752 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-29 14:38:35,753 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-29 14:38:35,753 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-29 14:38:35,753 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-29 14:38:35,753 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-29 14:38:35,753 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-29 14:38:35,753 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-29 14:38:35,753 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-29 14:38:35,753 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-29 14:38:35,753 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-29 14:38:35,753 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-29 14:38:35,753 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-29 14:38:35,753 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-29 14:38:35,753 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-29 14:42:31,615 - openlit - INFO - Starting openLIT initialization... -2025-07-29 14:42:31,630 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-29 14:42:32,264 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-29 14:42:32,312 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-29 14:42:32,312 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-29 14:42:32,312 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-29 14:42:32,665 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-29 14:42:32,767 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-29 14:42:32,767 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-29 14:42:33,313 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-29 14:42:33,314 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-29 14:42:34,528 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-29 14:42:34,531 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-29 14:42:34,531 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-29 14:42:34,531 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-29 14:42:34,531 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-29 14:42:34,531 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-29 14:42:34,531 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-29 14:42:34,532 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-29 14:42:34,532 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-29 14:42:34,532 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-29 14:42:34,532 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-29 14:42:34,532 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-29 14:42:34,532 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-29 14:42:34,532 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-29 14:42:34,532 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-29 14:42:34,532 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-29 14:42:34,532 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-29 14:42:34,533 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-29 14:42:34,533 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-29 14:44:27,624 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.seed' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:44:27,625 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.frequency_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:44:27,626 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.presence_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:44:27,626 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.stop_sequences' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:44:27,626 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.top_p' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:44:27,626 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.user' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:49:49,686 - openlit - INFO - Starting openLIT initialization... -2025-07-29 14:49:49,690 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-29 14:49:50,100 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-29 14:49:50,127 - openlit - INFO - Library for ollama (ollama) not found. Skipping instrumentation -2025-07-29 14:49:50,127 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-29 14:49:50,127 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-29 14:49:50,127 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-29 14:49:50,379 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-29 14:49:50,544 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-29 14:49:50,544 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-29 14:49:51,090 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-29 14:49:51,091 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-29 14:49:51,091 - openlit - INFO - Library for transformers (transformers) not found. Skipping instrumentation -2025-07-29 14:49:51,094 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-29 14:49:51,094 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-29 14:49:51,094 - openlit - INFO - Library for autogen (autogen) not found. Skipping instrumentation -2025-07-29 14:49:51,094 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-29 14:49:51,095 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-29 14:49:51,095 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-29 14:49:51,095 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-29 14:49:51,095 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-29 14:49:51,095 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-29 14:49:51,095 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-29 14:49:51,095 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-29 14:49:51,095 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-29 14:49:51,095 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-29 14:49:51,095 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-29 14:49:51,095 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-29 14:49:51,095 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-29 14:49:51,095 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-29 14:49:51,095 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-29 14:54:22,615 - openlit - INFO - Starting openLIT initialization... -2025-07-29 14:54:22,630 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-29 14:54:23,376 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-29 14:54:23,430 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-29 14:54:23,430 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-29 14:54:23,430 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-29 14:54:23,806 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-29 14:54:23,916 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-29 14:54:23,916 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-29 14:54:24,499 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-29 14:54:24,500 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-29 14:54:25,918 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-29 14:54:25,921 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-29 14:54:25,921 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-29 14:54:25,921 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-29 14:54:25,921 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-29 14:54:25,921 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-29 14:54:25,921 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-29 14:54:25,922 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-29 14:54:25,922 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-29 14:54:25,922 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-29 14:54:25,922 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-29 14:54:25,922 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-29 14:54:25,922 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-29 14:54:25,922 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-29 14:54:25,922 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-29 14:54:25,922 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-29 14:54:25,922 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-29 14:54:25,922 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-29 14:54:25,922 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-29 14:54:45,790 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.seed' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:54:45,792 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.frequency_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:54:45,792 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.presence_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:54:45,792 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.stop_sequences' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:54:45,792 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.top_p' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:54:45,792 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.user' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:55:10,396 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.seed' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:55:10,397 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.frequency_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:55:10,397 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.presence_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:55:10,397 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.stop_sequences' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:55:10,397 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.top_p' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:55:10,397 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.service_tier' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:55:10,397 - opentelemetry.attributes - WARNING - Invalid type NoneType for attribute 'gen_ai.response.system_fingerprint' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:55:10,397 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.user' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:56:09,817 - openlit - INFO - Starting openLIT initialization... -2025-07-29 14:56:09,835 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-29 14:56:10,446 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-29 14:56:10,493 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-29 14:56:10,493 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-29 14:56:10,493 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-29 14:56:10,841 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-29 14:56:10,943 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-29 14:56:10,943 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-29 14:56:11,487 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-29 14:56:11,488 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-29 14:56:12,743 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-29 14:56:12,746 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-29 14:56:12,746 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-29 14:56:12,746 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-29 14:56:12,746 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-29 14:56:12,746 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-29 14:56:12,746 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-29 14:56:12,747 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-29 14:56:12,747 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-29 14:56:12,747 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-29 14:56:12,747 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-29 14:56:12,747 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-29 14:56:12,747 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-29 14:56:12,747 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-29 14:56:12,747 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-29 14:56:12,747 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-29 14:56:12,747 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-29 14:56:12,747 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-29 14:56:12,747 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-29 14:56:42,157 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.seed' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:56:42,158 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.frequency_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:56:42,158 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.presence_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:56:42,158 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.stop_sequences' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:56:42,158 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.top_p' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:56:42,158 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.service_tier' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:56:42,158 - opentelemetry.attributes - WARNING - Invalid type NoneType for attribute 'gen_ai.response.system_fingerprint' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 14:56:42,158 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.user' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-29 15:15:07,007 - opentelemetry.sdk.metrics._internal.export - ERROR - Exception while exporting metrics -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 534, in _make_request - response = conn.getresponse() - ^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connection.py", line 516, in getresponse - httplib_response = super().getresponse() - ^^^^^^^^^^^^^^^^^^^^^ - File "/opt/homebrew/Cellar/python@3.12/3.12.9/Frameworks/Python.framework/Versions/3.12/lib/python3.12/http/client.py", line 1430, in getresponse - response.begin() - File "/opt/homebrew/Cellar/python@3.12/3.12.9/Frameworks/Python.framework/Versions/3.12/lib/python3.12/http/client.py", line 331, in begin - version, status, reason = self._read_status() - ^^^^^^^^^^^^^^^^^^^ - File "/opt/homebrew/Cellar/python@3.12/3.12.9/Frameworks/Python.framework/Versions/3.12/lib/python3.12/http/client.py", line 292, in _read_status - line = str(self.fp.readline(_MAXLINE + 1), "iso-8859-1") - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/opt/homebrew/Cellar/python@3.12/3.12.9/Frameworks/Python.framework/Versions/3.12/lib/python3.12/socket.py", line 720, in readinto - return self._sock.recv_into(b) - ^^^^^^^^^^^^^^^^^^^^^^^ - File "/opt/homebrew/Cellar/python@3.12/3.12.9/Frameworks/Python.framework/Versions/3.12/lib/python3.12/ssl.py", line 1251, in recv_into - return self.read(nbytes, buffer) - ^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/opt/homebrew/Cellar/python@3.12/3.12.9/Frameworks/Python.framework/Versions/3.12/lib/python3.12/ssl.py", line 1103, in read - return self._sslobj.read(len, buffer) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ -TimeoutError: The read operation timed out - -The above exception was the direct cause of the following exception: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/adapters.py", line 667, in send - resp = conn.urlopen( - ^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 841, in urlopen - retries = retries.increment( - ^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/util/retry.py", line 474, in increment - raise reraise(type(error), error, _stacktrace) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/util/util.py", line 39, in reraise - raise value - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 787, in urlopen - response = self._make_request( - ^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 536, in _make_request - self._raise_timeout(err=e, url=url, timeout_value=read_timeout) - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 367, in _raise_timeout - raise ReadTimeoutError( -urllib3.exceptions.ReadTimeoutError: HTTPSConnectionPool(host='cloud.langfuse.com', port=443): Read timed out. (read timeout=10) - -During handling of the above exception, another exception occurred: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/opentelemetry/sdk/metrics/_internal/export/__init__.py", line 550, in _receive_metrics - self._exporter.export( - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/opentelemetry/exporter/otlp/proto/http/metric_exporter/__init__.py", line 222, in export - resp = self._export(serialized_data.SerializeToString()) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/opentelemetry/exporter/otlp/proto/http/metric_exporter/__init__.py", line 184, in _export - resp = self._session.post( - ^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 637, in post - return self.request("POST", url, data=data, json=json, **kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 589, in request - resp = self.send(prep, **send_kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 703, in send - r = adapter.send(request, **kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/adapters.py", line 713, in send - raise ReadTimeout(e, request=request) -requests.exceptions.ReadTimeout: HTTPSConnectionPool(host='cloud.langfuse.com', port=443): Read timed out. (read timeout=10) -2025-07-29 15:16:19,942 - opentelemetry.sdk.metrics._internal.export - ERROR - Exception while exporting metrics -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 534, in _make_request - response = conn.getresponse() - ^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connection.py", line 516, in getresponse - httplib_response = super().getresponse() - ^^^^^^^^^^^^^^^^^^^^^ - File "/opt/homebrew/Cellar/python@3.12/3.12.9/Frameworks/Python.framework/Versions/3.12/lib/python3.12/http/client.py", line 1430, in getresponse - response.begin() - File "/opt/homebrew/Cellar/python@3.12/3.12.9/Frameworks/Python.framework/Versions/3.12/lib/python3.12/http/client.py", line 331, in begin - version, status, reason = self._read_status() - ^^^^^^^^^^^^^^^^^^^ - File "/opt/homebrew/Cellar/python@3.12/3.12.9/Frameworks/Python.framework/Versions/3.12/lib/python3.12/http/client.py", line 292, in _read_status - line = str(self.fp.readline(_MAXLINE + 1), "iso-8859-1") - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/opt/homebrew/Cellar/python@3.12/3.12.9/Frameworks/Python.framework/Versions/3.12/lib/python3.12/socket.py", line 720, in readinto - return self._sock.recv_into(b) - ^^^^^^^^^^^^^^^^^^^^^^^ - File "/opt/homebrew/Cellar/python@3.12/3.12.9/Frameworks/Python.framework/Versions/3.12/lib/python3.12/ssl.py", line 1251, in recv_into - return self.read(nbytes, buffer) - ^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/opt/homebrew/Cellar/python@3.12/3.12.9/Frameworks/Python.framework/Versions/3.12/lib/python3.12/ssl.py", line 1103, in read - return self._sslobj.read(len, buffer) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ -TimeoutError: The read operation timed out - -The above exception was the direct cause of the following exception: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/adapters.py", line 667, in send - resp = conn.urlopen( - ^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 841, in urlopen - retries = retries.increment( - ^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/util/retry.py", line 474, in increment - raise reraise(type(error), error, _stacktrace) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/util/util.py", line 39, in reraise - raise value - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 787, in urlopen - response = self._make_request( - ^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 536, in _make_request - self._raise_timeout(err=e, url=url, timeout_value=read_timeout) - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 367, in _raise_timeout - raise ReadTimeoutError( -urllib3.exceptions.ReadTimeoutError: HTTPSConnectionPool(host='cloud.langfuse.com', port=443): Read timed out. (read timeout=10) - -During handling of the above exception, another exception occurred: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/opentelemetry/sdk/metrics/_internal/export/__init__.py", line 550, in _receive_metrics - self._exporter.export( - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/opentelemetry/exporter/otlp/proto/http/metric_exporter/__init__.py", line 222, in export - resp = self._export(serialized_data.SerializeToString()) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/opentelemetry/exporter/otlp/proto/http/metric_exporter/__init__.py", line 184, in _export - resp = self._session.post( - ^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 637, in post - return self.request("POST", url, data=data, json=json, **kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 589, in request - resp = self.send(prep, **send_kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 703, in send - r = adapter.send(request, **kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/adapters.py", line 713, in send - raise ReadTimeout(e, request=request) -requests.exceptions.ReadTimeout: HTTPSConnectionPool(host='cloud.langfuse.com', port=443): Read timed out. (read timeout=10) -2025-07-29 18:13:18,384 - openlit - INFO - Starting openLIT initialization... -2025-07-29 18:13:18,389 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-29 18:13:18,402 - openlit.__helpers - ERROR - Unexpected error occurred while fetching pricing info: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: /openlit/openlit/main/assets/pricing.json (Caused by NameResolutionError(": Failed to resolve 'raw.githubusercontent.com' ([Errno 8] nodename nor servname provided, or not known)")) -2025-07-29 18:13:18,786 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-29 18:13:18,815 - openlit - INFO - Library for ollama (ollama) not found. Skipping instrumentation -2025-07-29 18:13:18,815 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-29 18:13:18,815 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-29 18:13:18,815 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-29 18:13:19,077 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-29 18:13:19,252 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-29 18:13:19,252 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-29 18:13:19,813 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-29 18:13:19,814 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-29 18:13:19,814 - openlit - INFO - Library for transformers (transformers) not found. Skipping instrumentation -2025-07-29 18:13:19,818 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-29 18:13:19,818 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-29 18:13:19,818 - openlit - INFO - Library for autogen (autogen) not found. Skipping instrumentation -2025-07-29 18:13:19,818 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-29 18:13:19,818 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-29 18:13:19,818 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-29 18:13:19,818 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-29 18:13:19,818 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-29 18:13:19,818 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-29 18:13:19,819 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-29 18:13:19,819 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-29 18:13:19,819 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-29 18:13:19,819 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-29 18:13:19,819 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-29 18:13:19,819 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-29 18:13:19,819 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-29 18:13:19,819 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-29 18:13:19,819 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-29 18:13:22,528 - opentelemetry.sdk._shared_internal - ERROR - Exception while exporting Span. -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connection.py", line 198, in _new_conn - sock = connection.create_connection( - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/util/connection.py", line 60, in create_connection - for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM): - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/opt/homebrew/Cellar/python@3.12/3.12.9/Frameworks/Python.framework/Versions/3.12/lib/python3.12/socket.py", line 978, in getaddrinfo - for res in _socket.getaddrinfo(host, port, family, type, proto, flags): - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ -socket.gaierror: [Errno 8] nodename nor servname provided, or not known - -The above exception was the direct cause of the following exception: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 787, in urlopen - response = self._make_request( - ^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 488, in _make_request - raise new_e - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 464, in _make_request - self._validate_conn(conn) - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 1093, in _validate_conn - conn.connect() - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connection.py", line 704, in connect - self.sock = sock = self._new_conn() - ^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connection.py", line 205, in _new_conn - raise NameResolutionError(self.host, self, e) from e -urllib3.exceptions.NameResolutionError: : Failed to resolve 'telemetry.crewai.com' ([Errno 8] nodename nor servname provided, or not known) - -The above exception was the direct cause of the following exception: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/adapters.py", line 667, in send - resp = conn.urlopen( - ^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 841, in urlopen - retries = retries.increment( - ^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/util/retry.py", line 519, in increment - raise MaxRetryError(_pool, url, reason) from reason # type: ignore[arg-type] - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ -urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='telemetry.crewai.com', port=4319): Max retries exceeded with url: /v1/traces (Caused by NameResolutionError(": Failed to resolve 'telemetry.crewai.com' ([Errno 8] nodename nor servname provided, or not known)")) - -During handling of the above exception, another exception occurred: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/opentelemetry/exporter/otlp/proto/http/trace_exporter/__init__.py", line 139, in _export - resp = self._session.post( - ^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 637, in post - return self.request("POST", url, data=data, json=json, **kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 589, in request - resp = self.send(prep, **send_kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 703, in send - r = adapter.send(request, **kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/adapters.py", line 700, in send - raise ConnectionError(e, request=request) -requests.exceptions.ConnectionError: HTTPSConnectionPool(host='telemetry.crewai.com', port=4319): Max retries exceeded with url: /v1/traces (Caused by NameResolutionError(": Failed to resolve 'telemetry.crewai.com' ([Errno 8] nodename nor servname provided, or not known)")) - -During handling of the above exception, another exception occurred: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connection.py", line 198, in _new_conn - sock = connection.create_connection( - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/util/connection.py", line 60, in create_connection - for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM): - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/opt/homebrew/Cellar/python@3.12/3.12.9/Frameworks/Python.framework/Versions/3.12/lib/python3.12/socket.py", line 978, in getaddrinfo - for res in _socket.getaddrinfo(host, port, family, type, proto, flags): - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ -socket.gaierror: [Errno 8] nodename nor servname provided, or not known - -The above exception was the direct cause of the following exception: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 787, in urlopen - response = self._make_request( - ^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 488, in _make_request - raise new_e - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 464, in _make_request - self._validate_conn(conn) - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 1093, in _validate_conn - conn.connect() - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connection.py", line 704, in connect - self.sock = sock = self._new_conn() - ^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connection.py", line 205, in _new_conn - raise NameResolutionError(self.host, self, e) from e -urllib3.exceptions.NameResolutionError: : Failed to resolve 'telemetry.crewai.com' ([Errno 8] nodename nor servname provided, or not known) - -The above exception was the direct cause of the following exception: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/adapters.py", line 667, in send - resp = conn.urlopen( - ^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 841, in urlopen - retries = retries.increment( - ^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/util/retry.py", line 519, in increment - raise MaxRetryError(_pool, url, reason) from reason # type: ignore[arg-type] - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ -urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='telemetry.crewai.com', port=4319): Max retries exceeded with url: /v1/traces (Caused by NameResolutionError(": Failed to resolve 'telemetry.crewai.com' ([Errno 8] nodename nor servname provided, or not known)")) - -During handling of the above exception, another exception occurred: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/opentelemetry/sdk/_shared_internal/__init__.py", line 152, in _export - self._exporter.export( - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/opentelemetry/exporter/otlp/proto/http/trace_exporter/__init__.py", line 204, in export - return self._export_serialized_spans(serialized_data) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/opentelemetry/exporter/otlp/proto/http/trace_exporter/__init__.py", line 174, in _export_serialized_spans - resp = self._export(serialized_data) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/opentelemetry/exporter/otlp/proto/http/trace_exporter/__init__.py", line 147, in _export - resp = self._session.post( - ^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 637, in post - return self.request("POST", url, data=data, json=json, **kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 589, in request - resp = self.send(prep, **send_kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 703, in send - r = adapter.send(request, **kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/adapters.py", line 700, in send - raise ConnectionError(e, request=request) -requests.exceptions.ConnectionError: HTTPSConnectionPool(host='telemetry.crewai.com', port=4319): Max retries exceeded with url: /v1/traces (Caused by NameResolutionError(": Failed to resolve 'telemetry.crewai.com' ([Errno 8] nodename nor servname provided, or not known)")) -2025-07-29 18:13:22,538 - opentelemetry.sdk._shared_internal - ERROR - Exception while exporting Span. -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connection.py", line 198, in _new_conn - sock = connection.create_connection( - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/util/connection.py", line 60, in create_connection - for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM): - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/opt/homebrew/Cellar/python@3.12/3.12.9/Frameworks/Python.framework/Versions/3.12/lib/python3.12/socket.py", line 978, in getaddrinfo - for res in _socket.getaddrinfo(host, port, family, type, proto, flags): - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ -socket.gaierror: [Errno 8] nodename nor servname provided, or not known - -The above exception was the direct cause of the following exception: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 787, in urlopen - response = self._make_request( - ^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 488, in _make_request - raise new_e - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 464, in _make_request - self._validate_conn(conn) - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 1093, in _validate_conn - conn.connect() - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connection.py", line 704, in connect - self.sock = sock = self._new_conn() - ^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connection.py", line 205, in _new_conn - raise NameResolutionError(self.host, self, e) from e -urllib3.exceptions.NameResolutionError: : Failed to resolve 'cloud.langfuse.com' ([Errno 8] nodename nor servname provided, or not known) - -The above exception was the direct cause of the following exception: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/adapters.py", line 667, in send - resp = conn.urlopen( - ^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 841, in urlopen - retries = retries.increment( - ^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/util/retry.py", line 519, in increment - raise MaxRetryError(_pool, url, reason) from reason # type: ignore[arg-type] - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ -urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='cloud.langfuse.com', port=443): Max retries exceeded with url: /api/public/otel/v1/traces (Caused by NameResolutionError(": Failed to resolve 'cloud.langfuse.com' ([Errno 8] nodename nor servname provided, or not known)")) - -During handling of the above exception, another exception occurred: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/opentelemetry/exporter/otlp/proto/http/trace_exporter/__init__.py", line 139, in _export - resp = self._session.post( - ^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 637, in post - return self.request("POST", url, data=data, json=json, **kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 589, in request - resp = self.send(prep, **send_kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 703, in send - r = adapter.send(request, **kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/adapters.py", line 700, in send - raise ConnectionError(e, request=request) -requests.exceptions.ConnectionError: HTTPSConnectionPool(host='cloud.langfuse.com', port=443): Max retries exceeded with url: /api/public/otel/v1/traces (Caused by NameResolutionError(": Failed to resolve 'cloud.langfuse.com' ([Errno 8] nodename nor servname provided, or not known)")) - -During handling of the above exception, another exception occurred: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connection.py", line 198, in _new_conn - sock = connection.create_connection( - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/util/connection.py", line 60, in create_connection - for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM): - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/opt/homebrew/Cellar/python@3.12/3.12.9/Frameworks/Python.framework/Versions/3.12/lib/python3.12/socket.py", line 978, in getaddrinfo - for res in _socket.getaddrinfo(host, port, family, type, proto, flags): - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ -socket.gaierror: [Errno 8] nodename nor servname provided, or not known - -The above exception was the direct cause of the following exception: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 787, in urlopen - response = self._make_request( - ^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 488, in _make_request - raise new_e - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 464, in _make_request - self._validate_conn(conn) - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 1093, in _validate_conn - conn.connect() - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connection.py", line 704, in connect - self.sock = sock = self._new_conn() - ^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connection.py", line 205, in _new_conn - raise NameResolutionError(self.host, self, e) from e -urllib3.exceptions.NameResolutionError: : Failed to resolve 'cloud.langfuse.com' ([Errno 8] nodename nor servname provided, or not known) - -The above exception was the direct cause of the following exception: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/adapters.py", line 667, in send - resp = conn.urlopen( - ^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 841, in urlopen - retries = retries.increment( - ^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/util/retry.py", line 519, in increment - raise MaxRetryError(_pool, url, reason) from reason # type: ignore[arg-type] - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ -urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='cloud.langfuse.com', port=443): Max retries exceeded with url: /api/public/otel/v1/traces (Caused by NameResolutionError(": Failed to resolve 'cloud.langfuse.com' ([Errno 8] nodename nor servname provided, or not known)")) - -During handling of the above exception, another exception occurred: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/opentelemetry/sdk/_shared_internal/__init__.py", line 152, in _export - self._exporter.export( - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/opentelemetry/exporter/otlp/proto/http/trace_exporter/__init__.py", line 204, in export - return self._export_serialized_spans(serialized_data) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/opentelemetry/exporter/otlp/proto/http/trace_exporter/__init__.py", line 174, in _export_serialized_spans - resp = self._export(serialized_data) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/opentelemetry/exporter/otlp/proto/http/trace_exporter/__init__.py", line 147, in _export - resp = self._session.post( - ^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 637, in post - return self.request("POST", url, data=data, json=json, **kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 589, in request - resp = self.send(prep, **send_kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 703, in send - r = adapter.send(request, **kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/adapters.py", line 700, in send - raise ConnectionError(e, request=request) -requests.exceptions.ConnectionError: HTTPSConnectionPool(host='cloud.langfuse.com', port=443): Max retries exceeded with url: /api/public/otel/v1/traces (Caused by NameResolutionError(": Failed to resolve 'cloud.langfuse.com' ([Errno 8] nodename nor servname provided, or not known)")) -2025-07-29 18:16:02,522 - openlit - INFO - Starting openLIT initialization... -2025-07-29 18:16:02,527 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-29 18:17:41,494 - openlit.__helpers - ERROR - Unexpected error occurred while fetching pricing info: HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: /openlit/openlit/main/assets/pricing.json (Caused by ConnectTimeoutError(, 'Connection to raw.githubusercontent.com timed out. (connect timeout=20)')) -2025-07-29 18:17:41,918 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-29 18:17:41,951 - openlit - INFO - Library for ollama (ollama) not found. Skipping instrumentation -2025-07-29 18:17:41,951 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-29 18:17:41,951 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-29 18:17:41,951 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-29 18:17:42,223 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-29 18:17:42,406 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-29 18:17:42,406 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-29 18:17:42,971 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-29 18:17:42,972 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-29 18:17:42,973 - openlit - INFO - Library for transformers (transformers) not found. Skipping instrumentation -2025-07-29 18:17:42,976 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-29 18:17:42,976 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-29 18:17:42,976 - openlit - INFO - Library for autogen (autogen) not found. Skipping instrumentation -2025-07-29 18:17:42,976 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-29 18:17:42,976 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-29 18:17:42,976 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-29 18:17:42,976 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-29 18:17:42,976 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-29 18:17:42,976 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-29 18:17:42,976 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-29 18:17:42,976 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-29 18:17:42,977 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-29 18:17:42,977 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-29 18:17:42,977 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-29 18:17:42,977 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-29 18:17:42,977 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-29 18:17:42,977 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-29 18:17:42,977 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-29 18:17:45,682 - opentelemetry.sdk._shared_internal - ERROR - Exception while exporting Span. -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connection.py", line 198, in _new_conn - sock = connection.create_connection( - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/util/connection.py", line 60, in create_connection - for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM): - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/opt/homebrew/Cellar/python@3.12/3.12.9/Frameworks/Python.framework/Versions/3.12/lib/python3.12/socket.py", line 978, in getaddrinfo - for res in _socket.getaddrinfo(host, port, family, type, proto, flags): - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ -socket.gaierror: [Errno 8] nodename nor servname provided, or not known - -The above exception was the direct cause of the following exception: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 787, in urlopen - response = self._make_request( - ^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 488, in _make_request - raise new_e - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 464, in _make_request - self._validate_conn(conn) - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 1093, in _validate_conn - conn.connect() - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connection.py", line 704, in connect - self.sock = sock = self._new_conn() - ^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connection.py", line 205, in _new_conn - raise NameResolutionError(self.host, self, e) from e -urllib3.exceptions.NameResolutionError: : Failed to resolve 'telemetry.crewai.com' ([Errno 8] nodename nor servname provided, or not known) - -The above exception was the direct cause of the following exception: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/adapters.py", line 667, in send - resp = conn.urlopen( - ^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 841, in urlopen - retries = retries.increment( - ^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/util/retry.py", line 519, in increment - raise MaxRetryError(_pool, url, reason) from reason # type: ignore[arg-type] - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ -urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='telemetry.crewai.com', port=4319): Max retries exceeded with url: /v1/traces (Caused by NameResolutionError(": Failed to resolve 'telemetry.crewai.com' ([Errno 8] nodename nor servname provided, or not known)")) - -During handling of the above exception, another exception occurred: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/opentelemetry/exporter/otlp/proto/http/trace_exporter/__init__.py", line 139, in _export - resp = self._session.post( - ^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 637, in post - return self.request("POST", url, data=data, json=json, **kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 589, in request - resp = self.send(prep, **send_kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 703, in send - r = adapter.send(request, **kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/adapters.py", line 700, in send - raise ConnectionError(e, request=request) -requests.exceptions.ConnectionError: HTTPSConnectionPool(host='telemetry.crewai.com', port=4319): Max retries exceeded with url: /v1/traces (Caused by NameResolutionError(": Failed to resolve 'telemetry.crewai.com' ([Errno 8] nodename nor servname provided, or not known)")) - -During handling of the above exception, another exception occurred: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connection.py", line 198, in _new_conn - sock = connection.create_connection( - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/util/connection.py", line 60, in create_connection - for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM): - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/opt/homebrew/Cellar/python@3.12/3.12.9/Frameworks/Python.framework/Versions/3.12/lib/python3.12/socket.py", line 978, in getaddrinfo - for res in _socket.getaddrinfo(host, port, family, type, proto, flags): - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ -socket.gaierror: [Errno 8] nodename nor servname provided, or not known - -The above exception was the direct cause of the following exception: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 787, in urlopen - response = self._make_request( - ^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 488, in _make_request - raise new_e - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 464, in _make_request - self._validate_conn(conn) - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 1093, in _validate_conn - conn.connect() - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connection.py", line 704, in connect - self.sock = sock = self._new_conn() - ^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connection.py", line 205, in _new_conn - raise NameResolutionError(self.host, self, e) from e -urllib3.exceptions.NameResolutionError: : Failed to resolve 'telemetry.crewai.com' ([Errno 8] nodename nor servname provided, or not known) - -The above exception was the direct cause of the following exception: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/adapters.py", line 667, in send - resp = conn.urlopen( - ^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 841, in urlopen - retries = retries.increment( - ^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/util/retry.py", line 519, in increment - raise MaxRetryError(_pool, url, reason) from reason # type: ignore[arg-type] - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ -urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='telemetry.crewai.com', port=4319): Max retries exceeded with url: /v1/traces (Caused by NameResolutionError(": Failed to resolve 'telemetry.crewai.com' ([Errno 8] nodename nor servname provided, or not known)")) - -During handling of the above exception, another exception occurred: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/opentelemetry/sdk/_shared_internal/__init__.py", line 152, in _export - self._exporter.export( - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/opentelemetry/exporter/otlp/proto/http/trace_exporter/__init__.py", line 204, in export - return self._export_serialized_spans(serialized_data) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/opentelemetry/exporter/otlp/proto/http/trace_exporter/__init__.py", line 174, in _export_serialized_spans - resp = self._export(serialized_data) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/opentelemetry/exporter/otlp/proto/http/trace_exporter/__init__.py", line 147, in _export - resp = self._session.post( - ^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 637, in post - return self.request("POST", url, data=data, json=json, **kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 589, in request - resp = self.send(prep, **send_kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 703, in send - r = adapter.send(request, **kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/adapters.py", line 700, in send - raise ConnectionError(e, request=request) -requests.exceptions.ConnectionError: HTTPSConnectionPool(host='telemetry.crewai.com', port=4319): Max retries exceeded with url: /v1/traces (Caused by NameResolutionError(": Failed to resolve 'telemetry.crewai.com' ([Errno 8] nodename nor servname provided, or not known)")) -2025-07-29 18:17:45,691 - opentelemetry.sdk._shared_internal - ERROR - Exception while exporting Span. -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connection.py", line 198, in _new_conn - sock = connection.create_connection( - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/util/connection.py", line 60, in create_connection - for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM): - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/opt/homebrew/Cellar/python@3.12/3.12.9/Frameworks/Python.framework/Versions/3.12/lib/python3.12/socket.py", line 978, in getaddrinfo - for res in _socket.getaddrinfo(host, port, family, type, proto, flags): - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ -socket.gaierror: [Errno 8] nodename nor servname provided, or not known - -The above exception was the direct cause of the following exception: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 787, in urlopen - response = self._make_request( - ^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 488, in _make_request - raise new_e - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 464, in _make_request - self._validate_conn(conn) - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 1093, in _validate_conn - conn.connect() - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connection.py", line 704, in connect - self.sock = sock = self._new_conn() - ^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connection.py", line 205, in _new_conn - raise NameResolutionError(self.host, self, e) from e -urllib3.exceptions.NameResolutionError: : Failed to resolve 'cloud.langfuse.com' ([Errno 8] nodename nor servname provided, or not known) - -The above exception was the direct cause of the following exception: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/adapters.py", line 667, in send - resp = conn.urlopen( - ^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 841, in urlopen - retries = retries.increment( - ^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/util/retry.py", line 519, in increment - raise MaxRetryError(_pool, url, reason) from reason # type: ignore[arg-type] - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ -urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='cloud.langfuse.com', port=443): Max retries exceeded with url: /api/public/otel/v1/traces (Caused by NameResolutionError(": Failed to resolve 'cloud.langfuse.com' ([Errno 8] nodename nor servname provided, or not known)")) - -During handling of the above exception, another exception occurred: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/opentelemetry/exporter/otlp/proto/http/trace_exporter/__init__.py", line 139, in _export - resp = self._session.post( - ^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 637, in post - return self.request("POST", url, data=data, json=json, **kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 589, in request - resp = self.send(prep, **send_kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 703, in send - r = adapter.send(request, **kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/adapters.py", line 700, in send - raise ConnectionError(e, request=request) -requests.exceptions.ConnectionError: HTTPSConnectionPool(host='cloud.langfuse.com', port=443): Max retries exceeded with url: /api/public/otel/v1/traces (Caused by NameResolutionError(": Failed to resolve 'cloud.langfuse.com' ([Errno 8] nodename nor servname provided, or not known)")) - -During handling of the above exception, another exception occurred: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connection.py", line 198, in _new_conn - sock = connection.create_connection( - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/util/connection.py", line 60, in create_connection - for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM): - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/opt/homebrew/Cellar/python@3.12/3.12.9/Frameworks/Python.framework/Versions/3.12/lib/python3.12/socket.py", line 978, in getaddrinfo - for res in _socket.getaddrinfo(host, port, family, type, proto, flags): - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ -socket.gaierror: [Errno 8] nodename nor servname provided, or not known - -The above exception was the direct cause of the following exception: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 787, in urlopen - response = self._make_request( - ^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 488, in _make_request - raise new_e - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 464, in _make_request - self._validate_conn(conn) - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 1093, in _validate_conn - conn.connect() - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connection.py", line 704, in connect - self.sock = sock = self._new_conn() - ^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connection.py", line 205, in _new_conn - raise NameResolutionError(self.host, self, e) from e -urllib3.exceptions.NameResolutionError: : Failed to resolve 'cloud.langfuse.com' ([Errno 8] nodename nor servname provided, or not known) - -The above exception was the direct cause of the following exception: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/adapters.py", line 667, in send - resp = conn.urlopen( - ^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/connectionpool.py", line 841, in urlopen - retries = retries.increment( - ^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/urllib3/util/retry.py", line 519, in increment - raise MaxRetryError(_pool, url, reason) from reason # type: ignore[arg-type] - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ -urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='cloud.langfuse.com', port=443): Max retries exceeded with url: /api/public/otel/v1/traces (Caused by NameResolutionError(": Failed to resolve 'cloud.langfuse.com' ([Errno 8] nodename nor servname provided, or not known)")) - -During handling of the above exception, another exception occurred: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/opentelemetry/sdk/_shared_internal/__init__.py", line 152, in _export - self._exporter.export( - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/opentelemetry/exporter/otlp/proto/http/trace_exporter/__init__.py", line 204, in export - return self._export_serialized_spans(serialized_data) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/opentelemetry/exporter/otlp/proto/http/trace_exporter/__init__.py", line 174, in _export_serialized_spans - resp = self._export(serialized_data) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/opentelemetry/exporter/otlp/proto/http/trace_exporter/__init__.py", line 147, in _export - resp = self._session.post( - ^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 637, in post - return self.request("POST", url, data=data, json=json, **kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 589, in request - resp = self.send(prep, **send_kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/sessions.py", line 703, in send - r = adapter.send(request, **kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/.venv/lib/python3.12/site-packages/requests/adapters.py", line 700, in send - raise ConnectionError(e, request=request) -requests.exceptions.ConnectionError: HTTPSConnectionPool(host='cloud.langfuse.com', port=443): Max retries exceeded with url: /api/public/otel/v1/traces (Caused by NameResolutionError(": Failed to resolve 'cloud.langfuse.com' ([Errno 8] nodename nor servname provided, or not known)")) -2025-07-31 12:35:00,314 - openlit - INFO - Starting openLIT initialization... -2025-07-31 12:35:00,331 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-31 12:35:01,077 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-31 12:35:01,133 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-31 12:35:01,133 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-31 12:35:01,133 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-31 12:35:01,543 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-31 12:35:01,667 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-31 12:35:01,667 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-31 12:35:02,331 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-31 12:35:02,332 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-31 12:35:03,875 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-31 12:35:03,878 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-31 12:35:03,878 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-31 12:35:03,879 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-31 12:35:03,879 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-31 12:35:03,879 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-31 12:35:03,879 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-31 12:35:03,879 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-31 12:35:03,880 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-31 12:35:03,880 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-31 12:35:03,880 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-31 12:35:03,880 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-31 12:35:03,880 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-31 12:35:03,880 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-31 12:35:03,880 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-31 12:35:03,880 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-31 12:35:03,880 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-31 12:35:03,880 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-31 12:35:03,880 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-31 12:35:48,088 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.seed' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:35:48,088 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.frequency_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:35:48,088 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.presence_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:35:48,088 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.stop_sequences' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:35:48,088 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.top_p' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:35:48,088 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.service_tier' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:35:48,088 - opentelemetry.attributes - WARNING - Invalid type NoneType for attribute 'gen_ai.response.system_fingerprint' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:35:48,088 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.user' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:36:07,264 - openlit - INFO - Starting openLIT initialization... -2025-07-31 12:36:07,279 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-31 12:36:07,908 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-31 12:36:08,000 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-31 12:36:08,000 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-31 12:36:08,000 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-31 12:36:08,433 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-31 12:36:08,552 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-31 12:36:08,552 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-31 12:36:09,134 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-31 12:36:09,136 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-31 12:36:10,449 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-31 12:36:10,452 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-31 12:36:10,452 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-31 12:36:10,452 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-31 12:36:10,452 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-31 12:36:10,452 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-31 12:36:10,453 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-31 12:36:10,453 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-31 12:36:10,453 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-31 12:36:10,453 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-31 12:36:10,453 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-31 12:36:10,453 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-31 12:36:10,453 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-31 12:36:10,453 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-31 12:36:10,453 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-31 12:36:10,453 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-31 12:36:10,454 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-31 12:36:10,454 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-31 12:36:10,454 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-31 12:42:09,278 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.seed' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:42:09,279 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.frequency_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:42:09,279 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.presence_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:42:09,279 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.stop_sequences' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:42:09,279 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.top_p' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:42:09,279 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.service_tier' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:42:09,279 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.user' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:47:30,194 - openlit.instrumentation.openai.async_openai - ERROR - Error in trace creation: 'content' -2025-07-31 12:48:24,829 - openlit - INFO - Starting openLIT initialization... -2025-07-31 12:48:24,847 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-31 12:48:25,660 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-31 12:48:25,712 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-31 12:48:25,712 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-31 12:48:25,712 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-31 12:48:26,087 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-31 12:48:26,247 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-31 12:48:26,247 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-31 12:48:26,849 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-31 12:48:26,850 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-31 12:48:28,315 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-31 12:48:28,318 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-31 12:48:28,318 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-31 12:48:28,319 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-31 12:48:28,319 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-31 12:48:28,319 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-31 12:48:28,319 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-31 12:48:28,319 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-31 12:48:28,319 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-31 12:48:28,319 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-31 12:48:28,319 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-31 12:48:28,319 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-31 12:48:28,319 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-31 12:48:28,320 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-31 12:48:28,320 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-31 12:48:28,320 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-31 12:48:28,320 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-31 12:48:28,320 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-31 12:48:28,320 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-31 12:53:48,928 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.seed' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:53:48,930 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.frequency_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:53:48,930 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.presence_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:53:48,930 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.stop_sequences' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:53:48,930 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.top_p' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:53:48,930 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.service_tier' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:53:48,930 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.user' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:58:46,669 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.seed' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:58:46,670 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.frequency_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:58:46,670 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.presence_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:58:46,670 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.stop_sequences' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:58:46,670 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.top_p' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:58:46,670 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.service_tier' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 12:58:46,670 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.user' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 13:04:04,130 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.seed' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 13:04:04,131 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.frequency_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 13:04:04,132 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.presence_penalty' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 13:04:04,132 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.stop_sequences' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 13:04:04,132 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.top_p' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 13:04:04,132 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.service_tier' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 13:04:04,132 - opentelemetry.attributes - WARNING - Invalid type NotGiven for attribute 'gen_ai.request.user' value. Expected one of ['bool', 'str', 'bytes', 'int', 'float'] or a sequence of those types -2025-07-31 13:09:42,758 - openlit.instrumentation.openai.async_openai - ERROR - Error in trace creation: 'content' -2025-07-31 17:02:47,598 - openlit - INFO - Starting openLIT initialization... -2025-07-31 17:02:47,613 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-31 17:02:48,239 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-31 17:02:48,292 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-31 17:02:48,292 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-31 17:02:48,292 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-31 17:02:48,658 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-31 17:02:48,776 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-31 17:02:48,776 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-31 17:02:49,388 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-31 17:02:49,389 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-31 17:02:51,186 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-31 17:02:51,188 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-31 17:02:51,188 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-31 17:02:51,189 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-31 17:02:51,189 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-31 17:02:51,189 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-31 17:02:51,189 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-31 17:02:51,189 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-31 17:02:51,189 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-31 17:02:51,189 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-31 17:02:51,189 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-31 17:02:51,190 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-31 17:02:51,190 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-31 17:02:51,190 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-31 17:02:51,190 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-31 17:02:51,190 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-31 17:02:51,190 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-31 17:02:51,190 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-31 17:02:51,190 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-31 17:03:07,672 - openlit - INFO - Starting openLIT initialization... -2025-07-31 17:03:07,688 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-31 17:03:08,644 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-31 17:03:08,687 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-31 17:03:08,687 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-31 17:03:08,687 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-31 17:03:09,021 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-31 17:03:09,132 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-31 17:03:09,132 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-31 17:03:09,697 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-31 17:03:09,698 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-31 17:03:11,016 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-31 17:03:11,018 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-31 17:03:11,018 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-31 17:03:11,019 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-31 17:03:11,019 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-31 17:03:11,019 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-31 17:03:11,019 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-31 17:03:11,019 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-31 17:03:11,019 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-31 17:03:11,019 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-31 17:03:11,019 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-31 17:03:11,019 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-31 17:03:11,019 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-31 17:03:11,019 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-31 17:03:11,020 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-31 17:03:11,020 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-31 17:03:11,020 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-31 17:03:11,020 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-31 17:03:11,020 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-31 17:03:11,983 - __main__ - INFO - Loaded 1 texts from logs/example_1.txt -2025-07-31 17:03:11,983 - agentgraph.input.text_processing.chunking_service - INFO - ChunkingService initialized with batch_size=3, model=gpt-4o-mini -2025-07-31 17:03:11,983 - __main__ - INFO - Initialized BatchKGExtractor: -2025-07-31 17:03:11,983 - __main__ - INFO - - Model: gpt-4o-mini, Method: production -2025-07-31 17:03:11,983 - __main__ - INFO - - Preprocessing: False -2025-07-31 17:03:11,983 - __main__ - INFO - - Line numbers: False -2025-07-31 17:03:11,983 - __main__ - INFO - Processing batch of 1 texts -2025-07-31 17:03:11,983 - __main__ - INFO - Processing text 1/1: text_0 -2025-07-31 17:03:11,983 - __main__ - INFO - Processing text text_0 (format: auto) -2025-07-31 17:03:11,983 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-31 17:03:11,983 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 8852 characters -2025-07-31 17:03:11,983 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-31 17:03:11,989 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-31 17:03:11,993 - agentgraph.input.content_analysis.semantic_analyzer - INFO - Detected agent trace type: crewai_execution (confidence: 0.90) -2025-07-31 17:03:12,687 - agentgraph.input.text_processing.chunking_service - INFO - Applied rule-based line splitting to 1 chunks (max_line_length=800) -2025-07-31 17:03:12,689 - agentgraph.input.text_processing.chunking_service - INFO - Assigning global line numbers to 1 chunks -2025-07-31 17:03:12,689 - agentgraph.input.text_processing.trace_line_processor - INFO - Added line numbers to 175 lines, starting from line 1 -2025-07-31 17:03:12,689 - agentgraph.input.text_processing.chunking_service - INFO - Successfully assigned global line numbers to all chunks -2025-07-31 17:03:12,690 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using agent_semantic splitter -2025-07-31 17:03:12,690 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=350000, overlap_size=17500 -2025-07-31 17:03:12,690 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context enhancement complete: 0 total documents -2025-07-31 17:03:12,690 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - SlidingWindowMonitor initialized with model: gpt-4o-mini, method: production -2025-07-31 17:03:12,690 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context documents: 0 provided -2025-07-31 17:03:12,690 - agentgraph.extraction.graph_utilities.knowledge_graph_merger - INFO - KnowledgeGraphMerger initialized with model: gpt-4o-mini -2025-07-31 17:03:12,691 - __main__ - INFO - Extracting knowledge graph from 1 chunks -2025-07-31 17:03:12,691 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks (splitter: agent_semantic, window_size=350000, overlap=0) -2025-07-31 17:03:12,691 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided source trace ID: text_0 -2025-07-31 17:03:12,691 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided processing run ID: batch_text_0 -2025-07-31 17:03:12,691 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks -2025-07-31 17:03:12,691 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using parallel processing with batch size 3 (1 batches) -2025-07-31 17:03:12,691 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 windows in 1 batches -2025-07-31 17:03:12,703 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks in 1 sub-batches (max 1 concurrent per sub-batch) -2025-07-31 17:03:12,704 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing window 0 -2025-07-31 17:03:12,704 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - 🤖 Starting async_crew extraction for window 0 with method: production using model: gpt-4o-mini -2025-07-31 17:03:12,704 - agentgraph.methods.production.multi_agent_knowledge_extractor - INFO - Creating agent monitoring crew with model: gpt-4o-mini -2025-07-31 17:04:16,182 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Result Generated (production method) -2025-07-31 17:04:16,183 - agentgraph.reconstruction.content_reference_resolver - INFO - Entity prompt resolution stats: {'total_entities': 5, 'entities_with_refs': 5, 'successful_resolutions': 5, 'failed_resolutions': 0} -2025-07-31 17:04:16,184 - agentgraph.reconstruction.content_reference_resolver - INFO - Relation prompt resolution stats: {'total_relations': 6, 'relations_with_refs': 5, 'successful_resolutions': 5, 'failed_resolutions': 0} -2025-07-31 17:04:16,184 - agentgraph.reconstruction.content_reference_resolver - INFO - Resolved content references for knowledge graph with 5 entities and 6 relations -2025-07-31 17:04:16,185 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Preparing window knowledge graphs with metadata -2025-07-31 17:04:16,185 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Merging knowledge graphs... -2025-07-31 17:04:16,185 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Only one knowledge graph generated, skipping merge process -2025-07-31 17:04:16,185 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing complete. Knowledge graph generated with 5 entities and 6 relations -2025-07-31 17:10:02,530 - openlit - INFO - Starting openLIT initialization... -2025-07-31 17:10:02,545 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-31 17:10:03,248 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-31 17:10:03,298 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-31 17:10:03,298 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-31 17:10:03,298 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-31 17:10:03,654 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-31 17:10:03,760 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-31 17:10:03,760 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-31 17:10:04,341 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-31 17:10:04,342 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-31 17:10:05,889 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-31 17:10:05,891 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-31 17:10:05,892 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-31 17:10:05,892 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-31 17:10:05,892 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-31 17:10:05,892 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-31 17:10:05,892 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-31 17:10:05,892 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-31 17:10:05,893 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-31 17:10:05,893 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-31 17:10:05,893 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-31 17:10:05,893 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-31 17:10:05,893 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-31 17:10:05,893 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-31 17:10:05,893 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-31 17:10:05,893 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-31 17:10:05,893 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-31 17:10:05,893 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-31 17:10:05,893 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-31 17:10:07,010 - __main__ - INFO - Loaded 1 texts from logs/example_1.txt -2025-07-31 17:10:07,010 - agentgraph.input.text_processing.chunking_service - INFO - ChunkingService initialized with batch_size=3, model=gpt-4o-mini -2025-07-31 17:10:07,010 - __main__ - INFO - Initialized BatchKGExtractor: -2025-07-31 17:10:07,010 - __main__ - INFO - - Model: gpt-4o-mini, Method: production -2025-07-31 17:10:07,010 - __main__ - INFO - - Preprocessing: False -2025-07-31 17:10:07,010 - __main__ - INFO - - Line numbers: False -2025-07-31 17:10:07,010 - __main__ - INFO - Processing batch of 1 texts -2025-07-31 17:10:07,010 - __main__ - INFO - Processing text 1/1: text_0 -2025-07-31 17:10:07,010 - __main__ - INFO - Processing text text_0 (format: auto) -2025-07-31 17:10:07,010 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-31 17:10:07,010 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 8852 characters -2025-07-31 17:10:07,010 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-31 17:10:07,016 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-31 17:10:07,020 - agentgraph.input.content_analysis.semantic_analyzer - INFO - Detected agent trace type: crewai_execution (confidence: 0.90) -2025-07-31 17:10:07,573 - agentgraph.input.text_processing.chunking_service - INFO - Applied rule-based line splitting to 1 chunks (max_line_length=800) -2025-07-31 17:10:07,575 - agentgraph.input.text_processing.chunking_service - INFO - Assigning global line numbers to 1 chunks -2025-07-31 17:10:07,576 - agentgraph.input.text_processing.trace_line_processor - INFO - Added line numbers to 175 lines, starting from line 1 -2025-07-31 17:10:07,576 - agentgraph.input.text_processing.chunking_service - INFO - Successfully assigned global line numbers to all chunks -2025-07-31 17:10:07,576 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using agent_semantic splitter -2025-07-31 17:10:07,576 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=350000, overlap_size=17500 -2025-07-31 17:10:07,576 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context enhancement complete: 0 total documents -2025-07-31 17:10:07,576 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - SlidingWindowMonitor initialized with model: gpt-4o-mini, method: production -2025-07-31 17:10:07,576 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context documents: 0 provided -2025-07-31 17:10:07,576 - agentgraph.extraction.graph_utilities.knowledge_graph_merger - INFO - KnowledgeGraphMerger initialized with model: gpt-4o-mini -2025-07-31 17:10:07,578 - __main__ - INFO - Extracting knowledge graph from 1 chunks -2025-07-31 17:10:07,578 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks (splitter: agent_semantic, window_size=350000, overlap=0) -2025-07-31 17:10:07,578 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided source trace ID: text_0 -2025-07-31 17:10:07,578 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided processing run ID: batch_text_0 -2025-07-31 17:10:07,578 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks -2025-07-31 17:10:07,578 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using parallel processing with batch size 3 (1 batches) -2025-07-31 17:10:07,578 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 windows in 1 batches -2025-07-31 17:10:07,589 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks in 1 sub-batches (max 1 concurrent per sub-batch) -2025-07-31 17:10:07,590 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing window 0 -2025-07-31 17:10:07,590 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - 🤖 Starting async_crew extraction for window 0 with method: production using model: gpt-4o-mini -2025-07-31 17:10:07,590 - agentgraph.methods.production.multi_agent_knowledge_extractor - INFO - Creating agent monitoring crew with model: gpt-4o-mini -2025-07-31 17:11:08,933 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Result Generated (production method) -2025-07-31 17:11:08,934 - agentgraph.reconstruction.content_reference_resolver - INFO - Entity prompt resolution stats: {'total_entities': 5, 'entities_with_refs': 5, 'successful_resolutions': 5, 'failed_resolutions': 0} -2025-07-31 17:11:08,935 - agentgraph.reconstruction.content_reference_resolver - INFO - Relation prompt resolution stats: {'total_relations': 7, 'relations_with_refs': 4, 'successful_resolutions': 4, 'failed_resolutions': 0} -2025-07-31 17:11:08,935 - agentgraph.reconstruction.content_reference_resolver - INFO - Resolved content references for knowledge graph with 5 entities and 7 relations -2025-07-31 17:11:08,935 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Preparing window knowledge graphs with metadata -2025-07-31 17:11:08,935 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Merging knowledge graphs... -2025-07-31 17:11:08,936 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Only one knowledge graph generated, skipping merge process -2025-07-31 17:11:08,936 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing complete. Knowledge graph generated with 5 entities and 7 relations -2025-07-31 17:17:21,808 - openlit - INFO - Starting openLIT initialization... -2025-07-31 17:17:21,824 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-31 17:17:22,601 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-31 17:17:22,652 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-31 17:17:22,652 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-31 17:17:22,652 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-31 17:17:23,025 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-31 17:17:23,136 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-31 17:17:23,136 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-31 17:17:23,725 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-31 17:17:23,725 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-31 17:17:25,089 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-31 17:17:25,091 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-31 17:17:25,091 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-31 17:17:25,092 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-31 17:17:25,092 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-31 17:17:25,092 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-31 17:17:25,092 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-31 17:17:25,092 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-31 17:17:25,092 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-31 17:17:25,092 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-31 17:17:25,092 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-31 17:17:25,092 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-31 17:17:25,092 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-31 17:17:25,092 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-31 17:17:25,093 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-31 17:17:25,093 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-31 17:17:25,093 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-31 17:17:25,093 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-31 17:17:25,093 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-31 17:17:26,041 - __main__ - INFO - Loaded 1 texts from logs/example_1.txt -2025-07-31 17:17:26,041 - agentgraph.input.text_processing.chunking_service - INFO - ChunkingService initialized with batch_size=3, model=gpt-4o-mini -2025-07-31 17:17:26,041 - __main__ - INFO - Initialized BatchKGExtractor: -2025-07-31 17:17:26,041 - __main__ - INFO - - Model: gpt-4o-mini, Method: production -2025-07-31 17:17:26,041 - __main__ - INFO - - Preprocessing: False -2025-07-31 17:17:26,041 - __main__ - INFO - - Line numbers: False -2025-07-31 17:17:26,041 - __main__ - INFO - Processing batch of 1 texts -2025-07-31 17:17:26,041 - __main__ - INFO - Processing text 1/1: text_0 -2025-07-31 17:17:26,041 - __main__ - INFO - Processing text text_0 (format: auto) -2025-07-31 17:17:26,041 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-31 17:17:26,041 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 8852 characters -2025-07-31 17:17:26,041 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-31 17:17:26,047 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-31 17:17:26,051 - agentgraph.input.content_analysis.semantic_analyzer - INFO - Detected agent trace type: crewai_execution (confidence: 0.90) -2025-07-31 17:17:26,565 - agentgraph.input.text_processing.chunking_service - INFO - Applied rule-based line splitting to 1 chunks (max_line_length=800) -2025-07-31 17:17:26,567 - agentgraph.input.text_processing.chunking_service - INFO - Assigning global line numbers to 1 chunks -2025-07-31 17:17:26,568 - agentgraph.input.text_processing.trace_line_processor - INFO - Added line numbers to 175 lines, starting from line 1 -2025-07-31 17:17:26,568 - agentgraph.input.text_processing.chunking_service - INFO - Successfully assigned global line numbers to all chunks -2025-07-31 17:17:26,568 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using agent_semantic splitter -2025-07-31 17:17:26,568 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=350000, overlap_size=17500 -2025-07-31 17:17:26,568 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context enhancement complete: 0 total documents -2025-07-31 17:17:26,568 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - SlidingWindowMonitor initialized with model: gpt-4o-mini, method: production -2025-07-31 17:17:26,568 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context documents: 0 provided -2025-07-31 17:17:26,568 - agentgraph.extraction.graph_utilities.knowledge_graph_merger - INFO - KnowledgeGraphMerger initialized with model: gpt-4o-mini -2025-07-31 17:17:26,570 - __main__ - INFO - Extracting knowledge graph from 1 chunks -2025-07-31 17:17:26,570 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks (splitter: agent_semantic, window_size=350000, overlap=0) -2025-07-31 17:17:26,570 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided source trace ID: text_0 -2025-07-31 17:17:26,570 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided processing run ID: batch_text_0 -2025-07-31 17:17:26,570 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks -2025-07-31 17:17:26,570 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using parallel processing with batch size 3 (1 batches) -2025-07-31 17:17:26,570 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 windows in 1 batches -2025-07-31 17:17:26,579 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks in 1 sub-batches (max 1 concurrent per sub-batch) -2025-07-31 17:17:26,580 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing window 0 -2025-07-31 17:17:26,580 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - 🤖 Starting async_crew extraction for window 0 with method: production using model: gpt-4o-mini -2025-07-31 17:17:26,580 - agentgraph.methods.production.multi_agent_knowledge_extractor - INFO - Creating agent monitoring crew with model: gpt-4o-mini -2025-07-31 17:20:05,281 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Result Generated (production method) -2025-07-31 17:20:05,282 - agentgraph.extraction.graph_processing.knowledge_graph_processor - WARNING - Failed to resolve content references for window 0: 1 validation error for Entity -type - Input should be 'Agent', 'Task', 'Tool', 'Input', 'Output' or 'Human' [type=literal_error, input_value='Crew', input_type=str] - For further information visit https://errors.pydantic.dev/2.11/v/literal_error -2025-07-31 17:20:05,282 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Preparing window knowledge graphs with metadata -2025-07-31 17:20:05,282 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Merging knowledge graphs... -2025-07-31 17:20:05,282 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Only one knowledge graph generated, skipping merge process -2025-07-31 17:20:05,282 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing complete. Knowledge graph generated with 7 entities and 7 relations -2025-07-31 17:24:03,556 - openlit - INFO - Starting openLIT initialization... -2025-07-31 17:24:03,573 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-31 17:24:04,232 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-31 17:24:04,288 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-31 17:24:04,289 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-31 17:24:04,289 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-31 17:24:04,663 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-31 17:24:04,777 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-31 17:24:04,777 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-31 17:24:05,376 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-31 17:24:05,377 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-31 17:24:06,759 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-31 17:24:06,761 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-31 17:24:06,762 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-31 17:24:06,762 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-31 17:24:06,762 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-31 17:24:06,762 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-31 17:24:06,762 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-31 17:24:06,762 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-31 17:24:06,762 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-31 17:24:06,763 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-31 17:24:06,763 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-31 17:24:06,763 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-31 17:24:06,763 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-31 17:24:06,763 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-31 17:24:06,763 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-31 17:24:06,763 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-31 17:24:06,763 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-31 17:24:06,763 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-31 17:24:06,763 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-31 17:24:07,907 - __main__ - INFO - Loaded 1 texts from logs/example_1.txt -2025-07-31 17:24:07,907 - agentgraph.input.text_processing.chunking_service - INFO - ChunkingService initialized with batch_size=3, model=gpt-4o-mini -2025-07-31 17:24:07,907 - __main__ - INFO - Initialized BatchKGExtractor: -2025-07-31 17:24:07,907 - __main__ - INFO - - Model: gpt-4o-mini, Method: production -2025-07-31 17:24:07,907 - __main__ - INFO - - Preprocessing: False -2025-07-31 17:24:07,907 - __main__ - INFO - - Line numbers: False -2025-07-31 17:24:07,907 - __main__ - INFO - Processing batch of 1 texts -2025-07-31 17:24:07,907 - __main__ - INFO - Processing text 1/1: text_0 -2025-07-31 17:24:07,907 - __main__ - INFO - Processing text text_0 (format: auto) -2025-07-31 17:24:07,907 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-31 17:24:07,907 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 8852 characters -2025-07-31 17:24:07,907 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-31 17:24:07,913 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-31 17:24:07,917 - agentgraph.input.content_analysis.semantic_analyzer - INFO - Detected agent trace type: crewai_execution (confidence: 0.90) -2025-07-31 17:24:08,408 - agentgraph.input.text_processing.chunking_service - INFO - Applied rule-based line splitting to 1 chunks (max_line_length=800) -2025-07-31 17:24:08,410 - agentgraph.input.text_processing.chunking_service - INFO - Assigning global line numbers to 1 chunks -2025-07-31 17:24:08,411 - agentgraph.input.text_processing.trace_line_processor - INFO - Added line numbers to 175 lines, starting from line 1 -2025-07-31 17:24:08,411 - agentgraph.input.text_processing.chunking_service - INFO - Successfully assigned global line numbers to all chunks -2025-07-31 17:24:08,411 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using agent_semantic splitter -2025-07-31 17:24:08,411 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=350000, overlap_size=17500 -2025-07-31 17:24:08,411 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context enhancement complete: 0 total documents -2025-07-31 17:24:08,411 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - SlidingWindowMonitor initialized with model: gpt-4o-mini, method: production -2025-07-31 17:24:08,411 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context documents: 0 provided -2025-07-31 17:24:08,411 - agentgraph.extraction.graph_utilities.knowledge_graph_merger - INFO - KnowledgeGraphMerger initialized with model: gpt-4o-mini -2025-07-31 17:24:08,413 - __main__ - INFO - Extracting knowledge graph from 1 chunks -2025-07-31 17:24:08,413 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks (splitter: agent_semantic, window_size=350000, overlap=0) -2025-07-31 17:24:08,413 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided source trace ID: text_0 -2025-07-31 17:24:08,413 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided processing run ID: batch_text_0 -2025-07-31 17:24:08,413 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks -2025-07-31 17:24:08,413 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using parallel processing with batch size 3 (1 batches) -2025-07-31 17:24:08,413 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 windows in 1 batches -2025-07-31 17:24:08,423 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks in 1 sub-batches (max 1 concurrent per sub-batch) -2025-07-31 17:24:08,424 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing window 0 -2025-07-31 17:24:08,424 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - 🤖 Starting async_crew extraction for window 0 with method: production using model: gpt-4o-mini -2025-07-31 17:24:08,424 - agentgraph.methods.production.multi_agent_knowledge_extractor - INFO - Creating agent monitoring crew with model: gpt-4o-mini -2025-07-31 17:27:53,945 - openlit - INFO - Starting openLIT initialization... -2025-07-31 17:27:53,961 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-31 17:27:54,547 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-31 17:27:54,595 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-31 17:27:54,595 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-31 17:27:54,595 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-31 17:27:55,030 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-31 17:27:55,136 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-31 17:27:55,136 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-31 17:27:55,708 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-31 17:27:55,708 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-31 17:27:57,272 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-31 17:27:57,274 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-31 17:27:57,275 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-31 17:27:57,275 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-31 17:27:57,275 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-31 17:27:57,275 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-31 17:27:57,275 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-31 17:27:57,275 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-31 17:27:57,275 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-31 17:27:57,275 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-31 17:27:57,275 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-31 17:27:57,275 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-31 17:27:57,276 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-31 17:27:57,276 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-31 17:27:57,276 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-31 17:27:57,276 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-31 17:27:57,276 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-31 17:27:57,276 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-31 17:27:57,276 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-31 17:27:58,261 - __main__ - INFO - Loaded 1 texts from logs/example_1.txt -2025-07-31 17:27:58,262 - agentgraph.input.text_processing.chunking_service - INFO - ChunkingService initialized with batch_size=3, model=gpt-4o-mini -2025-07-31 17:27:58,262 - __main__ - INFO - Initialized BatchKGExtractor: -2025-07-31 17:27:58,262 - __main__ - INFO - - Model: gpt-4o-mini, Method: production -2025-07-31 17:27:58,262 - __main__ - INFO - - Preprocessing: False -2025-07-31 17:27:58,262 - __main__ - INFO - - Line numbers: False -2025-07-31 17:27:58,262 - __main__ - INFO - Processing batch of 1 texts -2025-07-31 17:27:58,262 - __main__ - INFO - Processing text 1/1: text_0 -2025-07-31 17:27:58,262 - __main__ - INFO - Processing text text_0 (format: auto) -2025-07-31 17:27:58,262 - agentgraph.input.text_processing.chunking_service - INFO - Chunking trace content with agent_semantic splitter -2025-07-31 17:27:58,262 - agentgraph.input.text_processing.chunking_service - INFO - Content length: 8852 characters -2025-07-31 17:27:58,262 - agentgraph.input.text_processing.chunking_service - INFO - Using provided parameters: window_size=350000, overlap_size=17500 -2025-07-31 17:27:58,268 - agentgraph.input.text_processing.chunking_service - INFO - Created AgentAwareSemanticSplitter with window_size=350000, overlap_ratio=0.05 -2025-07-31 17:27:58,272 - agentgraph.input.content_analysis.semantic_analyzer - INFO - Detected agent trace type: crewai_execution (confidence: 0.90) -2025-07-31 17:27:58,758 - agentgraph.input.text_processing.chunking_service - INFO - Applied rule-based line splitting to 1 chunks (max_line_length=800) -2025-07-31 17:27:58,759 - agentgraph.input.text_processing.chunking_service - INFO - Assigning global line numbers to 1 chunks -2025-07-31 17:27:58,760 - agentgraph.input.text_processing.trace_line_processor - INFO - Added line numbers to 175 lines, starting from line 1 -2025-07-31 17:27:58,760 - agentgraph.input.text_processing.chunking_service - INFO - Successfully assigned global line numbers to all chunks -2025-07-31 17:27:58,760 - agentgraph.input.text_processing.chunking_service - INFO - Split content into 1 chunks using agent_semantic splitter -2025-07-31 17:27:58,760 - agentgraph.input.text_processing.chunking_service - INFO - Parameters used: window_size=350000, overlap_size=17500 -2025-07-31 17:27:58,760 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context enhancement complete: 0 total documents -2025-07-31 17:27:58,760 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - SlidingWindowMonitor initialized with model: gpt-4o-mini, method: production -2025-07-31 17:27:58,760 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Context documents: 0 provided -2025-07-31 17:27:58,761 - agentgraph.extraction.graph_utilities.knowledge_graph_merger - INFO - KnowledgeGraphMerger initialized with model: gpt-4o-mini -2025-07-31 17:27:58,763 - __main__ - INFO - Extracting knowledge graph from 1 chunks -2025-07-31 17:27:58,763 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks (splitter: agent_semantic, window_size=350000, overlap=0) -2025-07-31 17:27:58,763 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided source trace ID: text_0 -2025-07-31 17:27:58,763 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using provided processing run ID: batch_text_0 -2025-07-31 17:27:58,763 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks -2025-07-31 17:27:58,763 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Using parallel processing with batch size 3 (1 batches) -2025-07-31 17:27:58,763 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 windows in 1 batches -2025-07-31 17:27:58,774 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing 1 chunks in 1 sub-batches (max 1 concurrent per sub-batch) -2025-07-31 17:27:58,774 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing window 0 -2025-07-31 17:27:58,774 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - 🤖 Starting async_crew extraction for window 0 with method: production using model: gpt-4o-mini -2025-07-31 17:27:58,774 - agentgraph.methods.production.multi_agent_knowledge_extractor - INFO - Creating agent monitoring crew with model: gpt-4o-mini -2025-07-31 17:28:50,554 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Result Generated (production method) -2025-07-31 17:28:50,556 - agentgraph.reconstruction.content_reference_resolver - INFO - Entity prompt resolution stats: {'total_entities': 6, 'entities_with_refs': 6, 'successful_resolutions': 6, 'failed_resolutions': 0} -2025-07-31 17:28:50,556 - agentgraph.reconstruction.content_reference_resolver - INFO - Relation prompt resolution stats: {'total_relations': 7, 'relations_with_refs': 5, 'successful_resolutions': 5, 'failed_resolutions': 0} -2025-07-31 17:28:50,557 - agentgraph.reconstruction.content_reference_resolver - INFO - Resolved content references for knowledge graph with 6 entities and 7 relations -2025-07-31 17:28:50,557 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Preparing window knowledge graphs with metadata -2025-07-31 17:28:50,557 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Merging knowledge graphs... -2025-07-31 17:28:50,557 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Only one knowledge graph generated, skipping merge process -2025-07-31 17:28:50,557 - agentgraph.extraction.graph_processing.knowledge_graph_processor - INFO - Processing complete. Knowledge graph generated with 6 entities and 7 relations -2025-08-03 23:34:00,375 - openlit - INFO - Starting openLIT initialization... -2025-08-03 23:34:00,392 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-08-03 23:34:00,998 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-08-03 23:34:01,047 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-08-03 23:34:01,047 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-08-03 23:34:01,047 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-08-03 23:34:01,399 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-08-03 23:34:01,504 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-08-03 23:34:01,504 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-08-03 23:34:02,082 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-08-03 23:34:02,083 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-08-03 23:34:03,396 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-08-03 23:34:03,399 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-08-03 23:34:03,399 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-08-03 23:34:03,399 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-08-03 23:34:03,399 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-08-03 23:34:03,399 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-08-03 23:34:03,399 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-08-03 23:34:03,399 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-08-03 23:34:03,400 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-08-03 23:34:03,400 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-08-03 23:34:03,400 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-08-03 23:34:03,400 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-08-03 23:34:03,400 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-08-03 23:34:03,400 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-08-03 23:34:03,400 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-08-03 23:34:03,400 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-08-03 23:34:03,400 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-08-03 23:34:03,400 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-08-03 23:34:03,400 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-08-04 01:34:31,260 - openlit - INFO - Starting openLIT initialization... -2025-08-04 01:34:31,275 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-08-04 01:34:31,916 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-08-04 01:34:31,962 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-08-04 01:34:31,963 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-08-04 01:34:31,963 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-08-04 01:34:32,522 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-08-04 01:34:32,785 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-08-04 01:34:32,785 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-08-04 01:34:33,852 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-08-04 01:34:33,855 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-08-04 01:34:35,252 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-08-04 01:34:35,254 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-08-04 01:34:35,254 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-08-04 01:34:35,255 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-08-04 01:34:35,255 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-08-04 01:34:35,255 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-08-04 01:34:35,255 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-08-04 01:34:35,255 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-08-04 01:34:35,255 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-08-04 01:34:35,255 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-08-04 01:34:35,255 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-08-04 01:34:35,255 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-08-04 01:34:35,255 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-08-04 01:34:35,256 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-08-04 01:34:35,256 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-08-04 01:34:35,256 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-08-04 01:34:35,256 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-08-04 01:34:35,256 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-08-04 01:34:35,256 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-08-04 01:34:36,112 - agentgraph.reconstruction.prompt_reconstructor - INFO - Successfully initialized PromptReconstructor with 3 entities and 2 relations -2025-08-04 02:19:35,586 - openlit - INFO - Starting openLIT initialization... -2025-08-04 02:19:35,608 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-08-04 02:19:36,342 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-08-04 02:19:36,406 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-08-04 02:19:36,406 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-08-04 02:19:36,406 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-08-04 02:19:36,853 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-08-04 02:19:36,993 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-08-04 02:19:36,993 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-08-04 02:19:37,703 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-08-04 02:19:37,704 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-08-04 02:19:39,432 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-08-04 02:19:39,435 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-08-04 02:19:39,435 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-08-04 02:19:39,436 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-08-04 02:19:39,436 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-08-04 02:19:39,436 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-08-04 02:19:39,436 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-08-04 02:19:39,436 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-08-04 02:19:39,436 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-08-04 02:19:39,437 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-08-04 02:19:39,437 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-08-04 02:19:39,437 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-08-04 02:19:39,437 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-08-04 02:19:39,437 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-08-04 02:19:39,437 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-08-04 02:19:39,437 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-08-04 02:19:39,437 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-08-04 02:19:39,438 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-08-04 02:19:39,438 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-08-04 02:19:40,501 - agentgraph.reconstruction.prompt_reconstructor - INFO - Successfully initialized PromptReconstructor with 2 entities and 1 relations -2025-08-04 02:41:36,938 - openlit - INFO - Starting openLIT initialization... -2025-08-04 02:41:36,959 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-08-04 02:41:37,708 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-08-04 02:41:37,767 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-08-04 02:41:37,767 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-08-04 02:41:37,767 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-08-04 02:41:38,089 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-08-04 02:41:38,093 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-08-04 02:41:38,093 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-08-04 02:41:38,096 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-08-04 02:41:38,097 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-08-04 02:41:39,577 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-08-04 02:41:39,580 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-08-04 02:41:39,580 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-08-04 02:41:39,580 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-08-04 02:41:39,580 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-08-04 02:41:39,580 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-08-04 02:41:39,580 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-08-04 02:41:39,581 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-08-04 02:41:39,581 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-08-04 02:41:39,581 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-08-04 02:41:39,581 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-08-04 02:41:39,581 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-08-04 02:41:39,581 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-08-04 02:41:39,581 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-08-04 02:41:39,581 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-08-04 02:41:39,582 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-08-04 02:41:39,582 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-08-04 02:41:39,582 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-08-04 02:41:39,582 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-08-04 02:41:40,725 - alembic.runtime.migration - INFO - Context impl SQLiteImpl. -2025-08-04 02:41:40,726 - alembic.runtime.migration - INFO - Will assume non-transactional DDL. -2025-08-04 02:41:40,732 - agentgraph.reconstruction.rag_prompt_reconstructor - ERROR - Failed to initialize RAG tool: EmbedChain.add() missing 1 required positional argument: 'source' -2025-08-04 02:41:40,734 - agentgraph.reconstruction.prompt_reconstructor - INFO - Successfully initialized PromptReconstructor with 4 entities and 3 relations -2025-08-04 02:42:04,829 - openlit - INFO - Starting openLIT initialization... -2025-08-04 02:42:04,850 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-08-04 02:42:05,513 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-08-04 02:42:05,566 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-08-04 02:42:05,566 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-08-04 02:42:05,566 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-08-04 02:42:05,856 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-08-04 02:42:05,859 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-08-04 02:42:05,859 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-08-04 02:42:05,861 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-08-04 02:42:05,862 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-08-04 02:42:06,988 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-08-04 02:42:06,991 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-08-04 02:42:06,991 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-08-04 02:42:06,991 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-08-04 02:42:06,991 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-08-04 02:42:06,992 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-08-04 02:42:06,992 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-08-04 02:42:06,992 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-08-04 02:42:06,992 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-08-04 02:42:06,992 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-08-04 02:42:06,992 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-08-04 02:42:06,992 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-08-04 02:42:06,992 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-08-04 02:42:06,992 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-08-04 02:42:06,993 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-08-04 02:42:06,993 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-08-04 02:42:06,993 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-08-04 02:42:06,993 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-08-04 02:42:06,993 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-08-04 02:42:08,004 - alembic.runtime.migration - INFO - Context impl SQLiteImpl. -2025-08-04 02:42:08,005 - alembic.runtime.migration - INFO - Will assume non-transactional DDL. -2025-08-04 02:42:08,016 - embedchain.embedchain - INFO - Invalid data_type: 'file', using `custom` instead. - Check docs to pass the valid data type: `https://docs.embedchain.ai/data-sources/overview` -2025-08-04 02:42:08,016 - agentgraph.reconstruction.rag_prompt_reconstructor - ERROR - Failed to initialize RAG tool: Cant find the loader for DataType.CUSTOM. We recommend to pass the loader to use data_type: DataType.CUSTOM, check `https://docs.embedchain.ai/data-sources/overview`. -2025-08-04 02:42:08,018 - agentgraph.reconstruction.prompt_reconstructor - INFO - Successfully initialized PromptReconstructor with 4 entities and 3 relations -2025-08-04 02:42:34,783 - openlit - INFO - Starting openLIT initialization... -2025-08-04 02:42:34,803 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-08-04 02:42:35,466 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-08-04 02:42:35,518 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-08-04 02:42:35,519 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-08-04 02:42:35,519 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-08-04 02:42:35,809 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-08-04 02:42:35,812 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-08-04 02:42:35,812 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-08-04 02:42:35,815 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-08-04 02:42:35,815 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-08-04 02:42:36,949 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-08-04 02:42:36,951 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-08-04 02:42:36,951 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-08-04 02:42:36,952 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-08-04 02:42:36,952 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-08-04 02:42:36,952 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-08-04 02:42:36,952 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-08-04 02:42:36,952 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-08-04 02:42:36,952 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-08-04 02:42:36,952 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-08-04 02:42:36,953 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-08-04 02:42:36,953 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-08-04 02:42:36,953 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-08-04 02:42:36,953 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-08-04 02:42:36,953 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-08-04 02:42:36,953 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-08-04 02:42:36,953 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-08-04 02:42:36,954 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-08-04 02:42:36,954 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-08-04 02:42:37,901 - alembic.runtime.migration - INFO - Context impl SQLiteImpl. -2025-08-04 02:42:37,901 - alembic.runtime.migration - INFO - Will assume non-transactional DDL. -2025-08-04 02:42:37,908 - embedchain.chunkers.base_chunker - INFO - Skipping chunks smaller than 1 characters -2025-08-04 02:42:39,796 - embedchain.embedchain - INFO - Successfully saved === INPUT === -system: You are Organizer. You're an expert in organizing. -Your personal goal is: You (DataType.TEXT). New chunks count: 166 -2025-08-04 02:42:39,802 - agentgraph.reconstruction.rag_prompt_reconstructor - INFO - Successfully initialized RAG tool with trace content -2025-08-04 02:42:39,804 - agentgraph.reconstruction.rag_prompt_reconstructor - INFO - Initialized RagPromptReconstructor with 4 entities and 3 relations -2025-08-04 02:42:44,986 - embedchain.llm.base - INFO - Prompt: -You are a Q&A expert system. Your responses must always be rooted in the context provided for each query. Here are some guidelines to follow: - -1. Refrain from explicitly mentioning the context provided in your response. -2. The context should silently guide your answers without being directly acknowledged. -3. Do not use phrases such as 'According to the context provided', 'Based on the context, ...' etc. - -Context information: ----------------------- -=== INPUT === -system: You are Organizer. You're an expert in organizing. -Your personal goal is: You organize the plan and the tasks. -You ONLY have access to the following tools, and should NEVER make up tools that are not listed here: | Tool Description: Delegate a specific task to one of the following coworkers: Thinker | Tool Description: Ask a specific question to one of the following coworkers: Thinker ----------------------- - -Query: User Task Query Organizer Agent -Answer: - -2025-08-04 02:42:51,278 - embedchain.llm.base - INFO - Prompt: -You are a Q&A expert system. Your responses must always be rooted in the context provided for each query. Here are some guidelines to follow: - -1. Refrain from explicitly mentioning the context provided in your response. -2. The context should silently guide your answers without being directly acknowledged. -3. Do not use phrases such as 'According to the context provided', 'Based on the context, ...' etc. - -Context information: ----------------------- -Action: Search in a specific website -Action Input: {"search_query":"snow with 3 legs","website":"https://www.google.com"} -Observation: Relevant Content: | Thought: -assistant: Action: Search in a specific website -Action Input: {"search_query":"haiku about snow with 3 legs","website":"https://www.haiku-studio.com"} -Observation: Relevant Content: | Tool Description: A tool that can be used to semantic search a query from a specific URL content. ----------------------- - -Query: Organizer Agent Search in a specific website, uses Search in a specific website, interaction between Organizer Agent and Search in a specific website -Answer: - -2025-08-04 02:42:56,344 - embedchain.llm.base - INFO - Prompt: -You are a Q&A expert system. Your responses must always be rooted in the context provided for each query. Here are some guidelines to follow: - -1. Refrain from explicitly mentioning the context provided in your response. -2. The context should silently guide your answers without being directly acknowledged. -3. Do not use phrases such as 'According to the context provided', 'Based on the context, ...' etc. - -Context information: ----------------------- -=== INPUT === -['capital of the moon'] - -=== INPUT === -system: You are Organizer. You're an expert in organizing. -Your personal goal is: You organize the plan and the tasks. -You ONLY have access to the following tools, and should NEVER make up tools that are not listed here: | assistant: Thought: I need to clarify that the moon does not have a capital, as it does not function as a governed location like a country would. Instead, it is a celestial body. I can incorporate this information into a creative format. -Action: Ask question to coworker | === OUTPUT === -Thought: I need to clarify that the moon does not have a capital, as it does not function as a governed location like a country would. Instead, it is a celestial body. I can incorporate this information into a creative format. -Action: Ask question to coworker ----------------------- - -Query: Organizer Agent Answer Moon Capital Question -Answer: - -2025-08-04 02:42:58,528 - alembic.runtime.migration - INFO - Context impl SQLiteImpl. -2025-08-04 02:42:58,529 - alembic.runtime.migration - INFO - Will assume non-transactional DDL. -2025-08-04 02:42:58,534 - embedchain.chunkers.base_chunker - INFO - Skipping chunks smaller than 1 characters -2025-08-04 02:43:00,364 - embedchain.embedchain - INFO - Successfully saved === INPUT === -system: You are Organizer. You're an expert in organizing. -Your personal goal is: You (DataType.TEXT). New chunks count: 0 -2025-08-04 02:43:00,367 - agentgraph.reconstruction.rag_prompt_reconstructor - INFO - Successfully initialized RAG tool with trace content -2025-08-04 02:43:00,368 - agentgraph.reconstruction.rag_prompt_reconstructor - INFO - Initialized RagPromptReconstructor with 4 entities and 3 relations -2025-08-04 02:43:06,507 - embedchain.llm.base - INFO - Prompt: -You are a Q&A expert system. Your responses must always be rooted in the context provided for each query. Here are some guidelines to follow: - -1. Refrain from explicitly mentioning the context provided in your response. -2. The context should silently guide your answers without being directly acknowledged. -3. Do not use phrases such as 'According to the context provided', 'Based on the context, ...' etc. - -Context information: ----------------------- -=== INPUT === -system: You are Organizer. You're an expert in organizing. -Your personal goal is: You organize the plan and the tasks. -You ONLY have access to the following tools, and should NEVER make up tools that are not listed here: | Tool Description: Delegate a specific task to one of the following coworkers: Thinker | Tool Description: Ask a specific question to one of the following coworkers: Thinker ----------------------- - -Query: User Task Query Organizer Agent, performs Organizer Agent, interaction between User Task Query and Organizer Agent -Answer: - -2025-08-04 02:43:08,143 - agentgraph.reconstruction.prompt_reconstructor - INFO - Successfully initialized PromptReconstructor with 4 entities and 3 relations -2025-08-04 02:53:07,745 - openlit - INFO - Starting openLIT initialization... -2025-08-04 02:53:07,765 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-08-04 02:53:08,450 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-08-04 02:53:08,502 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-08-04 02:53:08,502 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-08-04 02:53:08,503 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-08-04 02:53:08,805 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-08-04 02:53:08,808 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-08-04 02:53:08,808 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-08-04 02:53:08,810 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-08-04 02:53:08,811 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-08-04 02:53:10,003 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-08-04 02:53:10,006 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-08-04 02:53:10,006 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-08-04 02:53:10,006 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-08-04 02:53:10,006 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-08-04 02:53:10,007 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-08-04 02:53:10,007 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-08-04 02:53:10,007 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-08-04 02:53:10,007 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-08-04 02:53:10,007 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-08-04 02:53:10,007 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-08-04 02:53:10,007 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-08-04 02:53:10,007 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-08-04 02:53:10,008 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-08-04 02:53:10,008 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-08-04 02:53:10,008 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-08-04 02:53:10,008 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-08-04 02:53:10,008 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-08-04 02:53:10,008 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-08-04 02:53:11,038 - alembic.runtime.migration - INFO - Context impl SQLiteImpl. -2025-08-04 02:53:11,038 - alembic.runtime.migration - INFO - Will assume non-transactional DDL. -2025-08-04 02:53:11,048 - embedchain.chunkers.base_chunker - INFO - Skipping chunks smaller than 1 characters -2025-08-04 02:53:13,016 - embedchain.embedchain - INFO - Successfully saved === INPUT === -system: You are Organizer. You're an expert in organizing. -Your personal goal is: You (DataType.TEXT). New chunks count: 0 -2025-08-04 02:53:13,024 - agentgraph.reconstruction.rag_prompt_reconstructor - INFO - Successfully initialized RAG tool with trace content -2025-08-04 02:53:13,025 - agentgraph.reconstruction.rag_prompt_reconstructor - INFO - Initialized RagPromptReconstructor with 8 entities and 7 relations -2025-08-04 02:53:25,334 - embedchain.llm.base - INFO - Prompt: -You are a Q&A expert system. Your responses must always be rooted in the context provided for each query. Here are some guidelines to follow: - -1. Refrain from explicitly mentioning the context provided in your response. -2. The context should silently guide your answers without being directly acknowledged. -3. Do not use phrases such as 'According to the context provided', 'Based on the context, ...' etc. - -Context information: ----------------------- -Thought: -assistant: Action: Search in a specific website -Action Input: {"search_query":"haiku about snow with 3 legs","website":"https://www.haiku-studio.com"} -Observation: Relevant Content: | Tool Description: A tool that can be used to semantic search a query from a specific URL content. -Tool Name: Delegate work to coworker | Thought: -assistant: Action: Search in a specific website -Action Input: {"search_query":"capital of the moon","website":"https://en.wikipedia.org"} -Observation: Relevant Content: ----------------------- - -Query: Organizer Agent Search in a specific website -Answer: - -2025-08-04 02:53:35,201 - embedchain.llm.base - INFO - Prompt: -You are a Q&A expert system. Your responses must always be rooted in the context provided for each query. Here are some guidelines to follow: - -1. Refrain from explicitly mentioning the context provided in your response. -2. The context should silently guide your answers without being directly acknowledged. -3. Do not use phrases such as 'According to the context provided', 'Based on the context, ...' etc. - -Context information: ----------------------- -=== INPUT === -['capital of the moon'] - -=== INPUT === -system: You are Organizer. You're an expert in organizing. -Your personal goal is: You organize the plan and the tasks. -You ONLY have access to the following tools, and should NEVER make up tools that are not listed here: | assistant: Thought: I need to clarify that the moon does not have a capital, as it does not function as a governed location like a country would. Instead, it is a celestial body. I can incorporate this information into a creative format. -Action: Ask question to coworker | === OUTPUT === -Thought: I need to clarify that the moon does not have a capital, as it does not function as a governed location like a country would. Instead, it is a celestial body. I can incorporate this information into a creative format. -Action: Ask question to coworker ----------------------- - -Query: Organizer Agent Answer Moon Capital Question, performs Answer Moon Capital Question, interaction between Organizer Agent and Answer Moon Capital Question -Answer: - -2025-08-04 02:53:40,622 - embedchain.llm.base - INFO - Prompt: -You are a Q&A expert system. Your responses must always be rooted in the context provided for each query. Here are some guidelines to follow: - -1. Refrain from explicitly mentioning the context provided in your response. -2. The context should silently guide your answers without being directly acknowledged. -3. Do not use phrases such as 'According to the context provided', 'Based on the context, ...' etc. - -Context information: ----------------------- -Action: Delegate work to coworker | Tool Description: Delegate a specific task to one of the following coworkers: Thinker | Tool Description: Delegate a specific task to one of the following coworkers: Writer ----------------------- - -Query: Organizer Agent Delegate work to coworker -Answer: - -2025-08-04 02:53:46,177 - embedchain.llm.base - INFO - Prompt: -You are a Q&A expert system. Your responses must always be rooted in the context provided for each query. Here are some guidelines to follow: - -1. Refrain from explicitly mentioning the context provided in your response. -2. The context should silently guide your answers without being directly acknowledged. -3. Do not use phrases such as 'According to the context provided', 'Based on the context, ...' etc. - -Context information: ----------------------- -Tool Description: Ask a specific question to one of the following coworkers: Writer | Tool Description: Ask a specific question to one of the following coworkers: Thinker | Action: Delegate work to coworker ----------------------- - -Query: Organizer Agent Ask question to coworker -Answer: - -2025-08-04 02:53:52,093 - embedchain.llm.base - INFO - Prompt: -You are a Q&A expert system. Your responses must always be rooted in the context provided for each query. Here are some guidelines to follow: - -1. Refrain from explicitly mentioning the context provided in your response. -2. The context should silently guide your answers without being directly acknowledged. -3. Do not use phrases such as 'According to the context provided', 'Based on the context, ...' etc. - -Context information: ----------------------- -Tool Description: Ask a specific question to one of the following coworkers: Thinker | Tool Description: Delegate a specific task to one of the following coworkers: Thinker | === INPUT === -system: You are Thinker. You're an expert in thinking. -Your personal goal is: You think about the question and come up with a plan. -You ONLY have access to the following tools, and should NEVER make up tools that are not listed here: ----------------------- - -Query: Thinker Agent Search Task -Answer: - -2025-08-04 02:53:59,069 - alembic.runtime.migration - INFO - Context impl SQLiteImpl. -2025-08-04 02:53:59,069 - alembic.runtime.migration - INFO - Will assume non-transactional DDL. -2025-08-04 02:53:59,074 - embedchain.chunkers.base_chunker - INFO - Skipping chunks smaller than 1 characters -2025-08-04 02:54:00,637 - embedchain.embedchain - INFO - Successfully saved === INPUT === -system: You are Organizer. You're an expert in organizing. -Your personal goal is: You (DataType.TEXT). New chunks count: 0 -2025-08-04 02:54:00,641 - agentgraph.reconstruction.rag_prompt_reconstructor - INFO - Successfully initialized RAG tool with trace content -2025-08-04 02:54:00,642 - agentgraph.reconstruction.rag_prompt_reconstructor - INFO - Initialized RagPromptReconstructor with 8 entities and 7 relations -2025-08-04 02:54:02,288 - embedchain.llm.base - INFO - Prompt: -You are a Q&A expert system. Your responses must always be rooted in the context provided for each query. Here are some guidelines to follow: - -1. Refrain from explicitly mentioning the context provided in your response. -2. The context should silently guide your answers without being directly acknowledged. -3. Do not use phrases such as 'According to the context provided', 'Based on the context, ...' etc. - -Context information: ----------------------- -Tool Description: A tool that can be used to semantic search a query from a specific URL content. -Tool Name: Delegate work to coworker | Tool Description: A tool that can be used to semantic search a query from a specific URL content. | Thought: -assistant: Action: Search in a specific website -Action Input: {"search_query":"haiku about snow with 3 legs","website":"https://www.haiku-studio.com"} -Observation: Relevant Content: ----------------------- - -Query: Generate semantic search queries for the relationship between User Task Query and Organizer Agent -Answer: - -2025-08-04 02:54:06,196 - embedchain.llm.base - INFO - Prompt: -You are a Q&A expert system. Your responses must always be rooted in the context provided for each query. Here are some guidelines to follow: - -1. Refrain from explicitly mentioning the context provided in your response. -2. The context should silently guide your answers without being directly acknowledged. -3. Do not use phrases such as 'According to the context provided', 'Based on the context, ...' etc. - -Context information: ----------------------- -=== INPUT === -system: You are Organizer. You're an expert in organizing. -Your personal goal is: You organize the plan and the tasks. -You ONLY have access to the following tools, and should NEVER make up tools that are not listed here: | Tool Description: Delegate a specific task to one of the following coworkers: Thinker | Tool Description: Ask a specific question to one of the following coworkers: Thinker ----------------------- - -Query: User Task Query Organizer Agent -Answer: - -2025-08-04 02:54:06,651 - embedchain.llm.base - INFO - Prompt: -You are a Q&A expert system. Your responses must always be rooted in the context provided for each query. Here are some guidelines to follow: - -1. Refrain from explicitly mentioning the context provided in your response. -2. The context should silently guide your answers without being directly acknowledged. -3. Do not use phrases such as 'According to the context provided', 'Based on the context, ...' etc. - -Context information: ----------------------- -Action: Search in a specific website -Action Input: {"search_query":"snow with 3 legs","website":"https://www.google.com"} -Observation: Relevant Content: | Thought: -assistant: Action: Search in a specific website -Action Input: {"search_query":"haiku about snow with 3 legs","website":"https://www.haiku-studio.com"} -Observation: Relevant Content: | Tool Description: A tool that can be used to semantic search a query from a specific URL content. ----------------------- - -Query: Organizer Agent Search in a specific website, uses Search in a specific website, interaction between Organizer Agent and Search in a specific website -Answer: - -2025-08-04 02:54:08,142 - agentgraph.reconstruction.rag_prompt_reconstructor - INFO - Completed parallel reconstruction for relation rel_002 -2025-08-04 02:54:08,207 - agentgraph.reconstruction.rag_prompt_reconstructor - INFO - Completed parallel reconstruction for relation rel_001 -2025-08-04 02:54:11,941 - embedchain.llm.base - INFO - Prompt: -You are a Q&A expert system. Your responses must always be rooted in the context provided for each query. Here are some guidelines to follow: - -1. Refrain from explicitly mentioning the context provided in your response. -2. The context should silently guide your answers without being directly acknowledged. -3. Do not use phrases such as 'According to the context provided', 'Based on the context, ...' etc. - -Context information: ----------------------- -=== INPUT === -['capital of the moon'] - -=== INPUT === -system: You are Organizer. You're an expert in organizing. -Your personal goal is: You organize the plan and the tasks. -You ONLY have access to the following tools, and should NEVER make up tools that are not listed here: | assistant: Thought: I need to clarify that the moon does not have a capital, as it does not function as a governed location like a country would. Instead, it is a celestial body. I can incorporate this information into a creative format. -Action: Ask question to coworker | === OUTPUT === -Thought: I need to clarify that the moon does not have a capital, as it does not function as a governed location like a country would. Instead, it is a celestial body. I can incorporate this information into a creative format. -Action: Ask question to coworker ----------------------- - -Query: Organizer Agent Answer Moon Capital Question -Answer: - -2025-08-04 02:54:12,808 - agentgraph.reconstruction.rag_prompt_reconstructor - INFO - Completed parallel reconstruction for relation rel_003 -2025-08-04 02:54:14,862 - embedchain.llm.base - INFO - Prompt: -You are a Q&A expert system. Your responses must always be rooted in the context provided for each query. Here are some guidelines to follow: - -1. Refrain from explicitly mentioning the context provided in your response. -2. The context should silently guide your answers without being directly acknowledged. -3. Do not use phrases such as 'According to the context provided', 'Based on the context, ...' etc. - -Context information: ----------------------- -Action: Delegate work to coworker | Tool Description: Delegate a specific task to one of the following coworkers: Thinker | Tool Description: Delegate a specific task to one of the following coworkers: Writer ----------------------- - -Query: Organizer Agent Delegate work to coworker -Answer: - -2025-08-04 02:54:14,903 - embedchain.llm.base - INFO - Prompt: -You are a Q&A expert system. Your responses must always be rooted in the context provided for each query. Here are some guidelines to follow: - -1. Refrain from explicitly mentioning the context provided in your response. -2. The context should silently guide your answers without being directly acknowledged. -3. Do not use phrases such as 'According to the context provided', 'Based on the context, ...' etc. - -Context information: ----------------------- -Tool Description: Ask a specific question to one of the following coworkers: Thinker | Tool Description: Ask a specific question to one of the following coworkers: Writer | The input to this tool should be the coworker, the task you want them to do, and ALL necessary context to execute the task, they know nothing about the task, so share absolutely everything you know, don't reference things but instead explain them. -Tool Name: Ask question to coworker ----------------------- - -Query: Organizer Agent uses tool to ask question to coworker -Answer: - -2025-08-04 02:54:16,474 - agentgraph.reconstruction.rag_prompt_reconstructor - INFO - Completed parallel reconstruction for relation rel_004 -2025-08-04 02:54:21,782 - embedchain.llm.base - INFO - Prompt: -You are a Q&A expert system. Your responses must always be rooted in the context provided for each query. Here are some guidelines to follow: - -1. Refrain from explicitly mentioning the context provided in your response. -2. The context should silently guide your answers without being directly acknowledged. -3. Do not use phrases such as 'According to the context provided', 'Based on the context, ...' etc. - -Context information: ----------------------- -Tool Description: Ask a specific question to one of the following coworkers: Writer | Tool Description: Ask a specific question to one of the following coworkers: Thinker | Action: Delegate work to coworker ----------------------- - -Query: Organizer Agent Ask question to coworker -Answer: - -2025-08-04 02:54:22,853 - agentgraph.reconstruction.rag_prompt_reconstructor - INFO - Completed parallel reconstruction for relation rel_005 -2025-08-04 02:54:23,146 - embedchain.llm.base - INFO - Prompt: -You are a Q&A expert system. Your responses must always be rooted in the context provided for each query. Here are some guidelines to follow: - -1. Refrain from explicitly mentioning the context provided in your response. -2. The context should silently guide your answers without being directly acknowledged. -3. Do not use phrases such as 'According to the context provided', 'Based on the context, ...' etc. - -Context information: ----------------------- -Tool Description: Ask a specific question to one of the following coworkers: Thinker | Tool Description: Delegate a specific task to one of the following coworkers: Thinker | === INPUT === -system: You are Thinker. You're an expert in thinking. -Your personal goal is: You think about the question and come up with a plan. -You ONLY have access to the following tools, and should NEVER make up tools that are not listed here: ----------------------- - -Query: Thinker Agent Search Task -Answer: - -2025-08-04 02:54:25,393 - agentgraph.reconstruction.rag_prompt_reconstructor - INFO - Completed parallel reconstruction for relation rel_006 -2025-08-04 02:54:26,064 - embedchain.llm.base - INFO - Prompt: -You are a Q&A expert system. Your responses must always be rooted in the context provided for each query. Here are some guidelines to follow: - -1. Refrain from explicitly mentioning the context provided in your response. -2. The context should silently guide your answers without being directly acknowledged. -3. Do not use phrases such as 'According to the context provided', 'Based on the context, ...' etc. - -Context information: ----------------------- -Tool Description: A tool that can be used to semantic search a query from a specific URL content. | Tool Description: A tool that can be used to semantic search a query from a specific URL content. -Tool Name: Delegate work to coworker | Thought: -assistant: Action: Search in a specific website -Action Input: {"search_query":"haiku about snow with 3 legs","website":"https://www.haiku-studio.com"} -Observation: Relevant Content: ----------------------- - -Query: Generate semantic search queries for retrieving trace content related to a specific website and search task -Answer: - -2025-08-04 02:54:31,326 - agentgraph.reconstruction.rag_prompt_reconstructor - INFO - Completed parallel reconstruction for relation rel_007 -2025-08-04 02:54:31,340 - alembic.runtime.migration - INFO - Context impl SQLiteImpl. -2025-08-04 02:54:31,340 - alembic.runtime.migration - INFO - Will assume non-transactional DDL. -2025-08-04 02:54:31,346 - embedchain.chunkers.base_chunker - INFO - Skipping chunks smaller than 1 characters -2025-08-04 02:54:32,979 - embedchain.embedchain - INFO - Successfully saved === INPUT === -system: You are Organizer. You're an expert in organizing. -Your personal goal is: You (DataType.TEXT). New chunks count: 0 -2025-08-04 02:54:32,983 - agentgraph.reconstruction.rag_prompt_reconstructor - INFO - Successfully initialized RAG tool with trace content -2025-08-04 02:54:32,984 - agentgraph.reconstruction.rag_prompt_reconstructor - INFO - Initialized RagPromptReconstructor with 8 entities and 7 relations -2025-08-04 02:54:34,577 - embedchain.llm.base - INFO - Prompt: -You are a Q&A expert system. Your responses must always be rooted in the context provided for each query. Here are some guidelines to follow: - -1. Refrain from explicitly mentioning the context provided in your response. -2. The context should silently guide your answers without being directly acknowledged. -3. Do not use phrases such as 'According to the context provided', 'Based on the context, ...' etc. - -Context information: ----------------------- -Tool Description: A tool that can be used to semantic search a query from a specific URL content. -Tool Name: Delegate work to coworker | Tool Description: A tool that can be used to semantic search a query from a specific URL content. | Thought: -assistant: Action: Search in a specific website -Action Input: {"search_query":"haiku about snow with 3 legs","website":"https://www.haiku-studio.com"} -Observation: Relevant Content: ----------------------- - -Query: Generate semantic search queries related to User Task Query and Organizer Agent -Answer: - -2025-08-04 02:54:34,841 - embedchain.llm.base - INFO - Prompt: -You are a Q&A expert system. Your responses must always be rooted in the context provided for each query. Here are some guidelines to follow: - -1. Refrain from explicitly mentioning the context provided in your response. -2. The context should silently guide your answers without being directly acknowledged. -3. Do not use phrases such as 'According to the context provided', 'Based on the context, ...' etc. - -Context information: ----------------------- -Tool Description: A tool that can be used to semantic search a query from a specific URL content. -Tool Name: Delegate work to coworker | Action: Delegate work to coworker | Tool Description: Delegate a specific task to one of the following coworkers: Thinker ----------------------- - -Query: Generate semantic search queries for the relationship between Organizer Agent and delegating work to coworker -Answer: - -2025-08-04 02:54:38,759 - embedchain.llm.base - INFO - Prompt: -You are a Q&A expert system. Your responses must always be rooted in the context provided for each query. Here are some guidelines to follow: - -1. Refrain from explicitly mentioning the context provided in your response. -2. The context should silently guide your answers without being directly acknowledged. -3. Do not use phrases such as 'According to the context provided', 'Based on the context, ...' etc. - -Context information: ----------------------- -=== INPUT === -system: You are Organizer. You're an expert in organizing. -Your personal goal is: You organize the plan and the tasks. -You ONLY have access to the following tools, and should NEVER make up tools that are not listed here: | Tool Description: Delegate a specific task to one of the following coworkers: Thinker | Tool Description: Ask a specific question to one of the following coworkers: Thinker ----------------------- - -Query: User Task Query Organizer Agent -Answer: - -2025-08-04 02:54:39,685 - embedchain.llm.base - INFO - Prompt: -You are a Q&A expert system. Your responses must always be rooted in the context provided for each query. Here are some guidelines to follow: - -1. Refrain from explicitly mentioning the context provided in your response. -2. The context should silently guide your answers without being directly acknowledged. -3. Do not use phrases such as 'According to the context provided', 'Based on the context, ...' etc. - -Context information: ----------------------- -Thought: -assistant: Action: Search in a specific website -Action Input: {"search_query":"haiku about snow with 3 legs","website":"https://www.haiku-studio.com"} -Observation: Relevant Content: | Tool Description: A tool that can be used to semantic search a query from a specific URL content. -Tool Name: Delegate work to coworker | Thought: -assistant: Action: Search in a specific website -Action Input: {"search_query":"capital of the moon","website":"https://en.wikipedia.org"} -Observation: Relevant Content: ----------------------- - -Query: Organizer Agent Search in a specific website -Answer: - -2025-08-04 02:54:39,894 - embedchain.llm.base - INFO - Prompt: -You are a Q&A expert system. Your responses must always be rooted in the context provided for each query. Here are some guidelines to follow: - -1. Refrain from explicitly mentioning the context provided in your response. -2. The context should silently guide your answers without being directly acknowledged. -3. Do not use phrases such as 'According to the context provided', 'Based on the context, ...' etc. - -Context information: ----------------------- -=== INPUT === -['capital of the moon'] - -=== INPUT === -system: You are Organizer. You're an expert in organizing. -Your personal goal is: You organize the plan and the tasks. -You ONLY have access to the following tools, and should NEVER make up tools that are not listed here: | assistant: Thought: I need to clarify that the moon does not have a capital, as it does not function as a governed location like a country would. Instead, it is a celestial body. I can incorporate this information into a creative format. -Action: Ask question to coworker | === OUTPUT === -Thought: I need to clarify that the moon does not have a capital, as it does not function as a governed location like a country would. Instead, it is a celestial body. I can incorporate this information into a creative format. -Action: Ask question to coworker ----------------------- - -Query: Organizer Agent Answer Moon Capital Question -Answer: - -2025-08-04 02:54:40,701 - agentgraph.reconstruction.rag_prompt_reconstructor - INFO - Completed parallel reconstruction for relation rel_001 -2025-08-04 02:54:41,005 - agentgraph.reconstruction.rag_prompt_reconstructor - INFO - Completed parallel reconstruction for relation rel_002 -2025-08-04 02:54:41,100 - agentgraph.reconstruction.rag_prompt_reconstructor - INFO - Completed parallel reconstruction for relation rel_003 -2025-08-04 02:54:41,727 - embedchain.llm.base - INFO - Prompt: -You are a Q&A expert system. Your responses must always be rooted in the context provided for each query. Here are some guidelines to follow: - -1. Refrain from explicitly mentioning the context provided in your response. -2. The context should silently guide your answers without being directly acknowledged. -3. Do not use phrases such as 'According to the context provided', 'Based on the context, ...' etc. - -Context information: ----------------------- -Action: Delegate work to coworker | Tool Description: Delegate a specific task to one of the following coworkers: Thinker | Tool Description: Delegate a specific task to one of the following coworkers: Writer ----------------------- - -Query: Organizer Agent Delegate work to coworker -Answer: - -2025-08-04 02:54:43,086 - agentgraph.reconstruction.rag_prompt_reconstructor - INFO - Completed parallel reconstruction for relation rel_004 -2025-08-04 02:54:47,864 - embedchain.llm.base - INFO - Prompt: -You are a Q&A expert system. Your responses must always be rooted in the context provided for each query. Here are some guidelines to follow: - -1. Refrain from explicitly mentioning the context provided in your response. -2. The context should silently guide your answers without being directly acknowledged. -3. Do not use phrases such as 'According to the context provided', 'Based on the context, ...' etc. - -Context information: ----------------------- -Tool Description: Ask a specific question to one of the following coworkers: Writer | Tool Description: Ask a specific question to one of the following coworkers: Thinker | Action: Delegate work to coworker ----------------------- - -Query: Organizer Agent Ask question to coworker -Answer: - -2025-08-04 02:54:48,937 - agentgraph.reconstruction.rag_prompt_reconstructor - INFO - Completed parallel reconstruction for relation rel_005 -2025-08-04 02:54:52,989 - embedchain.llm.base - INFO - Prompt: -You are a Q&A expert system. Your responses must always be rooted in the context provided for each query. Here are some guidelines to follow: - -1. Refrain from explicitly mentioning the context provided in your response. -2. The context should silently guide your answers without being directly acknowledged. -3. Do not use phrases such as 'According to the context provided', 'Based on the context, ...' etc. - -Context information: ----------------------- -Tool Description: Ask a specific question to one of the following coworkers: Thinker | Tool Description: Delegate a specific task to one of the following coworkers: Thinker | === INPUT === -system: You are Thinker. You're an expert in thinking. -Your personal goal is: You think about the question and come up with a plan. -You ONLY have access to the following tools, and should NEVER make up tools that are not listed here: ----------------------- - -Query: Thinker Agent Search Task -Answer: - -2025-08-04 02:54:54,781 - agentgraph.reconstruction.rag_prompt_reconstructor - INFO - Completed parallel reconstruction for relation rel_006 -2025-08-04 02:54:57,596 - embedchain.llm.base - INFO - Prompt: -You are a Q&A expert system. Your responses must always be rooted in the context provided for each query. Here are some guidelines to follow: - -1. Refrain from explicitly mentioning the context provided in your response. -2. The context should silently guide your answers without being directly acknowledged. -3. Do not use phrases such as 'According to the context provided', 'Based on the context, ...' etc. - -Context information: ----------------------- -Action: Search in a specific website -Action Input: {"search_query":"snow with 3 legs","website":"https://www.google.com"} -Observation: Relevant Content: | Tool Description: A tool that can be used to semantic search a query from a specific URL content. | Thought: -assistant: Action: Search in a specific website -Action Input: {"search_query":"haiku about snow with 3 legs","website":"https://www.haiku-studio.com"} -Observation: Relevant Content: ----------------------- - -Query: Search in a specific website Search Task, supports Search Task, interaction between Search in a specific website and Search Task -Answer: - -2025-08-04 02:54:59,015 - agentgraph.reconstruction.rag_prompt_reconstructor - INFO - Completed parallel reconstruction for relation rel_007 -2025-08-06 11:27:22,081 - openlit - INFO - Starting openLIT initialization... -2025-08-06 11:27:22,099 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-08-06 11:27:22,882 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-08-06 11:27:22,936 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-08-06 11:27:22,936 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-08-06 11:27:22,936 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-08-06 11:27:23,289 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-08-06 11:27:23,293 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-08-06 11:27:23,293 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-08-06 11:27:23,296 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-08-06 11:27:23,297 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-08-06 11:27:25,032 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-08-06 11:27:25,034 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-08-06 11:27:25,034 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-08-06 11:27:25,035 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-08-06 11:27:25,035 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-08-06 11:27:25,035 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-08-06 11:27:25,035 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-08-06 11:27:25,035 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-08-06 11:27:25,035 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-08-06 11:27:25,035 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-08-06 11:27:25,035 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-08-06 11:27:25,035 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-08-06 11:27:25,035 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-08-06 11:27:25,035 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-08-06 11:27:25,036 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-08-06 11:27:25,036 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-08-06 11:27:25,036 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-08-06 11:27:25,036 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-08-06 11:27:25,036 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-08-06 11:30:08,869 - openlit - INFO - Starting openLIT initialization... -2025-08-06 11:30:08,885 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-08-06 11:30:09,533 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-08-06 11:30:09,582 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-08-06 11:30:09,582 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-08-06 11:30:09,582 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-08-06 11:30:09,850 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-08-06 11:30:09,853 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-08-06 11:30:09,853 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-08-06 11:30:09,855 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-08-06 11:30:09,856 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-08-06 11:30:11,263 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-08-06 11:30:11,266 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-08-06 11:30:11,266 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-08-06 11:30:11,267 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-08-06 11:30:11,267 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-08-06 11:30:11,267 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-08-06 11:30:11,267 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-08-06 11:30:11,267 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-08-06 11:30:11,267 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-08-06 11:30:11,267 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-08-06 11:30:11,267 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-08-06 11:30:11,267 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-08-06 11:30:11,267 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-08-06 11:30:11,268 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-08-06 11:30:11,268 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-08-06 11:30:11,268 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-08-06 11:30:11,268 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-08-06 11:30:11,268 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-08-06 11:30:11,268 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-08-06 11:35:20,722 - openlit - INFO - Starting openLIT initialization... -2025-08-06 11:35:20,740 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-08-06 11:35:21,369 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-08-06 11:35:21,415 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-08-06 11:35:21,415 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-08-06 11:35:21,415 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-08-06 11:35:21,717 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-08-06 11:35:21,719 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-08-06 11:35:21,719 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-08-06 11:35:21,722 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-08-06 11:35:21,722 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-08-06 11:35:23,140 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-08-06 11:35:23,143 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-08-06 11:35:23,143 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-08-06 11:35:23,143 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-08-06 11:35:23,143 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-08-06 11:35:23,143 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-08-06 11:35:23,143 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-08-06 11:35:23,143 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-08-06 11:35:23,144 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-08-06 11:35:23,144 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-08-06 11:35:23,144 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-08-06 11:35:23,144 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-08-06 11:35:23,144 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-08-06 11:35:23,144 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-08-06 11:35:23,144 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-08-06 11:35:23,144 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-08-06 11:35:23,144 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-08-06 11:35:23,144 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-08-06 11:35:23,144 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-08-06 11:39:36,055 - openlit - INFO - Starting openLIT initialization... -2025-08-06 11:39:36,071 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-08-06 11:39:36,841 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-08-06 11:39:36,898 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-08-06 11:39:36,898 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-08-06 11:39:36,898 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-08-06 11:39:37,224 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-08-06 11:39:37,228 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-08-06 11:39:37,228 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-08-06 11:39:37,231 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-08-06 11:39:37,232 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-08-06 11:39:38,910 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-08-06 11:39:38,913 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-08-06 11:39:38,913 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-08-06 11:39:38,913 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-08-06 11:39:38,914 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-08-06 11:39:38,914 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-08-06 11:39:38,914 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-08-06 11:39:38,914 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-08-06 11:39:38,914 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-08-06 11:39:38,914 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-08-06 11:39:38,914 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-08-06 11:39:38,914 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-08-06 11:39:38,914 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-08-06 11:39:38,914 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-08-06 11:39:38,915 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-08-06 11:39:38,915 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-08-06 11:39:38,915 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-08-06 11:39:38,915 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-08-06 11:39:38,915 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-08-06 11:59:42,178 - openlit - INFO - Starting openLIT initialization... -2025-08-06 11:59:42,193 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-08-06 11:59:42,925 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-08-06 11:59:42,978 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-08-06 11:59:42,978 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-08-06 11:59:42,978 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-08-06 11:59:43,274 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-08-06 11:59:43,278 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-08-06 11:59:43,278 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-08-06 11:59:43,281 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-08-06 11:59:43,282 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-08-06 11:59:44,843 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-08-06 11:59:44,845 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-08-06 11:59:44,845 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-08-06 11:59:44,845 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-08-06 11:59:44,846 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-08-06 11:59:44,846 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-08-06 11:59:44,846 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-08-06 11:59:44,846 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-08-06 11:59:44,846 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-08-06 11:59:44,846 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-08-06 11:59:44,846 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-08-06 11:59:44,846 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-08-06 11:59:44,846 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-08-06 11:59:44,846 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-08-06 11:59:44,846 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-08-06 11:59:44,846 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-08-06 11:59:44,847 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-08-06 11:59:44,847 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-08-06 11:59:44,847 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-08-06 12:26:20,759 - openlit - INFO - Starting openLIT initialization... -2025-08-06 12:26:20,775 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-08-06 12:26:21,572 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-08-06 12:26:21,623 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-08-06 12:26:21,623 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-08-06 12:26:21,623 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-08-06 12:26:21,917 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-08-06 12:26:21,921 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-08-06 12:26:21,921 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-08-06 12:26:21,924 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-08-06 12:26:21,925 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-08-06 12:26:23,589 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-08-06 12:26:23,592 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-08-06 12:26:23,592 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-08-06 12:26:23,592 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-08-06 12:26:23,592 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-08-06 12:26:23,592 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-08-06 12:26:23,593 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-08-06 12:26:23,593 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-08-06 12:26:23,593 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-08-06 12:26:23,593 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-08-06 12:26:23,593 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-08-06 12:26:23,593 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-08-06 12:26:23,593 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-08-06 12:26:23,593 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-08-06 12:26:23,593 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-08-06 12:26:23,593 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-08-06 12:26:23,593 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-08-06 12:26:23,593 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-08-06 12:26:23,594 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-08-06 14:52:51,250 - openlit - INFO - Starting openLIT initialization... -2025-08-06 14:52:51,267 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-08-06 14:52:52,209 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-08-06 14:52:52,263 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-08-06 14:52:52,263 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-08-06 14:52:52,263 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-08-06 14:52:52,578 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-08-06 14:52:52,581 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-08-06 14:52:52,581 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-08-06 14:52:52,584 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-08-06 14:52:52,584 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-08-06 14:52:54,206 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-08-06 14:52:54,208 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-08-06 14:52:54,208 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-08-06 14:52:54,209 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-08-06 14:52:54,209 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-08-06 14:52:54,209 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-08-06 14:52:54,209 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-08-06 14:52:54,209 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-08-06 14:52:54,209 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-08-06 14:52:54,209 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-08-06 14:52:54,209 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-08-06 14:52:54,210 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-08-06 14:52:54,210 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-08-06 14:52:54,210 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-08-06 14:52:54,210 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-08-06 14:52:54,210 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-08-06 14:52:54,210 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-08-06 14:52:54,210 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-08-06 14:52:54,210 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-08-11 20:09:15,672 - openlit - INFO - Starting openLIT initialization... -2025-08-11 20:09:15,689 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-08-11 20:09:16,327 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-08-11 20:09:16,379 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-08-11 20:09:16,379 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-08-11 20:09:16,379 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-08-11 20:09:16,650 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-08-11 20:09:16,652 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-08-11 20:09:16,653 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-08-11 20:09:16,655 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-08-11 20:09:16,656 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-08-11 20:09:18,800 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-08-11 20:09:18,800 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-08-11 20:09:18,801 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-08-11 20:09:18,801 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-08-11 20:09:18,801 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-08-11 20:09:18,801 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-08-11 20:09:18,801 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-08-11 20:09:18,801 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-08-11 20:09:18,801 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-08-11 20:09:18,801 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-08-11 20:09:18,801 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-08-11 20:09:18,801 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-08-11 20:09:18,802 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-08-11 20:09:18,802 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-08-11 20:09:18,802 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-08-11 20:09:18,802 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-08-11 20:09:18,802 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-08-11 20:09:18,802 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-08-11 20:09:19,701 - agentgraph.methods.baseline.pydantic_method - INFO - Processing text with pydantic_ai method in parallel_3_stage mode (length: 1144) -2025-08-11 20:09:21,814 - agentgraph.methods.baseline.pydantic_method - ERROR - Error in pydantic_ai knowledge extraction: status_code: 429, model_name: gpt-4o-mini, body: {'message': 'You exceeded your current quota, please check your plan and billing details. For more information on this error, read the docs: https://platform.openai.com/docs/guides/error-codes/api-errors.', 'type': 'insufficient_quota', 'param': None, 'code': 'insufficient_quota'} -2025-08-11 20:09:21,819 - agentgraph.methods.baseline.pydantic_method - ERROR - Traceback: Traceback (most recent call last): - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/pydantic_ai/models/openai.py", line 319, in _completions_create - return await self.client.chat.completions.create( - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/openlit/instrumentation/openai/async_openai.py", line 722, in wrapper - response = await wrapped(*args, **kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/openai/resources/chat/completions/completions.py", line 2454, in create - return await self._post( - ^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/openai/_base_client.py", line 1791, in post - return await self.request(cast_to, opts, stream=stream, stream_cls=stream_cls) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/openai/_base_client.py", line 1591, in request - raise self._make_status_error_from_response(err.response) from None -openai.RateLimitError: Error code: 429 - {'error': {'message': 'You exceeded your current quota, please check your plan and billing details. For more information on this error, read the docs: https://platform.openai.com/docs/guides/error-codes/api-errors.', 'type': 'insufficient_quota', 'param': None, 'code': 'insufficient_quota'}} - -The above exception was the direct cause of the following exception: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/agentgraph/methods/baseline/pydantic_method.py", line 307, in process_text - kg_data, usage = asyncio.run(get_agent_graph(text, self.sequential, self.hybrid)) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/asyncio/runners.py", line 190, in run - return runner.run(main) - ^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/asyncio/runners.py", line 118, in run - return self._loop.run_until_complete(task) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/asyncio/base_events.py", line 654, in run_until_complete - return future.result() - ^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/agentgraph/methods/baseline/pydantic_method.py", line 252, in get_agent_graph - entity_result, relation_result = await gather( - ^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/agentgraph/methods/baseline/pydantic_method.py", line 86, in get_agent_graph_relations - relation_result: AgentRunResult[List[Relation]] = await relation_agent.run(instruction) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/pydantic_ai/agent.py", line 562, in run - async for _ in agent_run: - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/pydantic_ai/agent.py", line 2193, in __anext__ - next_node = await self._graph_run.__anext__() - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/pydantic_graph/graph.py", line 809, in __anext__ - return await self.next(self._next_node) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/pydantic_graph/graph.py", line 782, in next - self._next_node = await node.run(ctx) - ^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/pydantic_ai/_agent_graph.py", line 299, in run - return await self._make_request(ctx) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/pydantic_ai/_agent_graph.py", line 359, in _make_request - model_response = await ctx.deps.model.request(message_history, model_settings, model_request_parameters) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/pydantic_ai/models/openai.py", line 234, in request - response = await self._completions_create( - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/pydantic_ai/models/openai.py", line 348, in _completions_create - raise ModelHTTPError(status_code=status_code, model_name=self.model_name, body=e.body) from e -pydantic_ai.exceptions.ModelHTTPError: status_code: 429, model_name: gpt-4o-mini, body: {'message': 'You exceeded your current quota, please check your plan and billing details. For more information on this error, read the docs: https://platform.openai.com/docs/guides/error-codes/api-errors.', 'type': 'insufficient_quota', 'param': None, 'code': 'insufficient_quota'} - -2025-08-11 20:09:21,819 - agentgraph.methods.baseline.pydantic_method - INFO - Processing text with pydantic_ai method in sequential_3_stage mode (length: 1144) -2025-08-11 20:09:23,481 - agentgraph.methods.baseline.pydantic_method - ERROR - Error in pydantic_ai knowledge extraction: status_code: 429, model_name: gpt-4o-mini, body: {'message': 'You exceeded your current quota, please check your plan and billing details. For more information on this error, read the docs: https://platform.openai.com/docs/guides/error-codes/api-errors.', 'type': 'insufficient_quota', 'param': None, 'code': 'insufficient_quota'} -2025-08-11 20:09:23,482 - agentgraph.methods.baseline.pydantic_method - ERROR - Traceback: Traceback (most recent call last): - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/pydantic_ai/models/openai.py", line 319, in _completions_create - return await self.client.chat.completions.create( - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/openlit/instrumentation/openai/async_openai.py", line 722, in wrapper - response = await wrapped(*args, **kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/openai/resources/chat/completions/completions.py", line 2454, in create - return await self._post( - ^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/openai/_base_client.py", line 1791, in post - return await self.request(cast_to, opts, stream=stream, stream_cls=stream_cls) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/openai/_base_client.py", line 1591, in request - raise self._make_status_error_from_response(err.response) from None -openai.RateLimitError: Error code: 429 - {'error': {'message': 'You exceeded your current quota, please check your plan and billing details. For more information on this error, read the docs: https://platform.openai.com/docs/guides/error-codes/api-errors.', 'type': 'insufficient_quota', 'param': None, 'code': 'insufficient_quota'}} - -The above exception was the direct cause of the following exception: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/agentgraph/methods/baseline/pydantic_method.py", line 307, in process_text - kg_data, usage = asyncio.run(get_agent_graph(text, self.sequential, self.hybrid)) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/asyncio/runners.py", line 190, in run - return runner.run(main) - ^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/asyncio/runners.py", line 118, in run - return self._loop.run_until_complete(task) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/asyncio/base_events.py", line 654, in run_until_complete - return future.result() - ^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/agentgraph/methods/baseline/pydantic_method.py", line 244, in get_agent_graph - entity_result = await get_agent_graph_entities(trace_content) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/agentgraph/methods/baseline/pydantic_method.py", line 61, in get_agent_graph_entities - entity_result: AgentRunResult[List[Entity]] = await entity_agent.run(instruction_template.format(input_data=trace_content)) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/pydantic_ai/agent.py", line 562, in run - async for _ in agent_run: - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/pydantic_ai/agent.py", line 2193, in __anext__ - next_node = await self._graph_run.__anext__() - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/pydantic_graph/graph.py", line 809, in __anext__ - return await self.next(self._next_node) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/pydantic_graph/graph.py", line 782, in next - self._next_node = await node.run(ctx) - ^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/pydantic_ai/_agent_graph.py", line 299, in run - return await self._make_request(ctx) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/pydantic_ai/_agent_graph.py", line 359, in _make_request - model_response = await ctx.deps.model.request(message_history, model_settings, model_request_parameters) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/pydantic_ai/models/openai.py", line 234, in request - response = await self._completions_create( - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/pydantic_ai/models/openai.py", line 348, in _completions_create - raise ModelHTTPError(status_code=status_code, model_name=self.model_name, body=e.body) from e -pydantic_ai.exceptions.ModelHTTPError: status_code: 429, model_name: gpt-4o-mini, body: {'message': 'You exceeded your current quota, please check your plan and billing details. For more information on this error, read the docs: https://platform.openai.com/docs/guides/error-codes/api-errors.', 'type': 'insufficient_quota', 'param': None, 'code': 'insufficient_quota'} - -2025-08-11 20:09:23,482 - agentgraph.methods.baseline.pydantic_method - INFO - Processing text with pydantic_ai method in hybrid_2_stage mode (length: 1144) -2025-08-11 20:09:25,131 - agentgraph.methods.baseline.pydantic_method - ERROR - Error in pydantic_ai knowledge extraction: status_code: 429, model_name: gpt-4o-mini, body: {'message': 'You exceeded your current quota, please check your plan and billing details. For more information on this error, read the docs: https://platform.openai.com/docs/guides/error-codes/api-errors.', 'type': 'insufficient_quota', 'param': None, 'code': 'insufficient_quota'} -2025-08-11 20:09:25,132 - agentgraph.methods.baseline.pydantic_method - ERROR - Traceback: Traceback (most recent call last): - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/pydantic_ai/models/openai.py", line 319, in _completions_create - return await self.client.chat.completions.create( - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/openlit/instrumentation/openai/async_openai.py", line 722, in wrapper - response = await wrapped(*args, **kwargs) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/openai/resources/chat/completions/completions.py", line 2454, in create - return await self._post( - ^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/openai/_base_client.py", line 1791, in post - return await self.request(cast_to, opts, stream=stream, stream_cls=stream_cls) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/openai/_base_client.py", line 1591, in request - raise self._make_status_error_from_response(err.response) from None -openai.RateLimitError: Error code: 429 - {'error': {'message': 'You exceeded your current quota, please check your plan and billing details. For more information on this error, read the docs: https://platform.openai.com/docs/guides/error-codes/api-errors.', 'type': 'insufficient_quota', 'param': None, 'code': 'insufficient_quota'}} - -The above exception was the direct cause of the following exception: - -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/agentgraph/methods/baseline/pydantic_method.py", line 307, in process_text - kg_data, usage = asyncio.run(get_agent_graph(text, self.sequential, self.hybrid)) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/asyncio/runners.py", line 190, in run - return runner.run(main) - ^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/asyncio/runners.py", line 118, in run - return self._loop.run_until_complete(task) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/asyncio/base_events.py", line 654, in run_until_complete - return future.result() - ^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/agentgraph/methods/baseline/pydantic_method.py", line 229, in get_agent_graph - extraction_result = await get_hybrid_extraction(trace_content) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/Desktop/agent_monitoring/agentgraph/methods/baseline/pydantic_method.py", line 140, in get_hybrid_extraction - extraction_result: AgentRunResult[str] = await extraction_agent.run( - ^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/pydantic_ai/agent.py", line 562, in run - async for _ in agent_run: - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/pydantic_ai/agent.py", line 2193, in __anext__ - next_node = await self._graph_run.__anext__() - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/pydantic_graph/graph.py", line 809, in __anext__ - return await self.next(self._next_node) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/pydantic_graph/graph.py", line 782, in next - self._next_node = await node.run(ctx) - ^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/pydantic_ai/_agent_graph.py", line 299, in run - return await self._make_request(ctx) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/pydantic_ai/_agent_graph.py", line 359, in _make_request - model_response = await ctx.deps.model.request(message_history, model_settings, model_request_parameters) - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/pydantic_ai/models/openai.py", line 234, in request - response = await self._completions_create( - ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ - File "/Users/zekunwu/anaconda3/lib/python3.11/site-packages/pydantic_ai/models/openai.py", line 348, in _completions_create - raise ModelHTTPError(status_code=status_code, model_name=self.model_name, body=e.body) from e -pydantic_ai.exceptions.ModelHTTPError: status_code: 429, model_name: gpt-4o-mini, body: {'message': 'You exceeded your current quota, please check your plan and billing details. For more information on this error, read the docs: https://platform.openai.com/docs/guides/error-codes/api-errors.', 'type': 'insufficient_quota', 'param': None, 'code': 'insufficient_quota'} - -2025-08-22 11:32:38,819 - openlit - INFO - Starting openLIT initialization... -2025-08-22 11:32:38,836 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-08-22 11:32:39,651 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-08-22 11:32:39,713 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-08-22 11:32:39,713 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-08-22 11:32:39,713 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-08-22 11:32:40,050 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-08-22 11:32:40,053 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-08-22 11:32:40,053 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-08-22 11:32:40,056 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-08-22 11:32:40,057 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-08-22 11:32:42,703 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-08-22 11:32:42,704 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-08-22 11:32:42,704 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-08-22 11:32:42,705 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-08-22 11:32:42,705 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-08-22 11:32:42,705 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-08-22 11:32:42,705 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-08-22 11:32:42,705 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-08-22 11:32:42,705 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-08-22 11:32:42,705 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-08-22 11:32:42,705 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-08-22 11:32:42,705 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-08-22 11:32:42,706 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-08-22 11:32:42,706 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-08-22 11:32:42,706 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-08-22 11:32:42,706 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-08-22 11:32:42,706 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-08-22 11:32:42,706 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-08-22 11:32:44,258 - backend.routers.example_traces - INFO - Loading examples from /Users/zekunwu/Desktop/agent_monitoring/datasets/example_traces/algorithm-generated.jsonl -2025-08-22 11:32:44,264 - backend.routers.example_traces - INFO - Loading examples from /Users/zekunwu/Desktop/agent_monitoring/datasets/example_traces/hand-crafted.jsonl -2025-08-22 11:32:44,280 - backend.routers.example_traces - INFO - Loaded 184 examples from Who_and_When dataset diff --git a/agentgraph/methods/production/__pycache__/__init__.cpython-311.pyc b/agentgraph/methods/production/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index 236b422409ea5052f0ec3dd291fa1ccd4af446b6..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/production/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/methods/production/__pycache__/__init__.cpython-312.pyc b/agentgraph/methods/production/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index b54d5f6525484d7105401af5578e8226435a52c5..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/production/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/methods/production/__pycache__/langsmith_enhanced_extractor.cpython-312.pyc b/agentgraph/methods/production/__pycache__/langsmith_enhanced_extractor.cpython-312.pyc deleted file mode 100644 index 5d1707e329d00d4fcdca142318777208ba292e05..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/production/__pycache__/langsmith_enhanced_extractor.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/methods/production/__pycache__/multi_agent_knowledge_extractor.cpython-311.pyc b/agentgraph/methods/production/__pycache__/multi_agent_knowledge_extractor.cpython-311.pyc deleted file mode 100644 index 45758d75242ae6c379e7872c69a24dfd51d92f7f..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/production/__pycache__/multi_agent_knowledge_extractor.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/methods/production/__pycache__/multi_agent_knowledge_extractor.cpython-312.pyc b/agentgraph/methods/production/__pycache__/multi_agent_knowledge_extractor.cpython-312.pyc deleted file mode 100644 index 29b2e255bbeead8de212a8c968b08d3821dad33a..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/production/__pycache__/multi_agent_knowledge_extractor.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/methods/production/__pycache__/pydantic_multi_agent_knowledge_extractor.cpython-311.pyc b/agentgraph/methods/production/__pycache__/pydantic_multi_agent_knowledge_extractor.cpython-311.pyc deleted file mode 100644 index f2df98c999ea5cf26fbd29096b5cc50dd7fd6e64..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/production/__pycache__/pydantic_multi_agent_knowledge_extractor.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/methods/production/__pycache__/pydantic_multi_agent_knowledge_extractor.cpython-312.pyc b/agentgraph/methods/production/__pycache__/pydantic_multi_agent_knowledge_extractor.cpython-312.pyc deleted file mode 100644 index 8d4eb5ec5536a4e64209fc21cac347adcb30fbae..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/production/__pycache__/pydantic_multi_agent_knowledge_extractor.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/methods/production/__pycache__/task_prompts.cpython-311.pyc b/agentgraph/methods/production/__pycache__/task_prompts.cpython-311.pyc deleted file mode 100644 index 26654ef7661f1c5d7eefb863bec276ca6600acea..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/production/__pycache__/task_prompts.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/methods/production/__pycache__/task_prompts.cpython-312.pyc b/agentgraph/methods/production/__pycache__/task_prompts.cpython-312.pyc deleted file mode 100644 index 9eb417616d3389fb3145b0e8229c5c985fead408..0000000000000000000000000000000000000000 Binary files a/agentgraph/methods/production/__pycache__/task_prompts.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/reconstruction/__pycache__/__init__.cpython-311.pyc b/agentgraph/reconstruction/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index bcf24cbc7d62ce2d2206611de98aa890817de207..0000000000000000000000000000000000000000 Binary files a/agentgraph/reconstruction/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/reconstruction/__pycache__/__init__.cpython-312.pyc b/agentgraph/reconstruction/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index b76b854ef217122861207b8aaf3ccb519a7b17a0..0000000000000000000000000000000000000000 Binary files a/agentgraph/reconstruction/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/reconstruction/__pycache__/content_reference_resolver.cpython-311.pyc b/agentgraph/reconstruction/__pycache__/content_reference_resolver.cpython-311.pyc deleted file mode 100644 index 4e7ea83accc06506233792e9083f67d3bed8c07a..0000000000000000000000000000000000000000 Binary files a/agentgraph/reconstruction/__pycache__/content_reference_resolver.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/reconstruction/__pycache__/content_reference_resolver.cpython-312.pyc b/agentgraph/reconstruction/__pycache__/content_reference_resolver.cpython-312.pyc deleted file mode 100644 index eef61cbc0dac4b5ace428488942598ccf9cc7051..0000000000000000000000000000000000000000 Binary files a/agentgraph/reconstruction/__pycache__/content_reference_resolver.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/reconstruction/__pycache__/prompt_reconstructor.cpython-311.pyc b/agentgraph/reconstruction/__pycache__/prompt_reconstructor.cpython-311.pyc deleted file mode 100644 index 0873e11dd2dc8400e25e8ffb0a8deb44edf99914..0000000000000000000000000000000000000000 Binary files a/agentgraph/reconstruction/__pycache__/prompt_reconstructor.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/reconstruction/__pycache__/prompt_reconstructor.cpython-312.pyc b/agentgraph/reconstruction/__pycache__/prompt_reconstructor.cpython-312.pyc deleted file mode 100644 index 060749ff1658c518c1a1c1d3f58387bedd0e993d..0000000000000000000000000000000000000000 Binary files a/agentgraph/reconstruction/__pycache__/prompt_reconstructor.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/reconstruction/__pycache__/prompt_reconstructor.cpython-313.pyc b/agentgraph/reconstruction/__pycache__/prompt_reconstructor.cpython-313.pyc deleted file mode 100644 index b13740daa8082a707284a54a00c1baffb7dd6272..0000000000000000000000000000000000000000 Binary files a/agentgraph/reconstruction/__pycache__/prompt_reconstructor.cpython-313.pyc and /dev/null differ diff --git a/agentgraph/reconstruction/__pycache__/rag_prompt_reconstructor.cpython-311.pyc b/agentgraph/reconstruction/__pycache__/rag_prompt_reconstructor.cpython-311.pyc deleted file mode 100644 index d0cd630fd5529f87e68afd595df00f6f0d22a33c..0000000000000000000000000000000000000000 Binary files a/agentgraph/reconstruction/__pycache__/rag_prompt_reconstructor.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/reconstruction/__pycache__/rag_prompt_reconstructor.cpython-312.pyc b/agentgraph/reconstruction/__pycache__/rag_prompt_reconstructor.cpython-312.pyc deleted file mode 100644 index 4f05bbbc8e0fb3a7cf9c69cf944ca52cd03251ff..0000000000000000000000000000000000000000 Binary files a/agentgraph/reconstruction/__pycache__/rag_prompt_reconstructor.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/shared/__pycache__/__init__.cpython-311.pyc b/agentgraph/shared/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index 7a215523db59aa675a290fb680bbacaadb6cfff0..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/shared/__pycache__/__init__.cpython-312.pyc b/agentgraph/shared/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index 5edb2c95da23b3eeb4008782232ccfdab9a4a1a4..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/shared/__pycache__/extraction_factory.cpython-311.pyc b/agentgraph/shared/__pycache__/extraction_factory.cpython-311.pyc deleted file mode 100644 index a24b976d89eac34b22044e07d8e0e7598184f0ae..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/__pycache__/extraction_factory.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/shared/__pycache__/extraction_factory.cpython-312.pyc b/agentgraph/shared/__pycache__/extraction_factory.cpython-312.pyc deleted file mode 100644 index ac7540fda33b5634c6379a2eaad3dec6c952e730..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/__pycache__/extraction_factory.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/shared/__pycache__/flexible_processor.cpython-312.pyc b/agentgraph/shared/__pycache__/flexible_processor.cpython-312.pyc deleted file mode 100644 index 49e0ef15309f771dba069e770da86f42d219bdde..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/__pycache__/flexible_processor.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/shared/__pycache__/method_registry.cpython-311.pyc b/agentgraph/shared/__pycache__/method_registry.cpython-311.pyc deleted file mode 100644 index 5321300a0fdb0e241271810b51fd440ea7a5857d..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/__pycache__/method_registry.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/shared/__pycache__/method_registry.cpython-312.pyc b/agentgraph/shared/__pycache__/method_registry.cpython-312.pyc deleted file mode 100644 index 99b9e4487ce7449f05431770afef92b6391c0e70..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/__pycache__/method_registry.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/shared/__pycache__/schema_manager.cpython-312.pyc b/agentgraph/shared/__pycache__/schema_manager.cpython-312.pyc deleted file mode 100644 index b90f58315a7adbe85b632f5d43710844f408f468..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/__pycache__/schema_manager.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/shared/models/__pycache__/__init__.cpython-311.pyc b/agentgraph/shared/models/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index e29b505cbf0875adb9b5d3503e127408bd04e405..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/models/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/shared/models/__pycache__/__init__.cpython-312.pyc b/agentgraph/shared/models/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index a4b9287e87ec7e97afe4557292d35ce540f0bb9b..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/models/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/shared/models/__pycache__/content_reference.cpython-312.pyc b/agentgraph/shared/models/__pycache__/content_reference.cpython-312.pyc deleted file mode 100644 index 8e99a1f01af1d1d63c7ec811213553feb6a2cd1d..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/models/__pycache__/content_reference.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/shared/models/__pycache__/entity.cpython-312.pyc b/agentgraph/shared/models/__pycache__/entity.cpython-312.pyc deleted file mode 100644 index 45025ea94b1a2ae9f327adc8d244644e513d8415..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/models/__pycache__/entity.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/shared/models/__pycache__/failure.cpython-312.pyc b/agentgraph/shared/models/__pycache__/failure.cpython-312.pyc deleted file mode 100644 index ab0a608e0047d324d7ca424e0e350a75687d2b65..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/models/__pycache__/failure.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/shared/models/__pycache__/relation.cpython-312.pyc b/agentgraph/shared/models/__pycache__/relation.cpython-312.pyc deleted file mode 100644 index ba6f11c5966a59eb421d9de858d8911ffca42115..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/models/__pycache__/relation.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/shared/models/__pycache__/report.cpython-312.pyc b/agentgraph/shared/models/__pycache__/report.cpython-312.pyc deleted file mode 100644 index 771d6cba3cf023e6fb49531eb2de5ba2b1779741..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/models/__pycache__/report.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/shared/models/direct_based/__pycache__/__init__.cpython-312.pyc b/agentgraph/shared/models/direct_based/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index a7d9c34840b3e13df83fb3aa3dacfc0aafbd1878..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/models/direct_based/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/shared/models/direct_based/__pycache__/models.cpython-312.pyc b/agentgraph/shared/models/direct_based/__pycache__/models.cpython-312.pyc deleted file mode 100644 index 6293af942402c4dbe21337ccbab92ebac3febec8..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/models/direct_based/__pycache__/models.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/shared/models/platform_models/__pycache__/langsmith.cpython-311.pyc b/agentgraph/shared/models/platform_models/__pycache__/langsmith.cpython-311.pyc deleted file mode 100644 index 0859be987b96b8949831fbcbe4328a06ff67ee2a..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/models/platform_models/__pycache__/langsmith.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/shared/models/platform_models/__pycache__/langsmith.cpython-312.pyc b/agentgraph/shared/models/platform_models/__pycache__/langsmith.cpython-312.pyc deleted file mode 100644 index 69d3466444b0d945afd72dd4213c229747f501b1..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/models/platform_models/__pycache__/langsmith.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/shared/models/reference_based/__pycache__/__init__.cpython-311.pyc b/agentgraph/shared/models/reference_based/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index 456d97a48deec2fb1318a5c2380d6f6a55a78e89..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/models/reference_based/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/shared/models/reference_based/__pycache__/__init__.cpython-312.pyc b/agentgraph/shared/models/reference_based/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index 797e75606326d9248d80302561f291922d57fa8d..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/models/reference_based/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/shared/models/reference_based/__pycache__/content_reference.cpython-311.pyc b/agentgraph/shared/models/reference_based/__pycache__/content_reference.cpython-311.pyc deleted file mode 100644 index a5b28329d7c1001814e9f8b430f4e56553a1dbc1..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/models/reference_based/__pycache__/content_reference.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/shared/models/reference_based/__pycache__/content_reference.cpython-312.pyc b/agentgraph/shared/models/reference_based/__pycache__/content_reference.cpython-312.pyc deleted file mode 100644 index 5072fa2604024b6f6735f2c4cf046996460197cf..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/models/reference_based/__pycache__/content_reference.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/shared/models/reference_based/__pycache__/entity.cpython-311.pyc b/agentgraph/shared/models/reference_based/__pycache__/entity.cpython-311.pyc deleted file mode 100644 index 38b7ffec9b7d77e0880d9651e090f369372d9e13..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/models/reference_based/__pycache__/entity.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/shared/models/reference_based/__pycache__/entity.cpython-312.pyc b/agentgraph/shared/models/reference_based/__pycache__/entity.cpython-312.pyc deleted file mode 100644 index 736dbfaf3d7fd5193a067ab06750d8d416234560..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/models/reference_based/__pycache__/entity.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/shared/models/reference_based/__pycache__/failure.cpython-311.pyc b/agentgraph/shared/models/reference_based/__pycache__/failure.cpython-311.pyc deleted file mode 100644 index f18df11a4b86e131e8accd201a014c92d55e2a7e..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/models/reference_based/__pycache__/failure.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/shared/models/reference_based/__pycache__/failure.cpython-312.pyc b/agentgraph/shared/models/reference_based/__pycache__/failure.cpython-312.pyc deleted file mode 100644 index 1de9e73a594ebc0d19bff69ea33e0afcc5ee9bda..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/models/reference_based/__pycache__/failure.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/shared/models/reference_based/__pycache__/optimization_recommendation.cpython-311.pyc b/agentgraph/shared/models/reference_based/__pycache__/optimization_recommendation.cpython-311.pyc deleted file mode 100644 index 37f9525c15293ce44eb24c142860411a554cdd0a..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/models/reference_based/__pycache__/optimization_recommendation.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/shared/models/reference_based/__pycache__/optimization_recommendation.cpython-312.pyc b/agentgraph/shared/models/reference_based/__pycache__/optimization_recommendation.cpython-312.pyc deleted file mode 100644 index 7b46ea1d2dcfe2f2766f9d2fd3e0c871b1adc265..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/models/reference_based/__pycache__/optimization_recommendation.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/shared/models/reference_based/__pycache__/relation.cpython-311.pyc b/agentgraph/shared/models/reference_based/__pycache__/relation.cpython-311.pyc deleted file mode 100644 index edd39d8da9308d1f480428fae8966cdfbebc81f7..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/models/reference_based/__pycache__/relation.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/shared/models/reference_based/__pycache__/relation.cpython-312.pyc b/agentgraph/shared/models/reference_based/__pycache__/relation.cpython-312.pyc deleted file mode 100644 index e0ded1491a524051e2dfe4c07d85cd9d9c0ebb66..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/models/reference_based/__pycache__/relation.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/shared/models/reference_based/__pycache__/report.cpython-311.pyc b/agentgraph/shared/models/reference_based/__pycache__/report.cpython-311.pyc deleted file mode 100644 index 8bf5c1110ce43a9d78b9d24d952ccab1575223ca..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/models/reference_based/__pycache__/report.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/shared/models/reference_based/__pycache__/report.cpython-312.pyc b/agentgraph/shared/models/reference_based/__pycache__/report.cpython-312.pyc deleted file mode 100644 index f70fd20d6d02488b56831af580436da27272663d..0000000000000000000000000000000000000000 Binary files a/agentgraph/shared/models/reference_based/__pycache__/report.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/testing/__pycache__/__init__.cpython-311.pyc b/agentgraph/testing/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index 70d4b814df3f9f3f92917698d3330c1784e5c297..0000000000000000000000000000000000000000 Binary files a/agentgraph/testing/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/testing/__pycache__/__init__.cpython-312.pyc b/agentgraph/testing/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index 71a7af6bb0ed99297384fe76874fdcad3d10e67e..0000000000000000000000000000000000000000 Binary files a/agentgraph/testing/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/testing/__pycache__/knowledge_graph_tester.cpython-311.pyc b/agentgraph/testing/__pycache__/knowledge_graph_tester.cpython-311.pyc deleted file mode 100644 index 47be24ff9511e6752166c14168a3b1cf38786024..0000000000000000000000000000000000000000 Binary files a/agentgraph/testing/__pycache__/knowledge_graph_tester.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/testing/__pycache__/knowledge_graph_tester.cpython-312.pyc b/agentgraph/testing/__pycache__/knowledge_graph_tester.cpython-312.pyc deleted file mode 100644 index 2c16fd70b1f9aed14fa7f806db152bec1c26bddb..0000000000000000000000000000000000000000 Binary files a/agentgraph/testing/__pycache__/knowledge_graph_tester.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/testing/perturbation_types/__pycache__/__init__.cpython-311.pyc b/agentgraph/testing/perturbation_types/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index c5312724c63c7d07df34a01b6524142351ff2af7..0000000000000000000000000000000000000000 Binary files a/agentgraph/testing/perturbation_types/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/testing/perturbation_types/__pycache__/__init__.cpython-312.pyc b/agentgraph/testing/perturbation_types/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index d7530244f4f0862a087330e9758b75755a40eee5..0000000000000000000000000000000000000000 Binary files a/agentgraph/testing/perturbation_types/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/testing/perturbation_types/__pycache__/base.cpython-311.pyc b/agentgraph/testing/perturbation_types/__pycache__/base.cpython-311.pyc deleted file mode 100644 index 8e592919f88d9b1419a42e1e8e9e546026719124..0000000000000000000000000000000000000000 Binary files a/agentgraph/testing/perturbation_types/__pycache__/base.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/testing/perturbation_types/__pycache__/base.cpython-312.pyc b/agentgraph/testing/perturbation_types/__pycache__/base.cpython-312.pyc deleted file mode 100644 index e2fd5231550a1a1ecfabe05839772138aa9e70d1..0000000000000000000000000000000000000000 Binary files a/agentgraph/testing/perturbation_types/__pycache__/base.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/testing/perturbation_types/__pycache__/counterfactual_bias.cpython-311.pyc b/agentgraph/testing/perturbation_types/__pycache__/counterfactual_bias.cpython-311.pyc deleted file mode 100644 index 6b4616444984c1c44509757471bc0a090f7aa3a5..0000000000000000000000000000000000000000 Binary files a/agentgraph/testing/perturbation_types/__pycache__/counterfactual_bias.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/testing/perturbation_types/__pycache__/counterfactual_bias.cpython-312.pyc b/agentgraph/testing/perturbation_types/__pycache__/counterfactual_bias.cpython-312.pyc deleted file mode 100644 index 127aecad897e296b4b6e15301c7d37c237df267e..0000000000000000000000000000000000000000 Binary files a/agentgraph/testing/perturbation_types/__pycache__/counterfactual_bias.cpython-312.pyc and /dev/null differ diff --git a/agentgraph/testing/perturbation_types/__pycache__/jailbreak.cpython-311.pyc b/agentgraph/testing/perturbation_types/__pycache__/jailbreak.cpython-311.pyc deleted file mode 100644 index 542535ca66aa5569f78b1e68e622f31a9dcce24e..0000000000000000000000000000000000000000 Binary files a/agentgraph/testing/perturbation_types/__pycache__/jailbreak.cpython-311.pyc and /dev/null differ diff --git a/agentgraph/testing/perturbation_types/__pycache__/jailbreak.cpython-312.pyc b/agentgraph/testing/perturbation_types/__pycache__/jailbreak.cpython-312.pyc deleted file mode 100644 index bed3f6ff8ca89824db16324f7ecf9b94b2cb7ae6..0000000000000000000000000000000000000000 Binary files a/agentgraph/testing/perturbation_types/__pycache__/jailbreak.cpython-312.pyc and /dev/null differ diff --git a/backend/__pycache__/__init__.cpython-311.pyc b/backend/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index ab934a63024abae5eafe15de345d9ceb0d5026e6..0000000000000000000000000000000000000000 Binary files a/backend/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/backend/__pycache__/__init__.cpython-312.pyc b/backend/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index 58a486941efefa802a0697fa104c74106abc81b0..0000000000000000000000000000000000000000 Binary files a/backend/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/backend/__pycache__/__init__.cpython-313.pyc b/backend/__pycache__/__init__.cpython-313.pyc deleted file mode 100644 index fc392d99c75fe5f932d6b6bf40dff69d1ce695ee..0000000000000000000000000000000000000000 Binary files a/backend/__pycache__/__init__.cpython-313.pyc and /dev/null differ diff --git a/backend/__pycache__/app.cpython-311.pyc b/backend/__pycache__/app.cpython-311.pyc deleted file mode 100644 index 65c173395370584820cea7714d7bf388e65a5500..0000000000000000000000000000000000000000 Binary files a/backend/__pycache__/app.cpython-311.pyc and /dev/null differ diff --git a/backend/__pycache__/app.cpython-312.pyc b/backend/__pycache__/app.cpython-312.pyc deleted file mode 100644 index ba29b690e2aff1ed8e1f4c1e16ebc5a511a8e7e3..0000000000000000000000000000000000000000 Binary files a/backend/__pycache__/app.cpython-312.pyc and /dev/null differ diff --git a/backend/__pycache__/config.cpython-311.pyc b/backend/__pycache__/config.cpython-311.pyc deleted file mode 100644 index b07e5bcfaa007796864a7742abf9d7f4a8837808..0000000000000000000000000000000000000000 Binary files a/backend/__pycache__/config.cpython-311.pyc and /dev/null differ diff --git a/backend/__pycache__/config.cpython-312.pyc b/backend/__pycache__/config.cpython-312.pyc deleted file mode 100644 index b73937e10eb19311787da9914349fb3ad8a0e14e..0000000000000000000000000000000000000000 Binary files a/backend/__pycache__/config.cpython-312.pyc and /dev/null differ diff --git a/backend/__pycache__/config.cpython-313.pyc b/backend/__pycache__/config.cpython-313.pyc deleted file mode 100644 index fa5c84716feb3465983c035a2fda6c15ad673e3f..0000000000000000000000000000000000000000 Binary files a/backend/__pycache__/config.cpython-313.pyc and /dev/null differ diff --git a/backend/__pycache__/dependencies.cpython-311.pyc b/backend/__pycache__/dependencies.cpython-311.pyc deleted file mode 100644 index f1e356d106b7ad606b541e4854750366fbebdf66..0000000000000000000000000000000000000000 Binary files a/backend/__pycache__/dependencies.cpython-311.pyc and /dev/null differ diff --git a/backend/__pycache__/dependencies.cpython-312.pyc b/backend/__pycache__/dependencies.cpython-312.pyc deleted file mode 100644 index f0610c0fb5ffb7fec72cb07c17843bfdac5633d0..0000000000000000000000000000000000000000 Binary files a/backend/__pycache__/dependencies.cpython-312.pyc and /dev/null differ diff --git a/backend/__pycache__/main.cpython-311.pyc b/backend/__pycache__/main.cpython-311.pyc deleted file mode 100644 index ec7972add75fa2dee14b36438b00f7409273b0f9..0000000000000000000000000000000000000000 Binary files a/backend/__pycache__/main.cpython-311.pyc and /dev/null differ diff --git a/backend/__pycache__/main.cpython-312.pyc b/backend/__pycache__/main.cpython-312.pyc deleted file mode 100644 index ffb0a7392a15fd5c8f9191b6d093fa488edfe7a5..0000000000000000000000000000000000000000 Binary files a/backend/__pycache__/main.cpython-312.pyc and /dev/null differ diff --git a/backend/__pycache__/models.cpython-311.pyc b/backend/__pycache__/models.cpython-311.pyc deleted file mode 100644 index 9d6efc0025e6ea882831d91890a834e33e313417..0000000000000000000000000000000000000000 Binary files a/backend/__pycache__/models.cpython-311.pyc and /dev/null differ diff --git a/backend/__pycache__/models.cpython-312.pyc b/backend/__pycache__/models.cpython-312.pyc deleted file mode 100644 index aa7aa6927ac5977da5c41f11c9286f6d22943047..0000000000000000000000000000000000000000 Binary files a/backend/__pycache__/models.cpython-312.pyc and /dev/null differ diff --git a/backend/__pycache__/server_config.cpython-311.pyc b/backend/__pycache__/server_config.cpython-311.pyc deleted file mode 100644 index 3d1c98e752786041eb18fb7de0bd3dc2739703ba..0000000000000000000000000000000000000000 Binary files a/backend/__pycache__/server_config.cpython-311.pyc and /dev/null differ diff --git a/backend/__pycache__/server_config.cpython-312.pyc b/backend/__pycache__/server_config.cpython-312.pyc deleted file mode 100644 index 43ff08493e4fa8738da6c0af9352ac9b75bfea8f..0000000000000000000000000000000000000000 Binary files a/backend/__pycache__/server_config.cpython-312.pyc and /dev/null differ diff --git a/backend/__pycache__/server_config.cpython-313.pyc b/backend/__pycache__/server_config.cpython-313.pyc deleted file mode 100644 index 1634b84643f969bfbeef3f69b21e8c50033b9a9d..0000000000000000000000000000000000000000 Binary files a/backend/__pycache__/server_config.cpython-313.pyc and /dev/null differ diff --git a/backend/backend.log b/backend/backend.log deleted file mode 100644 index cd5dc3cd0adadd74c63716ddd60e7f353533a1e0..0000000000000000000000000000000000000000 --- a/backend/backend.log +++ /dev/null @@ -1,4 +0,0 @@ -Traceback (most recent call last): - File "/Users/zekunwu/Desktop/agent_monitoring/backend/app.py", line 22, in - from backend.config import ensure_directories -ModuleNotFoundError: No module named 'backend' diff --git a/backend/database/__pycache__/__init__.cpython-311.pyc b/backend/database/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index 90c4160c907f76866877511c731ca878ef3d27f8..0000000000000000000000000000000000000000 Binary files a/backend/database/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/backend/database/__pycache__/__init__.cpython-312.pyc b/backend/database/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index 386615851584630cdf4c1214ab4d6e9d591f374d..0000000000000000000000000000000000000000 Binary files a/backend/database/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/backend/database/__pycache__/__init__.cpython-313.pyc b/backend/database/__pycache__/__init__.cpython-313.pyc deleted file mode 100644 index 8ee6731ce044ee44388c3a33c6a0f6ba64f71951..0000000000000000000000000000000000000000 Binary files a/backend/database/__pycache__/__init__.cpython-313.pyc and /dev/null differ diff --git a/backend/database/__pycache__/init_db.cpython-311.pyc b/backend/database/__pycache__/init_db.cpython-311.pyc deleted file mode 100644 index 7dc0c1e97676c1e2a073f4eb079117a213b58a13..0000000000000000000000000000000000000000 Binary files a/backend/database/__pycache__/init_db.cpython-311.pyc and /dev/null differ diff --git a/backend/database/__pycache__/init_db.cpython-312.pyc b/backend/database/__pycache__/init_db.cpython-312.pyc deleted file mode 100644 index 4c60fe25fbada1d695f72792cc1cc9e79c02303d..0000000000000000000000000000000000000000 Binary files a/backend/database/__pycache__/init_db.cpython-312.pyc and /dev/null differ diff --git a/backend/database/__pycache__/models.cpython-311.pyc b/backend/database/__pycache__/models.cpython-311.pyc deleted file mode 100644 index 9e7dd5051b221904cd9175637c89c24ef5532cb9..0000000000000000000000000000000000000000 Binary files a/backend/database/__pycache__/models.cpython-311.pyc and /dev/null differ diff --git a/backend/database/__pycache__/models.cpython-312.pyc b/backend/database/__pycache__/models.cpython-312.pyc deleted file mode 100644 index f4396fff4b771a995bc203a4dbbed67d554edf9b..0000000000000000000000000000000000000000 Binary files a/backend/database/__pycache__/models.cpython-312.pyc and /dev/null differ diff --git a/backend/database/__pycache__/utils.cpython-311.pyc b/backend/database/__pycache__/utils.cpython-311.pyc deleted file mode 100644 index a3b596d7377e6971b62464e63b1709fde11bce88..0000000000000000000000000000000000000000 Binary files a/backend/database/__pycache__/utils.cpython-311.pyc and /dev/null differ diff --git a/backend/database/__pycache__/utils.cpython-312.pyc b/backend/database/__pycache__/utils.cpython-312.pyc deleted file mode 100644 index 31b9ae4c798462bb24f5069daa1ca8c942fa2dc4..0000000000000000000000000000000000000000 Binary files a/backend/database/__pycache__/utils.cpython-312.pyc and /dev/null differ diff --git a/backend/logs/agent_monitoring.log b/backend/logs/agent_monitoring.log deleted file mode 100644 index acd37ad969b56b0fa869676aba2486f3daf5c415..0000000000000000000000000000000000000000 --- a/backend/logs/agent_monitoring.log +++ /dev/null @@ -1,398 +0,0 @@ -2025-07-10 11:01:10,074 - __main__ - INFO - Creating tables -2025-07-10 11:01:10,077 - __main__ - INFO - Database initialization completed. Tables created: ['knowledge_graphs', 'sqlite_sequence', 'traces', 'entities', 'relations', 'prompt_reconstructions', 'perturbation_tests', 'causal_analyses', 'observability_connections', 'fetched_traces'] -2025-07-10 11:01:10,077 - __main__ - INFO - Database contains: 109 knowledge graphs, 52 entities, 42 relations, 15 traces -2025-07-10 16:49:25,427 - __main__ - INFO - Creating tables -2025-07-10 16:49:25,439 - __main__ - INFO - Checking for column migrations... -2025-07-10 16:49:25,439 - __main__ - INFO - Database initialization completed. Tables created: ['knowledge_graphs', 'sqlite_sequence', 'traces', 'entities', 'relations', 'prompt_reconstructions', 'perturbation_tests', 'observability_connections', 'fetched_traces', 'causal_analyses'] -2025-07-10 16:49:25,439 - __main__ - INFO - Database contains: 0 knowledge graphs, 0 entities, 0 relations, 0 traces -2025-07-10 21:10:16,166 - __main__ - INFO - Creating tables -2025-07-10 21:10:16,168 - __main__ - INFO - Checking for column migrations... -2025-07-10 21:10:16,168 - __main__ - INFO - Database initialization completed. Tables created: ['knowledge_graphs', 'sqlite_sequence', 'traces', 'entities', 'relations', 'prompt_reconstructions', 'perturbation_tests', 'observability_connections', 'fetched_traces', 'causal_analyses'] -2025-07-10 21:10:16,168 - __main__ - INFO - Database contains: 10 knowledge graphs, 50 entities, 40 relations, 3 traces -2025-07-17 11:20:43,239 - agent_monitoring_server.services.universal_parser - INFO - Running universal parser on trace 009e9a65-55e2-4476-bfa8-c84681b8cd40 -2025-07-17 11:20:43,240 - agent_monitoring_server.services.universal_parser - ERROR - Error creating context document: type object 'ContextDocumentType' has no attribute 'technical' -2025-07-17 11:20:43,240 - agent_monitoring_server.services.universal_parser - ERROR - Error creating context document: type object 'ContextDocumentType' has no attribute 'technical' -2025-07-17 11:20:43,240 - agent_monitoring_server.services.universal_parser - ERROR - Error creating context document: type object 'ContextDocumentType' has no attribute 'technical' -2025-07-17 11:20:43,240 - agent_monitoring_server.services.universal_parser - ERROR - Error creating context document: type object 'ContextDocumentType' has no attribute 'technical' -2025-07-17 11:20:43,240 - agent_monitoring_server.services.universal_parser - INFO - Successfully created 0 context documents for trace 009e9a65-55e2-4476-bfa8-c84681b8cd40 -2025-07-17 11:26:50,608 - openlit - INFO - Starting openLIT initialization... -2025-07-17 11:26:50,614 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-17 11:26:51,249 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-17 11:26:51,280 - openlit - INFO - Library for ollama (ollama) not found. Skipping instrumentation -2025-07-17 11:26:51,280 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-17 11:26:51,280 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-17 11:26:51,280 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-17 11:26:51,652 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-17 11:26:51,763 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-17 11:26:51,763 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-17 11:26:52,229 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-17 11:26:52,230 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-17 11:26:52,230 - openlit - INFO - Library for transformers (transformers) not found. Skipping instrumentation -2025-07-17 11:26:52,234 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-17 11:26:52,234 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-17 11:26:52,234 - openlit - INFO - Library for autogen (autogen) not found. Skipping instrumentation -2025-07-17 11:26:52,234 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-17 11:26:52,234 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-17 11:26:52,234 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-17 11:26:52,234 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-17 11:26:52,234 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-17 11:26:52,235 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-17 11:26:52,235 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-17 11:26:52,235 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-17 11:26:52,235 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-17 11:26:52,235 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-17 11:26:52,235 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-17 11:26:52,235 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-17 11:26:52,235 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-17 11:26:52,235 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-17 11:26:52,235 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-17 11:28:43,892 - openlit - INFO - Starting openLIT initialization... -2025-07-17 11:28:43,898 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-17 11:28:45,430 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-17 11:28:45,453 - openlit - INFO - Library for ollama (ollama) not found. Skipping instrumentation -2025-07-17 11:28:45,453 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-17 11:28:45,453 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-17 11:28:45,454 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-17 11:28:45,800 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-17 11:28:45,903 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-17 11:28:45,903 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-17 11:28:46,473 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-17 11:28:46,475 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-17 11:28:46,475 - openlit - INFO - Library for transformers (transformers) not found. Skipping instrumentation -2025-07-17 11:28:46,477 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-17 11:28:46,478 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-17 11:28:46,478 - openlit - INFO - Library for autogen (autogen) not found. Skipping instrumentation -2025-07-17 11:28:46,478 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-17 11:28:46,478 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-17 11:28:46,478 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-17 11:28:46,478 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-17 11:28:46,478 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-17 11:28:46,478 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-17 11:28:46,478 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-17 11:28:46,479 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-17 11:28:46,479 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-17 11:28:46,479 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-17 11:28:46,479 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-17 11:28:46,479 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-17 11:28:46,479 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-17 11:28:46,479 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-17 11:28:46,479 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-17 11:29:11,127 - openlit - INFO - Starting openLIT initialization... -2025-07-17 11:29:11,133 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-17 11:29:11,505 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-17 11:29:11,527 - openlit - INFO - Library for ollama (ollama) not found. Skipping instrumentation -2025-07-17 11:29:11,527 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-17 11:29:11,527 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-17 11:29:11,527 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-17 11:29:11,868 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-17 11:29:11,969 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-17 11:29:11,969 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-17 11:29:12,413 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-17 11:29:12,414 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-17 11:29:12,414 - openlit - INFO - Library for transformers (transformers) not found. Skipping instrumentation -2025-07-17 11:29:12,417 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-17 11:29:12,417 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-17 11:29:12,418 - openlit - INFO - Library for autogen (autogen) not found. Skipping instrumentation -2025-07-17 11:29:12,418 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-17 11:29:12,418 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-17 11:29:12,418 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-17 11:29:12,418 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-17 11:29:12,418 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-17 11:29:12,418 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-17 11:29:12,418 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-17 11:29:12,418 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-17 11:29:12,418 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-17 11:29:12,419 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-17 11:29:12,419 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-17 11:29:12,419 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-17 11:29:12,419 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-17 11:29:12,419 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-17 11:29:12,420 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-17 11:29:13,449 - agent_monitoring_server.services.universal_parser - INFO - Running universal parser on trace 009e9a65-55e2-4476-bfa8-c84681b8cd40 -2025-07-17 11:29:13,452 - agent_monitoring_server.services.universal_parser - INFO - Context document already exists: System Architecture Overview -2025-07-17 11:29:13,452 - agent_monitoring_server.services.universal_parser - INFO - Context document already exists: Execution Flow Analysis -2025-07-17 11:29:13,452 - agent_monitoring_server.services.universal_parser - INFO - Context document already exists: Performance and Resource Analysis -2025-07-17 11:29:13,452 - agent_monitoring_server.services.universal_parser - INFO - Context document already exists: Technical Framework Analysis -2025-07-17 11:29:13,453 - agent_monitoring_server.services.universal_parser - INFO - Successfully created 0 context documents for trace 009e9a65-55e2-4476-bfa8-c84681b8cd40 -2025-07-17 11:29:13,453 - agent_monitoring_server.services.universal_parser - ERROR - Error running universal parser on trace 009e9a65-55e2-4476-bfa8-c84681b8cd40: name 'os' is not defined -2025-07-17 11:35:24,232 - openlit - INFO - Starting openLIT initialization... -2025-07-17 11:35:24,237 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-17 11:35:24,638 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-17 11:35:24,660 - openlit - INFO - Library for ollama (ollama) not found. Skipping instrumentation -2025-07-17 11:35:24,661 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-17 11:35:24,661 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-17 11:35:24,661 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-17 11:35:25,019 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-17 11:35:25,125 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-17 11:35:25,125 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-17 11:35:25,579 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-17 11:35:25,580 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-17 11:35:25,580 - openlit - INFO - Library for transformers (transformers) not found. Skipping instrumentation -2025-07-17 11:35:25,583 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-17 11:35:25,583 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-17 11:35:25,583 - openlit - INFO - Library for autogen (autogen) not found. Skipping instrumentation -2025-07-17 11:35:25,583 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-17 11:35:25,583 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-17 11:35:25,583 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-17 11:35:25,583 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-17 11:35:25,583 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-17 11:35:25,584 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-17 11:35:25,584 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-17 11:35:25,584 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-17 11:35:25,584 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-17 11:35:25,585 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-17 11:35:25,585 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-17 11:35:25,585 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-17 11:35:25,585 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-17 11:35:25,585 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-17 11:35:25,585 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-17 11:51:38,967 - backend.database.utils - INFO - New trace saved to database (ID: 1, trace_id: da8396d1-3531-44f2-93b7-356059acec80) -2025-07-17 11:51:41,833 - openlit - INFO - Starting openLIT initialization... -2025-07-17 11:51:41,838 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-17 11:51:42,216 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-17 11:51:42,247 - openlit - INFO - Library for ollama (ollama) not found. Skipping instrumentation -2025-07-17 11:51:42,247 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-17 11:51:42,247 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-17 11:51:42,248 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-17 11:51:42,635 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-17 11:51:42,890 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-17 11:51:42,891 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-17 11:51:43,361 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-17 11:51:43,363 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-17 11:51:43,363 - openlit - INFO - Library for transformers (transformers) not found. Skipping instrumentation -2025-07-17 11:51:43,366 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-17 11:51:43,366 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-17 11:51:43,366 - openlit - INFO - Library for autogen (autogen) not found. Skipping instrumentation -2025-07-17 11:51:43,367 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-17 11:51:43,367 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-17 11:51:43,367 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-17 11:51:43,367 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-17 11:51:43,367 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-17 11:51:43,367 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-17 11:51:43,367 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-17 11:51:43,367 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-17 11:51:43,368 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-17 11:51:43,368 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-17 11:51:43,368 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-17 11:51:43,368 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-17 11:51:43,368 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-17 11:51:43,368 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-17 11:51:43,368 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-17 11:51:44,377 - agent_monitoring_server.services.universal_parser - INFO - Running universal parser on trace da8396d1-3531-44f2-93b7-356059acec80 -2025-07-17 11:51:44,383 - agent_monitoring_server.services.universal_parser - INFO - Successfully stored schema analytics metadata for trace da8396d1-3531-44f2-93b7-356059acec80 -2025-07-17 11:51:44,383 - agent_monitoring_server.services.universal_parser - INFO - Stored global schema view metadata for trace da8396d1-3531-44f2-93b7-356059acec80 -2025-07-17 11:51:44,384 - agent_monitoring_server.services.universal_parser - INFO - Created context document: Global Schema Architecture Overview -2025-07-17 11:51:44,385 - agent_monitoring_server.services.universal_parser - INFO - Created context document: Component-to-Entity Mapping Guide -2025-07-17 11:51:44,386 - agent_monitoring_server.services.universal_parser - INFO - Created context document: Relationship Pattern Guide -2025-07-17 11:51:44,386 - agent_monitoring_server.services.universal_parser - INFO - Created context document: Domain Classification Guide -2025-07-17 11:51:44,386 - agent_monitoring_server.services.universal_parser - INFO - Successfully created 4 context documents for trace da8396d1-3531-44f2-93b7-356059acec80 -2025-07-17 13:30:53,750 - openlit - INFO - Starting openLIT initialization... -2025-07-17 13:30:53,756 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-17 13:30:54,231 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-17 13:30:54,270 - openlit - INFO - Library for ollama (ollama) not found. Skipping instrumentation -2025-07-17 13:30:54,270 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-17 13:30:54,271 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-17 13:30:54,271 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-17 13:30:54,683 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-17 13:30:54,838 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-17 13:30:54,838 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-17 13:30:55,374 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-17 13:30:55,375 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-17 13:30:55,375 - openlit - INFO - Library for transformers (transformers) not found. Skipping instrumentation -2025-07-17 13:30:55,378 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-17 13:30:55,378 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-17 13:30:55,379 - openlit - INFO - Library for autogen (autogen) not found. Skipping instrumentation -2025-07-17 13:30:55,379 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-17 13:30:55,379 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-17 13:30:55,379 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-17 13:30:55,379 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-17 13:30:55,379 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-17 13:30:55,379 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-17 13:30:55,380 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-17 13:30:55,380 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-17 13:30:55,380 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-17 13:30:55,380 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-17 13:30:55,380 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-17 13:30:55,380 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-17 13:30:55,380 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-17 13:30:55,381 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-17 13:30:55,381 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-17 13:30:56,667 - agent_monitoring_server.services.universal_parser - INFO - Trace 93680da8-1397-45df-bffd-968500eff13e is not in a parseable format, skipping universal parser -2025-07-17 15:22:55,874 - backend.services.cost_calculation_service - INFO - Successfully fetched LiteLLM pricing data -2025-07-17 15:23:36,362 - backend.services.cost_calculation_service - INFO - Successfully fetched LiteLLM pricing data -2025-07-24 16:42:01,305 - numexpr.utils - INFO - NumExpr defaulting to 12 threads. -2025-07-24 16:42:02,058 - httpx - INFO - HTTP Request: GET https://raw.githubusercontent.com/BerriAI/litellm/main/model_prices_and_context_window.json "HTTP/1.1 200 OK" -2025-07-24 16:42:03,322 - openlit - INFO - Starting openLIT initialization... -2025-07-24 16:42:03,350 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 16:42:03,976 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 16:42:04,027 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 16:42:04,027 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 16:42:04,027 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 16:42:04,459 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 16:42:04,564 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 16:42:04,564 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 16:42:05,005 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 16:42:05,006 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 16:42:06,273 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 16:42:06,276 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 16:42:06,276 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 16:42:06,276 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 16:42:06,276 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 16:42:06,277 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 16:42:06,277 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 16:42:06,277 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 16:42:06,277 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 16:42:06,277 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 16:42:06,277 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 16:42:06,277 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 16:42:06,277 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 16:42:06,277 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 16:42:06,278 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 16:42:06,278 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 16:42:06,278 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 16:42:06,278 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 16:42:06,278 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 16:48:16,667 - numexpr.utils - INFO - NumExpr defaulting to 12 threads. -2025-07-24 16:48:17,424 - httpx - INFO - HTTP Request: GET https://raw.githubusercontent.com/BerriAI/litellm/main/model_prices_and_context_window.json "HTTP/1.1 200 OK" -2025-07-24 16:48:19,005 - openlit - INFO - Starting openLIT initialization... -2025-07-24 16:48:19,023 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 16:48:19,769 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 16:48:19,821 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 16:48:19,822 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 16:48:19,822 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 16:48:20,092 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 16:48:20,193 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 16:48:20,193 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 16:48:20,626 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 16:48:20,627 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 16:48:21,900 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 16:48:21,902 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 16:48:21,902 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 16:48:21,903 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 16:48:21,903 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 16:48:21,903 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 16:48:21,903 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 16:48:21,903 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 16:48:21,903 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 16:48:21,903 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 16:48:21,903 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 16:48:21,903 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 16:48:21,904 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 16:48:21,904 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 16:48:21,904 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 16:48:21,904 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 16:48:21,904 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 16:48:21,904 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 16:48:21,905 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 16:48:22,763 - agent_monitoring_server.services.universal_parser - INFO - Running universal parser on trace 851edfaf-be9c-4ab9-89ae-2270db3f78b0 -2025-07-24 16:48:22,764 - agent_monitoring_server.services.universal_parser - ERROR - Error running universal parser on trace 851edfaf-be9c-4ab9-89ae-2270db3f78b0: Unknown trace format in /var/folders/64/lfz3kc9x5x9cv5dk8bm5j0gm0000gn/T/tmpjt2wuj8h.json -2025-07-24 16:50:00,564 - numexpr.utils - INFO - NumExpr defaulting to 12 threads. -2025-07-24 16:50:01,249 - httpx - INFO - HTTP Request: GET https://raw.githubusercontent.com/BerriAI/litellm/main/model_prices_and_context_window.json "HTTP/1.1 200 OK" -2025-07-24 16:50:02,314 - openlit - INFO - Starting openLIT initialization... -2025-07-24 16:50:02,330 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 16:50:02,881 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 16:50:02,926 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 16:50:02,926 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 16:50:02,926 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 16:50:03,179 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 16:50:03,422 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 16:50:03,422 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 16:50:03,807 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 16:50:03,808 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 16:50:04,743 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 16:50:04,746 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 16:50:04,746 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 16:50:04,746 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 16:50:04,746 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 16:50:04,746 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 16:50:04,746 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 16:50:04,746 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 16:50:04,747 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 16:50:04,747 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 16:50:04,747 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 16:50:04,747 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 16:50:04,747 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 16:50:04,747 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 16:50:04,747 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 16:50:04,747 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 16:50:04,747 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 16:50:04,748 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 16:50:04,748 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 16:50:05,485 - agent_monitoring_server.services.universal_parser - INFO - Running universal parser on trace 851edfaf-be9c-4ab9-89ae-2270db3f78b0 -2025-07-24 16:50:05,485 - agent_monitoring_server.services.universal_parser - ERROR - Error running universal parser on trace 851edfaf-be9c-4ab9-89ae-2270db3f78b0: Unknown trace format in /var/folders/64/lfz3kc9x5x9cv5dk8bm5j0gm0000gn/T/tmpe2r85xk8.json -2025-07-24 16:50:59,939 - numexpr.utils - INFO - NumExpr defaulting to 12 threads. -2025-07-24 16:51:00,584 - httpx - INFO - HTTP Request: GET https://raw.githubusercontent.com/BerriAI/litellm/main/model_prices_and_context_window.json "HTTP/1.1 200 OK" -2025-07-24 16:51:01,619 - openlit - INFO - Starting openLIT initialization... -2025-07-24 16:51:01,638 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 16:51:02,195 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 16:51:02,238 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 16:51:02,238 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 16:51:02,238 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 16:51:02,483 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 16:51:02,573 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 16:51:02,573 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 16:51:02,957 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 16:51:02,958 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 16:51:03,899 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 16:51:03,901 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 16:51:03,901 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 16:51:03,902 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 16:51:03,902 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 16:51:03,902 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 16:51:03,902 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 16:51:03,902 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 16:51:03,902 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 16:51:03,902 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 16:51:03,902 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 16:51:03,902 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 16:51:03,903 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 16:51:03,903 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 16:51:03,903 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 16:51:03,903 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 16:51:03,903 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 16:51:03,903 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 16:51:03,903 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation -2025-07-24 16:51:04,639 - agent_monitoring_server.services.universal_parser - INFO - Running universal parser on trace 851edfaf-be9c-4ab9-89ae-2270db3f78b0 -2025-07-24 16:51:04,643 - agent_monitoring_server.services.universal_parser - INFO - Successfully stored schema analytics metadata for trace 851edfaf-be9c-4ab9-89ae-2270db3f78b0 -2025-07-24 16:51:04,643 - agent_monitoring_server.services.universal_parser - INFO - Stored global schema view metadata for trace 851edfaf-be9c-4ab9-89ae-2270db3f78b0 -2025-07-24 16:51:04,644 - agent_monitoring_server.services.universal_parser - INFO - Created context document: Global Schema Architecture Overview -2025-07-24 16:51:04,646 - agent_monitoring_server.services.universal_parser - INFO - Created context document: Component-to-Entity Mapping Guide -2025-07-24 16:51:04,647 - agent_monitoring_server.services.universal_parser - INFO - Created context document: Relationship Pattern Guide -2025-07-24 16:51:04,649 - agent_monitoring_server.services.universal_parser - INFO - Created context document: Domain Classification Guide -2025-07-24 16:51:04,649 - agent_monitoring_server.services.universal_parser - INFO - Successfully created 4 context documents for trace 851edfaf-be9c-4ab9-89ae-2270db3f78b0 -2025-07-24 17:10:08,628 - numexpr.utils - INFO - NumExpr defaulting to 12 threads. -2025-07-24 17:10:10,102 - openlit - INFO - Starting openLIT initialization... -2025-07-24 17:10:10,119 - opentelemetry.trace - WARNING - Overriding of current TracerProvider is not allowed -2025-07-24 17:10:10,675 - openlit - INFO - Library for vertexai (vertexai) not found. Skipping instrumentation -2025-07-24 17:10:10,816 - openlit - INFO - Library for gpt4all (gpt4all) not found. Skipping instrumentation -2025-07-24 17:10:10,816 - openlit - INFO - Library for elevenlabs (elevenlabs) not found. Skipping instrumentation -2025-07-24 17:10:10,816 - openlit - INFO - Library for vllm (vllm) not found. Skipping instrumentation -2025-07-24 17:10:11,100 - openlit - INFO - Library for azure-ai-inference (azure.ai.inference) not found. Skipping instrumentation -2025-07-24 17:10:11,203 - openlit - INFO - Library for llama_index (llama_index) not found. Skipping instrumentation -2025-07-24 17:10:11,204 - openlit - INFO - Library for haystack (haystack) not found. Skipping instrumentation -2025-07-24 17:10:11,662 - openlit - INFO - Library for pinecone (pinecone) not found. Skipping instrumentation -2025-07-24 17:10:11,663 - openlit - INFO - Library for milvus (pymilvus) not found. Skipping instrumentation -2025-07-24 17:10:12,934 - openlit - ERROR - Failed to instrument transformers: Failed to import transformers.pipelines because of the following error (look up to see its traceback): -module 'torch' has no attribute 'compiler' -2025-07-24 17:10:12,936 - openlit - INFO - Library for ag2 (ag2) not found. Skipping instrumentation -2025-07-24 17:10:12,937 - openlit - INFO - Library for multion (multion) not found. Skipping instrumentation -2025-07-24 17:10:12,937 - opentelemetry.instrumentation.instrumentor - ERROR - DependencyConflict: requested: "ag2 >= 0.3.2" but found: "None" -2025-07-24 17:10:12,937 - openlit - INFO - Library for pyautogen (pyautogen) not found. Skipping instrumentation -2025-07-24 17:10:12,937 - openlit - INFO - Library for dynamiq (dynamiq) not found. Skipping instrumentation -2025-07-24 17:10:12,937 - openlit - INFO - Library for phidata (phi) not found. Skipping instrumentation -2025-07-24 17:10:12,937 - openlit - INFO - Library for reka-api (reka) not found. Skipping instrumentation -2025-07-24 17:10:12,937 - openlit - INFO - Library for premai (premai) not found. Skipping instrumentation -2025-07-24 17:10:12,937 - openlit - INFO - Library for julep (julep) not found. Skipping instrumentation -2025-07-24 17:10:12,938 - openlit - INFO - Library for astra (astrapy) not found. Skipping instrumentation -2025-07-24 17:10:12,938 - openlit - INFO - Library for ai21 (ai21) not found. Skipping instrumentation -2025-07-24 17:10:12,938 - openlit - INFO - Library for controlflow (controlflow) not found. Skipping instrumentation -2025-07-24 17:10:12,938 - openlit - INFO - Library for assemblyai (assemblyai) not found. Skipping instrumentation -2025-07-24 17:10:12,938 - openlit - INFO - Library for crawl4ai (crawl4ai) not found. Skipping instrumentation -2025-07-24 17:10:12,938 - openlit - INFO - Library for firecrawl (firecrawl) not found. Skipping instrumentation -2025-07-24 17:10:12,938 - openlit - INFO - Library for letta (letta) not found. Skipping instrumentation -2025-07-24 17:10:12,939 - openlit - INFO - Library for together (together) not found. Skipping instrumentation -2025-07-24 17:10:12,939 - openlit - INFO - Library for openai-agents (agents) not found. Skipping instrumentation diff --git a/backend/routers/__pycache__/__init__.cpython-311.pyc b/backend/routers/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index 3bb9297f3ab5504d7d470574486db660e3773892..0000000000000000000000000000000000000000 Binary files a/backend/routers/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/backend/routers/__pycache__/__init__.cpython-312.pyc b/backend/routers/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index cee42c04440cd36ab3ae5427a7560840e4883340..0000000000000000000000000000000000000000 Binary files a/backend/routers/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/backend/routers/__pycache__/agentgraph.cpython-311.pyc b/backend/routers/__pycache__/agentgraph.cpython-311.pyc deleted file mode 100644 index 7852ed8e6794d437151c744ad78e1e9667b9b8d2..0000000000000000000000000000000000000000 Binary files a/backend/routers/__pycache__/agentgraph.cpython-311.pyc and /dev/null differ diff --git a/backend/routers/__pycache__/agentgraph.cpython-312.pyc b/backend/routers/__pycache__/agentgraph.cpython-312.pyc deleted file mode 100644 index deef01dad02a4799d31b946f7419bcddc08eb8c2..0000000000000000000000000000000000000000 Binary files a/backend/routers/__pycache__/agentgraph.cpython-312.pyc and /dev/null differ diff --git a/backend/routers/__pycache__/example_traces.cpython-311.pyc b/backend/routers/__pycache__/example_traces.cpython-311.pyc deleted file mode 100644 index 7fa5ea58d0bd0859ccefc5f1ee580ab11d6f2218..0000000000000000000000000000000000000000 Binary files a/backend/routers/__pycache__/example_traces.cpython-311.pyc and /dev/null differ diff --git a/backend/routers/__pycache__/example_traces.cpython-312.pyc b/backend/routers/__pycache__/example_traces.cpython-312.pyc deleted file mode 100644 index f9ecc270e050cdd6c84e3c48764eae90bb5d59fc..0000000000000000000000000000000000000000 Binary files a/backend/routers/__pycache__/example_traces.cpython-312.pyc and /dev/null differ diff --git a/backend/routers/__pycache__/files.cpython-311.pyc b/backend/routers/__pycache__/files.cpython-311.pyc deleted file mode 100644 index 69ca753148884a624bc26310fffa4321d84a28c7..0000000000000000000000000000000000000000 Binary files a/backend/routers/__pycache__/files.cpython-311.pyc and /dev/null differ diff --git a/backend/routers/__pycache__/files.cpython-312.pyc b/backend/routers/__pycache__/files.cpython-312.pyc deleted file mode 100644 index 24f0691e29b5435188bfd312909d591f1eee0a4e..0000000000000000000000000000000000000000 Binary files a/backend/routers/__pycache__/files.cpython-312.pyc and /dev/null differ diff --git a/backend/routers/__pycache__/graph_comparison.cpython-311.pyc b/backend/routers/__pycache__/graph_comparison.cpython-311.pyc deleted file mode 100644 index 090e031db68e212e00976678cde0363f843aa3cf..0000000000000000000000000000000000000000 Binary files a/backend/routers/__pycache__/graph_comparison.cpython-311.pyc and /dev/null differ diff --git a/backend/routers/__pycache__/graph_comparison.cpython-312.pyc b/backend/routers/__pycache__/graph_comparison.cpython-312.pyc deleted file mode 100644 index 1bb872bf1cf9f97c5bde12a8a4765977eb16834d..0000000000000000000000000000000000000000 Binary files a/backend/routers/__pycache__/graph_comparison.cpython-312.pyc and /dev/null differ diff --git a/backend/routers/__pycache__/knowledge_graphs.cpython-311.pyc b/backend/routers/__pycache__/knowledge_graphs.cpython-311.pyc deleted file mode 100644 index edb02ac368f1ec783269ef48ec3bcfc0482f832f..0000000000000000000000000000000000000000 Binary files a/backend/routers/__pycache__/knowledge_graphs.cpython-311.pyc and /dev/null differ diff --git a/backend/routers/__pycache__/knowledge_graphs.cpython-312.pyc b/backend/routers/__pycache__/knowledge_graphs.cpython-312.pyc deleted file mode 100644 index 1affbf871f156d4061f446e436a480e36efdca58..0000000000000000000000000000000000000000 Binary files a/backend/routers/__pycache__/knowledge_graphs.cpython-312.pyc and /dev/null differ diff --git a/backend/routers/__pycache__/langsmith_parser.cpython-312.pyc b/backend/routers/__pycache__/langsmith_parser.cpython-312.pyc deleted file mode 100644 index a28e76b3cf584bee660fe5880663bf5e7c874646..0000000000000000000000000000000000000000 Binary files a/backend/routers/__pycache__/langsmith_parser.cpython-312.pyc and /dev/null differ diff --git a/backend/routers/__pycache__/methods.cpython-311.pyc b/backend/routers/__pycache__/methods.cpython-311.pyc deleted file mode 100644 index 6a197feb7518e884e1209f15884dd940200cce8d..0000000000000000000000000000000000000000 Binary files a/backend/routers/__pycache__/methods.cpython-311.pyc and /dev/null differ diff --git a/backend/routers/__pycache__/methods.cpython-312.pyc b/backend/routers/__pycache__/methods.cpython-312.pyc deleted file mode 100644 index 226abaaba74c00d5c60fb72444ac34d8e511ebb0..0000000000000000000000000000000000000000 Binary files a/backend/routers/__pycache__/methods.cpython-312.pyc and /dev/null differ diff --git a/backend/routers/__pycache__/observability.cpython-311.pyc b/backend/routers/__pycache__/observability.cpython-311.pyc deleted file mode 100644 index bcaca088840bce5ea94d0e284ace2c4b865516c9..0000000000000000000000000000000000000000 Binary files a/backend/routers/__pycache__/observability.cpython-311.pyc and /dev/null differ diff --git a/backend/routers/__pycache__/observability.cpython-312.pyc b/backend/routers/__pycache__/observability.cpython-312.pyc deleted file mode 100644 index 1d6ce41c7b2c16acf464b31909d6c09212450d9a..0000000000000000000000000000000000000000 Binary files a/backend/routers/__pycache__/observability.cpython-312.pyc and /dev/null differ diff --git a/backend/routers/__pycache__/observe_models.cpython-311.pyc b/backend/routers/__pycache__/observe_models.cpython-311.pyc deleted file mode 100644 index 0212ec0daf517c943a6d4dfb205bde79e3aa6d13..0000000000000000000000000000000000000000 Binary files a/backend/routers/__pycache__/observe_models.cpython-311.pyc and /dev/null differ diff --git a/backend/routers/__pycache__/observe_models.cpython-312.pyc b/backend/routers/__pycache__/observe_models.cpython-312.pyc deleted file mode 100644 index c15ff6c5d430dcecda28e796f4affc6da319a8a5..0000000000000000000000000000000000000000 Binary files a/backend/routers/__pycache__/observe_models.cpython-312.pyc and /dev/null differ diff --git a/backend/routers/__pycache__/stage_processor.cpython-311.pyc b/backend/routers/__pycache__/stage_processor.cpython-311.pyc deleted file mode 100644 index 5c116dedf8d164b1925739f6c2abb7c8d0a2265f..0000000000000000000000000000000000000000 Binary files a/backend/routers/__pycache__/stage_processor.cpython-311.pyc and /dev/null differ diff --git a/backend/routers/__pycache__/stage_processor.cpython-312.pyc b/backend/routers/__pycache__/stage_processor.cpython-312.pyc deleted file mode 100644 index 77a226838bbdbf7251b13b3613fd6d21d36660ab..0000000000000000000000000000000000000000 Binary files a/backend/routers/__pycache__/stage_processor.cpython-312.pyc and /dev/null differ diff --git a/backend/routers/__pycache__/tasks.cpython-311.pyc b/backend/routers/__pycache__/tasks.cpython-311.pyc deleted file mode 100644 index 38d5a8ccb56cc12e2051aea84164473cb23cae05..0000000000000000000000000000000000000000 Binary files a/backend/routers/__pycache__/tasks.cpython-311.pyc and /dev/null differ diff --git a/backend/routers/__pycache__/tasks.cpython-312.pyc b/backend/routers/__pycache__/tasks.cpython-312.pyc deleted file mode 100644 index 730cd0d0b88215cf6db6bdaaf6cb3a95d9727fdc..0000000000000000000000000000000000000000 Binary files a/backend/routers/__pycache__/tasks.cpython-312.pyc and /dev/null differ diff --git a/backend/routers/__pycache__/temporal_graphs.cpython-311.pyc b/backend/routers/__pycache__/temporal_graphs.cpython-311.pyc deleted file mode 100644 index abae93ecdc75a5d57e7c7ee57f040e31d704cd62..0000000000000000000000000000000000000000 Binary files a/backend/routers/__pycache__/temporal_graphs.cpython-311.pyc and /dev/null differ diff --git a/backend/routers/__pycache__/temporal_graphs.cpython-312.pyc b/backend/routers/__pycache__/temporal_graphs.cpython-312.pyc deleted file mode 100644 index 6f018f8c53af50e4b2c42ec4dd4866f1298e0e6a..0000000000000000000000000000000000000000 Binary files a/backend/routers/__pycache__/temporal_graphs.cpython-312.pyc and /dev/null differ diff --git a/backend/routers/__pycache__/traces.cpython-311.pyc b/backend/routers/__pycache__/traces.cpython-311.pyc deleted file mode 100644 index 663eb207feef410a53390ba92bd5dd14bf2b7d3c..0000000000000000000000000000000000000000 Binary files a/backend/routers/__pycache__/traces.cpython-311.pyc and /dev/null differ diff --git a/backend/routers/__pycache__/traces.cpython-312.pyc b/backend/routers/__pycache__/traces.cpython-312.pyc deleted file mode 100644 index 0bbcb3b305b689e88b9ffb90536ddb7786171410..0000000000000000000000000000000000000000 Binary files a/backend/routers/__pycache__/traces.cpython-312.pyc and /dev/null differ diff --git a/backend/scripts/__init__.py b/backend/scripts/__init__.py deleted file mode 100644 index d20e178476eb33f532082ce609af410f7e1720c1..0000000000000000000000000000000000000000 --- a/backend/scripts/__init__.py +++ /dev/null @@ -1,4 +0,0 @@ -""" -CLI scripts for the agent monitoring server -This package contains command-line interface scripts. -""" \ No newline at end of file diff --git a/backend/scripts/__pycache__/__init__.cpython-312.pyc b/backend/scripts/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index c74440d0e955b1fff6d200ece60f754ebfbeae32..0000000000000000000000000000000000000000 Binary files a/backend/scripts/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/backend/scripts/__pycache__/fetch_example_dataset.cpython-312.pyc b/backend/scripts/__pycache__/fetch_example_dataset.cpython-312.pyc deleted file mode 100644 index b2a13731e6c0589a88b98bd71d88b698c2a583e4..0000000000000000000000000000000000000000 Binary files a/backend/scripts/__pycache__/fetch_example_dataset.cpython-312.pyc and /dev/null differ diff --git a/backend/scripts/fetch_example_dataset.py b/backend/scripts/fetch_example_dataset.py deleted file mode 100644 index 81ec787e07e83faebc45390bd0438c1bbb314ba6..0000000000000000000000000000000000000000 --- a/backend/scripts/fetch_example_dataset.py +++ /dev/null @@ -1,88 +0,0 @@ -#!/usr/bin/env python -"""Download Who_and_When dataset subsets and store canonicalised traces. -Run with `python -m server.scripts.fetch_example_dataset`. -""" -from __future__ import annotations - -import json -import os -from pathlib import Path -from typing import Literal - -from datasets import load_dataset - -SUBSETS: list[str] = ["Algorithm-Generated", "Hand-Crafted"] -OUTPUT_DIR = Path(__file__).resolve().parent.parent.parent / "datasets" / "example_traces" - - -def _trace_from_history(history: list[dict]) -> str: - """Convert HuggingFace chat history to plain multiline trace.""" - lines: list[str] = [] - for msg in history: - role = msg.get("role", msg.get("name", "assistant")) - content = (msg.get("content") or "").replace("\n", " ").strip() - role_lower = role.lower() if isinstance(role, str) else "assistant" - lines.append(f"{role_lower}: {content}") - return "\n".join(lines) - - -def dump_subset(subset: str, force: bool = False) -> None: - slug = subset.lower().replace(" ", "_") - out_path = OUTPUT_DIR / f"{slug}.jsonl" - if out_path.exists() and not force: - print(f"{out_path} already exists – skip. Use --force to overwrite.") - return - - ds = load_dataset("Kevin355/Who_and_When", subset, split="train") - OUTPUT_DIR.mkdir(parents=True, exist_ok=True) - with out_path.open("w", encoding="utf-8") as f: - for idx, row in enumerate(ds): - trace = json.dumps(row["history"], ensure_ascii=False) - # derive agent name (first assistant message prefix of content before colon) - agents_set: set[str] = set() - for m in row["history"]: - # Prefer explicit name field (used in algorithm-generated subset) - name = m.get("name") - if name: - agents_set.add(str(name)) - continue - - # Fallback: derive from role when name is absent (hand-crafted subset) - role = str(m.get("role", "")).strip() - # Skip human/user roles – we only care about agent/tool names - if role and role.lower() not in {"human", "user"}: - # Preserve original casing for readability - agents_set.add(role) - agent = sorted(agents_set)[0] if agents_set else None - obj = { - "id": idx, - "subset": subset, - "mistake_step": int(row.get("mistake_step", -1)), - "question": row.get("question", ""), - "agent": agent, - "agents": sorted(list(agents_set)), - "trace": trace, - # NEW: Add failure analysis fields - "is_correct": bool(row.get("is_correct", False)), - "question_id": row.get("question_ID", ""), - "ground_truth": row.get("ground_truth", ""), - "mistake_agent": row.get("mistake_agent", ""), - "mistake_reason": row.get("mistake_reason", ""), # This is the key field for failure analysis! - } - f.write(json.dumps(obj, ensure_ascii=False) + "\n") - print(f"Wrote {idx + 1} examples → {out_path.relative_to(Path.cwd())}") - - -def main(): - import argparse - - parser = argparse.ArgumentParser(description="Fetch Who_and_When example traces") - parser.add_argument("--force", action="store_true", help="Overwrite existing files") - args = parser.parse_args() - - for subset in SUBSETS: - dump_subset(subset, force=args.force) - - -if __name__ == "__main__": - main() \ No newline at end of file diff --git a/backend/scripts/fetch_langfuse_logs.py b/backend/scripts/fetch_langfuse_logs.py deleted file mode 100644 index 5fed0772f8ba277b5011f4749cdf06404d74b224..0000000000000000000000000000000000000000 --- a/backend/scripts/fetch_langfuse_logs.py +++ /dev/null @@ -1,263 +0,0 @@ -#!/usr/bin/env python3 -""" -Script to fetch all Langfuse logs including detailed observations for each trace. -""" -import os -import json -import argparse -import time -from datetime import datetime -from langfuse import Langfuse -from utils.config import LANGFUSE_PUBLIC_KEY, LANGFUSE_SECRET_KEY, LANGFUSE_HOST -import base64 - - -# Langfuse credentials -LANGFUSE_PUBLIC_KEY = "pk-lf-fb12410e-58ba-44b1-bc03-01cb26de076e" -LANGFUSE_SECRET_KEY = "sk-lf-a37fb02d-e1f3-4fe6-822b-21336aa946bd" -LANGFUSE_HOST = "https://cloud.langfuse.com" - -# Create base64 encoded auth string -LANGFUSE_AUTH = base64.b64encode(f"{LANGFUSE_PUBLIC_KEY}:{LANGFUSE_SECRET_KEY}".encode()).decode() - -# Set environment variables for OpenTelemetry -os.environ["OTEL_EXPORTER_OTLP_ENDPOINT"] = f"{LANGFUSE_HOST}/api/public/otel" -os.environ["OTEL_EXPORTER_OTLP_HEADERS"] = f"Authorization=Basic {LANGFUSE_AUTH}" - - -def initialize_langfuse(): - """Initialize and return the Langfuse client.""" - return Langfuse( - secret_key=LANGFUSE_SECRET_KEY, - public_key=LANGFUSE_PUBLIC_KEY, - host=LANGFUSE_HOST - ) - -# Helper function to convert objects to JSON-serializable format -def convert_to_serializable(obj): - """Convert an object to a JSON-serializable format.""" - if isinstance(obj, dict): - return {k: convert_to_serializable(v) for k, v in obj.items()} - elif isinstance(obj, list): - return [convert_to_serializable(item) for item in obj] - elif isinstance(obj, (str, int, float, bool, type(None))): - return obj - elif isinstance(obj, datetime): - return obj.isoformat() - elif hasattr(obj, 'model_dump'): - return convert_to_serializable(obj.model_dump()) - elif hasattr(obj, 'to_dict'): - return convert_to_serializable(obj.to_dict()) - elif hasattr(obj, '__dict__'): - return convert_to_serializable(obj.__dict__) - else: - return str(obj) - -def fetch_trace_with_details(trace_id, langfuse, retry_delay=1.0, max_retries=3): - """Fetch detailed information for a specific trace including all observations.""" - for attempt in range(max_retries): - try: - # First, get the full trace details - trace_response = langfuse.fetch_trace(trace_id) - trace = convert_to_serializable(trace_response) - - # Add a delay to avoid hitting rate limits - time.sleep(retry_delay) - - # Then, fetch all observations for this trace - try: - observations_response = langfuse.fetch_observations(trace_id=trace_id, limit=100) - - # Process the observations response - observations = [] - if hasattr(observations_response, 'data'): - observations = convert_to_serializable(observations_response.data) - elif hasattr(observations_response, 'model_dump'): - observations_dict = convert_to_serializable(observations_response.model_dump()) - if isinstance(observations_dict, dict) and 'data' in observations_dict: - observations = observations_dict['data'] - - # Add only the first observation to the trace - if observations and len(observations) > 0: - trace['observations'] = [observations[0]] - else: - trace['observations'] = [] - except Exception as e: - print(f"Warning: Could not fetch observations for trace {trace_id}: {str(e)}") - trace['observations'] = [] - - return trace - except Exception as e: - if "rate limit exceeded" in str(e).lower() and attempt < max_retries - 1: - wait_time = retry_delay * (attempt + 1) # Exponential backoff - print(f"Rate limit hit, waiting {wait_time} seconds before retry {attempt + 1}/{max_retries}") - time.sleep(wait_time) - else: - print(f"Error fetching details for trace {trace_id}: {str(e)}") - return None - -def fetch_all_logs(output_dir="./logs", limit=20, fetch_details=True, retry_delay=1.0, keys=None): - """ - Fetch all traces from Langfuse, including their detailed observations. - - Args: - output_dir (str): Directory to save the logs - limit (int): Maximum number of traces to fetch - fetch_details (bool): Whether to fetch detailed observations for each trace - retry_delay (float): Delay between API calls to avoid rate limiting - keys (list): List of keys to extract and save in the output - - Returns: - tuple: (filepath, traces) - """ - # Create output directory if it doesn't exist - os.makedirs(output_dir, exist_ok=True) - - # Initialize Langfuse client - langfuse = initialize_langfuse() - - # Langfuse API has a maximum limit of 100 per request, so we need to paginate - max_api_limit = 100 - current_limit = min(max_api_limit, limit) - total_fetched = 0 - page = 1 # Start with page 1 (Langfuse uses 1-indexed pages) - all_traces = [] - - # Fetch traces in batches until we reach the limit - print(f"Fetching up to {limit} traces from Langfuse (max 100 per request)...") - - while total_fetched < limit: - print(f"Fetching traces {total_fetched+1}-{min(total_fetched+current_limit, limit)} (page {page})...") - try: - # Use page parameter instead of offset - traces_response = langfuse.fetch_traces(limit=current_limit, page=page) - - # Convert the response to a list of dictionaries - batch_traces = [] - - try: - if hasattr(traces_response, 'data'): - batch_traces = convert_to_serializable(traces_response.data) - elif hasattr(traces_response, 'model_dump'): - traces_dict = convert_to_serializable(traces_response.model_dump()) - if isinstance(traces_dict, dict) and 'data' in traces_dict: - batch_traces = traces_dict['data'] - else: - batch_traces = [traces_dict] - elif hasattr(traces_response, 'to_dict'): - traces_dict = convert_to_serializable(traces_response.to_dict()) - if isinstance(traces_dict, dict) and 'data' in traces_dict: - batch_traces = traces_dict['data'] - else: - batch_traces = [traces_dict] - else: - batch_traces = [{'response': convert_to_serializable(traces_response)}] - except Exception as e: - print(f"Warning: Exception during conversion: {str(e)}") - batch_traces = [{'error': 'Could not convert response'}] - - # Add the traces from this batch to our list - all_traces.extend(batch_traces) - - # If we got fewer traces than requested, we've reached the end - if len(batch_traces) < current_limit: - break - - # Update for the next batch - total_fetched += len(batch_traces) - page += 1 - - # If we have enough traces, stop - if len(all_traces) >= limit: - all_traces = all_traces[:limit] # Trim to exact limit - break - - # Add a delay to avoid rate limits - time.sleep(retry_delay) - - except Exception as e: - print(f"Error fetching traces: {str(e)}") - break - - print(f"Found {len(all_traces)} traces") - - # If requested, fetch detailed information for each trace - if fetch_details and all_traces: - print(f"Fetching detailed information for each trace (including observations)...") - print(f"Using a {retry_delay} second delay between API calls to avoid rate limiting...") - detailed_traces = [] - - for i, trace in enumerate(all_traces): - if 'id' in trace: - print(f"Fetching details for trace {i+1}/{len(all_traces)}: {trace['id']}") - detailed_trace = fetch_trace_with_details(trace['id'], langfuse, retry_delay) - if detailed_trace: - detailed_traces.append(detailed_trace) - else: - # If details fetch failed, keep the original trace and add empty observations - trace['observations'] = [] - detailed_traces.append(trace) - else: - print(f"Warning: Trace at index {i} has no ID, skipping details") - trace['observations'] = [] - detailed_traces.append(trace) - - # Add a delay between fetches to avoid rate limits - if i < len(all_traces) - 1: - time.sleep(retry_delay) - - all_traces = detailed_traces - - # # Filter the traces to only include specified keys if keys are provided - # if keys: - # filtered_traces = [] - # for trace in all_traces: - # filtered_trace = {} - # if 'data' in trace: - # filtered_trace['data'] = {k: v for k, v in trace['data'].items() if k in keys} - # if 'observations' in trace and trace['observations']: - # filtered_trace['observations'] = [] - # for obs in trace['observations']: - # filtered_obs = {k: v for k, v in obs.items() if k in keys} - # if filtered_obs: # Only add if there are matching keys - # filtered_trace['observations'].append(filtered_obs) - # if filtered_trace: # Only add if there are matching keys - # filtered_traces.append(filtered_trace) - # all_traces = filtered_traces - - # Save traces to a JSON file - timestamp = datetime.now().strftime("%Y%m%d_%H%M%S") - filename = f"langfuse_traces_with_details_{timestamp}.json" - - filepath = os.path.join(output_dir, filename) - - with open(filepath, 'w') as f: - json.dump(all_traces, f, indent=2) - - print(f"Saved {len(all_traces)} traces with all details to {filepath}") - - return filepath, all_traces - - -def main(): - parser = argparse.ArgumentParser(description='Fetch all Langfuse logs with detailed observations') - parser.add_argument('--limit', type=int, default=5, help='Maximum number of traces to fetch') - parser.add_argument('--output', default='./logs', help='Output directory') - parser.add_argument('--no-details', action='store_true', help='Skip fetching detailed observations for each trace') - parser.add_argument('--delay', type=float, default=2.0, help='Delay between API calls in seconds (to avoid rate limiting)') - #parser.add_argument('--keys', nargs='+', default=['latency', 'usage', 'model', 'input', 'output', 'model_parameters'], - # help='Specific keys to extract from the logs (e.g., "latency usage model input output model_parameters")') - - args = parser.parse_args() - - filepath, traces = fetch_all_logs( - output_dir=args.output, - limit=args.limit, - fetch_details=not args.no_details, - retry_delay=args.delay, - #keys=args.keys - ) - - -if __name__ == "__main__": - main() \ No newline at end of file diff --git a/backend/scripts/langsmith_exporter.py b/backend/scripts/langsmith_exporter.py deleted file mode 100644 index 9fb5baa2f55aac0c06a385ba3903f32d9ee1825b..0000000000000000000000000000000000000000 --- a/backend/scripts/langsmith_exporter.py +++ /dev/null @@ -1,212 +0,0 @@ -#!/usr/bin/env python3 -""" -LangSmith Trace Exporter -Export all traces from LangSmith account to individual JSON files -""" - -import argparse -import json -import sys -from datetime import datetime -from pathlib import Path -from typing import Optional - -try: - from langsmith import Client -except ImportError: - print("Error: langsmith package not found. Install with: pip install langsmith") - sys.exit(1) - - -def setup_export_directory() -> Path: - """Create export directory with timestamp""" - timestamp = datetime.now().strftime("%Y%m%d_%H%M%S") - export_dir = Path(f"langsmith_export_{timestamp}") - export_dir.mkdir(exist_ok=True) - return export_dir - - -def sanitize_filename(name: str) -> str: - """Sanitize string for use as filename""" - # Replace invalid characters with underscore - invalid_chars = '<>:"/\\|?*' - for char in invalid_chars: - name = name.replace(char, '_') - return name[:50] # Limit length - - -def export_project_traces(client: Client, project_name: str, export_dir: Path) -> int: - """Export all runs from a specific project - each run with all its traces""" - print(f"Exporting project: {project_name}") - - project_dir = export_dir / sanitize_filename(project_name) - project_dir.mkdir(exist_ok=True) - - run_count = 0 - try: - # Get all runs (top-level runs only) - force no limit to get everything - traces = client.list_runs( - project_name=project_name, - is_root=True, - limit=None # Explicitly no limit to get all runs - ) - - for trace in traces: - run_count += 1 - - # Create filename with run info - trace_name = getattr(trace, 'name', 'unnamed') - trace_id = str(trace.id) - filename = f"{sanitize_filename(trace_name)}_{trace_id[:8]}.json" - - # Get all traces for this run (including nested children) - all_runs = [] - try: - # Get the root run and all its children - trace_runs = client.list_runs(project_name=project_name, trace_id=trace_id) - trace_list = list(trace_runs) - - # Sort traces by start_time descending (latest first) - sorted_traces = sorted(trace_list, key=lambda t: getattr(t, 'start_time', None) or datetime.min) - - for trace_run in sorted_traces: - trace_data = trace_run.dict() if hasattr(trace_run, 'dict') else dict(trace_run) - all_runs.append(trace_data) - except Exception as e: - print(f" Warning: Could not get child traces for run {trace_id}: {e}") - # Fallback to just the main run - trace_data = trace.dict() if hasattr(trace, 'dict') else dict(trace) - all_runs = [trace_data] - - # Save run with all its traces as JSON list - run_export = { - "trace_id": trace_id, - "trace_name": trace_name, - "project_name": project_name, - "export_time": datetime.now().isoformat(), - "total_runs": len(all_runs), - "runs": all_runs - } - - file_path = project_dir / filename - with open(file_path, 'w', encoding='utf-8') as f: - json.dump(run_export, f, ensure_ascii=False, indent=2, default=str) - - if run_count % 10 == 0: - print(f" Exported {run_count} runs...") - - except Exception as e: - print(f"Error exporting project {project_name}: {e}") - return run_count - - print(f" Total runs exported: {run_count}") - return run_count - - -def export_all_traces(api_key: str, base_url: Optional[str] = None, project_name: Optional[str] = None) -> None: - """Export traces from LangSmith account""" - - # Initialize LangSmith client - try: - if base_url: - client = Client(api_key=api_key, api_url=base_url) - else: - client = Client(api_key=api_key) - print("Connected to LangSmith successfully") - except Exception as e: - print(f"Error connecting to LangSmith: {e}") - sys.exit(1) - - # Setup export directory - export_dir = setup_export_directory() - print(f"Export directory: {export_dir}") - - # Get projects to export - total_runs = 0 - if project_name: - # Export specific project - try: - print(f"Exporting specific project: {project_name}") - run_count = export_project_traces(client, project_name, export_dir) - total_runs += run_count - projects = [type('Project', (), {'name': project_name, 'id': 'unknown'})()] - except Exception as e: - print(f"Error exporting project {project_name}: {e}") - sys.exit(1) - else: - try: - projects = list(client.list_projects()) - print(f"Found {len(projects)} projects") - except Exception as e: - print(f"Error listing projects: {e}") - sys.exit(1) - - for project in projects: - proj_name = project.name - if not proj_name: - print(f"Skipping project with no name (ID: {project.id})") - continue - run_count = export_project_traces(client, proj_name, export_dir) - total_runs += run_count - - # Create summary file - summary = { - "export_timestamp": datetime.now().isoformat(), - "total_projects": len(projects), - "total_runs_exported": total_runs, - "projects": [{"name": p.name, "id": str(p.id)} for p in projects] - } - - summary_path = export_dir / "export_summary.json" - with open(summary_path, 'w', encoding='utf-8') as f: - json.dump(summary, f, ensure_ascii=False, indent=2) - - print("\n✅ Export completed!") - print(f"Total projects: {len(projects)}") - print(f"Total runs exported: {total_runs}") - print(f"Export directory: {export_dir}") - - -def main(): - parser = argparse.ArgumentParser( - description="Export all traces from LangSmith account to JSON files" - ) - - parser.add_argument( - "--token", - required=True, - help="LangSmith API token" - ) - - parser.add_argument( - "--base-url", - help="LangSmith base URL (optional, defaults to cloud)" - ) - - parser.add_argument( - "--project", - help="Specific project name to export (optional, exports all projects if not specified)" - ) - - args = parser.parse_args() - - # Validate token - if not args.token or len(args.token.strip()) == 0: - print("Error: Valid API token is required") - sys.exit(1) - - print("🚀 Starting LangSmith trace export...") - print(f"Token: {args.token[:10]}..." if len(args.token) > 10 else f"Token: {args.token}") - - try: - export_all_traces(args.token.strip(), args.base_url, args.project) - except KeyboardInterrupt: - print("\n❌ Export cancelled by user") - sys.exit(1) - except Exception as e: - print(f"❌ Export failed: {e}") - sys.exit(1) - - -if __name__ == "__main__": - main() diff --git a/backend/services/__pycache__/__init__.cpython-311.pyc b/backend/services/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index 318d75e2200cb0e5a9f5d14cd0d5ba750487aa7e..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/backend/services/__pycache__/__init__.cpython-312.pyc b/backend/services/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index 4c39f594be0cf7f9b7eb276cd71bec6bcba0a2bc..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/backend/services/__pycache__/__init__.cpython-313.pyc b/backend/services/__pycache__/__init__.cpython-313.pyc deleted file mode 100644 index 50d70822ad32e2bcb952c5eb40fef648e3a85c97..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/__init__.cpython-313.pyc and /dev/null differ diff --git a/backend/services/__pycache__/base_service.cpython-311.pyc b/backend/services/__pycache__/base_service.cpython-311.pyc deleted file mode 100644 index 7a207f099a9716729e84c8f66ecd333bece0c939..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/base_service.cpython-311.pyc and /dev/null differ diff --git a/backend/services/__pycache__/base_service.cpython-312.pyc b/backend/services/__pycache__/base_service.cpython-312.pyc deleted file mode 100644 index 6289443cdfd244c5b368b98a405e896eacad1f39..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/base_service.cpython-312.pyc and /dev/null differ diff --git a/backend/services/__pycache__/causal_service.cpython-311.pyc b/backend/services/__pycache__/causal_service.cpython-311.pyc deleted file mode 100644 index 92068e0b17043769e5ae021bcc594462980e798e..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/causal_service.cpython-311.pyc and /dev/null differ diff --git a/backend/services/__pycache__/causal_service.cpython-312.pyc b/backend/services/__pycache__/causal_service.cpython-312.pyc deleted file mode 100644 index 4b50859205f3e5a58857e17ef35eaa39384452bb..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/causal_service.cpython-312.pyc and /dev/null differ diff --git a/backend/services/__pycache__/context_service.cpython-311.pyc b/backend/services/__pycache__/context_service.cpython-311.pyc deleted file mode 100644 index 7132adfd3737f756e3932bf75d2831ac1d33204e..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/context_service.cpython-311.pyc and /dev/null differ diff --git a/backend/services/__pycache__/context_service.cpython-312.pyc b/backend/services/__pycache__/context_service.cpython-312.pyc deleted file mode 100644 index 3ff5fa0055aa7930e23979d4a340d263cd891eae..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/context_service.cpython-312.pyc and /dev/null differ diff --git a/backend/services/__pycache__/cost_calculation_service.cpython-311.pyc b/backend/services/__pycache__/cost_calculation_service.cpython-311.pyc deleted file mode 100644 index 92b5edd21d9c2ea17b49144471a8719352cfcc26..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/cost_calculation_service.cpython-311.pyc and /dev/null differ diff --git a/backend/services/__pycache__/cost_calculation_service.cpython-312.pyc b/backend/services/__pycache__/cost_calculation_service.cpython-312.pyc deleted file mode 100644 index 5c2c506f9e8b40bd6dce327798e0a562dfe5bfe1..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/cost_calculation_service.cpython-312.pyc and /dev/null differ diff --git a/backend/services/__pycache__/file_service.cpython-311.pyc b/backend/services/__pycache__/file_service.cpython-311.pyc deleted file mode 100644 index 4f09c1661913a9a78b120ceb37f15b0832ccd959..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/file_service.cpython-311.pyc and /dev/null differ diff --git a/backend/services/__pycache__/file_service.cpython-312.pyc b/backend/services/__pycache__/file_service.cpython-312.pyc deleted file mode 100644 index eed0da9213338173b76705b10bab529b0e8b3851..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/file_service.cpython-312.pyc and /dev/null differ diff --git a/backend/services/__pycache__/knowledge_graph_service.cpython-311.pyc b/backend/services/__pycache__/knowledge_graph_service.cpython-311.pyc deleted file mode 100644 index 550a243e1b85b37448acbcd016f8a865e97ea17b..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/knowledge_graph_service.cpython-311.pyc and /dev/null differ diff --git a/backend/services/__pycache__/knowledge_graph_service.cpython-312.pyc b/backend/services/__pycache__/knowledge_graph_service.cpython-312.pyc deleted file mode 100644 index 9ea24338a5033e8c27db6e1dcd44f9ad2c9eafbd..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/knowledge_graph_service.cpython-312.pyc and /dev/null differ diff --git a/backend/services/__pycache__/knowledge_graph_service.cpython-313.pyc b/backend/services/__pycache__/knowledge_graph_service.cpython-313.pyc deleted file mode 100644 index 039a8eb5a50d2f3c6fb25791d4fe11dcdaa095f8..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/knowledge_graph_service.cpython-313.pyc and /dev/null differ diff --git a/backend/services/__pycache__/method_service.cpython-311.pyc b/backend/services/__pycache__/method_service.cpython-311.pyc deleted file mode 100644 index 5c266505abea5dd1e606af6b8669e1d61f2413de..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/method_service.cpython-311.pyc and /dev/null differ diff --git a/backend/services/__pycache__/method_service.cpython-312.pyc b/backend/services/__pycache__/method_service.cpython-312.pyc deleted file mode 100644 index 1adff05f59b26c88d549c685bbd0c8b9edc7f249..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/method_service.cpython-312.pyc and /dev/null differ diff --git a/backend/services/__pycache__/processing_service.cpython-311.pyc b/backend/services/__pycache__/processing_service.cpython-311.pyc deleted file mode 100644 index e80b50ca92a1e8c2b3dabd0abdafddfa4cd234ca..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/processing_service.cpython-311.pyc and /dev/null differ diff --git a/backend/services/__pycache__/processing_service.cpython-312.pyc b/backend/services/__pycache__/processing_service.cpython-312.pyc deleted file mode 100644 index e1d5074645bfc34219c93885e0255a184fd3d2a9..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/processing_service.cpython-312.pyc and /dev/null differ diff --git a/backend/services/__pycache__/reconstruction_service.cpython-311.pyc b/backend/services/__pycache__/reconstruction_service.cpython-311.pyc deleted file mode 100644 index 318042e934360b20bbd23224dcf3d71bf8baf21c..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/reconstruction_service.cpython-311.pyc and /dev/null differ diff --git a/backend/services/__pycache__/reconstruction_service.cpython-312.pyc b/backend/services/__pycache__/reconstruction_service.cpython-312.pyc deleted file mode 100644 index 78b5da509abe20e83b18529cc8d50fc727c78368..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/reconstruction_service.cpython-312.pyc and /dev/null differ diff --git a/backend/services/__pycache__/scheduler_service.cpython-312.pyc b/backend/services/__pycache__/scheduler_service.cpython-312.pyc deleted file mode 100644 index e25b33a9701b05c3df8b7d31d17f85d521180f9e..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/scheduler_service.cpython-312.pyc and /dev/null differ diff --git a/backend/services/__pycache__/task_queue.cpython-311.pyc b/backend/services/__pycache__/task_queue.cpython-311.pyc deleted file mode 100644 index 58449f69f01f652750314344fbc295d1ed3b2b0d..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/task_queue.cpython-311.pyc and /dev/null differ diff --git a/backend/services/__pycache__/task_queue.cpython-312.pyc b/backend/services/__pycache__/task_queue.cpython-312.pyc deleted file mode 100644 index 4584cc3fe149cbae138a8bd3d77383cdb58c847e..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/task_queue.cpython-312.pyc and /dev/null differ diff --git a/backend/services/__pycache__/task_service.cpython-311.pyc b/backend/services/__pycache__/task_service.cpython-311.pyc deleted file mode 100644 index 05acc71fb1618826b7b33f07d46786f0b001222b..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/task_service.cpython-311.pyc and /dev/null differ diff --git a/backend/services/__pycache__/task_service.cpython-312.pyc b/backend/services/__pycache__/task_service.cpython-312.pyc deleted file mode 100644 index 21a44f3bbaf96f87d1950e86e3931eee779c8fd5..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/task_service.cpython-312.pyc and /dev/null differ diff --git a/backend/services/__pycache__/task_store_service.cpython-311.pyc b/backend/services/__pycache__/task_store_service.cpython-311.pyc deleted file mode 100644 index 81ac0f572e1542337f173863230b06e47ee3b5dc..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/task_store_service.cpython-311.pyc and /dev/null differ diff --git a/backend/services/__pycache__/task_store_service.cpython-312.pyc b/backend/services/__pycache__/task_store_service.cpython-312.pyc deleted file mode 100644 index ad0e09dce6348b644c6e20a78e918ba6a7f0d3b2..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/task_store_service.cpython-312.pyc and /dev/null differ diff --git a/backend/services/__pycache__/test_service.cpython-311.pyc b/backend/services/__pycache__/test_service.cpython-311.pyc deleted file mode 100644 index 6424283f1f4d5d49d87fdb783eef358927c7d309..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/test_service.cpython-311.pyc and /dev/null differ diff --git a/backend/services/__pycache__/test_service.cpython-312.pyc b/backend/services/__pycache__/test_service.cpython-312.pyc deleted file mode 100644 index b0a91723347594151cf08f2a6ab14f40a11e701a..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/test_service.cpython-312.pyc and /dev/null differ diff --git a/backend/services/__pycache__/testing_service.cpython-311.pyc b/backend/services/__pycache__/testing_service.cpython-311.pyc deleted file mode 100644 index 95b3cb486022b00f65b56588929de9d28a03e723..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/testing_service.cpython-311.pyc and /dev/null differ diff --git a/backend/services/__pycache__/testing_service.cpython-312.pyc b/backend/services/__pycache__/testing_service.cpython-312.pyc deleted file mode 100644 index d3abf1db4c21f96802022b8514af4a24ce6720c3..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/testing_service.cpython-312.pyc and /dev/null differ diff --git a/backend/services/__pycache__/trace_management_service.cpython-311.pyc b/backend/services/__pycache__/trace_management_service.cpython-311.pyc deleted file mode 100644 index 47207a8be0896f27502afad9cf325784261b8e46..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/trace_management_service.cpython-311.pyc and /dev/null differ diff --git a/backend/services/__pycache__/trace_management_service.cpython-312.pyc b/backend/services/__pycache__/trace_management_service.cpython-312.pyc deleted file mode 100644 index 7c6a0ee2e33f04217f758197c8b8f1c94eaa91bc..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/trace_management_service.cpython-312.pyc and /dev/null differ diff --git a/backend/services/__pycache__/universal_parser_service.cpython-311.pyc b/backend/services/__pycache__/universal_parser_service.cpython-311.pyc deleted file mode 100644 index f096074f661462bebdfb975a06f547f3a5111955..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/universal_parser_service.cpython-311.pyc and /dev/null differ diff --git a/backend/services/__pycache__/universal_parser_service.cpython-312.pyc b/backend/services/__pycache__/universal_parser_service.cpython-312.pyc deleted file mode 100644 index 5fb57db7b2fa5e8cd149191ecaedab167e218e66..0000000000000000000000000000000000000000 Binary files a/backend/services/__pycache__/universal_parser_service.cpython-312.pyc and /dev/null differ diff --git a/backend/services/platform/__pycache__/__init__.cpython-311.pyc b/backend/services/platform/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index aa80d180cdc412a53a5a88aa51930ad99a51e4ec..0000000000000000000000000000000000000000 Binary files a/backend/services/platform/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/backend/services/platform/__pycache__/__init__.cpython-312.pyc b/backend/services/platform/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index 6d796c739b2abadfe7d1308fce4138283c83d395..0000000000000000000000000000000000000000 Binary files a/backend/services/platform/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/backend/services/platform/__pycache__/connection_service.cpython-311.pyc b/backend/services/platform/__pycache__/connection_service.cpython-311.pyc deleted file mode 100644 index f67954fa01a63da81c85389dd2ba9c5c352c1133..0000000000000000000000000000000000000000 Binary files a/backend/services/platform/__pycache__/connection_service.cpython-311.pyc and /dev/null differ diff --git a/backend/services/platform/__pycache__/connection_service.cpython-312.pyc b/backend/services/platform/__pycache__/connection_service.cpython-312.pyc deleted file mode 100644 index 5ad9e5125467adee58085b8d782a6f841af84134..0000000000000000000000000000000000000000 Binary files a/backend/services/platform/__pycache__/connection_service.cpython-312.pyc and /dev/null differ diff --git a/backend/services/platform/__pycache__/import_service.cpython-311.pyc b/backend/services/platform/__pycache__/import_service.cpython-311.pyc deleted file mode 100644 index f979e3e146a0ed3c3be3f037403ae4af99f86de5..0000000000000000000000000000000000000000 Binary files a/backend/services/platform/__pycache__/import_service.cpython-311.pyc and /dev/null differ diff --git a/backend/services/platform/__pycache__/import_service.cpython-312.pyc b/backend/services/platform/__pycache__/import_service.cpython-312.pyc deleted file mode 100644 index a1832a61f08521f36004bb5735ac9442caa4192a..0000000000000000000000000000000000000000 Binary files a/backend/services/platform/__pycache__/import_service.cpython-312.pyc and /dev/null differ diff --git a/backend/services/platform/__pycache__/langfuse_downloader.cpython-311.pyc b/backend/services/platform/__pycache__/langfuse_downloader.cpython-311.pyc deleted file mode 100644 index 01fd4f872eb6a33e6b7b5bf3432d23f39d93dfe6..0000000000000000000000000000000000000000 Binary files a/backend/services/platform/__pycache__/langfuse_downloader.cpython-311.pyc and /dev/null differ diff --git a/backend/services/platform/__pycache__/langfuse_downloader.cpython-312.pyc b/backend/services/platform/__pycache__/langfuse_downloader.cpython-312.pyc deleted file mode 100644 index 233890035d77b24fb316cb286e142267cb18c1eb..0000000000000000000000000000000000000000 Binary files a/backend/services/platform/__pycache__/langfuse_downloader.cpython-312.pyc and /dev/null differ diff --git a/backend/services/platform/__pycache__/trace_service.cpython-311.pyc b/backend/services/platform/__pycache__/trace_service.cpython-311.pyc deleted file mode 100644 index c7d3f490617c3064b8890a150a878462d8b74295..0000000000000000000000000000000000000000 Binary files a/backend/services/platform/__pycache__/trace_service.cpython-311.pyc and /dev/null differ diff --git a/backend/services/platform/__pycache__/trace_service.cpython-312.pyc b/backend/services/platform/__pycache__/trace_service.cpython-312.pyc deleted file mode 100644 index 502965e0cdf04969af2edc224d274d8df78096f9..0000000000000000000000000000000000000000 Binary files a/backend/services/platform/__pycache__/trace_service.cpython-312.pyc and /dev/null differ diff --git a/dependency-analysis.md b/dependency-analysis.md deleted file mode 100644 index c4b2cc5df107e99d468b532748f796f0760f48f5..0000000000000000000000000000000000000000 --- a/dependency-analysis.md +++ /dev/null @@ -1,150 +0,0 @@ -# 🔍 AgentGraph 依赖项分析报告 - -## 📊 现状概览 - -- **总依赖数**: 30 个包 -- **当前镜像大小**: 4.27GB -- **部署环境**: Hugging Face Spaces (Docker) - -## 📋 依赖使用情况分析 - -### ✅ 高使用率 - 核心必需 - -| 依赖包 | 使用文件数 | 状态 | 说明 | -| ------------------- | ---------- | ------- | ------------ | -| `fastapi>=0.115.0` | 20+ | 🟢 必需 | Web 框架核心 | -| `sqlalchemy>=2.0.0` | 36+ | 🟢 必需 | 数据库 ORM | -| `pydantic>=2.10.0` | 36+ | 🟢 必需 | 数据验证 | -| `langfuse>=3.0.0` | 15+ | 🟢 必需 | AI 监控 | -| `langsmith>=0.3.38` | 12+ | 🟢 必需 | LLM 追踪 | -| `pandas>=1.3` | 10+ | 🟢 必需 | 数据处理 | -| `scikit-learn>=1.0` | 6+ | 🟢 必需 | 机器学习 | -| `openai>=1.76.2` | 7+ | 🟢 必需 | LLM API | - -### 🟡 中等使用率 - 功能相关 - -| 依赖包 | 使用文件数 | 状态 | 说明 | -| ----------------- | ---------- | --------- | ------------- | -| `dowhy>=0.12` | 3+ | 🟡 可选 | 因果分析功能 | -| `openlit>=1.33.0` | 3+ | 🟡 可选 | 监控工具 | -| `crewai>=0.108.0` | 2+ | 🟡 功能性 | 多 agent 系统 | - -### ❌ 低使用率 - 可优化 - -| 依赖包 | 使用文件数 | 状态 | 建议 | -| ------------------------ | ---------- | ----------- | ---------------------------- | -| `bottleneck>=1.3,<2.0.0` | 0 | 🔴 移除 | 未使用的 pandas 加速 | -| `pydot>=3.0.4` | 0 | 🔴 移除 | 图形绘制未使用 | -| `openai-agents==0.2.4` | 0 | 🔴 移除 | 未使用的 agent 框架 | -| `fire>=0.7.0` | 1 | 🟡 考虑移除 | CLI 工具,部署环境不需要 | -| `datasets>=3.6.0` | 1 | 🟡 考虑移除 | HuggingFace 数据集,使用率低 | - -## 🎯 优化建议 - -### 📦 立即移除 (节省 ~500MB) - -```toml -# 移除这些未使用的依赖 -# "bottleneck>=1.3,<2.0.0", # 未使用 -# "pydot>=3.0.4", # 未使用 -# "openai-agents==0.2.4", # 未使用 -``` - -### 🔧 条件移除 (节省 ~200MB) - -```toml -# 如果不需要CLI功能,可移除 -# "fire>=0.7.0", # 仅CLI使用 - -# 如果不使用HuggingFace数据集加载,可移除 -# "datasets>=3.6.0", # 使用率低 -``` - -### 🎨 功能性可选 (节省 ~800MB) - -```toml -# 如果不需要因果分析功能,可移除 -# "dowhy>=0.12", # 因果分析 -# "scipy>=1.7,<2.0.0", # dowhy依赖 - -# 如果不需要额外监控,可移除 -# "openlit>=1.33.0", # 监控工具 -``` - -## 📋 优化后的最小依赖集 - -### 🎯 核心运行时依赖 (约 2GB) - -```toml -dependencies = [ - # Web框架核心 - "fastapi>=0.115.0", - "uvicorn>=0.34.0", - "httpx>=0.27.0", - "python-multipart>=0.0.6,<1.0.0", - - # 数据层 - "sqlalchemy>=2.0.0", - "pydantic>=2.10.0", - - # AI/LLM核心 - "openai>=1.76.2", - "tiktoken>=0.9.0", - "langfuse>=3.0.0", - "langsmith>=0.3.38", - "litellm>=1.60.0", - "crewai>=0.108.0", - "crewai-tools>=0.38.1", - - # 数据处理核心 - "pandas>=1.3", - "numpy>=1.23,<2.0.0", - "scikit-learn>=1.0", - - # 基础工具 - "python-dotenv>=1.0.0", - "jinja2>=3.1.2", - "psutil>=7.0.0", -] -``` - -## 🚀 部署优化步骤 - -### 1️⃣ 立即优化 (安全移除) - -```bash -# 移除确认未使用的依赖 -# 预计节省: 500MB, 构建时间: -2分钟 -``` - -### 2️⃣ 功能评估优化 - -```bash -# 评估是否需要因果分析、CLI工具等功能 -# 预计节省: 1GB, 构建时间: -3分钟 -``` - -### 3️⃣ 最小化部署 - -```bash -# 仅保留核心运行时依赖 -# 预计最终镜像: ~2GB (当前: 4.27GB) -# 构建时间: ~3-4分钟 (当前: 5-8分钟) -``` - -## 💡 建议的实施策略 - -1. **阶段 1**: 立即移除未使用依赖 (`bottleneck`, `pydot`, `openai-agents`) -2. **阶段 2**: 评估功能需求,移除可选依赖 -3. **阶段 3**: 创建生产环境专用的最小依赖配置 - -## ⚠️ 注意事项 - -- 移除依赖前先在本地测试 -- 某些依赖可能被间接使用 -- 考虑创建开发和生产两套依赖配置 -- 监控移除后的功能完整性 - ---- - -**总结**: 通过合理的依赖优化,可以将 Docker 镜像从 4.27GB 缩减到约 2GB,构建时间从 5-8 分钟缩减到 3-4 分钟。 diff --git a/deploy.sh b/deploy.sh old mode 100755 new mode 100644 index 81b9d5143ec1aec605ae70ba9eb87f42d7720950..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 --- a/deploy.sh +++ b/deploy.sh @@ -1,47 +0,0 @@ -#!/bin/bash - -# 🚀 AgentGraph 部署到 Hugging Face Spaces - -echo "🚀 开始部署 AgentGraph 到 Hugging Face Spaces..." -echo "" - -# 检查是否在正确目录 -if [ ! -f "Dockerfile" ] || [ ! -f "main.py" ]; then - echo "❌ 错误:请在 huggingface/AgentGraph 目录中运行此脚本" - exit 1 -fi - -echo "📋 准备提交文件..." - -# 添加所有文件到git -git add . - -echo "💬 创建提交..." -git commit -m "🚀 Deploy AgentGraph: Complete agent monitoring and knowledge graph system - -Features: -- 📊 Real-time agent monitoring dashboard -- 🕸️ Knowledge graph extraction from traces -- 📈 Interactive visualizations and analytics -- 🔄 Multi-agent system with CrewAI -- 🎨 Modern React + FastAPI architecture - -Ready for Docker deployment on HF Spaces (port 7860)" - -echo "🔄 推送到 Hugging Face..." -git push - -echo "" -echo "✅ 部署完成!" -echo "" -echo "🔗 你的应用将在几分钟内在以下地址可用:" -echo " https://huggingface.co/spaces/holistic-ai/AgentGraph" -echo "" -echo "📝 部署状态:" -echo " • ✅ Docker 配置完成" -echo " • ✅ 端口 7860 已配置" -echo " • ✅ 前后端已合并" -echo " • ✅ 文件已推送" -echo "" -echo "⏳ 等待 HF Spaces 构建容器..." -echo "" diff --git a/utils/__pycache__/__init__.cpython-311.pyc b/utils/__pycache__/__init__.cpython-311.pyc deleted file mode 100644 index df27512dfc6c4ad45d665b7736896ae7016833e2..0000000000000000000000000000000000000000 Binary files a/utils/__pycache__/__init__.cpython-311.pyc and /dev/null differ diff --git a/utils/__pycache__/__init__.cpython-312.pyc b/utils/__pycache__/__init__.cpython-312.pyc deleted file mode 100644 index 91755ecdb68b1362e72396c8132496fb9d1442d4..0000000000000000000000000000000000000000 Binary files a/utils/__pycache__/__init__.cpython-312.pyc and /dev/null differ diff --git a/utils/__pycache__/__init__.cpython-313.pyc b/utils/__pycache__/__init__.cpython-313.pyc deleted file mode 100644 index cae02e68fc20cda09d550cfcf8f03925eecdd630..0000000000000000000000000000000000000000 Binary files a/utils/__pycache__/__init__.cpython-313.pyc and /dev/null differ diff --git a/utils/__pycache__/config.cpython-311.pyc b/utils/__pycache__/config.cpython-311.pyc deleted file mode 100644 index 7df79e1f563a54243e43a29ce17e99ca2b0ab8e0..0000000000000000000000000000000000000000 Binary files a/utils/__pycache__/config.cpython-311.pyc and /dev/null differ diff --git a/utils/__pycache__/config.cpython-312.pyc b/utils/__pycache__/config.cpython-312.pyc deleted file mode 100644 index 2877a6ccec3878d19d319552d3528d0e69b651df..0000000000000000000000000000000000000000 Binary files a/utils/__pycache__/config.cpython-312.pyc and /dev/null differ diff --git a/utils/__pycache__/config.cpython-313.pyc b/utils/__pycache__/config.cpython-313.pyc deleted file mode 100644 index 51c86c0b6a4d6fef97c81cdf127767a858a5de55..0000000000000000000000000000000000000000 Binary files a/utils/__pycache__/config.cpython-313.pyc and /dev/null differ diff --git a/utils/__pycache__/fetch_langfuse_logs.cpython-311.pyc b/utils/__pycache__/fetch_langfuse_logs.cpython-311.pyc deleted file mode 100644 index 4b3bc4438c4c83edd8d9e982e0e20f3e450ebd74..0000000000000000000000000000000000000000 Binary files a/utils/__pycache__/fetch_langfuse_logs.cpython-311.pyc and /dev/null differ diff --git a/utils/__pycache__/fetch_langfuse_logs.cpython-312.pyc b/utils/__pycache__/fetch_langfuse_logs.cpython-312.pyc deleted file mode 100644 index 9ff5c20ab5ec36bcf745117b1d4b78607389f5de..0000000000000000000000000000000000000000 Binary files a/utils/__pycache__/fetch_langfuse_logs.cpython-312.pyc and /dev/null differ diff --git a/utils/__pycache__/fix_litellm_stop_param.cpython-311.pyc b/utils/__pycache__/fix_litellm_stop_param.cpython-311.pyc deleted file mode 100644 index 1146ff23b47b29a59badfdad8ee8536881542509..0000000000000000000000000000000000000000 Binary files a/utils/__pycache__/fix_litellm_stop_param.cpython-311.pyc and /dev/null differ diff --git a/utils/__pycache__/fix_litellm_stop_param.cpython-312.pyc b/utils/__pycache__/fix_litellm_stop_param.cpython-312.pyc deleted file mode 100644 index f5f80a1373c520721397891f1a17f5301f35f3ae..0000000000000000000000000000000000000000 Binary files a/utils/__pycache__/fix_litellm_stop_param.cpython-312.pyc and /dev/null differ diff --git a/utils/__pycache__/langfuse_downloader.cpython-311.pyc b/utils/__pycache__/langfuse_downloader.cpython-311.pyc deleted file mode 100644 index 3649e4575d20a9a568b5ef1fb1414af0247b3828..0000000000000000000000000000000000000000 Binary files a/utils/__pycache__/langfuse_downloader.cpython-311.pyc and /dev/null differ diff --git a/utils/__pycache__/langfuse_downloader.cpython-312.pyc b/utils/__pycache__/langfuse_downloader.cpython-312.pyc deleted file mode 100644 index 365651a5530dddf08d76c1a7980aca5d12758a2c..0000000000000000000000000000000000000000 Binary files a/utils/__pycache__/langfuse_downloader.cpython-312.pyc and /dev/null differ diff --git a/utils/__pycache__/langsmith_exporter.cpython-311.pyc b/utils/__pycache__/langsmith_exporter.cpython-311.pyc deleted file mode 100644 index 053b844a5c96b49f637e6454486cff45fb35af42..0000000000000000000000000000000000000000 Binary files a/utils/__pycache__/langsmith_exporter.cpython-311.pyc and /dev/null differ diff --git a/utils/__pycache__/langsmith_exporter.cpython-312.pyc b/utils/__pycache__/langsmith_exporter.cpython-312.pyc deleted file mode 100644 index 80227bbd584dddc14cf44adf092cbe29b0f7657f..0000000000000000000000000000000000000000 Binary files a/utils/__pycache__/langsmith_exporter.cpython-312.pyc and /dev/null differ diff --git a/utils/__pycache__/preprocess_traces.cpython-311.pyc b/utils/__pycache__/preprocess_traces.cpython-311.pyc deleted file mode 100644 index eb4d065186794432c87efce393242d58a40be772..0000000000000000000000000000000000000000 Binary files a/utils/__pycache__/preprocess_traces.cpython-311.pyc and /dev/null differ diff --git a/utils/__pycache__/preprocess_traces.cpython-312.pyc b/utils/__pycache__/preprocess_traces.cpython-312.pyc deleted file mode 100644 index 74fe90ffd718526e8c009d9df00730c9091926be..0000000000000000000000000000000000000000 Binary files a/utils/__pycache__/preprocess_traces.cpython-312.pyc and /dev/null differ diff --git a/utils/__pycache__/preprocess_traces_aggressive.cpython-311.pyc b/utils/__pycache__/preprocess_traces_aggressive.cpython-311.pyc deleted file mode 100644 index 78cfeab978a10d3a606b9fd6485889e36ab45beb..0000000000000000000000000000000000000000 Binary files a/utils/__pycache__/preprocess_traces_aggressive.cpython-311.pyc and /dev/null differ diff --git a/utils/__pycache__/preprocess_traces_raw.cpython-312.pyc b/utils/__pycache__/preprocess_traces_raw.cpython-312.pyc deleted file mode 100644 index b0820dee0d477fea550c065933a9f955b521d355..0000000000000000000000000000000000000000 Binary files a/utils/__pycache__/preprocess_traces_raw.cpython-312.pyc and /dev/null differ diff --git a/utils/__pycache__/preprocess_traces_variable.cpython-311.pyc b/utils/__pycache__/preprocess_traces_variable.cpython-311.pyc deleted file mode 100644 index fd3c344dd9a188309654bba80be4b2df72b8b129..0000000000000000000000000000000000000000 Binary files a/utils/__pycache__/preprocess_traces_variable.cpython-311.pyc and /dev/null differ diff --git a/utils/__pycache__/setup_env.cpython-311.pyc b/utils/__pycache__/setup_env.cpython-311.pyc deleted file mode 100644 index 0eda37a16e1d99376574585307497cc3a14426a2..0000000000000000000000000000000000000000 Binary files a/utils/__pycache__/setup_env.cpython-311.pyc and /dev/null differ diff --git a/utils/__pycache__/setup_env.cpython-312.pyc b/utils/__pycache__/setup_env.cpython-312.pyc deleted file mode 100644 index f6001d4472d50e40905f080be566e535042f2035..0000000000000000000000000000000000000000 Binary files a/utils/__pycache__/setup_env.cpython-312.pyc and /dev/null differ