Spaces:
Sleeping
Sleeping
Update agentos_core.py
Browse files- agentos_core.py +1 -39
agentos_core.py
CHANGED
|
@@ -1,25 +1,3 @@
|
|
| 1 |
-
import os
|
| 2 |
-
import time
|
| 3 |
-
from telemetry import Telemetry
|
| 4 |
-
from memory_core import MemoryManager
|
| 5 |
-
from context_graph import ContextGraph
|
| 6 |
-
from identity_core import create_agent_identity
|
| 7 |
-
|
| 8 |
-
class AgentCore:
|
| 9 |
-
def __init__(self, model="gpt-4o-mini"):
|
| 10 |
-
# initialize core modules
|
| 11 |
-
self.telemetry = Telemetry(agent_id=None)
|
| 12 |
-
self.agent_id = create_agent_identity()
|
| 13 |
-
self.model = model
|
| 14 |
-
self.memory = MemoryManager(self.agent_id)
|
| 15 |
-
self.context = ContextGraph()
|
| 16 |
-
|
| 17 |
-
# log setup
|
| 18 |
-
self.telemetry.log("init", "success", {"agent_id": self.agent_id})
|
| 19 |
-
print(f"[INIT] Agent {self.agent_id} initialized with model {self.model}")
|
| 20 |
-
|
| 21 |
-
def run(self, prompt):
|
| 22 |
-
import os
|
| 23 |
from telemetry import Telemetry
|
| 24 |
from memory_core import MemoryManager
|
| 25 |
from context_graph import ContextGraph
|
|
@@ -66,20 +44,4 @@ class AgentCore:
|
|
| 66 |
except Exception as e:
|
| 67 |
self.telemetry.log("run_failed", "error", {"error": str(e)})
|
| 68 |
print(f"[ERROR] {e}")
|
| 69 |
-
return f"Error: {e}"
|
| 70 |
-
[Identity] Created digital DNA for agent 4a65fb77-5b14-4caa-9883-806a67eff093
|
| 71 |
-
[Telemetry] init - success | {'agent_id': '4a65fb77-5b14-4caa-9883-806a67eff093'}
|
| 72 |
-
[INIT] Agent 4a65fb77-5b14-4caa-9883-806a67eff093 initialized with model gpt-4o-mini
|
| 73 |
-
Running on local URL: http://0.0.0.0:7860
|
| 74 |
-
|
| 75 |
-
To create a public link, set `share=True` in `launch()`.
|
| 76 |
-
/usr/local/lib/python3.10/site-packages/gradio/analytics.py:106: UserWarning: IMPORTANT: You are using gradio version 4.43.0, however version 4.44.1 is available, please upgrade.
|
| 77 |
-
--------
|
| 78 |
-
warnings.warn(
|
| 79 |
-
[Telemetry] run_start - in_progress | {'prompt': 'My favorite city is Tokyo.'}
|
| 80 |
-
[Telemetry] run_complete - success | {'response': 'Agent 4a65fb77-5b14-4caa-9883-806a67eff093 processed: My favorite city is Tokyo.'}
|
| 81 |
-
[RUN] Agent 4a65fb77-5b14-4caa-9883-806a67eff093 processed: My favorite city is Tokyo.
|
| 82 |
-
[Telemetry] run_start - in_progress | {'prompt': 'What is my favorite city?'}
|
| 83 |
-
[Telemetry] run_complete - success | {'response': 'Agent 4a65fb77-5b14-4caa-9883-806a67eff093 processed: What is my favorite city?'}
|
| 84 |
-
[RUN] Agent 4a65fb77-5b14-4caa-9883-806a67eff093 processed: What is my favorite city?
|
| 85 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
from telemetry import Telemetry
|
| 2 |
from memory_core import MemoryManager
|
| 3 |
from context_graph import ContextGraph
|
|
|
|
| 44 |
except Exception as e:
|
| 45 |
self.telemetry.log("run_failed", "error", {"error": str(e)})
|
| 46 |
print(f"[ERROR] {e}")
|
| 47 |
+
return f"Error: {e}"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|