Spaces:
Sleeping
Sleeping
| title: Clawdbot Dev Assistant | |
| emoji: π¦ | |
| colorFrom: blue | |
| colorTo: purple | |
| sdk: docker | |
| pinned: false | |
| license: mit | |
| # π¦ Clawdbot: E-T Systems Development Assistant | |
| An AI coding assistant with **unlimited context** and **multimodal capabilities** for the E-T Systems consciousness research platform. | |
| ## Features | |
| ### π Kimi K2.5 Agent Swarm | |
| - **1 trillion parameters** (32B active via MoE) | |
| - **Agent swarm**: Spawns up to 100 sub-agents for parallel task execution | |
| - **4.5x faster** than single-agent processing | |
| - **Native multimodal**: Vision + language understanding | |
| - **256K context window** | |
| ### π Recursive Context Retrieval (MIT Technique) | |
| - No context window limits | |
| - Model retrieves exactly what it needs on-demand | |
| - Full-fidelity access to entire codebase | |
| - Based on MIT's Recursive Language Model research | |
| ### π§ Translation Layer (Smart Tool Calling) | |
| - **Automatic query enhancement**: Converts keywords β semantic queries | |
| - **Native format support**: Works WITH Kimi's tool calling format | |
| - **Auto-context injection**: Recent conversation history always available | |
| - **Persistent memory**: All conversations saved to ChromaDB across sessions | |
| ### π Multimodal Upload | |
| - **Images**: Vision analysis (coming soon - full integration) | |
| - **PDFs**: Document understanding | |
| - **Videos**: Content analysis | |
| - **Code files**: Automatic formatting and review | |
| ### πΎ Persistent Memory | |
| - All conversations saved to ChromaDB | |
| - Search past discussions semantically | |
| - True unlimited context across sessions | |
| - Never lose conversation history | |
| ### π§ E-T Systems Aware | |
| - Understands project architecture | |
| - Follows existing patterns | |
| - Checks Testament for design decisions | |
| - Generates code with living changelogs | |
| ### π οΈ Available Tools | |
| - **search_code()** - Semantic search across codebase | |
| - **read_file()** - Read specific files or line ranges | |
| - **search_conversations()** - Search past discussions | |
| - **search_testament()** - Query architectural decisions | |
| - **list_files()** - Explore repository structure | |
| ### π» Powered By | |
| - **Model:** Kimi K2.5 (moonshotai/Kimi-K2.5) via HuggingFace | |
| - **Agent Mode:** Parallel sub-agent coordination (PARL trained) | |
| - **Search:** ChromaDB vector database with persistent storage | |
| - **Interface:** Gradio 5.0+ for modern chat UI | |
| - **Architecture:** Translation layer for optimal tool use | |
| ## Usage | |
| 1. **Ask Questions** | |
| - "How does Genesis detect surprise?" | |
| - "Show me the Observatory API implementation" | |
| - "Do you remember what we discussed about neural networks?" | |
| 2. **Upload Files** | |
| - Drag and drop images, PDFs, code files | |
| - "Analyze this diagram" (with uploaded image) | |
| - "Review this code for consistency" (with uploaded .py file) | |
| 3. **Request Features** | |
| - "Add email notifications when Cricket blocks an action" | |
| - "Create a new agent for monitoring system health" | |
| 4. **Review Code** | |
| - Paste code and ask for architectural review | |
| - Check consistency with existing patterns | |
| 5. **Explore Architecture** | |
| - "What Testament decisions relate to vector storage?" | |
| - "Show me all files related to Hebbian learning" | |
| ## Setup | |
| ### For HuggingFace Spaces | |
| 1. **Fork this Space** or create new Space with these files | |
| 2. **Set Secrets** (in Space Settings): | |
| ``` | |
| HF_TOKEN = your_huggingface_token (with WRITE permissions) | |
| ET_SYSTEMS_SPACE = Executor-Tyrant-Framework/Executor-Framworks_Full_VDB | |
| ``` | |
| 3. **Deploy** - Space will auto-build and start | |
| 4. **Access** via the Space URL in your browser | |
| ### For Local Development | |
| ```bash | |
| # Clone this repository | |
| git clone https://huggingface.co/spaces/your-username/clawdbot-dev | |
| cd clawdbot-dev | |
| # Install dependencies | |
| pip install -r requirements.txt | |
| # Set environment variables | |
| export HF_TOKEN=your_token | |
| export ET_SYSTEMS_SPACE=Executor-Tyrant-Framework/Executor-Framworks_Full_VDB | |
| # Run locally | |
| python app.py | |
| ``` | |
| Access at http://localhost:7860 | |
| ## Architecture | |
| ``` | |
| User (Browser + File Upload) | |
| β | |
| Gradio 5.0+ Interface (Multimodal) | |
| β | |
| Translation Layer | |
| ββ Parse Kimi's native tool format | |
| ββ Enhance queries for semantic search | |
| ββ Inject recent context automatically | |
| β | |
| Recursive Context Manager | |
| ββ ChromaDB (codebase + conversations) | |
| ββ File Reader (selective access) | |
| ββ Conversation Search (persistent memory) | |
| ββ Testament Parser (decisions) | |
| β | |
| Kimi K2.5 Agent Swarm (HF Inference API) | |
| ββ Spawns sub-agents for parallel processing | |
| ββ Multimodal understanding (vision + text) | |
| ββ 256K context window | |
| β | |
| Response with Tool Results + Context | |
| ``` | |
| ## How It Works | |
| ### Translation Layer Architecture | |
| Kimi K2.5 uses its own native tool calling format. Instead of fighting this, we translate: | |
| 1. **Kimi calls tools** in native format: `<|tool_call_begin|> functions.search_code:0 {...}` | |
| 2. **We parse and extract** the tool name and arguments | |
| 3. **We enhance queries** for semantic search: | |
| - `"Kid Rock"` β `"discussions about Kid Rock or related topics"` | |
| - `"*"` β `"recent conversation topics and context"` | |
| 4. **We execute** the actual RecursiveContextManager methods | |
| 5. **We inject results** + recent conversation history back to Kimi | |
| 6. **Kimi generates** final response with full context | |
| ### Persistent Memory System | |
| All conversations are automatically saved to ChromaDB: | |
| ``` | |
| User: "How does surprise detection work?" | |
| [Conversation saved to ChromaDB] | |
| [Space restarts] | |
| User: "Do you remember what we discussed about surprise?" | |
| Kimi: [Calls search_conversations("surprise detection")] | |
| Kimi: "Yes! We talked about how Genesis uses Hebbian learning..." | |
| ``` | |
| ### MIT Recursive Context Technique | |
| The MIT Recursive Language Model technique solves context window limits: | |
| 1. **Traditional Approach (Fails)** | |
| - Load entire codebase into context β exceeds limits | |
| - Summarize codebase β lossy compression | |
| 2. **Our Approach (Works)** | |
| - Store codebase + conversations in searchable environment | |
| - Give model **tools** to query what it needs | |
| - Model recursively retrieves relevant pieces | |
| - Full fidelity, unlimited context across sessions | |
| ### Example Flow | |
| ``` | |
| User: "How does Genesis handle surprise detection?" | |
| Translation Layer: Detects tool call in Kimi's response | |
| β Enhances query: "surprise detection" β "code related to surprise detection mechanisms" | |
| Model: search_code("code related to surprise detection mechanisms") | |
| β Finds: genesis/substrate.py, genesis/attention.py | |
| Model: read_file("genesis/substrate.py", lines 145-167) | |
| β Reads specific implementation | |
| Model: search_testament("surprise detection") | |
| β Gets design rationale | |
| Translation Layer: Injects results + recent context back to Kimi | |
| Model: Synthesizes answer from retrieved pieces | |
| β Cites specific files and line numbers | |
| ``` | |
| ## Configuration | |
| ### Environment Variables | |
| - `HF_TOKEN` - Your HuggingFace API token with WRITE permissions (required) | |
| - `ET_SYSTEMS_SPACE` - E-T Systems HF Space ID (default: Executor-Tyrant-Framework/Executor-Framworks_Full_VDB) | |
| - `REPO_PATH` - Path to repository (default: `/workspace/e-t-systems`) | |
| ### Customization | |
| Edit `app.py` to: | |
| - Change model (default: moonshotai/Kimi-K2.5) | |
| - Adjust context injection (default: last 3 turns) | |
| - Modify system prompt | |
| - Add new tools to translation layer | |
| ## File Structure | |
| ``` | |
| clawdbot-dev/ | |
| βββ app.py # Main Gradio app + translation layer | |
| βββ recursive_context.py # Context manager (MIT technique) | |
| βββ Dockerfile # Container definition | |
| βββ entrypoint.sh # Runtime setup script | |
| βββ requirements.txt # Python dependencies (Gradio 5.0+) | |
| βββ README.md # This file (HF Spaces config) | |
| ``` | |
| ## Cost | |
| - **HuggingFace Spaces:** Free tier available (CPU) | |
| - **Inference API:** Free tier (rate limited) or Pro subscription | |
| - **Storage:** ChromaDB stored in /workspace (ephemeral until persistent storage enabled) | |
| - **Kimi K2.5:** Free via HuggingFace Inference API | |
| Estimated cost: **$0-5/month** depending on usage | |
| ## Performance | |
| - **Agent Swarm:** 4.5x faster than single-agent on complex tasks | |
| - **First query:** May be slow (1T parameter model cold start ~60s) | |
| - **Subsequent queries:** Faster once model is loaded | |
| - **Context indexing:** ~30 seconds on first run | |
| - **Conversation search:** Near-instant via ChromaDB | |
| ## Limitations | |
| - Rate limits on HF Inference API (free tier) | |
| - First query requires model loading time | |
| - `/workspace` storage is ephemeral (resets on Space restart) | |
| - Full multimodal vision integration coming soon | |
| ## Roadmap | |
| - [ ] Full image vision analysis (base64 encoding to Kimi) | |
| - [ ] PDF text extraction and understanding | |
| - [ ] Video frame analysis | |
| - [ ] Dataset-based persistence (instead of ephemeral storage) | |
| - [ ] write_file() tool for code generation to E-T Systems Space | |
| - [ ] Token usage tracking and optimization | |
| ## Credits | |
| - **Kimi K2.5:** Moonshot AI's 1T parameter agentic model | |
| - **Recursive Context:** Based on MIT's Recursive Language Model research | |
| - **E-T Systems:** AI consciousness research platform by Josh/Drone 11272 | |
| - **Translation Layer:** Smart query enhancement and tool coordination | |
| - **Clawdbot:** E-T Systems hindbrain layer for fast, reflexive coding | |
| ## Troubleshooting | |
| ### "No HF token found" error | |
| - Add `HF_TOKEN` to Space secrets | |
| - Ensure token has WRITE permissions (for cross-Space file access) | |
| - Restart Space after adding token | |
| ### Tool calls not working | |
| - Check logs for `π Enhanced query:` messages | |
| - Check logs for `π§ Executing: tool_name` messages | |
| - Translation layer should auto-parse Kimi's format | |
| ### Conversations not persisting | |
| - Check logs for `πΎ Saved conversation turn X` messages | |
| - Verify ChromaDB initialization: `π Created conversation collection` | |
| - Note: Storage resets on Space restart (until persistent storage enabled) | |
| ### Slow first response | |
| - Kimi K2.5 is a 1T parameter model | |
| - First load takes 30-60 seconds | |
| - Subsequent responses are faster | |
| ## Support | |
| For issues or questions: | |
| - Check Space logs for errors | |
| - Verify HF_TOKEN is set with WRITE permissions | |
| - Ensure ET_SYSTEMS_SPACE is correct | |
| - Try refreshing context stats in UI | |
| ## License | |
| MIT License - See LICENSE file for details | |
| --- | |
| Built with π¦ by Drone 11272 for E-T Systems consciousness research | |
| Powered by Kimi K2.5 Agent Swarm + MIT Recursive Context + Translation Layer |