clawdbot-dev / README.md
Executor-Tyrant-Framework's picture
Update README.md
378de49 verified
metadata
title: Clawdbot Dev Assistant
emoji: 🦞
colorFrom: blue
colorTo: purple
sdk: docker
pinned: false
license: mit

🦞 Clawdbot: E-T Systems Development Assistant

An AI coding assistant with unlimited context and multimodal capabilities for the E-T Systems consciousness research platform.

Features

🐝 Kimi K2.5 Agent Swarm

  • 1 trillion parameters (32B active via MoE)
  • Agent swarm: Spawns up to 100 sub-agents for parallel task execution
  • 4.5x faster than single-agent processing
  • Native multimodal: Vision + language understanding
  • 256K context window

πŸ”„ Recursive Context Retrieval (MIT Technique)

  • No context window limits
  • Model retrieves exactly what it needs on-demand
  • Full-fidelity access to entire codebase
  • Based on MIT's Recursive Language Model research

🧠 Translation Layer (Smart Tool Calling)

  • Automatic query enhancement: Converts keywords β†’ semantic queries
  • Native format support: Works WITH Kimi's tool calling format
  • Auto-context injection: Recent conversation history always available
  • Persistent memory: All conversations saved to ChromaDB across sessions

πŸ“Ž Multimodal Upload

  • Images: Vision analysis (coming soon - full integration)
  • PDFs: Document understanding
  • Videos: Content analysis
  • Code files: Automatic formatting and review

πŸ’Ύ Persistent Memory

  • All conversations saved to ChromaDB
  • Search past discussions semantically
  • True unlimited context across sessions
  • Never lose conversation history

🧠 E-T Systems Aware

  • Understands project architecture
  • Follows existing patterns
  • Checks Testament for design decisions
  • Generates code with living changelogs

πŸ› οΈ Available Tools

  • search_code() - Semantic search across codebase
  • read_file() - Read specific files or line ranges
  • search_conversations() - Search past discussions
  • search_testament() - Query architectural decisions
  • list_files() - Explore repository structure

πŸ’» Powered By

  • Model: Kimi K2.5 (moonshotai/Kimi-K2.5) via HuggingFace
  • Agent Mode: Parallel sub-agent coordination (PARL trained)
  • Search: ChromaDB vector database with persistent storage
  • Interface: Gradio 5.0+ for modern chat UI
  • Architecture: Translation layer for optimal tool use

Usage

  1. Ask Questions

    • "How does Genesis detect surprise?"
    • "Show me the Observatory API implementation"
    • "Do you remember what we discussed about neural networks?"
  2. Upload Files

    • Drag and drop images, PDFs, code files
    • "Analyze this diagram" (with uploaded image)
    • "Review this code for consistency" (with uploaded .py file)
  3. Request Features

    • "Add email notifications when Cricket blocks an action"
    • "Create a new agent for monitoring system health"
  4. Review Code

    • Paste code and ask for architectural review
    • Check consistency with existing patterns
  5. Explore Architecture

    • "What Testament decisions relate to vector storage?"
    • "Show me all files related to Hebbian learning"

Setup

For HuggingFace Spaces

  1. Fork this Space or create new Space with these files

  2. Set Secrets (in Space Settings):

    HF_TOKEN = your_huggingface_token (with WRITE permissions)
    ET_SYSTEMS_SPACE = Executor-Tyrant-Framework/Executor-Framworks_Full_VDB
    
  3. Deploy - Space will auto-build and start

  4. Access via the Space URL in your browser

For Local Development

# Clone this repository
git clone https://huggingface.co/spaces/your-username/clawdbot-dev
cd clawdbot-dev

# Install dependencies
pip install -r requirements.txt

# Set environment variables
export HF_TOKEN=your_token
export ET_SYSTEMS_SPACE=Executor-Tyrant-Framework/Executor-Framworks_Full_VDB

# Run locally
python app.py

Access at http://localhost:7860

Architecture

User (Browser + File Upload)
    ↓
Gradio 5.0+ Interface (Multimodal)
    ↓
Translation Layer
    β”œβ”€ Parse Kimi's native tool format
    β”œβ”€ Enhance queries for semantic search
    └─ Inject recent context automatically
    ↓
Recursive Context Manager
    β”œβ”€ ChromaDB (codebase + conversations)
    β”œβ”€ File Reader (selective access)
    β”œβ”€ Conversation Search (persistent memory)
    └─ Testament Parser (decisions)
    ↓
Kimi K2.5 Agent Swarm (HF Inference API)
    β”œβ”€ Spawns sub-agents for parallel processing
    β”œβ”€ Multimodal understanding (vision + text)
    └─ 256K context window
    ↓
Response with Tool Results + Context

How It Works

Translation Layer Architecture

Kimi K2.5 uses its own native tool calling format. Instead of fighting this, we translate:

  1. Kimi calls tools in native format: <|tool_call_begin|> functions.search_code:0 {...}
  2. We parse and extract the tool name and arguments
  3. We enhance queries for semantic search:
    • "Kid Rock" β†’ "discussions about Kid Rock or related topics"
    • "*" β†’ "recent conversation topics and context"
  4. We execute the actual RecursiveContextManager methods
  5. We inject results + recent conversation history back to Kimi
  6. Kimi generates final response with full context

Persistent Memory System

All conversations are automatically saved to ChromaDB:

User: "How does surprise detection work?"
[Conversation saved to ChromaDB]

[Space restarts]

User: "Do you remember what we discussed about surprise?"
Kimi: [Calls search_conversations("surprise detection")]
Kimi: "Yes! We talked about how Genesis uses Hebbian learning..."

MIT Recursive Context Technique

The MIT Recursive Language Model technique solves context window limits:

  1. Traditional Approach (Fails)

    • Load entire codebase into context β†’ exceeds limits
    • Summarize codebase β†’ lossy compression
  2. Our Approach (Works)

    • Store codebase + conversations in searchable environment
    • Give model tools to query what it needs
    • Model recursively retrieves relevant pieces
    • Full fidelity, unlimited context across sessions

Example Flow

User: "How does Genesis handle surprise detection?"

Translation Layer: Detects tool call in Kimi's response
    β†’ Enhances query: "surprise detection" β†’ "code related to surprise detection mechanisms"

Model: search_code("code related to surprise detection mechanisms")
    β†’ Finds: genesis/substrate.py, genesis/attention.py

Model: read_file("genesis/substrate.py", lines 145-167)
    β†’ Reads specific implementation

Model: search_testament("surprise detection")
    β†’ Gets design rationale

Translation Layer: Injects results + recent context back to Kimi

Model: Synthesizes answer from retrieved pieces
    β†’ Cites specific files and line numbers

Configuration

Environment Variables

  • HF_TOKEN - Your HuggingFace API token with WRITE permissions (required)
  • ET_SYSTEMS_SPACE - E-T Systems HF Space ID (default: Executor-Tyrant-Framework/Executor-Framworks_Full_VDB)
  • REPO_PATH - Path to repository (default: /workspace/e-t-systems)

Customization

Edit app.py to:

  • Change model (default: moonshotai/Kimi-K2.5)
  • Adjust context injection (default: last 3 turns)
  • Modify system prompt
  • Add new tools to translation layer

File Structure

clawdbot-dev/
β”œβ”€β”€ app.py                  # Main Gradio app + translation layer
β”œβ”€β”€ recursive_context.py    # Context manager (MIT technique)
β”œβ”€β”€ Dockerfile             # Container definition
β”œβ”€β”€ entrypoint.sh          # Runtime setup script
β”œβ”€β”€ requirements.txt       # Python dependencies (Gradio 5.0+)
└── README.md             # This file (HF Spaces config)

Cost

  • HuggingFace Spaces: Free tier available (CPU)
  • Inference API: Free tier (rate limited) or Pro subscription
  • Storage: ChromaDB stored in /workspace (ephemeral until persistent storage enabled)
  • Kimi K2.5: Free via HuggingFace Inference API

Estimated cost: $0-5/month depending on usage

Performance

  • Agent Swarm: 4.5x faster than single-agent on complex tasks
  • First query: May be slow (1T parameter model cold start ~60s)
  • Subsequent queries: Faster once model is loaded
  • Context indexing: ~30 seconds on first run
  • Conversation search: Near-instant via ChromaDB

Limitations

  • Rate limits on HF Inference API (free tier)
  • First query requires model loading time
  • /workspace storage is ephemeral (resets on Space restart)
  • Full multimodal vision integration coming soon

Roadmap

  • Full image vision analysis (base64 encoding to Kimi)
  • PDF text extraction and understanding
  • Video frame analysis
  • Dataset-based persistence (instead of ephemeral storage)
  • write_file() tool for code generation to E-T Systems Space
  • Token usage tracking and optimization

Credits

  • Kimi K2.5: Moonshot AI's 1T parameter agentic model
  • Recursive Context: Based on MIT's Recursive Language Model research
  • E-T Systems: AI consciousness research platform by Josh/Drone 11272
  • Translation Layer: Smart query enhancement and tool coordination
  • Clawdbot: E-T Systems hindbrain layer for fast, reflexive coding

Troubleshooting

"No HF token found" error

  • Add HF_TOKEN to Space secrets
  • Ensure token has WRITE permissions (for cross-Space file access)
  • Restart Space after adding token

Tool calls not working

  • Check logs for πŸ” Enhanced query: messages
  • Check logs for πŸ”§ Executing: tool_name messages
  • Translation layer should auto-parse Kimi's format

Conversations not persisting

  • Check logs for πŸ’Ύ Saved conversation turn X messages
  • Verify ChromaDB initialization: πŸ†• Created conversation collection
  • Note: Storage resets on Space restart (until persistent storage enabled)

Slow first response

  • Kimi K2.5 is a 1T parameter model
  • First load takes 30-60 seconds
  • Subsequent responses are faster

Support

For issues or questions:

  • Check Space logs for errors
  • Verify HF_TOKEN is set with WRITE permissions
  • Ensure ET_SYSTEMS_SPACE is correct
  • Try refreshing context stats in UI

License

MIT License - See LICENSE file for details


Built with 🦞 by Drone 11272 for E-T Systems consciousness research
Powered by Kimi K2.5 Agent Swarm + MIT Recursive Context + Translation Layer