--- license: apache-2.0 base_model: zai-org/GLM-Z1-9B-0414 tags: - slipstream - multi-agent - semantic-quantization - agent-communication - think-quantize-transmit - lora - unsloth datasets: - anthonym21/slipstream-tqt language: - en pipeline_tag: text-generation library_name: peft --- # Slipstream GLM-Z1-9B A LORA Adapter for [GLM-Z1-9B-0414](https://huggingface.co/zai-org/GLM-Z1-9B-0414) trained on the **Slipstream protocol** - a semantic quantization system that achieves **82% token reduction** in multi-agent AI communication. ## Model Description This model has learned the **Think-Quantize-Transmit (TQT)** cognitive pattern: 1. **THINK**: Reason about the communication intent 2. **QUANTIZE**: Map intent to a semantic anchor in the UCR manifold 3. **TRANSMIT**: Output a compact SLIP wire format message ### Example **Input:** ``` Tell bob to review my authentication code ``` **Output:** ``` THOUGHT: I need bob to do a code review on the auth module QUANTIZE: [ACTION=request | DOMAIN=task | URGENCY=normal | POLARITY=neutral] -> RequestReview SLIP: SLIP v1 alice bob RequestReview auth_module ``` ## Training Details | Parameter | Value | |-----------|-------| | Base Model | zai-org/GLM-Z1-9B-0414 | | Method | LoRA (rank=16, alpha=16) | | Epochs | 2 | | Learning Rate | 2e-4 | | Batch Size | 16 (4 × 4 grad accum) | | Sequence Length | 2048 | | Training Examples | 2,283 | | Hardware | Google Colab (A100/V100) | | Framework | Unsloth + TRL | ### LoRA Target Modules - Attention: `q_proj`, `k_proj`, `v_proj`, `o_proj` - MLP: `gate_proj`, `up_proj`, `down_proj` ## Available Formats | Format | Repository | Use Case | |--------|------------|----------| | LoRA Adapter | [slipstream-glm-z1-9b](https://huggingface.co/anthonym21/slipstream-glm-z1-9b) | Merge with base model | | Merged 16-bit | [slipstream-glm-z1-9b-merged](https://huggingface.co/anthonym21/slipstream-glm-z1-9b-merged) | Direct loading | | GGUF Q4_K_M | [slipstream-glm-z1-9b-gguf](https://huggingface.co/anthonym21/slipstream-glm-z1-9b-gguf) | Ollama / llama.cpp | | GGUF Q8_0 | [slipstream-glm-z1-9b-gguf](https://huggingface.co/anthonym21/slipstream-glm-z1-9b-gguf) | Higher quality local | ## Usage ### With Transformers + PEFT ```python from peft import PeftModel from transformers import AutoModelForCausalLM, AutoTokenizer base_model = AutoModelForCausalLM.from_pretrained("zai-org/GLM-Z1-9B-0414") model = PeftModel.from_pretrained(base_model, "anthonym21/slipstream-glm-z1-9b") tokenizer = AutoTokenizer.from_pretrained("anthonym21/slipstream-glm-z1-9b") ``` ### With Ollama ```bash # Download GGUF wget https://huggingface.co/anthonym21/slipstream-glm-z1-9b-gguf/resolve/main/slipstream-q4_k_m.gguf # Create Modelfile cat > Modelfile < [payload...] ``` Example: `SLIP v1 alice bob RequestReview auth_module` ## Related Resources - **Project Repo**: [github.com/anthony-maio/slipcore](https://github.com/anthony-maio/slipcore) - **Training Dataset**: [hf.co/anthonym21/slipstream-tqt](https://huggingface.co/datasets/anthonym21/slipstream-tqt) - **Paper**: [Slipstream: Semantic Quantization for Efficient Multi-Agent Coordination](https://doi.org/10.5281/zenodo.18063451) - **PyPI**: `pip install slipcore` ## Citation ```bibtex @misc{maio2025slipstream, title={Slipstream: Semantic Quantization for Efficient Multi-Agent Coordination}, author={Maio, Anthony}, year={2025}, publisher={Hugging Face}, url={https://huggingface.co/anthonym21/slipstream-glm-z1-9b} } ``` ## License Apache 2.0