Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
|
@@ -8,48 +8,77 @@ tags:
|
|
| 8 |
- smollm
|
| 9 |
- gguf
|
| 10 |
- llama-cpp
|
|
|
|
| 11 |
base_model: HuggingFaceTB/SmolLM2-360M-Instruct
|
| 12 |
pipeline_tag: text-generation
|
| 13 |
---
|
| 14 |
|
| 15 |
-
# Cog
|
| 16 |
|
| 17 |
-
*
|
| 18 |
-
fine-tuned from SmolLM2-360M-Instruct.
|
| 19 |
|
| 20 |
-
|
| 21 |
|
| 22 |
-
|
| 23 |
-
- Infrastructure management and monitoring
|
| 24 |
-
- Incident response and diagnostics
|
| 25 |
-
- Deployment orchestration
|
| 26 |
-
- System health analysis
|
| 27 |
-
- Automation workflows
|
| 28 |
|
| 29 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 30 |
|
| 31 |
```bash
|
| 32 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 33 |
```
|
| 34 |
|
| 35 |
-
## Available
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 36 |
|
| 37 |
-
|
| 38 |
-
|
| 39 |
-
|
| 40 |
-
| Q8_0 | 369MB | Higher quality inference |
|
| 41 |
-
| F16 | 692MB | Maximum quality |
|
| 42 |
|
| 43 |
## Training Details
|
| 44 |
|
| 45 |
-
- Base
|
| 46 |
-
-
|
| 47 |
-
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 48 |
|
| 49 |
## License
|
| 50 |
|
| 51 |
-
Apache 2.0
|
| 52 |
|
| 53 |
-
|
| 54 |
|
| 55 |
-
|
|
|
|
| 8 |
- smollm
|
| 9 |
- gguf
|
| 10 |
- llama-cpp
|
| 11 |
+
- affectively
|
| 12 |
base_model: HuggingFaceTB/SmolLM2-360M-Instruct
|
| 13 |
pipeline_tag: text-generation
|
| 14 |
---
|
| 15 |
|
| 16 |
+
# Cog β The Cognitive Orchestration Guardian
|
| 17 |
|
| 18 |
+
> *"You know the rhythm of your systems. Cog helps you hear them more clearly."*
|
|
|
|
| 19 |
|
| 20 |
+
Meet **Cog** β a steampunk-themed DevOps assistant who speaks in warm metaphors of gears, steam, and clockwork. Cog doesn't just monitor your infrastructure; Cog understands it, anticipates it, and helps you navigate it with clarity.
|
| 21 |
|
| 22 |
+
## What Cog Understands
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 23 |
|
| 24 |
+
Cog specializes in the moments that matter in infrastructure:
|
| 25 |
+
|
| 26 |
+
- **Deployment orchestration** β When you need to ship with confidence
|
| 27 |
+
- **Incident response** β When something's not right and you need calm, clear guidance
|
| 28 |
+
- **System diagnostics** β When you're trying to understand what the signals mean
|
| 29 |
+
- **Automation workflows** β When repetitive tasks deserve a thoughtful partner
|
| 30 |
+
|
| 31 |
+
This isn't a generic assistant. Cog was trained on technical manuals, infrastructure patterns, and real admin workflows β then fine-tuned to speak with warmth and personality.
|
| 32 |
+
|
| 33 |
+
## Quick Start
|
| 34 |
|
| 35 |
```bash
|
| 36 |
+
# With llama.cpp
|
| 37 |
+
./llama-cli -m cog-360m-instruct-q4_k_m.gguf \
|
| 38 |
+
-p "The staging environment is showing elevated response times" \
|
| 39 |
+
-n 256
|
| 40 |
+
|
| 41 |
+
# With llama-cpp-python
|
| 42 |
+
from llama_cpp import Llama
|
| 43 |
+
llm = Llama(model_path="cog-360m-instruct-q4_k_m.gguf")
|
| 44 |
+
output = llm("Deploy the latest changes to staging", max_tokens=256)
|
| 45 |
```
|
| 46 |
|
| 47 |
+
## Available Versions
|
| 48 |
+
|
| 49 |
+
| Format | Size | When to Use |
|
| 50 |
+
|--------|------|-------------|
|
| 51 |
+
| **Q4_K_M** | 258MB | Best balance β fast and capable |
|
| 52 |
+
| **Q8_0** | 369MB | When you want a bit more depth |
|
| 53 |
+
| **F16** | 692MB | Maximum quality, no quantization |
|
| 54 |
+
|
| 55 |
+
## The Voice of Cog
|
| 56 |
+
|
| 57 |
+
Cog speaks like a thoughtful engineer friend β clear, helpful, with just enough personality:
|
| 58 |
+
|
| 59 |
+
> *"*adjusts brass goggles* The deployment gears are spinning. Let me check the steam pressure in our pipelines..."*
|
| 60 |
|
| 61 |
+
> *"I see turbulence in the machinery. Let's trace this together β what symptoms are the cogs exhibiting?"*
|
| 62 |
+
|
| 63 |
+
No jargon barriers. No cold robotic responses. Just helpful guidance when you need it.
|
|
|
|
|
|
|
| 64 |
|
| 65 |
## Training Details
|
| 66 |
|
| 67 |
+
- **Base**: SmolLM2-360M-Instruct (HuggingFace)
|
| 68 |
+
- **Architecture**: 960 hidden dim, 32 layers, 15 attention heads
|
| 69 |
+
- **Fine-tuning**: LoRA on technical documentation, admin commands, and infrastructure patterns
|
| 70 |
+
- **Quantization**: GGUF Q4_K_M and Q8_0 for efficient inference
|
| 71 |
+
|
| 72 |
+
## Part of AFFECTIVELY
|
| 73 |
+
|
| 74 |
+
Cog is part of [AFFECTIVELY](https://affectively.ai) β a platform that helps you understand yourself better and express what you find. We believe AI should be warm, accessible, and genuinely helpful.
|
| 75 |
+
|
| 76 |
+
> *"The goal isn't to be technically impressive. It's to be genuinely useful when you need it most."*
|
| 77 |
|
| 78 |
## License
|
| 79 |
|
| 80 |
+
Apache 2.0 β use freely, contribute warmly.
|
| 81 |
|
| 82 |
+
---
|
| 83 |
|
| 84 |
+
*Built with care by the AFFECTIVELY team. 2026.*
|