File size: 4,113 Bytes
680a754 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 |
---
language:
- en
license: mit
library_name: founder-game-classifier
tags:
- text-classification
- founders
- content-analysis
- sentence-transformers
datasets:
- custom
metrics:
- accuracy
pipeline_tag: text-classification
---
# Founder Game Classifier
A trained classifier that identifies which of **6 founder games** a piece of content belongs to.
## Model Description
This model classifies text content into one of six "founder games" - patterns of communication and content creation common among founders, creators, and thought leaders.
### The 6 Games
| Game | Name | Description |
|------|------|-------------|
| **G1** | Identity/Canon | Recruiting into identity, lineage, belonging, status, canon formation |
| **G2** | Ideas/Play Mining | Extracting reusable plays, tactics, heuristics; "do this / steal this" |
| **G3** | Models/Understanding | Building mental models, frameworks, mechanisms, explanations |
| **G4** | Performance/Competition | Winning, dominance, execution, metrics, endurance, zero-sum edges |
| **G5** | Meaning/Therapy | Healing, values, emotional processing, personal transformation |
| **G6** | Network/Coordination | Community building, protocols, collaboration, collective action |
## Usage
### Installation
```bash
pip install founder-game-classifier
```
### Basic Usage
```python
from founder_game_classifier import GameClassifier
# Load the model (downloads from Hub on first use)
classifier = GameClassifier.from_pretrained("leoguinan/founder-game-classifier")
# Classify a single text
result = classifier.predict("Here's a tactic you can steal for your next launch...")
print(result["primary_game"]) # "G2"
print(result["confidence"]) # 0.72
print(result["probabilities"]) # {"G1": 0.05, "G2": 0.72, "G3": 0.10, ...}
```
### Batch Classification
```python
texts = [
"Here's the mental model I use for thinking about systems...",
"Join our community of builders who are changing the world...",
"I tried 47 different tactics. Here's what actually worked...",
]
results = classifier.predict_batch(texts)
for text, result in zip(texts, results):
print(f"{result['primary_game']}: {text[:50]}...")
```
### Get Aggregate Signature
Useful for analyzing a corpus of content:
```python
texts = load_my_blog_posts() # List of strings
signature = classifier.get_game_signature(texts)
print(signature)
# {'G1': 0.05, 'G2': 0.42, 'G3': 0.18, 'G4': 0.20, 'G5': 0.08, 'G6': 0.07}
```
## Model Architecture
- **Embedding Model**: `all-MiniLM-L6-v2` (384 dimensions)
- **Classifier**: Logistic Regression (sklearn)
- **Manifold System**: Mahalanobis distance to game centroids (optional)
## Training Data
The model was trained on labeled founder content spanning:
- Podcast transcripts
- Blog posts
- Twitter threads
- Newsletter content
Training used a multi-stage pipeline:
1. Text chunking and span extraction
2. LLM-assisted labeling with human verification
3. Embedding generation
4. Classifier training with cross-validation
## Performance
Validated on held-out test set:
| Metric | Score |
|--------|-------|
| Accuracy | 0.78 |
| Macro F1 | 0.74 |
| Top-2 Accuracy | 0.91 |
The model performs best on clear examples of each game and may show lower confidence on boundary cases or mixed content.
## Limitations
- Trained primarily on English content from tech/startup domain
- May not generalize well to non-business contexts
- Short texts (<50 words) may have lower accuracy
- Cultural and domain biases from training data
## Citation
```bibtex
@misc{guinan2024foundergameclassifier,
title={Founder Game Classifier},
author={Leo Guinan},
year={2024},
publisher={Hugging Face},
url={https://huggingface.co/leoguinan/founder-game-classifier}
}
```
## License
MIT License - free for commercial and non-commercial use.
## Files
- `classifier.pkl` - Trained LogisticRegression model (19KB)
- `label_encoder.pkl` - Label encoder for game classes (375B)
- `metadata.json` - Model metadata and configuration (143B)
- `game_manifolds.json` - Manifold centroids and covariances for geometric analysis (29MB)
|