leoguinan's picture
Upload model card
680a754 verified
---
language:
- en
license: mit
library_name: founder-game-classifier
tags:
- text-classification
- founders
- content-analysis
- sentence-transformers
datasets:
- custom
metrics:
- accuracy
pipeline_tag: text-classification
---
# Founder Game Classifier
A trained classifier that identifies which of **6 founder games** a piece of content belongs to.
## Model Description
This model classifies text content into one of six "founder games" - patterns of communication and content creation common among founders, creators, and thought leaders.
### The 6 Games
| Game | Name | Description |
|------|------|-------------|
| **G1** | Identity/Canon | Recruiting into identity, lineage, belonging, status, canon formation |
| **G2** | Ideas/Play Mining | Extracting reusable plays, tactics, heuristics; "do this / steal this" |
| **G3** | Models/Understanding | Building mental models, frameworks, mechanisms, explanations |
| **G4** | Performance/Competition | Winning, dominance, execution, metrics, endurance, zero-sum edges |
| **G5** | Meaning/Therapy | Healing, values, emotional processing, personal transformation |
| **G6** | Network/Coordination | Community building, protocols, collaboration, collective action |
## Usage
### Installation
```bash
pip install founder-game-classifier
```
### Basic Usage
```python
from founder_game_classifier import GameClassifier
# Load the model (downloads from Hub on first use)
classifier = GameClassifier.from_pretrained("leoguinan/founder-game-classifier")
# Classify a single text
result = classifier.predict("Here's a tactic you can steal for your next launch...")
print(result["primary_game"]) # "G2"
print(result["confidence"]) # 0.72
print(result["probabilities"]) # {"G1": 0.05, "G2": 0.72, "G3": 0.10, ...}
```
### Batch Classification
```python
texts = [
"Here's the mental model I use for thinking about systems...",
"Join our community of builders who are changing the world...",
"I tried 47 different tactics. Here's what actually worked...",
]
results = classifier.predict_batch(texts)
for text, result in zip(texts, results):
print(f"{result['primary_game']}: {text[:50]}...")
```
### Get Aggregate Signature
Useful for analyzing a corpus of content:
```python
texts = load_my_blog_posts() # List of strings
signature = classifier.get_game_signature(texts)
print(signature)
# {'G1': 0.05, 'G2': 0.42, 'G3': 0.18, 'G4': 0.20, 'G5': 0.08, 'G6': 0.07}
```
## Model Architecture
- **Embedding Model**: `all-MiniLM-L6-v2` (384 dimensions)
- **Classifier**: Logistic Regression (sklearn)
- **Manifold System**: Mahalanobis distance to game centroids (optional)
## Training Data
The model was trained on labeled founder content spanning:
- Podcast transcripts
- Blog posts
- Twitter threads
- Newsletter content
Training used a multi-stage pipeline:
1. Text chunking and span extraction
2. LLM-assisted labeling with human verification
3. Embedding generation
4. Classifier training with cross-validation
## Performance
Validated on held-out test set:
| Metric | Score |
|--------|-------|
| Accuracy | 0.78 |
| Macro F1 | 0.74 |
| Top-2 Accuracy | 0.91 |
The model performs best on clear examples of each game and may show lower confidence on boundary cases or mixed content.
## Limitations
- Trained primarily on English content from tech/startup domain
- May not generalize well to non-business contexts
- Short texts (<50 words) may have lower accuracy
- Cultural and domain biases from training data
## Citation
```bibtex
@misc{guinan2024foundergameclassifier,
title={Founder Game Classifier},
author={Leo Guinan},
year={2024},
publisher={Hugging Face},
url={https://huggingface.co/leoguinan/founder-game-classifier}
}
```
## License
MIT License - free for commercial and non-commercial use.
## Files
- `classifier.pkl` - Trained LogisticRegression model (19KB)
- `label_encoder.pkl` - Label encoder for game classes (375B)
- `metadata.json` - Model metadata and configuration (143B)
- `game_manifolds.json` - Manifold centroids and covariances for geometric analysis (29MB)