File size: 4,113 Bytes
680a754
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
---
language:
- en
license: mit
library_name: founder-game-classifier
tags:
- text-classification
- founders
- content-analysis
- sentence-transformers
datasets:
- custom
metrics:
- accuracy
pipeline_tag: text-classification
---

# Founder Game Classifier

A trained classifier that identifies which of **6 founder games** a piece of content belongs to.

## Model Description

This model classifies text content into one of six "founder games" - patterns of communication and content creation common among founders, creators, and thought leaders.

### The 6 Games

| Game | Name | Description |
|------|------|-------------|
| **G1** | Identity/Canon | Recruiting into identity, lineage, belonging, status, canon formation |
| **G2** | Ideas/Play Mining | Extracting reusable plays, tactics, heuristics; "do this / steal this" |
| **G3** | Models/Understanding | Building mental models, frameworks, mechanisms, explanations |
| **G4** | Performance/Competition | Winning, dominance, execution, metrics, endurance, zero-sum edges |
| **G5** | Meaning/Therapy | Healing, values, emotional processing, personal transformation |
| **G6** | Network/Coordination | Community building, protocols, collaboration, collective action |

## Usage

### Installation

```bash
pip install founder-game-classifier
```

### Basic Usage

```python
from founder_game_classifier import GameClassifier

# Load the model (downloads from Hub on first use)
classifier = GameClassifier.from_pretrained("leoguinan/founder-game-classifier")

# Classify a single text
result = classifier.predict("Here's a tactic you can steal for your next launch...")

print(result["primary_game"])      # "G2"
print(result["confidence"])        # 0.72
print(result["probabilities"])     # {"G1": 0.05, "G2": 0.72, "G3": 0.10, ...}
```

### Batch Classification

```python
texts = [
    "Here's the mental model I use for thinking about systems...",
    "Join our community of builders who are changing the world...",
    "I tried 47 different tactics. Here's what actually worked...",
]

results = classifier.predict_batch(texts)

for text, result in zip(texts, results):
    print(f"{result['primary_game']}: {text[:50]}...")
```

### Get Aggregate Signature

Useful for analyzing a corpus of content:

```python
texts = load_my_blog_posts()  # List of strings
signature = classifier.get_game_signature(texts)

print(signature)
# {'G1': 0.05, 'G2': 0.42, 'G3': 0.18, 'G4': 0.20, 'G5': 0.08, 'G6': 0.07}
```

## Model Architecture

- **Embedding Model**: `all-MiniLM-L6-v2` (384 dimensions)
- **Classifier**: Logistic Regression (sklearn)
- **Manifold System**: Mahalanobis distance to game centroids (optional)

## Training Data

The model was trained on labeled founder content spanning:
- Podcast transcripts
- Blog posts
- Twitter threads
- Newsletter content

Training used a multi-stage pipeline:
1. Text chunking and span extraction
2. LLM-assisted labeling with human verification
3. Embedding generation
4. Classifier training with cross-validation

## Performance

Validated on held-out test set:

| Metric | Score |
|--------|-------|
| Accuracy | 0.78 |
| Macro F1 | 0.74 |
| Top-2 Accuracy | 0.91 |

The model performs best on clear examples of each game and may show lower confidence on boundary cases or mixed content.

## Limitations

- Trained primarily on English content from tech/startup domain
- May not generalize well to non-business contexts
- Short texts (<50 words) may have lower accuracy
- Cultural and domain biases from training data

## Citation

```bibtex
@misc{guinan2024foundergameclassifier,
  title={Founder Game Classifier},
  author={Leo Guinan},
  year={2024},
  publisher={Hugging Face},
  url={https://huggingface.co/leoguinan/founder-game-classifier}
}
```

## License

MIT License - free for commercial and non-commercial use.

## Files

- `classifier.pkl` - Trained LogisticRegression model (19KB)
- `label_encoder.pkl` - Label encoder for game classes (375B)
- `metadata.json` - Model metadata and configuration (143B)
- `game_manifolds.json` - Manifold centroids and covariances for geometric analysis (29MB)