--- language: - en license: mit library_name: founder-game-classifier tags: - text-classification - founders - content-analysis - sentence-transformers datasets: - custom metrics: - accuracy pipeline_tag: text-classification --- # Founder Game Classifier A trained classifier that identifies which of **6 founder games** a piece of content belongs to. ## Model Description This model classifies text content into one of six "founder games" - patterns of communication and content creation common among founders, creators, and thought leaders. ### The 6 Games | Game | Name | Description | |------|------|-------------| | **G1** | Identity/Canon | Recruiting into identity, lineage, belonging, status, canon formation | | **G2** | Ideas/Play Mining | Extracting reusable plays, tactics, heuristics; "do this / steal this" | | **G3** | Models/Understanding | Building mental models, frameworks, mechanisms, explanations | | **G4** | Performance/Competition | Winning, dominance, execution, metrics, endurance, zero-sum edges | | **G5** | Meaning/Therapy | Healing, values, emotional processing, personal transformation | | **G6** | Network/Coordination | Community building, protocols, collaboration, collective action | ## Usage ### Installation ```bash pip install founder-game-classifier ``` ### Basic Usage ```python from founder_game_classifier import GameClassifier # Load the model (downloads from Hub on first use) classifier = GameClassifier.from_pretrained("leoguinan/founder-game-classifier") # Classify a single text result = classifier.predict("Here's a tactic you can steal for your next launch...") print(result["primary_game"]) # "G2" print(result["confidence"]) # 0.72 print(result["probabilities"]) # {"G1": 0.05, "G2": 0.72, "G3": 0.10, ...} ``` ### Batch Classification ```python texts = [ "Here's the mental model I use for thinking about systems...", "Join our community of builders who are changing the world...", "I tried 47 different tactics. Here's what actually worked...", ] results = classifier.predict_batch(texts) for text, result in zip(texts, results): print(f"{result['primary_game']}: {text[:50]}...") ``` ### Get Aggregate Signature Useful for analyzing a corpus of content: ```python texts = load_my_blog_posts() # List of strings signature = classifier.get_game_signature(texts) print(signature) # {'G1': 0.05, 'G2': 0.42, 'G3': 0.18, 'G4': 0.20, 'G5': 0.08, 'G6': 0.07} ``` ## Model Architecture - **Embedding Model**: `all-MiniLM-L6-v2` (384 dimensions) - **Classifier**: Logistic Regression (sklearn) - **Manifold System**: Mahalanobis distance to game centroids (optional) ## Training Data The model was trained on labeled founder content spanning: - Podcast transcripts - Blog posts - Twitter threads - Newsletter content Training used a multi-stage pipeline: 1. Text chunking and span extraction 2. LLM-assisted labeling with human verification 3. Embedding generation 4. Classifier training with cross-validation ## Performance Validated on held-out test set: | Metric | Score | |--------|-------| | Accuracy | 0.78 | | Macro F1 | 0.74 | | Top-2 Accuracy | 0.91 | The model performs best on clear examples of each game and may show lower confidence on boundary cases or mixed content. ## Limitations - Trained primarily on English content from tech/startup domain - May not generalize well to non-business contexts - Short texts (<50 words) may have lower accuracy - Cultural and domain biases from training data ## Citation ```bibtex @misc{guinan2024foundergameclassifier, title={Founder Game Classifier}, author={Leo Guinan}, year={2024}, publisher={Hugging Face}, url={https://huggingface.co/leoguinan/founder-game-classifier} } ``` ## License MIT License - free for commercial and non-commercial use. ## Files - `classifier.pkl` - Trained LogisticRegression model (19KB) - `label_encoder.pkl` - Label encoder for game classes (375B) - `metadata.json` - Model metadata and configuration (143B) - `game_manifolds.json` - Manifold centroids and covariances for geometric analysis (29MB)