Create README.md
Browse files
README.md
ADDED
|
@@ -0,0 +1,99 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# YCbench
|
| 2 |
+
|
| 3 |
+
**A dataset for benchmarking models that predict Y Combinator (YC) startup performance.**
|
| 4 |
+
|
| 5 |
+
This dataset collects structured data related to Y Combinator startups, primarily focused on **W26** (Winter 2026) batch and related signals. It supports research and development of forecasting models that aim to identify which early-stage startups will **outperform their batch peers** in the short term (e.g., next 90 days), as explored in the [YC Bench](https://ycbench.com/) live benchmark.
|
| 6 |
+
|
| 7 |
+
## Dataset Overview
|
| 8 |
+
|
| 9 |
+
YCbench aggregates public and derived signals for YC companies, including:
|
| 10 |
+
- Google/web mentions (as a proxy for visibility/traction)
|
| 11 |
+
- Pre-demo day scores and predictions
|
| 12 |
+
- Startup profiles
|
| 13 |
+
- Traction metrics
|
| 14 |
+
|
| 15 |
+
The data is useful for:
|
| 16 |
+
- Training/evaluating ML models for startup success prediction
|
| 17 |
+
- Cohort-relative benchmarking (comparing startups within the same YC batch)
|
| 18 |
+
- Analyzing early signals of velocity and outperformance
|
| 19 |
+
- Research on founder signals, traction, and short-term startup dynamics
|
| 20 |
+
|
| 21 |
+
## Files
|
| 22 |
+
|
| 23 |
+
The dataset repository contains the following CSV files:
|
| 24 |
+
|
| 25 |
+
- `yc_mentions.csv` — Google/web mention counts for YC-related domains
|
| 26 |
+
- `yc_mentions_early.csv` — Early-stage mention data
|
| 27 |
+
- `yc_w26_startups.csv` — List of W26 startups with basic information
|
| 28 |
+
- `yc_w26_traction.csv` — Traction metrics for W26 companies
|
| 29 |
+
- `yc_w26_pre_demo_scores.csv` — Pre-demo day scores, velocity scores, hybrid scores, etc.
|
| 30 |
+
|
| 31 |
+
**Note:** These files have varying schemas (different columns). When loading with the `datasets` library, you may need to load them individually or as separate configurations rather than as a single default split.
|
| 32 |
+
|
| 33 |
+
## Loading the Dataset
|
| 34 |
+
|
| 35 |
+
You can load individual files using pandas or the Hugging Face `datasets` library with manual configuration:
|
| 36 |
+
|
| 37 |
+
```python
|
| 38 |
+
import pandas as pd
|
| 39 |
+
from datasets import load_dataset
|
| 40 |
+
|
| 41 |
+
# Load with pandas (recommended for now due to schema differences)
|
| 42 |
+
df_mentions = pd.read_csv("hf://datasets/benstaf/ycbench/yc_mentions.csv")
|
| 43 |
+
df_w26 = pd.read_csv("hf://datasets/benstaf/ycbench/yc_w26_startups.csv")
|
| 44 |
+
|
| 45 |
+
# Or load specific files via datasets (if you define configs)
|
| 46 |
+
# dataset = load_dataset("benstaf/ycbench", "w26_startups") # example config
|
| 47 |
+
```
|
| 48 |
+
|
| 49 |
+
A preview of the data shows columns such as:
|
| 50 |
+
- `domain` (string)
|
| 51 |
+
- `google_mentions` (float)
|
| 52 |
+
|
| 53 |
+
Other files include scores like `mentions_score`, `Velocity_Score`, `hybrid_score`, etc.
|
| 54 |
+
|
| 55 |
+
## Motivation & Related Work
|
| 56 |
+
|
| 57 |
+
This dataset is connected to **YC Bench** — a live benchmark that evaluates forecasting models on their ability to predict which YC startups will show the strongest execution velocity in the 90 days following the start of a batch.
|
| 58 |
+
|
| 59 |
+
Key advantages of using YC batches for benchmarking:
|
| 60 |
+
- Startups enter at roughly the same time and stage
|
| 61 |
+
- Similar funding environment within a batch
|
| 62 |
+
- Clean peer comparison (cohort-relative ranking)
|
| 63 |
+
|
| 64 |
+
Instead of waiting years for exits or unicorn status, models are scored on measurable short-term outperformance.
|
| 65 |
+
|
| 66 |
+
## Usage Ideas
|
| 67 |
+
|
| 68 |
+
- Build baseline models using mention counts and traction signals
|
| 69 |
+
- Experiment with hybrid scoring (mentions + velocity + other features)
|
| 70 |
+
- Create leaderboards for LLM-based or traditional ML predictors
|
| 71 |
+
- Analyze correlations between early signals and later performance
|
| 72 |
+
|
| 73 |
+
## Limitations
|
| 74 |
+
|
| 75 |
+
- Data is currently focused on W26 and general YC mentions
|
| 76 |
+
- Column schemas are not fully unified across files
|
| 77 |
+
- Some fields may contain null values
|
| 78 |
+
- No official splits (train/validation/test) are provided yet
|
| 79 |
+
|
| 80 |
+
## License
|
| 81 |
+
|
| 82 |
+
No license specified. Please check with the uploader (`benstaf`) for intended use. For research and non-commercial purposes, it is generally safe to assume open use, but verify before commercial applications.
|
| 83 |
+
|
| 84 |
+
## Author
|
| 85 |
+
|
| 86 |
+
- **Ben Staf** ([benstaf on Hugging Face](https://huggingface.co/benstaf))
|
| 87 |
+
|
| 88 |
+
## Contributing / Improvements
|
| 89 |
+
|
| 90 |
+
Feel free to open a discussion or pull request if you want to:
|
| 91 |
+
- Unify schemas into a single consistent dataset
|
| 92 |
+
- Add more batches (S25, W27, etc.)
|
| 93 |
+
- Include additional signals (LinkedIn activity, GitHub stars, funding news, etc.)
|
| 94 |
+
- Add proper dataset cards, splits, or evaluation scripts
|
| 95 |
+
|
| 96 |
+
## Related Links
|
| 97 |
+
|
| 98 |
+
- [YC Bench website](https://ycbench.com/) — Live leaderboard and benchmark
|
| 99 |
+
- [Ben Staf's other datasets](https://huggingface.co/benstaf/datasets)
|