Spaces:
Running
Running
Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
|
@@ -1,10 +1,40 @@
|
|
| 1 |
---
|
| 2 |
-
title:
|
| 3 |
-
emoji:
|
| 4 |
-
colorFrom: indigo
|
| 5 |
-
colorTo: purple
|
| 6 |
-
sdk: static
|
| 7 |
-
pinned: false
|
| 8 |
---
|
| 9 |
|
| 10 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
+
title: Lightning Rod Labs
|
| 3 |
+
emoji: "\u26A1"
|
|
|
|
|
|
|
|
|
|
|
|
|
| 4 |
---
|
| 5 |
|
| 6 |
+
# Lightning Rod Labs
|
| 7 |
+
|
| 8 |
+
**Train with Timestamps, Not Labels.**
|
| 9 |
+
|
| 10 |
+
Lightning Rod Labs automatically generates high-quality training data from your documents or public sources — no labeling or extraction required. Define your criteria in Python, and our SDK treats real-world outcomes as the label, producing high-signal supervision at scale. Models learn causal factors, not just tokens. Raw data to deployable specialized models in hours.
|
| 11 |
+
|
| 12 |
+
[Website](https://lightningrod.ai/) · [SDK](https://github.com/lightning-rod-labs/lightningrod-python-sdk) · [Blog](https://blog.lightningrod.ai/)
|
| 13 |
+
|
| 14 |
+
---
|
| 15 |
+
|
| 16 |
+
## How It Works
|
| 17 |
+
|
| 18 |
+
We generate grounded, model-ready training data from documents or public sources (Google News, SEC filings, market data). You define your criteria in Python, and our SDK uses the **future as the label** — turning messy, timestamped history into training signal automatically. No labeling pipelines, no extraction, no human annotation.
|
| 19 |
+
|
| 20 |
+
This approach has been used to beat frontier AIs 100x larger on prediction-market benchmarks, and has demonstrated success in financial forecasting, risk estimation, and policy prediction.
|
| 21 |
+
|
| 22 |
+
---
|
| 23 |
+
|
| 24 |
+
## Research & Results
|
| 25 |
+
|
| 26 |
+
- **[SEC Risk Prediction](https://arxiv.org/abs/2601.19189)**: Foresight learning on raw SEC filings trains a 32B model to outperform GPT-5 at predicting public company risks.
|
| 27 |
+
- **[Future-as-Label](https://arxiv.org/abs/2601.06336)**: AI learns directly from raw chronological news data at unlimited scale, no human annotation.
|
| 28 |
+
- **[Outcome-based RL](https://arxiv.org/abs/2505.17989)** (TMLR): Using RL to improve LLM forecasting ability from real-world outcomes.
|
| 29 |
+
- **[Foresight-32B vs. Frontier LLMs](https://blog.lightningrod.ai/p/foresight-32b-beats-frontier-llms-on-live-polymarket-predictions)**: Live demonstration beating frontier models on Polymarket predictions.
|
| 30 |
+
|
| 31 |
+
Foresight-32B is consistently top-ranked on [ForecastBench](https://www.forecastbench.org/tournament/) and [ProphetArena Sports](https://www.prophetarena.co/leaderboard).
|
| 32 |
+
|
| 33 |
+
---
|
| 34 |
+
|
| 35 |
+
## Models & Datasets
|
| 36 |
+
|
| 37 |
+
| Resource | Description |
|
| 38 |
+
|----------|-------------|
|
| 39 |
+
| [Trump-Forecaster](https://huggingface.co/LightningRodLabs/Trump-Forecaster) | RL-tuned gpt-oss-120b LoRA adapter for predicting Trump administration actions. Beats GPT-5 (Brier 0.194 vs 0.200). |
|
| 40 |
+
| [WWTD-2025](https://huggingface.co/datasets/LightningRodLabs/WWTD-2025) | 2,790 binary forecasting questions about U.S. policy under the Trump administration, with news context and ground-truth resolutions. |
|