Bturtel commited on
Commit
3201e44
·
verified ·
1 Parent(s): 22356b9

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +37 -7
README.md CHANGED
@@ -1,10 +1,40 @@
1
  ---
2
- title: README
3
- emoji: 📈
4
- colorFrom: indigo
5
- colorTo: purple
6
- sdk: static
7
- pinned: false
8
  ---
9
 
10
- Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ title: Lightning Rod Labs
3
+ emoji: "\u26A1"
 
 
 
 
4
  ---
5
 
6
+ # Lightning Rod Labs
7
+
8
+ **Train with Timestamps, Not Labels.**
9
+
10
+ Lightning Rod Labs automatically generates high-quality training data from your documents or public sources — no labeling or extraction required. Define your criteria in Python, and our SDK treats real-world outcomes as the label, producing high-signal supervision at scale. Models learn causal factors, not just tokens. Raw data to deployable specialized models in hours.
11
+
12
+ [Website](https://lightningrod.ai/) · [SDK](https://github.com/lightning-rod-labs/lightningrod-python-sdk) · [Blog](https://blog.lightningrod.ai/)
13
+
14
+ ---
15
+
16
+ ## How It Works
17
+
18
+ We generate grounded, model-ready training data from documents or public sources (Google News, SEC filings, market data). You define your criteria in Python, and our SDK uses the **future as the label** — turning messy, timestamped history into training signal automatically. No labeling pipelines, no extraction, no human annotation.
19
+
20
+ This approach has been used to beat frontier AIs 100x larger on prediction-market benchmarks, and has demonstrated success in financial forecasting, risk estimation, and policy prediction.
21
+
22
+ ---
23
+
24
+ ## Research & Results
25
+
26
+ - **[SEC Risk Prediction](https://arxiv.org/abs/2601.19189)**: Foresight learning on raw SEC filings trains a 32B model to outperform GPT-5 at predicting public company risks.
27
+ - **[Future-as-Label](https://arxiv.org/abs/2601.06336)**: AI learns directly from raw chronological news data at unlimited scale, no human annotation.
28
+ - **[Outcome-based RL](https://arxiv.org/abs/2505.17989)** (TMLR): Using RL to improve LLM forecasting ability from real-world outcomes.
29
+ - **[Foresight-32B vs. Frontier LLMs](https://blog.lightningrod.ai/p/foresight-32b-beats-frontier-llms-on-live-polymarket-predictions)**: Live demonstration beating frontier models on Polymarket predictions.
30
+
31
+ Foresight-32B is consistently top-ranked on [ForecastBench](https://www.forecastbench.org/tournament/) and [ProphetArena Sports](https://www.prophetarena.co/leaderboard).
32
+
33
+ ---
34
+
35
+ ## Models & Datasets
36
+
37
+ | Resource | Description |
38
+ |----------|-------------|
39
+ | [Trump-Forecaster](https://huggingface.co/LightningRodLabs/Trump-Forecaster) | RL-tuned gpt-oss-120b LoRA adapter for predicting Trump administration actions. Beats GPT-5 (Brier 0.194 vs 0.200). |
40
+ | [WWTD-2025](https://huggingface.co/datasets/LightningRodLabs/WWTD-2025) | 2,790 binary forecasting questions about U.S. policy under the Trump administration, with news context and ground-truth resolutions. |