|
|
--- |
|
|
license: mit |
|
|
configs: |
|
|
- config_name: default |
|
|
data_files: |
|
|
- split: ratio1_v1 |
|
|
path: ratio1_v1.csv |
|
|
- split: ratio1_v2 |
|
|
path: ratio1_v2.csv |
|
|
- split: ratio2_v1 |
|
|
path: ratio2_v2.csv |
|
|
- split: ratio3_v1 |
|
|
path: ratio3_v2.csv |
|
|
- split: ratio6_v2 |
|
|
path: ratio6_v2.csv |
|
|
- split: ratio10_v1 |
|
|
path: ratio10_v1.csv |
|
|
- split: ratio30_v1 |
|
|
path: ratio30_v1.csv |
|
|
- split: ratio50_v1 |
|
|
path: ratio50_v1.csv |
|
|
- split: test |
|
|
path: test.csv |
|
|
--- |
|
|
|
|
|
This dataset is composed of Claude-labelled [fineweb](https://huggingface.co/datasets/HuggingFaceFW/fineweb/) documents. |
|
|
|
|
|
For each document, Claude is asked if it is 'forecastable' (i.e. would be a reasonable seed for a pastcasting question) and to estimate the date the document was published. |
|
|
V1 splits were generated by having Claude label ~50K random fineweb documents and v2 splits were augmented with labels on ~30K additional documents that a DebertaV3 classifier finetuned on ratio10_v1 thought were forecastable (Claude thought ~1/3 of these additional documents were forecastable). |
|
|
|
|
|
Splits are of the form `ratio{negative_to_positive_ratio}_v{1 or 2}`. For example `ratio6_v2` is ~6 negative examples for each positive example. Splits do overlap. |
|
|
|
|
|
`ratio6_v2` was used to train https://huggingface.co/noanabeshima/forecastability-classifier-v1. |
|
|
|
|
|
Prompt can be found in `prompt.txt`. It was iterated on slightly using a small set of ground-truth human labels. GPT-4.1 performed slightly better, but Claude had a more favorable TOS for open-sourcing data/models. |
|
|
|
|
|
Claude considers ~2% of fineweb documents to be forecastable. |
|
|
|
|
|
Made with the help of Collin Gray. |