CoT Oracle Paper Ablations And Baselines
Collection
All models used for my LessWrong post • 8 items • Updated
This repo contains the 1-layer paper ablation for the CoT Oracle recipe: on-policy lens tasks, chunked ConvQA, FineWeb lens readouts, and classification, without LatentQA.
Qwen/Qwen3-8B[18]shuffled4250M input tokens22.5M logged training tokensfuturelens: enabled, n: 30000pastlens: enabled, n: 30000chunked_convqa: enabled, n: -1 (all available examples)classification: enabled, n: 20000, datasets = sst2, ag_news, snlifineweb: enabled, n: 60000, variants = futurelens_fineweb,pastlens_fineweblatentqa: disabledconfigs/train.yaml: disabled50M input-token budget in the YAML.