metadata
base_model: Qwen/Qwen3-8B
library_name: peft
pipeline_tag: text-generation
tags:
- base_model:adapter:Qwen/Qwen3-8B
- lora
- transformers
- cot-oracle
- activation-oracle
- paper-ablation
- ours
- 1-layer
- 22.5m-train-tokens
CoT Oracle Paper Ablation: Ours, 1 Layer
This repo contains the 1-layer paper ablation for the CoT Oracle recipe: on-policy lens tasks, chunked ConvQA, FineWeb lens readouts, and classification, without LatentQA.
What This Checkpoint Is
- Base model:
Qwen/Qwen3-8B - Adapter format: PEFT LoRA
- Activation readout layers:
[18] - Task order:
shuffled - Seed:
42 - Planned budget:
50Minput tokens - Paper label:
22.5Mlogged training tokens
Exact Training Mixture
- On-policy
futurelens: enabled,n: 30000 - On-policy
pastlens: enabled,n: 30000 chunked_convqa: enabled,n: -1(all available examples)classification: enabled,n: 20000, datasets =sst2,ag_news,snlifineweb: enabled,n: 60000, variants =futurelens_fineweb,pastlens_fineweblatentqa: disabled- All other tasks in
configs/train.yaml: disabled
Notes
- This is the 1-layer counterpart to the 3-layer paper ablations.
- The token label follows the paper bookkeeping from the run logs rather than the planned
50Minput-token budget in the YAML.