jkminder commited on
Commit
668cb22
·
verified ·
1 Parent(s): 67d244c

Training in progress, step 54

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +1 -0
  2. 2025-10-10/23-16-13/.hydra/config.yaml +45 -0
  3. 2025-10-10/23-16-13/.hydra/hydra.yaml +168 -0
  4. 2025-10-10/23-16-13/.hydra/overrides.yaml +7 -0
  5. 2025-10-10/23-16-13/train.log +0 -0
  6. 2025-10-10/23-16-52/.hydra/config.yaml +45 -0
  7. 2025-10-10/23-16-52/.hydra/hydra.yaml +168 -0
  8. 2025-10-10/23-16-52/.hydra/overrides.yaml +7 -0
  9. 2025-10-10/23-16-52/train.log +0 -0
  10. 2025-10-10/23-18-13/.hydra/config.yaml +45 -0
  11. 2025-10-10/23-18-13/.hydra/hydra.yaml +167 -0
  12. 2025-10-10/23-18-13/.hydra/overrides.yaml +6 -0
  13. 2025-10-10/23-18-13/train.log +0 -0
  14. 2025-10-10/23-21-52/.hydra/config.yaml +45 -0
  15. 2025-10-10/23-21-52/.hydra/hydra.yaml +167 -0
  16. 2025-10-10/23-21-52/.hydra/overrides.yaml +6 -0
  17. 2025-10-10/23-21-52/train.log +0 -0
  18. 2025-10-10/23-23-18/.hydra/config.yaml +45 -0
  19. 2025-10-10/23-23-18/.hydra/hydra.yaml +168 -0
  20. 2025-10-10/23-23-18/.hydra/overrides.yaml +7 -0
  21. 2025-10-10/23-23-18/train.log +0 -0
  22. 2025-10-10/23-23-47/.hydra/config.yaml +45 -0
  23. 2025-10-10/23-23-47/.hydra/hydra.yaml +167 -0
  24. 2025-10-10/23-23-47/.hydra/overrides.yaml +6 -0
  25. 2025-10-10/23-23-47/train.log +0 -0
  26. 2025-10-10/23-28-17/.hydra/config.yaml +45 -0
  27. 2025-10-10/23-28-17/.hydra/hydra.yaml +167 -0
  28. 2025-10-10/23-28-17/.hydra/overrides.yaml +6 -0
  29. 2025-10-10/23-28-17/train.log +0 -0
  30. 2025-10-10/23-33-34/.hydra/config.yaml +45 -0
  31. 2025-10-10/23-33-34/.hydra/hydra.yaml +167 -0
  32. 2025-10-10/23-33-34/.hydra/overrides.yaml +6 -0
  33. 2025-10-10/23-33-34/train.log +0 -0
  34. 2025-10-10/23-35-10/.hydra/config.yaml +45 -0
  35. 2025-10-10/23-35-10/.hydra/hydra.yaml +167 -0
  36. 2025-10-10/23-35-10/.hydra/overrides.yaml +6 -0
  37. 2025-10-10/23-35-10/train.log +0 -0
  38. 2025-10-10/23-36-21/.hydra/config.yaml +45 -0
  39. 2025-10-10/23-36-21/.hydra/hydra.yaml +167 -0
  40. 2025-10-10/23-36-21/.hydra/overrides.yaml +6 -0
  41. 2025-10-10/23-36-21/train.log +0 -0
  42. 2025-10-10/23-40-03/.hydra/config.yaml +45 -0
  43. 2025-10-10/23-40-03/.hydra/hydra.yaml +167 -0
  44. 2025-10-10/23-40-03/.hydra/overrides.yaml +6 -0
  45. 2025-10-10/23-40-03/train.log +0 -0
  46. 2025-10-10/23-44-29/.hydra/config.yaml +45 -0
  47. 2025-10-10/23-44-29/.hydra/hydra.yaml +167 -0
  48. 2025-10-10/23-44-29/.hydra/overrides.yaml +6 -0
  49. 2025-10-10/23-44-29/train.log +0 -0
  50. 2025-10-10/23-47-52/.hydra/config.yaml +45 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
2025-10-10/23-16-13/.hydra/config.yaml ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ pretrained: gpt2
3
+ resize_embeddings: true
4
+ dataset:
5
+ name: science-of-finetuning/fineweb-1m-sample
6
+ config: ''
7
+ seq_len: 256
8
+ context_len: 224
9
+ num_train_samples: 100000
10
+ training:
11
+ output_dir: outputs
12
+ num_train_epochs: 1
13
+ per_device_train_batch_size: 4
14
+ gradient_accumulation_steps: 1
15
+ learning_rate: 5.0e-05
16
+ weight_decay: 0.0
17
+ warmup_steps: 0
18
+ save_steps: 500
19
+ logging_steps: 50
20
+ seed: 42
21
+ fp16: false
22
+ bf16: true
23
+ calibration_batch_size: 128
24
+ experiment:
25
+ stage: 1
26
+ p_intervene: 0.1
27
+ use_measurement: true
28
+ force_text: null
29
+ init_from:
30
+ hub_repo: null
31
+ local_ckpt: null
32
+ wandb:
33
+ project: lorentz-forcing
34
+ entity: jkminder
35
+ hfhub:
36
+ push_to_hub: true
37
+ repo_id: your-org/lorentz-poc-stage1
38
+ eval:
39
+ output_path: eval_results.json
40
+ max_eval_samples: 128
41
+ seed: 42
42
+ ci:
43
+ method: bootstrap
44
+ level: 0.95
45
+ bootstrap_iters: 1000
2025-10-10/23-16-13/.hydra/hydra.yaml ADDED
@@ -0,0 +1,168 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - experiment=stage1
116
+ - model.pretrained=gpt2
117
+ - hfhub.repo_id=your-org/lorentz-poc-stage1
118
+ - wandb.project=lorentz-forcing
119
+ - training.num_train_epochs=1
120
+ - dataset.seq_len=256
121
+ - dataset.context_len=224
122
+ job:
123
+ name: train
124
+ chdir: null
125
+ override_dirname: dataset.context_len=224,dataset.seq_len=256,experiment=stage1,hfhub.repo_id=your-org/lorentz-poc-stage1,model.pretrained=gpt2,training.num_train_epochs=1,wandb.project=lorentz-forcing
126
+ id: ???
127
+ num: ???
128
+ config_name: config
129
+ env_set: {}
130
+ env_copy: []
131
+ config:
132
+ override_dirname:
133
+ kv_sep: '='
134
+ item_sep: ','
135
+ exclude_keys: []
136
+ runtime:
137
+ version: 1.3.2
138
+ version_base: '1.3'
139
+ cwd: /mnt/nw/home/j.minder/repositories/lorentz_forcing
140
+ config_sources:
141
+ - path: hydra.conf
142
+ schema: pkg
143
+ provider: hydra
144
+ - path: /mnt/nw/home/j.minder/repositories/lorentz_forcing/conf
145
+ schema: file
146
+ provider: main
147
+ - path: ''
148
+ schema: structured
149
+ provider: schema
150
+ output_dir: /mnt/nw/home/j.minder/repositories/lorentz_forcing/outputs/2025-10-10/23-16-13
151
+ choices:
152
+ eval: default
153
+ hfhub: default
154
+ wandb: default
155
+ experiment: stage1
156
+ training: default
157
+ dataset: fineweb
158
+ model: llama32_1B
159
+ hydra/env: default
160
+ hydra/callbacks: null
161
+ hydra/job_logging: default
162
+ hydra/hydra_logging: default
163
+ hydra/hydra_help: default
164
+ hydra/help: default
165
+ hydra/sweeper: basic
166
+ hydra/launcher: basic
167
+ hydra/output: default
168
+ verbose: false
2025-10-10/23-16-13/.hydra/overrides.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ - experiment=stage1
2
+ - model.pretrained=gpt2
3
+ - hfhub.repo_id=your-org/lorentz-poc-stage1
4
+ - wandb.project=lorentz-forcing
5
+ - training.num_train_epochs=1
6
+ - dataset.seq_len=256
7
+ - dataset.context_len=224
2025-10-10/23-16-13/train.log ADDED
File without changes
2025-10-10/23-16-52/.hydra/config.yaml ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ pretrained: gpt2
3
+ resize_embeddings: true
4
+ dataset:
5
+ name: science-of-finetuning/fineweb-1m-sample
6
+ config: ''
7
+ seq_len: 256
8
+ context_len: 224
9
+ num_train_samples: 100000
10
+ training:
11
+ output_dir: outputs
12
+ num_train_epochs: 1
13
+ per_device_train_batch_size: 4
14
+ gradient_accumulation_steps: 1
15
+ learning_rate: 5.0e-05
16
+ weight_decay: 0.0
17
+ warmup_steps: 0
18
+ save_steps: 500
19
+ logging_steps: 50
20
+ seed: 42
21
+ fp16: false
22
+ bf16: true
23
+ calibration_batch_size: 128
24
+ experiment:
25
+ stage: 1
26
+ p_intervene: 0.1
27
+ use_measurement: true
28
+ force_text: null
29
+ init_from:
30
+ hub_repo: null
31
+ local_ckpt: null
32
+ wandb:
33
+ project: lorentz-forcing
34
+ entity: jkminder
35
+ hfhub:
36
+ push_to_hub: true
37
+ repo_id: your-org/lorentz-poc-stage1
38
+ eval:
39
+ output_path: eval_results.json
40
+ max_eval_samples: 128
41
+ seed: 42
42
+ ci:
43
+ method: bootstrap
44
+ level: 0.95
45
+ bootstrap_iters: 1000
2025-10-10/23-16-52/.hydra/hydra.yaml ADDED
@@ -0,0 +1,168 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - experiment=stage1
116
+ - model.pretrained=gpt2
117
+ - hfhub.repo_id=your-org/lorentz-poc-stage1
118
+ - wandb.project=lorentz-forcing
119
+ - training.num_train_epochs=1
120
+ - dataset.seq_len=256
121
+ - dataset.context_len=224
122
+ job:
123
+ name: train
124
+ chdir: null
125
+ override_dirname: dataset.context_len=224,dataset.seq_len=256,experiment=stage1,hfhub.repo_id=your-org/lorentz-poc-stage1,model.pretrained=gpt2,training.num_train_epochs=1,wandb.project=lorentz-forcing
126
+ id: ???
127
+ num: ???
128
+ config_name: config
129
+ env_set: {}
130
+ env_copy: []
131
+ config:
132
+ override_dirname:
133
+ kv_sep: '='
134
+ item_sep: ','
135
+ exclude_keys: []
136
+ runtime:
137
+ version: 1.3.2
138
+ version_base: '1.3'
139
+ cwd: /mnt/nw/home/j.minder/repositories/lorentz_forcing
140
+ config_sources:
141
+ - path: hydra.conf
142
+ schema: pkg
143
+ provider: hydra
144
+ - path: /mnt/nw/home/j.minder/repositories/lorentz_forcing/conf
145
+ schema: file
146
+ provider: main
147
+ - path: ''
148
+ schema: structured
149
+ provider: schema
150
+ output_dir: /mnt/nw/home/j.minder/repositories/lorentz_forcing/outputs/2025-10-10/23-16-52
151
+ choices:
152
+ eval: default
153
+ hfhub: default
154
+ wandb: default
155
+ experiment: stage1
156
+ training: default
157
+ dataset: fineweb
158
+ model: llama32_1B
159
+ hydra/env: default
160
+ hydra/callbacks: null
161
+ hydra/job_logging: default
162
+ hydra/hydra_logging: default
163
+ hydra/hydra_help: default
164
+ hydra/help: default
165
+ hydra/sweeper: basic
166
+ hydra/launcher: basic
167
+ hydra/output: default
168
+ verbose: false
2025-10-10/23-16-52/.hydra/overrides.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ - experiment=stage1
2
+ - model.pretrained=gpt2
3
+ - hfhub.repo_id=your-org/lorentz-poc-stage1
4
+ - wandb.project=lorentz-forcing
5
+ - training.num_train_epochs=1
6
+ - dataset.seq_len=256
7
+ - dataset.context_len=224
2025-10-10/23-16-52/train.log ADDED
File without changes
2025-10-10/23-18-13/.hydra/config.yaml ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ pretrained: meta-llama/Llama-3.2-1B
3
+ resize_embeddings: true
4
+ dataset:
5
+ name: science-of-finetuning/fineweb-1m-sample
6
+ config: ''
7
+ seq_len: 512
8
+ context_len: 512
9
+ num_train_samples: 100000
10
+ training:
11
+ output_dir: outputs
12
+ num_train_epochs: 1
13
+ per_device_train_batch_size: 4
14
+ gradient_accumulation_steps: 1
15
+ learning_rate: 5.0e-05
16
+ weight_decay: 0.0
17
+ warmup_steps: 0
18
+ save_steps: 500
19
+ logging_steps: 50
20
+ seed: 42
21
+ fp16: false
22
+ bf16: true
23
+ calibration_batch_size: 128
24
+ experiment:
25
+ stage: 1
26
+ p_intervene: 0.1
27
+ use_measurement: true
28
+ force_text: null
29
+ init_from:
30
+ hub_repo: null
31
+ local_ckpt: null
32
+ wandb:
33
+ project: lorentz-forcing
34
+ entity: jkminder
35
+ hfhub:
36
+ push_to_hub: true
37
+ repo_id: jkminder/lorentz-poc-stage1
38
+ eval:
39
+ output_path: eval_results.json
40
+ max_eval_samples: 128
41
+ seed: 42
42
+ ci:
43
+ method: bootstrap
44
+ level: 0.95
45
+ bootstrap_iters: 1000
2025-10-10/23-18-13/.hydra/hydra.yaml ADDED
@@ -0,0 +1,167 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - experiment=stage1
116
+ - hfhub.repo_id=jkminder/lorentz-poc-stage1
117
+ - wandb.project=lorentz-forcing
118
+ - training.num_train_epochs=1
119
+ - dataset.seq_len=512
120
+ - dataset.context_len=512
121
+ job:
122
+ name: train
123
+ chdir: null
124
+ override_dirname: dataset.context_len=512,dataset.seq_len=512,experiment=stage1,hfhub.repo_id=jkminder/lorentz-poc-stage1,training.num_train_epochs=1,wandb.project=lorentz-forcing
125
+ id: ???
126
+ num: ???
127
+ config_name: config
128
+ env_set: {}
129
+ env_copy: []
130
+ config:
131
+ override_dirname:
132
+ kv_sep: '='
133
+ item_sep: ','
134
+ exclude_keys: []
135
+ runtime:
136
+ version: 1.3.2
137
+ version_base: '1.3'
138
+ cwd: /mnt/nw/home/j.minder/repositories/lorentz_forcing
139
+ config_sources:
140
+ - path: hydra.conf
141
+ schema: pkg
142
+ provider: hydra
143
+ - path: /mnt/nw/home/j.minder/repositories/lorentz_forcing/conf
144
+ schema: file
145
+ provider: main
146
+ - path: ''
147
+ schema: structured
148
+ provider: schema
149
+ output_dir: /mnt/nw/home/j.minder/repositories/lorentz_forcing/outputs/2025-10-10/23-18-13
150
+ choices:
151
+ eval: default
152
+ hfhub: default
153
+ wandb: default
154
+ experiment: stage1
155
+ training: default
156
+ dataset: fineweb
157
+ model: llama32_1B
158
+ hydra/env: default
159
+ hydra/callbacks: null
160
+ hydra/job_logging: default
161
+ hydra/hydra_logging: default
162
+ hydra/hydra_help: default
163
+ hydra/help: default
164
+ hydra/sweeper: basic
165
+ hydra/launcher: basic
166
+ hydra/output: default
167
+ verbose: false
2025-10-10/23-18-13/.hydra/overrides.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ - experiment=stage1
2
+ - hfhub.repo_id=jkminder/lorentz-poc-stage1
3
+ - wandb.project=lorentz-forcing
4
+ - training.num_train_epochs=1
5
+ - dataset.seq_len=512
6
+ - dataset.context_len=512
2025-10-10/23-18-13/train.log ADDED
File without changes
2025-10-10/23-21-52/.hydra/config.yaml ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ pretrained: meta-llama/Llama-3.2-1B
3
+ resize_embeddings: true
4
+ dataset:
5
+ name: science-of-finetuning/fineweb-1m-sample
6
+ config: ''
7
+ seq_len: 512
8
+ context_len: 512
9
+ num_train_samples: 100000
10
+ training:
11
+ output_dir: outputs
12
+ num_train_epochs: 1
13
+ per_device_train_batch_size: 4
14
+ gradient_accumulation_steps: 1
15
+ learning_rate: 5.0e-05
16
+ weight_decay: 0.0
17
+ warmup_steps: 0
18
+ save_steps: 500
19
+ logging_steps: 50
20
+ seed: 42
21
+ fp16: false
22
+ bf16: true
23
+ calibration_batch_size: 128
24
+ experiment:
25
+ stage: 1
26
+ p_intervene: 0.1
27
+ use_measurement: true
28
+ force_text: null
29
+ init_from:
30
+ hub_repo: null
31
+ local_ckpt: null
32
+ wandb:
33
+ project: lorentz-forcing
34
+ entity: jkminder
35
+ hfhub:
36
+ push_to_hub: true
37
+ repo_id: jkminder/lorentz-poc-stage1
38
+ eval:
39
+ output_path: eval_results.json
40
+ max_eval_samples: 128
41
+ seed: 42
42
+ ci:
43
+ method: bootstrap
44
+ level: 0.95
45
+ bootstrap_iters: 1000
2025-10-10/23-21-52/.hydra/hydra.yaml ADDED
@@ -0,0 +1,167 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - experiment=stage1
116
+ - hfhub.repo_id=jkminder/lorentz-poc-stage1
117
+ - wandb.project=lorentz-forcing
118
+ - training.num_train_epochs=1
119
+ - dataset.seq_len=512
120
+ - dataset.context_len=512
121
+ job:
122
+ name: train
123
+ chdir: null
124
+ override_dirname: dataset.context_len=512,dataset.seq_len=512,experiment=stage1,hfhub.repo_id=jkminder/lorentz-poc-stage1,training.num_train_epochs=1,wandb.project=lorentz-forcing
125
+ id: ???
126
+ num: ???
127
+ config_name: config
128
+ env_set: {}
129
+ env_copy: []
130
+ config:
131
+ override_dirname:
132
+ kv_sep: '='
133
+ item_sep: ','
134
+ exclude_keys: []
135
+ runtime:
136
+ version: 1.3.2
137
+ version_base: '1.3'
138
+ cwd: /mnt/nw/home/j.minder/repositories/lorentz_forcing
139
+ config_sources:
140
+ - path: hydra.conf
141
+ schema: pkg
142
+ provider: hydra
143
+ - path: /mnt/nw/home/j.minder/repositories/lorentz_forcing/conf
144
+ schema: file
145
+ provider: main
146
+ - path: ''
147
+ schema: structured
148
+ provider: schema
149
+ output_dir: /mnt/nw/home/j.minder/repositories/lorentz_forcing/outputs/2025-10-10/23-21-52
150
+ choices:
151
+ eval: default
152
+ hfhub: default
153
+ wandb: default
154
+ experiment: stage1
155
+ training: default
156
+ dataset: fineweb
157
+ model: llama32_1B
158
+ hydra/env: default
159
+ hydra/callbacks: null
160
+ hydra/job_logging: default
161
+ hydra/hydra_logging: default
162
+ hydra/hydra_help: default
163
+ hydra/help: default
164
+ hydra/sweeper: basic
165
+ hydra/launcher: basic
166
+ hydra/output: default
167
+ verbose: false
2025-10-10/23-21-52/.hydra/overrides.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ - experiment=stage1
2
+ - hfhub.repo_id=jkminder/lorentz-poc-stage1
3
+ - wandb.project=lorentz-forcing
4
+ - training.num_train_epochs=1
5
+ - dataset.seq_len=512
6
+ - dataset.context_len=512
2025-10-10/23-21-52/train.log ADDED
File without changes
2025-10-10/23-23-18/.hydra/config.yaml ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ pretrained: jkminder/lorentz-poc-stage1
3
+ resize_embeddings: true
4
+ dataset:
5
+ name: science-of-finetuning/fineweb-1m-sample
6
+ config: ''
7
+ seq_len: 512
8
+ context_len: 512
9
+ num_train_samples: 1000
10
+ training:
11
+ output_dir: outputs
12
+ num_train_epochs: 1
13
+ per_device_train_batch_size: 4
14
+ gradient_accumulation_steps: 1
15
+ learning_rate: 5.0e-05
16
+ weight_decay: 0.0
17
+ warmup_steps: 0
18
+ save_steps: 500
19
+ logging_steps: 50
20
+ seed: 42
21
+ fp16: false
22
+ bf16: true
23
+ calibration_batch_size: 128
24
+ experiment:
25
+ stage: 2
26
+ p_intervene: 0.1
27
+ use_measurement: false
28
+ force_text: <intro>I am uncertain.</intro>
29
+ init_from:
30
+ hub_repo: null
31
+ local_ckpt: null
32
+ wandb:
33
+ project: lorentz-forcing
34
+ entity: jkminder
35
+ hfhub:
36
+ push_to_hub: true
37
+ repo_id: jkminder/lorentz-poc-stage2
38
+ eval:
39
+ output_path: eval_results.json
40
+ max_eval_samples: 128
41
+ seed: 42
42
+ ci:
43
+ method: bootstrap
44
+ level: 0.95
45
+ bootstrap_iters: 1000
2025-10-10/23-23-18/.hydra/hydra.yaml ADDED
@@ -0,0 +1,168 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - experiment=stage2
116
+ - model.pretrained=jkminder/lorentz-poc-stage1
117
+ - hfhub.repo_id=jkminder/lorentz-poc-stage2
118
+ - wandb.project=lorentz-forcing
119
+ - training.num_train_epochs=1
120
+ - dataset.seq_len=512
121
+ - dataset.context_len=512
122
+ job:
123
+ name: train
124
+ chdir: null
125
+ override_dirname: dataset.context_len=512,dataset.seq_len=512,experiment=stage2,hfhub.repo_id=jkminder/lorentz-poc-stage2,model.pretrained=jkminder/lorentz-poc-stage1,training.num_train_epochs=1,wandb.project=lorentz-forcing
126
+ id: ???
127
+ num: ???
128
+ config_name: config
129
+ env_set: {}
130
+ env_copy: []
131
+ config:
132
+ override_dirname:
133
+ kv_sep: '='
134
+ item_sep: ','
135
+ exclude_keys: []
136
+ runtime:
137
+ version: 1.3.2
138
+ version_base: '1.3'
139
+ cwd: /mnt/nw/home/j.minder/repositories/lorentz_forcing
140
+ config_sources:
141
+ - path: hydra.conf
142
+ schema: pkg
143
+ provider: hydra
144
+ - path: /mnt/nw/home/j.minder/repositories/lorentz_forcing/conf
145
+ schema: file
146
+ provider: main
147
+ - path: ''
148
+ schema: structured
149
+ provider: schema
150
+ output_dir: /mnt/nw/home/j.minder/repositories/lorentz_forcing/outputs/2025-10-10/23-23-18
151
+ choices:
152
+ eval: default
153
+ hfhub: default
154
+ wandb: default
155
+ experiment: stage2
156
+ training: default
157
+ dataset: fineweb
158
+ model: llama32_1B
159
+ hydra/env: default
160
+ hydra/callbacks: null
161
+ hydra/job_logging: default
162
+ hydra/hydra_logging: default
163
+ hydra/hydra_help: default
164
+ hydra/help: default
165
+ hydra/sweeper: basic
166
+ hydra/launcher: basic
167
+ hydra/output: default
168
+ verbose: false
2025-10-10/23-23-18/.hydra/overrides.yaml ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ - experiment=stage2
2
+ - model.pretrained=jkminder/lorentz-poc-stage1
3
+ - hfhub.repo_id=jkminder/lorentz-poc-stage2
4
+ - wandb.project=lorentz-forcing
5
+ - training.num_train_epochs=1
6
+ - dataset.seq_len=512
7
+ - dataset.context_len=512
2025-10-10/23-23-18/train.log ADDED
File without changes
2025-10-10/23-23-47/.hydra/config.yaml ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ pretrained: meta-llama/Llama-3.2-1B
3
+ resize_embeddings: true
4
+ dataset:
5
+ name: science-of-finetuning/fineweb-1m-sample
6
+ config: ''
7
+ seq_len: 512
8
+ context_len: 512
9
+ num_train_samples: 1000
10
+ training:
11
+ output_dir: outputs
12
+ num_train_epochs: 1
13
+ per_device_train_batch_size: 4
14
+ gradient_accumulation_steps: 1
15
+ learning_rate: 5.0e-05
16
+ weight_decay: 0.0
17
+ warmup_steps: 0
18
+ save_steps: 500
19
+ logging_steps: 50
20
+ seed: 42
21
+ fp16: false
22
+ bf16: true
23
+ calibration_batch_size: 128
24
+ experiment:
25
+ stage: 1
26
+ p_intervene: 0.1
27
+ use_measurement: true
28
+ force_text: null
29
+ init_from:
30
+ hub_repo: null
31
+ local_ckpt: null
32
+ wandb:
33
+ project: lorentz-forcing
34
+ entity: jkminder
35
+ hfhub:
36
+ push_to_hub: true
37
+ repo_id: jkminder/lorentz-poc-stage1
38
+ eval:
39
+ output_path: eval_results.json
40
+ max_eval_samples: 128
41
+ seed: 42
42
+ ci:
43
+ method: bootstrap
44
+ level: 0.95
45
+ bootstrap_iters: 1000
2025-10-10/23-23-47/.hydra/hydra.yaml ADDED
@@ -0,0 +1,167 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - experiment=stage1
116
+ - hfhub.repo_id=jkminder/lorentz-poc-stage1
117
+ - wandb.project=lorentz-forcing
118
+ - training.num_train_epochs=1
119
+ - dataset.seq_len=512
120
+ - dataset.context_len=512
121
+ job:
122
+ name: train
123
+ chdir: null
124
+ override_dirname: dataset.context_len=512,dataset.seq_len=512,experiment=stage1,hfhub.repo_id=jkminder/lorentz-poc-stage1,training.num_train_epochs=1,wandb.project=lorentz-forcing
125
+ id: ???
126
+ num: ???
127
+ config_name: config
128
+ env_set: {}
129
+ env_copy: []
130
+ config:
131
+ override_dirname:
132
+ kv_sep: '='
133
+ item_sep: ','
134
+ exclude_keys: []
135
+ runtime:
136
+ version: 1.3.2
137
+ version_base: '1.3'
138
+ cwd: /mnt/nw/home/j.minder/repositories/lorentz_forcing
139
+ config_sources:
140
+ - path: hydra.conf
141
+ schema: pkg
142
+ provider: hydra
143
+ - path: /mnt/nw/home/j.minder/repositories/lorentz_forcing/conf
144
+ schema: file
145
+ provider: main
146
+ - path: ''
147
+ schema: structured
148
+ provider: schema
149
+ output_dir: /mnt/nw/home/j.minder/repositories/lorentz_forcing/outputs/2025-10-10/23-23-47
150
+ choices:
151
+ eval: default
152
+ hfhub: default
153
+ wandb: default
154
+ experiment: stage1
155
+ training: default
156
+ dataset: fineweb
157
+ model: llama32_1B
158
+ hydra/env: default
159
+ hydra/callbacks: null
160
+ hydra/job_logging: default
161
+ hydra/hydra_logging: default
162
+ hydra/hydra_help: default
163
+ hydra/help: default
164
+ hydra/sweeper: basic
165
+ hydra/launcher: basic
166
+ hydra/output: default
167
+ verbose: false
2025-10-10/23-23-47/.hydra/overrides.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ - experiment=stage1
2
+ - hfhub.repo_id=jkminder/lorentz-poc-stage1
3
+ - wandb.project=lorentz-forcing
4
+ - training.num_train_epochs=1
5
+ - dataset.seq_len=512
6
+ - dataset.context_len=512
2025-10-10/23-23-47/train.log ADDED
File without changes
2025-10-10/23-28-17/.hydra/config.yaml ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ pretrained: meta-llama/Llama-3.2-1B
3
+ resize_embeddings: true
4
+ dataset:
5
+ name: science-of-finetuning/fineweb-1m-sample
6
+ config: ''
7
+ seq_len: 512
8
+ context_len: 512
9
+ num_train_samples: 1000
10
+ training:
11
+ output_dir: outputs
12
+ num_train_epochs: 1
13
+ per_device_train_batch_size: 4
14
+ gradient_accumulation_steps: 1
15
+ learning_rate: 5.0e-05
16
+ weight_decay: 0.0
17
+ warmup_steps: 0
18
+ save_steps: 500
19
+ logging_steps: 50
20
+ seed: 42
21
+ fp16: false
22
+ bf16: true
23
+ calibration_batch_size: 128
24
+ experiment:
25
+ stage: 1
26
+ p_intervene: 0.1
27
+ use_measurement: true
28
+ force_text: null
29
+ init_from:
30
+ hub_repo: null
31
+ local_ckpt: null
32
+ wandb:
33
+ project: lorentz-forcing
34
+ entity: jkminder
35
+ hfhub:
36
+ push_to_hub: true
37
+ repo_id: jkminder/lorentz-poc-stage1
38
+ eval:
39
+ output_path: eval_results.json
40
+ max_eval_samples: 128
41
+ seed: 42
42
+ ci:
43
+ method: bootstrap
44
+ level: 0.95
45
+ bootstrap_iters: 1000
2025-10-10/23-28-17/.hydra/hydra.yaml ADDED
@@ -0,0 +1,167 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - experiment=stage1
116
+ - hfhub.repo_id=jkminder/lorentz-poc-stage1
117
+ - wandb.project=lorentz-forcing
118
+ - training.num_train_epochs=1
119
+ - dataset.seq_len=512
120
+ - dataset.context_len=512
121
+ job:
122
+ name: train
123
+ chdir: null
124
+ override_dirname: dataset.context_len=512,dataset.seq_len=512,experiment=stage1,hfhub.repo_id=jkminder/lorentz-poc-stage1,training.num_train_epochs=1,wandb.project=lorentz-forcing
125
+ id: ???
126
+ num: ???
127
+ config_name: config
128
+ env_set: {}
129
+ env_copy: []
130
+ config:
131
+ override_dirname:
132
+ kv_sep: '='
133
+ item_sep: ','
134
+ exclude_keys: []
135
+ runtime:
136
+ version: 1.3.2
137
+ version_base: '1.3'
138
+ cwd: /mnt/nw/home/j.minder/repositories/lorentz_forcing
139
+ config_sources:
140
+ - path: hydra.conf
141
+ schema: pkg
142
+ provider: hydra
143
+ - path: /mnt/nw/home/j.minder/repositories/lorentz_forcing/conf
144
+ schema: file
145
+ provider: main
146
+ - path: ''
147
+ schema: structured
148
+ provider: schema
149
+ output_dir: /mnt/nw/home/j.minder/repositories/lorentz_forcing/outputs/2025-10-10/23-28-17
150
+ choices:
151
+ eval: default
152
+ hfhub: default
153
+ wandb: default
154
+ experiment: stage1
155
+ training: default
156
+ dataset: fineweb
157
+ model: llama32_1B
158
+ hydra/env: default
159
+ hydra/callbacks: null
160
+ hydra/job_logging: default
161
+ hydra/hydra_logging: default
162
+ hydra/hydra_help: default
163
+ hydra/help: default
164
+ hydra/sweeper: basic
165
+ hydra/launcher: basic
166
+ hydra/output: default
167
+ verbose: false
2025-10-10/23-28-17/.hydra/overrides.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ - experiment=stage1
2
+ - hfhub.repo_id=jkminder/lorentz-poc-stage1
3
+ - wandb.project=lorentz-forcing
4
+ - training.num_train_epochs=1
5
+ - dataset.seq_len=512
6
+ - dataset.context_len=512
2025-10-10/23-28-17/train.log ADDED
File without changes
2025-10-10/23-33-34/.hydra/config.yaml ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ pretrained: meta-llama/Llama-3.2-1B
3
+ resize_embeddings: true
4
+ dataset:
5
+ name: science-of-finetuning/fineweb-1m-sample
6
+ config: ''
7
+ seq_len: 512
8
+ context_len: 512
9
+ num_train_samples: 1000
10
+ training:
11
+ output_dir: outputs
12
+ num_train_epochs: 1
13
+ per_device_train_batch_size: 4
14
+ gradient_accumulation_steps: 1
15
+ learning_rate: 5.0e-05
16
+ weight_decay: 0.0
17
+ warmup_steps: 0
18
+ save_steps: 500
19
+ logging_steps: 50
20
+ seed: 42
21
+ fp16: false
22
+ bf16: true
23
+ calibration_batch_size: 128
24
+ experiment:
25
+ stage: 1
26
+ p_intervene: 0.1
27
+ use_measurement: true
28
+ force_text: null
29
+ init_from:
30
+ hub_repo: null
31
+ local_ckpt: null
32
+ wandb:
33
+ project: lorentz-forcing
34
+ entity: jkminder
35
+ hfhub:
36
+ push_to_hub: true
37
+ repo_id: jkminder/lorentz-poc-stage1
38
+ eval:
39
+ output_path: eval_results.json
40
+ max_eval_samples: 128
41
+ seed: 42
42
+ ci:
43
+ method: bootstrap
44
+ level: 0.95
45
+ bootstrap_iters: 1000
2025-10-10/23-33-34/.hydra/hydra.yaml ADDED
@@ -0,0 +1,167 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - experiment=stage1
116
+ - hfhub.repo_id=jkminder/lorentz-poc-stage1
117
+ - wandb.project=lorentz-forcing
118
+ - training.num_train_epochs=1
119
+ - dataset.seq_len=512
120
+ - dataset.context_len=512
121
+ job:
122
+ name: train
123
+ chdir: null
124
+ override_dirname: dataset.context_len=512,dataset.seq_len=512,experiment=stage1,hfhub.repo_id=jkminder/lorentz-poc-stage1,training.num_train_epochs=1,wandb.project=lorentz-forcing
125
+ id: ???
126
+ num: ???
127
+ config_name: config
128
+ env_set: {}
129
+ env_copy: []
130
+ config:
131
+ override_dirname:
132
+ kv_sep: '='
133
+ item_sep: ','
134
+ exclude_keys: []
135
+ runtime:
136
+ version: 1.3.2
137
+ version_base: '1.3'
138
+ cwd: /mnt/nw/home/j.minder/repositories/lorentz_forcing
139
+ config_sources:
140
+ - path: hydra.conf
141
+ schema: pkg
142
+ provider: hydra
143
+ - path: /mnt/nw/home/j.minder/repositories/lorentz_forcing/conf
144
+ schema: file
145
+ provider: main
146
+ - path: ''
147
+ schema: structured
148
+ provider: schema
149
+ output_dir: /mnt/nw/home/j.minder/repositories/lorentz_forcing/outputs/2025-10-10/23-33-34
150
+ choices:
151
+ eval: default
152
+ hfhub: default
153
+ wandb: default
154
+ experiment: stage1
155
+ training: default
156
+ dataset: fineweb
157
+ model: llama32_1B
158
+ hydra/env: default
159
+ hydra/callbacks: null
160
+ hydra/job_logging: default
161
+ hydra/hydra_logging: default
162
+ hydra/hydra_help: default
163
+ hydra/help: default
164
+ hydra/sweeper: basic
165
+ hydra/launcher: basic
166
+ hydra/output: default
167
+ verbose: false
2025-10-10/23-33-34/.hydra/overrides.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ - experiment=stage1
2
+ - hfhub.repo_id=jkminder/lorentz-poc-stage1
3
+ - wandb.project=lorentz-forcing
4
+ - training.num_train_epochs=1
5
+ - dataset.seq_len=512
6
+ - dataset.context_len=512
2025-10-10/23-33-34/train.log ADDED
File without changes
2025-10-10/23-35-10/.hydra/config.yaml ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ pretrained: meta-llama/Llama-3.2-1B
3
+ resize_embeddings: true
4
+ dataset:
5
+ name: science-of-finetuning/fineweb-1m-sample
6
+ config: ''
7
+ seq_len: 512
8
+ context_len: 512
9
+ num_train_samples: 1000
10
+ training:
11
+ output_dir: outputs
12
+ num_train_epochs: 1
13
+ per_device_train_batch_size: 4
14
+ gradient_accumulation_steps: 1
15
+ learning_rate: 5.0e-05
16
+ weight_decay: 0.0
17
+ warmup_steps: 0
18
+ save_steps: 500
19
+ logging_steps: 50
20
+ seed: 42
21
+ fp16: false
22
+ bf16: true
23
+ calibration_batch_size: 128
24
+ experiment:
25
+ stage: 1
26
+ p_intervene: 0.1
27
+ use_measurement: true
28
+ force_text: null
29
+ init_from:
30
+ hub_repo: null
31
+ local_ckpt: null
32
+ wandb:
33
+ project: lorentz-forcing
34
+ entity: jkminder
35
+ hfhub:
36
+ push_to_hub: true
37
+ repo_id: jkminder/lorentz-poc-stage1
38
+ eval:
39
+ output_path: eval_results.json
40
+ max_eval_samples: 128
41
+ seed: 42
42
+ ci:
43
+ method: bootstrap
44
+ level: 0.95
45
+ bootstrap_iters: 1000
2025-10-10/23-35-10/.hydra/hydra.yaml ADDED
@@ -0,0 +1,167 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - experiment=stage1
116
+ - hfhub.repo_id=jkminder/lorentz-poc-stage1
117
+ - wandb.project=lorentz-forcing
118
+ - training.num_train_epochs=1
119
+ - dataset.seq_len=512
120
+ - dataset.context_len=512
121
+ job:
122
+ name: train
123
+ chdir: null
124
+ override_dirname: dataset.context_len=512,dataset.seq_len=512,experiment=stage1,hfhub.repo_id=jkminder/lorentz-poc-stage1,training.num_train_epochs=1,wandb.project=lorentz-forcing
125
+ id: ???
126
+ num: ???
127
+ config_name: config
128
+ env_set: {}
129
+ env_copy: []
130
+ config:
131
+ override_dirname:
132
+ kv_sep: '='
133
+ item_sep: ','
134
+ exclude_keys: []
135
+ runtime:
136
+ version: 1.3.2
137
+ version_base: '1.3'
138
+ cwd: /mnt/nw/home/j.minder/repositories/lorentz_forcing
139
+ config_sources:
140
+ - path: hydra.conf
141
+ schema: pkg
142
+ provider: hydra
143
+ - path: /mnt/nw/home/j.minder/repositories/lorentz_forcing/conf
144
+ schema: file
145
+ provider: main
146
+ - path: ''
147
+ schema: structured
148
+ provider: schema
149
+ output_dir: /mnt/nw/home/j.minder/repositories/lorentz_forcing/outputs/2025-10-10/23-35-10
150
+ choices:
151
+ eval: default
152
+ hfhub: default
153
+ wandb: default
154
+ experiment: stage1
155
+ training: default
156
+ dataset: fineweb
157
+ model: llama32_1B
158
+ hydra/env: default
159
+ hydra/callbacks: null
160
+ hydra/job_logging: default
161
+ hydra/hydra_logging: default
162
+ hydra/hydra_help: default
163
+ hydra/help: default
164
+ hydra/sweeper: basic
165
+ hydra/launcher: basic
166
+ hydra/output: default
167
+ verbose: false
2025-10-10/23-35-10/.hydra/overrides.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ - experiment=stage1
2
+ - hfhub.repo_id=jkminder/lorentz-poc-stage1
3
+ - wandb.project=lorentz-forcing
4
+ - training.num_train_epochs=1
5
+ - dataset.seq_len=512
6
+ - dataset.context_len=512
2025-10-10/23-35-10/train.log ADDED
File without changes
2025-10-10/23-36-21/.hydra/config.yaml ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ pretrained: meta-llama/Llama-3.2-1B
3
+ resize_embeddings: true
4
+ dataset:
5
+ name: science-of-finetuning/fineweb-1m-sample
6
+ config: ''
7
+ seq_len: 512
8
+ context_len: 512
9
+ num_train_samples: 1000
10
+ training:
11
+ output_dir: outputs
12
+ num_train_epochs: 1
13
+ per_device_train_batch_size: 4
14
+ gradient_accumulation_steps: 1
15
+ learning_rate: 5.0e-05
16
+ weight_decay: 0.0
17
+ warmup_steps: 0
18
+ save_steps: 500
19
+ logging_steps: 50
20
+ seed: 42
21
+ fp16: false
22
+ bf16: true
23
+ calibration_batch_size: 128
24
+ experiment:
25
+ stage: 1
26
+ p_intervene: 0.1
27
+ use_measurement: true
28
+ force_text: null
29
+ init_from:
30
+ hub_repo: null
31
+ local_ckpt: null
32
+ wandb:
33
+ project: lorentz-forcing
34
+ entity: jkminder
35
+ hfhub:
36
+ push_to_hub: true
37
+ repo_id: jkminder/lorentz-poc-stage1
38
+ eval:
39
+ output_path: eval_results.json
40
+ max_eval_samples: 128
41
+ seed: 42
42
+ ci:
43
+ method: bootstrap
44
+ level: 0.95
45
+ bootstrap_iters: 1000
2025-10-10/23-36-21/.hydra/hydra.yaml ADDED
@@ -0,0 +1,167 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - experiment=stage1
116
+ - hfhub.repo_id=jkminder/lorentz-poc-stage1
117
+ - wandb.project=lorentz-forcing
118
+ - training.num_train_epochs=1
119
+ - dataset.seq_len=512
120
+ - dataset.context_len=512
121
+ job:
122
+ name: train
123
+ chdir: null
124
+ override_dirname: dataset.context_len=512,dataset.seq_len=512,experiment=stage1,hfhub.repo_id=jkminder/lorentz-poc-stage1,training.num_train_epochs=1,wandb.project=lorentz-forcing
125
+ id: ???
126
+ num: ???
127
+ config_name: config
128
+ env_set: {}
129
+ env_copy: []
130
+ config:
131
+ override_dirname:
132
+ kv_sep: '='
133
+ item_sep: ','
134
+ exclude_keys: []
135
+ runtime:
136
+ version: 1.3.2
137
+ version_base: '1.3'
138
+ cwd: /mnt/nw/home/j.minder/repositories/lorentz_forcing
139
+ config_sources:
140
+ - path: hydra.conf
141
+ schema: pkg
142
+ provider: hydra
143
+ - path: /mnt/nw/home/j.minder/repositories/lorentz_forcing/conf
144
+ schema: file
145
+ provider: main
146
+ - path: ''
147
+ schema: structured
148
+ provider: schema
149
+ output_dir: /mnt/nw/home/j.minder/repositories/lorentz_forcing/outputs/2025-10-10/23-36-21
150
+ choices:
151
+ eval: default
152
+ hfhub: default
153
+ wandb: default
154
+ experiment: stage1
155
+ training: default
156
+ dataset: fineweb
157
+ model: llama32_1B
158
+ hydra/env: default
159
+ hydra/callbacks: null
160
+ hydra/job_logging: default
161
+ hydra/hydra_logging: default
162
+ hydra/hydra_help: default
163
+ hydra/help: default
164
+ hydra/sweeper: basic
165
+ hydra/launcher: basic
166
+ hydra/output: default
167
+ verbose: false
2025-10-10/23-36-21/.hydra/overrides.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ - experiment=stage1
2
+ - hfhub.repo_id=jkminder/lorentz-poc-stage1
3
+ - wandb.project=lorentz-forcing
4
+ - training.num_train_epochs=1
5
+ - dataset.seq_len=512
6
+ - dataset.context_len=512
2025-10-10/23-36-21/train.log ADDED
File without changes
2025-10-10/23-40-03/.hydra/config.yaml ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ pretrained: meta-llama/Llama-3.2-1B
3
+ resize_embeddings: true
4
+ dataset:
5
+ name: science-of-finetuning/fineweb-1m-sample
6
+ config: ''
7
+ seq_len: 512
8
+ context_len: 512
9
+ num_train_samples: 1000
10
+ training:
11
+ output_dir: outputs
12
+ num_train_epochs: 1
13
+ per_device_train_batch_size: 4
14
+ gradient_accumulation_steps: 1
15
+ learning_rate: 5.0e-05
16
+ weight_decay: 0.0
17
+ warmup_steps: 0
18
+ save_steps: 500
19
+ logging_steps: 50
20
+ seed: 42
21
+ fp16: false
22
+ bf16: true
23
+ calibration_batch_size: 128
24
+ experiment:
25
+ stage: 1
26
+ p_intervene: 0.1
27
+ use_measurement: true
28
+ force_text: null
29
+ init_from:
30
+ hub_repo: null
31
+ local_ckpt: null
32
+ wandb:
33
+ project: lorentz-forcing
34
+ entity: jkminder
35
+ hfhub:
36
+ push_to_hub: true
37
+ repo_id: jkminder/lorentz-poc-stage1
38
+ eval:
39
+ output_path: eval_results.json
40
+ max_eval_samples: 128
41
+ seed: 42
42
+ ci:
43
+ method: bootstrap
44
+ level: 0.95
45
+ bootstrap_iters: 1000
2025-10-10/23-40-03/.hydra/hydra.yaml ADDED
@@ -0,0 +1,167 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - experiment=stage1
116
+ - hfhub.repo_id=jkminder/lorentz-poc-stage1
117
+ - wandb.project=lorentz-forcing
118
+ - training.num_train_epochs=1
119
+ - dataset.seq_len=512
120
+ - dataset.context_len=512
121
+ job:
122
+ name: train
123
+ chdir: null
124
+ override_dirname: dataset.context_len=512,dataset.seq_len=512,experiment=stage1,hfhub.repo_id=jkminder/lorentz-poc-stage1,training.num_train_epochs=1,wandb.project=lorentz-forcing
125
+ id: ???
126
+ num: ???
127
+ config_name: config
128
+ env_set: {}
129
+ env_copy: []
130
+ config:
131
+ override_dirname:
132
+ kv_sep: '='
133
+ item_sep: ','
134
+ exclude_keys: []
135
+ runtime:
136
+ version: 1.3.2
137
+ version_base: '1.3'
138
+ cwd: /mnt/nw/home/j.minder/repositories/lorentz_forcing
139
+ config_sources:
140
+ - path: hydra.conf
141
+ schema: pkg
142
+ provider: hydra
143
+ - path: /mnt/nw/home/j.minder/repositories/lorentz_forcing/conf
144
+ schema: file
145
+ provider: main
146
+ - path: ''
147
+ schema: structured
148
+ provider: schema
149
+ output_dir: /mnt/nw/home/j.minder/repositories/lorentz_forcing/outputs/2025-10-10/23-40-03
150
+ choices:
151
+ eval: default
152
+ hfhub: default
153
+ wandb: default
154
+ experiment: stage1
155
+ training: default
156
+ dataset: fineweb
157
+ model: llama32_1B
158
+ hydra/env: default
159
+ hydra/callbacks: null
160
+ hydra/job_logging: default
161
+ hydra/hydra_logging: default
162
+ hydra/hydra_help: default
163
+ hydra/help: default
164
+ hydra/sweeper: basic
165
+ hydra/launcher: basic
166
+ hydra/output: default
167
+ verbose: false
2025-10-10/23-40-03/.hydra/overrides.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ - experiment=stage1
2
+ - hfhub.repo_id=jkminder/lorentz-poc-stage1
3
+ - wandb.project=lorentz-forcing
4
+ - training.num_train_epochs=1
5
+ - dataset.seq_len=512
6
+ - dataset.context_len=512
2025-10-10/23-40-03/train.log ADDED
File without changes
2025-10-10/23-44-29/.hydra/config.yaml ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ pretrained: meta-llama/Llama-3.2-1B
3
+ resize_embeddings: true
4
+ dataset:
5
+ name: science-of-finetuning/fineweb-1m-sample
6
+ config: ''
7
+ seq_len: 512
8
+ context_len: 512
9
+ num_train_samples: 1000
10
+ training:
11
+ output_dir: outputs
12
+ num_train_epochs: 1
13
+ per_device_train_batch_size: 4
14
+ gradient_accumulation_steps: 1
15
+ learning_rate: 5.0e-05
16
+ weight_decay: 0.0
17
+ warmup_steps: 0
18
+ save_steps: 500
19
+ logging_steps: 50
20
+ seed: 42
21
+ fp16: false
22
+ bf16: true
23
+ calibration_batch_size: 128
24
+ experiment:
25
+ stage: 1
26
+ p_intervene: 0.1
27
+ use_measurement: true
28
+ force_text: null
29
+ init_from:
30
+ hub_repo: null
31
+ local_ckpt: null
32
+ wandb:
33
+ project: lorentz-forcing
34
+ entity: jkminder
35
+ hfhub:
36
+ push_to_hub: true
37
+ repo_id: jkminder/lorentz-poc-stage1
38
+ eval:
39
+ output_path: eval_results.json
40
+ max_eval_samples: 128
41
+ seed: 42
42
+ ci:
43
+ method: bootstrap
44
+ level: 0.95
45
+ bootstrap_iters: 1000
2025-10-10/23-44-29/.hydra/hydra.yaml ADDED
@@ -0,0 +1,167 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - experiment=stage1
116
+ - hfhub.repo_id=jkminder/lorentz-poc-stage1
117
+ - wandb.project=lorentz-forcing
118
+ - training.num_train_epochs=1
119
+ - dataset.seq_len=512
120
+ - dataset.context_len=512
121
+ job:
122
+ name: train
123
+ chdir: null
124
+ override_dirname: dataset.context_len=512,dataset.seq_len=512,experiment=stage1,hfhub.repo_id=jkminder/lorentz-poc-stage1,training.num_train_epochs=1,wandb.project=lorentz-forcing
125
+ id: ???
126
+ num: ???
127
+ config_name: config
128
+ env_set: {}
129
+ env_copy: []
130
+ config:
131
+ override_dirname:
132
+ kv_sep: '='
133
+ item_sep: ','
134
+ exclude_keys: []
135
+ runtime:
136
+ version: 1.3.2
137
+ version_base: '1.3'
138
+ cwd: /mnt/nw/home/j.minder/repositories/lorentz_forcing
139
+ config_sources:
140
+ - path: hydra.conf
141
+ schema: pkg
142
+ provider: hydra
143
+ - path: /mnt/nw/home/j.minder/repositories/lorentz_forcing/conf
144
+ schema: file
145
+ provider: main
146
+ - path: ''
147
+ schema: structured
148
+ provider: schema
149
+ output_dir: /mnt/nw/home/j.minder/repositories/lorentz_forcing/outputs/2025-10-10/23-44-29
150
+ choices:
151
+ eval: default
152
+ hfhub: default
153
+ wandb: default
154
+ experiment: stage1
155
+ training: default
156
+ dataset: fineweb
157
+ model: llama32_1B
158
+ hydra/env: default
159
+ hydra/callbacks: null
160
+ hydra/job_logging: default
161
+ hydra/hydra_logging: default
162
+ hydra/hydra_help: default
163
+ hydra/help: default
164
+ hydra/sweeper: basic
165
+ hydra/launcher: basic
166
+ hydra/output: default
167
+ verbose: false
2025-10-10/23-44-29/.hydra/overrides.yaml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ - experiment=stage1
2
+ - hfhub.repo_id=jkminder/lorentz-poc-stage1
3
+ - wandb.project=lorentz-forcing
4
+ - training.num_train_epochs=1
5
+ - dataset.seq_len=512
6
+ - dataset.context_len=512
2025-10-10/23-44-29/train.log ADDED
File without changes
2025-10-10/23-47-52/.hydra/config.yaml ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ pretrained: meta-llama/Llama-3.2-1B
3
+ resize_embeddings: true
4
+ dataset:
5
+ name: science-of-finetuning/fineweb-1m-sample
6
+ config: ''
7
+ seq_len: 512
8
+ context_len: 512
9
+ num_train_samples: 1000
10
+ training:
11
+ output_dir: outputs
12
+ num_train_epochs: 1
13
+ per_device_train_batch_size: 4
14
+ gradient_accumulation_steps: 1
15
+ learning_rate: 5.0e-05
16
+ weight_decay: 0.0
17
+ warmup_steps: 0
18
+ save_steps: 500
19
+ logging_steps: 50
20
+ seed: 42
21
+ fp16: false
22
+ bf16: true
23
+ calibration_batch_size: 128
24
+ experiment:
25
+ stage: 1
26
+ p_intervene: 0.1
27
+ use_measurement: true
28
+ force_text: null
29
+ init_from:
30
+ hub_repo: null
31
+ local_ckpt: null
32
+ wandb:
33
+ project: lorentz-forcing
34
+ entity: jkminder
35
+ hfhub:
36
+ push_to_hub: true
37
+ repo_id: jkminder/lorentz-poc-stage1
38
+ eval:
39
+ output_path: eval_results.json
40
+ max_eval_samples: 128
41
+ seed: 42
42
+ ci:
43
+ method: bootstrap
44
+ level: 0.95
45
+ bootstrap_iters: 1000