khazic commited on
Commit
3e6d033
Β·
verified Β·
1 Parent(s): 704a564

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes. Β  See raw diff
Files changed (50) hide show
  1. .gitattributes +5 -0
  2. outputs/2025-11-18/exp1_len1024/dream-inst/dllm/gsm8k/.hydra/config.yaml +53 -0
  3. outputs/2025-11-18/exp1_len1024/dream-inst/dllm/gsm8k/.hydra/hydra.yaml +167 -0
  4. outputs/2025-11-18/exp1_len1024/dream-inst/dllm/gsm8k/.hydra/overrides.yaml +9 -0
  5. outputs/2025-11-18/exp1_len1024/dream-inst/dllm/gsm8k/eval.log +41 -0
  6. outputs/2025-11-18/exp1_len1024/dream-inst/dllm/gsm8k/results.json +3 -0
  7. outputs/2025-11-18/exp1_len1024/dream-inst/dllm/gsm8k/stderr.log +63 -0
  8. outputs/2025-11-18/exp1_len1024/dream-inst/dllm/humaneval/.hydra/config.yaml +53 -0
  9. outputs/2025-11-18/exp1_len1024/dream-inst/dllm/humaneval/.hydra/hydra.yaml +167 -0
  10. outputs/2025-11-18/exp1_len1024/dream-inst/dllm/humaneval/.hydra/overrides.yaml +9 -0
  11. outputs/2025-11-18/exp1_len1024/dream-inst/dllm/humaneval/eval.log +41 -0
  12. outputs/2025-11-18/exp1_len1024/dream-inst/dllm/humaneval/results.json +0 -0
  13. outputs/2025-11-18/exp1_len1024/dream-inst/dllm/humaneval/stderr.log +63 -0
  14. outputs/2025-11-18/exp1_len1024/dream-inst/dllm/math-500/.hydra/config.yaml +53 -0
  15. outputs/2025-11-18/exp1_len1024/dream-inst/dllm/math-500/.hydra/hydra.yaml +167 -0
  16. outputs/2025-11-18/exp1_len1024/dream-inst/dllm/math-500/.hydra/overrides.yaml +9 -0
  17. outputs/2025-11-18/exp1_len1024/dream-inst/dllm/math-500/eval.log +41 -0
  18. outputs/2025-11-18/exp1_len1024/dream-inst/dllm/math-500/results.json +0 -0
  19. outputs/2025-11-18/exp1_len1024/dream-inst/dllm/math-500/stderr.log +103 -0
  20. outputs/2025-11-18/exp1_len1024/dream-inst/dllm/mbpp/.hydra/config.yaml +53 -0
  21. outputs/2025-11-18/exp1_len1024/dream-inst/dllm/mbpp/.hydra/hydra.yaml +167 -0
  22. outputs/2025-11-18/exp1_len1024/dream-inst/dllm/mbpp/.hydra/overrides.yaml +9 -0
  23. outputs/2025-11-18/exp1_len1024/dream-inst/dllm/mbpp/eval.log +41 -0
  24. outputs/2025-11-18/exp1_len1024/dream-inst/dllm/mbpp/results.json +0 -0
  25. outputs/2025-11-18/exp1_len1024/dream-inst/dllm/mbpp/stderr.log +63 -0
  26. outputs/2025-11-18/exp1_len1024/dream-inst/heat/gsm8k/.hydra/config.yaml +54 -0
  27. outputs/2025-11-18/exp1_len1024/dream-inst/heat/gsm8k/.hydra/hydra.yaml +168 -0
  28. outputs/2025-11-18/exp1_len1024/dream-inst/heat/gsm8k/.hydra/overrides.yaml +10 -0
  29. outputs/2025-11-18/exp1_len1024/dream-inst/heat/gsm8k/eval.log +41 -0
  30. outputs/2025-11-18/exp1_len1024/dream-inst/heat/gsm8k/results.json +3 -0
  31. outputs/2025-11-18/exp1_len1024/dream-inst/heat/gsm8k/stderr.log +63 -0
  32. outputs/2025-11-18/exp1_len1024/dream-inst/heat/humaneval/.hydra/config.yaml +54 -0
  33. outputs/2025-11-18/exp1_len1024/dream-inst/heat/humaneval/.hydra/hydra.yaml +168 -0
  34. outputs/2025-11-18/exp1_len1024/dream-inst/heat/humaneval/.hydra/overrides.yaml +10 -0
  35. outputs/2025-11-18/exp1_len1024/dream-inst/heat/humaneval/eval.log +41 -0
  36. outputs/2025-11-18/exp1_len1024/dream-inst/heat/humaneval/results.json +0 -0
  37. outputs/2025-11-18/exp1_len1024/dream-inst/heat/humaneval/stderr.log +63 -0
  38. outputs/2025-11-18/exp1_len1024/dream-inst/heat/math-500/.hydra/config.yaml +54 -0
  39. outputs/2025-11-18/exp1_len1024/dream-inst/heat/math-500/.hydra/hydra.yaml +168 -0
  40. outputs/2025-11-18/exp1_len1024/dream-inst/heat/math-500/.hydra/overrides.yaml +10 -0
  41. outputs/2025-11-18/exp1_len1024/dream-inst/heat/math-500/eval.log +41 -0
  42. outputs/2025-11-18/exp1_len1024/dream-inst/heat/math-500/results.json +0 -0
  43. outputs/2025-11-18/exp1_len1024/dream-inst/heat/math-500/stderr.log +103 -0
  44. outputs/2025-11-18/exp1_len1024/dream-inst/heat/mbpp/.hydra/config.yaml +54 -0
  45. outputs/2025-11-18/exp1_len1024/dream-inst/heat/mbpp/.hydra/hydra.yaml +168 -0
  46. outputs/2025-11-18/exp1_len1024/dream-inst/heat/mbpp/.hydra/overrides.yaml +10 -0
  47. outputs/2025-11-18/exp1_len1024/dream-inst/heat/mbpp/eval.log +41 -0
  48. outputs/2025-11-18/exp1_len1024/dream-inst/heat/mbpp/results.json +0 -0
  49. outputs/2025-11-18/exp1_len1024/dream-inst/heat/mbpp/stderr.log +63 -0
  50. outputs/2025-11-18/exp1_len1024/dream-inst/no_cache/gsm8k/.hydra/config.yaml +48 -0
.gitattributes CHANGED
@@ -33,3 +33,8 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ outputs/2025-11-18/exp1_len1024/dream-inst/dllm/gsm8k/results.json filter=lfs diff=lfs merge=lfs -text
37
+ outputs/2025-11-18/exp1_len1024/dream-inst/heat/gsm8k/results.json filter=lfs diff=lfs merge=lfs -text
38
+ outputs/2025-11-18/exp1_len1024/dream-inst/no_cache/gsm8k/results.json filter=lfs diff=lfs merge=lfs -text
39
+ outputs/2025-11-18/exp1_len1024/dream-inst/prefix/gsm8k/results.json filter=lfs diff=lfs merge=lfs -text
40
+ outputs/2025-11-18/exp2_random_gen/dream-inst/gsm8k_len256/results.json filter=lfs diff=lfs merge=lfs -text
outputs/2025-11-18/exp1_len1024/dream-inst/dllm/gsm8k/.hydra/config.yaml ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ generation:
3
+ mask_token_id: 151666
4
+ eot_token_id: 151643
5
+ pad_token_id: 151643
6
+ add_bos_token: true
7
+ alg: maskgit_plus
8
+ name: dream-inst
9
+ path: ${oc.env:DREAM_INST_PATH}
10
+ generation:
11
+ strategy: vanilla
12
+ threshold: null
13
+ factor: null
14
+ alg: maskgit_plus
15
+ gen_length: 1024
16
+ block_length: 32
17
+ steps: 1024
18
+ temperature: 0.0
19
+ top_p: null
20
+ top_k: null
21
+ debias: false
22
+ output_probs: false
23
+ cache:
24
+ _target_: src.cache.dLLMCache
25
+ kr: 1
26
+ kp: 50
27
+ rou: 0.25
28
+ seed: 1234
29
+ batch_size: 1
30
+ attn_implementation: eager
31
+ dataset:
32
+ name: gsm8k
33
+ size: null
34
+ n_shot: null
35
+ system_prompt: null
36
+ batch_size: 1
37
+ mc_num: null
38
+ max_length: 4096
39
+ is_check_greedy: true
40
+ add_bos_token: true
41
+ nll_type: mc
42
+ log_type: ftb
43
+ eval_args:
44
+ log_samples: true
45
+ tasks: ${..dataset.name}
46
+ num_fewshot: ${..dataset.n_shot}
47
+ batch_size: ${..batch_size}
48
+ limit: ${..dataset.size}
49
+ confirm_run_unsafe_code: true
50
+ random_seed: ${..seed}
51
+ fewshot_random_seed: ${..seed}
52
+ numpy_random_seed: ${..seed}
53
+ torch_random_seed: ${..seed}
outputs/2025-11-18/exp1_len1024/dream-inst/dllm/gsm8k/.hydra/hydra.yaml ADDED
@@ -0,0 +1,167 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/2025-11-18/exp1_len1024/dream-inst/dllm/gsm8k
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.run.dir=outputs/2025-11-18/exp1_len1024/dream-inst/dllm/gsm8k
114
+ - hydra.mode=RUN
115
+ task:
116
+ - dataset.name=gsm8k
117
+ - model=dream-inst
118
+ - cache=dllm
119
+ - generation=vanilla
120
+ - generation.gen_length=1024
121
+ - generation.steps=1024
122
+ - generation.block_length=32
123
+ - batch_size=1
124
+ - seed=1234
125
+ job:
126
+ name: eval
127
+ chdir: null
128
+ override_dirname: batch_size=1,cache=dllm,dataset.name=gsm8k,generation.block_length=32,generation.gen_length=1024,generation.steps=1024,generation=vanilla,model=dream-inst,seed=1234
129
+ id: ???
130
+ num: ???
131
+ config_name: eval
132
+ env_set: {}
133
+ env_copy: []
134
+ config:
135
+ override_dirname:
136
+ kv_sep: '='
137
+ item_sep: ','
138
+ exclude_keys: []
139
+ runtime:
140
+ version: 1.3.2
141
+ version_base: '1.3'
142
+ cwd: /xfr_ceph_sh/liuchonghan/HEAT/heat
143
+ config_sources:
144
+ - path: hydra.conf
145
+ schema: pkg
146
+ provider: hydra
147
+ - path: /xfr_ceph_sh/liuchonghan/HEAT/heat/configs
148
+ schema: file
149
+ provider: main
150
+ - path: ''
151
+ schema: structured
152
+ provider: schema
153
+ output_dir: /xfr_ceph_sh/liuchonghan/HEAT/heat/outputs/2025-11-18/exp1_len1024/dream-inst/dllm/gsm8k
154
+ choices:
155
+ cache: dllm
156
+ generation: vanilla
157
+ model: dream-inst
158
+ hydra/env: default
159
+ hydra/callbacks: null
160
+ hydra/job_logging: default
161
+ hydra/hydra_logging: default
162
+ hydra/hydra_help: default
163
+ hydra/help: default
164
+ hydra/sweeper: basic
165
+ hydra/launcher: basic
166
+ hydra/output: default
167
+ verbose: false
outputs/2025-11-18/exp1_len1024/dream-inst/dllm/gsm8k/.hydra/overrides.yaml ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ - dataset.name=gsm8k
2
+ - model=dream-inst
3
+ - cache=dllm
4
+ - generation=vanilla
5
+ - generation.gen_length=1024
6
+ - generation.steps=1024
7
+ - generation.block_length=32
8
+ - batch_size=1
9
+ - seed=1234
outputs/2025-11-18/exp1_len1024/dream-inst/dllm/gsm8k/eval.log ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-11-18 21:17:27,700][accelerate.utils.other][WARNING] - Detected kernel version 5.4.0, which is below the recommended minimum of 5.5.0; this can cause the process to hang. It is recommended to upgrade the kernel to the minimum version or higher.
2
+ [2025-11-18 21:17:35,559][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
3
+ [2025-11-18 21:17:35,559][lm_eval.evaluator][INFO] - Using pre-initialized model
4
+ [2025-11-18 21:17:35,663][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
5
+ [2025-11-18 21:17:35,664][lm_eval.evaluator][INFO] - Using pre-initialized model
6
+ [2025-11-18 21:17:35,918][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
7
+ [2025-11-18 21:17:35,918][lm_eval.evaluator][INFO] - Using pre-initialized model
8
+ [2025-11-18 21:17:36,015][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
9
+ [2025-11-18 21:17:36,015][lm_eval.evaluator][INFO] - Using pre-initialized model
10
+ [2025-11-18 21:17:37,961][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
11
+ [2025-11-18 21:17:37,961][lm_eval.evaluator][INFO] - Using pre-initialized model
12
+ [2025-11-18 21:17:38,027][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
13
+ [2025-11-18 21:17:38,027][lm_eval.evaluator][INFO] - Using pre-initialized model
14
+ [2025-11-18 21:17:38,231][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
15
+ [2025-11-18 21:17:38,231][lm_eval.evaluator][INFO] - Using pre-initialized model
16
+ [2025-11-18 21:17:39,335][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
17
+ [2025-11-18 21:17:39,335][lm_eval.evaluator][INFO] - Using pre-initialized model
18
+ [2025-11-18 21:17:46,329][lm_eval.evaluator][INFO] - gsm8k: Using gen_kwargs: {'until': ['Question:', '</s>', '<|im_end|>'], 'do_sample': False, 'temperature': 0.0}
19
+ [2025-11-18 21:17:46,330][lm_eval.api.task][INFO] - Building contexts for gsm8k on rank 5...
20
+ [2025-11-18 21:17:46,906][lm_eval.evaluator][INFO] - gsm8k: Using gen_kwargs: {'until': ['Question:', '</s>', '<|im_end|>'], 'do_sample': False, 'temperature': 0.0}
21
+ [2025-11-18 21:17:46,907][lm_eval.api.task][INFO] - Building contexts for gsm8k on rank 7...
22
+ [2025-11-18 21:17:47,086][lm_eval.evaluator][INFO] - gsm8k: Using gen_kwargs: {'until': ['Question:', '</s>', '<|im_end|>'], 'do_sample': False, 'temperature': 0.0}
23
+ [2025-11-18 21:17:47,087][lm_eval.api.task][INFO] - Building contexts for gsm8k on rank 3...
24
+ [2025-11-18 21:17:48,048][lm_eval.evaluator][INFO] - gsm8k: Using gen_kwargs: {'until': ['Question:', '</s>', '<|im_end|>'], 'do_sample': False, 'temperature': 0.0}
25
+ [2025-11-18 21:17:48,048][lm_eval.api.task][INFO] - Building contexts for gsm8k on rank 6...
26
+ [2025-11-18 21:17:51,594][lm_eval.evaluator][INFO] - gsm8k: Using gen_kwargs: {'until': ['Question:', '</s>', '<|im_end|>'], 'do_sample': False, 'temperature': 0.0}
27
+ [2025-11-18 21:17:51,595][lm_eval.api.task][INFO] - Building contexts for gsm8k on rank 0...
28
+ [2025-11-18 21:17:51,886][lm_eval.evaluator][INFO] - gsm8k: Using gen_kwargs: {'until': ['Question:', '</s>', '<|im_end|>'], 'do_sample': False, 'temperature': 0.0}
29
+ [2025-11-18 21:17:51,886][lm_eval.api.task][INFO] - Building contexts for gsm8k on rank 4...
30
+ [2025-11-18 21:17:53,012][lm_eval.evaluator][INFO] - gsm8k: Using gen_kwargs: {'until': ['Question:', '</s>', '<|im_end|>'], 'do_sample': False, 'temperature': 0.0}
31
+ [2025-11-18 21:17:53,012][lm_eval.api.task][INFO] - Building contexts for gsm8k on rank 1...
32
+ [2025-11-18 21:17:53,340][lm_eval.evaluator][INFO] - gsm8k: Using gen_kwargs: {'until': ['Question:', '</s>', '<|im_end|>'], 'do_sample': False, 'temperature': 0.0}
33
+ [2025-11-18 21:17:53,341][lm_eval.api.task][INFO] - Building contexts for gsm8k on rank 2...
34
+ [2025-11-18 21:17:57,676][lm_eval.evaluator][INFO] - Running generate_until requests
35
+ [2025-11-18 21:17:57,676][lm_eval.evaluator][INFO] - Running generate_until requests
36
+ [2025-11-18 21:17:57,676][lm_eval.evaluator][INFO] - Running generate_until requests
37
+ [2025-11-18 21:17:57,676][lm_eval.evaluator][INFO] - Running generate_until requests
38
+ [2025-11-18 21:17:57,676][lm_eval.evaluator][INFO] - Running generate_until requests
39
+ [2025-11-18 21:17:57,677][lm_eval.evaluator][INFO] - Running generate_until requests
40
+ [2025-11-18 21:17:57,677][lm_eval.evaluator][INFO] - Running generate_until requests
41
+ [2025-11-18 21:17:57,677][lm_eval.evaluator][INFO] - Running generate_until requests
outputs/2025-11-18/exp1_len1024/dream-inst/dllm/gsm8k/results.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c31ab00416ae76d358c5c97124fbbca1c6b43cf1b1d325f07b3ed2e163bb17e
3
+ size 13676970
outputs/2025-11-18/exp1_len1024/dream-inst/dllm/gsm8k/stderr.log ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0
  0%| | 0/165 [00:00<?, ?it/s]
1
  17%|β–ˆβ–‹ | 28/165 [00:00<00:00, 271.20it/s]
2
  34%|β–ˆβ–ˆβ–ˆβ– | 56/165 [00:00<00:00, 271.85it/s]
3
  53%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 87/165 [00:00<00:00, 286.62it/s]
4
  77%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 127/165 [00:00<00:00, 327.90it/s]
 
5
  0%| | 0/164 [00:00<?, ?it/s]
6
  13%|β–ˆβ–Ž | 22/164 [00:00<00:00, 216.66it/s]
7
  0%| | 0/165 [00:00<?, ?it/s]
8
  37%|β–ˆβ–ˆβ–ˆβ–‹ | 61/164 [00:00<00:00, 315.68it/s]
9
  24%|β–ˆβ–ˆβ–Ž | 39/165 [00:00<00:00, 386.72it/s]
10
  61%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 100/164 [00:00<00:00, 346.89it/s]
11
  47%|β–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 78/165 [00:00<00:00, 387.03it/s]
12
  85%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 139/164 [00:00<00:00, 363.02it/s]
13
  71%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 117/165 [00:00<00:00, 387.94it/s]
 
14
  95%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 156/165 [00:00<00:00, 387.82it/s]
 
15
  0%| | 0/165 [00:00<?, ?it/s]
16
  24%|β–ˆβ–ˆβ–Ž | 39/165 [00:00<00:00, 385.35it/s]
17
  47%|β–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 78/165 [00:00<00:00, 387.84it/s]
18
  72%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 118/165 [00:00<00:00, 388.85it/s]
19
  96%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ| 158/165 [00:00<00:00, 389.70it/s]
 
20
  0%| | 0/165 [00:00<?, ?it/s]
21
  21%|β–ˆβ–ˆ | 35/165 [00:00<00:00, 348.57it/s]
22
  45%|β–ˆβ–ˆβ–ˆβ–ˆβ– | 74/165 [00:00<00:00, 368.97it/s]
23
  0%| | 0/165 [00:00<?, ?it/s]
24
  68%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 112/165 [00:00<00:00, 373.57it/s]
25
  24%|β–ˆβ–ˆβ–Ž | 39/165 [00:00<00:00, 385.26it/s]
26
  92%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 151/165 [00:00<00:00, 377.34it/s]
 
27
  47%|β–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 78/165 [00:00<00:00, 386.80it/s]
28
  72%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 118/165 [00:00<00:00, 388.67it/s]
29
  96%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ| 158/165 [00:00<00:00, 390.94it/s]
 
30
  0%| | 0/165 [00:00<?, ?it/s]
31
  24%|β–ˆβ–ˆβ–Ž | 39/165 [00:00<00:00, 385.52it/s]
32
  47%|β–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 78/165 [00:00<00:00, 388.01it/s]
33
  72%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 118/165 [00:00<00:00, 390.90it/s]
34
  0%| | 0/165 [00:00<?, ?it/s]
35
  96%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ| 158/165 [00:00<00:00, 392.38it/s]
 
36
  24%|β–ˆβ–ˆβ–Ž | 39/165 [00:00<00:00, 386.65it/s]
37
  47%|β–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 78/165 [00:00<00:00, 386.51it/s]
38
  72%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 118/165 [00:00<00:00, 388.81it/s]
39
  96%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ| 158/165 [00:00<00:00, 390.15it/s]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ipex flag is deprecated, will be removed in Accelerate v1.10. From 2.7.0, PyTorch has all needed optimizations for Intel CPU and XPU.
2
+ The following values were not passed to `accelerate launch` and had defaults used instead:
3
+ More than one GPU was found, enabling multi-GPU training.
4
+ If this was unintended please pass in `--num_processes=1`.
5
+ `--mixed_precision` was set to a value of `'no'`
6
+ `--dynamo_backend` was set to a value of `'no'`
7
+ To avoid this warning pass in values for each of the problematic parameters or run `accelerate config`.
8
+ [W1118 21:17:11.095728867 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
9
+ 2025-11-18 21:17:19.569 | INFO | src.utils:pre_initialize:603 - {'strategy': 'vanilla', 'threshold': None, 'factor': None, 'alg': 'maskgit_plus', 'gen_length': 1024, 'block_length': 32, 'steps': 1024, 'temperature': 0.0, 'top_p': 0.9, 'top_k': None, 'debias': False, 'output_probs': False, 'mask_token_id': 151666, 'eot_token_id': 151643, 'pad_token_id': 151643, 'add_bos_token': True, 'sigma': None}
10
+ 2025-11-18 21:17:19.569 | INFO | src.utils:pre_initialize:618 - Using cache with args: {'kp': 25, 'kr': 2}
11
+ [W1118 21:17:24.959332404 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
12
+ [W1118 21:17:24.316010842 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
13
+ [W1118 21:17:24.483325854 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
14
+ [W1118 21:17:24.510713611 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
15
+ [W1118 21:17:24.551847406 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
16
+ [W1118 21:17:24.622360312 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
17
+ [W1118 21:17:24.658556962 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
18
+ [W1118 21:17:25.013811805 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
19
+
20
+
21
+
22
+
23
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
24
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
25
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
26
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
27
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
28
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
29
+
30
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
31
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
32
+
33
+
34
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
35
+
36
+
37
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
38
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
39
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
40
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
41
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
42
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
43
+
44
  0%| | 0/165 [00:00<?, ?it/s]
45
  17%|β–ˆβ–‹ | 28/165 [00:00<00:00, 271.20it/s]
46
  34%|β–ˆβ–ˆβ–ˆβ– | 56/165 [00:00<00:00, 271.85it/s]
47
  53%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 87/165 [00:00<00:00, 286.62it/s]
48
  77%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 127/165 [00:00<00:00, 327.90it/s]
49
+
50
  0%| | 0/164 [00:00<?, ?it/s]
51
  13%|β–ˆβ–Ž | 22/164 [00:00<00:00, 216.66it/s]
52
  0%| | 0/165 [00:00<?, ?it/s]
53
  37%|β–ˆβ–ˆβ–ˆβ–‹ | 61/164 [00:00<00:00, 315.68it/s]
54
  24%|β–ˆβ–ˆβ–Ž | 39/165 [00:00<00:00, 386.72it/s]
55
  61%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 100/164 [00:00<00:00, 346.89it/s]
56
  47%|β–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 78/165 [00:00<00:00, 387.03it/s]
57
  85%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 139/164 [00:00<00:00, 363.02it/s]
58
  71%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 117/165 [00:00<00:00, 387.94it/s]
59
+
60
  95%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 156/165 [00:00<00:00, 387.82it/s]
61
+
62
  0%| | 0/165 [00:00<?, ?it/s]
63
  24%|β–ˆβ–ˆβ–Ž | 39/165 [00:00<00:00, 385.35it/s]
64
  47%|β–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 78/165 [00:00<00:00, 387.84it/s]
65
  72%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 118/165 [00:00<00:00, 388.85it/s]
66
  96%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ| 158/165 [00:00<00:00, 389.70it/s]
67
+
68
  0%| | 0/165 [00:00<?, ?it/s]
69
  21%|β–ˆβ–ˆ | 35/165 [00:00<00:00, 348.57it/s]
70
  45%|β–ˆβ–ˆβ–ˆβ–ˆβ– | 74/165 [00:00<00:00, 368.97it/s]
71
  0%| | 0/165 [00:00<?, ?it/s]
72
  68%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 112/165 [00:00<00:00, 373.57it/s]
73
  24%|β–ˆβ–ˆβ–Ž | 39/165 [00:00<00:00, 385.26it/s]
74
  92%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 151/165 [00:00<00:00, 377.34it/s]
75
+
76
  47%|β–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 78/165 [00:00<00:00, 386.80it/s]
77
  72%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 118/165 [00:00<00:00, 388.67it/s]
78
  96%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ| 158/165 [00:00<00:00, 390.94it/s]
79
+
80
  0%| | 0/165 [00:00<?, ?it/s]
81
  24%|β–ˆβ–ˆβ–Ž | 39/165 [00:00<00:00, 385.52it/s]
82
  47%|β–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 78/165 [00:00<00:00, 388.01it/s]
83
  72%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 118/165 [00:00<00:00, 390.90it/s]
84
  0%| | 0/165 [00:00<?, ?it/s]
85
  96%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ| 158/165 [00:00<00:00, 392.38it/s]
86
+
87
  24%|β–ˆβ–ˆβ–Ž | 39/165 [00:00<00:00, 386.65it/s]
88
  47%|β–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 78/165 [00:00<00:00, 386.51it/s]
89
  72%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 118/165 [00:00<00:00, 388.81it/s]
90
  96%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ| 158/165 [00:00<00:00, 390.15it/s]
91
+
92
+ 2025-11-18 21:17:57.690 | WARNING | src.generation:generate:53 - The arguments ('sigma', 'eot_token_id', 'add_bos_token') are not supported by the generation strategy 'vanilla'.
93
+ 2025-11-18 21:17:57.690 | WARNING | src.generation:generate:53 - The arguments ('add_bos_token', 'eot_token_id', 'sigma') are not supported by the generation strategy 'vanilla'.
94
+ 2025-11-18 21:17:57.691 | WARNING | src.generation:generate:53 - The arguments ('eot_token_id', 'add_bos_token', 'sigma') are not supported by the generation strategy 'vanilla'.
95
+ 2025-11-18 21:17:57.692 | WARNING | src.generation:generate:53 - The arguments ('eot_token_id', 'sigma', 'add_bos_token') are not supported by the generation strategy 'vanilla'.
96
+ 2025-11-18 21:17:57.694 | WARNING | src.generation:generate:53 - The arguments ('add_bos_token', 'eot_token_id', 'sigma') are not supported by the generation strategy 'vanilla'.
97
+ 2025-11-18 21:17:57.697 | WARNING | src.generation:generate:53 - The arguments ('add_bos_token', 'eot_token_id', 'sigma') are not supported by the generation strategy 'vanilla'.
98
+ 2025-11-18 21:17:57.700 | WARNING | src.generation:generate:53 - The arguments ('eot_token_id', 'add_bos_token', 'sigma') are not supported by the generation strategy 'vanilla'.
99
+
100
+ 2025-11-19 01:01:00.021 | INFO | __main__:main:81 - Throughput: 1.63 tokens/sec, Tokens per step: 0.96 tokens/step (full: 12.98 tokens/sec, 1.00 tokens/step), Latency: 79.11 s, Total time: 12965.69 s
101
+ 2025-11-19 01:01:00.135 | INFO | __main__:main:98 - Results saved to /xfr_ceph_sh/liuchonghan/HEAT/heat/outputs/2025-11-18/exp1_len1024/dream-inst/dllm/gsm8k/results.json
102
+ 2025-11-19 01:01:00.135 | INFO | __main__:main:101 - eval time: 12965.69 seconds
103
+ [rank0]:[W1119 01:01:00.284901363 ProcessGroupNCCL.cpp:1479] Warning: WARNING: destroy_process_group() was not called before program exit, which can leak resources. For more info, please see https://pytorch.org/docs/stable/distributed.html#shutdown (function operator())
outputs/2025-11-18/exp1_len1024/dream-inst/dllm/humaneval/.hydra/config.yaml ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ generation:
3
+ mask_token_id: 151666
4
+ eot_token_id: 151643
5
+ pad_token_id: 151643
6
+ add_bos_token: true
7
+ alg: maskgit_plus
8
+ name: dream-inst
9
+ path: ${oc.env:DREAM_INST_PATH}
10
+ generation:
11
+ strategy: vanilla
12
+ threshold: null
13
+ factor: null
14
+ alg: maskgit_plus
15
+ gen_length: 1024
16
+ block_length: 32
17
+ steps: 1024
18
+ temperature: 0.0
19
+ top_p: null
20
+ top_k: null
21
+ debias: false
22
+ output_probs: false
23
+ cache:
24
+ _target_: src.cache.dLLMCache
25
+ kr: 1
26
+ kp: 50
27
+ rou: 0.25
28
+ seed: 1234
29
+ batch_size: 1
30
+ attn_implementation: eager
31
+ dataset:
32
+ name: humaneval
33
+ size: null
34
+ n_shot: null
35
+ system_prompt: null
36
+ batch_size: 1
37
+ mc_num: null
38
+ max_length: 4096
39
+ is_check_greedy: true
40
+ add_bos_token: true
41
+ nll_type: mc
42
+ log_type: ftb
43
+ eval_args:
44
+ log_samples: true
45
+ tasks: ${..dataset.name}
46
+ num_fewshot: ${..dataset.n_shot}
47
+ batch_size: ${..batch_size}
48
+ limit: ${..dataset.size}
49
+ confirm_run_unsafe_code: true
50
+ random_seed: ${..seed}
51
+ fewshot_random_seed: ${..seed}
52
+ numpy_random_seed: ${..seed}
53
+ torch_random_seed: ${..seed}
outputs/2025-11-18/exp1_len1024/dream-inst/dllm/humaneval/.hydra/hydra.yaml ADDED
@@ -0,0 +1,167 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/2025-11-18/exp1_len1024/dream-inst/dllm/humaneval
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.run.dir=outputs/2025-11-18/exp1_len1024/dream-inst/dllm/humaneval
114
+ - hydra.mode=RUN
115
+ task:
116
+ - dataset.name=humaneval
117
+ - model=dream-inst
118
+ - cache=dllm
119
+ - generation=vanilla
120
+ - generation.gen_length=1024
121
+ - generation.steps=1024
122
+ - generation.block_length=32
123
+ - batch_size=1
124
+ - seed=1234
125
+ job:
126
+ name: eval
127
+ chdir: null
128
+ override_dirname: batch_size=1,cache=dllm,dataset.name=humaneval,generation.block_length=32,generation.gen_length=1024,generation.steps=1024,generation=vanilla,model=dream-inst,seed=1234
129
+ id: ???
130
+ num: ???
131
+ config_name: eval
132
+ env_set: {}
133
+ env_copy: []
134
+ config:
135
+ override_dirname:
136
+ kv_sep: '='
137
+ item_sep: ','
138
+ exclude_keys: []
139
+ runtime:
140
+ version: 1.3.2
141
+ version_base: '1.3'
142
+ cwd: /xfr_ceph_sh/liuchonghan/HEAT/heat
143
+ config_sources:
144
+ - path: hydra.conf
145
+ schema: pkg
146
+ provider: hydra
147
+ - path: /xfr_ceph_sh/liuchonghan/HEAT/heat/configs
148
+ schema: file
149
+ provider: main
150
+ - path: ''
151
+ schema: structured
152
+ provider: schema
153
+ output_dir: /xfr_ceph_sh/liuchonghan/HEAT/heat/outputs/2025-11-18/exp1_len1024/dream-inst/dllm/humaneval
154
+ choices:
155
+ cache: dllm
156
+ generation: vanilla
157
+ model: dream-inst
158
+ hydra/env: default
159
+ hydra/callbacks: null
160
+ hydra/job_logging: default
161
+ hydra/hydra_logging: default
162
+ hydra/hydra_help: default
163
+ hydra/help: default
164
+ hydra/sweeper: basic
165
+ hydra/launcher: basic
166
+ hydra/output: default
167
+ verbose: false
outputs/2025-11-18/exp1_len1024/dream-inst/dllm/humaneval/.hydra/overrides.yaml ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ - dataset.name=humaneval
2
+ - model=dream-inst
3
+ - cache=dllm
4
+ - generation=vanilla
5
+ - generation.gen_length=1024
6
+ - generation.steps=1024
7
+ - generation.block_length=32
8
+ - batch_size=1
9
+ - seed=1234
outputs/2025-11-18/exp1_len1024/dream-inst/dllm/humaneval/eval.log ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-11-18 11:28:45,856][accelerate.utils.other][WARNING] - Detected kernel version 5.4.0, which is below the recommended minimum of 5.5.0; this can cause the process to hang. It is recommended to upgrade the kernel to the minimum version or higher.
2
+ [2025-11-18 11:29:09,035][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
3
+ [2025-11-18 11:29:09,035][lm_eval.evaluator][INFO] - Using pre-initialized model
4
+ [2025-11-18 11:29:09,290][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
5
+ [2025-11-18 11:29:09,290][lm_eval.evaluator][INFO] - Using pre-initialized model
6
+ [2025-11-18 11:29:09,542][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
7
+ [2025-11-18 11:29:09,542][lm_eval.evaluator][INFO] - Using pre-initialized model
8
+ [2025-11-18 11:29:09,655][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
9
+ [2025-11-18 11:29:09,655][lm_eval.evaluator][INFO] - Using pre-initialized model
10
+ [2025-11-18 11:29:10,420][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
11
+ [2025-11-18 11:29:10,420][lm_eval.evaluator][INFO] - Using pre-initialized model
12
+ [2025-11-18 11:29:11,324][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
13
+ [2025-11-18 11:29:11,325][lm_eval.evaluator][INFO] - Using pre-initialized model
14
+ [2025-11-18 11:29:11,664][lm_eval.evaluator][INFO] - humaneval: Using gen_kwargs: {'until': ['\nclass', '\ndef', '\n#', '\nif', '\nprint'], 'max_gen_toks': 1024, 'do_sample': False}
15
+ [2025-11-18 11:29:11,665][lm_eval.api.task][INFO] - Building contexts for humaneval on rank 5...
16
+ [2025-11-18 11:29:11,944][lm_eval.evaluator][INFO] - humaneval: Using gen_kwargs: {'until': ['\nclass', '\ndef', '\n#', '\nif', '\nprint'], 'max_gen_toks': 1024, 'do_sample': False}
17
+ [2025-11-18 11:29:11,946][lm_eval.api.task][INFO] - Building contexts for humaneval on rank 4...
18
+ [2025-11-18 11:29:12,224][lm_eval.evaluator][INFO] - humaneval: Using gen_kwargs: {'until': ['\nclass', '\ndef', '\n#', '\nif', '\nprint'], 'max_gen_toks': 1024, 'do_sample': False}
19
+ [2025-11-18 11:29:12,226][lm_eval.api.task][INFO] - Building contexts for humaneval on rank 6...
20
+ [2025-11-18 11:29:12,299][lm_eval.evaluator][INFO] - humaneval: Using gen_kwargs: {'until': ['\nclass', '\ndef', '\n#', '\nif', '\nprint'], 'max_gen_toks': 1024, 'do_sample': False}
21
+ [2025-11-18 11:29:12,300][lm_eval.api.task][INFO] - Building contexts for humaneval on rank 2...
22
+ [2025-11-18 11:29:12,987][lm_eval.evaluator][INFO] - humaneval: Using gen_kwargs: {'until': ['\nclass', '\ndef', '\n#', '\nif', '\nprint'], 'max_gen_toks': 1024, 'do_sample': False}
23
+ [2025-11-18 11:29:12,988][lm_eval.api.task][INFO] - Building contexts for humaneval on rank 1...
24
+ [2025-11-18 11:29:13,185][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
25
+ [2025-11-18 11:29:13,185][lm_eval.evaluator][INFO] - Using pre-initialized model
26
+ [2025-11-18 11:29:13,388][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
27
+ [2025-11-18 11:29:13,388][lm_eval.evaluator][INFO] - Using pre-initialized model
28
+ [2025-11-18 11:29:13,802][lm_eval.evaluator][INFO] - humaneval: Using gen_kwargs: {'until': ['\nclass', '\ndef', '\n#', '\nif', '\nprint'], 'max_gen_toks': 1024, 'do_sample': False}
29
+ [2025-11-18 11:29:13,803][lm_eval.api.task][INFO] - Building contexts for humaneval on rank 0...
30
+ [2025-11-18 11:29:15,749][lm_eval.evaluator][INFO] - humaneval: Using gen_kwargs: {'until': ['\nclass', '\ndef', '\n#', '\nif', '\nprint'], 'max_gen_toks': 1024, 'do_sample': False}
31
+ [2025-11-18 11:29:15,750][lm_eval.api.task][INFO] - Building contexts for humaneval on rank 7...
32
+ [2025-11-18 11:29:16,000][lm_eval.evaluator][INFO] - humaneval: Using gen_kwargs: {'until': ['\nclass', '\ndef', '\n#', '\nif', '\nprint'], 'max_gen_toks': 1024, 'do_sample': False}
33
+ [2025-11-18 11:29:16,001][lm_eval.api.task][INFO] - Building contexts for humaneval on rank 3...
34
+ [2025-11-18 11:29:19,932][lm_eval.evaluator][INFO] - Running generate_until requests
35
+ [2025-11-18 11:29:19,932][lm_eval.evaluator][INFO] - Running generate_until requests
36
+ [2025-11-18 11:29:19,932][lm_eval.evaluator][INFO] - Running generate_until requests
37
+ [2025-11-18 11:29:19,932][lm_eval.evaluator][INFO] - Running generate_until requests
38
+ [2025-11-18 11:29:19,933][lm_eval.evaluator][INFO] - Running generate_until requests
39
+ [2025-11-18 11:29:19,933][lm_eval.evaluator][INFO] - Running generate_until requests
40
+ [2025-11-18 11:29:19,933][lm_eval.evaluator][INFO] - Running generate_until requests
41
+ [2025-11-18 11:29:19,934][lm_eval.evaluator][INFO] - Running generate_until requests
outputs/2025-11-18/exp1_len1024/dream-inst/dllm/humaneval/results.json ADDED
The diff for this file is too large to render. See raw diff
 
outputs/2025-11-18/exp1_len1024/dream-inst/dllm/humaneval/stderr.log ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0
  0%| | 0/20 [00:00<?, ?it/s]
 
1
  0%| | 0/20 [00:00<?, ?it/s]
 
2
  0%| | 0/20 [00:00<?, ?it/s]
 
3
  0%| | 0/21 [00:00<?, ?it/s]
 
4
  0%| | 0/21 [00:00<?, ?it/s]
 
5
  0%| | 0/21 [00:00<?, ?it/s]
 
6
  0%| | 0/20 [00:00<?, ?it/s]
 
7
  0%| | 0/21 [00:00<?, ?it/s]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ipex flag is deprecated, will be removed in Accelerate v1.10. From 2.7.0, PyTorch has all needed optimizations for Intel CPU and XPU.
2
+ The following values were not passed to `accelerate launch` and had defaults used instead:
3
+ More than one GPU was found, enabling multi-GPU training.
4
+ If this was unintended please pass in `--num_processes=1`.
5
+ `--mixed_precision` was set to a value of `'no'`
6
+ `--dynamo_backend` was set to a value of `'no'`
7
+ To avoid this warning pass in values for each of the problematic parameters or run `accelerate config`.
8
+ [W1118 11:28:30.873002893 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
9
+ [W1118 11:28:39.149288617 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
10
+ 2025-11-18 11:28:39.480 | INFO | src.utils:pre_initialize:603 - {'strategy': 'vanilla', 'threshold': None, 'factor': None, 'alg': 'maskgit_plus', 'gen_length': 1024, 'block_length': 32, 'steps': 1024, 'temperature': 0.0, 'top_p': 0.9, 'top_k': None, 'debias': False, 'output_probs': False, 'mask_token_id': 151666, 'eot_token_id': 151643, 'pad_token_id': 151643, 'add_bos_token': True, 'sigma': None}
11
+ 2025-11-18 11:28:39.481 | INFO | src.utils:pre_initialize:618 - Using cache with args: {'kp': 50, 'kr': 1}
12
+ [W1118 11:28:40.236613715 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
13
+ [W1118 11:28:42.690415311 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
14
+ [W1118 11:28:43.286219076 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
15
+ [W1118 11:28:43.303679181 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
16
+ [W1118 11:28:43.346367847 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
17
+ [W1118 11:28:43.441263228 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
18
+ [W1118 11:28:43.473438887 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
19
+
20
+
21
+
22
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
23
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
24
+
25
+
26
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
27
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
28
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
29
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
30
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
31
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
32
+
33
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
34
+
35
+
36
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
37
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
38
+
39
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
40
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
41
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
42
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
43
+
44
  0%| | 0/20 [00:00<?, ?it/s]
45
+
46
  0%| | 0/20 [00:00<?, ?it/s]
47
+
48
  0%| | 0/20 [00:00<?, ?it/s]
49
+
50
  0%| | 0/21 [00:00<?, ?it/s]
51
+
52
  0%| | 0/21 [00:00<?, ?it/s]
53
+
54
  0%| | 0/21 [00:00<?, ?it/s]
55
+
56
  0%| | 0/20 [00:00<?, ?it/s]
57
+
58
  0%| | 0/21 [00:00<?, ?it/s]
59
+
60
+ 2025-11-18 11:29:19.940 | WARNING | src.generation:generate:53 - The arguments ('sigma', 'eot_token_id', 'add_bos_token') are not supported by the generation strategy 'vanilla'.
61
+ 2025-11-18 11:29:19.940 | WARNING | src.generation:generate:53 - The arguments ('sigma', 'add_bos_token', 'eot_token_id') are not supported by the generation strategy 'vanilla'.
62
+ 2025-11-18 11:29:19.941 | WARNING | src.generation:generate:53 - The arguments ('sigma', 'add_bos_token', 'eot_token_id') are not supported by the generation strategy 'vanilla'.
63
+ 2025-11-18 11:29:19.942 | WARNING | src.generation:generate:53 - The arguments ('eot_token_id', 'add_bos_token', 'sigma') are not supported by the generation strategy 'vanilla'.
64
+ 2025-11-18 11:29:19.944 | WARNING | src.generation:generate:53 - The arguments ('sigma', 'add_bos_token', 'eot_token_id') are not supported by the generation strategy 'vanilla'.
65
+ 2025-11-18 11:29:19.949 | WARNING | src.generation:generate:53 - The arguments ('eot_token_id', 'sigma', 'add_bos_token') are not supported by the generation strategy 'vanilla'.
66
+ 2025-11-18 11:29:19.950 | WARNING | src.generation:generate:53 - The arguments ('sigma', 'eot_token_id', 'add_bos_token') are not supported by the generation strategy 'vanilla'.
67
+
68
+ 2025-11-18 11:56:59.412 | INFO | __main__:main:81 - Throughput: 2.40 tokens/sec, Tokens per step: 0.94 tokens/step (full: 13.58 tokens/sec, 1.00 tokens/step), Latency: 75.51 s, Total time: 1593.89 s
69
+ 2025-11-18 11:56:59.424 | INFO | __main__:main:98 - Results saved to /xfr_ceph_sh/liuchonghan/HEAT/heat/outputs/2025-11-18/exp1_len1024/dream-inst/dllm/humaneval/results.json
70
+ 2025-11-18 11:56:59.424 | INFO | __main__:main:101 - eval time: 1593.89 seconds
71
+ [rank0]:[W1118 11:56:59.614479582 ProcessGroupNCCL.cpp:1479] Warning: WARNING: destroy_process_group() was not called before program exit, which can leak resources. For more info, please see https://pytorch.org/docs/stable/distributed.html#shutdown (function operator())
outputs/2025-11-18/exp1_len1024/dream-inst/dllm/math-500/.hydra/config.yaml ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ generation:
3
+ mask_token_id: 151666
4
+ eot_token_id: 151643
5
+ pad_token_id: 151643
6
+ add_bos_token: true
7
+ alg: maskgit_plus
8
+ name: dream-inst
9
+ path: ${oc.env:DREAM_INST_PATH}
10
+ generation:
11
+ strategy: vanilla
12
+ threshold: null
13
+ factor: null
14
+ alg: maskgit_plus
15
+ gen_length: 1024
16
+ block_length: 32
17
+ steps: 1024
18
+ temperature: 0.0
19
+ top_p: null
20
+ top_k: null
21
+ debias: false
22
+ output_probs: false
23
+ cache:
24
+ _target_: src.cache.dLLMCache
25
+ kr: 1
26
+ kp: 50
27
+ rou: 0.25
28
+ seed: 1234
29
+ batch_size: 1
30
+ attn_implementation: eager
31
+ dataset:
32
+ name: math-500
33
+ size: null
34
+ n_shot: null
35
+ system_prompt: null
36
+ batch_size: 1
37
+ mc_num: null
38
+ max_length: 4096
39
+ is_check_greedy: true
40
+ add_bos_token: true
41
+ nll_type: mc
42
+ log_type: ftb
43
+ eval_args:
44
+ log_samples: true
45
+ tasks: ${..dataset.name}
46
+ num_fewshot: ${..dataset.n_shot}
47
+ batch_size: ${..batch_size}
48
+ limit: ${..dataset.size}
49
+ confirm_run_unsafe_code: true
50
+ random_seed: ${..seed}
51
+ fewshot_random_seed: ${..seed}
52
+ numpy_random_seed: ${..seed}
53
+ torch_random_seed: ${..seed}
outputs/2025-11-18/exp1_len1024/dream-inst/dllm/math-500/.hydra/hydra.yaml ADDED
@@ -0,0 +1,167 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/2025-11-18/exp1_len1024/dream-inst/dllm/math-500
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.run.dir=outputs/2025-11-18/exp1_len1024/dream-inst/dllm/math-500
114
+ - hydra.mode=RUN
115
+ task:
116
+ - dataset.name=math-500
117
+ - model=dream-inst
118
+ - cache=dllm
119
+ - generation=vanilla
120
+ - generation.gen_length=1024
121
+ - generation.steps=1024
122
+ - generation.block_length=32
123
+ - batch_size=1
124
+ - seed=1234
125
+ job:
126
+ name: eval
127
+ chdir: null
128
+ override_dirname: batch_size=1,cache=dllm,dataset.name=math-500,generation.block_length=32,generation.gen_length=1024,generation.steps=1024,generation=vanilla,model=dream-inst,seed=1234
129
+ id: ???
130
+ num: ???
131
+ config_name: eval
132
+ env_set: {}
133
+ env_copy: []
134
+ config:
135
+ override_dirname:
136
+ kv_sep: '='
137
+ item_sep: ','
138
+ exclude_keys: []
139
+ runtime:
140
+ version: 1.3.2
141
+ version_base: '1.3'
142
+ cwd: /xfr_ceph_sh/liuchonghan/HEAT/heat
143
+ config_sources:
144
+ - path: hydra.conf
145
+ schema: pkg
146
+ provider: hydra
147
+ - path: /xfr_ceph_sh/liuchonghan/HEAT/heat/configs
148
+ schema: file
149
+ provider: main
150
+ - path: ''
151
+ schema: structured
152
+ provider: schema
153
+ output_dir: /xfr_ceph_sh/liuchonghan/HEAT/heat/outputs/2025-11-18/exp1_len1024/dream-inst/dllm/math-500
154
+ choices:
155
+ cache: dllm
156
+ generation: vanilla
157
+ model: dream-inst
158
+ hydra/env: default
159
+ hydra/callbacks: null
160
+ hydra/job_logging: default
161
+ hydra/hydra_logging: default
162
+ hydra/hydra_help: default
163
+ hydra/help: default
164
+ hydra/sweeper: basic
165
+ hydra/launcher: basic
166
+ hydra/output: default
167
+ verbose: false
outputs/2025-11-18/exp1_len1024/dream-inst/dllm/math-500/.hydra/overrides.yaml ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ - dataset.name=math-500
2
+ - model=dream-inst
3
+ - cache=dllm
4
+ - generation=vanilla
5
+ - generation.gen_length=1024
6
+ - generation.steps=1024
7
+ - generation.block_length=32
8
+ - batch_size=1
9
+ - seed=1234
outputs/2025-11-18/exp1_len1024/dream-inst/dllm/math-500/eval.log ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-11-19 11:50:26,990][accelerate.utils.other][WARNING] - Detected kernel version 5.4.0, which is below the recommended minimum of 5.5.0; this can cause the process to hang. It is recommended to upgrade the kernel to the minimum version or higher.
2
+ [2025-11-19 11:50:38,980][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
3
+ [2025-11-19 11:50:38,980][lm_eval.evaluator][INFO] - Using pre-initialized model
4
+ [2025-11-19 11:50:39,186][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
5
+ [2025-11-19 11:50:39,186][lm_eval.evaluator][INFO] - Using pre-initialized model
6
+ [2025-11-19 11:50:39,639][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
7
+ [2025-11-19 11:50:39,640][lm_eval.evaluator][INFO] - Using pre-initialized model
8
+ [2025-11-19 11:50:39,771][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
9
+ [2025-11-19 11:50:39,771][lm_eval.evaluator][INFO] - Using pre-initialized model
10
+ [2025-11-19 11:50:39,881][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
11
+ [2025-11-19 11:50:39,881][lm_eval.evaluator][INFO] - Using pre-initialized model
12
+ [2025-11-19 11:50:39,888][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
13
+ [2025-11-19 11:50:39,888][lm_eval.evaluator][INFO] - Using pre-initialized model
14
+ [2025-11-19 11:50:40,118][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
15
+ [2025-11-19 11:50:40,118][lm_eval.evaluator][INFO] - Using pre-initialized model
16
+ [2025-11-19 11:50:40,213][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
17
+ [2025-11-19 11:50:40,214][lm_eval.evaluator][INFO] - Using pre-initialized model
18
+ [2025-11-19 11:50:44,328][lm_eval.evaluator][INFO] - math-500: Using gen_kwargs: {'until': ['Problem:'], 'do_sample': False, 'temperature': 0.0}
19
+ [2025-11-19 11:50:44,331][lm_eval.api.task][INFO] - Building contexts for math-500 on rank 1...
20
+ [2025-11-19 11:50:44,654][lm_eval.evaluator][INFO] - math-500: Using gen_kwargs: {'until': ['Problem:'], 'do_sample': False, 'temperature': 0.0}
21
+ [2025-11-19 11:50:44,655][lm_eval.api.task][INFO] - Building contexts for math-500 on rank 0...
22
+ [2025-11-19 11:50:44,984][lm_eval.evaluator][INFO] - math-500: Using gen_kwargs: {'until': ['Problem:'], 'do_sample': False, 'temperature': 0.0}
23
+ [2025-11-19 11:50:44,988][lm_eval.api.task][INFO] - Building contexts for math-500 on rank 7...
24
+ [2025-11-19 11:50:45,168][lm_eval.evaluator][INFO] - math-500: Using gen_kwargs: {'until': ['Problem:'], 'do_sample': False, 'temperature': 0.0}
25
+ [2025-11-19 11:50:45,170][lm_eval.api.task][INFO] - Building contexts for math-500 on rank 3...
26
+ [2025-11-19 11:50:45,328][lm_eval.evaluator][INFO] - math-500: Using gen_kwargs: {'until': ['Problem:'], 'do_sample': False, 'temperature': 0.0}
27
+ [2025-11-19 11:50:45,330][lm_eval.api.task][INFO] - Building contexts for math-500 on rank 2...
28
+ [2025-11-19 11:50:45,344][lm_eval.evaluator][INFO] - math-500: Using gen_kwargs: {'until': ['Problem:'], 'do_sample': False, 'temperature': 0.0}
29
+ [2025-11-19 11:50:45,345][lm_eval.api.task][INFO] - Building contexts for math-500 on rank 6...
30
+ [2025-11-19 11:50:45,439][lm_eval.evaluator][INFO] - math-500: Using gen_kwargs: {'until': ['Problem:'], 'do_sample': False, 'temperature': 0.0}
31
+ [2025-11-19 11:50:45,442][lm_eval.api.task][INFO] - Building contexts for math-500 on rank 5...
32
+ [2025-11-19 11:50:45,451][lm_eval.evaluator][INFO] - math-500: Using gen_kwargs: {'until': ['Problem:'], 'do_sample': False, 'temperature': 0.0}
33
+ [2025-11-19 11:50:45,452][lm_eval.api.task][INFO] - Building contexts for math-500 on rank 4...
34
+ [2025-11-19 11:50:49,775][lm_eval.evaluator][INFO] - Running generate_until requests
35
+ [2025-11-19 11:50:49,775][lm_eval.evaluator][INFO] - Running generate_until requests
36
+ [2025-11-19 11:50:49,775][lm_eval.evaluator][INFO] - Running generate_until requests
37
+ [2025-11-19 11:50:49,775][lm_eval.evaluator][INFO] - Running generate_until requests
38
+ [2025-11-19 11:50:49,775][lm_eval.evaluator][INFO] - Running generate_until requests
39
+ [2025-11-19 11:50:49,775][lm_eval.evaluator][INFO] - Running generate_until requests
40
+ [2025-11-19 11:50:49,775][lm_eval.evaluator][INFO] - Running generate_until requests
41
+ [2025-11-19 11:50:49,775][lm_eval.evaluator][INFO] - Running generate_until requests
outputs/2025-11-18/exp1_len1024/dream-inst/dllm/math-500/results.json ADDED
The diff for this file is too large to render. See raw diff
 
outputs/2025-11-18/exp1_len1024/dream-inst/dllm/math-500/stderr.log ADDED
@@ -0,0 +1,103 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0
  0%| | 0/63 [00:00<?, ?it/s]
1
  79%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 50/63 [00:00<00:00, 498.82it/s]
 
2
  0%| | 0/63 [00:00<?, ?it/s]
3
  81%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 51/63 [00:00<00:00, 494.69it/s]
 
4
  0%| | 0/62 [00:00<?, ?it/s]
5
  53%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 33/62 [00:00<00:00, 319.86it/s]
6
  0%| | 0/63 [00:00<?, ?it/s]
 
7
  68%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 43/63 [00:00<00:00, 414.51it/s]
8
  0%| | 0/63 [00:00<?, ?it/s]
9
  0%| | 0/62 [00:00<?, ?it/s]
 
10
  86%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 54/63 [00:00<00:00, 537.73it/s]
11
  0%| | 0/62 [00:00<?, ?it/s]
12
  90%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 56/62 [00:00<00:00, 550.59it/s]
 
13
  0%| | 0/62 [00:00<?, ?it/s]
 
14
  89%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 55/62 [00:00<00:00, 548.88it/s]
15
  90%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 56/62 [00:00<00:00, 549.94it/s]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ipex flag is deprecated, will be removed in Accelerate v1.10. From 2.7.0, PyTorch has all needed optimizations for Intel CPU and XPU.
2
+ The following values were not passed to `accelerate launch` and had defaults used instead:
3
+ More than one GPU was found, enabling multi-GPU training.
4
+ If this was unintended please pass in `--num_processes=1`.
5
+ `--mixed_precision` was set to a value of `'no'`
6
+ `--dynamo_backend` was set to a value of `'no'`
7
+ To avoid this warning pass in values for each of the problematic parameters or run `accelerate config`.
8
+ [W1119 11:50:12.876749115 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
9
+ 2025-11-19 11:50:20.128 | INFO | src.utils:pre_initialize:603 - {'strategy': 'vanilla', 'threshold': None, 'factor': None, 'alg': 'maskgit_plus', 'gen_length': 1024, 'block_length': 32, 'steps': 1024, 'temperature': 0.0, 'top_p': 0.9, 'top_k': None, 'debias': False, 'output_probs': False, 'mask_token_id': 151666, 'eot_token_id': 151643, 'pad_token_id': 151643, 'add_bos_token': True, 'sigma': None}
10
+ 2025-11-19 11:50:20.129 | INFO | src.utils:pre_initialize:618 - Using cache with args: {'kp': 50, 'kr': 1}
11
+ [W1119 11:50:20.738238144 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
12
+ [W1119 11:50:24.879620258 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
13
+ [W1119 11:50:24.888070672 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
14
+ [W1119 11:50:24.180661413 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
15
+ [W1119 11:50:24.384845106 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
16
+ [W1119 11:50:24.449183109 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
17
+ [W1119 11:50:24.576892757 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
18
+ [W1119 11:50:25.843752097 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
19
+
20
+
21
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
22
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
23
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
24
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
25
+
26
+
27
+
28
+
29
+
30
+
31
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
32
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
33
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
34
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
35
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
36
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
37
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
38
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
39
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
40
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
41
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
42
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
43
+ 2025-11-19 11:50:34.709 | INFO | src.utils:sympy_antlr_patcher:462 - Detected antlr4-python3-runtime version 4.9.3. Temporarily switching to 4.11.0...
44
+ 2025-11-19 11:50:34.714 | INFO | src.utils:sympy_antlr_patcher:473 - Downloading antlr4-python3-runtime==4.11.0 to /tmp/isolated_antlr_2s57y1bk...
45
+ 2025-11-19 11:50:34.877 | INFO | src.utils:sympy_antlr_patcher:462 - Detected antlr4-python3-runtime version 4.9.3. Temporarily switching to 4.11.0...
46
+ 2025-11-19 11:50:34.890 | INFO | src.utils:sympy_antlr_patcher:473 - Downloading antlr4-python3-runtime==4.11.0 to /tmp/isolated_antlr_aos902xl...
47
+ 2025-11-19 11:50:34.944 | INFO | src.utils:sympy_antlr_patcher:462 - Detected antlr4-python3-runtime version 4.9.3. Temporarily switching to 4.11.0...
48
+ 2025-11-19 11:50:34.950 | INFO | src.utils:sympy_antlr_patcher:473 - Downloading antlr4-python3-runtime==4.11.0 to /tmp/isolated_antlr_9dpjelay...
49
+ 2025-11-19 11:50:35.294 | INFO | src.utils:sympy_antlr_patcher:462 - Detected antlr4-python3-runtime version 4.9.3. Temporarily switching to 4.11.0...
50
+ 2025-11-19 11:50:35.299 | INFO | src.utils:sympy_antlr_patcher:473 - Downloading antlr4-python3-runtime==4.11.0 to /tmp/isolated_antlr_tljtvj6v...
51
+ 2025-11-19 11:50:35.406 | INFO | src.utils:sympy_antlr_patcher:462 - Detected antlr4-python3-runtime version 4.9.3. Temporarily switching to 4.11.0...
52
+ 2025-11-19 11:50:35.414 | INFO | src.utils:sympy_antlr_patcher:462 - Detected antlr4-python3-runtime version 4.9.3. Temporarily switching to 4.11.0...
53
+ 2025-11-19 11:50:35.418 | INFO | src.utils:sympy_antlr_patcher:473 - Downloading antlr4-python3-runtime==4.11.0 to /tmp/isolated_antlr_vn27d0lc...
54
+ 2025-11-19 11:50:35.422 | INFO | src.utils:sympy_antlr_patcher:473 - Downloading antlr4-python3-runtime==4.11.0 to /tmp/isolated_antlr_ozxomwwo...
55
+ 2025-11-19 11:50:35.471 | INFO | src.utils:sympy_antlr_patcher:462 - Detected antlr4-python3-runtime version 4.9.3. Temporarily switching to 4.11.0...
56
+ 2025-11-19 11:50:35.476 | INFO | src.utils:sympy_antlr_patcher:473 - Downloading antlr4-python3-runtime==4.11.0 to /tmp/isolated_antlr_98204mei...
57
+ 2025-11-19 11:50:35.631 | INFO | src.utils:sympy_antlr_patcher:500 - Unpacking antlr4_python3_runtime-4.11.0-py3-none-any.whl...
58
+ 2025-11-19 11:50:35.772 | INFO | src.utils:sympy_antlr_patcher:500 - Unpacking antlr4_python3_runtime-4.11.0-py3-none-any.whl...
59
+ 2025-11-19 11:50:35.846 | INFO | src.utils:sympy_antlr_patcher:500 - Unpacking antlr4_python3_runtime-4.11.0-py3-none-any.whl...
60
+ 2025-11-19 11:50:35.916 | INFO | src.utils:sympy_antlr_patcher:462 - Detected antlr4-python3-runtime version 4.9.3. Temporarily switching to 4.11.0...
61
+ 2025-11-19 11:50:35.920 | INFO | src.utils:sympy_antlr_patcher:473 - Downloading antlr4-python3-runtime==4.11.0 to /tmp/isolated_antlr_456oqdwt...
62
+ 2025-11-19 11:50:36.233 | INFO | src.utils:sympy_antlr_patcher:500 - Unpacking antlr4_python3_runtime-4.11.0-py3-none-any.whl...
63
+ 2025-11-19 11:50:36.303 | INFO | src.utils:sympy_antlr_patcher:500 - Unpacking antlr4_python3_runtime-4.11.0-py3-none-any.whl...
64
+ 2025-11-19 11:50:36.343 | INFO | src.utils:sympy_antlr_patcher:500 - Unpacking antlr4_python3_runtime-4.11.0-py3-none-any.whl...
65
+ 2025-11-19 11:50:36.364 | INFO | src.utils:sympy_antlr_patcher:500 - Unpacking antlr4_python3_runtime-4.11.0-py3-none-any.whl...
66
+ 2025-11-19 11:50:36.783 | INFO | src.utils:sympy_antlr_patcher:500 - Unpacking antlr4_python3_runtime-4.11.0-py3-none-any.whl...
67
+
68
  0%| | 0/63 [00:00<?, ?it/s]
69
  79%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 50/63 [00:00<00:00, 498.82it/s]
70
+
71
  0%| | 0/63 [00:00<?, ?it/s]
72
  81%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 51/63 [00:00<00:00, 494.69it/s]
73
+
74
  0%| | 0/62 [00:00<?, ?it/s]
75
  53%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 33/62 [00:00<00:00, 319.86it/s]
76
  0%| | 0/63 [00:00<?, ?it/s]
77
+
78
  68%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 43/63 [00:00<00:00, 414.51it/s]
79
  0%| | 0/63 [00:00<?, ?it/s]
80
  0%| | 0/62 [00:00<?, ?it/s]
81
+
82
  86%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 54/63 [00:00<00:00, 537.73it/s]
83
  0%| | 0/62 [00:00<?, ?it/s]
84
  90%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 56/62 [00:00<00:00, 550.59it/s]
85
+
86
  0%| | 0/62 [00:00<?, ?it/s]
87
+
88
  89%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 55/62 [00:00<00:00, 548.88it/s]
89
  90%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 56/62 [00:00<00:00, 549.94it/s]
90
+
91
+
92
+ 2025-11-19 11:50:49.788 | WARNING | src.generation:generate:53 - The arguments ('add_bos_token', 'eot_token_id', 'sigma') are not supported by the generation strategy 'vanilla'.
93
+ 2025-11-19 11:50:49.789 | WARNING | src.generation:generate:53 - The arguments ('sigma', 'eot_token_id', 'add_bos_token') are not supported by the generation strategy 'vanilla'.
94
+ 2025-11-19 11:50:49.789 | WARNING | src.generation:generate:53 - The arguments ('sigma', 'eot_token_id', 'add_bos_token') are not supported by the generation strategy 'vanilla'.
95
+ 2025-11-19 11:50:49.789 | WARNING | src.generation:generate:53 - The arguments ('add_bos_token', 'eot_token_id', 'sigma') are not supported by the generation strategy 'vanilla'.
96
+ 2025-11-19 11:50:49.790 | WARNING | src.generation:generate:53 - The arguments ('sigma', 'eot_token_id', 'add_bos_token') are not supported by the generation strategy 'vanilla'.
97
+ 2025-11-19 11:50:49.790 | WARNING | src.generation:generate:53 - The arguments ('add_bos_token', 'eot_token_id', 'sigma') are not supported by the generation strategy 'vanilla'.
98
+ 2025-11-19 11:50:49.792 | WARNING | src.generation:generate:53 - The arguments ('sigma', 'eot_token_id', 'add_bos_token') are not supported by the generation strategy 'vanilla'.
99
+
100
+ 2025-11-19 13:22:16.866 | INFO | src.utils:sympy_antlr_patcher:513 - Restoring original environment...
101
+ 2025-11-19 13:22:16.866 | INFO | src.utils:sympy_antlr_patcher:513 - Restoring original environment...
102
+ 2025-11-19 13:22:16.868 | INFO | src.utils:sympy_antlr_patcher:513 - Restoring original environment...
103
+ 2025-11-19 13:22:16.870 | INFO | src.utils:sympy_antlr_patcher:513 - Restoring original environment...
104
+ 2025-11-19 13:22:16.870 | INFO | src.utils:sympy_antlr_patcher:513 - Restoring original environment...
105
+ 2025-11-19 13:22:16.870 | INFO | src.utils:sympy_antlr_patcher:513 - Restoring original environment...
106
+ 2025-11-19 13:22:16.872 | INFO | src.utils:sympy_antlr_patcher:513 - Restoring original environment...
107
+ 2025-11-19 13:22:16.874 | INFO | src.utils:sympy_antlr_patcher:522 - Environment restored.
108
+ 2025-11-19 13:22:16.884 | INFO | src.utils:sympy_antlr_patcher:522 - Environment restored.
109
+ 2025-11-19 13:22:16.885 | INFO | src.utils:sympy_antlr_patcher:522 - Environment restored.
110
+ 2025-11-19 13:22:16.887 | INFO | src.utils:sympy_antlr_patcher:522 - Environment restored.
111
+ 2025-11-19 13:22:16.889 | INFO | src.utils:sympy_antlr_patcher:522 - Environment restored.
112
+ 2025-11-19 13:22:16.891 | INFO | src.utils:sympy_antlr_patcher:522 - Environment restored.
113
+ 2025-11-19 13:22:16.892 | INFO | src.utils:sympy_antlr_patcher:522 - Environment restored.
114
+ 2025-11-19 13:22:22.355 | INFO | src.utils:sympy_antlr_patcher:513 - Restoring original environment...
115
+ 2025-11-19 13:22:22.361 | INFO | src.utils:sympy_antlr_patcher:522 - Environment restored.
116
+ 2025-11-19 13:22:22.361 | INFO | __main__:main:81 - Throughput: 11.82 tokens/sec, Tokens per step: 1.00 tokens/step (full: 12.03 tokens/sec, 1.00 tokens/step), Latency: 85.26 s, Total time: 5443.58 s
117
+ 2025-11-19 13:22:22.390 | INFO | __main__:main:98 - Results saved to /xfr_ceph_sh/liuchonghan/HEAT/heat/outputs/2025-11-18/exp1_len1024/dream-inst/dllm/math-500/results.json
118
+ 2025-11-19 13:22:22.390 | INFO | __main__:main:101 - eval time: 5443.58 seconds
119
+ [rank0]:[W1119 13:22:22.537894604 ProcessGroupNCCL.cpp:1479] Warning: WARNING: destroy_process_group() was not called before program exit, which can leak resources. For more info, please see https://pytorch.org/docs/stable/distributed.html#shutdown (function operator())
outputs/2025-11-18/exp1_len1024/dream-inst/dllm/mbpp/.hydra/config.yaml ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ generation:
3
+ mask_token_id: 151666
4
+ eot_token_id: 151643
5
+ pad_token_id: 151643
6
+ add_bos_token: true
7
+ alg: maskgit_plus
8
+ name: dream-inst
9
+ path: ${oc.env:DREAM_INST_PATH}
10
+ generation:
11
+ strategy: vanilla
12
+ threshold: null
13
+ factor: null
14
+ alg: maskgit_plus
15
+ gen_length: 1024
16
+ block_length: 32
17
+ steps: 1024
18
+ temperature: 0.0
19
+ top_p: null
20
+ top_k: null
21
+ debias: false
22
+ output_probs: false
23
+ cache:
24
+ _target_: src.cache.dLLMCache
25
+ kr: 1
26
+ kp: 50
27
+ rou: 0.25
28
+ seed: 1234
29
+ batch_size: 1
30
+ attn_implementation: eager
31
+ dataset:
32
+ name: mbpp
33
+ size: null
34
+ n_shot: null
35
+ system_prompt: null
36
+ batch_size: 1
37
+ mc_num: null
38
+ max_length: 4096
39
+ is_check_greedy: true
40
+ add_bos_token: true
41
+ nll_type: mc
42
+ log_type: ftb
43
+ eval_args:
44
+ log_samples: true
45
+ tasks: ${..dataset.name}
46
+ num_fewshot: ${..dataset.n_shot}
47
+ batch_size: ${..batch_size}
48
+ limit: ${..dataset.size}
49
+ confirm_run_unsafe_code: true
50
+ random_seed: ${..seed}
51
+ fewshot_random_seed: ${..seed}
52
+ numpy_random_seed: ${..seed}
53
+ torch_random_seed: ${..seed}
outputs/2025-11-18/exp1_len1024/dream-inst/dllm/mbpp/.hydra/hydra.yaml ADDED
@@ -0,0 +1,167 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/2025-11-18/exp1_len1024/dream-inst/dllm/mbpp
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.run.dir=outputs/2025-11-18/exp1_len1024/dream-inst/dllm/mbpp
114
+ - hydra.mode=RUN
115
+ task:
116
+ - dataset.name=mbpp
117
+ - model=dream-inst
118
+ - cache=dllm
119
+ - generation=vanilla
120
+ - generation.gen_length=1024
121
+ - generation.steps=1024
122
+ - generation.block_length=32
123
+ - batch_size=1
124
+ - seed=1234
125
+ job:
126
+ name: eval
127
+ chdir: null
128
+ override_dirname: batch_size=1,cache=dllm,dataset.name=mbpp,generation.block_length=32,generation.gen_length=1024,generation.steps=1024,generation=vanilla,model=dream-inst,seed=1234
129
+ id: ???
130
+ num: ???
131
+ config_name: eval
132
+ env_set: {}
133
+ env_copy: []
134
+ config:
135
+ override_dirname:
136
+ kv_sep: '='
137
+ item_sep: ','
138
+ exclude_keys: []
139
+ runtime:
140
+ version: 1.3.2
141
+ version_base: '1.3'
142
+ cwd: /xfr_ceph_sh/liuchonghan/HEAT/heat
143
+ config_sources:
144
+ - path: hydra.conf
145
+ schema: pkg
146
+ provider: hydra
147
+ - path: /xfr_ceph_sh/liuchonghan/HEAT/heat/configs
148
+ schema: file
149
+ provider: main
150
+ - path: ''
151
+ schema: structured
152
+ provider: schema
153
+ output_dir: /xfr_ceph_sh/liuchonghan/HEAT/heat/outputs/2025-11-18/exp1_len1024/dream-inst/dllm/mbpp
154
+ choices:
155
+ cache: dllm
156
+ generation: vanilla
157
+ model: dream-inst
158
+ hydra/env: default
159
+ hydra/callbacks: null
160
+ hydra/job_logging: default
161
+ hydra/hydra_logging: default
162
+ hydra/hydra_help: default
163
+ hydra/help: default
164
+ hydra/sweeper: basic
165
+ hydra/launcher: basic
166
+ hydra/output: default
167
+ verbose: false
outputs/2025-11-18/exp1_len1024/dream-inst/dllm/mbpp/.hydra/overrides.yaml ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ - dataset.name=mbpp
2
+ - model=dream-inst
3
+ - cache=dllm
4
+ - generation=vanilla
5
+ - generation.gen_length=1024
6
+ - generation.steps=1024
7
+ - generation.block_length=32
8
+ - batch_size=1
9
+ - seed=1234
outputs/2025-11-18/exp1_len1024/dream-inst/dllm/mbpp/eval.log ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-11-19 06:39:17,131][accelerate.utils.other][WARNING] - Detected kernel version 5.4.0, which is below the recommended minimum of 5.5.0; this can cause the process to hang. It is recommended to upgrade the kernel to the minimum version or higher.
2
+ [2025-11-19 06:39:24,192][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
3
+ [2025-11-19 06:39:24,192][lm_eval.evaluator][INFO] - Using pre-initialized model
4
+ [2025-11-19 06:39:24,598][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
5
+ [2025-11-19 06:39:24,598][lm_eval.evaluator][INFO] - Using pre-initialized model
6
+ [2025-11-19 06:39:24,635][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
7
+ [2025-11-19 06:39:24,636][lm_eval.evaluator][INFO] - Using pre-initialized model
8
+ [2025-11-19 06:39:24,675][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
9
+ [2025-11-19 06:39:24,675][lm_eval.evaluator][INFO] - Using pre-initialized model
10
+ [2025-11-19 06:39:25,170][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
11
+ [2025-11-19 06:39:25,171][lm_eval.evaluator][INFO] - Using pre-initialized model
12
+ [2025-11-19 06:39:25,174][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
13
+ [2025-11-19 06:39:25,174][lm_eval.evaluator][INFO] - Using pre-initialized model
14
+ [2025-11-19 06:39:25,182][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
15
+ [2025-11-19 06:39:25,183][lm_eval.evaluator][INFO] - Using pre-initialized model
16
+ [2025-11-19 06:39:25,398][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
17
+ [2025-11-19 06:39:25,398][lm_eval.evaluator][INFO] - Using pre-initialized model
18
+ [2025-11-19 06:39:39,093][lm_eval.evaluator][INFO] - mbpp: Using gen_kwargs: {'until': ['[DONE]'], 'do_sample': False}
19
+ [2025-11-19 06:39:39,094][lm_eval.api.task][INFO] - Building contexts for mbpp on rank 4...
20
+ [2025-11-19 06:39:39,887][lm_eval.evaluator][INFO] - mbpp: Using gen_kwargs: {'until': ['[DONE]'], 'do_sample': False}
21
+ [2025-11-19 06:39:39,888][lm_eval.api.task][INFO] - Building contexts for mbpp on rank 1...
22
+ [2025-11-19 06:39:40,208][lm_eval.evaluator][INFO] - mbpp: Using gen_kwargs: {'until': ['[DONE]'], 'do_sample': False}
23
+ [2025-11-19 06:39:40,209][lm_eval.api.task][INFO] - Building contexts for mbpp on rank 5...
24
+ [2025-11-19 06:39:40,371][lm_eval.evaluator][INFO] - mbpp: Using gen_kwargs: {'until': ['[DONE]'], 'do_sample': False}
25
+ [2025-11-19 06:39:40,372][lm_eval.api.task][INFO] - Building contexts for mbpp on rank 7...
26
+ [2025-11-19 06:39:40,544][lm_eval.evaluator][INFO] - mbpp: Using gen_kwargs: {'until': ['[DONE]'], 'do_sample': False}
27
+ [2025-11-19 06:39:40,546][lm_eval.api.task][INFO] - Building contexts for mbpp on rank 6...
28
+ [2025-11-19 06:39:40,705][lm_eval.evaluator][INFO] - mbpp: Using gen_kwargs: {'until': ['[DONE]'], 'do_sample': False}
29
+ [2025-11-19 06:39:40,707][lm_eval.api.task][INFO] - Building contexts for mbpp on rank 3...
30
+ [2025-11-19 06:39:41,241][lm_eval.evaluator][INFO] - mbpp: Using gen_kwargs: {'until': ['[DONE]'], 'do_sample': False}
31
+ [2025-11-19 06:39:41,242][lm_eval.api.task][INFO] - Building contexts for mbpp on rank 0...
32
+ [2025-11-19 06:39:42,146][lm_eval.evaluator][INFO] - mbpp: Using gen_kwargs: {'until': ['[DONE]'], 'do_sample': False}
33
+ [2025-11-19 06:39:42,147][lm_eval.api.task][INFO] - Building contexts for mbpp on rank 2...
34
+ [2025-11-19 06:39:46,593][lm_eval.evaluator][INFO] - Running generate_until requests
35
+ [2025-11-19 06:39:46,593][lm_eval.evaluator][INFO] - Running generate_until requests
36
+ [2025-11-19 06:39:46,593][lm_eval.evaluator][INFO] - Running generate_until requests
37
+ [2025-11-19 06:39:46,594][lm_eval.evaluator][INFO] - Running generate_until requests
38
+ [2025-11-19 06:39:46,594][lm_eval.evaluator][INFO] - Running generate_until requests
39
+ [2025-11-19 06:39:46,594][lm_eval.evaluator][INFO] - Running generate_until requests
40
+ [2025-11-19 06:39:46,594][lm_eval.evaluator][INFO] - Running generate_until requests
41
+ [2025-11-19 06:39:46,594][lm_eval.evaluator][INFO] - Running generate_until requests
outputs/2025-11-18/exp1_len1024/dream-inst/dllm/mbpp/results.json ADDED
The diff for this file is too large to render. See raw diff
 
outputs/2025-11-18/exp1_len1024/dream-inst/dllm/mbpp/stderr.log ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0
  0%| | 0/62 [00:00<?, ?it/s]
1
  32%|β–ˆβ–ˆβ–ˆβ– | 20/62 [00:00<00:00, 184.79it/s]
2
  63%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 39/62 [00:00<00:00, 133.65it/s]
3
  90%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 56/62 [00:00<00:00, 144.26it/s]
 
4
  0%| | 0/63 [00:00<?, ?it/s]
5
  24%|β–ˆβ–ˆβ– | 15/63 [00:00<00:00, 132.04it/s]
6
  46%|β–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 29/63 [00:00<00:00, 119.63it/s]
7
  0%| | 0/62 [00:00<?, ?it/s]
8
  67%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 42/63 [00:00<00:00, 122.31it/s]
9
  32%|β–ˆβ–ˆβ–ˆβ– | 20/62 [00:00<00:00, 196.39it/s]
10
  87%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 55/63 [00:00<00:00, 115.72it/s]
11
  0%| | 0/62 [00:00<?, ?it/s]
 
12
  65%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 40/62 [00:00<00:00, 197.85it/s]
13
  32%|β–ˆβ–ˆβ–ˆβ– | 20/62 [00:00<00:00, 193.20it/s]
14
  97%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹| 60/62 [00:00<00:00, 198.37it/s]
 
15
  0%| | 0/62 [00:00<?, ?it/s]
16
  65%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 40/62 [00:00<00:00, 193.98it/s]
17
  32%|β–ˆβ–ˆβ–ˆβ– | 20/62 [00:00<00:00, 197.83it/s]
18
  97%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹| 60/62 [00:00<00:00, 194.85it/s]
 
19
  0%| | 0/63 [00:00<?, ?it/s]
20
  66%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 41/62 [00:00<00:00, 199.22it/s]
21
  32%|β–ˆβ–ˆβ–ˆβ– | 20/63 [00:00<00:00, 196.64it/s]
 
22
  63%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 40/63 [00:00<00:00, 197.73it/s]
23
  95%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ| 60/63 [00:00<00:00, 198.02it/s]
 
24
  0%| | 0/63 [00:00<?, ?it/s]
25
  29%|β–ˆβ–ˆβ–Š | 18/63 [00:00<00:00, 176.35it/s]
26
  60%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 38/63 [00:00<00:00, 188.52it/s]
27
  92%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 58/63 [00:00<00:00, 192.50it/s]
 
28
  0%| | 0/63 [00:00<?, ?it/s]
29
  32%|β–ˆβ–ˆβ–ˆβ– | 20/63 [00:00<00:00, 193.28it/s]
30
  63%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 40/63 [00:00<00:00, 195.90it/s]
31
  95%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ| 60/63 [00:00<00:00, 197.29it/s]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ipex flag is deprecated, will be removed in Accelerate v1.10. From 2.7.0, PyTorch has all needed optimizations for Intel CPU and XPU.
2
+ The following values were not passed to `accelerate launch` and had defaults used instead:
3
+ More than one GPU was found, enabling multi-GPU training.
4
+ If this was unintended please pass in `--num_processes=1`.
5
+ `--mixed_precision` was set to a value of `'no'`
6
+ `--dynamo_backend` was set to a value of `'no'`
7
+ To avoid this warning pass in values for each of the problematic parameters or run `accelerate config`.
8
+ [W1119 06:39:00.544153052 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
9
+ 2025-11-19 06:39:08.594 | INFO | src.utils:pre_initialize:603 - {'strategy': 'vanilla', 'threshold': None, 'factor': None, 'alg': 'maskgit_plus', 'gen_length': 1024, 'block_length': 32, 'steps': 1024, 'temperature': 0.0, 'top_p': 0.9, 'top_k': None, 'debias': False, 'output_probs': False, 'mask_token_id': 151666, 'eot_token_id': 151643, 'pad_token_id': 151643, 'add_bos_token': True, 'sigma': None}
10
+ 2025-11-19 06:39:08.594 | INFO | src.utils:pre_initialize:618 - Using cache with args: {'kp': 10, 'kr': 8}
11
+ [W1119 06:39:09.079402470 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
12
+ [W1119 06:39:13.471812944 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
13
+ [W1119 06:39:13.518657843 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
14
+ [W1119 06:39:14.961188753 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
15
+ [W1119 06:39:14.176243946 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
16
+ [W1119 06:39:14.237022163 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
17
+ [W1119 06:39:14.289035287 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
18
+ [W1119 06:39:14.549959783 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
19
+
20
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
21
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
22
+
23
+
24
+
25
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
26
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
27
+
28
+
29
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
30
+
31
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
32
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
33
+
34
+
35
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
36
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
37
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
38
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
39
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
40
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
41
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
42
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
43
+
44
  0%| | 0/62 [00:00<?, ?it/s]
45
  32%|β–ˆβ–ˆβ–ˆβ– | 20/62 [00:00<00:00, 184.79it/s]
46
  63%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 39/62 [00:00<00:00, 133.65it/s]
47
  90%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 56/62 [00:00<00:00, 144.26it/s]
48
+
49
  0%| | 0/63 [00:00<?, ?it/s]
50
  24%|β–ˆβ–ˆβ– | 15/63 [00:00<00:00, 132.04it/s]
51
  46%|β–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 29/63 [00:00<00:00, 119.63it/s]
52
  0%| | 0/62 [00:00<?, ?it/s]
53
  67%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 42/63 [00:00<00:00, 122.31it/s]
54
  32%|β–ˆβ–ˆβ–ˆβ– | 20/62 [00:00<00:00, 196.39it/s]
55
  87%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 55/63 [00:00<00:00, 115.72it/s]
56
  0%| | 0/62 [00:00<?, ?it/s]
57
+
58
  65%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 40/62 [00:00<00:00, 197.85it/s]
59
  32%|β–ˆβ–ˆβ–ˆβ– | 20/62 [00:00<00:00, 193.20it/s]
60
  97%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹| 60/62 [00:00<00:00, 198.37it/s]
61
+
62
  0%| | 0/62 [00:00<?, ?it/s]
63
  65%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 40/62 [00:00<00:00, 193.98it/s]
64
  32%|β–ˆβ–ˆβ–ˆβ– | 20/62 [00:00<00:00, 197.83it/s]
65
  97%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹| 60/62 [00:00<00:00, 194.85it/s]
66
+
67
  0%| | 0/63 [00:00<?, ?it/s]
68
  66%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 41/62 [00:00<00:00, 199.22it/s]
69
  32%|β–ˆβ–ˆβ–ˆβ– | 20/63 [00:00<00:00, 196.64it/s]
70
+
71
  63%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 40/63 [00:00<00:00, 197.73it/s]
72
  95%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ| 60/63 [00:00<00:00, 198.02it/s]
73
+
74
  0%| | 0/63 [00:00<?, ?it/s]
75
  29%|β–ˆβ–ˆβ–Š | 18/63 [00:00<00:00, 176.35it/s]
76
  60%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 38/63 [00:00<00:00, 188.52it/s]
77
  92%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 58/63 [00:00<00:00, 192.50it/s]
78
+
79
  0%| | 0/63 [00:00<?, ?it/s]
80
  32%|β–ˆβ–ˆβ–ˆβ– | 20/63 [00:00<00:00, 193.28it/s]
81
  63%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 40/63 [00:00<00:00, 195.90it/s]
82
  95%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ| 60/63 [00:00<00:00, 197.29it/s]
83
+
84
+ 2025-11-19 06:39:46.604 | WARNING | src.generation:generate:53 - The arguments ('add_bos_token', 'eot_token_id', 'sigma') are not supported by the generation strategy 'vanilla'.
85
+ 2025-11-19 06:39:46.604 | WARNING | src.generation:generate:53 - The arguments ('eot_token_id', 'add_bos_token', 'sigma') are not supported by the generation strategy 'vanilla'.
86
+ 2025-11-19 06:39:46.605 | WARNING | src.generation:generate:53 - The arguments ('eot_token_id', 'add_bos_token', 'sigma') are not supported by the generation strategy 'vanilla'.
87
+ 2025-11-19 06:39:46.606 | WARNING | src.generation:generate:53 - The arguments ('add_bos_token', 'sigma', 'eot_token_id') are not supported by the generation strategy 'vanilla'.
88
+ 2025-11-19 06:39:46.607 | WARNING | src.generation:generate:53 - The arguments ('sigma', 'add_bos_token', 'eot_token_id') are not supported by the generation strategy 'vanilla'.
89
+ 2025-11-19 06:39:46.613 | WARNING | src.generation:generate:53 - The arguments ('add_bos_token', 'eot_token_id', 'sigma') are not supported by the generation strategy 'vanilla'.
90
+ 2025-11-19 06:39:46.614 | WARNING | src.generation:generate:53 - The arguments ('sigma', 'eot_token_id', 'add_bos_token') are not supported by the generation strategy 'vanilla'.
91
+
92
+ 2025-11-19 07:50:55.778 | INFO | __main__:main:81 - Throughput: 6.68 tokens/sec, Tokens per step: 0.87 tokens/step (full: 15.81 tokens/sec, 1.00 tokens/step), Latency: 64.90 s, Total time: 3990.63 s
93
+ 2025-11-19 07:50:55.803 | INFO | __main__:main:98 - Results saved to /xfr_ceph_sh/liuchonghan/HEAT/heat/outputs/2025-11-18/exp1_len1024/dream-inst/dllm/mbpp/results.json
94
+ 2025-11-19 07:50:55.803 | INFO | __main__:main:101 - eval time: 3990.63 seconds
95
+ [rank0]:[W1119 07:50:56.058581021 ProcessGroupNCCL.cpp:1479] Warning: WARNING: destroy_process_group() was not called before program exit, which can leak resources. For more info, please see https://pytorch.org/docs/stable/distributed.html#shutdown (function operator())
outputs/2025-11-18/exp1_len1024/dream-inst/heat/gsm8k/.hydra/config.yaml ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ generation:
3
+ mask_token_id: 151666
4
+ eot_token_id: 151643
5
+ pad_token_id: 151643
6
+ add_bos_token: true
7
+ alg: maskgit_plus
8
+ name: dream-inst
9
+ path: ${oc.env:DREAM_INST_PATH}
10
+ generation:
11
+ strategy: dyna
12
+ threshold: null
13
+ factor: null
14
+ alg: maskgit_plus
15
+ gen_length: 1024
16
+ block_length: null
17
+ steps: 1024
18
+ temperature: 0.0
19
+ top_p: null
20
+ top_k: null
21
+ debias: false
22
+ output_probs: false
23
+ cache:
24
+ _target_: src.cache.HeatCache
25
+ rollout_p: 0.1
26
+ current_k: 32
27
+ num_rollout_layers: null
28
+ sigma: 10.0
29
+ seed: 1234
30
+ batch_size: 1
31
+ attn_implementation: eager
32
+ dataset:
33
+ name: gsm8k
34
+ size: null
35
+ n_shot: null
36
+ system_prompt: null
37
+ batch_size: 1
38
+ mc_num: null
39
+ max_length: 4096
40
+ is_check_greedy: true
41
+ add_bos_token: true
42
+ nll_type: mc
43
+ log_type: ftb
44
+ eval_args:
45
+ log_samples: true
46
+ tasks: ${..dataset.name}
47
+ num_fewshot: ${..dataset.n_shot}
48
+ batch_size: ${..batch_size}
49
+ limit: ${..dataset.size}
50
+ confirm_run_unsafe_code: true
51
+ random_seed: ${..seed}
52
+ fewshot_random_seed: ${..seed}
53
+ numpy_random_seed: ${..seed}
54
+ torch_random_seed: ${..seed}
outputs/2025-11-18/exp1_len1024/dream-inst/heat/gsm8k/.hydra/hydra.yaml ADDED
@@ -0,0 +1,168 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/2025-11-18/exp1_len1024/dream-inst/heat/gsm8k
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.run.dir=outputs/2025-11-18/exp1_len1024/dream-inst/heat/gsm8k
114
+ - hydra.mode=RUN
115
+ task:
116
+ - dataset.name=gsm8k
117
+ - model=dream-inst
118
+ - cache=heat
119
+ - cache.current_k=32
120
+ - cache.rollout_p=0.1
121
+ - generation=dyna
122
+ - generation.gen_length=1024
123
+ - generation.steps=1024
124
+ - batch_size=1
125
+ - seed=1234
126
+ job:
127
+ name: eval
128
+ chdir: null
129
+ override_dirname: batch_size=1,cache.current_k=32,cache.rollout_p=0.1,cache=heat,dataset.name=gsm8k,generation.gen_length=1024,generation.steps=1024,generation=dyna,model=dream-inst,seed=1234
130
+ id: ???
131
+ num: ???
132
+ config_name: eval
133
+ env_set: {}
134
+ env_copy: []
135
+ config:
136
+ override_dirname:
137
+ kv_sep: '='
138
+ item_sep: ','
139
+ exclude_keys: []
140
+ runtime:
141
+ version: 1.3.2
142
+ version_base: '1.3'
143
+ cwd: /xfr_ceph_sh/liuchonghan/HEAT/heat
144
+ config_sources:
145
+ - path: hydra.conf
146
+ schema: pkg
147
+ provider: hydra
148
+ - path: /xfr_ceph_sh/liuchonghan/HEAT/heat/configs
149
+ schema: file
150
+ provider: main
151
+ - path: ''
152
+ schema: structured
153
+ provider: schema
154
+ output_dir: /xfr_ceph_sh/liuchonghan/HEAT/heat/outputs/2025-11-18/exp1_len1024/dream-inst/heat/gsm8k
155
+ choices:
156
+ cache: heat
157
+ generation: dyna
158
+ model: dream-inst
159
+ hydra/env: default
160
+ hydra/callbacks: null
161
+ hydra/job_logging: default
162
+ hydra/hydra_logging: default
163
+ hydra/hydra_help: default
164
+ hydra/help: default
165
+ hydra/sweeper: basic
166
+ hydra/launcher: basic
167
+ hydra/output: default
168
+ verbose: false
outputs/2025-11-18/exp1_len1024/dream-inst/heat/gsm8k/.hydra/overrides.yaml ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ - dataset.name=gsm8k
2
+ - model=dream-inst
3
+ - cache=heat
4
+ - cache.current_k=32
5
+ - cache.rollout_p=0.1
6
+ - generation=dyna
7
+ - generation.gen_length=1024
8
+ - generation.steps=1024
9
+ - batch_size=1
10
+ - seed=1234
outputs/2025-11-18/exp1_len1024/dream-inst/heat/gsm8k/eval.log ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-11-19 01:01:25,028][accelerate.utils.other][WARNING] - Detected kernel version 5.4.0, which is below the recommended minimum of 5.5.0; this can cause the process to hang. It is recommended to upgrade the kernel to the minimum version or higher.
2
+ [2025-11-19 01:01:34,671][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
3
+ [2025-11-19 01:01:34,672][lm_eval.evaluator][INFO] - Using pre-initialized model
4
+ [2025-11-19 01:01:34,736][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
5
+ [2025-11-19 01:01:34,736][lm_eval.evaluator][INFO] - Using pre-initialized model
6
+ [2025-11-19 01:01:34,745][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
7
+ [2025-11-19 01:01:34,745][lm_eval.evaluator][INFO] - Using pre-initialized model
8
+ [2025-11-19 01:01:35,036][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
9
+ [2025-11-19 01:01:35,036][lm_eval.evaluator][INFO] - Using pre-initialized model
10
+ [2025-11-19 01:01:35,146][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
11
+ [2025-11-19 01:01:35,146][lm_eval.evaluator][INFO] - Using pre-initialized model
12
+ [2025-11-19 01:01:35,507][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
13
+ [2025-11-19 01:01:35,508][lm_eval.evaluator][INFO] - Using pre-initialized model
14
+ [2025-11-19 01:01:36,429][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
15
+ [2025-11-19 01:01:36,429][lm_eval.evaluator][INFO] - Using pre-initialized model
16
+ [2025-11-19 01:01:37,173][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
17
+ [2025-11-19 01:01:37,173][lm_eval.evaluator][INFO] - Using pre-initialized model
18
+ [2025-11-19 01:01:45,968][lm_eval.evaluator][INFO] - gsm8k: Using gen_kwargs: {'until': ['Question:', '</s>', '<|im_end|>'], 'do_sample': False, 'temperature': 0.0}
19
+ [2025-11-19 01:01:45,968][lm_eval.api.task][INFO] - Building contexts for gsm8k on rank 5...
20
+ [2025-11-19 01:01:46,129][lm_eval.evaluator][INFO] - gsm8k: Using gen_kwargs: {'until': ['Question:', '</s>', '<|im_end|>'], 'do_sample': False, 'temperature': 0.0}
21
+ [2025-11-19 01:01:46,130][lm_eval.api.task][INFO] - Building contexts for gsm8k on rank 6...
22
+ [2025-11-19 01:01:46,432][lm_eval.evaluator][INFO] - gsm8k: Using gen_kwargs: {'until': ['Question:', '</s>', '<|im_end|>'], 'do_sample': False, 'temperature': 0.0}
23
+ [2025-11-19 01:01:46,433][lm_eval.api.task][INFO] - Building contexts for gsm8k on rank 2...
24
+ [2025-11-19 01:01:47,228][lm_eval.evaluator][INFO] - gsm8k: Using gen_kwargs: {'until': ['Question:', '</s>', '<|im_end|>'], 'do_sample': False, 'temperature': 0.0}
25
+ [2025-11-19 01:01:47,229][lm_eval.api.task][INFO] - Building contexts for gsm8k on rank 7...
26
+ [2025-11-19 01:01:47,349][lm_eval.evaluator][INFO] - gsm8k: Using gen_kwargs: {'until': ['Question:', '</s>', '<|im_end|>'], 'do_sample': False, 'temperature': 0.0}
27
+ [2025-11-19 01:01:47,349][lm_eval.api.task][INFO] - Building contexts for gsm8k on rank 4...
28
+ [2025-11-19 01:01:50,407][lm_eval.evaluator][INFO] - gsm8k: Using gen_kwargs: {'until': ['Question:', '</s>', '<|im_end|>'], 'do_sample': False, 'temperature': 0.0}
29
+ [2025-11-19 01:01:50,407][lm_eval.api.task][INFO] - Building contexts for gsm8k on rank 0...
30
+ [2025-11-19 01:01:52,592][lm_eval.evaluator][INFO] - gsm8k: Using gen_kwargs: {'until': ['Question:', '</s>', '<|im_end|>'], 'do_sample': False, 'temperature': 0.0}
31
+ [2025-11-19 01:01:52,593][lm_eval.api.task][INFO] - Building contexts for gsm8k on rank 3...
32
+ [2025-11-19 01:01:53,960][lm_eval.evaluator][INFO] - gsm8k: Using gen_kwargs: {'until': ['Question:', '</s>', '<|im_end|>'], 'do_sample': False, 'temperature': 0.0}
33
+ [2025-11-19 01:01:53,961][lm_eval.api.task][INFO] - Building contexts for gsm8k on rank 1...
34
+ [2025-11-19 01:01:57,903][lm_eval.evaluator][INFO] - Running generate_until requests
35
+ [2025-11-19 01:01:57,903][lm_eval.evaluator][INFO] - Running generate_until requests
36
+ [2025-11-19 01:01:57,903][lm_eval.evaluator][INFO] - Running generate_until requests
37
+ [2025-11-19 01:01:57,903][lm_eval.evaluator][INFO] - Running generate_until requests
38
+ [2025-11-19 01:01:57,903][lm_eval.evaluator][INFO] - Running generate_until requests
39
+ [2025-11-19 01:01:57,903][lm_eval.evaluator][INFO] - Running generate_until requests
40
+ [2025-11-19 01:01:57,903][lm_eval.evaluator][INFO] - Running generate_until requests
41
+ [2025-11-19 01:01:57,904][lm_eval.evaluator][INFO] - Running generate_until requests
outputs/2025-11-18/exp1_len1024/dream-inst/heat/gsm8k/results.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c65ec61f9d9cdf56e7b11fde503c84051a0e4952b2546e7350a3993df34c34a3
3
+ size 13622720
outputs/2025-11-18/exp1_len1024/dream-inst/heat/gsm8k/stderr.log ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0
  0%| | 0/165 [00:00<?, ?it/s]
1
  15%|β–ˆβ– | 24/165 [00:00<00:00, 236.64it/s]
2
  0%| | 0/165 [00:00<?, ?it/s]
3
  33%|β–ˆβ–ˆβ–ˆβ–Ž | 55/165 [00:00<00:00, 276.52it/s]
4
  23%|β–ˆβ–ˆβ–Ž | 38/165 [00:00<00:00, 372.06it/s]
5
  50%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 83/165 [00:00<00:00, 271.90it/s]
6
  46%|β–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 76/165 [00:00<00:00, 376.30it/s]
7
  67%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 111/165 [00:00<00:00, 238.36it/s]
8
  0%| | 0/165 [00:00<?, ?it/s]
9
  69%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 114/165 [00:00<00:00, 345.60it/s]
10
  85%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 140/165 [00:00<00:00, 254.33it/s]
11
  24%|β–ˆβ–ˆβ–Ž | 39/165 [00:00<00:00, 381.69it/s]
12
  92%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 152/165 [00:00<00:00, 357.59it/s]
 
 
13
  47%|β–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 78/165 [00:00<00:00, 378.82it/s]
14
  71%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 117/165 [00:00<00:00, 382.86it/s]
15
  95%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 156/165 [00:00<00:00, 380.81it/s]
 
16
  0%| | 0/164 [00:00<?, ?it/s]
17
  23%|β–ˆβ–ˆβ–Ž | 38/164 [00:00<00:00, 378.96it/s]
18
  0%| | 0/165 [00:00<?, ?it/s]
19
  47%|β–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 77/164 [00:00<00:00, 383.29it/s]
20
  24%|β–ˆβ–ˆβ–Ž | 39/165 [00:00<00:00, 381.01it/s]
21
  71%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 116/164 [00:00<00:00, 382.83it/s]
22
  48%|β–ˆβ–ˆβ–ˆβ–ˆβ–Š | 79/165 [00:00<00:00, 386.84it/s]
23
  95%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 155/164 [00:00<00:00, 383.20it/s]
24
  72%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 118/165 [00:00<00:00, 387.08it/s]
 
25
  95%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ| 157/165 [00:00<00:00, 386.93it/s]
 
26
  0%| | 0/165 [00:00<?, ?it/s]
27
  12%|β–ˆβ– | 20/165 [00:00<00:00, 192.79it/s]
28
  34%|β–ˆβ–ˆβ–ˆβ– | 56/165 [00:00<00:00, 286.45it/s]
29
  58%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 95/165 [00:00<00:00, 330.03it/s]
30
  81%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 134/165 [00:00<00:00, 351.06it/s]
 
31
  0%| | 0/165 [00:00<?, ?it/s]
32
  23%|β–ˆβ–ˆβ–Ž | 38/165 [00:00<00:00, 371.37it/s]
33
  47%|β–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 77/165 [00:00<00:00, 381.96it/s]
34
  71%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 117/165 [00:00<00:00, 385.84it/s]
35
  95%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ| 157/165 [00:00<00:00, 388.45it/s]
 
36
  0%| | 0/165 [00:00<?, ?it/s]
37
  24%|β–ˆβ–ˆβ–Ž | 39/165 [00:00<00:00, 383.02it/s]
38
  47%|β–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 78/165 [00:00<00:00, 378.08it/s]
39
  71%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 117/165 [00:00<00:00, 381.32it/s]
40
  95%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 156/165 [00:00<00:00, 379.33it/s]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ipex flag is deprecated, will be removed in Accelerate v1.10. From 2.7.0, PyTorch has all needed optimizations for Intel CPU and XPU.
2
+ The following values were not passed to `accelerate launch` and had defaults used instead:
3
+ More than one GPU was found, enabling multi-GPU training.
4
+ If this was unintended please pass in `--num_processes=1`.
5
+ `--mixed_precision` was set to a value of `'no'`
6
+ `--dynamo_backend` was set to a value of `'no'`
7
+ To avoid this warning pass in values for each of the problematic parameters or run `accelerate config`.
8
+ [W1119 01:01:11.003348596 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
9
+ 2025-11-19 01:01:19.008 | INFO | src.utils:pre_initialize:603 - {'strategy': 'dyna', 'threshold': None, 'factor': None, 'alg': 'maskgit_plus', 'gen_length': 1024, 'block_length': 256, 'steps': 1024, 'temperature': 0.0, 'top_p': 0.9, 'top_k': None, 'debias': False, 'output_probs': False, 'mask_token_id': 151666, 'eot_token_id': 151643, 'pad_token_id': 151643, 'add_bos_token': True, 'sigma': 10.0}
10
+ 2025-11-19 01:01:19.009 | INFO | src.utils:pre_initialize:618 - Using cache with args: {'rollout_p': 0.1, 'current_k': 32, 'sigma': 10.0}
11
+ [W1119 01:01:21.010138023 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
12
+ [W1119 01:01:21.683502785 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
13
+ [W1119 01:01:23.239356235 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
14
+ [W1119 01:01:23.367433167 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
15
+ [W1119 01:01:24.077399338 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
16
+ [W1119 01:01:24.115102464 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
17
+ [W1119 01:01:24.135317722 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
18
+ [W1119 01:01:24.152975229 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
19
+
20
+
21
+
22
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
23
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
24
+
25
+
26
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
27
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
28
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
29
+
30
+
31
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
32
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
33
+
34
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
35
+
36
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
37
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
38
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
39
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
40
+
41
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
42
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
43
+
44
  0%| | 0/165 [00:00<?, ?it/s]
45
  15%|β–ˆβ– | 24/165 [00:00<00:00, 236.64it/s]
46
  0%| | 0/165 [00:00<?, ?it/s]
47
  33%|β–ˆβ–ˆβ–ˆβ–Ž | 55/165 [00:00<00:00, 276.52it/s]
48
  23%|β–ˆβ–ˆβ–Ž | 38/165 [00:00<00:00, 372.06it/s]
49
  50%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 83/165 [00:00<00:00, 271.90it/s]
50
  46%|β–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 76/165 [00:00<00:00, 376.30it/s]
51
  67%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 111/165 [00:00<00:00, 238.36it/s]
52
  0%| | 0/165 [00:00<?, ?it/s]
53
  69%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 114/165 [00:00<00:00, 345.60it/s]
54
  85%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 140/165 [00:00<00:00, 254.33it/s]
55
  24%|β–ˆβ–ˆβ–Ž | 39/165 [00:00<00:00, 381.69it/s]
56
  92%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 152/165 [00:00<00:00, 357.59it/s]
57
+
58
+
59
  47%|β–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 78/165 [00:00<00:00, 378.82it/s]
60
  71%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 117/165 [00:00<00:00, 382.86it/s]
61
  95%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 156/165 [00:00<00:00, 380.81it/s]
62
+
63
  0%| | 0/164 [00:00<?, ?it/s]
64
  23%|β–ˆβ–ˆβ–Ž | 38/164 [00:00<00:00, 378.96it/s]
65
  0%| | 0/165 [00:00<?, ?it/s]
66
  47%|β–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 77/164 [00:00<00:00, 383.29it/s]
67
  24%|β–ˆβ–ˆβ–Ž | 39/165 [00:00<00:00, 381.01it/s]
68
  71%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 116/164 [00:00<00:00, 382.83it/s]
69
  48%|β–ˆβ–ˆβ–ˆβ–ˆβ–Š | 79/165 [00:00<00:00, 386.84it/s]
70
  95%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 155/164 [00:00<00:00, 383.20it/s]
71
  72%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 118/165 [00:00<00:00, 387.08it/s]
72
+
73
  95%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ| 157/165 [00:00<00:00, 386.93it/s]
74
+
75
  0%| | 0/165 [00:00<?, ?it/s]
76
  12%|β–ˆβ– | 20/165 [00:00<00:00, 192.79it/s]
77
  34%|β–ˆβ–ˆβ–ˆβ– | 56/165 [00:00<00:00, 286.45it/s]
78
  58%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 95/165 [00:00<00:00, 330.03it/s]
79
  81%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 134/165 [00:00<00:00, 351.06it/s]
80
+
81
  0%| | 0/165 [00:00<?, ?it/s]
82
  23%|β–ˆβ–ˆβ–Ž | 38/165 [00:00<00:00, 371.37it/s]
83
  47%|β–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 77/165 [00:00<00:00, 381.96it/s]
84
  71%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 117/165 [00:00<00:00, 385.84it/s]
85
  95%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ| 157/165 [00:00<00:00, 388.45it/s]
86
+
87
  0%| | 0/165 [00:00<?, ?it/s]
88
  24%|β–ˆβ–ˆβ–Ž | 39/165 [00:00<00:00, 383.02it/s]
89
  47%|β–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 78/165 [00:00<00:00, 378.08it/s]
90
  71%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 117/165 [00:00<00:00, 381.32it/s]
91
  95%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 156/165 [00:00<00:00, 379.33it/s]
92
+
93
+ 2025-11-19 01:01:57.918 | WARNING | src.generation:generate:53 - The arguments ('add_bos_token', 'sigma', 'eot_token_id', 'block_length') are not supported by the generation strategy 'dyna'.
94
+ 2025-11-19 01:01:57.919 | WARNING | src.generation:generate:53 - The arguments ('add_bos_token', 'block_length', 'sigma', 'eot_token_id') are not supported by the generation strategy 'dyna'.
95
+ 2025-11-19 01:01:57.920 | WARNING | src.generation:generate:53 - The arguments ('eot_token_id', 'sigma', 'block_length', 'add_bos_token') are not supported by the generation strategy 'dyna'.
96
+ 2025-11-19 01:01:57.921 | WARNING | src.generation:generate:53 - The arguments ('eot_token_id', 'sigma', 'add_bos_token', 'block_length') are not supported by the generation strategy 'dyna'.
97
+ 2025-11-19 01:01:57.922 | WARNING | src.generation:generate:53 - The arguments ('sigma', 'add_bos_token', 'eot_token_id', 'block_length') are not supported by the generation strategy 'dyna'.
98
+ 2025-11-19 01:01:57.922 | WARNING | src.generation:generate:53 - The arguments ('add_bos_token', 'eot_token_id', 'sigma', 'block_length') are not supported by the generation strategy 'dyna'.
99
+ 2025-11-19 01:01:57.927 | WARNING | src.generation:generate:53 - The arguments ('block_length', 'sigma', 'add_bos_token', 'eot_token_id') are not supported by the generation strategy 'dyna'.
100
+
101
+ 2025-11-19 03:35:49.710 | INFO | __main__:main:81 - Throughput: 2.28 tokens/sec, Tokens per step: 0.99 tokens/step (full: 19.12 tokens/sec, 1.00 tokens/step), Latency: 53.66 s, Total time: 8646.34 s
102
+ 2025-11-19 03:35:49.824 | INFO | __main__:main:98 - Results saved to /xfr_ceph_sh/liuchonghan/HEAT/heat/outputs/2025-11-18/exp1_len1024/dream-inst/heat/gsm8k/results.json
103
+ 2025-11-19 03:35:49.824 | INFO | __main__:main:101 - eval time: 8646.34 seconds
104
+ [rank0]:[W1119 03:35:50.020992224 ProcessGroupNCCL.cpp:1479] Warning: WARNING: destroy_process_group() was not called before program exit, which can leak resources. For more info, please see https://pytorch.org/docs/stable/distributed.html#shutdown (function operator())
outputs/2025-11-18/exp1_len1024/dream-inst/heat/humaneval/.hydra/config.yaml ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ generation:
3
+ mask_token_id: 151666
4
+ eot_token_id: 151643
5
+ pad_token_id: 151643
6
+ add_bos_token: true
7
+ alg: maskgit_plus
8
+ name: dream-inst
9
+ path: ${oc.env:DREAM_INST_PATH}
10
+ generation:
11
+ strategy: dyna
12
+ threshold: null
13
+ factor: null
14
+ alg: maskgit_plus
15
+ gen_length: 1024
16
+ block_length: null
17
+ steps: 1024
18
+ temperature: 0.0
19
+ top_p: null
20
+ top_k: null
21
+ debias: false
22
+ output_probs: false
23
+ cache:
24
+ _target_: src.cache.HeatCache
25
+ rollout_p: 0.1
26
+ current_k: 32
27
+ num_rollout_layers: null
28
+ sigma: 10.0
29
+ seed: 1234
30
+ batch_size: 1
31
+ attn_implementation: eager
32
+ dataset:
33
+ name: humaneval
34
+ size: null
35
+ n_shot: null
36
+ system_prompt: null
37
+ batch_size: 1
38
+ mc_num: null
39
+ max_length: 4096
40
+ is_check_greedy: true
41
+ add_bos_token: true
42
+ nll_type: mc
43
+ log_type: ftb
44
+ eval_args:
45
+ log_samples: true
46
+ tasks: ${..dataset.name}
47
+ num_fewshot: ${..dataset.n_shot}
48
+ batch_size: ${..batch_size}
49
+ limit: ${..dataset.size}
50
+ confirm_run_unsafe_code: true
51
+ random_seed: ${..seed}
52
+ fewshot_random_seed: ${..seed}
53
+ numpy_random_seed: ${..seed}
54
+ torch_random_seed: ${..seed}
outputs/2025-11-18/exp1_len1024/dream-inst/heat/humaneval/.hydra/hydra.yaml ADDED
@@ -0,0 +1,168 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/2025-11-18/exp1_len1024/dream-inst/heat/humaneval
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.run.dir=outputs/2025-11-18/exp1_len1024/dream-inst/heat/humaneval
114
+ - hydra.mode=RUN
115
+ task:
116
+ - dataset.name=humaneval
117
+ - model=dream-inst
118
+ - cache=heat
119
+ - cache.current_k=32
120
+ - cache.rollout_p=0.1
121
+ - generation=dyna
122
+ - generation.gen_length=1024
123
+ - generation.steps=1024
124
+ - batch_size=1
125
+ - seed=1234
126
+ job:
127
+ name: eval
128
+ chdir: null
129
+ override_dirname: batch_size=1,cache.current_k=32,cache.rollout_p=0.1,cache=heat,dataset.name=humaneval,generation.gen_length=1024,generation.steps=1024,generation=dyna,model=dream-inst,seed=1234
130
+ id: ???
131
+ num: ???
132
+ config_name: eval
133
+ env_set: {}
134
+ env_copy: []
135
+ config:
136
+ override_dirname:
137
+ kv_sep: '='
138
+ item_sep: ','
139
+ exclude_keys: []
140
+ runtime:
141
+ version: 1.3.2
142
+ version_base: '1.3'
143
+ cwd: /xfr_ceph_sh/liuchonghan/HEAT/heat
144
+ config_sources:
145
+ - path: hydra.conf
146
+ schema: pkg
147
+ provider: hydra
148
+ - path: /xfr_ceph_sh/liuchonghan/HEAT/heat/configs
149
+ schema: file
150
+ provider: main
151
+ - path: ''
152
+ schema: structured
153
+ provider: schema
154
+ output_dir: /xfr_ceph_sh/liuchonghan/HEAT/heat/outputs/2025-11-18/exp1_len1024/dream-inst/heat/humaneval
155
+ choices:
156
+ cache: heat
157
+ generation: dyna
158
+ model: dream-inst
159
+ hydra/env: default
160
+ hydra/callbacks: null
161
+ hydra/job_logging: default
162
+ hydra/hydra_logging: default
163
+ hydra/hydra_help: default
164
+ hydra/help: default
165
+ hydra/sweeper: basic
166
+ hydra/launcher: basic
167
+ hydra/output: default
168
+ verbose: false
outputs/2025-11-18/exp1_len1024/dream-inst/heat/humaneval/.hydra/overrides.yaml ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ - dataset.name=humaneval
2
+ - model=dream-inst
3
+ - cache=heat
4
+ - cache.current_k=32
5
+ - cache.rollout_p=0.1
6
+ - generation=dyna
7
+ - generation.gen_length=1024
8
+ - generation.steps=1024
9
+ - batch_size=1
10
+ - seed=1234
outputs/2025-11-18/exp1_len1024/dream-inst/heat/humaneval/eval.log ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-11-18 11:57:27,511][accelerate.utils.other][WARNING] - Detected kernel version 5.4.0, which is below the recommended minimum of 5.5.0; this can cause the process to hang. It is recommended to upgrade the kernel to the minimum version or higher.
2
+ [2025-11-18 11:57:50,836][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
3
+ [2025-11-18 11:57:50,836][lm_eval.evaluator][INFO] - Using pre-initialized model
4
+ [2025-11-18 11:57:51,428][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
5
+ [2025-11-18 11:57:51,428][lm_eval.evaluator][INFO] - Using pre-initialized model
6
+ [2025-11-18 11:57:51,507][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
7
+ [2025-11-18 11:57:51,507][lm_eval.evaluator][INFO] - Using pre-initialized model
8
+ [2025-11-18 11:57:51,531][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
9
+ [2025-11-18 11:57:51,531][lm_eval.evaluator][INFO] - Using pre-initialized model
10
+ [2025-11-18 11:57:51,599][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
11
+ [2025-11-18 11:57:51,599][lm_eval.evaluator][INFO] - Using pre-initialized model
12
+ [2025-11-18 11:57:51,849][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
13
+ [2025-11-18 11:57:51,849][lm_eval.evaluator][INFO] - Using pre-initialized model
14
+ [2025-11-18 11:57:52,077][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
15
+ [2025-11-18 11:57:52,077][lm_eval.evaluator][INFO] - Using pre-initialized model
16
+ [2025-11-18 11:57:52,094][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
17
+ [2025-11-18 11:57:52,094][lm_eval.evaluator][INFO] - Using pre-initialized model
18
+ [2025-11-18 11:57:54,019][lm_eval.evaluator][INFO] - humaneval: Using gen_kwargs: {'until': ['\nclass', '\ndef', '\n#', '\nif', '\nprint'], 'max_gen_toks': 1024, 'do_sample': False}
19
+ [2025-11-18 11:57:54,021][lm_eval.api.task][INFO] - Building contexts for humaneval on rank 0...
20
+ [2025-11-18 11:57:55,009][lm_eval.evaluator][INFO] - humaneval: Using gen_kwargs: {'until': ['\nclass', '\ndef', '\n#', '\nif', '\nprint'], 'max_gen_toks': 1024, 'do_sample': False}
21
+ [2025-11-18 11:57:55,011][lm_eval.api.task][INFO] - Building contexts for humaneval on rank 5...
22
+ [2025-11-18 11:57:55,040][lm_eval.evaluator][INFO] - humaneval: Using gen_kwargs: {'until': ['\nclass', '\ndef', '\n#', '\nif', '\nprint'], 'max_gen_toks': 1024, 'do_sample': False}
23
+ [2025-11-18 11:57:55,041][lm_eval.api.task][INFO] - Building contexts for humaneval on rank 6...
24
+ [2025-11-18 11:57:55,060][lm_eval.evaluator][INFO] - humaneval: Using gen_kwargs: {'until': ['\nclass', '\ndef', '\n#', '\nif', '\nprint'], 'max_gen_toks': 1024, 'do_sample': False}
25
+ [2025-11-18 11:57:55,061][lm_eval.api.task][INFO] - Building contexts for humaneval on rank 1...
26
+ [2025-11-18 11:57:55,341][lm_eval.evaluator][INFO] - humaneval: Using gen_kwargs: {'until': ['\nclass', '\ndef', '\n#', '\nif', '\nprint'], 'max_gen_toks': 1024, 'do_sample': False}
27
+ [2025-11-18 11:57:55,343][lm_eval.api.task][INFO] - Building contexts for humaneval on rank 2...
28
+ [2025-11-18 11:57:55,498][lm_eval.evaluator][INFO] - humaneval: Using gen_kwargs: {'until': ['\nclass', '\ndef', '\n#', '\nif', '\nprint'], 'max_gen_toks': 1024, 'do_sample': False}
29
+ [2025-11-18 11:57:55,499][lm_eval.api.task][INFO] - Building contexts for humaneval on rank 4...
30
+ [2025-11-18 11:57:55,593][lm_eval.evaluator][INFO] - humaneval: Using gen_kwargs: {'until': ['\nclass', '\ndef', '\n#', '\nif', '\nprint'], 'max_gen_toks': 1024, 'do_sample': False}
31
+ [2025-11-18 11:57:55,594][lm_eval.api.task][INFO] - Building contexts for humaneval on rank 3...
32
+ [2025-11-18 11:57:55,597][lm_eval.evaluator][INFO] - humaneval: Using gen_kwargs: {'until': ['\nclass', '\ndef', '\n#', '\nif', '\nprint'], 'max_gen_toks': 1024, 'do_sample': False}
33
+ [2025-11-18 11:57:55,598][lm_eval.api.task][INFO] - Building contexts for humaneval on rank 7...
34
+ [2025-11-18 11:57:59,590][lm_eval.evaluator][INFO] - Running generate_until requests
35
+ [2025-11-18 11:57:59,590][lm_eval.evaluator][INFO] - Running generate_until requests
36
+ [2025-11-18 11:57:59,590][lm_eval.evaluator][INFO] - Running generate_until requests
37
+ [2025-11-18 11:57:59,590][lm_eval.evaluator][INFO] - Running generate_until requests
38
+ [2025-11-18 11:57:59,590][lm_eval.evaluator][INFO] - Running generate_until requests
39
+ [2025-11-18 11:57:59,590][lm_eval.evaluator][INFO] - Running generate_until requests
40
+ [2025-11-18 11:57:59,590][lm_eval.evaluator][INFO] - Running generate_until requests
41
+ [2025-11-18 11:57:59,591][lm_eval.evaluator][INFO] - Running generate_until requests
outputs/2025-11-18/exp1_len1024/dream-inst/heat/humaneval/results.json ADDED
The diff for this file is too large to render. See raw diff
 
outputs/2025-11-18/exp1_len1024/dream-inst/heat/humaneval/stderr.log ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0
  0%| | 0/21 [00:00<?, ?it/s]
 
1
  0%| | 0/20 [00:00<?, ?it/s]
 
2
  0%| | 0/20 [00:00<?, ?it/s]
 
3
  0%| | 0/21 [00:00<?, ?it/s]
 
4
  0%| | 0/21 [00:00<?, ?it/s]
 
5
  0%| | 0/20 [00:00<?, ?it/s]
 
6
  0%| | 0/21 [00:00<?, ?it/s]
7
  0%| | 0/20 [00:00<?, ?it/s]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ipex flag is deprecated, will be removed in Accelerate v1.10. From 2.7.0, PyTorch has all needed optimizations for Intel CPU and XPU.
2
+ The following values were not passed to `accelerate launch` and had defaults used instead:
3
+ More than one GPU was found, enabling multi-GPU training.
4
+ If this was unintended please pass in `--num_processes=1`.
5
+ `--mixed_precision` was set to a value of `'no'`
6
+ `--dynamo_backend` was set to a value of `'no'`
7
+ To avoid this warning pass in values for each of the problematic parameters or run `accelerate config`.
8
+ [W1118 11:57:09.587421033 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
9
+ 2025-11-18 11:57:19.084 | INFO | src.utils:pre_initialize:603 - {'strategy': 'dyna', 'threshold': None, 'factor': None, 'alg': 'maskgit_plus', 'gen_length': 1024, 'block_length': 512, 'steps': 1024, 'temperature': 0.0, 'top_p': 0.9, 'top_k': None, 'debias': False, 'output_probs': False, 'mask_token_id': 151666, 'eot_token_id': 151643, 'pad_token_id': 151643, 'add_bos_token': True, 'sigma': 10.0}
10
+ 2025-11-18 11:57:19.084 | INFO | src.utils:pre_initialize:618 - Using cache with args: {'rollout_p': 0.1, 'current_k': 32, 'sigma': 10.0}
11
+ [W1118 11:57:24.909665331 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
12
+ [W1118 11:57:24.996962462 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
13
+ [W1118 11:57:24.158513485 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
14
+ [W1118 11:57:24.202768949 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
15
+ [W1118 11:57:24.221986689 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
16
+ [W1118 11:57:24.233998212 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
17
+ [W1118 11:57:25.783812119 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
18
+ [W1118 11:57:25.854676461 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
19
+
20
+
21
+
22
+
23
+
24
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
25
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
26
+
27
+
28
+
29
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
30
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
31
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
32
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
33
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
34
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
35
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
36
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
37
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
38
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
39
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
40
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
41
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
42
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
43
+
44
  0%| | 0/21 [00:00<?, ?it/s]
45
+
46
  0%| | 0/20 [00:00<?, ?it/s]
47
+
48
  0%| | 0/20 [00:00<?, ?it/s]
49
+
50
  0%| | 0/21 [00:00<?, ?it/s]
51
+
52
  0%| | 0/21 [00:00<?, ?it/s]
53
+
54
  0%| | 0/20 [00:00<?, ?it/s]
55
+
56
  0%| | 0/21 [00:00<?, ?it/s]
57
  0%| | 0/20 [00:00<?, ?it/s]
58
+
59
+
60
+ 2025-11-18 11:57:59.597 | WARNING | src.generation:generate:53 - The arguments ('block_length', 'eot_token_id', 'add_bos_token', 'sigma') are not supported by the generation strategy 'dyna'.
61
+ 2025-11-18 11:57:59.599 | WARNING | src.generation:generate:53 - The arguments ('add_bos_token', 'sigma', 'eot_token_id', 'block_length') are not supported by the generation strategy 'dyna'.
62
+ 2025-11-18 11:57:59.601 | WARNING | src.generation:generate:53 - The arguments ('add_bos_token', 'eot_token_id', 'sigma', 'block_length') are not supported by the generation strategy 'dyna'.
63
+ 2025-11-18 11:57:59.602 | WARNING | src.generation:generate:53 - The arguments ('add_bos_token', 'block_length', 'sigma', 'eot_token_id') are not supported by the generation strategy 'dyna'.
64
+ 2025-11-18 11:57:59.603 | WARNING | src.generation:generate:53 - The arguments ('block_length', 'add_bos_token', 'sigma', 'eot_token_id') are not supported by the generation strategy 'dyna'.
65
+ 2025-11-18 11:57:59.603 | WARNING | src.generation:generate:53 - The arguments ('sigma', 'block_length', 'eot_token_id', 'add_bos_token') are not supported by the generation strategy 'dyna'.
66
+ 2025-11-18 11:57:59.605 | WARNING | src.generation:generate:53 - The arguments ('block_length', 'eot_token_id', 'sigma', 'add_bos_token') are not supported by the generation strategy 'dyna'.
67
+
68
+ 2025-11-18 12:16:38.734 | INFO | __main__:main:81 - Throughput: 2.86 tokens/sec, Tokens per step: 0.99 tokens/step (full: 20.39 tokens/sec, 1.00 tokens/step), Latency: 50.24 s, Total time: 1055.85 s
69
+ 2025-11-18 12:16:38.745 | INFO | __main__:main:98 - Results saved to /xfr_ceph_sh/liuchonghan/HEAT/heat/outputs/2025-11-18/exp1_len1024/dream-inst/heat/humaneval/results.json
70
+ 2025-11-18 12:16:38.745 | INFO | __main__:main:101 - eval time: 1055.85 seconds
71
+ [rank0]:[W1118 12:16:39.004768498 ProcessGroupNCCL.cpp:1479] Warning: WARNING: destroy_process_group() was not called before program exit, which can leak resources. For more info, please see https://pytorch.org/docs/stable/distributed.html#shutdown (function operator())
outputs/2025-11-18/exp1_len1024/dream-inst/heat/math-500/.hydra/config.yaml ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ generation:
3
+ mask_token_id: 151666
4
+ eot_token_id: 151643
5
+ pad_token_id: 151643
6
+ add_bos_token: true
7
+ alg: maskgit_plus
8
+ name: dream-inst
9
+ path: ${oc.env:DREAM_INST_PATH}
10
+ generation:
11
+ strategy: dyna
12
+ threshold: null
13
+ factor: null
14
+ alg: maskgit_plus
15
+ gen_length: 1024
16
+ block_length: null
17
+ steps: 1024
18
+ temperature: 0.0
19
+ top_p: null
20
+ top_k: null
21
+ debias: false
22
+ output_probs: false
23
+ cache:
24
+ _target_: src.cache.HeatCache
25
+ rollout_p: 0.1
26
+ current_k: 32
27
+ num_rollout_layers: null
28
+ sigma: 10.0
29
+ seed: 1234
30
+ batch_size: 1
31
+ attn_implementation: eager
32
+ dataset:
33
+ name: math-500
34
+ size: null
35
+ n_shot: null
36
+ system_prompt: null
37
+ batch_size: 1
38
+ mc_num: null
39
+ max_length: 4096
40
+ is_check_greedy: true
41
+ add_bos_token: true
42
+ nll_type: mc
43
+ log_type: ftb
44
+ eval_args:
45
+ log_samples: true
46
+ tasks: ${..dataset.name}
47
+ num_fewshot: ${..dataset.n_shot}
48
+ batch_size: ${..batch_size}
49
+ limit: ${..dataset.size}
50
+ confirm_run_unsafe_code: true
51
+ random_seed: ${..seed}
52
+ fewshot_random_seed: ${..seed}
53
+ numpy_random_seed: ${..seed}
54
+ torch_random_seed: ${..seed}
outputs/2025-11-18/exp1_len1024/dream-inst/heat/math-500/.hydra/hydra.yaml ADDED
@@ -0,0 +1,168 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/2025-11-18/exp1_len1024/dream-inst/heat/math-500
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.run.dir=outputs/2025-11-18/exp1_len1024/dream-inst/heat/math-500
114
+ - hydra.mode=RUN
115
+ task:
116
+ - dataset.name=math-500
117
+ - model=dream-inst
118
+ - cache=heat
119
+ - cache.current_k=32
120
+ - cache.rollout_p=0.1
121
+ - generation=dyna
122
+ - generation.gen_length=1024
123
+ - generation.steps=1024
124
+ - batch_size=1
125
+ - seed=1234
126
+ job:
127
+ name: eval
128
+ chdir: null
129
+ override_dirname: batch_size=1,cache.current_k=32,cache.rollout_p=0.1,cache=heat,dataset.name=math-500,generation.gen_length=1024,generation.steps=1024,generation=dyna,model=dream-inst,seed=1234
130
+ id: ???
131
+ num: ???
132
+ config_name: eval
133
+ env_set: {}
134
+ env_copy: []
135
+ config:
136
+ override_dirname:
137
+ kv_sep: '='
138
+ item_sep: ','
139
+ exclude_keys: []
140
+ runtime:
141
+ version: 1.3.2
142
+ version_base: '1.3'
143
+ cwd: /xfr_ceph_sh/liuchonghan/HEAT/heat
144
+ config_sources:
145
+ - path: hydra.conf
146
+ schema: pkg
147
+ provider: hydra
148
+ - path: /xfr_ceph_sh/liuchonghan/HEAT/heat/configs
149
+ schema: file
150
+ provider: main
151
+ - path: ''
152
+ schema: structured
153
+ provider: schema
154
+ output_dir: /xfr_ceph_sh/liuchonghan/HEAT/heat/outputs/2025-11-18/exp1_len1024/dream-inst/heat/math-500
155
+ choices:
156
+ cache: heat
157
+ generation: dyna
158
+ model: dream-inst
159
+ hydra/env: default
160
+ hydra/callbacks: null
161
+ hydra/job_logging: default
162
+ hydra/hydra_logging: default
163
+ hydra/hydra_help: default
164
+ hydra/help: default
165
+ hydra/sweeper: basic
166
+ hydra/launcher: basic
167
+ hydra/output: default
168
+ verbose: false
outputs/2025-11-18/exp1_len1024/dream-inst/heat/math-500/.hydra/overrides.yaml ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ - dataset.name=math-500
2
+ - model=dream-inst
3
+ - cache=heat
4
+ - cache.current_k=32
5
+ - cache.rollout_p=0.1
6
+ - generation=dyna
7
+ - generation.gen_length=1024
8
+ - generation.steps=1024
9
+ - batch_size=1
10
+ - seed=1234
outputs/2025-11-18/exp1_len1024/dream-inst/heat/math-500/eval.log ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-11-19 13:22:56,217][accelerate.utils.other][WARNING] - Detected kernel version 5.4.0, which is below the recommended minimum of 5.5.0; this can cause the process to hang. It is recommended to upgrade the kernel to the minimum version or higher.
2
+ [2025-11-19 13:23:06,661][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
3
+ [2025-11-19 13:23:06,661][lm_eval.evaluator][INFO] - Using pre-initialized model
4
+ [2025-11-19 13:23:07,528][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
5
+ [2025-11-19 13:23:07,529][lm_eval.evaluator][INFO] - Using pre-initialized model
6
+ [2025-11-19 13:23:07,673][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
7
+ [2025-11-19 13:23:07,674][lm_eval.evaluator][INFO] - Using pre-initialized model
8
+ [2025-11-19 13:23:07,776][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
9
+ [2025-11-19 13:23:07,776][lm_eval.evaluator][INFO] - Using pre-initialized model
10
+ [2025-11-19 13:23:07,851][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
11
+ [2025-11-19 13:23:07,851][lm_eval.evaluator][INFO] - Using pre-initialized model
12
+ [2025-11-19 13:23:07,914][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
13
+ [2025-11-19 13:23:07,914][lm_eval.evaluator][INFO] - Using pre-initialized model
14
+ [2025-11-19 13:23:07,964][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
15
+ [2025-11-19 13:23:07,964][lm_eval.evaluator][INFO] - Using pre-initialized model
16
+ [2025-11-19 13:23:09,917][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
17
+ [2025-11-19 13:23:09,917][lm_eval.evaluator][INFO] - Using pre-initialized model
18
+ [2025-11-19 13:23:11,870][lm_eval.evaluator][INFO] - math-500: Using gen_kwargs: {'until': ['Problem:'], 'do_sample': False, 'temperature': 0.0}
19
+ [2025-11-19 13:23:11,871][lm_eval.api.task][INFO] - Building contexts for math-500 on rank 2...
20
+ [2025-11-19 13:23:13,115][lm_eval.evaluator][INFO] - math-500: Using gen_kwargs: {'until': ['Problem:'], 'do_sample': False, 'temperature': 0.0}
21
+ [2025-11-19 13:23:13,118][lm_eval.api.task][INFO] - Building contexts for math-500 on rank 1...
22
+ [2025-11-19 13:23:13,142][lm_eval.evaluator][INFO] - math-500: Using gen_kwargs: {'until': ['Problem:'], 'do_sample': False, 'temperature': 0.0}
23
+ [2025-11-19 13:23:13,144][lm_eval.api.task][INFO] - Building contexts for math-500 on rank 4...
24
+ [2025-11-19 13:23:13,160][lm_eval.evaluator][INFO] - math-500: Using gen_kwargs: {'until': ['Problem:'], 'do_sample': False, 'temperature': 0.0}
25
+ [2025-11-19 13:23:13,162][lm_eval.api.task][INFO] - Building contexts for math-500 on rank 0...
26
+ [2025-11-19 13:23:13,450][lm_eval.evaluator][INFO] - math-500: Using gen_kwargs: {'until': ['Problem:'], 'do_sample': False, 'temperature': 0.0}
27
+ [2025-11-19 13:23:13,451][lm_eval.api.task][INFO] - Building contexts for math-500 on rank 6...
28
+ [2025-11-19 13:23:13,458][lm_eval.evaluator][INFO] - math-500: Using gen_kwargs: {'until': ['Problem:'], 'do_sample': False, 'temperature': 0.0}
29
+ [2025-11-19 13:23:13,459][lm_eval.api.task][INFO] - Building contexts for math-500 on rank 3...
30
+ [2025-11-19 13:23:13,552][lm_eval.evaluator][INFO] - math-500: Using gen_kwargs: {'until': ['Problem:'], 'do_sample': False, 'temperature': 0.0}
31
+ [2025-11-19 13:23:13,553][lm_eval.api.task][INFO] - Building contexts for math-500 on rank 5...
32
+ [2025-11-19 13:23:16,828][lm_eval.evaluator][INFO] - math-500: Using gen_kwargs: {'until': ['Problem:'], 'do_sample': False, 'temperature': 0.0}
33
+ [2025-11-19 13:23:16,830][lm_eval.api.task][INFO] - Building contexts for math-500 on rank 7...
34
+ [2025-11-19 13:23:20,968][lm_eval.evaluator][INFO] - Running generate_until requests
35
+ [2025-11-19 13:23:20,968][lm_eval.evaluator][INFO] - Running generate_until requests
36
+ [2025-11-19 13:23:20,968][lm_eval.evaluator][INFO] - Running generate_until requests
37
+ [2025-11-19 13:23:20,968][lm_eval.evaluator][INFO] - Running generate_until requests
38
+ [2025-11-19 13:23:20,968][lm_eval.evaluator][INFO] - Running generate_until requests
39
+ [2025-11-19 13:23:20,968][lm_eval.evaluator][INFO] - Running generate_until requests
40
+ [2025-11-19 13:23:20,969][lm_eval.evaluator][INFO] - Running generate_until requests
41
+ [2025-11-19 13:23:20,969][lm_eval.evaluator][INFO] - Running generate_until requests
outputs/2025-11-18/exp1_len1024/dream-inst/heat/math-500/results.json ADDED
The diff for this file is too large to render. See raw diff
 
outputs/2025-11-18/exp1_len1024/dream-inst/heat/math-500/stderr.log ADDED
@@ -0,0 +1,103 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0
  0%| | 0/63 [00:00<?, ?it/s]
1
  89%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 56/63 [00:00<00:00, 550.79it/s]
 
2
  0%| | 0/63 [00:00<?, ?it/s]
3
  0%| | 0/62 [00:00<?, ?it/s]
4
  0%| | 0/63 [00:00<?, ?it/s]
5
  86%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 54/63 [00:00<00:00, 537.94it/s]
 
6
  89%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 55/62 [00:00<00:00, 545.71it/s]
 
7
  75%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 47/63 [00:00<00:00, 469.70it/s]
 
8
  0%| | 0/62 [00:00<?, ?it/s]
9
  0%| | 0/63 [00:00<?, ?it/s]
10
  0%| | 0/62 [00:00<?, ?it/s]
11
  87%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 54/62 [00:00<00:00, 539.38it/s]
12
  87%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 55/63 [00:00<00:00, 547.05it/s]
 
 
13
  87%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 54/62 [00:00<00:00, 539.81it/s]
 
14
  0%| | 0/62 [00:00<?, ?it/s]
15
  90%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 56/62 [00:00<00:00, 550.85it/s]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ipex flag is deprecated, will be removed in Accelerate v1.10. From 2.7.0, PyTorch has all needed optimizations for Intel CPU and XPU.
2
+ The following values were not passed to `accelerate launch` and had defaults used instead:
3
+ More than one GPU was found, enabling multi-GPU training.
4
+ If this was unintended please pass in `--num_processes=1`.
5
+ `--mixed_precision` was set to a value of `'no'`
6
+ `--dynamo_backend` was set to a value of `'no'`
7
+ To avoid this warning pass in values for each of the problematic parameters or run `accelerate config`.
8
+ [W1119 13:22:41.887213418 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
9
+ 2025-11-19 13:22:49.453 | INFO | src.utils:pre_initialize:603 - {'strategy': 'dyna', 'threshold': None, 'factor': None, 'alg': 'maskgit_plus', 'gen_length': 1024, 'block_length': 256, 'steps': 1024, 'temperature': 0.0, 'top_p': 0.9, 'top_k': None, 'debias': False, 'output_probs': False, 'mask_token_id': 151666, 'eot_token_id': 151643, 'pad_token_id': 151643, 'add_bos_token': True, 'sigma': 10.0}
10
+ 2025-11-19 13:22:49.454 | INFO | src.utils:pre_initialize:618 - Using cache with args: {'rollout_p': 0.1, 'current_k': 32, 'sigma': 10.0}
11
+ [W1119 13:22:50.209196182 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
12
+ [W1119 13:22:51.118344731 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
13
+ [W1119 13:22:52.247364390 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
14
+ [W1119 13:22:53.481626799 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
15
+ [W1119 13:22:54.841878670 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
16
+ [W1119 13:22:54.861399448 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
17
+ [W1119 13:22:54.915315879 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
18
+ [W1119 13:22:54.940491560 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
19
+
20
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
21
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
22
+
23
+
24
+
25
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
26
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
27
+
28
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
29
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
30
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
31
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
32
+
33
+
34
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
35
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
36
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
37
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
38
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
39
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
40
+
41
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
42
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
43
+ 2025-11-19 13:23:02.814 | INFO | src.utils:sympy_antlr_patcher:462 - Detected antlr4-python3-runtime version 4.9.3. Temporarily switching to 4.11.0...
44
+ 2025-11-19 13:23:02.819 | INFO | src.utils:sympy_antlr_patcher:473 - Downloading antlr4-python3-runtime==4.11.0 to /tmp/isolated_antlr_i1m7kavk...
45
+ 2025-11-19 13:23:03.396 | INFO | src.utils:sympy_antlr_patcher:462 - Detected antlr4-python3-runtime version 4.9.3. Temporarily switching to 4.11.0...
46
+ 2025-11-19 13:23:03.400 | INFO | src.utils:sympy_antlr_patcher:473 - Downloading antlr4-python3-runtime==4.11.0 to /tmp/isolated_antlr_hgrcnp6w...
47
+ 2025-11-19 13:23:03.483 | INFO | src.utils:sympy_antlr_patcher:462 - Detected antlr4-python3-runtime version 4.9.3. Temporarily switching to 4.11.0...
48
+ 2025-11-19 13:23:03.498 | INFO | src.utils:sympy_antlr_patcher:473 - Downloading antlr4-python3-runtime==4.11.0 to /tmp/isolated_antlr_msyv5ata...
49
+ 2025-11-19 13:23:03.613 | INFO | src.utils:sympy_antlr_patcher:462 - Detected antlr4-python3-runtime version 4.9.3. Temporarily switching to 4.11.0...
50
+ 2025-11-19 13:23:03.617 | INFO | src.utils:sympy_antlr_patcher:473 - Downloading antlr4-python3-runtime==4.11.0 to /tmp/isolated_antlr_8t27kby9...
51
+ 2025-11-19 13:23:03.645 | INFO | src.utils:sympy_antlr_patcher:462 - Detected antlr4-python3-runtime version 4.9.3. Temporarily switching to 4.11.0...
52
+ 2025-11-19 13:23:03.650 | INFO | src.utils:sympy_antlr_patcher:473 - Downloading antlr4-python3-runtime==4.11.0 to /tmp/isolated_antlr_yffd15jm...
53
+ 2025-11-19 13:23:03.690 | INFO | src.utils:sympy_antlr_patcher:500 - Unpacking antlr4_python3_runtime-4.11.0-py3-none-any.whl...
54
+ 2025-11-19 13:23:03.709 | INFO | src.utils:sympy_antlr_patcher:462 - Detected antlr4-python3-runtime version 4.9.3. Temporarily switching to 4.11.0...
55
+ 2025-11-19 13:23:03.716 | INFO | src.utils:sympy_antlr_patcher:473 - Downloading antlr4-python3-runtime==4.11.0 to /tmp/isolated_antlr_ous84by6...
56
+ 2025-11-19 13:23:03.764 | INFO | src.utils:sympy_antlr_patcher:462 - Detected antlr4-python3-runtime version 4.9.3. Temporarily switching to 4.11.0...
57
+ 2025-11-19 13:23:03.768 | INFO | src.utils:sympy_antlr_patcher:473 - Downloading antlr4-python3-runtime==4.11.0 to /tmp/isolated_antlr_5ko680x0...
58
+ 2025-11-19 13:23:04.277 | INFO | src.utils:sympy_antlr_patcher:500 - Unpacking antlr4_python3_runtime-4.11.0-py3-none-any.whl...
59
+ 2025-11-19 13:23:04.330 | INFO | src.utils:sympy_antlr_patcher:500 - Unpacking antlr4_python3_runtime-4.11.0-py3-none-any.whl...
60
+ 2025-11-19 13:23:04.469 | INFO | src.utils:sympy_antlr_patcher:500 - Unpacking antlr4_python3_runtime-4.11.0-py3-none-any.whl...
61
+ 2025-11-19 13:23:04.557 | INFO | src.utils:sympy_antlr_patcher:500 - Unpacking antlr4_python3_runtime-4.11.0-py3-none-any.whl...
62
+ 2025-11-19 13:23:04.598 | INFO | src.utils:sympy_antlr_patcher:500 - Unpacking antlr4_python3_runtime-4.11.0-py3-none-any.whl...
63
+ 2025-11-19 13:23:04.648 | INFO | src.utils:sympy_antlr_patcher:500 - Unpacking antlr4_python3_runtime-4.11.0-py3-none-any.whl...
64
+ 2025-11-19 13:23:06.596 | INFO | src.utils:sympy_antlr_patcher:462 - Detected antlr4-python3-runtime version 4.9.3. Temporarily switching to 4.11.0...
65
+ 2025-11-19 13:23:06.600 | INFO | src.utils:sympy_antlr_patcher:473 - Downloading antlr4-python3-runtime==4.11.0 to /tmp/isolated_antlr_snuey55k...
66
+ 2025-11-19 13:23:07.558 | INFO | src.utils:sympy_antlr_patcher:500 - Unpacking antlr4_python3_runtime-4.11.0-py3-none-any.whl...
67
+
68
  0%| | 0/63 [00:00<?, ?it/s]
69
  89%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 56/63 [00:00<00:00, 550.79it/s]
70
+
71
  0%| | 0/63 [00:00<?, ?it/s]
72
  0%| | 0/62 [00:00<?, ?it/s]
73
  0%| | 0/63 [00:00<?, ?it/s]
74
  86%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 54/63 [00:00<00:00, 537.94it/s]
75
+
76
  89%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 55/62 [00:00<00:00, 545.71it/s]
77
+
78
  75%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 47/63 [00:00<00:00, 469.70it/s]
79
+
80
  0%| | 0/62 [00:00<?, ?it/s]
81
  0%| | 0/63 [00:00<?, ?it/s]
82
  0%| | 0/62 [00:00<?, ?it/s]
83
  87%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 54/62 [00:00<00:00, 539.38it/s]
84
  87%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 55/63 [00:00<00:00, 547.05it/s]
85
+
86
+
87
  87%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 54/62 [00:00<00:00, 539.81it/s]
88
+
89
  0%| | 0/62 [00:00<?, ?it/s]
90
  90%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 56/62 [00:00<00:00, 550.85it/s]
91
+
92
+ 2025-11-19 13:23:20.979 | WARNING | src.generation:generate:53 - The arguments ('sigma', 'add_bos_token', 'eot_token_id', 'block_length') are not supported by the generation strategy 'dyna'.
93
+ 2025-11-19 13:23:20.982 | WARNING | src.generation:generate:53 - The arguments ('sigma', 'eot_token_id', 'add_bos_token', 'block_length') are not supported by the generation strategy 'dyna'.
94
+ 2025-11-19 13:23:20.984 | WARNING | src.generation:generate:53 - The arguments ('sigma', 'add_bos_token', 'block_length', 'eot_token_id') are not supported by the generation strategy 'dyna'.
95
+ 2025-11-19 13:23:20.986 | WARNING | src.generation:generate:53 - The arguments ('eot_token_id', 'add_bos_token', 'block_length', 'sigma') are not supported by the generation strategy 'dyna'.
96
+ 2025-11-19 13:23:20.987 | WARNING | src.generation:generate:53 - The arguments ('add_bos_token', 'block_length', 'sigma', 'eot_token_id') are not supported by the generation strategy 'dyna'.
97
+ 2025-11-19 13:23:20.993 | WARNING | src.generation:generate:53 - The arguments ('sigma', 'block_length', 'add_bos_token', 'eot_token_id') are not supported by the generation strategy 'dyna'.
98
+ 2025-11-19 13:23:20.997 | WARNING | src.generation:generate:53 - The arguments ('block_length', 'eot_token_id', 'sigma', 'add_bos_token') are not supported by the generation strategy 'dyna'.
99
+
100
+ 2025-11-19 14:19:09.786 | INFO | src.utils:sympy_antlr_patcher:513 - Restoring original environment...
101
+ 2025-11-19 14:19:09.788 | INFO | src.utils:sympy_antlr_patcher:513 - Restoring original environment...
102
+ 2025-11-19 14:19:09.788 | INFO | src.utils:sympy_antlr_patcher:513 - Restoring original environment...
103
+ 2025-11-19 14:19:09.788 | INFO | src.utils:sympy_antlr_patcher:513 - Restoring original environment...
104
+ 2025-11-19 14:19:09.788 | INFO | src.utils:sympy_antlr_patcher:513 - Restoring original environment...
105
+ 2025-11-19 14:19:09.789 | INFO | src.utils:sympy_antlr_patcher:513 - Restoring original environment...
106
+ 2025-11-19 14:19:09.793 | INFO | src.utils:sympy_antlr_patcher:513 - Restoring original environment...
107
+ 2025-11-19 14:19:09.798 | INFO | src.utils:sympy_antlr_patcher:522 - Environment restored.
108
+ 2025-11-19 14:19:09.801 | INFO | src.utils:sympy_antlr_patcher:522 - Environment restored.
109
+ 2025-11-19 14:19:09.801 | INFO | src.utils:sympy_antlr_patcher:522 - Environment restored.
110
+ 2025-11-19 14:19:09.802 | INFO | src.utils:sympy_antlr_patcher:522 - Environment restored.
111
+ 2025-11-19 14:19:09.802 | INFO | src.utils:sympy_antlr_patcher:522 - Environment restored.
112
+ 2025-11-19 14:19:09.804 | INFO | src.utils:sympy_antlr_patcher:522 - Environment restored.
113
+ 2025-11-19 14:19:09.809 | INFO | src.utils:sympy_antlr_patcher:522 - Environment restored.
114
+ 2025-11-19 14:19:16.088 | INFO | src.utils:sympy_antlr_patcher:513 - Restoring original environment...
115
+ 2025-11-19 14:19:16.101 | INFO | src.utils:sympy_antlr_patcher:522 - Environment restored.
116
+ 2025-11-19 14:19:16.102 | INFO | __main__:main:81 - Throughput: 19.64 tokens/sec, Tokens per step: 1.00 tokens/step (full: 19.93 tokens/sec, 1.00 tokens/step), Latency: 51.42 s, Total time: 3188.57 s
117
+ 2025-11-19 14:19:16.130 | INFO | __main__:main:98 - Results saved to /xfr_ceph_sh/liuchonghan/HEAT/heat/outputs/2025-11-18/exp1_len1024/dream-inst/heat/math-500/results.json
118
+ 2025-11-19 14:19:16.130 | INFO | __main__:main:101 - eval time: 3188.57 seconds
119
+ [rank0]:[W1119 14:19:16.383022793 ProcessGroupNCCL.cpp:1479] Warning: WARNING: destroy_process_group() was not called before program exit, which can leak resources. For more info, please see https://pytorch.org/docs/stable/distributed.html#shutdown (function operator())
outputs/2025-11-18/exp1_len1024/dream-inst/heat/mbpp/.hydra/config.yaml ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ generation:
3
+ mask_token_id: 151666
4
+ eot_token_id: 151643
5
+ pad_token_id: 151643
6
+ add_bos_token: true
7
+ alg: maskgit_plus
8
+ name: dream-inst
9
+ path: ${oc.env:DREAM_INST_PATH}
10
+ generation:
11
+ strategy: dyna
12
+ threshold: null
13
+ factor: null
14
+ alg: maskgit_plus
15
+ gen_length: 1024
16
+ block_length: null
17
+ steps: 1024
18
+ temperature: 0.0
19
+ top_p: null
20
+ top_k: null
21
+ debias: false
22
+ output_probs: false
23
+ cache:
24
+ _target_: src.cache.HeatCache
25
+ rollout_p: 0.1
26
+ current_k: 32
27
+ num_rollout_layers: null
28
+ sigma: 10.0
29
+ seed: 1234
30
+ batch_size: 1
31
+ attn_implementation: eager
32
+ dataset:
33
+ name: mbpp
34
+ size: null
35
+ n_shot: null
36
+ system_prompt: null
37
+ batch_size: 1
38
+ mc_num: null
39
+ max_length: 4096
40
+ is_check_greedy: true
41
+ add_bos_token: true
42
+ nll_type: mc
43
+ log_type: ftb
44
+ eval_args:
45
+ log_samples: true
46
+ tasks: ${..dataset.name}
47
+ num_fewshot: ${..dataset.n_shot}
48
+ batch_size: ${..batch_size}
49
+ limit: ${..dataset.size}
50
+ confirm_run_unsafe_code: true
51
+ random_seed: ${..seed}
52
+ fewshot_random_seed: ${..seed}
53
+ numpy_random_seed: ${..seed}
54
+ torch_random_seed: ${..seed}
outputs/2025-11-18/exp1_len1024/dream-inst/heat/mbpp/.hydra/hydra.yaml ADDED
@@ -0,0 +1,168 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/2025-11-18/exp1_len1024/dream-inst/heat/mbpp
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.run.dir=outputs/2025-11-18/exp1_len1024/dream-inst/heat/mbpp
114
+ - hydra.mode=RUN
115
+ task:
116
+ - dataset.name=mbpp
117
+ - model=dream-inst
118
+ - cache=heat
119
+ - cache.current_k=32
120
+ - cache.rollout_p=0.1
121
+ - generation=dyna
122
+ - generation.gen_length=1024
123
+ - generation.steps=1024
124
+ - batch_size=1
125
+ - seed=1234
126
+ job:
127
+ name: eval
128
+ chdir: null
129
+ override_dirname: batch_size=1,cache.current_k=32,cache.rollout_p=0.1,cache=heat,dataset.name=mbpp,generation.gen_length=1024,generation.steps=1024,generation=dyna,model=dream-inst,seed=1234
130
+ id: ???
131
+ num: ???
132
+ config_name: eval
133
+ env_set: {}
134
+ env_copy: []
135
+ config:
136
+ override_dirname:
137
+ kv_sep: '='
138
+ item_sep: ','
139
+ exclude_keys: []
140
+ runtime:
141
+ version: 1.3.2
142
+ version_base: '1.3'
143
+ cwd: /xfr_ceph_sh/liuchonghan/HEAT/heat
144
+ config_sources:
145
+ - path: hydra.conf
146
+ schema: pkg
147
+ provider: hydra
148
+ - path: /xfr_ceph_sh/liuchonghan/HEAT/heat/configs
149
+ schema: file
150
+ provider: main
151
+ - path: ''
152
+ schema: structured
153
+ provider: schema
154
+ output_dir: /xfr_ceph_sh/liuchonghan/HEAT/heat/outputs/2025-11-18/exp1_len1024/dream-inst/heat/mbpp
155
+ choices:
156
+ cache: heat
157
+ generation: dyna
158
+ model: dream-inst
159
+ hydra/env: default
160
+ hydra/callbacks: null
161
+ hydra/job_logging: default
162
+ hydra/hydra_logging: default
163
+ hydra/hydra_help: default
164
+ hydra/help: default
165
+ hydra/sweeper: basic
166
+ hydra/launcher: basic
167
+ hydra/output: default
168
+ verbose: false
outputs/2025-11-18/exp1_len1024/dream-inst/heat/mbpp/.hydra/overrides.yaml ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ - dataset.name=mbpp
2
+ - model=dream-inst
3
+ - cache=heat
4
+ - cache.current_k=32
5
+ - cache.rollout_p=0.1
6
+ - generation=dyna
7
+ - generation.gen_length=1024
8
+ - generation.steps=1024
9
+ - batch_size=1
10
+ - seed=1234
outputs/2025-11-18/exp1_len1024/dream-inst/heat/mbpp/eval.log ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-11-19 07:51:25,028][accelerate.utils.other][WARNING] - Detected kernel version 5.4.0, which is below the recommended minimum of 5.5.0; this can cause the process to hang. It is recommended to upgrade the kernel to the minimum version or higher.
2
+ [2025-11-19 07:51:32,868][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
3
+ [2025-11-19 07:51:32,868][lm_eval.evaluator][INFO] - Using pre-initialized model
4
+ [2025-11-19 07:51:33,237][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
5
+ [2025-11-19 07:51:33,237][lm_eval.evaluator][INFO] - Using pre-initialized model
6
+ [2025-11-19 07:51:34,154][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
7
+ [2025-11-19 07:51:34,154][lm_eval.evaluator][INFO] - Using pre-initialized model
8
+ [2025-11-19 07:51:34,224][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
9
+ [2025-11-19 07:51:34,224][lm_eval.evaluator][INFO] - Using pre-initialized model
10
+ [2025-11-19 07:51:34,331][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
11
+ [2025-11-19 07:51:34,331][lm_eval.evaluator][INFO] - Using pre-initialized model
12
+ [2025-11-19 07:51:34,343][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
13
+ [2025-11-19 07:51:34,343][lm_eval.evaluator][INFO] - Using pre-initialized model
14
+ [2025-11-19 07:51:34,407][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
15
+ [2025-11-19 07:51:34,407][lm_eval.evaluator][INFO] - Using pre-initialized model
16
+ [2025-11-19 07:51:34,533][lm_eval.evaluator][INFO] - Setting random seed to 1234 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 | Setting fewshot manual seed to 1234
17
+ [2025-11-19 07:51:34,533][lm_eval.evaluator][INFO] - Using pre-initialized model
18
+ [2025-11-19 07:51:47,455][lm_eval.evaluator][INFO] - mbpp: Using gen_kwargs: {'until': ['[DONE]'], 'do_sample': False}
19
+ [2025-11-19 07:51:47,457][lm_eval.api.task][INFO] - Building contexts for mbpp on rank 0...
20
+ [2025-11-19 07:51:48,141][lm_eval.evaluator][INFO] - mbpp: Using gen_kwargs: {'until': ['[DONE]'], 'do_sample': False}
21
+ [2025-11-19 07:51:48,142][lm_eval.api.task][INFO] - Building contexts for mbpp on rank 5...
22
+ [2025-11-19 07:51:49,468][lm_eval.evaluator][INFO] - mbpp: Using gen_kwargs: {'until': ['[DONE]'], 'do_sample': False}
23
+ [2025-11-19 07:51:49,469][lm_eval.api.task][INFO] - Building contexts for mbpp on rank 1...
24
+ [2025-11-19 07:51:49,639][lm_eval.evaluator][INFO] - mbpp: Using gen_kwargs: {'until': ['[DONE]'], 'do_sample': False}
25
+ [2025-11-19 07:51:49,640][lm_eval.api.task][INFO] - Building contexts for mbpp on rank 2...
26
+ [2025-11-19 07:51:49,706][lm_eval.evaluator][INFO] - mbpp: Using gen_kwargs: {'until': ['[DONE]'], 'do_sample': False}
27
+ [2025-11-19 07:51:49,707][lm_eval.api.task][INFO] - Building contexts for mbpp on rank 4...
28
+ [2025-11-19 07:51:49,709][lm_eval.evaluator][INFO] - mbpp: Using gen_kwargs: {'until': ['[DONE]'], 'do_sample': False}
29
+ [2025-11-19 07:51:49,710][lm_eval.api.task][INFO] - Building contexts for mbpp on rank 6...
30
+ [2025-11-19 07:51:50,065][lm_eval.evaluator][INFO] - mbpp: Using gen_kwargs: {'until': ['[DONE]'], 'do_sample': False}
31
+ [2025-11-19 07:51:50,067][lm_eval.api.task][INFO] - Building contexts for mbpp on rank 7...
32
+ [2025-11-19 07:51:50,067][lm_eval.evaluator][INFO] - mbpp: Using gen_kwargs: {'until': ['[DONE]'], 'do_sample': False}
33
+ [2025-11-19 07:51:50,068][lm_eval.api.task][INFO] - Building contexts for mbpp on rank 3...
34
+ [2025-11-19 07:51:54,467][lm_eval.evaluator][INFO] - Running generate_until requests
35
+ [2025-11-19 07:51:54,467][lm_eval.evaluator][INFO] - Running generate_until requests
36
+ [2025-11-19 07:51:54,467][lm_eval.evaluator][INFO] - Running generate_until requests
37
+ [2025-11-19 07:51:54,467][lm_eval.evaluator][INFO] - Running generate_until requests
38
+ [2025-11-19 07:51:54,467][lm_eval.evaluator][INFO] - Running generate_until requests
39
+ [2025-11-19 07:51:54,467][lm_eval.evaluator][INFO] - Running generate_until requests
40
+ [2025-11-19 07:51:54,467][lm_eval.evaluator][INFO] - Running generate_until requests
41
+ [2025-11-19 07:51:54,469][lm_eval.evaluator][INFO] - Running generate_until requests
outputs/2025-11-18/exp1_len1024/dream-inst/heat/mbpp/results.json ADDED
The diff for this file is too large to render. See raw diff
 
outputs/2025-11-18/exp1_len1024/dream-inst/heat/mbpp/stderr.log ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0
  0%| | 0/63 [00:00<?, ?it/s]
1
  32%|β–ˆβ–ˆβ–ˆβ– | 20/63 [00:00<00:00, 192.09it/s]
2
  63%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 40/63 [00:00<00:00, 195.01it/s]
3
  95%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ| 60/63 [00:00<00:00, 196.83it/s]
 
4
  0%| | 0/62 [00:00<?, ?it/s]
5
  27%|β–ˆβ–ˆβ–‹ | 17/62 [00:00<00:00, 166.68it/s]
6
  60%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 37/62 [00:00<00:00, 181.32it/s]
7
  92%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 57/62 [00:00<00:00, 189.04it/s]
 
8
  0%| | 0/63 [00:00<?, ?it/s]
9
  32%|β–ˆβ–ˆβ–ˆβ– | 20/63 [00:00<00:00, 198.33it/s]
10
  0%| | 0/63 [00:00<?, ?it/s]
11
  0%| | 0/62 [00:00<?, ?it/s]
12
  0%| | 0/62 [00:00<?, ?it/s]
13
  63%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 40/63 [00:00<00:00, 152.60it/s]
14
  32%|β–ˆβ–ˆβ–ˆβ– | 20/63 [00:00<00:00, 197.48it/s]
15
  32%|β–ˆβ–ˆβ–ˆβ– | 20/62 [00:00<00:00, 197.64it/s]
16
  32%|β–ˆβ–ˆβ–ˆβ– | 20/62 [00:00<00:00, 188.25it/s]
17
  90%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 57/63 [00:00<00:00, 156.83it/s]
18
  65%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 41/63 [00:00<00:00, 199.17it/s]
 
19
  65%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 40/62 [00:00<00:00, 198.45it/s]
20
  65%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 40/62 [00:00<00:00, 193.49it/s]
21
  98%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š| 62/63 [00:00<00:00, 199.65it/s]
 
22
  97%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹| 60/62 [00:00<00:00, 199.02it/s]
23
  97%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹| 60/62 [00:00<00:00, 195.92it/s]
 
 
24
  0%| | 0/62 [00:00<?, ?it/s]
25
  0%| | 0/63 [00:00<?, ?it/s]
26
  32%|β–ˆβ–ˆβ–ˆβ– | 20/62 [00:00<00:00, 198.11it/s]
27
  29%|β–ˆβ–ˆβ–Š | 18/63 [00:00<00:00, 173.29it/s]
28
  65%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 40/62 [00:00<00:00, 199.09it/s]
29
  59%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 37/63 [00:00<00:00, 178.84it/s]
30
  98%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š| 61/62 [00:00<00:00, 199.65it/s]
 
31
  90%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 57/63 [00:00<00:00, 188.11it/s]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ipex flag is deprecated, will be removed in Accelerate v1.10. From 2.7.0, PyTorch has all needed optimizations for Intel CPU and XPU.
2
+ The following values were not passed to `accelerate launch` and had defaults used instead:
3
+ More than one GPU was found, enabling multi-GPU training.
4
+ If this was unintended please pass in `--num_processes=1`.
5
+ `--mixed_precision` was set to a value of `'no'`
6
+ `--dynamo_backend` was set to a value of `'no'`
7
+ To avoid this warning pass in values for each of the problematic parameters or run `accelerate config`.
8
+ [W1119 07:51:11.683567084 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
9
+ 2025-11-19 07:51:19.460 | INFO | src.utils:pre_initialize:603 - {'strategy': 'dyna', 'threshold': None, 'factor': None, 'alg': 'maskgit_plus', 'gen_length': 1024, 'block_length': 512, 'steps': 1024, 'temperature': 0.0, 'top_p': 0.9, 'top_k': None, 'debias': False, 'output_probs': False, 'mask_token_id': 151666, 'eot_token_id': 151643, 'pad_token_id': 151643, 'add_bos_token': True, 'sigma': 10.0}
10
+ 2025-11-19 07:51:19.460 | INFO | src.utils:pre_initialize:618 - Using cache with args: {'rollout_p': 0.1, 'current_k': 32, 'sigma': 10.0}
11
+ [W1119 07:51:21.592036312 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
12
+ [W1119 07:51:21.593016469 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
13
+ [W1119 07:51:21.693916007 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
14
+ [W1119 07:51:24.814455070 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
15
+ [W1119 07:51:24.438362045 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
16
+ [W1119 07:51:24.687729340 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
17
+ [W1119 07:51:25.782100378 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
18
+ [W1119 07:51:25.816994034 socket.cpp:755] [c10d] The client socket cannot be initialized to connect to [localhost]:29500 (errno: 97 - Address family not supported by protocol).
19
+
20
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
21
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
22
+
23
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
24
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
25
+
26
+
27
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
28
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
29
+
30
+
31
+
32
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
33
+
34
+
35
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
36
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
37
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
38
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
39
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
40
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
41
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
42
+ The following generation flags are not valid and may be ignored: ['temperature']. Set `TRANSFORMERS_VERBOSITY=info` for more details.
43
+
44
  0%| | 0/63 [00:00<?, ?it/s]
45
  32%|β–ˆβ–ˆβ–ˆβ– | 20/63 [00:00<00:00, 192.09it/s]
46
  63%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 40/63 [00:00<00:00, 195.01it/s]
47
  95%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ| 60/63 [00:00<00:00, 196.83it/s]
48
+
49
  0%| | 0/62 [00:00<?, ?it/s]
50
  27%|β–ˆβ–ˆβ–‹ | 17/62 [00:00<00:00, 166.68it/s]
51
  60%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 37/62 [00:00<00:00, 181.32it/s]
52
  92%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 57/62 [00:00<00:00, 189.04it/s]
53
+
54
  0%| | 0/63 [00:00<?, ?it/s]
55
  32%|β–ˆβ–ˆβ–ˆβ– | 20/63 [00:00<00:00, 198.33it/s]
56
  0%| | 0/63 [00:00<?, ?it/s]
57
  0%| | 0/62 [00:00<?, ?it/s]
58
  0%| | 0/62 [00:00<?, ?it/s]
59
  63%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 40/63 [00:00<00:00, 152.60it/s]
60
  32%|β–ˆβ–ˆβ–ˆβ– | 20/63 [00:00<00:00, 197.48it/s]
61
  32%|β–ˆβ–ˆβ–ˆβ– | 20/62 [00:00<00:00, 197.64it/s]
62
  32%|β–ˆβ–ˆβ–ˆβ– | 20/62 [00:00<00:00, 188.25it/s]
63
  90%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 57/63 [00:00<00:00, 156.83it/s]
64
  65%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 41/63 [00:00<00:00, 199.17it/s]
65
+
66
  65%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 40/62 [00:00<00:00, 198.45it/s]
67
  65%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 40/62 [00:00<00:00, 193.49it/s]
68
  98%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š| 62/63 [00:00<00:00, 199.65it/s]
69
+
70
  97%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹| 60/62 [00:00<00:00, 199.02it/s]
71
  97%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹| 60/62 [00:00<00:00, 195.92it/s]
72
+
73
+
74
  0%| | 0/62 [00:00<?, ?it/s]
75
  0%| | 0/63 [00:00<?, ?it/s]
76
  32%|β–ˆβ–ˆβ–ˆβ– | 20/62 [00:00<00:00, 198.11it/s]
77
  29%|β–ˆβ–ˆβ–Š | 18/63 [00:00<00:00, 173.29it/s]
78
  65%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 40/62 [00:00<00:00, 199.09it/s]
79
  59%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 37/63 [00:00<00:00, 178.84it/s]
80
  98%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š| 61/62 [00:00<00:00, 199.65it/s]
81
+
82
  90%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 57/63 [00:00<00:00, 188.11it/s]
83
+
84
+ 2025-11-19 07:51:54.479 | WARNING | src.generation:generate:53 - The arguments ('block_length', 'sigma', 'eot_token_id', 'add_bos_token') are not supported by the generation strategy 'dyna'.
85
+ 2025-11-19 07:51:54.484 | WARNING | src.generation:generate:53 - The arguments ('eot_token_id', 'add_bos_token', 'block_length', 'sigma') are not supported by the generation strategy 'dyna'.
86
+ 2025-11-19 07:51:54.484 | WARNING | src.generation:generate:53 - The arguments ('eot_token_id', 'add_bos_token', 'sigma', 'block_length') are not supported by the generation strategy 'dyna'.
87
+ 2025-11-19 07:51:54.484 | WARNING | src.generation:generate:53 - The arguments ('eot_token_id', 'block_length', 'sigma', 'add_bos_token') are not supported by the generation strategy 'dyna'.
88
+ 2025-11-19 07:51:54.485 | WARNING | src.generation:generate:53 - The arguments ('block_length', 'eot_token_id', 'sigma', 'add_bos_token') are not supported by the generation strategy 'dyna'.
89
+ 2025-11-19 07:51:54.486 | WARNING | src.generation:generate:53 - The arguments ('sigma', 'eot_token_id', 'add_bos_token', 'block_length') are not supported by the generation strategy 'dyna'.
90
+ 2025-11-19 07:51:54.504 | WARNING | src.generation:generate:53 - The arguments ('sigma', 'block_length', 'eot_token_id', 'add_bos_token') are not supported by the generation strategy 'dyna'.
91
+
92
+ 2025-11-19 08:49:24.011 | INFO | __main__:main:81 - Throughput: 7.07 tokens/sec, Tokens per step: 0.93 tokens/step (full: 20.06 tokens/sec, 1.00 tokens/step), Latency: 51.12 s, Total time: 3208.85 s
93
+ 2025-11-19 08:49:24.040 | INFO | __main__:main:98 - Results saved to /xfr_ceph_sh/liuchonghan/HEAT/heat/outputs/2025-11-18/exp1_len1024/dream-inst/heat/mbpp/results.json
94
+ 2025-11-19 08:49:24.041 | INFO | __main__:main:101 - eval time: 3208.85 seconds
95
+ [rank0]:[W1119 08:49:24.234766482 ProcessGroupNCCL.cpp:1479] Warning: WARNING: destroy_process_group() was not called before program exit, which can leak resources. For more info, please see https://pytorch.org/docs/stable/distributed.html#shutdown (function operator())
outputs/2025-11-18/exp1_len1024/dream-inst/no_cache/gsm8k/.hydra/config.yaml ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ generation:
3
+ mask_token_id: 151666
4
+ eot_token_id: 151643
5
+ pad_token_id: 151643
6
+ add_bos_token: true
7
+ alg: maskgit_plus
8
+ name: dream-inst
9
+ path: ${oc.env:DREAM_INST_PATH}
10
+ generation:
11
+ strategy: vanilla
12
+ threshold: null
13
+ factor: null
14
+ alg: maskgit_plus
15
+ gen_length: 1024
16
+ block_length: 32
17
+ steps: 1024
18
+ temperature: 0.0
19
+ top_p: null
20
+ top_k: null
21
+ debias: false
22
+ output_probs: false
23
+ seed: 1234
24
+ batch_size: 1
25
+ attn_implementation: eager
26
+ dataset:
27
+ name: gsm8k
28
+ size: null
29
+ n_shot: null
30
+ system_prompt: null
31
+ batch_size: 1
32
+ mc_num: null
33
+ max_length: 4096
34
+ is_check_greedy: true
35
+ add_bos_token: true
36
+ nll_type: mc
37
+ log_type: ftb
38
+ eval_args:
39
+ log_samples: true
40
+ tasks: ${..dataset.name}
41
+ num_fewshot: ${..dataset.n_shot}
42
+ batch_size: ${..batch_size}
43
+ limit: ${..dataset.size}
44
+ confirm_run_unsafe_code: true
45
+ random_seed: ${..seed}
46
+ fewshot_random_seed: ${..seed}
47
+ numpy_random_seed: ${..seed}
48
+ torch_random_seed: ${..seed}