Yaning1001 commited on
Commit
8faa740
·
verified ·
1 Parent(s): 4471134

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +7 -0
  2. wandb/run-20241030_010641-ig1px172/files/config.yaml +47 -0
  3. wandb/run-20241030_010641-ig1px172/files/output.log +4 -0
  4. wandb/run-20241030_010641-ig1px172/files/wandb-metadata.json +29 -0
  5. wandb/run-20241030_010641-ig1px172/files/wandb-summary.json +1 -0
  6. wandb/run-20241030_010641-ig1px172/logs/debug-internal.log +16 -0
  7. wandb/run-20241030_010641-ig1px172/logs/debug.log +27 -0
  8. wandb/run-20241030_010641-ig1px172/run-ig1px172.wandb +0 -0
  9. wandb/run-20241030_222932-73t76wgo/files/output.log +40 -0
  10. wandb/run-20241030_222932-73t76wgo/files/requirements.txt +147 -0
  11. wandb/run-20241030_222932-73t76wgo/files/wandb-metadata.json +97 -0
  12. wandb/run-20241030_222932-73t76wgo/logs/debug-internal.log +10 -0
  13. wandb/run-20241030_222932-73t76wgo/logs/debug.log +26 -0
  14. wandb/run-20241030_225833-5mf4xvmz/run-5mf4xvmz.wandb +3 -0
  15. wandb/run-20241030_225833-a1ftigpc/run-a1ftigpc.wandb +3 -0
  16. wandb/run-20241030_225833-h0mm818w/files/output.log +38 -0
  17. wandb/run-20241030_225833-h0mm818w/files/requirements.txt +147 -0
  18. wandb/run-20241030_225833-h0mm818w/files/wandb-metadata.json +97 -0
  19. wandb/run-20241030_225833-h0mm818w/logs/debug-internal.log +8 -0
  20. wandb/run-20241030_225833-h0mm818w/logs/debug.log +26 -0
  21. wandb/run-20241030_231835-aqeiow5c/run-aqeiow5c.wandb +3 -0
  22. wandb/run-20241030_233740-a8ghkt6q/run-a8ghkt6q.wandb +3 -0
  23. wandb/run-20241031_122113-f74sq78x/files/config.yaml +531 -0
  24. wandb/run-20241031_122113-f74sq78x/files/output.log +0 -0
  25. wandb/run-20241031_122113-f74sq78x/files/wandb-metadata.json +97 -0
  26. wandb/run-20241031_122113-f74sq78x/files/wandb-summary.json +1 -0
  27. wandb/run-20241031_122113-f74sq78x/logs/debug-internal.log +17 -0
  28. wandb/run-20241031_122113-f74sq78x/logs/debug.log +36 -0
  29. wandb/run-20241101_093116-hq0hplpe/files/config.yaml +481 -0
  30. wandb/run-20241101_093116-hq0hplpe/files/output.log +52 -0
  31. wandb/run-20241101_093116-hq0hplpe/files/wandb-metadata.json +97 -0
  32. wandb/run-20241101_093116-hq0hplpe/files/wandb-summary.json +1 -0
  33. wandb/run-20241101_093116-hq0hplpe/logs/debug-internal.log +16 -0
  34. wandb/run-20241101_093116-hq0hplpe/logs/debug.log +30 -0
  35. wandb/run-20241101_093116-hq0hplpe/run-hq0hplpe.wandb +0 -0
  36. wandb/run-20241101_200535-k9brpado/run-k9brpado.wandb +3 -0
  37. wandb/run-20241101_200535-lnp8ii96/run-lnp8ii96.wandb +3 -0
  38. wandb/run-20241105_155954-baeezzpr/files/config.yaml +49 -0
  39. wandb/run-20241105_155954-baeezzpr/files/output.log +19 -0
  40. wandb/run-20241105_155954-baeezzpr/files/requirements.txt +147 -0
  41. wandb/run-20241105_155954-baeezzpr/files/wandb-metadata.json +44 -0
  42. wandb/run-20241105_155954-baeezzpr/files/wandb-summary.json +1 -0
  43. wandb/run-20241105_155954-baeezzpr/logs/debug-internal.log +17 -0
  44. wandb/run-20241105_155954-baeezzpr/logs/debug.log +27 -0
  45. wandb/run-20241105_155954-baeezzpr/run-baeezzpr.wandb +0 -0
  46. wandb/run-20241106_224518-az8kcj8i/files/config.yaml +49 -0
  47. wandb/run-20241106_224518-az8kcj8i/files/output.log +57 -0
  48. wandb/run-20241106_224518-az8kcj8i/files/wandb-metadata.json +97 -0
  49. wandb/run-20241106_224518-az8kcj8i/files/wandb-summary.json +1 -0
  50. wandb/run-20241106_224518-az8kcj8i/logs/debug-internal.log +16 -0
.gitattributes CHANGED
@@ -69,3 +69,10 @@ wandb/run-20241113_180154-b24s43he/run-b24s43he.wandb filter=lfs diff=lfs merge=
69
  wandb/run-20241030_233740-0fagh7s8/run-0fagh7s8.wandb filter=lfs diff=lfs merge=lfs -text
70
  wandb/run-20241031_001055-dua2g15g/run-dua2g15g.wandb filter=lfs diff=lfs merge=lfs -text
71
  wandb/run-20241101_200535-xloij0da/run-xloij0da.wandb filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
69
  wandb/run-20241030_233740-0fagh7s8/run-0fagh7s8.wandb filter=lfs diff=lfs merge=lfs -text
70
  wandb/run-20241031_001055-dua2g15g/run-dua2g15g.wandb filter=lfs diff=lfs merge=lfs -text
71
  wandb/run-20241101_200535-xloij0da/run-xloij0da.wandb filter=lfs diff=lfs merge=lfs -text
72
+ wandb/run-20241101_200535-k9brpado/run-k9brpado.wandb filter=lfs diff=lfs merge=lfs -text
73
+ wandb/run-20241030_225833-5mf4xvmz/run-5mf4xvmz.wandb filter=lfs diff=lfs merge=lfs -text
74
+ wandb/run-20241030_225833-a1ftigpc/run-a1ftigpc.wandb filter=lfs diff=lfs merge=lfs -text
75
+ wandb/run-20241130_000210-r9oa6v8c/run-r9oa6v8c.wandb filter=lfs diff=lfs merge=lfs -text
76
+ wandb/run-20241030_231835-aqeiow5c/run-aqeiow5c.wandb filter=lfs diff=lfs merge=lfs -text
77
+ wandb/run-20241030_233740-a8ghkt6q/run-a8ghkt6q.wandb filter=lfs diff=lfs merge=lfs -text
78
+ wandb/run-20241101_200535-lnp8ii96/run-lnp8ii96.wandb filter=lfs diff=lfs merge=lfs -text
wandb/run-20241030_010641-ig1px172/files/config.yaml ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.18.5
4
+ m: []
5
+ python_version: 3.9.19
6
+ t:
7
+ "1":
8
+ - 1
9
+ - 5
10
+ - 11
11
+ - 49
12
+ - 51
13
+ - 53
14
+ - 55
15
+ - 71
16
+ - 98
17
+ "2":
18
+ - 1
19
+ - 5
20
+ - 11
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 98
27
+ "3":
28
+ - 13
29
+ - 23
30
+ - 55
31
+ "4": 3.9.19
32
+ "5": 0.18.5
33
+ "6": 4.45.1
34
+ "8":
35
+ - 5
36
+ "12": 0.18.5
37
+ "13": linux-x86_64
38
+ batch_size:
39
+ value: 3
40
+ epoch:
41
+ value: 7
42
+ perturbation:
43
+ value: reverse_control
44
+ seed:
45
+ value: 0
46
+ train_set:
47
+ value: 10M
wandb/run-20241030_010641-ig1px172/files/output.log ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ Traceback (most recent call last):
2
+ File "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py", line 162, in <module>
3
+ dataset_name = f"babylm_{args.perturbation}_{args.train_zset}_seed{args.seed}"
4
+ AttributeError: 'Namespace' object has no attribute 'train_zset'
wandb/run-20241030_010641-ig1px172/files/wandb-metadata.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-162-generic-x86_64-with-glibc2.31",
3
+ "python": "3.9.19",
4
+ "startedAt": "2024-10-30T05:06:41.443612Z",
5
+ "args": [
6
+ "--perturbation",
7
+ "reverse_control",
8
+ "--train_set",
9
+ "10M",
10
+ "--batch_size",
11
+ "3",
12
+ "--epoch",
13
+ "7",
14
+ "--seed",
15
+ "0"
16
+ ],
17
+ "program": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py",
18
+ "codePath": "train/train_deep_wandb.py",
19
+ "git": {
20
+ "remote": "git@hf.co:Yaning1001/Impossible_llm.git",
21
+ "commit": "ed716cdcfcdea02b67f7ed0f3504c2b1c8b737c4"
22
+ },
23
+ "email": "yaning1001@gmail.com",
24
+ "root": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train",
25
+ "host": "mms-large-2",
26
+ "username": "chunhui",
27
+ "executable": "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/bin/python",
28
+ "codePathLocal": "train_deep_wandb.py"
29
+ }
wandb/run-20241030_010641-ig1px172/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_wandb":{"runtime":1}}
wandb/run-20241030_010641-ig1px172/logs/debug-internal.log ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2024-10-30T01:06:41.44738347-04:00","level":"INFO","msg":"using version","core version":"0.18.5"}
2
+ {"time":"2024-10-30T01:06:41.44739976-04:00","level":"INFO","msg":"created symlink","path":"/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241030_010641-ig1px172/logs/debug-core.log"}
3
+ {"time":"2024-10-30T01:06:41.555521578-04:00","level":"INFO","msg":"created new stream","id":"ig1px172"}
4
+ {"time":"2024-10-30T01:06:41.555572528-04:00","level":"INFO","msg":"stream: started","id":"ig1px172"}
5
+ {"time":"2024-10-30T01:06:41.555614238-04:00","level":"INFO","msg":"sender: started","stream_id":"ig1px172"}
6
+ {"time":"2024-10-30T01:06:41.555601318-04:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"ig1px172"}}
7
+ {"time":"2024-10-30T01:06:41.555602858-04:00","level":"INFO","msg":"handler: started","stream_id":{"value":"ig1px172"}}
8
+ {"time":"2024-10-30T01:06:43.050607579-04:00","level":"INFO","msg":"Starting system monitor"}
9
+ {"time":"2024-10-30T01:06:43.145734168-04:00","level":"INFO","msg":"stream: closing","id":"ig1px172"}
10
+ {"time":"2024-10-30T01:06:43.145770938-04:00","level":"INFO","msg":"Stopping system monitor"}
11
+ {"time":"2024-10-30T01:06:43.240066421-04:00","level":"INFO","msg":"Stopped system monitor"}
12
+ {"time":"2024-10-30T01:06:43.773982834-04:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
13
+ {"time":"2024-10-30T01:06:43.883577491-04:00","level":"INFO","msg":"handler: closed","stream_id":{"value":"ig1px172"}}
14
+ {"time":"2024-10-30T01:06:43.883615942-04:00","level":"INFO","msg":"writer: Close: closed","stream_id":{"value":"ig1px172"}}
15
+ {"time":"2024-10-30T01:06:43.883633552-04:00","level":"INFO","msg":"sender: closed","stream_id":"ig1px172"}
16
+ {"time":"2024-10-30T01:06:43.883683032-04:00","level":"INFO","msg":"stream: closed","id":"ig1px172"}
wandb/run-20241030_010641-ig1px172/logs/debug.log ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-10-30 01:06:41,442 INFO MainThread:321596 [wandb_setup.py:_flush():79] Current SDK version is 0.18.5
2
+ 2024-10-30 01:06:41,442 INFO MainThread:321596 [wandb_setup.py:_flush():79] Configure stats pid to 321596
3
+ 2024-10-30 01:06:41,442 INFO MainThread:321596 [wandb_setup.py:_flush():79] Loading settings from /home/chunhui/.config/wandb/settings
4
+ 2024-10-30 01:06:41,442 INFO MainThread:321596 [wandb_setup.py:_flush():79] Loading settings from /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/settings
5
+ 2024-10-30 01:06:41,442 INFO MainThread:321596 [wandb_setup.py:_flush():79] Loading settings from environment variables: {}
6
+ 2024-10-30 01:06:41,442 INFO MainThread:321596 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': None, '_disable_service': None}
7
+ 2024-10-30 01:06:41,442 INFO MainThread:321596 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': 'train/train_deep_wandb.py', 'program_abspath': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py', 'program': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py'}
8
+ 2024-10-30 01:06:41,442 INFO MainThread:321596 [wandb_setup.py:_flush():79] Applying login settings: {}
9
+ 2024-10-30 01:06:41,442 INFO MainThread:321596 [wandb_init.py:_log_setup():534] Logging user logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241030_010641-ig1px172/logs/debug.log
10
+ 2024-10-30 01:06:41,442 INFO MainThread:321596 [wandb_init.py:_log_setup():535] Logging internal logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241030_010641-ig1px172/logs/debug-internal.log
11
+ 2024-10-30 01:06:41,442 INFO MainThread:321596 [wandb_init.py:init():621] calling init triggers
12
+ 2024-10-30 01:06:41,442 INFO MainThread:321596 [wandb_init.py:init():628] wandb.init called with sweep_config: {}
13
+ config: {}
14
+ 2024-10-30 01:06:41,442 INFO MainThread:321596 [wandb_init.py:init():671] starting backend
15
+ 2024-10-30 01:06:41,442 INFO MainThread:321596 [wandb_init.py:init():675] sending inform_init request
16
+ 2024-10-30 01:06:41,443 INFO MainThread:321596 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2024-10-30 01:06:41,443 INFO MainThread:321596 [wandb_init.py:init():688] backend started and connected
18
+ 2024-10-30 01:06:41,446 INFO MainThread:321596 [wandb_init.py:init():783] updated telemetry
19
+ 2024-10-30 01:06:41,466 INFO MainThread:321596 [wandb_init.py:init():816] communicating run to backend with 90.0 second timeout
20
+ 2024-10-30 01:06:43,047 INFO MainThread:321596 [wandb_init.py:init():867] starting run threads in backend
21
+ 2024-10-30 01:06:43,143 INFO MainThread:321596 [wandb_run.py:_console_start():2463] atexit reg
22
+ 2024-10-30 01:06:43,143 INFO MainThread:321596 [wandb_run.py:_redirect():2311] redirect: wrap_raw
23
+ 2024-10-30 01:06:43,143 INFO MainThread:321596 [wandb_run.py:_redirect():2376] Wrapping output streams.
24
+ 2024-10-30 01:06:43,143 INFO MainThread:321596 [wandb_run.py:_redirect():2401] Redirects installed.
25
+ 2024-10-30 01:06:43,144 INFO MainThread:321596 [wandb_init.py:init():911] run started, returning control to user process
26
+ 2024-10-30 01:06:43,145 INFO MainThread:321596 [wandb_run.py:_config_callback():1390] config_cb None None {'perturbation': 'reverse_control', 'train_set': '10M', 'batch_size': 3, 'epoch': 7, 'seed': 0}
27
+ 2024-10-30 01:06:43,145 WARNING MsgRouterThr:321596 [router.py:message_loop():77] message_loop has been closed
wandb/run-20241030_010641-ig1px172/run-ig1px172.wandb ADDED
Binary file (1.57 kB). View file
 
wandb/run-20241030_222932-73t76wgo/files/output.log ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Loading checkpoint shards: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 2/2 [00:18<00:00, 9.40s/it]
2
+ tokenized_valid: Dataset({
3
+ features: ['input_ids', 'attention_mask'],
4
+ num_rows: 600
5
+ })
6
+ /mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/training_args.py:1545: FutureWarning: `evaluation_strategy` is deprecated and will be removed in version 4.46 of 🤗 Transformers. Use `eval_strategy` instead
7
+ warnings.warn(
8
+ [2024-10-30 22:29:54,266] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect)
9
+ [2024-10-30 22:30:04,072] [INFO] [comm.py:652:init_distributed] cdb=None
10
+ Installed CUDA version 11.8 does not match the version torch was compiled with 11.7 but since the APIs are compatible, accepting this combination
11
+ Using /home/chunhui/.cache/torch_extensions/py39_cu117 as PyTorch extensions root...
12
+ Loading extension module cpu_adam...
13
+ Time to load cpu_adam op: 5.045851945877075 seconds
14
+ wandb: WARNING Fatal error while uploading data. Some run data will not be synced, but it will still be written to disk. Use `wandb sync` at the end of the run to try uploading.
15
+ Traceback (most recent call last):
16
+ File "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py", line 219, in <module>
17
+ trainer.train()
18
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/trainer.py", line 2052, in train
19
+ return inner_training_loop(
20
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/trainer.py", line 2388, in _inner_training_loop
21
+ tr_loss_step = self.training_step(model, inputs)
22
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/trainer.py", line 3518, in training_step
23
+ self.accelerator.backward(loss, **kwargs)
24
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/accelerate/accelerator.py", line 2238, in backward
25
+ self.deepspeed_engine_wrapped.backward(loss, **kwargs)
26
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/accelerate/utils/deepspeed.py", line 186, in backward
27
+ self.engine.backward(loss, **kwargs)
28
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/deepspeed/utils/nvtx.py", line 18, in wrapped_fn
29
+ ret_val = func(*args, **kwargs)
30
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/deepspeed/runtime/engine.py", line 2020, in backward
31
+ self.optimizer.backward(loss, retain_graph=retain_graph)
32
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/deepspeed/runtime/zero/stage_1_and_2.py", line 2063, in backward
33
+ self.loss_scaler.backward(loss.float(), retain_graph=retain_graph)
34
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/deepspeed/runtime/fp16/loss_scaler.py", line 63, in backward
35
+ scaled_loss.backward(retain_graph=retain_graph)
36
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/torch/_tensor.py", line 487, in backward
37
+ torch.autograd.backward(
38
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/torch/autograd/__init__.py", line 200, in backward
39
+ Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
40
+ KeyboardInterrupt
wandb/run-20241030_222932-73t76wgo/files/requirements.txt ADDED
@@ -0,0 +1,147 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ funcsigs==1.0.2
2
+ sentry-sdk==2.17.0
3
+ multiprocess==0.70.16
4
+ numpy==1.26.2
5
+ pluralizer==1.2.0
6
+ debugpy==1.6.7
7
+ nvidia-cudnn-cu11==8.5.0.96
8
+ deepspeed==0.15.2
9
+ data==0.4
10
+ pandas==2.1.3
11
+ tomli==2.0.1
12
+ charset-normalizer==3.3.2
13
+ attrs==24.2.0
14
+ aiosignal==1.3.1
15
+ fsspec==2023.10.0
16
+ nvidia-cusparse-cu11==11.7.4.91
17
+ zipp==3.12.0
18
+ mypy-extensions==1.0.0
19
+ datasets==3.0.1
20
+ joblib==1.3.2
21
+ hjson==3.1.0
22
+ traitlets==5.7.1
23
+ stack-data==0.6.0
24
+ transformers==4.45.1
25
+ sympy==1.11.1
26
+ Pygments==2.15.0
27
+ docker-pycreds==0.4.0
28
+ dill==0.3.8
29
+ wheel==0.44.0
30
+ prompt-toolkit==3.0.30
31
+ parso==0.8.3
32
+ ipykernel==6.23.1
33
+ pyarrow==17.0.0
34
+ certifi==2023.11.17
35
+ nvidia-cufft-cu11==10.9.0.58
36
+ six==1.16.0
37
+ pydantic==2.9.2
38
+ click==8.1.7
39
+ nest-asyncio==1.5.6
40
+ gmpy2==2.1.0
41
+ matplotlib==3.8.2
42
+ scipy==1.11.4
43
+ typing_extensions==4.12.2
44
+ statsmodels==0.14.0
45
+ huggingface-hub==0.25.0
46
+ frozenlist==1.4.1
47
+ gpustat==1.1.1
48
+ nvidia-nvtx-cu11==11.7.91
49
+ safetensors==0.4.5
50
+ stanza==1.9.2
51
+ decorator==5.1.1
52
+ seaborn==0.13.0
53
+ sentencepiece==0.2.0
54
+ PyYAML==6.0.1
55
+ black==24.8.0
56
+ protobuf==4.25.1
57
+ pickleshare==0.7.5
58
+ peft==0.13.0
59
+ triton==2.0.0
60
+ nvidia-cuda-runtime-cu11==11.7.99
61
+ Jinja2==3.1.2
62
+ nvidia-cusolver-cu11==11.4.0.1
63
+ executing==1.2.0
64
+ jupyter_client==8.1.0
65
+ pluggy==1.3.0
66
+ cmake==3.30.3
67
+ pytz==2023.3.post1
68
+ aiohappyeyeballs==2.4.2
69
+ kiwisolver==1.4.5
70
+ py-cpuinfo==9.0.0
71
+ Pillow==10.1.0
72
+ ptyprocess==0.7.0
73
+ importlib_resources==6.4.5
74
+ GitPython==3.1.43
75
+ importlib-metadata==6.0.0
76
+ iniconfig==2.0.0
77
+ scikit-learn==1.3.2
78
+ exceptiongroup==1.1.0
79
+ networkx==2.8.6
80
+ accelerate==1.0.0
81
+ nltk==3.8.1
82
+ shutilwhich==1.1.0
83
+ fonttools==4.45.1
84
+ future==0.18.3
85
+ aiohttp==3.10.6
86
+ wcwidth==0.2.5
87
+ idna==3.6
88
+ filelock==3.12.2
89
+ pathspec==0.12.1
90
+ jupyter_core==5.1.0
91
+ lit==18.1.8
92
+ nvidia-curand-cu11==10.2.10.91
93
+ nvidia-cublas-cu11==11.10.3.66
94
+ nvidia-ml-py==12.560.30
95
+ msgpack==1.1.0
96
+ python-dateutil==2.8.2
97
+ blessed==1.20.0
98
+ packaging==23.0
99
+ gitdb==4.0.11
100
+ yarl==1.13.0
101
+ emoji==2.8.0
102
+ tzdata==2023.3
103
+ cycler==0.12.1
104
+ tornado==6.2
105
+ backcall==0.2.0
106
+ plotnine==0.12.4
107
+ ninja==1.11.1.1
108
+ latex==0.7.0
109
+ wandb==0.18.5
110
+ setproctitle==1.3.3
111
+ threadpoolctl==3.2.0
112
+ requests==2.32.3
113
+ pyparsing==3.1.1
114
+ smmap==5.0.1
115
+ pyzmq==23.0.0
116
+ async-timeout==4.0.3
117
+ annotated-types==0.7.0
118
+ matplotlib-inline==0.1.6
119
+ latexcodec==1.0.0
120
+ ipython==8.0.0
121
+ patsy==0.5.3
122
+ contourpy==1.2.0
123
+ multidict==6.1.0
124
+ mizani==0.9.3
125
+ urllib3==2.1.0
126
+ tokenizers==0.20.0
127
+ MarkupSafe==2.1.2
128
+ pip==24.2
129
+ pexpect==4.8.0
130
+ tqdm==4.66.5
131
+ jedi==0.18.2
132
+ pydantic_core==2.23.4
133
+ tempdir==0.7.1
134
+ mpmath==1.2.1
135
+ setuptools==72.1.0
136
+ pytest==7.4.3
137
+ pure-eval==0.2.2
138
+ psutil==5.9.1
139
+ comm==0.1.2
140
+ nvidia-cuda-cupti-cu11==11.7.101
141
+ nvidia-cuda-nvrtc-cu11==11.7.99
142
+ regex==2023.10.3
143
+ platformdirs==2.5.2
144
+ asttokens==2.2.1
145
+ torch==2.0.0
146
+ nvidia-nccl-cu11==2.14.3
147
+ xxhash==3.5.0
wandb/run-20241030_222932-73t76wgo/files/wandb-metadata.json ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-162-generic-x86_64-with-glibc2.31",
3
+ "python": "3.9.19",
4
+ "startedAt": "2024-10-31T02:29:32.412130Z",
5
+ "args": [
6
+ "--perturbation",
7
+ "reverse_control",
8
+ "--train_set",
9
+ "10M",
10
+ "--batch_size",
11
+ "3",
12
+ "--epoch",
13
+ "3",
14
+ "--seed",
15
+ "0"
16
+ ],
17
+ "program": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py",
18
+ "codePath": "train/train_deep_wandb.py",
19
+ "git": {
20
+ "remote": "git@hf.co:Yaning1001/Impossible_llm.git",
21
+ "commit": "ed716cdcfcdea02b67f7ed0f3504c2b1c8b737c4"
22
+ },
23
+ "email": "yaning1001@gmail.com",
24
+ "root": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train",
25
+ "host": "mms-large-2",
26
+ "username": "chunhui",
27
+ "executable": "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/bin/python",
28
+ "codePathLocal": "train_deep_wandb.py",
29
+ "cpu_count": 32,
30
+ "cpu_count_logical": 64,
31
+ "gpu": "NVIDIA RTX A6000",
32
+ "gpu_count": 8,
33
+ "disk": {
34
+ "/": {
35
+ "total": "1888559353856",
36
+ "used": "1710969503744"
37
+ }
38
+ },
39
+ "memory": {
40
+ "total": "202617098240"
41
+ },
42
+ "cpu": {
43
+ "count": 32,
44
+ "countLogical": 64
45
+ },
46
+ "gpu_nvidia": [
47
+ {
48
+ "name": "NVIDIA RTX A6000",
49
+ "memoryTotal": "51527024640",
50
+ "cudaCores": 10752,
51
+ "architecture": "Ampere"
52
+ },
53
+ {
54
+ "name": "NVIDIA RTX A6000",
55
+ "memoryTotal": "51527024640",
56
+ "cudaCores": 10752,
57
+ "architecture": "Ampere"
58
+ },
59
+ {
60
+ "name": "NVIDIA RTX A6000",
61
+ "memoryTotal": "51527024640",
62
+ "cudaCores": 10752,
63
+ "architecture": "Ampere"
64
+ },
65
+ {
66
+ "name": "NVIDIA RTX A6000",
67
+ "memoryTotal": "51527024640",
68
+ "cudaCores": 10752,
69
+ "architecture": "Ampere"
70
+ },
71
+ {
72
+ "name": "NVIDIA RTX A6000",
73
+ "memoryTotal": "51527024640",
74
+ "cudaCores": 10752,
75
+ "architecture": "Ampere"
76
+ },
77
+ {
78
+ "name": "NVIDIA RTX A6000",
79
+ "memoryTotal": "51527024640",
80
+ "cudaCores": 10752,
81
+ "architecture": "Ampere"
82
+ },
83
+ {
84
+ "name": "NVIDIA RTX A6000",
85
+ "memoryTotal": "51527024640",
86
+ "cudaCores": 10752,
87
+ "architecture": "Ampere"
88
+ },
89
+ {
90
+ "name": "NVIDIA RTX A6000",
91
+ "memoryTotal": "51527024640",
92
+ "cudaCores": 10752,
93
+ "architecture": "Ampere"
94
+ }
95
+ ],
96
+ "cudaVersion": "11.8"
97
+ }
wandb/run-20241030_222932-73t76wgo/logs/debug-internal.log ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2024-10-30T22:29:32.414170083-04:00","level":"INFO","msg":"using version","core version":"0.18.5"}
2
+ {"time":"2024-10-30T22:29:32.414182363-04:00","level":"INFO","msg":"created symlink","path":"/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241030_222932-73t76wgo/logs/debug-core.log"}
3
+ {"time":"2024-10-30T22:29:32.521624282-04:00","level":"INFO","msg":"created new stream","id":"73t76wgo"}
4
+ {"time":"2024-10-30T22:29:32.521728433-04:00","level":"INFO","msg":"stream: started","id":"73t76wgo"}
5
+ {"time":"2024-10-30T22:29:32.521755523-04:00","level":"INFO","msg":"sender: started","stream_id":"73t76wgo"}
6
+ {"time":"2024-10-30T22:29:32.521740263-04:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"73t76wgo"}}
7
+ {"time":"2024-10-30T22:29:32.521770803-04:00","level":"INFO","msg":"handler: started","stream_id":{"value":"73t76wgo"}}
8
+ {"time":"2024-10-30T22:29:32.718393175-04:00","level":"INFO","msg":"Starting system monitor"}
9
+ {"time":"2024-10-30T22:55:33.122240017-04:00","level":"ERROR","msg":"HTTP error","status":404,"method":"POST","url":"https://api.wandb.ai/files/yaning1001-dartmouth-college/impossible_llm_reverse/73t76wgo/file_stream"}
10
+ {"time":"2024-10-30T22:55:33.126562902-04:00","level":"ERROR+4","msg":"filestream: fatal error: filestream: failed to upload: 404 Not Found path=files/yaning1001-dartmouth-college/impossible_llm_reverse/73t76wgo/file_stream: {\"error\":\"run impossible_llm_reverse/73t76wgo not found while streaming file\"}"}
wandb/run-20241030_222932-73t76wgo/logs/debug.log ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-10-30 22:29:32,410 INFO MainThread:447698 [wandb_setup.py:_flush():79] Current SDK version is 0.18.5
2
+ 2024-10-30 22:29:32,410 INFO MainThread:447698 [wandb_setup.py:_flush():79] Configure stats pid to 447698
3
+ 2024-10-30 22:29:32,410 INFO MainThread:447698 [wandb_setup.py:_flush():79] Loading settings from /home/chunhui/.config/wandb/settings
4
+ 2024-10-30 22:29:32,410 INFO MainThread:447698 [wandb_setup.py:_flush():79] Loading settings from /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/settings
5
+ 2024-10-30 22:29:32,410 INFO MainThread:447698 [wandb_setup.py:_flush():79] Loading settings from environment variables: {}
6
+ 2024-10-30 22:29:32,410 INFO MainThread:447698 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': None, '_disable_service': None}
7
+ 2024-10-30 22:29:32,410 INFO MainThread:447698 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': 'train/train_deep_wandb.py', 'program_abspath': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py', 'program': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py'}
8
+ 2024-10-30 22:29:32,410 INFO MainThread:447698 [wandb_setup.py:_flush():79] Applying login settings: {}
9
+ 2024-10-30 22:29:32,410 INFO MainThread:447698 [wandb_init.py:_log_setup():534] Logging user logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241030_222932-73t76wgo/logs/debug.log
10
+ 2024-10-30 22:29:32,410 INFO MainThread:447698 [wandb_init.py:_log_setup():535] Logging internal logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241030_222932-73t76wgo/logs/debug-internal.log
11
+ 2024-10-30 22:29:32,410 INFO MainThread:447698 [wandb_init.py:init():621] calling init triggers
12
+ 2024-10-30 22:29:32,410 INFO MainThread:447698 [wandb_init.py:init():628] wandb.init called with sweep_config: {}
13
+ config: {}
14
+ 2024-10-30 22:29:32,410 INFO MainThread:447698 [wandb_init.py:init():671] starting backend
15
+ 2024-10-30 22:29:32,410 INFO MainThread:447698 [wandb_init.py:init():675] sending inform_init request
16
+ 2024-10-30 22:29:32,411 INFO MainThread:447698 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2024-10-30 22:29:32,411 INFO MainThread:447698 [wandb_init.py:init():688] backend started and connected
18
+ 2024-10-30 22:29:32,415 INFO MainThread:447698 [wandb_init.py:init():783] updated telemetry
19
+ 2024-10-30 22:29:32,447 INFO MainThread:447698 [wandb_init.py:init():816] communicating run to backend with 90.0 second timeout
20
+ 2024-10-30 22:29:32,714 INFO MainThread:447698 [wandb_init.py:init():867] starting run threads in backend
21
+ 2024-10-30 22:29:32,823 INFO MainThread:447698 [wandb_run.py:_console_start():2463] atexit reg
22
+ 2024-10-30 22:29:32,823 INFO MainThread:447698 [wandb_run.py:_redirect():2311] redirect: wrap_raw
23
+ 2024-10-30 22:29:32,823 INFO MainThread:447698 [wandb_run.py:_redirect():2376] Wrapping output streams.
24
+ 2024-10-30 22:29:32,823 INFO MainThread:447698 [wandb_run.py:_redirect():2401] Redirects installed.
25
+ 2024-10-30 22:29:32,825 INFO MainThread:447698 [wandb_init.py:init():911] run started, returning control to user process
26
+ 2024-10-30 22:29:32,825 INFO MainThread:447698 [wandb_run.py:_config_callback():1390] config_cb None None {'perturbation': 'reverse_control', 'train_set': '10M', 'batch_size': 3, 'epoch': 3, 'seed': 0}
wandb/run-20241030_225833-5mf4xvmz/run-5mf4xvmz.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff1e5d3401f4c043bfc35fdb9ffe7a85d9356ed56e86883fa1d4efd013cf8529
3
+ size 720896
wandb/run-20241030_225833-a1ftigpc/run-a1ftigpc.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ceda6532ed28cc374b6ea26a5f7ee5b5e7080c57164f2dbcf50c6442b3792596
3
+ size 1114112
wandb/run-20241030_225833-h0mm818w/files/output.log ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ config.json: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 844/844 [00:00<00:00, 353kB/s]
2
+ Downloading shards: 100%|████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 2/2 [02:32<00:00, 76.31s/it]
3
+ Loading checkpoint shards: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 2/2 [00:05<00:00, 2.52s/it]
4
+ Map: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 18140/18140 [00:57<00:00, 315.81 examples/s]
5
+ tokenized_valid: Dataset({
6
+ features: ['input_ids', 'attention_mask'],
7
+ num_rows: 600
8
+ })
9
+ /mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/training_args.py:1545: FutureWarning: `evaluation_strategy` is deprecated and will be removed in version 4.46 of 🤗 Transformers. Use `eval_strategy` instead
10
+ warnings.warn(
11
+ [2024-10-30 23:02:11,098] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect)
12
+ [2024-10-30 23:02:20,163] [INFO] [comm.py:652:init_distributed] cdb=None
13
+ Installed CUDA version 11.8 does not match the version torch was compiled with 11.7 but since the APIs are compatible, accepting this combination
14
+ Using /home/chunhui/.cache/torch_extensions/py39_cu117 as PyTorch extensions root...
15
+ Loading extension module cpu_adam...
16
+ Time to load cpu_adam op: 4.8812315464019775 seconds
17
+ Traceback (most recent call last):
18
+ File "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py", line 219, in <module>
19
+ trainer.train()
20
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/trainer.py", line 2052, in train
21
+ return inner_training_loop(
22
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/trainer.py", line 2388, in _inner_training_loop
23
+ tr_loss_step = self.training_step(model, inputs)
24
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/trainer.py", line 3518, in training_step
25
+ self.accelerator.backward(loss, **kwargs)
26
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/accelerate/accelerator.py", line 2238, in backward
27
+ self.deepspeed_engine_wrapped.backward(loss, **kwargs)
28
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/accelerate/utils/deepspeed.py", line 195, in backward
29
+ self.engine.step()
30
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/deepspeed/runtime/engine.py", line 2217, in step
31
+ self.tput_timer.stop(global_step=self.is_gradient_accumulation_boundary(), report_speed=report_progress)
32
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/deepspeed/utils/timer.py", line 256, in stop
33
+ get_accelerator().synchronize()
34
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/deepspeed/accelerator/cuda_accelerator.py", line 79, in synchronize
35
+ return torch.cuda.synchronize(device_index)
36
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/torch/cuda/__init__.py", line 688, in synchronize
37
+ return torch._C._cuda_synchronize()
38
+ KeyboardInterrupt
wandb/run-20241030_225833-h0mm818w/files/requirements.txt ADDED
@@ -0,0 +1,147 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ funcsigs==1.0.2
2
+ sentry-sdk==2.17.0
3
+ multiprocess==0.70.16
4
+ numpy==1.26.2
5
+ pluralizer==1.2.0
6
+ debugpy==1.6.7
7
+ nvidia-cudnn-cu11==8.5.0.96
8
+ deepspeed==0.15.2
9
+ data==0.4
10
+ pandas==2.1.3
11
+ tomli==2.0.1
12
+ charset-normalizer==3.3.2
13
+ attrs==24.2.0
14
+ aiosignal==1.3.1
15
+ fsspec==2023.10.0
16
+ nvidia-cusparse-cu11==11.7.4.91
17
+ zipp==3.12.0
18
+ mypy-extensions==1.0.0
19
+ datasets==3.0.1
20
+ joblib==1.3.2
21
+ hjson==3.1.0
22
+ traitlets==5.7.1
23
+ stack-data==0.6.0
24
+ transformers==4.45.1
25
+ sympy==1.11.1
26
+ Pygments==2.15.0
27
+ docker-pycreds==0.4.0
28
+ dill==0.3.8
29
+ wheel==0.44.0
30
+ prompt-toolkit==3.0.30
31
+ parso==0.8.3
32
+ ipykernel==6.23.1
33
+ pyarrow==17.0.0
34
+ certifi==2023.11.17
35
+ nvidia-cufft-cu11==10.9.0.58
36
+ six==1.16.0
37
+ pydantic==2.9.2
38
+ click==8.1.7
39
+ nest-asyncio==1.5.6
40
+ gmpy2==2.1.0
41
+ matplotlib==3.8.2
42
+ scipy==1.11.4
43
+ typing_extensions==4.12.2
44
+ statsmodels==0.14.0
45
+ huggingface-hub==0.25.0
46
+ frozenlist==1.4.1
47
+ gpustat==1.1.1
48
+ nvidia-nvtx-cu11==11.7.91
49
+ safetensors==0.4.5
50
+ stanza==1.9.2
51
+ decorator==5.1.1
52
+ seaborn==0.13.0
53
+ sentencepiece==0.2.0
54
+ PyYAML==6.0.1
55
+ black==24.8.0
56
+ protobuf==4.25.1
57
+ pickleshare==0.7.5
58
+ peft==0.13.0
59
+ triton==2.0.0
60
+ nvidia-cuda-runtime-cu11==11.7.99
61
+ Jinja2==3.1.2
62
+ nvidia-cusolver-cu11==11.4.0.1
63
+ executing==1.2.0
64
+ jupyter_client==8.1.0
65
+ pluggy==1.3.0
66
+ cmake==3.30.3
67
+ pytz==2023.3.post1
68
+ aiohappyeyeballs==2.4.2
69
+ kiwisolver==1.4.5
70
+ py-cpuinfo==9.0.0
71
+ Pillow==10.1.0
72
+ ptyprocess==0.7.0
73
+ importlib_resources==6.4.5
74
+ GitPython==3.1.43
75
+ importlib-metadata==6.0.0
76
+ iniconfig==2.0.0
77
+ scikit-learn==1.3.2
78
+ exceptiongroup==1.1.0
79
+ networkx==2.8.6
80
+ accelerate==1.0.0
81
+ nltk==3.8.1
82
+ shutilwhich==1.1.0
83
+ fonttools==4.45.1
84
+ future==0.18.3
85
+ aiohttp==3.10.6
86
+ wcwidth==0.2.5
87
+ idna==3.6
88
+ filelock==3.12.2
89
+ pathspec==0.12.1
90
+ jupyter_core==5.1.0
91
+ lit==18.1.8
92
+ nvidia-curand-cu11==10.2.10.91
93
+ nvidia-cublas-cu11==11.10.3.66
94
+ nvidia-ml-py==12.560.30
95
+ msgpack==1.1.0
96
+ python-dateutil==2.8.2
97
+ blessed==1.20.0
98
+ packaging==23.0
99
+ gitdb==4.0.11
100
+ yarl==1.13.0
101
+ emoji==2.8.0
102
+ tzdata==2023.3
103
+ cycler==0.12.1
104
+ tornado==6.2
105
+ backcall==0.2.0
106
+ plotnine==0.12.4
107
+ ninja==1.11.1.1
108
+ latex==0.7.0
109
+ wandb==0.18.5
110
+ setproctitle==1.3.3
111
+ threadpoolctl==3.2.0
112
+ requests==2.32.3
113
+ pyparsing==3.1.1
114
+ smmap==5.0.1
115
+ pyzmq==23.0.0
116
+ async-timeout==4.0.3
117
+ annotated-types==0.7.0
118
+ matplotlib-inline==0.1.6
119
+ latexcodec==1.0.0
120
+ ipython==8.0.0
121
+ patsy==0.5.3
122
+ contourpy==1.2.0
123
+ multidict==6.1.0
124
+ mizani==0.9.3
125
+ urllib3==2.1.0
126
+ tokenizers==0.20.0
127
+ MarkupSafe==2.1.2
128
+ pip==24.2
129
+ pexpect==4.8.0
130
+ tqdm==4.66.5
131
+ jedi==0.18.2
132
+ pydantic_core==2.23.4
133
+ tempdir==0.7.1
134
+ mpmath==1.2.1
135
+ setuptools==72.1.0
136
+ pytest==7.4.3
137
+ pure-eval==0.2.2
138
+ psutil==5.9.1
139
+ comm==0.1.2
140
+ nvidia-cuda-cupti-cu11==11.7.101
141
+ nvidia-cuda-nvrtc-cu11==11.7.99
142
+ regex==2023.10.3
143
+ platformdirs==2.5.2
144
+ asttokens==2.2.1
145
+ torch==2.0.0
146
+ nvidia-nccl-cu11==2.14.3
147
+ xxhash==3.5.0
wandb/run-20241030_225833-h0mm818w/files/wandb-metadata.json ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-162-generic-x86_64-with-glibc2.31",
3
+ "python": "3.9.19",
4
+ "startedAt": "2024-10-31T02:58:33.435474Z",
5
+ "args": [
6
+ "--perturbation",
7
+ "reverse_full",
8
+ "--train_set",
9
+ "10M",
10
+ "--batch_size",
11
+ "3",
12
+ "--epoch",
13
+ "3",
14
+ "--seed",
15
+ "0"
16
+ ],
17
+ "program": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py",
18
+ "codePath": "train/train_deep_wandb.py",
19
+ "git": {
20
+ "remote": "git@hf.co:Yaning1001/Impossible_llm.git",
21
+ "commit": "ed716cdcfcdea02b67f7ed0f3504c2b1c8b737c4"
22
+ },
23
+ "email": "yaning1001@gmail.com",
24
+ "root": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train",
25
+ "host": "mms-large-2",
26
+ "username": "chunhui",
27
+ "executable": "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/bin/python",
28
+ "codePathLocal": "train_deep_wandb.py",
29
+ "cpu_count": 32,
30
+ "cpu_count_logical": 64,
31
+ "gpu": "NVIDIA RTX A6000",
32
+ "gpu_count": 8,
33
+ "disk": {
34
+ "/": {
35
+ "total": "1888559353856",
36
+ "used": "1710970511360"
37
+ }
38
+ },
39
+ "memory": {
40
+ "total": "202617098240"
41
+ },
42
+ "cpu": {
43
+ "count": 32,
44
+ "countLogical": 64
45
+ },
46
+ "gpu_nvidia": [
47
+ {
48
+ "name": "NVIDIA RTX A6000",
49
+ "memoryTotal": "51527024640",
50
+ "cudaCores": 10752,
51
+ "architecture": "Ampere"
52
+ },
53
+ {
54
+ "name": "NVIDIA RTX A6000",
55
+ "memoryTotal": "51527024640",
56
+ "cudaCores": 10752,
57
+ "architecture": "Ampere"
58
+ },
59
+ {
60
+ "name": "NVIDIA RTX A6000",
61
+ "memoryTotal": "51527024640",
62
+ "cudaCores": 10752,
63
+ "architecture": "Ampere"
64
+ },
65
+ {
66
+ "name": "NVIDIA RTX A6000",
67
+ "memoryTotal": "51527024640",
68
+ "cudaCores": 10752,
69
+ "architecture": "Ampere"
70
+ },
71
+ {
72
+ "name": "NVIDIA RTX A6000",
73
+ "memoryTotal": "51527024640",
74
+ "cudaCores": 10752,
75
+ "architecture": "Ampere"
76
+ },
77
+ {
78
+ "name": "NVIDIA RTX A6000",
79
+ "memoryTotal": "51527024640",
80
+ "cudaCores": 10752,
81
+ "architecture": "Ampere"
82
+ },
83
+ {
84
+ "name": "NVIDIA RTX A6000",
85
+ "memoryTotal": "51527024640",
86
+ "cudaCores": 10752,
87
+ "architecture": "Ampere"
88
+ },
89
+ {
90
+ "name": "NVIDIA RTX A6000",
91
+ "memoryTotal": "51527024640",
92
+ "cudaCores": 10752,
93
+ "architecture": "Ampere"
94
+ }
95
+ ],
96
+ "cudaVersion": "11.8"
97
+ }
wandb/run-20241030_225833-h0mm818w/logs/debug-internal.log ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {"time":"2024-10-30T22:58:33.43741463-04:00","level":"INFO","msg":"using version","core version":"0.18.5"}
2
+ {"time":"2024-10-30T22:58:33.437437561-04:00","level":"INFO","msg":"created symlink","path":"/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241030_225833-h0mm818w/logs/debug-core.log"}
3
+ {"time":"2024-10-30T22:58:33.545709195-04:00","level":"INFO","msg":"created new stream","id":"h0mm818w"}
4
+ {"time":"2024-10-30T22:58:33.545745915-04:00","level":"INFO","msg":"stream: started","id":"h0mm818w"}
5
+ {"time":"2024-10-30T22:58:33.545768925-04:00","level":"INFO","msg":"sender: started","stream_id":"h0mm818w"}
6
+ {"time":"2024-10-30T22:58:33.545802915-04:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"h0mm818w"}}
7
+ {"time":"2024-10-30T22:58:33.545762365-04:00","level":"INFO","msg":"handler: started","stream_id":{"value":"h0mm818w"}}
8
+ {"time":"2024-10-30T22:58:33.742833825-04:00","level":"INFO","msg":"Starting system monitor"}
wandb/run-20241030_225833-h0mm818w/logs/debug.log ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-10-30 22:58:33,433 INFO MainThread:451910 [wandb_setup.py:_flush():79] Current SDK version is 0.18.5
2
+ 2024-10-30 22:58:33,433 INFO MainThread:451910 [wandb_setup.py:_flush():79] Configure stats pid to 451910
3
+ 2024-10-30 22:58:33,433 INFO MainThread:451910 [wandb_setup.py:_flush():79] Loading settings from /home/chunhui/.config/wandb/settings
4
+ 2024-10-30 22:58:33,433 INFO MainThread:451910 [wandb_setup.py:_flush():79] Loading settings from /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/settings
5
+ 2024-10-30 22:58:33,433 INFO MainThread:451910 [wandb_setup.py:_flush():79] Loading settings from environment variables: {}
6
+ 2024-10-30 22:58:33,433 INFO MainThread:451910 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': None, '_disable_service': None}
7
+ 2024-10-30 22:58:33,433 INFO MainThread:451910 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': 'train/train_deep_wandb.py', 'program_abspath': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py', 'program': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py'}
8
+ 2024-10-30 22:58:33,433 INFO MainThread:451910 [wandb_setup.py:_flush():79] Applying login settings: {}
9
+ 2024-10-30 22:58:33,434 INFO MainThread:451910 [wandb_init.py:_log_setup():534] Logging user logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241030_225833-h0mm818w/logs/debug.log
10
+ 2024-10-30 22:58:33,434 INFO MainThread:451910 [wandb_init.py:_log_setup():535] Logging internal logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241030_225833-h0mm818w/logs/debug-internal.log
11
+ 2024-10-30 22:58:33,434 INFO MainThread:451910 [wandb_init.py:init():621] calling init triggers
12
+ 2024-10-30 22:58:33,434 INFO MainThread:451910 [wandb_init.py:init():628] wandb.init called with sweep_config: {}
13
+ config: {}
14
+ 2024-10-30 22:58:33,434 INFO MainThread:451910 [wandb_init.py:init():671] starting backend
15
+ 2024-10-30 22:58:33,434 INFO MainThread:451910 [wandb_init.py:init():675] sending inform_init request
16
+ 2024-10-30 22:58:33,435 INFO MainThread:451910 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2024-10-30 22:58:33,435 INFO MainThread:451910 [wandb_init.py:init():688] backend started and connected
18
+ 2024-10-30 22:58:33,437 INFO MainThread:451910 [wandb_init.py:init():783] updated telemetry
19
+ 2024-10-30 22:58:33,465 INFO MainThread:451910 [wandb_init.py:init():816] communicating run to backend with 90.0 second timeout
20
+ 2024-10-30 22:58:33,739 INFO MainThread:451910 [wandb_init.py:init():867] starting run threads in backend
21
+ 2024-10-30 22:58:33,852 INFO MainThread:451910 [wandb_run.py:_console_start():2463] atexit reg
22
+ 2024-10-30 22:58:33,852 INFO MainThread:451910 [wandb_run.py:_redirect():2311] redirect: wrap_raw
23
+ 2024-10-30 22:58:33,852 INFO MainThread:451910 [wandb_run.py:_redirect():2376] Wrapping output streams.
24
+ 2024-10-30 22:58:33,852 INFO MainThread:451910 [wandb_run.py:_redirect():2401] Redirects installed.
25
+ 2024-10-30 22:58:33,853 INFO MainThread:451910 [wandb_init.py:init():911] run started, returning control to user process
26
+ 2024-10-30 22:58:33,853 INFO MainThread:451910 [wandb_run.py:_config_callback():1390] config_cb None None {'perturbation': 'reverse_full', 'train_set': '10M', 'batch_size': 3, 'epoch': 3, 'seed': 0}
wandb/run-20241030_231835-aqeiow5c/run-aqeiow5c.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b876b97d67b9487b44cd4c5e28efb8a5bf360988426b01686c35b0986916a243
3
+ size 655360
wandb/run-20241030_233740-a8ghkt6q/run-a8ghkt6q.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ddc4d6b02cbced2fdd39b8df729f1bf87631bf68cfc846a22b3e1599a95dbc73
3
+ size 851968
wandb/run-20241031_122113-f74sq78x/files/config.yaml ADDED
@@ -0,0 +1,531 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _name_or_path:
2
+ value: meta-llama/Llama-3.2-3B
3
+ _wandb:
4
+ value:
5
+ cli_version: 0.18.5
6
+ m:
7
+ - "1": train/epoch
8
+ "5": 2
9
+ "6":
10
+ - 1
11
+ - 3
12
+ "7": []
13
+ - "1": train/global_step
14
+ "6":
15
+ - 3
16
+ "7": []
17
+ - "1": eval/loss
18
+ "5": 2
19
+ "6":
20
+ - 1
21
+ - 3
22
+ "7": []
23
+ - "1": eval/runtime
24
+ "5": 2
25
+ "6":
26
+ - 1
27
+ - 3
28
+ "7": []
29
+ - "1": eval/samples_per_second
30
+ "5": 2
31
+ "6":
32
+ - 1
33
+ - 3
34
+ "7": []
35
+ - "1": eval/steps_per_second
36
+ "5": 2
37
+ "6":
38
+ - 1
39
+ - 3
40
+ "7": []
41
+ - "1": train/grad_norm
42
+ "5": 2
43
+ "6":
44
+ - 1
45
+ - 3
46
+ "7": []
47
+ - "1": train/learning_rate
48
+ "5": 2
49
+ "6":
50
+ - 1
51
+ - 3
52
+ "7": []
53
+ - "1": train/loss
54
+ "5": 2
55
+ "6":
56
+ - 1
57
+ - 3
58
+ "7": []
59
+ python_version: 3.9.19
60
+ t:
61
+ "1":
62
+ - 1
63
+ - 5
64
+ - 11
65
+ - 49
66
+ - 51
67
+ - 53
68
+ - 55
69
+ - 71
70
+ - 98
71
+ "2":
72
+ - 1
73
+ - 5
74
+ - 11
75
+ - 49
76
+ - 51
77
+ - 53
78
+ - 55
79
+ - 71
80
+ - 98
81
+ "3":
82
+ - 2
83
+ - 7
84
+ - 13
85
+ - 19
86
+ - 23
87
+ - 55
88
+ - 62
89
+ - 66
90
+ "4": 3.9.19
91
+ "5": 0.18.5
92
+ "6": 4.45.1
93
+ "8":
94
+ - 5
95
+ "9":
96
+ "1": transformers_trainer
97
+ "12": 0.18.5
98
+ "13": linux-x86_64
99
+ accelerator_config:
100
+ value:
101
+ dispatch_batches: null
102
+ even_batches: true
103
+ gradient_accumulation_kwargs: null
104
+ non_blocking: false
105
+ split_batches: false
106
+ use_seedable_sampler: true
107
+ adafactor:
108
+ value: false
109
+ adam_beta1:
110
+ value: 0.9
111
+ adam_beta2:
112
+ value: 0.999
113
+ adam_epsilon:
114
+ value: 1e-08
115
+ add_cross_attention:
116
+ value: false
117
+ architectures:
118
+ value:
119
+ - LlamaForCausalLM
120
+ attention_bias:
121
+ value: false
122
+ attention_dropout:
123
+ value: 0
124
+ auto_find_batch_size:
125
+ value: false
126
+ bad_words_ids:
127
+ value: null
128
+ batch_eval_metrics:
129
+ value: false
130
+ batch_size:
131
+ value: 3
132
+ begin_suppress_tokens:
133
+ value: null
134
+ bf16:
135
+ value: false
136
+ bf16_full_eval:
137
+ value: false
138
+ bos_token_id:
139
+ value: 128000
140
+ chunk_size_feed_forward:
141
+ value: 0
142
+ cross_attention_hidden_size:
143
+ value: null
144
+ data_seed:
145
+ value: null
146
+ dataloader_drop_last:
147
+ value: false
148
+ dataloader_num_workers:
149
+ value: 0
150
+ dataloader_persistent_workers:
151
+ value: false
152
+ dataloader_pin_memory:
153
+ value: true
154
+ dataloader_prefetch_factor:
155
+ value: null
156
+ ddp_backend:
157
+ value: null
158
+ ddp_broadcast_buffers:
159
+ value: null
160
+ ddp_bucket_cap_mb:
161
+ value: null
162
+ ddp_find_unused_parameters:
163
+ value: null
164
+ ddp_timeout:
165
+ value: 1800
166
+ debug:
167
+ value: []
168
+ decoder_start_token_id:
169
+ value: null
170
+ deepspeed:
171
+ value: deepspeed_config/train_dp_config.json
172
+ disable_tqdm:
173
+ value: false
174
+ dispatch_batches:
175
+ value: null
176
+ diversity_penalty:
177
+ value: 0
178
+ do_eval:
179
+ value: true
180
+ do_predict:
181
+ value: false
182
+ do_sample:
183
+ value: false
184
+ do_train:
185
+ value: false
186
+ early_stopping:
187
+ value: false
188
+ encoder_no_repeat_ngram_size:
189
+ value: 0
190
+ eos_token_id:
191
+ value: 128001
192
+ epoch:
193
+ value: 6
194
+ eval_accumulation_steps:
195
+ value: null
196
+ eval_delay:
197
+ value: 0
198
+ eval_do_concat_batches:
199
+ value: true
200
+ eval_on_start:
201
+ value: false
202
+ eval_steps:
203
+ value: 10
204
+ eval_strategy:
205
+ value: steps
206
+ eval_use_gather_object:
207
+ value: false
208
+ evaluation_strategy:
209
+ value: steps
210
+ exponential_decay_length_penalty:
211
+ value: null
212
+ finetuning_task:
213
+ value: null
214
+ forced_bos_token_id:
215
+ value: null
216
+ forced_eos_token_id:
217
+ value: null
218
+ fp16:
219
+ value: true
220
+ fp16_backend:
221
+ value: auto
222
+ fp16_full_eval:
223
+ value: false
224
+ fp16_opt_level:
225
+ value: O1
226
+ fsdp:
227
+ value: []
228
+ fsdp_config:
229
+ value:
230
+ min_num_params: 0
231
+ xla: false
232
+ xla_fsdp_grad_ckpt: false
233
+ xla_fsdp_v2: false
234
+ fsdp_min_num_params:
235
+ value: 0
236
+ fsdp_transformer_layer_cls_to_wrap:
237
+ value: null
238
+ full_determinism:
239
+ value: false
240
+ gradient_accumulation_steps:
241
+ value: 2
242
+ gradient_checkpointing:
243
+ value: false
244
+ gradient_checkpointing_kwargs:
245
+ value: null
246
+ greater_is_better:
247
+ value: null
248
+ group_by_length:
249
+ value: false
250
+ half_precision_backend:
251
+ value: auto
252
+ head_dim:
253
+ value: 128
254
+ hidden_act:
255
+ value: silu
256
+ hidden_size:
257
+ value: 3072
258
+ hub_always_push:
259
+ value: false
260
+ hub_model_id:
261
+ value: null
262
+ hub_private_repo:
263
+ value: false
264
+ hub_strategy:
265
+ value: every_save
266
+ hub_token:
267
+ value: <HUB_TOKEN>
268
+ id2label:
269
+ value:
270
+ "0": LABEL_0
271
+ "1": LABEL_1
272
+ ignore_data_skip:
273
+ value: false
274
+ include_inputs_for_metrics:
275
+ value: false
276
+ include_num_input_tokens_seen:
277
+ value: false
278
+ include_tokens_per_second:
279
+ value: false
280
+ initializer_range:
281
+ value: 0.02
282
+ intermediate_size:
283
+ value: 8192
284
+ is_decoder:
285
+ value: false
286
+ is_encoder_decoder:
287
+ value: false
288
+ jit_mode_eval:
289
+ value: false
290
+ label_names:
291
+ value: null
292
+ label_smoothing_factor:
293
+ value: 0
294
+ label2id:
295
+ value:
296
+ LABEL_0: 0
297
+ LABEL_1: 1
298
+ learning_rate:
299
+ value: 5e-06
300
+ length_column_name:
301
+ value: length
302
+ length_penalty:
303
+ value: 1
304
+ load_best_model_at_end:
305
+ value: false
306
+ local_rank:
307
+ value: 0
308
+ log_level:
309
+ value: passive
310
+ log_level_replica:
311
+ value: warning
312
+ log_on_each_node:
313
+ value: true
314
+ logging_dir:
315
+ value: ./logs
316
+ logging_first_step:
317
+ value: false
318
+ logging_nan_inf_filter:
319
+ value: true
320
+ logging_steps:
321
+ value: 1
322
+ logging_strategy:
323
+ value: steps
324
+ lr:
325
+ value: 5e-06
326
+ lr_scheduler_type:
327
+ value: linear
328
+ max_grad_norm:
329
+ value: 1
330
+ max_length:
331
+ value: 20
332
+ max_position_embeddings:
333
+ value: 131072
334
+ max_steps:
335
+ value: -1
336
+ metric_for_best_model:
337
+ value: null
338
+ min_length:
339
+ value: 0
340
+ mlp_bias:
341
+ value: false
342
+ model/num_parameters:
343
+ value: 3212749824
344
+ model_type:
345
+ value: llama
346
+ mp_parameters:
347
+ value: ""
348
+ neftune_noise_alpha:
349
+ value: null
350
+ no_cuda:
351
+ value: false
352
+ no_repeat_ngram_size:
353
+ value: 0
354
+ num_attention_heads:
355
+ value: 24
356
+ num_beam_groups:
357
+ value: 1
358
+ num_beams:
359
+ value: 1
360
+ num_hidden_layers:
361
+ value: 28
362
+ num_key_value_heads:
363
+ value: 8
364
+ num_return_sequences:
365
+ value: 1
366
+ num_train_epochs:
367
+ value: 6
368
+ optim:
369
+ value: adamw_torch
370
+ optim_args:
371
+ value: null
372
+ optim_target_modules:
373
+ value: null
374
+ output_attentions:
375
+ value: false
376
+ output_dir:
377
+ value: ./checkpoints/Llama-3.2-3B/babylm_reverse_full_10M_seed0/runs
378
+ output_hidden_states:
379
+ value: false
380
+ output_scores:
381
+ value: false
382
+ overwrite_output_dir:
383
+ value: false
384
+ pad_token_id:
385
+ value: null
386
+ past_index:
387
+ value: -1
388
+ per_device_eval_batch_size:
389
+ value: 8
390
+ per_device_train_batch_size:
391
+ value: 3
392
+ per_gpu_eval_batch_size:
393
+ value: null
394
+ per_gpu_train_batch_size:
395
+ value: null
396
+ perturbation:
397
+ value: reverse_full
398
+ prediction_loss_only:
399
+ value: false
400
+ prefix:
401
+ value: null
402
+ pretraining_tp:
403
+ value: 1
404
+ problem_type:
405
+ value: null
406
+ push_to_hub:
407
+ value: false
408
+ push_to_hub_model_id:
409
+ value: null
410
+ push_to_hub_organization:
411
+ value: null
412
+ push_to_hub_token:
413
+ value: <PUSH_TO_HUB_TOKEN>
414
+ ray_scope:
415
+ value: last
416
+ remove_invalid_values:
417
+ value: false
418
+ remove_unused_columns:
419
+ value: true
420
+ repetition_penalty:
421
+ value: 1
422
+ report_to:
423
+ value:
424
+ - wandb
425
+ restore_callback_states_from_checkpoint:
426
+ value: false
427
+ resume_from_checkpoint:
428
+ value: null
429
+ return_dict:
430
+ value: true
431
+ return_dict_in_generate:
432
+ value: false
433
+ rms_norm_eps:
434
+ value: 1e-05
435
+ rope_scaling:
436
+ value:
437
+ factor: 32
438
+ high_freq_factor: 4
439
+ low_freq_factor: 1
440
+ original_max_position_embeddings: 8192
441
+ rope_type: llama3
442
+ rope_theta:
443
+ value: 500000
444
+ run_name:
445
+ value: ./checkpoints/Llama-3.2-3B/babylm_reverse_full_10M_seed0/runs
446
+ save_on_each_node:
447
+ value: false
448
+ save_only_model:
449
+ value: false
450
+ save_safetensors:
451
+ value: true
452
+ save_steps:
453
+ value: 150
454
+ save_strategy:
455
+ value: steps
456
+ save_total_limit:
457
+ value: null
458
+ seed:
459
+ value: 0
460
+ sep_token_id:
461
+ value: null
462
+ skip_memory_metrics:
463
+ value: true
464
+ split_batches:
465
+ value: null
466
+ suppress_tokens:
467
+ value: null
468
+ task_specific_params:
469
+ value: null
470
+ temperature:
471
+ value: 1
472
+ tf_legacy_loss:
473
+ value: false
474
+ tf32:
475
+ value: null
476
+ tie_encoder_decoder:
477
+ value: false
478
+ tie_word_embeddings:
479
+ value: true
480
+ tokenizer_class:
481
+ value: null
482
+ top_k:
483
+ value: 50
484
+ top_p:
485
+ value: 1
486
+ torch_compile:
487
+ value: false
488
+ torch_compile_backend:
489
+ value: null
490
+ torch_compile_mode:
491
+ value: null
492
+ torch_dtype:
493
+ value: bfloat16
494
+ torch_empty_cache_steps:
495
+ value: null
496
+ torchdynamo:
497
+ value: null
498
+ torchscript:
499
+ value: false
500
+ tpu_metrics_debug:
501
+ value: false
502
+ tpu_num_cores:
503
+ value: null
504
+ train_set:
505
+ value: 10M
506
+ transformers_version:
507
+ value: 4.45.1
508
+ typical_p:
509
+ value: 1
510
+ use_bfloat16:
511
+ value: false
512
+ use_cache:
513
+ value: true
514
+ use_cpu:
515
+ value: false
516
+ use_ipex:
517
+ value: false
518
+ use_legacy_prediction_loop:
519
+ value: false
520
+ use_liger_kernel:
521
+ value: false
522
+ use_mps_device:
523
+ value: false
524
+ vocab_size:
525
+ value: 128256
526
+ warmup_ratio:
527
+ value: 0
528
+ warmup_steps:
529
+ value: 0
530
+ weight_decay:
531
+ value: 0
wandb/run-20241031_122113-f74sq78x/files/output.log ADDED
The diff for this file is too large to render. See raw diff
 
wandb/run-20241031_122113-f74sq78x/files/wandb-metadata.json ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-162-generic-x86_64-with-glibc2.31",
3
+ "python": "3.9.19",
4
+ "startedAt": "2024-10-31T16:21:13.916582Z",
5
+ "args": [
6
+ "--perturbation",
7
+ "reverse_full",
8
+ "--train_set",
9
+ "10M",
10
+ "--batch_size",
11
+ "3",
12
+ "--epoch",
13
+ "6",
14
+ "--seed",
15
+ "0"
16
+ ],
17
+ "program": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py",
18
+ "codePath": "train/train_deep_wandb.py",
19
+ "git": {
20
+ "remote": "git@hf.co:Yaning1001/Impossible_llm.git",
21
+ "commit": "ed716cdcfcdea02b67f7ed0f3504c2b1c8b737c4"
22
+ },
23
+ "email": "yaning1001@gmail.com",
24
+ "root": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train",
25
+ "host": "mms-large-2",
26
+ "username": "chunhui",
27
+ "executable": "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/bin/python",
28
+ "codePathLocal": "train_deep_wandb.py",
29
+ "cpu_count": 32,
30
+ "cpu_count_logical": 64,
31
+ "gpu": "NVIDIA RTX A6000",
32
+ "gpu_count": 8,
33
+ "disk": {
34
+ "/": {
35
+ "total": "1888559353856",
36
+ "used": "1753159962624"
37
+ }
38
+ },
39
+ "memory": {
40
+ "total": "202617098240"
41
+ },
42
+ "cpu": {
43
+ "count": 32,
44
+ "countLogical": 64
45
+ },
46
+ "gpu_nvidia": [
47
+ {
48
+ "name": "NVIDIA RTX A6000",
49
+ "memoryTotal": "51527024640",
50
+ "cudaCores": 10752,
51
+ "architecture": "Ampere"
52
+ },
53
+ {
54
+ "name": "NVIDIA RTX A6000",
55
+ "memoryTotal": "51527024640",
56
+ "cudaCores": 10752,
57
+ "architecture": "Ampere"
58
+ },
59
+ {
60
+ "name": "NVIDIA RTX A6000",
61
+ "memoryTotal": "51527024640",
62
+ "cudaCores": 10752,
63
+ "architecture": "Ampere"
64
+ },
65
+ {
66
+ "name": "NVIDIA RTX A6000",
67
+ "memoryTotal": "51527024640",
68
+ "cudaCores": 10752,
69
+ "architecture": "Ampere"
70
+ },
71
+ {
72
+ "name": "NVIDIA RTX A6000",
73
+ "memoryTotal": "51527024640",
74
+ "cudaCores": 10752,
75
+ "architecture": "Ampere"
76
+ },
77
+ {
78
+ "name": "NVIDIA RTX A6000",
79
+ "memoryTotal": "51527024640",
80
+ "cudaCores": 10752,
81
+ "architecture": "Ampere"
82
+ },
83
+ {
84
+ "name": "NVIDIA RTX A6000",
85
+ "memoryTotal": "51527024640",
86
+ "cudaCores": 10752,
87
+ "architecture": "Ampere"
88
+ },
89
+ {
90
+ "name": "NVIDIA RTX A6000",
91
+ "memoryTotal": "51527024640",
92
+ "cudaCores": 10752,
93
+ "architecture": "Ampere"
94
+ }
95
+ ],
96
+ "cudaVersion": "11.8"
97
+ }
wandb/run-20241031_122113-f74sq78x/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"eval/runtime":11.6069,"_runtime":32015.260425775,"train/loss":1.1762,"train/grad_norm":3.5276362895965576,"eval/loss":1.8133738040924072,"train_loss":1.3733420533158043,"_timestamp":1.7304236891768625e+09,"train_samples_per_second":3.301,"train/learning_rate":3.080082135523614e-08,"train_steps_per_second":0.092,"eval/steps_per_second":1.12,"eval/samples_per_second":51.693,"total_flos":1.821755618831106e+18,"_wandb":{"runtime":32015},"train/epoch":6,"train_runtime":31844.4073,"_step":3214,"train/global_step":2922}
wandb/run-20241031_122113-f74sq78x/logs/debug-internal.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2024-10-31T12:21:13.919461979-04:00","level":"INFO","msg":"using version","core version":"0.18.5"}
2
+ {"time":"2024-10-31T12:21:13.919481669-04:00","level":"INFO","msg":"created symlink","path":"/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241031_122113-f74sq78x/logs/debug-core.log"}
3
+ {"time":"2024-10-31T12:21:14.028278673-04:00","level":"INFO","msg":"created new stream","id":"f74sq78x"}
4
+ {"time":"2024-10-31T12:21:14.028319283-04:00","level":"INFO","msg":"stream: started","id":"f74sq78x"}
5
+ {"time":"2024-10-31T12:21:14.028368513-04:00","level":"INFO","msg":"sender: started","stream_id":"f74sq78x"}
6
+ {"time":"2024-10-31T12:21:14.028354123-04:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"f74sq78x"}}
7
+ {"time":"2024-10-31T12:21:14.028408043-04:00","level":"INFO","msg":"handler: started","stream_id":{"value":"f74sq78x"}}
8
+ {"time":"2024-10-31T12:21:14.316281133-04:00","level":"INFO","msg":"Starting system monitor"}
9
+ {"time":"2024-10-31T21:14:49.439293898-04:00","level":"INFO","msg":"Stopping system monitor"}
10
+ {"time":"2024-10-31T21:14:49.510534489-04:00","level":"INFO","msg":"Stopped system monitor"}
11
+ {"time":"2024-10-31T21:14:50.197747935-04:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
12
+ {"time":"2024-10-31T21:14:50.329551846-04:00","level":"INFO","msg":"handler: operation stats","stats":{}}
13
+ {"time":"2024-10-31T21:14:51.422197422-04:00","level":"INFO","msg":"stream: closing","id":"f74sq78x"}
14
+ {"time":"2024-10-31T21:14:51.422215052-04:00","level":"INFO","msg":"handler: closed","stream_id":{"value":"f74sq78x"}}
15
+ {"time":"2024-10-31T21:14:51.422230202-04:00","level":"INFO","msg":"sender: closed","stream_id":"f74sq78x"}
16
+ {"time":"2024-10-31T21:14:51.422225522-04:00","level":"INFO","msg":"writer: Close: closed","stream_id":{"value":"f74sq78x"}}
17
+ {"time":"2024-10-31T21:14:51.422273323-04:00","level":"INFO","msg":"stream: closed","id":"f74sq78x"}
wandb/run-20241031_122113-f74sq78x/logs/debug.log ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-10-31 12:21:13,914 INFO MainThread:558430 [wandb_setup.py:_flush():79] Current SDK version is 0.18.5
2
+ 2024-10-31 12:21:13,914 INFO MainThread:558430 [wandb_setup.py:_flush():79] Configure stats pid to 558430
3
+ 2024-10-31 12:21:13,914 INFO MainThread:558430 [wandb_setup.py:_flush():79] Loading settings from /home/chunhui/.config/wandb/settings
4
+ 2024-10-31 12:21:13,914 INFO MainThread:558430 [wandb_setup.py:_flush():79] Loading settings from /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/settings
5
+ 2024-10-31 12:21:13,914 INFO MainThread:558430 [wandb_setup.py:_flush():79] Loading settings from environment variables: {}
6
+ 2024-10-31 12:21:13,914 INFO MainThread:558430 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': None, '_disable_service': None}
7
+ 2024-10-31 12:21:13,914 INFO MainThread:558430 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': 'train/train_deep_wandb.py', 'program_abspath': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py', 'program': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py'}
8
+ 2024-10-31 12:21:13,914 INFO MainThread:558430 [wandb_setup.py:_flush():79] Applying login settings: {}
9
+ 2024-10-31 12:21:13,914 INFO MainThread:558430 [wandb_init.py:_log_setup():534] Logging user logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241031_122113-f74sq78x/logs/debug.log
10
+ 2024-10-31 12:21:13,914 INFO MainThread:558430 [wandb_init.py:_log_setup():535] Logging internal logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241031_122113-f74sq78x/logs/debug-internal.log
11
+ 2024-10-31 12:21:13,914 INFO MainThread:558430 [wandb_init.py:init():621] calling init triggers
12
+ 2024-10-31 12:21:13,914 INFO MainThread:558430 [wandb_init.py:init():628] wandb.init called with sweep_config: {}
13
+ config: {}
14
+ 2024-10-31 12:21:13,914 INFO MainThread:558430 [wandb_init.py:init():671] starting backend
15
+ 2024-10-31 12:21:13,915 INFO MainThread:558430 [wandb_init.py:init():675] sending inform_init request
16
+ 2024-10-31 12:21:13,915 INFO MainThread:558430 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2024-10-31 12:21:13,916 INFO MainThread:558430 [wandb_init.py:init():688] backend started and connected
18
+ 2024-10-31 12:21:13,919 INFO MainThread:558430 [wandb_init.py:init():783] updated telemetry
19
+ 2024-10-31 12:21:13,952 INFO MainThread:558430 [wandb_init.py:init():816] communicating run to backend with 90.0 second timeout
20
+ 2024-10-31 12:21:14,313 INFO MainThread:558430 [wandb_init.py:init():867] starting run threads in backend
21
+ 2024-10-31 12:21:14,398 INFO MainThread:558430 [wandb_run.py:_console_start():2463] atexit reg
22
+ 2024-10-31 12:21:14,398 INFO MainThread:558430 [wandb_run.py:_redirect():2311] redirect: wrap_raw
23
+ 2024-10-31 12:21:14,398 INFO MainThread:558430 [wandb_run.py:_redirect():2376] Wrapping output streams.
24
+ 2024-10-31 12:21:14,398 INFO MainThread:558430 [wandb_run.py:_redirect():2401] Redirects installed.
25
+ 2024-10-31 12:21:14,401 INFO MainThread:558430 [wandb_init.py:init():911] run started, returning control to user process
26
+ 2024-10-31 12:21:14,401 INFO MainThread:558430 [wandb_run.py:_config_callback():1390] config_cb None None {'perturbation': 'reverse_full', 'train_set': '10M', 'batch_size': 3, 'epoch': 6, 'seed': 0, 'lr': 5e-06}
27
+ 2024-10-31 12:24:04,771 INFO MainThread:558430 [wandb_run.py:_config_callback():1390] config_cb None None {'vocab_size': 128256, 'max_position_embeddings': 131072, 'hidden_size': 3072, 'intermediate_size': 8192, 'num_hidden_layers': 28, 'num_attention_heads': 24, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'pretraining_tp': 1, 'use_cache': True, 'rope_theta': 500000.0, 'rope_scaling': {'factor': 32.0, 'high_freq_factor': 4.0, 'low_freq_factor': 1.0, 'original_max_position_embeddings': 8192, 'rope_type': 'llama3'}, 'attention_bias': False, 'attention_dropout': 0.0, 'mlp_bias': False, 'head_dim': 128, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['LlamaForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 128000, 'pad_token_id': None, 'eos_token_id': 128001, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'meta-llama/Llama-3.2-3B', 'transformers_version': '4.45.1', 'model_type': 'llama', 'output_dir': './checkpoints/Llama-3.2-3B/babylm_reverse_full_10M_seed0/runs', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 3, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': None, 'eval_delay': 0, 'torch_empty_cache_steps': None, 'learning_rate': 5e-06, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 6, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './logs', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 150, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 0, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 10, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': './checkpoints/Llama-3.2-3B/babylm_reverse_full_10M_seed0/runs', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': 'deepspeed_config/train_dp_config.json', 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': 'steps', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False, 'eval_on_start': False, 'use_liger_kernel': False, 'eval_use_gather_object': False}
28
+ 2024-10-31 12:24:04,773 INFO MainThread:558430 [wandb_config.py:__setitem__():154] config set model/num_parameters = 3212749824 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x7f57bac78d60>>
29
+ 2024-10-31 12:24:04,773 INFO MainThread:558430 [wandb_run.py:_config_callback():1390] config_cb model/num_parameters 3212749824 None
30
+ 2024-10-31 21:14:49,288 INFO MainThread:558430 [wandb_run.py:_finish():2158] finishing run yaning1001-dartmouth-college/impossible_llm_reverse/f74sq78x
31
+ 2024-10-31 21:14:49,399 INFO MainThread:558430 [wandb_run.py:_atexit_cleanup():2426] got exitcode: 0
32
+ 2024-10-31 21:14:49,400 INFO MainThread:558430 [wandb_run.py:_restore():2408] restore
33
+ 2024-10-31 21:14:49,400 INFO MainThread:558430 [wandb_run.py:_restore():2414] restore done
34
+ 2024-10-31 21:14:51,336 INFO MainThread:558430 [wandb_run.py:_footer_history_summary_info():3975] rendering history
35
+ 2024-10-31 21:14:51,337 INFO MainThread:558430 [wandb_run.py:_footer_history_summary_info():4007] rendering summary
36
+ 2024-10-31 21:14:51,421 INFO MainThread:558430 [wandb_run.py:_footer_sync_info():3934] logging synced files
wandb/run-20241101_093116-hq0hplpe/files/config.yaml ADDED
@@ -0,0 +1,481 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _name_or_path:
2
+ value: meta-llama/Llama-3.2-3B
3
+ _wandb:
4
+ value:
5
+ cli_version: 0.18.5
6
+ m:
7
+ - "1": train/global_step
8
+ "6":
9
+ - 3
10
+ "7": []
11
+ python_version: 3.9.19
12
+ t:
13
+ "1":
14
+ - 1
15
+ - 5
16
+ - 11
17
+ - 49
18
+ - 51
19
+ - 53
20
+ - 55
21
+ - 71
22
+ - 98
23
+ "2":
24
+ - 1
25
+ - 5
26
+ - 11
27
+ - 49
28
+ - 51
29
+ - 53
30
+ - 55
31
+ - 71
32
+ - 98
33
+ "3":
34
+ - 7
35
+ - 13
36
+ - 19
37
+ - 23
38
+ - 55
39
+ - 66
40
+ "4": 3.9.19
41
+ "5": 0.18.5
42
+ "6": 4.45.1
43
+ "8":
44
+ - 5
45
+ "9":
46
+ "1": transformers_trainer
47
+ "12": 0.18.5
48
+ "13": linux-x86_64
49
+ accelerator_config:
50
+ value:
51
+ dispatch_batches: null
52
+ even_batches: true
53
+ gradient_accumulation_kwargs: null
54
+ non_blocking: false
55
+ split_batches: false
56
+ use_seedable_sampler: true
57
+ adafactor:
58
+ value: false
59
+ adam_beta1:
60
+ value: 0.9
61
+ adam_beta2:
62
+ value: 0.999
63
+ adam_epsilon:
64
+ value: 1e-08
65
+ add_cross_attention:
66
+ value: false
67
+ architectures:
68
+ value:
69
+ - LlamaForCausalLM
70
+ attention_bias:
71
+ value: false
72
+ attention_dropout:
73
+ value: 0
74
+ auto_find_batch_size:
75
+ value: false
76
+ bad_words_ids:
77
+ value: null
78
+ batch_eval_metrics:
79
+ value: false
80
+ batch_size:
81
+ value: 3
82
+ begin_suppress_tokens:
83
+ value: null
84
+ bf16:
85
+ value: false
86
+ bf16_full_eval:
87
+ value: false
88
+ bos_token_id:
89
+ value: 128000
90
+ chunk_size_feed_forward:
91
+ value: 0
92
+ cross_attention_hidden_size:
93
+ value: null
94
+ data_seed:
95
+ value: null
96
+ dataloader_drop_last:
97
+ value: false
98
+ dataloader_num_workers:
99
+ value: 0
100
+ dataloader_persistent_workers:
101
+ value: false
102
+ dataloader_pin_memory:
103
+ value: true
104
+ dataloader_prefetch_factor:
105
+ value: null
106
+ ddp_backend:
107
+ value: null
108
+ ddp_broadcast_buffers:
109
+ value: null
110
+ ddp_bucket_cap_mb:
111
+ value: null
112
+ ddp_find_unused_parameters:
113
+ value: null
114
+ ddp_timeout:
115
+ value: 1800
116
+ debug:
117
+ value: []
118
+ decoder_start_token_id:
119
+ value: null
120
+ deepspeed:
121
+ value: deepspeed_config/train_dp_config.json
122
+ disable_tqdm:
123
+ value: false
124
+ dispatch_batches:
125
+ value: null
126
+ diversity_penalty:
127
+ value: 0
128
+ do_eval:
129
+ value: true
130
+ do_predict:
131
+ value: false
132
+ do_sample:
133
+ value: false
134
+ do_train:
135
+ value: false
136
+ early_stopping:
137
+ value: false
138
+ encoder_no_repeat_ngram_size:
139
+ value: 0
140
+ eos_token_id:
141
+ value: 128001
142
+ epoch:
143
+ value: 7
144
+ eval_accumulation_steps:
145
+ value: null
146
+ eval_delay:
147
+ value: 0
148
+ eval_do_concat_batches:
149
+ value: true
150
+ eval_on_start:
151
+ value: false
152
+ eval_steps:
153
+ value: 10
154
+ eval_strategy:
155
+ value: steps
156
+ eval_use_gather_object:
157
+ value: false
158
+ evaluation_strategy:
159
+ value: steps
160
+ exponential_decay_length_penalty:
161
+ value: null
162
+ finetuning_task:
163
+ value: null
164
+ forced_bos_token_id:
165
+ value: null
166
+ forced_eos_token_id:
167
+ value: null
168
+ fp16:
169
+ value: true
170
+ fp16_backend:
171
+ value: auto
172
+ fp16_full_eval:
173
+ value: false
174
+ fp16_opt_level:
175
+ value: O1
176
+ fsdp:
177
+ value: []
178
+ fsdp_config:
179
+ value:
180
+ min_num_params: 0
181
+ xla: false
182
+ xla_fsdp_grad_ckpt: false
183
+ xla_fsdp_v2: false
184
+ fsdp_min_num_params:
185
+ value: 0
186
+ fsdp_transformer_layer_cls_to_wrap:
187
+ value: null
188
+ full_determinism:
189
+ value: false
190
+ gradient_accumulation_steps:
191
+ value: 2
192
+ gradient_checkpointing:
193
+ value: false
194
+ gradient_checkpointing_kwargs:
195
+ value: null
196
+ greater_is_better:
197
+ value: null
198
+ group_by_length:
199
+ value: false
200
+ half_precision_backend:
201
+ value: auto
202
+ head_dim:
203
+ value: 128
204
+ hidden_act:
205
+ value: silu
206
+ hidden_size:
207
+ value: 3072
208
+ hub_always_push:
209
+ value: false
210
+ hub_model_id:
211
+ value: null
212
+ hub_private_repo:
213
+ value: false
214
+ hub_strategy:
215
+ value: every_save
216
+ hub_token:
217
+ value: <HUB_TOKEN>
218
+ id2label:
219
+ value:
220
+ "0": LABEL_0
221
+ "1": LABEL_1
222
+ ignore_data_skip:
223
+ value: false
224
+ include_inputs_for_metrics:
225
+ value: false
226
+ include_num_input_tokens_seen:
227
+ value: false
228
+ include_tokens_per_second:
229
+ value: false
230
+ initializer_range:
231
+ value: 0.02
232
+ intermediate_size:
233
+ value: 8192
234
+ is_decoder:
235
+ value: false
236
+ is_encoder_decoder:
237
+ value: false
238
+ jit_mode_eval:
239
+ value: false
240
+ label_names:
241
+ value: null
242
+ label_smoothing_factor:
243
+ value: 0
244
+ label2id:
245
+ value:
246
+ LABEL_0: 0
247
+ LABEL_1: 1
248
+ learning_rate:
249
+ value: 5e-06
250
+ length_column_name:
251
+ value: length
252
+ length_penalty:
253
+ value: 1
254
+ load_best_model_at_end:
255
+ value: false
256
+ local_rank:
257
+ value: 0
258
+ log_level:
259
+ value: passive
260
+ log_level_replica:
261
+ value: warning
262
+ log_on_each_node:
263
+ value: true
264
+ logging_dir:
265
+ value: ./logs
266
+ logging_first_step:
267
+ value: false
268
+ logging_nan_inf_filter:
269
+ value: true
270
+ logging_steps:
271
+ value: 1
272
+ logging_strategy:
273
+ value: steps
274
+ lr:
275
+ value: 5e-06
276
+ lr_scheduler_type:
277
+ value: linear
278
+ max_grad_norm:
279
+ value: 1
280
+ max_length:
281
+ value: 20
282
+ max_position_embeddings:
283
+ value: 131072
284
+ max_steps:
285
+ value: -1
286
+ metric_for_best_model:
287
+ value: null
288
+ min_length:
289
+ value: 0
290
+ mlp_bias:
291
+ value: false
292
+ model/num_parameters:
293
+ value: 3212749824
294
+ model_type:
295
+ value: llama
296
+ mp_parameters:
297
+ value: ""
298
+ neftune_noise_alpha:
299
+ value: null
300
+ no_cuda:
301
+ value: false
302
+ no_repeat_ngram_size:
303
+ value: 0
304
+ num_attention_heads:
305
+ value: 24
306
+ num_beam_groups:
307
+ value: 1
308
+ num_beams:
309
+ value: 1
310
+ num_hidden_layers:
311
+ value: 28
312
+ num_key_value_heads:
313
+ value: 8
314
+ num_return_sequences:
315
+ value: 1
316
+ num_train_epochs:
317
+ value: 7
318
+ optim:
319
+ value: adamw_torch
320
+ optim_args:
321
+ value: null
322
+ optim_target_modules:
323
+ value: null
324
+ output_attentions:
325
+ value: false
326
+ output_dir:
327
+ value: ./checkpoints/Llama-3.2-3B/babylm_reverse_control_10M_seed0/runs
328
+ output_hidden_states:
329
+ value: false
330
+ output_scores:
331
+ value: false
332
+ overwrite_output_dir:
333
+ value: false
334
+ pad_token_id:
335
+ value: null
336
+ past_index:
337
+ value: -1
338
+ per_device_eval_batch_size:
339
+ value: 8
340
+ per_device_train_batch_size:
341
+ value: 3
342
+ per_gpu_eval_batch_size:
343
+ value: null
344
+ per_gpu_train_batch_size:
345
+ value: null
346
+ perturbation:
347
+ value: reverse_control
348
+ prediction_loss_only:
349
+ value: false
350
+ prefix:
351
+ value: null
352
+ pretraining_tp:
353
+ value: 1
354
+ problem_type:
355
+ value: null
356
+ push_to_hub:
357
+ value: false
358
+ push_to_hub_model_id:
359
+ value: null
360
+ push_to_hub_organization:
361
+ value: null
362
+ push_to_hub_token:
363
+ value: <PUSH_TO_HUB_TOKEN>
364
+ ray_scope:
365
+ value: last
366
+ remove_invalid_values:
367
+ value: false
368
+ remove_unused_columns:
369
+ value: true
370
+ repetition_penalty:
371
+ value: 1
372
+ report_to:
373
+ value:
374
+ - wandb
375
+ restore_callback_states_from_checkpoint:
376
+ value: false
377
+ resume_from_checkpoint:
378
+ value: null
379
+ return_dict:
380
+ value: true
381
+ return_dict_in_generate:
382
+ value: false
383
+ rms_norm_eps:
384
+ value: 1e-05
385
+ rope_scaling:
386
+ value:
387
+ factor: 32
388
+ high_freq_factor: 4
389
+ low_freq_factor: 1
390
+ original_max_position_embeddings: 8192
391
+ rope_type: llama3
392
+ rope_theta:
393
+ value: 500000
394
+ run_name:
395
+ value: ./checkpoints/Llama-3.2-3B/babylm_reverse_control_10M_seed0/runs
396
+ save_on_each_node:
397
+ value: false
398
+ save_only_model:
399
+ value: false
400
+ save_safetensors:
401
+ value: true
402
+ save_steps:
403
+ value: 150
404
+ save_strategy:
405
+ value: steps
406
+ save_total_limit:
407
+ value: null
408
+ seed:
409
+ value: 0
410
+ sep_token_id:
411
+ value: null
412
+ skip_memory_metrics:
413
+ value: true
414
+ split_batches:
415
+ value: null
416
+ suppress_tokens:
417
+ value: null
418
+ task_specific_params:
419
+ value: null
420
+ temperature:
421
+ value: 1
422
+ tf_legacy_loss:
423
+ value: false
424
+ tf32:
425
+ value: null
426
+ tie_encoder_decoder:
427
+ value: false
428
+ tie_word_embeddings:
429
+ value: true
430
+ tokenizer_class:
431
+ value: null
432
+ top_k:
433
+ value: 50
434
+ top_p:
435
+ value: 1
436
+ torch_compile:
437
+ value: false
438
+ torch_compile_backend:
439
+ value: null
440
+ torch_compile_mode:
441
+ value: null
442
+ torch_dtype:
443
+ value: bfloat16
444
+ torch_empty_cache_steps:
445
+ value: null
446
+ torchdynamo:
447
+ value: null
448
+ torchscript:
449
+ value: false
450
+ tpu_metrics_debug:
451
+ value: false
452
+ tpu_num_cores:
453
+ value: null
454
+ train_set:
455
+ value: 10M
456
+ transformers_version:
457
+ value: 4.45.1
458
+ typical_p:
459
+ value: 1
460
+ use_bfloat16:
461
+ value: false
462
+ use_cache:
463
+ value: true
464
+ use_cpu:
465
+ value: false
466
+ use_ipex:
467
+ value: false
468
+ use_legacy_prediction_loop:
469
+ value: false
470
+ use_liger_kernel:
471
+ value: false
472
+ use_mps_device:
473
+ value: false
474
+ vocab_size:
475
+ value: 128256
476
+ warmup_ratio:
477
+ value: 0
478
+ warmup_steps:
479
+ value: 0
480
+ weight_decay:
481
+ value: 0
wandb/run-20241101_093116-hq0hplpe/files/output.log ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Loading checkpoint shards: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 2/2 [00:18<00:00, 9.04s/it]
2
+ tokenized_valid: Dataset({
3
+ features: ['input_ids', 'attention_mask'],
4
+ num_rows: 600
5
+ })
6
+ /mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/training_args.py:1545: FutureWarning: `evaluation_strategy` is deprecated and will be removed in version 4.46 of 🤗 Transformers. Use `eval_strategy` instead
7
+ warnings.warn(
8
+ [2024-11-01 09:31:37,005] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect)
9
+ [2024-11-01 09:31:45,762] [INFO] [comm.py:652:init_distributed] cdb=None
10
+ [2024-11-01 09:31:45,762] [INFO] [comm.py:683:init_distributed] Initializing TorchBackend in DeepSpeed with backend nccl
11
+ Detected kernel version 5.4.0, which is below the recommended minimum of 5.5.0; this can cause the process to hang. It is recommended to upgrade the kernel to the minimum version or higher.
12
+ Installed CUDA version 11.8 does not match the version torch was compiled with 11.7 but since the APIs are compatible, accepting this combination
13
+ Using /home/chunhui/.cache/torch_extensions/py39_cu117 as PyTorch extensions root...
14
+ Loading extension module cpu_adam...
15
+ Time to load cpu_adam op: 4.8145647048950195 seconds
16
+ wandb: WARNING The `run_name` is currently set to the same value as `TrainingArguments.output_dir`. If this was not intended, please specify a different run name by setting the `TrainingArguments.run_name` parameter.
17
+ 0%| | 0/3409 [00:00<?, ?it/s]Traceback (most recent call last):
18
+ File "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py", line 220, in <module>
19
+ trainer.train()
20
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/trainer.py", line 2052, in train
21
+ return inner_training_loop(
22
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/trainer.py", line 2388, in _inner_training_loop
23
+ tr_loss_step = self.training_step(model, inputs)
24
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/trainer.py", line 3485, in training_step
25
+ loss = self.compute_loss(model, inputs)
26
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/trainer.py", line 3532, in compute_loss
27
+ outputs = model(**inputs)
28
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1501, in _call_impl
29
+ return forward_call(*args, **kwargs)
30
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/deepspeed/utils/nvtx.py", line 18, in wrapped_fn
31
+ ret_val = func(*args, **kwargs)
32
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/deepspeed/runtime/engine.py", line 1899, in forward
33
+ loss = self.module(*inputs, **kwargs)
34
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1501, in _call_impl
35
+ return forward_call(*args, **kwargs)
36
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/models/llama/modeling_llama.py", line 1189, in forward
37
+ outputs = self.model(
38
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1501, in _call_impl
39
+ return forward_call(*args, **kwargs)
40
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/models/llama/modeling_llama.py", line 1000, in forward
41
+ layer_outputs = decoder_layer(
42
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1501, in _call_impl
43
+ return forward_call(*args, **kwargs)
44
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/models/llama/modeling_llama.py", line 729, in forward
45
+ hidden_states, self_attn_weights, present_key_value = self.self_attn(
46
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1501, in _call_impl
47
+ return forward_call(*args, **kwargs)
48
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/models/llama/modeling_llama.py", line 426, in forward
49
+ attn_weights = nn.functional.softmax(attn_weights, dim=-1, dtype=torch.float32).to(query_states.dtype)
50
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/torch/nn/functional.py", line 1845, in softmax
51
+ ret = input.softmax(dim, dtype=dtype)
52
+ torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 288.00 MiB (GPU 0; 47.54 GiB total capacity; 11.61 GiB already allocated; 236.56 MiB free; 11.64 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
wandb/run-20241101_093116-hq0hplpe/files/wandb-metadata.json ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-162-generic-x86_64-with-glibc2.31",
3
+ "python": "3.9.19",
4
+ "startedAt": "2024-11-01T13:31:16.509911Z",
5
+ "args": [
6
+ "--perturbation",
7
+ "reverse_control",
8
+ "--train_set",
9
+ "10M",
10
+ "--batch_size",
11
+ "3",
12
+ "--epoch",
13
+ "7",
14
+ "--seed",
15
+ "0"
16
+ ],
17
+ "program": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py",
18
+ "codePath": "train/train_deep_wandb.py",
19
+ "git": {
20
+ "remote": "git@hf.co:Yaning1001/Impossible_llm.git",
21
+ "commit": "ed716cdcfcdea02b67f7ed0f3504c2b1c8b737c4"
22
+ },
23
+ "email": "yaning1001@gmail.com",
24
+ "root": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train",
25
+ "host": "mms-large-2",
26
+ "username": "chunhui",
27
+ "executable": "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/bin/python",
28
+ "codePathLocal": "train_deep_wandb.py",
29
+ "cpu_count": 32,
30
+ "cpu_count_logical": 64,
31
+ "gpu": "NVIDIA RTX A6000",
32
+ "gpu_count": 8,
33
+ "disk": {
34
+ "/": {
35
+ "total": "1888559353856",
36
+ "used": "1754716262400"
37
+ }
38
+ },
39
+ "memory": {
40
+ "total": "202617098240"
41
+ },
42
+ "cpu": {
43
+ "count": 32,
44
+ "countLogical": 64
45
+ },
46
+ "gpu_nvidia": [
47
+ {
48
+ "name": "NVIDIA RTX A6000",
49
+ "memoryTotal": "51527024640",
50
+ "cudaCores": 10752,
51
+ "architecture": "Ampere"
52
+ },
53
+ {
54
+ "name": "NVIDIA RTX A6000",
55
+ "memoryTotal": "51527024640",
56
+ "cudaCores": 10752,
57
+ "architecture": "Ampere"
58
+ },
59
+ {
60
+ "name": "NVIDIA RTX A6000",
61
+ "memoryTotal": "51527024640",
62
+ "cudaCores": 10752,
63
+ "architecture": "Ampere"
64
+ },
65
+ {
66
+ "name": "NVIDIA RTX A6000",
67
+ "memoryTotal": "51527024640",
68
+ "cudaCores": 10752,
69
+ "architecture": "Ampere"
70
+ },
71
+ {
72
+ "name": "NVIDIA RTX A6000",
73
+ "memoryTotal": "51527024640",
74
+ "cudaCores": 10752,
75
+ "architecture": "Ampere"
76
+ },
77
+ {
78
+ "name": "NVIDIA RTX A6000",
79
+ "memoryTotal": "51527024640",
80
+ "cudaCores": 10752,
81
+ "architecture": "Ampere"
82
+ },
83
+ {
84
+ "name": "NVIDIA RTX A6000",
85
+ "memoryTotal": "51527024640",
86
+ "cudaCores": 10752,
87
+ "architecture": "Ampere"
88
+ },
89
+ {
90
+ "name": "NVIDIA RTX A6000",
91
+ "memoryTotal": "51527024640",
92
+ "cudaCores": 10752,
93
+ "architecture": "Ampere"
94
+ }
95
+ ],
96
+ "cudaVersion": "11.8"
97
+ }
wandb/run-20241101_093116-hq0hplpe/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_wandb":{"runtime":66}}
wandb/run-20241101_093116-hq0hplpe/logs/debug-internal.log ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2024-11-01T09:31:16.512440398-04:00","level":"INFO","msg":"using version","core version":"0.18.5"}
2
+ {"time":"2024-11-01T09:31:16.512456718-04:00","level":"INFO","msg":"created symlink","path":"/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241101_093116-hq0hplpe/logs/debug-core.log"}
3
+ {"time":"2024-11-01T09:31:16.620796284-04:00","level":"INFO","msg":"created new stream","id":"hq0hplpe"}
4
+ {"time":"2024-11-01T09:31:16.620829364-04:00","level":"INFO","msg":"stream: started","id":"hq0hplpe"}
5
+ {"time":"2024-11-01T09:31:16.620890365-04:00","level":"INFO","msg":"sender: started","stream_id":"hq0hplpe"}
6
+ {"time":"2024-11-01T09:31:16.620848285-04:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"hq0hplpe"}}
7
+ {"time":"2024-11-01T09:31:16.620867935-04:00","level":"INFO","msg":"handler: started","stream_id":{"value":"hq0hplpe"}}
8
+ {"time":"2024-11-01T09:31:16.815911104-04:00","level":"INFO","msg":"Starting system monitor"}
9
+ {"time":"2024-11-01T09:32:23.327932773-04:00","level":"INFO","msg":"stream: closing","id":"hq0hplpe"}
10
+ {"time":"2024-11-01T09:32:23.327979063-04:00","level":"INFO","msg":"Stopping system monitor"}
11
+ {"time":"2024-11-01T09:32:23.32897053-04:00","level":"INFO","msg":"Stopped system monitor"}
12
+ {"time":"2024-11-01T09:32:23.707909885-04:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
13
+ {"time":"2024-11-01T09:32:23.839808194-04:00","level":"INFO","msg":"handler: closed","stream_id":{"value":"hq0hplpe"}}
14
+ {"time":"2024-11-01T09:32:23.839873144-04:00","level":"INFO","msg":"writer: Close: closed","stream_id":{"value":"hq0hplpe"}}
15
+ {"time":"2024-11-01T09:32:23.839896164-04:00","level":"INFO","msg":"sender: closed","stream_id":"hq0hplpe"}
16
+ {"time":"2024-11-01T09:32:23.839956275-04:00","level":"INFO","msg":"stream: closed","id":"hq0hplpe"}
wandb/run-20241101_093116-hq0hplpe/logs/debug.log ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-11-01 09:31:16,508 INFO MainThread:781945 [wandb_setup.py:_flush():79] Current SDK version is 0.18.5
2
+ 2024-11-01 09:31:16,508 INFO MainThread:781945 [wandb_setup.py:_flush():79] Configure stats pid to 781945
3
+ 2024-11-01 09:31:16,508 INFO MainThread:781945 [wandb_setup.py:_flush():79] Loading settings from /home/chunhui/.config/wandb/settings
4
+ 2024-11-01 09:31:16,508 INFO MainThread:781945 [wandb_setup.py:_flush():79] Loading settings from /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/settings
5
+ 2024-11-01 09:31:16,508 INFO MainThread:781945 [wandb_setup.py:_flush():79] Loading settings from environment variables: {}
6
+ 2024-11-01 09:31:16,508 INFO MainThread:781945 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': None, '_disable_service': None}
7
+ 2024-11-01 09:31:16,508 INFO MainThread:781945 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': 'train/train_deep_wandb.py', 'program_abspath': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py', 'program': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py'}
8
+ 2024-11-01 09:31:16,508 INFO MainThread:781945 [wandb_setup.py:_flush():79] Applying login settings: {}
9
+ 2024-11-01 09:31:16,508 INFO MainThread:781945 [wandb_init.py:_log_setup():534] Logging user logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241101_093116-hq0hplpe/logs/debug.log
10
+ 2024-11-01 09:31:16,508 INFO MainThread:781945 [wandb_init.py:_log_setup():535] Logging internal logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241101_093116-hq0hplpe/logs/debug-internal.log
11
+ 2024-11-01 09:31:16,508 INFO MainThread:781945 [wandb_init.py:init():621] calling init triggers
12
+ 2024-11-01 09:31:16,508 INFO MainThread:781945 [wandb_init.py:init():628] wandb.init called with sweep_config: {}
13
+ config: {}
14
+ 2024-11-01 09:31:16,508 INFO MainThread:781945 [wandb_init.py:init():671] starting backend
15
+ 2024-11-01 09:31:16,508 INFO MainThread:781945 [wandb_init.py:init():675] sending inform_init request
16
+ 2024-11-01 09:31:16,509 INFO MainThread:781945 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2024-11-01 09:31:16,509 INFO MainThread:781945 [wandb_init.py:init():688] backend started and connected
18
+ 2024-11-01 09:31:16,512 INFO MainThread:781945 [wandb_init.py:init():783] updated telemetry
19
+ 2024-11-01 09:31:16,545 INFO MainThread:781945 [wandb_init.py:init():816] communicating run to backend with 90.0 second timeout
20
+ 2024-11-01 09:31:16,812 INFO MainThread:781945 [wandb_init.py:init():867] starting run threads in backend
21
+ 2024-11-01 09:31:16,944 INFO MainThread:781945 [wandb_run.py:_console_start():2463] atexit reg
22
+ 2024-11-01 09:31:16,944 INFO MainThread:781945 [wandb_run.py:_redirect():2311] redirect: wrap_raw
23
+ 2024-11-01 09:31:16,944 INFO MainThread:781945 [wandb_run.py:_redirect():2376] Wrapping output streams.
24
+ 2024-11-01 09:31:16,944 INFO MainThread:781945 [wandb_run.py:_redirect():2401] Redirects installed.
25
+ 2024-11-01 09:31:16,946 INFO MainThread:781945 [wandb_init.py:init():911] run started, returning control to user process
26
+ 2024-11-01 09:31:16,946 INFO MainThread:781945 [wandb_run.py:_config_callback():1390] config_cb None None {'perturbation': 'reverse_control', 'train_set': '10M', 'batch_size': 3, 'epoch': 7, 'seed': 0, 'lr': 5e-06}
27
+ 2024-11-01 09:32:20,281 INFO MainThread:781945 [wandb_run.py:_config_callback():1390] config_cb None None {'vocab_size': 128256, 'max_position_embeddings': 131072, 'hidden_size': 3072, 'intermediate_size': 8192, 'num_hidden_layers': 28, 'num_attention_heads': 24, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'pretraining_tp': 1, 'use_cache': True, 'rope_theta': 500000.0, 'rope_scaling': {'factor': 32.0, 'high_freq_factor': 4.0, 'low_freq_factor': 1.0, 'original_max_position_embeddings': 8192, 'rope_type': 'llama3'}, 'attention_bias': False, 'attention_dropout': 0.0, 'mlp_bias': False, 'head_dim': 128, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['LlamaForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 128000, 'pad_token_id': None, 'eos_token_id': 128001, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'meta-llama/Llama-3.2-3B', 'transformers_version': '4.45.1', 'model_type': 'llama', 'output_dir': './checkpoints/Llama-3.2-3B/babylm_reverse_control_10M_seed0/runs', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 3, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': None, 'eval_delay': 0, 'torch_empty_cache_steps': None, 'learning_rate': 5e-06, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 7, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './logs', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 150, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 0, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 10, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': './checkpoints/Llama-3.2-3B/babylm_reverse_control_10M_seed0/runs', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': 'deepspeed_config/train_dp_config.json', 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': 'steps', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False, 'eval_on_start': False, 'use_liger_kernel': False, 'eval_use_gather_object': False}
28
+ 2024-11-01 09:32:20,287 INFO MainThread:781945 [wandb_config.py:__setitem__():154] config set model/num_parameters = 3212749824 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x7fd6d4899d90>>
29
+ 2024-11-01 09:32:20,287 INFO MainThread:781945 [wandb_run.py:_config_callback():1390] config_cb model/num_parameters 3212749824 None
30
+ 2024-11-01 09:32:23,328 WARNING MsgRouterThr:781945 [router.py:message_loop():77] message_loop has been closed
wandb/run-20241101_093116-hq0hplpe/run-hq0hplpe.wandb ADDED
Binary file (54.5 kB). View file
 
wandb/run-20241101_200535-k9brpado/run-k9brpado.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e20145dca2740f037eb66fb9a8d9baff6bfca31133f1a084bd6cd7114c4048a
3
+ size 131273
wandb/run-20241101_200535-lnp8ii96/run-lnp8ii96.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f46e28d2b91058be18522431c03400139d8a1d74afa9ab1353a8ed260417dd19
3
+ size 425984
wandb/run-20241105_155954-baeezzpr/files/config.yaml ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.18.5
4
+ m: []
5
+ python_version: 3.9.19
6
+ t:
7
+ "1":
8
+ - 1
9
+ - 5
10
+ - 11
11
+ - 49
12
+ - 51
13
+ - 53
14
+ - 55
15
+ - 71
16
+ - 98
17
+ "2":
18
+ - 1
19
+ - 5
20
+ - 11
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 98
27
+ "3":
28
+ - 13
29
+ - 23
30
+ - 55
31
+ "4": 3.9.19
32
+ "5": 0.18.5
33
+ "6": 4.45.1
34
+ "8":
35
+ - 5
36
+ "12": 0.18.5
37
+ "13": linux-x86_64
38
+ batch_size:
39
+ value: 3
40
+ epoch:
41
+ value: 3
42
+ lr:
43
+ value: 5e-06
44
+ perturbation:
45
+ value: shuffle_deterministic21
46
+ seed:
47
+ value: 0
48
+ train_set:
49
+ value: 10M
wandb/run-20241105_155954-baeezzpr/files/output.log ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Traceback (most recent call last):
2
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/pathlib.py", line 1323, in mkdir
3
+ self._accessor.mkdir(self, mode)
4
+ FileNotFoundError: [Errno 2] No such file or directory: '/home/chunhui/.cache/huggingface/datasets/babylm_dataset_test/babylm_shuffle_deterministic21_10M_seed0/0.0.0'
5
+
6
+ During handling of the above exception, another exception occurred:
7
+
8
+ Traceback (most recent call last):
9
+ File "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py", line 165, in <module>
10
+ dataset = load_dataset('babylm_dataset_test.py', name=dataset_name, trust_remote_code=True)
11
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/datasets/load.py", line 2096, in load_dataset
12
+ builder_instance.download_and_prepare(
13
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/datasets/builder.py", line 855, in download_and_prepare
14
+ Path(self._output_dir).parent.mkdir(parents=True, exist_ok=True)
15
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/pathlib.py", line 1327, in mkdir
16
+ self.parent.mkdir(parents=True, exist_ok=True)
17
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/pathlib.py", line 1323, in mkdir
18
+ self._accessor.mkdir(self, mode)
19
+ OSError: [Errno 28] No space left on device: '/home/chunhui/.cache/huggingface/datasets/babylm_dataset_test/babylm_shuffle_deterministic21_10M_seed0'
wandb/run-20241105_155954-baeezzpr/files/requirements.txt ADDED
@@ -0,0 +1,147 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ funcsigs==1.0.2
2
+ sentry-sdk==2.17.0
3
+ multiprocess==0.70.16
4
+ numpy==1.26.2
5
+ pluralizer==1.2.0
6
+ debugpy==1.6.7
7
+ nvidia-cudnn-cu11==8.5.0.96
8
+ deepspeed==0.15.2
9
+ data==0.4
10
+ pandas==2.1.3
11
+ tomli==2.0.1
12
+ charset-normalizer==3.3.2
13
+ attrs==24.2.0
14
+ aiosignal==1.3.1
15
+ fsspec==2023.10.0
16
+ nvidia-cusparse-cu11==11.7.4.91
17
+ zipp==3.12.0
18
+ mypy-extensions==1.0.0
19
+ datasets==3.0.1
20
+ joblib==1.3.2
21
+ hjson==3.1.0
22
+ traitlets==5.7.1
23
+ stack-data==0.6.0
24
+ transformers==4.45.1
25
+ sympy==1.11.1
26
+ Pygments==2.15.0
27
+ docker-pycreds==0.4.0
28
+ dill==0.3.8
29
+ wheel==0.44.0
30
+ prompt-toolkit==3.0.30
31
+ parso==0.8.3
32
+ ipykernel==6.23.1
33
+ pyarrow==17.0.0
34
+ certifi==2023.11.17
35
+ nvidia-cufft-cu11==10.9.0.58
36
+ six==1.16.0
37
+ pydantic==2.9.2
38
+ click==8.1.7
39
+ nest-asyncio==1.5.6
40
+ gmpy2==2.1.0
41
+ matplotlib==3.8.2
42
+ scipy==1.11.4
43
+ typing_extensions==4.12.2
44
+ statsmodels==0.14.0
45
+ huggingface-hub==0.25.0
46
+ frozenlist==1.4.1
47
+ gpustat==1.1.1
48
+ nvidia-nvtx-cu11==11.7.91
49
+ safetensors==0.4.5
50
+ stanza==1.9.2
51
+ decorator==5.1.1
52
+ seaborn==0.13.0
53
+ sentencepiece==0.2.0
54
+ PyYAML==6.0.1
55
+ black==24.8.0
56
+ protobuf==4.25.1
57
+ pickleshare==0.7.5
58
+ peft==0.13.0
59
+ triton==2.0.0
60
+ nvidia-cuda-runtime-cu11==11.7.99
61
+ Jinja2==3.1.2
62
+ nvidia-cusolver-cu11==11.4.0.1
63
+ executing==1.2.0
64
+ jupyter_client==8.1.0
65
+ pluggy==1.3.0
66
+ cmake==3.30.3
67
+ pytz==2023.3.post1
68
+ aiohappyeyeballs==2.4.2
69
+ kiwisolver==1.4.5
70
+ py-cpuinfo==9.0.0
71
+ Pillow==10.1.0
72
+ ptyprocess==0.7.0
73
+ importlib_resources==6.4.5
74
+ GitPython==3.1.43
75
+ importlib-metadata==6.0.0
76
+ iniconfig==2.0.0
77
+ scikit-learn==1.3.2
78
+ exceptiongroup==1.1.0
79
+ networkx==2.8.6
80
+ accelerate==1.0.0
81
+ nltk==3.8.1
82
+ shutilwhich==1.1.0
83
+ fonttools==4.45.1
84
+ future==0.18.3
85
+ aiohttp==3.10.6
86
+ wcwidth==0.2.5
87
+ idna==3.6
88
+ filelock==3.12.2
89
+ pathspec==0.12.1
90
+ jupyter_core==5.1.0
91
+ lit==18.1.8
92
+ nvidia-curand-cu11==10.2.10.91
93
+ nvidia-cublas-cu11==11.10.3.66
94
+ nvidia-ml-py==12.560.30
95
+ msgpack==1.1.0
96
+ python-dateutil==2.8.2
97
+ blessed==1.20.0
98
+ packaging==23.0
99
+ gitdb==4.0.11
100
+ yarl==1.13.0
101
+ emoji==2.8.0
102
+ tzdata==2023.3
103
+ cycler==0.12.1
104
+ tornado==6.2
105
+ backcall==0.2.0
106
+ plotnine==0.12.4
107
+ ninja==1.11.1.1
108
+ latex==0.7.0
109
+ wandb==0.18.5
110
+ setproctitle==1.3.3
111
+ threadpoolctl==3.2.0
112
+ requests==2.32.3
113
+ pyparsing==3.1.1
114
+ smmap==5.0.1
115
+ pyzmq==23.0.0
116
+ async-timeout==4.0.3
117
+ annotated-types==0.7.0
118
+ matplotlib-inline==0.1.6
119
+ latexcodec==1.0.0
120
+ ipython==8.0.0
121
+ patsy==0.5.3
122
+ contourpy==1.2.0
123
+ multidict==6.1.0
124
+ mizani==0.9.3
125
+ urllib3==2.1.0
126
+ tokenizers==0.20.0
127
+ MarkupSafe==2.1.2
128
+ pip==24.2
129
+ pexpect==4.8.0
130
+ tqdm==4.66.5
131
+ jedi==0.18.2
132
+ pydantic_core==2.23.4
133
+ tempdir==0.7.1
134
+ mpmath==1.2.1
135
+ setuptools==72.1.0
136
+ pytest==7.4.3
137
+ pure-eval==0.2.2
138
+ psutil==5.9.1
139
+ comm==0.1.2
140
+ nvidia-cuda-cupti-cu11==11.7.101
141
+ nvidia-cuda-nvrtc-cu11==11.7.99
142
+ regex==2023.10.3
143
+ platformdirs==2.5.2
144
+ asttokens==2.2.1
145
+ torch==2.0.0
146
+ nvidia-nccl-cu11==2.14.3
147
+ xxhash==3.5.0
wandb/run-20241105_155954-baeezzpr/files/wandb-metadata.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-162-generic-x86_64-with-glibc2.31",
3
+ "python": "3.9.19",
4
+ "startedAt": "2024-11-05T20:59:54.274726Z",
5
+ "args": [
6
+ "--perturbation",
7
+ "shuffle_deterministic21",
8
+ "--train_set",
9
+ "10M",
10
+ "--batch_size",
11
+ "3",
12
+ "--epoch",
13
+ "3",
14
+ "--seed",
15
+ "0"
16
+ ],
17
+ "program": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py",
18
+ "codePath": "train/train_deep_wandb.py",
19
+ "git": {
20
+ "remote": "git@hf.co:Yaning1001/Impossible_llm.git",
21
+ "commit": "ed716cdcfcdea02b67f7ed0f3504c2b1c8b737c4"
22
+ },
23
+ "email": "yaning1001@gmail.com",
24
+ "root": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train",
25
+ "host": "mms-large-2",
26
+ "username": "chunhui",
27
+ "executable": "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/bin/python",
28
+ "codePathLocal": "train_deep_wandb.py",
29
+ "cpu_count": 32,
30
+ "cpu_count_logical": 64,
31
+ "disk": {
32
+ "/": {
33
+ "total": "1888559353856",
34
+ "used": "1792550322176"
35
+ }
36
+ },
37
+ "memory": {
38
+ "total": "202617098240"
39
+ },
40
+ "cpu": {
41
+ "count": 32,
42
+ "countLogical": 64
43
+ }
44
+ }
wandb/run-20241105_155954-baeezzpr/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_wandb":{"runtime":5}}
wandb/run-20241105_155954-baeezzpr/logs/debug-internal.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2024-11-05T15:59:54.276983395-05:00","level":"INFO","msg":"using version","core version":"0.18.5"}
2
+ {"time":"2024-11-05T15:59:54.276994555-05:00","level":"INFO","msg":"created symlink","path":"/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241105_155954-baeezzpr/logs/debug-core.log"}
3
+ {"time":"2024-11-05T15:59:59.30189816-05:00","level":"INFO","msg":"created new stream","id":"baeezzpr"}
4
+ {"time":"2024-11-05T15:59:59.30193775-05:00","level":"INFO","msg":"stream: started","id":"baeezzpr"}
5
+ {"time":"2024-11-05T15:59:59.30200871-05:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"baeezzpr"}}
6
+ {"time":"2024-11-05T15:59:59.30205932-05:00","level":"INFO","msg":"handler: started","stream_id":{"value":"baeezzpr"}}
7
+ {"time":"2024-11-05T15:59:59.30206516-05:00","level":"INFO","msg":"sender: started","stream_id":"baeezzpr"}
8
+ {"time":"2024-11-05T15:59:59.498122433-05:00","level":"INFO","msg":"Starting system monitor"}
9
+ {"time":"2024-11-05T15:59:59.604097578-05:00","level":"INFO","msg":"stream: closing","id":"baeezzpr"}
10
+ {"time":"2024-11-05T15:59:59.604120108-05:00","level":"INFO","msg":"Stopping system monitor"}
11
+ {"time":"2024-11-05T15:59:59.604168778-05:00","level":"INFO","msg":"Stopped system monitor"}
12
+ {"time":"2024-11-05T15:59:59.779215414-05:00","level":"ERROR","msg":"sender: sendDefer: failed to build job artifact","error":"failed to write data to file: write /tmp/tmpfile-915389451: no space left on device"}
13
+ {"time":"2024-11-05T16:00:00.039118999-05:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
14
+ {"time":"2024-11-05T16:00:00.178684337-05:00","level":"INFO","msg":"handler: closed","stream_id":{"value":"baeezzpr"}}
15
+ {"time":"2024-11-05T16:00:00.178749977-05:00","level":"INFO","msg":"sender: closed","stream_id":"baeezzpr"}
16
+ {"time":"2024-11-05T16:00:00.178740807-05:00","level":"INFO","msg":"writer: Close: closed","stream_id":{"value":"baeezzpr"}}
17
+ {"time":"2024-11-05T16:00:00.178845027-05:00","level":"INFO","msg":"stream: closed","id":"baeezzpr"}
wandb/run-20241105_155954-baeezzpr/logs/debug.log ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-11-05 15:59:54,271 INFO MainThread:1769192 [wandb_setup.py:_flush():79] Current SDK version is 0.18.5
2
+ 2024-11-05 15:59:54,272 INFO MainThread:1769192 [wandb_setup.py:_flush():79] Configure stats pid to 1769192
3
+ 2024-11-05 15:59:54,272 INFO MainThread:1769192 [wandb_setup.py:_flush():79] Loading settings from /home/chunhui/.config/wandb/settings
4
+ 2024-11-05 15:59:54,272 INFO MainThread:1769192 [wandb_setup.py:_flush():79] Loading settings from /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/settings
5
+ 2024-11-05 15:59:54,272 INFO MainThread:1769192 [wandb_setup.py:_flush():79] Loading settings from environment variables: {}
6
+ 2024-11-05 15:59:54,272 INFO MainThread:1769192 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': None, '_disable_service': None}
7
+ 2024-11-05 15:59:54,272 INFO MainThread:1769192 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': 'train/train_deep_wandb.py', 'program_abspath': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py', 'program': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py'}
8
+ 2024-11-05 15:59:54,272 INFO MainThread:1769192 [wandb_setup.py:_flush():79] Applying login settings: {}
9
+ 2024-11-05 15:59:54,272 INFO MainThread:1769192 [wandb_init.py:_log_setup():534] Logging user logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241105_155954-baeezzpr/logs/debug.log
10
+ 2024-11-05 15:59:54,272 INFO MainThread:1769192 [wandb_init.py:_log_setup():535] Logging internal logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241105_155954-baeezzpr/logs/debug-internal.log
11
+ 2024-11-05 15:59:54,272 INFO MainThread:1769192 [wandb_init.py:init():621] calling init triggers
12
+ 2024-11-05 15:59:54,272 INFO MainThread:1769192 [wandb_init.py:init():628] wandb.init called with sweep_config: {}
13
+ config: {}
14
+ 2024-11-05 15:59:54,272 INFO MainThread:1769192 [wandb_init.py:init():671] starting backend
15
+ 2024-11-05 15:59:54,272 INFO MainThread:1769192 [wandb_init.py:init():675] sending inform_init request
16
+ 2024-11-05 15:59:54,274 INFO MainThread:1769192 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2024-11-05 15:59:54,274 INFO MainThread:1769192 [wandb_init.py:init():688] backend started and connected
18
+ 2024-11-05 15:59:54,278 INFO MainThread:1769192 [wandb_init.py:init():783] updated telemetry
19
+ 2024-11-05 15:59:54,310 INFO MainThread:1769192 [wandb_init.py:init():816] communicating run to backend with 90.0 second timeout
20
+ 2024-11-05 15:59:59,495 INFO MainThread:1769192 [wandb_init.py:init():867] starting run threads in backend
21
+ 2024-11-05 15:59:59,579 INFO MainThread:1769192 [wandb_run.py:_console_start():2463] atexit reg
22
+ 2024-11-05 15:59:59,579 INFO MainThread:1769192 [wandb_run.py:_redirect():2311] redirect: wrap_raw
23
+ 2024-11-05 15:59:59,579 INFO MainThread:1769192 [wandb_run.py:_redirect():2376] Wrapping output streams.
24
+ 2024-11-05 15:59:59,580 INFO MainThread:1769192 [wandb_run.py:_redirect():2401] Redirects installed.
25
+ 2024-11-05 15:59:59,581 INFO MainThread:1769192 [wandb_init.py:init():911] run started, returning control to user process
26
+ 2024-11-05 15:59:59,581 INFO MainThread:1769192 [wandb_run.py:_config_callback():1390] config_cb None None {'perturbation': 'shuffle_deterministic21', 'train_set': '10M', 'batch_size': 3, 'epoch': 3, 'seed': 0, 'lr': 5e-06}
27
+ 2024-11-05 15:59:59,604 WARNING MsgRouterThr:1769192 [router.py:message_loop():77] message_loop has been closed
wandb/run-20241105_155954-baeezzpr/run-baeezzpr.wandb ADDED
Binary file (3.78 kB). View file
 
wandb/run-20241106_224518-az8kcj8i/files/config.yaml ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.18.5
4
+ m: []
5
+ python_version: 3.9.19
6
+ t:
7
+ "1":
8
+ - 1
9
+ - 5
10
+ - 11
11
+ - 49
12
+ - 51
13
+ - 53
14
+ - 55
15
+ - 71
16
+ - 98
17
+ "2":
18
+ - 1
19
+ - 5
20
+ - 11
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 98
27
+ "3":
28
+ - 13
29
+ - 23
30
+ - 55
31
+ "4": 3.9.19
32
+ "5": 0.18.5
33
+ "6": 4.45.1
34
+ "8":
35
+ - 5
36
+ "12": 0.18.5
37
+ "13": linux-x86_64
38
+ batch_size:
39
+ value: 3
40
+ epoch:
41
+ value: 3
42
+ lr:
43
+ value: 5e-06
44
+ perturbation:
45
+ value: shuffle_deterministic84
46
+ seed:
47
+ value: 0
48
+ train_set:
49
+ value: 10M
wandb/run-20241106_224518-az8kcj8i/files/output.log ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Downloading shards: 0%| | 0/2 [00:00<?, ?it/s]Exception ignored in: <generator object tqdm.__iter__ at 0x7f8ad8067cf0>
2
+ Traceback (most recent call last):
3
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/tqdm/std.py", line 1196, in __iter__
4
+ self.close()
5
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/tqdm/std.py", line 1302, in close
6
+ self.display(pos=0)
7
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/tqdm/std.py", line 1495, in display
8
+ self.sp(self.__str__() if msg is None else msg)
9
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/tqdm/std.py", line 459, in print_status
10
+ fp_write('\r' + s + (' ' * max(last_len[0] - len_s, 0)))
11
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/tqdm/std.py", line 452, in fp_write
12
+ fp.write(str(s))
13
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/tqdm/utils.py", line 196, in inner
14
+ return func(*args, **kwargs)
15
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/wandb/sdk/lib/redirect.py", line 648, in write
16
+ cb(data)
17
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/wandb/sdk/wandb_run.py", line 2386, in <lambda>
18
+ lambda data: self._console_raw_callback("stderr", data),
19
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/wandb/sdk/wandb_run.py", line 403, in wrapper_fn
20
+ return func(self, *args, **kwargs)
21
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/wandb/sdk/wandb_run.py", line 1547, in _console_raw_callback
22
+ self._backend.interface.publish_output_raw(name, data)
23
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/wandb/sdk/interface/interface.py", line 721, in publish_output_raw
24
+ self._publish_output_raw(o)
25
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/wandb/sdk/interface/interface_shared.py", line 79, in _publish_output_raw
26
+ self._publish(rec)
27
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/wandb/sdk/interface/interface_sock.py", line 51, in _publish
28
+ self._sock_client.send_record_publish(record)
29
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/wandb/sdk/lib/sock_client.py", line 224, in send_record_publish
30
+ server_req.record_publish.CopyFrom(record)
31
+ KeyboardInterrupt:
32
+ Traceback (most recent call last):
33
+ File "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py", line 174, in <module>
34
+ model = AutoModelForCausalLM.from_pretrained(model_name,
35
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/models/auto/auto_factory.py", line 564, in from_pretrained
36
+ return model_class.from_pretrained(
37
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/modeling_utils.py", line 3769, in from_pretrained
38
+ resolved_archive_file, sharded_metadata = get_checkpoint_shard_files(
39
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/utils/hub.py", line 1098, in get_checkpoint_shard_files
40
+ cached_filename = cached_file(
41
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/utils/hub.py", line 403, in cached_file
42
+ resolved_file = hf_hub_download(
43
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/utils/_deprecation.py", line 101, in inner_f
44
+ return f(*args, **kwargs)
45
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn
46
+ return fn(*args, **kwargs)
47
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/file_download.py", line 1232, in hf_hub_download
48
+ return _hf_hub_download_to_cache_dir(
49
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/file_download.py", line 1380, in _hf_hub_download_to_cache_dir
50
+ with WeakFileLock(lock_path):
51
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/contextlib.py", line 119, in __enter__
52
+ return next(self.gen)
53
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/utils/_fixes.py", line 98, in WeakFileLock
54
+ lock.acquire()
55
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/filelock/_api.py", line 225, in acquire
56
+ time.sleep(poll_interval)
57
+ KeyboardInterrupt
wandb/run-20241106_224518-az8kcj8i/files/wandb-metadata.json ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-162-generic-x86_64-with-glibc2.31",
3
+ "python": "3.9.19",
4
+ "startedAt": "2024-11-07T03:45:18.943895Z",
5
+ "args": [
6
+ "--perturbation",
7
+ "shuffle_deterministic84",
8
+ "--train_set",
9
+ "10M",
10
+ "--batch_size",
11
+ "3",
12
+ "--epoch",
13
+ "3",
14
+ "--seed",
15
+ "0"
16
+ ],
17
+ "program": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py",
18
+ "codePath": "train/train_deep_wandb.py",
19
+ "git": {
20
+ "remote": "git@hf.co:Yaning1001/Impossible_llm.git",
21
+ "commit": "ed716cdcfcdea02b67f7ed0f3504c2b1c8b737c4"
22
+ },
23
+ "email": "yaning1001@gmail.com",
24
+ "root": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train",
25
+ "host": "mms-large-2",
26
+ "username": "chunhui",
27
+ "executable": "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/bin/python",
28
+ "codePathLocal": "train_deep_wandb.py",
29
+ "cpu_count": 32,
30
+ "cpu_count_logical": 64,
31
+ "gpu": "NVIDIA RTX A6000",
32
+ "gpu_count": 8,
33
+ "disk": {
34
+ "/": {
35
+ "total": "1888559353856",
36
+ "used": "1774852702208"
37
+ }
38
+ },
39
+ "memory": {
40
+ "total": "202617098240"
41
+ },
42
+ "cpu": {
43
+ "count": 32,
44
+ "countLogical": 64
45
+ },
46
+ "gpu_nvidia": [
47
+ {
48
+ "name": "NVIDIA RTX A6000",
49
+ "memoryTotal": "51527024640",
50
+ "cudaCores": 10752,
51
+ "architecture": "Ampere"
52
+ },
53
+ {
54
+ "name": "NVIDIA RTX A6000",
55
+ "memoryTotal": "51527024640",
56
+ "cudaCores": 10752,
57
+ "architecture": "Ampere"
58
+ },
59
+ {
60
+ "name": "NVIDIA RTX A6000",
61
+ "memoryTotal": "51527024640",
62
+ "cudaCores": 10752,
63
+ "architecture": "Ampere"
64
+ },
65
+ {
66
+ "name": "NVIDIA RTX A6000",
67
+ "memoryTotal": "51527024640",
68
+ "cudaCores": 10752,
69
+ "architecture": "Ampere"
70
+ },
71
+ {
72
+ "name": "NVIDIA RTX A6000",
73
+ "memoryTotal": "51527024640",
74
+ "cudaCores": 10752,
75
+ "architecture": "Ampere"
76
+ },
77
+ {
78
+ "name": "NVIDIA RTX A6000",
79
+ "memoryTotal": "51527024640",
80
+ "cudaCores": 10752,
81
+ "architecture": "Ampere"
82
+ },
83
+ {
84
+ "name": "NVIDIA RTX A6000",
85
+ "memoryTotal": "51527024640",
86
+ "cudaCores": 10752,
87
+ "architecture": "Ampere"
88
+ },
89
+ {
90
+ "name": "NVIDIA RTX A6000",
91
+ "memoryTotal": "51527024640",
92
+ "cudaCores": 10752,
93
+ "architecture": "Ampere"
94
+ }
95
+ ],
96
+ "cudaVersion": "11.8"
97
+ }
wandb/run-20241106_224518-az8kcj8i/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_wandb":{"runtime":11}}
wandb/run-20241106_224518-az8kcj8i/logs/debug-internal.log ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2024-11-06T22:45:18.945770007-05:00","level":"INFO","msg":"using version","core version":"0.18.5"}
2
+ {"time":"2024-11-06T22:45:18.945781737-05:00","level":"INFO","msg":"created symlink","path":"/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241106_224518-az8kcj8i/logs/debug-core.log"}
3
+ {"time":"2024-11-06T22:45:19.053571616-05:00","level":"INFO","msg":"created new stream","id":"az8kcj8i"}
4
+ {"time":"2024-11-06T22:45:19.053631907-05:00","level":"INFO","msg":"stream: started","id":"az8kcj8i"}
5
+ {"time":"2024-11-06T22:45:19.053738127-05:00","level":"INFO","msg":"sender: started","stream_id":"az8kcj8i"}
6
+ {"time":"2024-11-06T22:45:19.053685487-05:00","level":"INFO","msg":"handler: started","stream_id":{"value":"az8kcj8i"}}
7
+ {"time":"2024-11-06T22:45:19.053654937-05:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"az8kcj8i"}}
8
+ {"time":"2024-11-06T22:45:19.214166646-05:00","level":"INFO","msg":"Starting system monitor"}
9
+ {"time":"2024-11-06T22:45:30.55948124-05:00","level":"INFO","msg":"stream: closing","id":"az8kcj8i"}
10
+ {"time":"2024-11-06T22:45:30.55952735-05:00","level":"INFO","msg":"Stopping system monitor"}
11
+ {"time":"2024-11-06T22:45:30.560064464-05:00","level":"INFO","msg":"Stopped system monitor"}
12
+ {"time":"2024-11-06T22:45:30.951268783-05:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
13
+ {"time":"2024-11-06T22:45:31.066261822-05:00","level":"INFO","msg":"handler: closed","stream_id":{"value":"az8kcj8i"}}
14
+ {"time":"2024-11-06T22:45:31.066309642-05:00","level":"INFO","msg":"sender: closed","stream_id":"az8kcj8i"}
15
+ {"time":"2024-11-06T22:45:31.066301022-05:00","level":"INFO","msg":"writer: Close: closed","stream_id":{"value":"az8kcj8i"}}
16
+ {"time":"2024-11-06T22:45:31.066383533-05:00","level":"INFO","msg":"stream: closed","id":"az8kcj8i"}