Yaning1001 commited on
Commit
dd71330
·
verified ·
1 Parent(s): 907dcdb

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +4 -0
  2. wandb/run-20241030_010759-v4mz2z5o/files/config.yaml +47 -0
  3. wandb/run-20241030_010759-v4mz2z5o/files/output.log +4 -0
  4. wandb/run-20241030_010759-v4mz2z5o/files/wandb-metadata.json +97 -0
  5. wandb/run-20241030_010759-v4mz2z5o/files/wandb-summary.json +1 -0
  6. wandb/run-20241030_010759-v4mz2z5o/logs/debug-internal.log +16 -0
  7. wandb/run-20241030_010759-v4mz2z5o/logs/debug.log +27 -0
  8. wandb/run-20241030_010759-v4mz2z5o/run-v4mz2z5o.wandb +0 -0
  9. wandb/run-20241030_231835-o1t74f3e/files/output.log +13 -0
  10. wandb/run-20241030_231835-o1t74f3e/files/requirements.txt +147 -0
  11. wandb/run-20241030_231835-o1t74f3e/files/wandb-metadata.json +97 -0
  12. wandb/run-20241030_231835-o1t74f3e/logs/debug-internal.log +8 -0
  13. wandb/run-20241030_231835-o1t74f3e/logs/debug.log +26 -0
  14. wandb/run-20241030_233740-98qje3cr/files/output.log +13 -0
  15. wandb/run-20241030_233740-98qje3cr/files/requirements.txt +147 -0
  16. wandb/run-20241030_233740-98qje3cr/files/wandb-metadata.json +97 -0
  17. wandb/run-20241030_233740-98qje3cr/logs/debug-internal.log +8 -0
  18. wandb/run-20241030_233740-98qje3cr/logs/debug.log +26 -0
  19. wandb/run-20241101_093116-n3c1zje4/files/output.log +13 -0
  20. wandb/run-20241101_093116-n3c1zje4/files/requirements.txt +147 -0
  21. wandb/run-20241101_093116-n3c1zje4/files/wandb-metadata.json +97 -0
  22. wandb/run-20241101_093116-n3c1zje4/logs/debug-internal.log +8 -0
  23. wandb/run-20241101_093116-n3c1zje4/logs/debug.log +26 -0
  24. wandb/run-20241101_093116-n3c1zje4/run-n3c1zje4.wandb +0 -0
  25. wandb/run-20241101_200535-kh416n4m/run-kh416n4m.wandb +3 -0
  26. wandb/run-20241101_201927-k1t1l8u4/run-k1t1l8u4.wandb +0 -0
  27. wandb/run-20241101_202058-jijqbvs1/run-jijqbvs1.wandb +3 -0
  28. wandb/run-20241105_160652-il3id7gi/files/config.yaml +49 -0
  29. wandb/run-20241105_160652-il3id7gi/files/output.log +8 -0
  30. wandb/run-20241105_160652-il3id7gi/files/requirements.txt +147 -0
  31. wandb/run-20241105_160652-il3id7gi/files/wandb-metadata.json +97 -0
  32. wandb/run-20241105_160652-il3id7gi/files/wandb-summary.json +1 -0
  33. wandb/run-20241105_160652-il3id7gi/logs/debug-internal.log +17 -0
  34. wandb/run-20241105_160652-il3id7gi/logs/debug.log +27 -0
  35. wandb/run-20241105_160652-il3id7gi/run-il3id7gi.wandb +0 -0
  36. wandb/run-20241105_163039-q4e8d8hm/files/config.yaml +49 -0
  37. wandb/run-20241105_163039-q4e8d8hm/files/wandb-metadata.json +97 -0
  38. wandb/run-20241105_163039-q4e8d8hm/logs/debug.log +27 -0
  39. wandb/run-20241105_163039-q4e8d8hm/run-q4e8d8hm.wandb +0 -0
  40. wandb/run-20241105_163244-o1vw2gev/logs/debug-internal.log +8 -0
  41. wandb/run-20241105_223736-2kagvv0a/files/config.yaml +50 -0
  42. wandb/run-20241105_223736-2kagvv0a/files/output.log +18 -0
  43. wandb/run-20241105_223736-2kagvv0a/files/wandb-metadata.json +97 -0
  44. wandb/run-20241105_223736-2kagvv0a/files/wandb-summary.json +1 -0
  45. wandb/run-20241105_223736-2kagvv0a/logs/debug-internal.log +18 -0
  46. wandb/run-20241105_223736-2kagvv0a/logs/debug.log +33 -0
  47. wandb/run-20241106_203647-jauklpbe/files/config.yaml +49 -0
  48. wandb/run-20241106_203647-jauklpbe/files/output.log +60 -0
  49. wandb/run-20241106_203647-jauklpbe/files/wandb-metadata.json +97 -0
  50. wandb/run-20241106_203647-jauklpbe/files/wandb-summary.json +1 -0
.gitattributes CHANGED
@@ -97,3 +97,7 @@ wandb/run-20241101_202058-hjyig8so/run-hjyig8so.wandb filter=lfs diff=lfs merge=
97
  wandb/run-20241101_012733-3tsgnm2p/run-3tsgnm2p.wandb filter=lfs diff=lfs merge=lfs -text
98
  wandb/run-20241030_112852-av3r7rx8/run-av3r7rx8.wandb filter=lfs diff=lfs merge=lfs -text
99
  wandb/run-20241115_125218-rrve0rbk/run-rrve0rbk.wandb filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
97
  wandb/run-20241101_012733-3tsgnm2p/run-3tsgnm2p.wandb filter=lfs diff=lfs merge=lfs -text
98
  wandb/run-20241030_112852-av3r7rx8/run-av3r7rx8.wandb filter=lfs diff=lfs merge=lfs -text
99
  wandb/run-20241115_125218-rrve0rbk/run-rrve0rbk.wandb filter=lfs diff=lfs merge=lfs -text
100
+ wandb/run-20241129_083813-1lx1o6j4/run-1lx1o6j4.wandb filter=lfs diff=lfs merge=lfs -text
101
+ wandb/run-20241101_200535-kh416n4m/run-kh416n4m.wandb filter=lfs diff=lfs merge=lfs -text
102
+ wandb/run-20241101_202058-jijqbvs1/run-jijqbvs1.wandb filter=lfs diff=lfs merge=lfs -text
103
+ wandb/run-20241129_083813-gsvlu1z8/run-gsvlu1z8.wandb filter=lfs diff=lfs merge=lfs -text
wandb/run-20241030_010759-v4mz2z5o/files/config.yaml ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.18.5
4
+ m: []
5
+ python_version: 3.9.19
6
+ t:
7
+ "1":
8
+ - 1
9
+ - 5
10
+ - 11
11
+ - 49
12
+ - 51
13
+ - 53
14
+ - 55
15
+ - 71
16
+ - 98
17
+ "2":
18
+ - 1
19
+ - 5
20
+ - 11
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 98
27
+ "3":
28
+ - 13
29
+ - 23
30
+ - 55
31
+ "4": 3.9.19
32
+ "5": 0.18.5
33
+ "6": 4.45.1
34
+ "8":
35
+ - 5
36
+ "12": 0.18.5
37
+ "13": linux-x86_64
38
+ batch_size:
39
+ value: 3
40
+ epoch:
41
+ value: 7
42
+ perturbation:
43
+ value: reverse_control
44
+ seed:
45
+ value: 0
46
+ train_set:
47
+ value: 10M
wandb/run-20241030_010759-v4mz2z5o/files/output.log ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ Traceback (most recent call last):
2
+ File "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py", line 162, in <module>
3
+ dataset_name = f"babylm_{args.perturbation}_{args.train_zset}_seed{args.seed}"
4
+ AttributeError: 'Namespace' object has no attribute 'train_zset'
wandb/run-20241030_010759-v4mz2z5o/files/wandb-metadata.json ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-162-generic-x86_64-with-glibc2.31",
3
+ "python": "3.9.19",
4
+ "startedAt": "2024-10-30T05:07:59.035692Z",
5
+ "args": [
6
+ "--perturbation",
7
+ "reverse_control",
8
+ "--train_set",
9
+ "10M",
10
+ "--batch_size",
11
+ "3",
12
+ "--epoch",
13
+ "7",
14
+ "--seed",
15
+ "0"
16
+ ],
17
+ "program": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py",
18
+ "codePath": "train/train_deep_wandb.py",
19
+ "git": {
20
+ "remote": "git@hf.co:Yaning1001/Impossible_llm.git",
21
+ "commit": "ed716cdcfcdea02b67f7ed0f3504c2b1c8b737c4"
22
+ },
23
+ "email": "yaning1001@gmail.com",
24
+ "root": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train",
25
+ "host": "mms-large-2",
26
+ "username": "chunhui",
27
+ "executable": "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/bin/python",
28
+ "codePathLocal": "train_deep_wandb.py",
29
+ "cpu_count": 32,
30
+ "cpu_count_logical": 64,
31
+ "gpu": "NVIDIA RTX A6000",
32
+ "gpu_count": 8,
33
+ "disk": {
34
+ "/": {
35
+ "total": "1888559353856",
36
+ "used": "1719200268288"
37
+ }
38
+ },
39
+ "memory": {
40
+ "total": "202617098240"
41
+ },
42
+ "cpu": {
43
+ "count": 32,
44
+ "countLogical": 64
45
+ },
46
+ "gpu_nvidia": [
47
+ {
48
+ "name": "NVIDIA RTX A6000",
49
+ "memoryTotal": "51527024640",
50
+ "cudaCores": 10752,
51
+ "architecture": "Ampere"
52
+ },
53
+ {
54
+ "name": "NVIDIA RTX A6000",
55
+ "memoryTotal": "51527024640",
56
+ "cudaCores": 10752,
57
+ "architecture": "Ampere"
58
+ },
59
+ {
60
+ "name": "NVIDIA RTX A6000",
61
+ "memoryTotal": "51527024640",
62
+ "cudaCores": 10752,
63
+ "architecture": "Ampere"
64
+ },
65
+ {
66
+ "name": "NVIDIA RTX A6000",
67
+ "memoryTotal": "51527024640",
68
+ "cudaCores": 10752,
69
+ "architecture": "Ampere"
70
+ },
71
+ {
72
+ "name": "NVIDIA RTX A6000",
73
+ "memoryTotal": "51527024640",
74
+ "cudaCores": 10752,
75
+ "architecture": "Ampere"
76
+ },
77
+ {
78
+ "name": "NVIDIA RTX A6000",
79
+ "memoryTotal": "51527024640",
80
+ "cudaCores": 10752,
81
+ "architecture": "Ampere"
82
+ },
83
+ {
84
+ "name": "NVIDIA RTX A6000",
85
+ "memoryTotal": "51527024640",
86
+ "cudaCores": 10752,
87
+ "architecture": "Ampere"
88
+ },
89
+ {
90
+ "name": "NVIDIA RTX A6000",
91
+ "memoryTotal": "51527024640",
92
+ "cudaCores": 10752,
93
+ "architecture": "Ampere"
94
+ }
95
+ ],
96
+ "cudaVersion": "11.8"
97
+ }
wandb/run-20241030_010759-v4mz2z5o/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_wandb":{"runtime":0}}
wandb/run-20241030_010759-v4mz2z5o/logs/debug-internal.log ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2024-10-30T01:07:59.037884269-04:00","level":"INFO","msg":"using version","core version":"0.18.5"}
2
+ {"time":"2024-10-30T01:07:59.037895309-04:00","level":"INFO","msg":"created symlink","path":"/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241030_010759-v4mz2z5o/logs/debug-core.log"}
3
+ {"time":"2024-10-30T01:07:59.144238393-04:00","level":"INFO","msg":"created new stream","id":"v4mz2z5o"}
4
+ {"time":"2024-10-30T01:07:59.144273343-04:00","level":"INFO","msg":"stream: started","id":"v4mz2z5o"}
5
+ {"time":"2024-10-30T01:07:59.144297273-04:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"v4mz2z5o"}}
6
+ {"time":"2024-10-30T01:07:59.144344244-04:00","level":"INFO","msg":"handler: started","stream_id":{"value":"v4mz2z5o"}}
7
+ {"time":"2024-10-30T01:07:59.144382634-04:00","level":"INFO","msg":"sender: started","stream_id":"v4mz2z5o"}
8
+ {"time":"2024-10-30T01:07:59.298984547-04:00","level":"INFO","msg":"Starting system monitor"}
9
+ {"time":"2024-10-30T01:07:59.387117067-04:00","level":"INFO","msg":"stream: closing","id":"v4mz2z5o"}
10
+ {"time":"2024-10-30T01:07:59.387151617-04:00","level":"INFO","msg":"Stopping system monitor"}
11
+ {"time":"2024-10-30T01:07:59.387463689-04:00","level":"INFO","msg":"Stopped system monitor"}
12
+ {"time":"2024-10-30T01:07:59.920617799-04:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
13
+ {"time":"2024-10-30T01:08:00.033586148-04:00","level":"INFO","msg":"handler: closed","stream_id":{"value":"v4mz2z5o"}}
14
+ {"time":"2024-10-30T01:08:00.033615528-04:00","level":"INFO","msg":"sender: closed","stream_id":"v4mz2z5o"}
15
+ {"time":"2024-10-30T01:08:00.033608968-04:00","level":"INFO","msg":"writer: Close: closed","stream_id":{"value":"v4mz2z5o"}}
16
+ {"time":"2024-10-30T01:08:00.033726869-04:00","level":"INFO","msg":"stream: closed","id":"v4mz2z5o"}
wandb/run-20241030_010759-v4mz2z5o/logs/debug.log ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-10-30 01:07:59,034 INFO MainThread:322461 [wandb_setup.py:_flush():79] Current SDK version is 0.18.5
2
+ 2024-10-30 01:07:59,034 INFO MainThread:322461 [wandb_setup.py:_flush():79] Configure stats pid to 322461
3
+ 2024-10-30 01:07:59,034 INFO MainThread:322461 [wandb_setup.py:_flush():79] Loading settings from /home/chunhui/.config/wandb/settings
4
+ 2024-10-30 01:07:59,034 INFO MainThread:322461 [wandb_setup.py:_flush():79] Loading settings from /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/settings
5
+ 2024-10-30 01:07:59,034 INFO MainThread:322461 [wandb_setup.py:_flush():79] Loading settings from environment variables: {}
6
+ 2024-10-30 01:07:59,034 INFO MainThread:322461 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': None, '_disable_service': None}
7
+ 2024-10-30 01:07:59,034 INFO MainThread:322461 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': 'train/train_deep_wandb.py', 'program_abspath': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py', 'program': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py'}
8
+ 2024-10-30 01:07:59,034 INFO MainThread:322461 [wandb_setup.py:_flush():79] Applying login settings: {}
9
+ 2024-10-30 01:07:59,034 INFO MainThread:322461 [wandb_init.py:_log_setup():534] Logging user logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241030_010759-v4mz2z5o/logs/debug.log
10
+ 2024-10-30 01:07:59,034 INFO MainThread:322461 [wandb_init.py:_log_setup():535] Logging internal logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241030_010759-v4mz2z5o/logs/debug-internal.log
11
+ 2024-10-30 01:07:59,034 INFO MainThread:322461 [wandb_init.py:init():621] calling init triggers
12
+ 2024-10-30 01:07:59,034 INFO MainThread:322461 [wandb_init.py:init():628] wandb.init called with sweep_config: {}
13
+ config: {}
14
+ 2024-10-30 01:07:59,034 INFO MainThread:322461 [wandb_init.py:init():671] starting backend
15
+ 2024-10-30 01:07:59,034 INFO MainThread:322461 [wandb_init.py:init():675] sending inform_init request
16
+ 2024-10-30 01:07:59,035 INFO MainThread:322461 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2024-10-30 01:07:59,035 INFO MainThread:322461 [wandb_init.py:init():688] backend started and connected
18
+ 2024-10-30 01:07:59,038 INFO MainThread:322461 [wandb_init.py:init():783] updated telemetry
19
+ 2024-10-30 01:07:59,069 INFO MainThread:322461 [wandb_init.py:init():816] communicating run to backend with 90.0 second timeout
20
+ 2024-10-30 01:07:59,295 INFO MainThread:322461 [wandb_init.py:init():867] starting run threads in backend
21
+ 2024-10-30 01:07:59,384 INFO MainThread:322461 [wandb_run.py:_console_start():2463] atexit reg
22
+ 2024-10-30 01:07:59,384 INFO MainThread:322461 [wandb_run.py:_redirect():2311] redirect: wrap_raw
23
+ 2024-10-30 01:07:59,384 INFO MainThread:322461 [wandb_run.py:_redirect():2376] Wrapping output streams.
24
+ 2024-10-30 01:07:59,384 INFO MainThread:322461 [wandb_run.py:_redirect():2401] Redirects installed.
25
+ 2024-10-30 01:07:59,385 INFO MainThread:322461 [wandb_init.py:init():911] run started, returning control to user process
26
+ 2024-10-30 01:07:59,386 INFO MainThread:322461 [wandb_run.py:_config_callback():1390] config_cb None None {'perturbation': 'reverse_control', 'train_set': '10M', 'batch_size': 3, 'epoch': 7, 'seed': 0}
27
+ 2024-10-30 01:07:59,387 WARNING MsgRouterThr:322461 [router.py:message_loop():77] message_loop has been closed
wandb/run-20241030_010759-v4mz2z5o/run-v4mz2z5o.wandb ADDED
Binary file (1.6 kB). View file
 
wandb/run-20241030_231835-o1t74f3e/files/output.log ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Loading checkpoint shards: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 2/2 [00:05<00:00, 2.78s/it]
2
+ tokenized_valid: Dataset({
3
+ features: ['input_ids', 'attention_mask'],
4
+ num_rows: 600
5
+ })
6
+ /mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/training_args.py:1545: FutureWarning: `evaluation_strategy` is deprecated and will be removed in version 4.46 of 🤗 Transformers. Use `eval_strategy` instead
7
+ warnings.warn(
8
+ [2024-10-30 23:18:42,945] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect)
9
+ [2024-10-30 23:18:52,440] [INFO] [comm.py:652:init_distributed] cdb=None
10
+ Installed CUDA version 11.8 does not match the version torch was compiled with 11.7 but since the APIs are compatible, accepting this combination
11
+ Using /home/chunhui/.cache/torch_extensions/py39_cu117 as PyTorch extensions root...
12
+ Loading extension module cpu_adam...
13
+ Time to load cpu_adam op: 5.8972320556640625 seconds
wandb/run-20241030_231835-o1t74f3e/files/requirements.txt ADDED
@@ -0,0 +1,147 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ funcsigs==1.0.2
2
+ sentry-sdk==2.17.0
3
+ multiprocess==0.70.16
4
+ numpy==1.26.2
5
+ pluralizer==1.2.0
6
+ debugpy==1.6.7
7
+ nvidia-cudnn-cu11==8.5.0.96
8
+ deepspeed==0.15.2
9
+ data==0.4
10
+ pandas==2.1.3
11
+ tomli==2.0.1
12
+ charset-normalizer==3.3.2
13
+ attrs==24.2.0
14
+ aiosignal==1.3.1
15
+ fsspec==2023.10.0
16
+ nvidia-cusparse-cu11==11.7.4.91
17
+ zipp==3.12.0
18
+ mypy-extensions==1.0.0
19
+ datasets==3.0.1
20
+ joblib==1.3.2
21
+ hjson==3.1.0
22
+ traitlets==5.7.1
23
+ stack-data==0.6.0
24
+ transformers==4.45.1
25
+ sympy==1.11.1
26
+ Pygments==2.15.0
27
+ docker-pycreds==0.4.0
28
+ dill==0.3.8
29
+ wheel==0.44.0
30
+ prompt-toolkit==3.0.30
31
+ parso==0.8.3
32
+ ipykernel==6.23.1
33
+ pyarrow==17.0.0
34
+ certifi==2023.11.17
35
+ nvidia-cufft-cu11==10.9.0.58
36
+ six==1.16.0
37
+ pydantic==2.9.2
38
+ click==8.1.7
39
+ nest-asyncio==1.5.6
40
+ gmpy2==2.1.0
41
+ matplotlib==3.8.2
42
+ scipy==1.11.4
43
+ typing_extensions==4.12.2
44
+ statsmodels==0.14.0
45
+ huggingface-hub==0.25.0
46
+ frozenlist==1.4.1
47
+ gpustat==1.1.1
48
+ nvidia-nvtx-cu11==11.7.91
49
+ safetensors==0.4.5
50
+ stanza==1.9.2
51
+ decorator==5.1.1
52
+ seaborn==0.13.0
53
+ sentencepiece==0.2.0
54
+ PyYAML==6.0.1
55
+ black==24.8.0
56
+ protobuf==4.25.1
57
+ pickleshare==0.7.5
58
+ peft==0.13.0
59
+ triton==2.0.0
60
+ nvidia-cuda-runtime-cu11==11.7.99
61
+ Jinja2==3.1.2
62
+ nvidia-cusolver-cu11==11.4.0.1
63
+ executing==1.2.0
64
+ jupyter_client==8.1.0
65
+ pluggy==1.3.0
66
+ cmake==3.30.3
67
+ pytz==2023.3.post1
68
+ aiohappyeyeballs==2.4.2
69
+ kiwisolver==1.4.5
70
+ py-cpuinfo==9.0.0
71
+ Pillow==10.1.0
72
+ ptyprocess==0.7.0
73
+ importlib_resources==6.4.5
74
+ GitPython==3.1.43
75
+ importlib-metadata==6.0.0
76
+ iniconfig==2.0.0
77
+ scikit-learn==1.3.2
78
+ exceptiongroup==1.1.0
79
+ networkx==2.8.6
80
+ accelerate==1.0.0
81
+ nltk==3.8.1
82
+ shutilwhich==1.1.0
83
+ fonttools==4.45.1
84
+ future==0.18.3
85
+ aiohttp==3.10.6
86
+ wcwidth==0.2.5
87
+ idna==3.6
88
+ filelock==3.12.2
89
+ pathspec==0.12.1
90
+ jupyter_core==5.1.0
91
+ lit==18.1.8
92
+ nvidia-curand-cu11==10.2.10.91
93
+ nvidia-cublas-cu11==11.10.3.66
94
+ nvidia-ml-py==12.560.30
95
+ msgpack==1.1.0
96
+ python-dateutil==2.8.2
97
+ blessed==1.20.0
98
+ packaging==23.0
99
+ gitdb==4.0.11
100
+ yarl==1.13.0
101
+ emoji==2.8.0
102
+ tzdata==2023.3
103
+ cycler==0.12.1
104
+ tornado==6.2
105
+ backcall==0.2.0
106
+ plotnine==0.12.4
107
+ ninja==1.11.1.1
108
+ latex==0.7.0
109
+ wandb==0.18.5
110
+ setproctitle==1.3.3
111
+ threadpoolctl==3.2.0
112
+ requests==2.32.3
113
+ pyparsing==3.1.1
114
+ smmap==5.0.1
115
+ pyzmq==23.0.0
116
+ async-timeout==4.0.3
117
+ annotated-types==0.7.0
118
+ matplotlib-inline==0.1.6
119
+ latexcodec==1.0.0
120
+ ipython==8.0.0
121
+ patsy==0.5.3
122
+ contourpy==1.2.0
123
+ multidict==6.1.0
124
+ mizani==0.9.3
125
+ urllib3==2.1.0
126
+ tokenizers==0.20.0
127
+ MarkupSafe==2.1.2
128
+ pip==24.2
129
+ pexpect==4.8.0
130
+ tqdm==4.66.5
131
+ jedi==0.18.2
132
+ pydantic_core==2.23.4
133
+ tempdir==0.7.1
134
+ mpmath==1.2.1
135
+ setuptools==72.1.0
136
+ pytest==7.4.3
137
+ pure-eval==0.2.2
138
+ psutil==5.9.1
139
+ comm==0.1.2
140
+ nvidia-cuda-cupti-cu11==11.7.101
141
+ nvidia-cuda-nvrtc-cu11==11.7.99
142
+ regex==2023.10.3
143
+ platformdirs==2.5.2
144
+ asttokens==2.2.1
145
+ torch==2.0.0
146
+ nvidia-nccl-cu11==2.14.3
147
+ xxhash==3.5.0
wandb/run-20241030_231835-o1t74f3e/files/wandb-metadata.json ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-162-generic-x86_64-with-glibc2.31",
3
+ "python": "3.9.19",
4
+ "startedAt": "2024-10-31T03:18:35.151538Z",
5
+ "args": [
6
+ "--perturbation",
7
+ "reverse_full",
8
+ "--train_set",
9
+ "10M",
10
+ "--batch_size",
11
+ "3",
12
+ "--epoch",
13
+ "3",
14
+ "--seed",
15
+ "0"
16
+ ],
17
+ "program": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py",
18
+ "codePath": "train/train_deep_wandb.py",
19
+ "git": {
20
+ "remote": "git@hf.co:Yaning1001/Impossible_llm.git",
21
+ "commit": "ed716cdcfcdea02b67f7ed0f3504c2b1c8b737c4"
22
+ },
23
+ "email": "yaning1001@gmail.com",
24
+ "root": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train",
25
+ "host": "mms-large-2",
26
+ "username": "chunhui",
27
+ "executable": "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/bin/python",
28
+ "codePathLocal": "train_deep_wandb.py",
29
+ "cpu_count": 32,
30
+ "cpu_count_logical": 64,
31
+ "gpu": "NVIDIA RTX A6000",
32
+ "gpu_count": 8,
33
+ "disk": {
34
+ "/": {
35
+ "total": "1888559353856",
36
+ "used": "1711064555520"
37
+ }
38
+ },
39
+ "memory": {
40
+ "total": "202617098240"
41
+ },
42
+ "cpu": {
43
+ "count": 32,
44
+ "countLogical": 64
45
+ },
46
+ "gpu_nvidia": [
47
+ {
48
+ "name": "NVIDIA RTX A6000",
49
+ "memoryTotal": "51527024640",
50
+ "cudaCores": 10752,
51
+ "architecture": "Ampere"
52
+ },
53
+ {
54
+ "name": "NVIDIA RTX A6000",
55
+ "memoryTotal": "51527024640",
56
+ "cudaCores": 10752,
57
+ "architecture": "Ampere"
58
+ },
59
+ {
60
+ "name": "NVIDIA RTX A6000",
61
+ "memoryTotal": "51527024640",
62
+ "cudaCores": 10752,
63
+ "architecture": "Ampere"
64
+ },
65
+ {
66
+ "name": "NVIDIA RTX A6000",
67
+ "memoryTotal": "51527024640",
68
+ "cudaCores": 10752,
69
+ "architecture": "Ampere"
70
+ },
71
+ {
72
+ "name": "NVIDIA RTX A6000",
73
+ "memoryTotal": "51527024640",
74
+ "cudaCores": 10752,
75
+ "architecture": "Ampere"
76
+ },
77
+ {
78
+ "name": "NVIDIA RTX A6000",
79
+ "memoryTotal": "51527024640",
80
+ "cudaCores": 10752,
81
+ "architecture": "Ampere"
82
+ },
83
+ {
84
+ "name": "NVIDIA RTX A6000",
85
+ "memoryTotal": "51527024640",
86
+ "cudaCores": 10752,
87
+ "architecture": "Ampere"
88
+ },
89
+ {
90
+ "name": "NVIDIA RTX A6000",
91
+ "memoryTotal": "51527024640",
92
+ "cudaCores": 10752,
93
+ "architecture": "Ampere"
94
+ }
95
+ ],
96
+ "cudaVersion": "11.8"
97
+ }
wandb/run-20241030_231835-o1t74f3e/logs/debug-internal.log ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {"time":"2024-10-30T23:18:35.153366138-04:00","level":"INFO","msg":"using version","core version":"0.18.5"}
2
+ {"time":"2024-10-30T23:18:35.153377718-04:00","level":"INFO","msg":"created symlink","path":"/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241030_231835-o1t74f3e/logs/debug-core.log"}
3
+ {"time":"2024-10-30T23:18:35.259200227-04:00","level":"INFO","msg":"created new stream","id":"o1t74f3e"}
4
+ {"time":"2024-10-30T23:18:35.259239637-04:00","level":"INFO","msg":"stream: started","id":"o1t74f3e"}
5
+ {"time":"2024-10-30T23:18:35.259262897-04:00","level":"INFO","msg":"sender: started","stream_id":"o1t74f3e"}
6
+ {"time":"2024-10-30T23:18:35.259258987-04:00","level":"INFO","msg":"handler: started","stream_id":{"value":"o1t74f3e"}}
7
+ {"time":"2024-10-30T23:18:35.259247477-04:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"o1t74f3e"}}
8
+ {"time":"2024-10-30T23:18:35.435477397-04:00","level":"INFO","msg":"Starting system monitor"}
wandb/run-20241030_231835-o1t74f3e/logs/debug.log ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-10-30 23:18:35,149 INFO MainThread:457831 [wandb_setup.py:_flush():79] Current SDK version is 0.18.5
2
+ 2024-10-30 23:18:35,149 INFO MainThread:457831 [wandb_setup.py:_flush():79] Configure stats pid to 457831
3
+ 2024-10-30 23:18:35,149 INFO MainThread:457831 [wandb_setup.py:_flush():79] Loading settings from /home/chunhui/.config/wandb/settings
4
+ 2024-10-30 23:18:35,149 INFO MainThread:457831 [wandb_setup.py:_flush():79] Loading settings from /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/settings
5
+ 2024-10-30 23:18:35,149 INFO MainThread:457831 [wandb_setup.py:_flush():79] Loading settings from environment variables: {}
6
+ 2024-10-30 23:18:35,149 INFO MainThread:457831 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': None, '_disable_service': None}
7
+ 2024-10-30 23:18:35,149 INFO MainThread:457831 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': 'train/train_deep_wandb.py', 'program_abspath': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py', 'program': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py'}
8
+ 2024-10-30 23:18:35,149 INFO MainThread:457831 [wandb_setup.py:_flush():79] Applying login settings: {}
9
+ 2024-10-30 23:18:35,149 INFO MainThread:457831 [wandb_init.py:_log_setup():534] Logging user logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241030_231835-o1t74f3e/logs/debug.log
10
+ 2024-10-30 23:18:35,150 INFO MainThread:457831 [wandb_init.py:_log_setup():535] Logging internal logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241030_231835-o1t74f3e/logs/debug-internal.log
11
+ 2024-10-30 23:18:35,150 INFO MainThread:457831 [wandb_init.py:init():621] calling init triggers
12
+ 2024-10-30 23:18:35,150 INFO MainThread:457831 [wandb_init.py:init():628] wandb.init called with sweep_config: {}
13
+ config: {}
14
+ 2024-10-30 23:18:35,150 INFO MainThread:457831 [wandb_init.py:init():671] starting backend
15
+ 2024-10-30 23:18:35,150 INFO MainThread:457831 [wandb_init.py:init():675] sending inform_init request
16
+ 2024-10-30 23:18:35,151 INFO MainThread:457831 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2024-10-30 23:18:35,151 INFO MainThread:457831 [wandb_init.py:init():688] backend started and connected
18
+ 2024-10-30 23:18:35,154 INFO MainThread:457831 [wandb_init.py:init():783] updated telemetry
19
+ 2024-10-30 23:18:35,179 INFO MainThread:457831 [wandb_init.py:init():816] communicating run to backend with 90.0 second timeout
20
+ 2024-10-30 23:18:35,432 INFO MainThread:457831 [wandb_init.py:init():867] starting run threads in backend
21
+ 2024-10-30 23:18:35,527 INFO MainThread:457831 [wandb_run.py:_console_start():2463] atexit reg
22
+ 2024-10-30 23:18:35,527 INFO MainThread:457831 [wandb_run.py:_redirect():2311] redirect: wrap_raw
23
+ 2024-10-30 23:18:35,527 INFO MainThread:457831 [wandb_run.py:_redirect():2376] Wrapping output streams.
24
+ 2024-10-30 23:18:35,527 INFO MainThread:457831 [wandb_run.py:_redirect():2401] Redirects installed.
25
+ 2024-10-30 23:18:35,528 INFO MainThread:457831 [wandb_init.py:init():911] run started, returning control to user process
26
+ 2024-10-30 23:18:35,529 INFO MainThread:457831 [wandb_run.py:_config_callback():1390] config_cb None None {'perturbation': 'reverse_full', 'train_set': '10M', 'batch_size': 3, 'epoch': 3, 'seed': 0}
wandb/run-20241030_233740-98qje3cr/files/output.log ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Loading checkpoint shards: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 2/2 [00:05<00:00, 2.98s/it]
2
+ tokenized_valid: Dataset({
3
+ features: ['input_ids', 'attention_mask'],
4
+ num_rows: 600
5
+ })
6
+ /mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/training_args.py:1545: FutureWarning: `evaluation_strategy` is deprecated and will be removed in version 4.46 of 🤗 Transformers. Use `eval_strategy` instead
7
+ warnings.warn(
8
+ [2024-10-30 23:37:50,812] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect)
9
+ [2024-10-30 23:37:58,524] [INFO] [comm.py:652:init_distributed] cdb=None
10
+ Installed CUDA version 11.8 does not match the version torch was compiled with 11.7 but since the APIs are compatible, accepting this combination
11
+ Using /home/chunhui/.cache/torch_extensions/py39_cu117 as PyTorch extensions root...
12
+ Loading extension module cpu_adam...
13
+ Time to load cpu_adam op: 4.6087868213653564 seconds
wandb/run-20241030_233740-98qje3cr/files/requirements.txt ADDED
@@ -0,0 +1,147 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ funcsigs==1.0.2
2
+ sentry-sdk==2.17.0
3
+ multiprocess==0.70.16
4
+ numpy==1.26.2
5
+ pluralizer==1.2.0
6
+ debugpy==1.6.7
7
+ nvidia-cudnn-cu11==8.5.0.96
8
+ deepspeed==0.15.2
9
+ data==0.4
10
+ pandas==2.1.3
11
+ tomli==2.0.1
12
+ charset-normalizer==3.3.2
13
+ attrs==24.2.0
14
+ aiosignal==1.3.1
15
+ fsspec==2023.10.0
16
+ nvidia-cusparse-cu11==11.7.4.91
17
+ zipp==3.12.0
18
+ mypy-extensions==1.0.0
19
+ datasets==3.0.1
20
+ joblib==1.3.2
21
+ hjson==3.1.0
22
+ traitlets==5.7.1
23
+ stack-data==0.6.0
24
+ transformers==4.45.1
25
+ sympy==1.11.1
26
+ Pygments==2.15.0
27
+ docker-pycreds==0.4.0
28
+ dill==0.3.8
29
+ wheel==0.44.0
30
+ prompt-toolkit==3.0.30
31
+ parso==0.8.3
32
+ ipykernel==6.23.1
33
+ pyarrow==17.0.0
34
+ certifi==2023.11.17
35
+ nvidia-cufft-cu11==10.9.0.58
36
+ six==1.16.0
37
+ pydantic==2.9.2
38
+ click==8.1.7
39
+ nest-asyncio==1.5.6
40
+ gmpy2==2.1.0
41
+ matplotlib==3.8.2
42
+ scipy==1.11.4
43
+ typing_extensions==4.12.2
44
+ statsmodels==0.14.0
45
+ huggingface-hub==0.25.0
46
+ frozenlist==1.4.1
47
+ gpustat==1.1.1
48
+ nvidia-nvtx-cu11==11.7.91
49
+ safetensors==0.4.5
50
+ stanza==1.9.2
51
+ decorator==5.1.1
52
+ seaborn==0.13.0
53
+ sentencepiece==0.2.0
54
+ PyYAML==6.0.1
55
+ black==24.8.0
56
+ protobuf==4.25.1
57
+ pickleshare==0.7.5
58
+ peft==0.13.0
59
+ triton==2.0.0
60
+ nvidia-cuda-runtime-cu11==11.7.99
61
+ Jinja2==3.1.2
62
+ nvidia-cusolver-cu11==11.4.0.1
63
+ executing==1.2.0
64
+ jupyter_client==8.1.0
65
+ pluggy==1.3.0
66
+ cmake==3.30.3
67
+ pytz==2023.3.post1
68
+ aiohappyeyeballs==2.4.2
69
+ kiwisolver==1.4.5
70
+ py-cpuinfo==9.0.0
71
+ Pillow==10.1.0
72
+ ptyprocess==0.7.0
73
+ importlib_resources==6.4.5
74
+ GitPython==3.1.43
75
+ importlib-metadata==6.0.0
76
+ iniconfig==2.0.0
77
+ scikit-learn==1.3.2
78
+ exceptiongroup==1.1.0
79
+ networkx==2.8.6
80
+ accelerate==1.0.0
81
+ nltk==3.8.1
82
+ shutilwhich==1.1.0
83
+ fonttools==4.45.1
84
+ future==0.18.3
85
+ aiohttp==3.10.6
86
+ wcwidth==0.2.5
87
+ idna==3.6
88
+ filelock==3.12.2
89
+ pathspec==0.12.1
90
+ jupyter_core==5.1.0
91
+ lit==18.1.8
92
+ nvidia-curand-cu11==10.2.10.91
93
+ nvidia-cublas-cu11==11.10.3.66
94
+ nvidia-ml-py==12.560.30
95
+ msgpack==1.1.0
96
+ python-dateutil==2.8.2
97
+ blessed==1.20.0
98
+ packaging==23.0
99
+ gitdb==4.0.11
100
+ yarl==1.13.0
101
+ emoji==2.8.0
102
+ tzdata==2023.3
103
+ cycler==0.12.1
104
+ tornado==6.2
105
+ backcall==0.2.0
106
+ plotnine==0.12.4
107
+ ninja==1.11.1.1
108
+ latex==0.7.0
109
+ wandb==0.18.5
110
+ setproctitle==1.3.3
111
+ threadpoolctl==3.2.0
112
+ requests==2.32.3
113
+ pyparsing==3.1.1
114
+ smmap==5.0.1
115
+ pyzmq==23.0.0
116
+ async-timeout==4.0.3
117
+ annotated-types==0.7.0
118
+ matplotlib-inline==0.1.6
119
+ latexcodec==1.0.0
120
+ ipython==8.0.0
121
+ patsy==0.5.3
122
+ contourpy==1.2.0
123
+ multidict==6.1.0
124
+ mizani==0.9.3
125
+ urllib3==2.1.0
126
+ tokenizers==0.20.0
127
+ MarkupSafe==2.1.2
128
+ pip==24.2
129
+ pexpect==4.8.0
130
+ tqdm==4.66.5
131
+ jedi==0.18.2
132
+ pydantic_core==2.23.4
133
+ tempdir==0.7.1
134
+ mpmath==1.2.1
135
+ setuptools==72.1.0
136
+ pytest==7.4.3
137
+ pure-eval==0.2.2
138
+ psutil==5.9.1
139
+ comm==0.1.2
140
+ nvidia-cuda-cupti-cu11==11.7.101
141
+ nvidia-cuda-nvrtc-cu11==11.7.99
142
+ regex==2023.10.3
143
+ platformdirs==2.5.2
144
+ asttokens==2.2.1
145
+ torch==2.0.0
146
+ nvidia-nccl-cu11==2.14.3
147
+ xxhash==3.5.0
wandb/run-20241030_233740-98qje3cr/files/wandb-metadata.json ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-162-generic-x86_64-with-glibc2.31",
3
+ "python": "3.9.19",
4
+ "startedAt": "2024-10-31T03:37:40.841144Z",
5
+ "args": [
6
+ "--perturbation",
7
+ "reverse_control",
8
+ "--train_set",
9
+ "10M",
10
+ "--batch_size",
11
+ "3",
12
+ "--epoch",
13
+ "3",
14
+ "--seed",
15
+ "0"
16
+ ],
17
+ "program": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py",
18
+ "codePath": "train/train_deep_wandb.py",
19
+ "git": {
20
+ "remote": "git@hf.co:Yaning1001/Impossible_llm.git",
21
+ "commit": "ed716cdcfcdea02b67f7ed0f3504c2b1c8b737c4"
22
+ },
23
+ "email": "yaning1001@gmail.com",
24
+ "root": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train",
25
+ "host": "mms-large-2",
26
+ "username": "chunhui",
27
+ "executable": "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/bin/python",
28
+ "codePathLocal": "train_deep_wandb.py",
29
+ "cpu_count": 32,
30
+ "cpu_count_logical": 64,
31
+ "gpu": "NVIDIA RTX A6000",
32
+ "gpu_count": 8,
33
+ "disk": {
34
+ "/": {
35
+ "total": "1888559353856",
36
+ "used": "1711065919488"
37
+ }
38
+ },
39
+ "memory": {
40
+ "total": "202617098240"
41
+ },
42
+ "cpu": {
43
+ "count": 32,
44
+ "countLogical": 64
45
+ },
46
+ "gpu_nvidia": [
47
+ {
48
+ "name": "NVIDIA RTX A6000",
49
+ "memoryTotal": "51527024640",
50
+ "cudaCores": 10752,
51
+ "architecture": "Ampere"
52
+ },
53
+ {
54
+ "name": "NVIDIA RTX A6000",
55
+ "memoryTotal": "51527024640",
56
+ "cudaCores": 10752,
57
+ "architecture": "Ampere"
58
+ },
59
+ {
60
+ "name": "NVIDIA RTX A6000",
61
+ "memoryTotal": "51527024640",
62
+ "cudaCores": 10752,
63
+ "architecture": "Ampere"
64
+ },
65
+ {
66
+ "name": "NVIDIA RTX A6000",
67
+ "memoryTotal": "51527024640",
68
+ "cudaCores": 10752,
69
+ "architecture": "Ampere"
70
+ },
71
+ {
72
+ "name": "NVIDIA RTX A6000",
73
+ "memoryTotal": "51527024640",
74
+ "cudaCores": 10752,
75
+ "architecture": "Ampere"
76
+ },
77
+ {
78
+ "name": "NVIDIA RTX A6000",
79
+ "memoryTotal": "51527024640",
80
+ "cudaCores": 10752,
81
+ "architecture": "Ampere"
82
+ },
83
+ {
84
+ "name": "NVIDIA RTX A6000",
85
+ "memoryTotal": "51527024640",
86
+ "cudaCores": 10752,
87
+ "architecture": "Ampere"
88
+ },
89
+ {
90
+ "name": "NVIDIA RTX A6000",
91
+ "memoryTotal": "51527024640",
92
+ "cudaCores": 10752,
93
+ "architecture": "Ampere"
94
+ }
95
+ ],
96
+ "cudaVersion": "11.8"
97
+ }
wandb/run-20241030_233740-98qje3cr/logs/debug-internal.log ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {"time":"2024-10-30T23:37:40.845317238-04:00","level":"INFO","msg":"using version","core version":"0.18.5"}
2
+ {"time":"2024-10-30T23:37:40.845328408-04:00","level":"INFO","msg":"created symlink","path":"/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241030_233740-98qje3cr/logs/debug-core.log"}
3
+ {"time":"2024-10-30T23:37:40.951923871-04:00","level":"INFO","msg":"created new stream","id":"98qje3cr"}
4
+ {"time":"2024-10-30T23:37:40.951953631-04:00","level":"INFO","msg":"stream: started","id":"98qje3cr"}
5
+ {"time":"2024-10-30T23:37:40.952025931-04:00","level":"INFO","msg":"sender: started","stream_id":"98qje3cr"}
6
+ {"time":"2024-10-30T23:37:40.951990871-04:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"98qje3cr"}}
7
+ {"time":"2024-10-30T23:37:40.952005801-04:00","level":"INFO","msg":"handler: started","stream_id":{"value":"98qje3cr"}}
8
+ {"time":"2024-10-30T23:37:41.405269516-04:00","level":"INFO","msg":"Starting system monitor"}
wandb/run-20241030_233740-98qje3cr/logs/debug.log ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-10-30 23:37:40,839 INFO MainThread:464534 [wandb_setup.py:_flush():79] Current SDK version is 0.18.5
2
+ 2024-10-30 23:37:40,839 INFO MainThread:464534 [wandb_setup.py:_flush():79] Configure stats pid to 464534
3
+ 2024-10-30 23:37:40,839 INFO MainThread:464534 [wandb_setup.py:_flush():79] Loading settings from /home/chunhui/.config/wandb/settings
4
+ 2024-10-30 23:37:40,839 INFO MainThread:464534 [wandb_setup.py:_flush():79] Loading settings from /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/settings
5
+ 2024-10-30 23:37:40,839 INFO MainThread:464534 [wandb_setup.py:_flush():79] Loading settings from environment variables: {}
6
+ 2024-10-30 23:37:40,839 INFO MainThread:464534 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': None, '_disable_service': None}
7
+ 2024-10-30 23:37:40,839 INFO MainThread:464534 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': 'train/train_deep_wandb.py', 'program_abspath': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py', 'program': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py'}
8
+ 2024-10-30 23:37:40,839 INFO MainThread:464534 [wandb_setup.py:_flush():79] Applying login settings: {}
9
+ 2024-10-30 23:37:40,839 INFO MainThread:464534 [wandb_init.py:_log_setup():534] Logging user logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241030_233740-98qje3cr/logs/debug.log
10
+ 2024-10-30 23:37:40,839 INFO MainThread:464534 [wandb_init.py:_log_setup():535] Logging internal logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241030_233740-98qje3cr/logs/debug-internal.log
11
+ 2024-10-30 23:37:40,839 INFO MainThread:464534 [wandb_init.py:init():621] calling init triggers
12
+ 2024-10-30 23:37:40,839 INFO MainThread:464534 [wandb_init.py:init():628] wandb.init called with sweep_config: {}
13
+ config: {}
14
+ 2024-10-30 23:37:40,839 INFO MainThread:464534 [wandb_init.py:init():671] starting backend
15
+ 2024-10-30 23:37:40,839 INFO MainThread:464534 [wandb_init.py:init():675] sending inform_init request
16
+ 2024-10-30 23:37:40,840 INFO MainThread:464534 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2024-10-30 23:37:40,840 INFO MainThread:464534 [wandb_init.py:init():688] backend started and connected
18
+ 2024-10-30 23:37:40,844 INFO MainThread:464534 [wandb_init.py:init():783] updated telemetry
19
+ 2024-10-30 23:37:40,871 INFO MainThread:464534 [wandb_init.py:init():816] communicating run to backend with 90.0 second timeout
20
+ 2024-10-30 23:37:41,402 INFO MainThread:464534 [wandb_init.py:init():867] starting run threads in backend
21
+ 2024-10-30 23:37:42,837 INFO MainThread:464534 [wandb_run.py:_console_start():2463] atexit reg
22
+ 2024-10-30 23:37:42,837 INFO MainThread:464534 [wandb_run.py:_redirect():2311] redirect: wrap_raw
23
+ 2024-10-30 23:37:42,837 INFO MainThread:464534 [wandb_run.py:_redirect():2376] Wrapping output streams.
24
+ 2024-10-30 23:37:42,837 INFO MainThread:464534 [wandb_run.py:_redirect():2401] Redirects installed.
25
+ 2024-10-30 23:37:42,860 INFO MainThread:464534 [wandb_init.py:init():911] run started, returning control to user process
26
+ 2024-10-30 23:37:42,861 INFO MainThread:464534 [wandb_run.py:_config_callback():1390] config_cb None None {'perturbation': 'reverse_control', 'train_set': '10M', 'batch_size': 3, 'epoch': 3, 'seed': 0}
wandb/run-20241101_093116-n3c1zje4/files/output.log ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Loading checkpoint shards: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 2/2 [00:18<00:00, 9.08s/it]
2
+ tokenized_valid: Dataset({
3
+ features: ['input_ids', 'attention_mask'],
4
+ num_rows: 600
5
+ })
6
+ /mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/training_args.py:1545: FutureWarning: `evaluation_strategy` is deprecated and will be removed in version 4.46 of 🤗 Transformers. Use `eval_strategy` instead
7
+ warnings.warn(
8
+ [2024-11-01 09:31:37,042] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect)
9
+ [2024-11-01 09:31:45,660] [INFO] [comm.py:652:init_distributed] cdb=None
10
+ Installed CUDA version 11.8 does not match the version torch was compiled with 11.7 but since the APIs are compatible, accepting this combination
11
+ Using /home/chunhui/.cache/torch_extensions/py39_cu117 as PyTorch extensions root...
12
+ Loading extension module cpu_adam...
13
+ Time to load cpu_adam op: 4.823585748672485 seconds
wandb/run-20241101_093116-n3c1zje4/files/requirements.txt ADDED
@@ -0,0 +1,147 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ funcsigs==1.0.2
2
+ sentry-sdk==2.17.0
3
+ multiprocess==0.70.16
4
+ numpy==1.26.2
5
+ pluralizer==1.2.0
6
+ debugpy==1.6.7
7
+ nvidia-cudnn-cu11==8.5.0.96
8
+ deepspeed==0.15.2
9
+ data==0.4
10
+ pandas==2.1.3
11
+ tomli==2.0.1
12
+ charset-normalizer==3.3.2
13
+ attrs==24.2.0
14
+ aiosignal==1.3.1
15
+ fsspec==2023.10.0
16
+ nvidia-cusparse-cu11==11.7.4.91
17
+ zipp==3.12.0
18
+ mypy-extensions==1.0.0
19
+ datasets==3.0.1
20
+ joblib==1.3.2
21
+ hjson==3.1.0
22
+ traitlets==5.7.1
23
+ stack-data==0.6.0
24
+ transformers==4.45.1
25
+ sympy==1.11.1
26
+ Pygments==2.15.0
27
+ docker-pycreds==0.4.0
28
+ dill==0.3.8
29
+ wheel==0.44.0
30
+ prompt-toolkit==3.0.30
31
+ parso==0.8.3
32
+ ipykernel==6.23.1
33
+ pyarrow==17.0.0
34
+ certifi==2023.11.17
35
+ nvidia-cufft-cu11==10.9.0.58
36
+ six==1.16.0
37
+ pydantic==2.9.2
38
+ click==8.1.7
39
+ nest-asyncio==1.5.6
40
+ gmpy2==2.1.0
41
+ matplotlib==3.8.2
42
+ scipy==1.11.4
43
+ typing_extensions==4.12.2
44
+ statsmodels==0.14.0
45
+ huggingface-hub==0.25.0
46
+ frozenlist==1.4.1
47
+ gpustat==1.1.1
48
+ nvidia-nvtx-cu11==11.7.91
49
+ safetensors==0.4.5
50
+ stanza==1.9.2
51
+ decorator==5.1.1
52
+ seaborn==0.13.0
53
+ sentencepiece==0.2.0
54
+ PyYAML==6.0.1
55
+ black==24.8.0
56
+ protobuf==4.25.1
57
+ pickleshare==0.7.5
58
+ peft==0.13.0
59
+ triton==2.0.0
60
+ nvidia-cuda-runtime-cu11==11.7.99
61
+ Jinja2==3.1.2
62
+ nvidia-cusolver-cu11==11.4.0.1
63
+ executing==1.2.0
64
+ jupyter_client==8.1.0
65
+ pluggy==1.3.0
66
+ cmake==3.30.3
67
+ pytz==2023.3.post1
68
+ aiohappyeyeballs==2.4.2
69
+ kiwisolver==1.4.5
70
+ py-cpuinfo==9.0.0
71
+ Pillow==10.1.0
72
+ ptyprocess==0.7.0
73
+ importlib_resources==6.4.5
74
+ GitPython==3.1.43
75
+ importlib-metadata==6.0.0
76
+ iniconfig==2.0.0
77
+ scikit-learn==1.3.2
78
+ exceptiongroup==1.1.0
79
+ networkx==2.8.6
80
+ accelerate==1.0.0
81
+ nltk==3.8.1
82
+ shutilwhich==1.1.0
83
+ fonttools==4.45.1
84
+ future==0.18.3
85
+ aiohttp==3.10.6
86
+ wcwidth==0.2.5
87
+ idna==3.6
88
+ filelock==3.12.2
89
+ pathspec==0.12.1
90
+ jupyter_core==5.1.0
91
+ lit==18.1.8
92
+ nvidia-curand-cu11==10.2.10.91
93
+ nvidia-cublas-cu11==11.10.3.66
94
+ nvidia-ml-py==12.560.30
95
+ msgpack==1.1.0
96
+ python-dateutil==2.8.2
97
+ blessed==1.20.0
98
+ packaging==23.0
99
+ gitdb==4.0.11
100
+ yarl==1.13.0
101
+ emoji==2.8.0
102
+ tzdata==2023.3
103
+ cycler==0.12.1
104
+ tornado==6.2
105
+ backcall==0.2.0
106
+ plotnine==0.12.4
107
+ ninja==1.11.1.1
108
+ latex==0.7.0
109
+ wandb==0.18.5
110
+ setproctitle==1.3.3
111
+ threadpoolctl==3.2.0
112
+ requests==2.32.3
113
+ pyparsing==3.1.1
114
+ smmap==5.0.1
115
+ pyzmq==23.0.0
116
+ async-timeout==4.0.3
117
+ annotated-types==0.7.0
118
+ matplotlib-inline==0.1.6
119
+ latexcodec==1.0.0
120
+ ipython==8.0.0
121
+ patsy==0.5.3
122
+ contourpy==1.2.0
123
+ multidict==6.1.0
124
+ mizani==0.9.3
125
+ urllib3==2.1.0
126
+ tokenizers==0.20.0
127
+ MarkupSafe==2.1.2
128
+ pip==24.2
129
+ pexpect==4.8.0
130
+ tqdm==4.66.5
131
+ jedi==0.18.2
132
+ pydantic_core==2.23.4
133
+ tempdir==0.7.1
134
+ mpmath==1.2.1
135
+ setuptools==72.1.0
136
+ pytest==7.4.3
137
+ pure-eval==0.2.2
138
+ psutil==5.9.1
139
+ comm==0.1.2
140
+ nvidia-cuda-cupti-cu11==11.7.101
141
+ nvidia-cuda-nvrtc-cu11==11.7.99
142
+ regex==2023.10.3
143
+ platformdirs==2.5.2
144
+ asttokens==2.2.1
145
+ torch==2.0.0
146
+ nvidia-nccl-cu11==2.14.3
147
+ xxhash==3.5.0
wandb/run-20241101_093116-n3c1zje4/files/wandb-metadata.json ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-162-generic-x86_64-with-glibc2.31",
3
+ "python": "3.9.19",
4
+ "startedAt": "2024-11-01T13:31:16.510021Z",
5
+ "args": [
6
+ "--perturbation",
7
+ "reverse_control",
8
+ "--train_set",
9
+ "10M",
10
+ "--batch_size",
11
+ "3",
12
+ "--epoch",
13
+ "7",
14
+ "--seed",
15
+ "0"
16
+ ],
17
+ "program": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py",
18
+ "codePath": "train/train_deep_wandb.py",
19
+ "git": {
20
+ "remote": "git@hf.co:Yaning1001/Impossible_llm.git",
21
+ "commit": "ed716cdcfcdea02b67f7ed0f3504c2b1c8b737c4"
22
+ },
23
+ "email": "yaning1001@gmail.com",
24
+ "root": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train",
25
+ "host": "mms-large-2",
26
+ "username": "chunhui",
27
+ "executable": "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/bin/python",
28
+ "codePathLocal": "train_deep_wandb.py",
29
+ "cpu_count": 32,
30
+ "cpu_count_logical": 64,
31
+ "gpu": "NVIDIA RTX A6000",
32
+ "gpu_count": 8,
33
+ "disk": {
34
+ "/": {
35
+ "total": "1888559353856",
36
+ "used": "1754716262400"
37
+ }
38
+ },
39
+ "memory": {
40
+ "total": "202617098240"
41
+ },
42
+ "cpu": {
43
+ "count": 32,
44
+ "countLogical": 64
45
+ },
46
+ "gpu_nvidia": [
47
+ {
48
+ "name": "NVIDIA RTX A6000",
49
+ "memoryTotal": "51527024640",
50
+ "cudaCores": 10752,
51
+ "architecture": "Ampere"
52
+ },
53
+ {
54
+ "name": "NVIDIA RTX A6000",
55
+ "memoryTotal": "51527024640",
56
+ "cudaCores": 10752,
57
+ "architecture": "Ampere"
58
+ },
59
+ {
60
+ "name": "NVIDIA RTX A6000",
61
+ "memoryTotal": "51527024640",
62
+ "cudaCores": 10752,
63
+ "architecture": "Ampere"
64
+ },
65
+ {
66
+ "name": "NVIDIA RTX A6000",
67
+ "memoryTotal": "51527024640",
68
+ "cudaCores": 10752,
69
+ "architecture": "Ampere"
70
+ },
71
+ {
72
+ "name": "NVIDIA RTX A6000",
73
+ "memoryTotal": "51527024640",
74
+ "cudaCores": 10752,
75
+ "architecture": "Ampere"
76
+ },
77
+ {
78
+ "name": "NVIDIA RTX A6000",
79
+ "memoryTotal": "51527024640",
80
+ "cudaCores": 10752,
81
+ "architecture": "Ampere"
82
+ },
83
+ {
84
+ "name": "NVIDIA RTX A6000",
85
+ "memoryTotal": "51527024640",
86
+ "cudaCores": 10752,
87
+ "architecture": "Ampere"
88
+ },
89
+ {
90
+ "name": "NVIDIA RTX A6000",
91
+ "memoryTotal": "51527024640",
92
+ "cudaCores": 10752,
93
+ "architecture": "Ampere"
94
+ }
95
+ ],
96
+ "cudaVersion": "11.8"
97
+ }
wandb/run-20241101_093116-n3c1zje4/logs/debug-internal.log ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {"time":"2024-11-01T09:31:16.512866181-04:00","level":"INFO","msg":"using version","core version":"0.18.5"}
2
+ {"time":"2024-11-01T09:31:16.512881061-04:00","level":"INFO","msg":"created symlink","path":"/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241101_093116-n3c1zje4/logs/debug-core.log"}
3
+ {"time":"2024-11-01T09:31:16.623420531-04:00","level":"INFO","msg":"created new stream","id":"n3c1zje4"}
4
+ {"time":"2024-11-01T09:31:16.623477982-04:00","level":"INFO","msg":"stream: started","id":"n3c1zje4"}
5
+ {"time":"2024-11-01T09:31:16.623527602-04:00","level":"INFO","msg":"handler: started","stream_id":{"value":"n3c1zje4"}}
6
+ {"time":"2024-11-01T09:31:16.623492722-04:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"n3c1zje4"}}
7
+ {"time":"2024-11-01T09:31:16.623569862-04:00","level":"INFO","msg":"sender: started","stream_id":"n3c1zje4"}
8
+ {"time":"2024-11-01T09:31:16.840771598-04:00","level":"INFO","msg":"Starting system monitor"}
wandb/run-20241101_093116-n3c1zje4/logs/debug.log ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-11-01 09:31:16,508 INFO MainThread:781950 [wandb_setup.py:_flush():79] Current SDK version is 0.18.5
2
+ 2024-11-01 09:31:16,508 INFO MainThread:781950 [wandb_setup.py:_flush():79] Configure stats pid to 781950
3
+ 2024-11-01 09:31:16,508 INFO MainThread:781950 [wandb_setup.py:_flush():79] Loading settings from /home/chunhui/.config/wandb/settings
4
+ 2024-11-01 09:31:16,508 INFO MainThread:781950 [wandb_setup.py:_flush():79] Loading settings from /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/settings
5
+ 2024-11-01 09:31:16,508 INFO MainThread:781950 [wandb_setup.py:_flush():79] Loading settings from environment variables: {}
6
+ 2024-11-01 09:31:16,508 INFO MainThread:781950 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': None, '_disable_service': None}
7
+ 2024-11-01 09:31:16,508 INFO MainThread:781950 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': 'train/train_deep_wandb.py', 'program_abspath': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py', 'program': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py'}
8
+ 2024-11-01 09:31:16,508 INFO MainThread:781950 [wandb_setup.py:_flush():79] Applying login settings: {}
9
+ 2024-11-01 09:31:16,508 INFO MainThread:781950 [wandb_init.py:_log_setup():534] Logging user logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241101_093116-n3c1zje4/logs/debug.log
10
+ 2024-11-01 09:31:16,508 INFO MainThread:781950 [wandb_init.py:_log_setup():535] Logging internal logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241101_093116-n3c1zje4/logs/debug-internal.log
11
+ 2024-11-01 09:31:16,508 INFO MainThread:781950 [wandb_init.py:init():621] calling init triggers
12
+ 2024-11-01 09:31:16,508 INFO MainThread:781950 [wandb_init.py:init():628] wandb.init called with sweep_config: {}
13
+ config: {}
14
+ 2024-11-01 09:31:16,508 INFO MainThread:781950 [wandb_init.py:init():671] starting backend
15
+ 2024-11-01 09:31:16,508 INFO MainThread:781950 [wandb_init.py:init():675] sending inform_init request
16
+ 2024-11-01 09:31:16,509 INFO MainThread:781950 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2024-11-01 09:31:16,509 INFO MainThread:781950 [wandb_init.py:init():688] backend started and connected
18
+ 2024-11-01 09:31:16,513 INFO MainThread:781950 [wandb_init.py:init():783] updated telemetry
19
+ 2024-11-01 09:31:16,546 INFO MainThread:781950 [wandb_init.py:init():816] communicating run to backend with 90.0 second timeout
20
+ 2024-11-01 09:31:16,836 INFO MainThread:781950 [wandb_init.py:init():867] starting run threads in backend
21
+ 2024-11-01 09:31:16,962 INFO MainThread:781950 [wandb_run.py:_console_start():2463] atexit reg
22
+ 2024-11-01 09:31:16,962 INFO MainThread:781950 [wandb_run.py:_redirect():2311] redirect: wrap_raw
23
+ 2024-11-01 09:31:16,963 INFO MainThread:781950 [wandb_run.py:_redirect():2376] Wrapping output streams.
24
+ 2024-11-01 09:31:16,963 INFO MainThread:781950 [wandb_run.py:_redirect():2401] Redirects installed.
25
+ 2024-11-01 09:31:16,964 INFO MainThread:781950 [wandb_init.py:init():911] run started, returning control to user process
26
+ 2024-11-01 09:31:16,964 INFO MainThread:781950 [wandb_run.py:_config_callback():1390] config_cb None None {'perturbation': 'reverse_control', 'train_set': '10M', 'batch_size': 3, 'epoch': 7, 'seed': 0, 'lr': 5e-06}
wandb/run-20241101_093116-n3c1zje4/run-n3c1zje4.wandb ADDED
Binary file (32.8 kB). View file
 
wandb/run-20241101_200535-kh416n4m/run-kh416n4m.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:262a348816d9419ec3403966b5e76fe2cb0a701d902903bc04dc131eabf5c8f4
3
+ size 131072
wandb/run-20241101_201927-k1t1l8u4/run-k1t1l8u4.wandb ADDED
Binary file (32.8 kB). View file
 
wandb/run-20241101_202058-jijqbvs1/run-jijqbvs1.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6daebef05a7995ff2c2936db987cd658c3b9b8ee650c82f84f0b76c724dcff4
3
+ size 13287862
wandb/run-20241105_160652-il3id7gi/files/config.yaml ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.18.5
4
+ m: []
5
+ python_version: 3.9.19
6
+ t:
7
+ "1":
8
+ - 1
9
+ - 5
10
+ - 11
11
+ - 49
12
+ - 51
13
+ - 53
14
+ - 55
15
+ - 71
16
+ - 98
17
+ "2":
18
+ - 1
19
+ - 5
20
+ - 11
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 98
27
+ "3":
28
+ - 13
29
+ - 23
30
+ - 55
31
+ "4": 3.9.19
32
+ "5": 0.18.5
33
+ "6": 4.45.1
34
+ "8":
35
+ - 5
36
+ "12": 0.18.5
37
+ "13": linux-x86_64
38
+ batch_size:
39
+ value: 3
40
+ epoch:
41
+ value: 3
42
+ lr:
43
+ value: 5e-06
44
+ perturbation:
45
+ value: shuffle_deterministic21
46
+ seed:
47
+ value: 0
48
+ train_set:
49
+ value: 10M
wandb/run-20241105_160652-il3id7gi/files/output.log ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ Traceback (most recent call last):
2
+ File "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py", line 165, in <module>
3
+ dataset = load_dataset('babylm_dataset_test.py', name=dataset_name, trust_remote_code=True)
4
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/datasets/load.py", line 2096, in load_dataset
5
+ builder_instance.download_and_prepare(
6
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/datasets/builder.py", line 875, in download_and_prepare
7
+ raise OSError(
8
+ OSError: Not enough disk space. Needed: Unknown size (download: Unknown size, generated: Unknown size, post-processed: Unknown size)
wandb/run-20241105_160652-il3id7gi/files/requirements.txt ADDED
@@ -0,0 +1,147 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ funcsigs==1.0.2
2
+ sentry-sdk==2.17.0
3
+ multiprocess==0.70.16
4
+ numpy==1.26.2
5
+ pluralizer==1.2.0
6
+ debugpy==1.6.7
7
+ nvidia-cudnn-cu11==8.5.0.96
8
+ deepspeed==0.15.2
9
+ data==0.4
10
+ pandas==2.1.3
11
+ tomli==2.0.1
12
+ charset-normalizer==3.3.2
13
+ attrs==24.2.0
14
+ aiosignal==1.3.1
15
+ fsspec==2023.10.0
16
+ nvidia-cusparse-cu11==11.7.4.91
17
+ zipp==3.12.0
18
+ mypy-extensions==1.0.0
19
+ datasets==3.0.1
20
+ joblib==1.3.2
21
+ hjson==3.1.0
22
+ traitlets==5.7.1
23
+ stack-data==0.6.0
24
+ transformers==4.45.1
25
+ sympy==1.11.1
26
+ Pygments==2.15.0
27
+ docker-pycreds==0.4.0
28
+ dill==0.3.8
29
+ wheel==0.44.0
30
+ prompt-toolkit==3.0.30
31
+ parso==0.8.3
32
+ ipykernel==6.23.1
33
+ pyarrow==17.0.0
34
+ certifi==2023.11.17
35
+ nvidia-cufft-cu11==10.9.0.58
36
+ six==1.16.0
37
+ pydantic==2.9.2
38
+ click==8.1.7
39
+ nest-asyncio==1.5.6
40
+ gmpy2==2.1.0
41
+ matplotlib==3.8.2
42
+ scipy==1.11.4
43
+ typing_extensions==4.12.2
44
+ statsmodels==0.14.0
45
+ huggingface-hub==0.25.0
46
+ frozenlist==1.4.1
47
+ gpustat==1.1.1
48
+ nvidia-nvtx-cu11==11.7.91
49
+ safetensors==0.4.5
50
+ stanza==1.9.2
51
+ decorator==5.1.1
52
+ seaborn==0.13.0
53
+ sentencepiece==0.2.0
54
+ PyYAML==6.0.1
55
+ black==24.8.0
56
+ protobuf==4.25.1
57
+ pickleshare==0.7.5
58
+ peft==0.13.0
59
+ triton==2.0.0
60
+ nvidia-cuda-runtime-cu11==11.7.99
61
+ Jinja2==3.1.2
62
+ nvidia-cusolver-cu11==11.4.0.1
63
+ executing==1.2.0
64
+ jupyter_client==8.1.0
65
+ pluggy==1.3.0
66
+ cmake==3.30.3
67
+ pytz==2023.3.post1
68
+ aiohappyeyeballs==2.4.2
69
+ kiwisolver==1.4.5
70
+ py-cpuinfo==9.0.0
71
+ Pillow==10.1.0
72
+ ptyprocess==0.7.0
73
+ importlib_resources==6.4.5
74
+ GitPython==3.1.43
75
+ importlib-metadata==6.0.0
76
+ iniconfig==2.0.0
77
+ scikit-learn==1.3.2
78
+ exceptiongroup==1.1.0
79
+ networkx==2.8.6
80
+ accelerate==1.0.0
81
+ nltk==3.8.1
82
+ shutilwhich==1.1.0
83
+ fonttools==4.45.1
84
+ future==0.18.3
85
+ aiohttp==3.10.6
86
+ wcwidth==0.2.5
87
+ idna==3.6
88
+ filelock==3.12.2
89
+ pathspec==0.12.1
90
+ jupyter_core==5.1.0
91
+ lit==18.1.8
92
+ nvidia-curand-cu11==10.2.10.91
93
+ nvidia-cublas-cu11==11.10.3.66
94
+ nvidia-ml-py==12.560.30
95
+ msgpack==1.1.0
96
+ python-dateutil==2.8.2
97
+ blessed==1.20.0
98
+ packaging==23.0
99
+ gitdb==4.0.11
100
+ yarl==1.13.0
101
+ emoji==2.8.0
102
+ tzdata==2023.3
103
+ cycler==0.12.1
104
+ tornado==6.2
105
+ backcall==0.2.0
106
+ plotnine==0.12.4
107
+ ninja==1.11.1.1
108
+ latex==0.7.0
109
+ wandb==0.18.5
110
+ setproctitle==1.3.3
111
+ threadpoolctl==3.2.0
112
+ requests==2.32.3
113
+ pyparsing==3.1.1
114
+ smmap==5.0.1
115
+ pyzmq==23.0.0
116
+ async-timeout==4.0.3
117
+ annotated-types==0.7.0
118
+ matplotlib-inline==0.1.6
119
+ latexcodec==1.0.0
120
+ ipython==8.0.0
121
+ patsy==0.5.3
122
+ contourpy==1.2.0
123
+ multidict==6.1.0
124
+ mizani==0.9.3
125
+ urllib3==2.1.0
126
+ tokenizers==0.20.0
127
+ MarkupSafe==2.1.2
128
+ pip==24.2
129
+ pexpect==4.8.0
130
+ tqdm==4.66.5
131
+ jedi==0.18.2
132
+ pydantic_core==2.23.4
133
+ tempdir==0.7.1
134
+ mpmath==1.2.1
135
+ setuptools==72.1.0
136
+ pytest==7.4.3
137
+ pure-eval==0.2.2
138
+ psutil==5.9.1
139
+ comm==0.1.2
140
+ nvidia-cuda-cupti-cu11==11.7.101
141
+ nvidia-cuda-nvrtc-cu11==11.7.99
142
+ regex==2023.10.3
143
+ platformdirs==2.5.2
144
+ asttokens==2.2.1
145
+ torch==2.0.0
146
+ nvidia-nccl-cu11==2.14.3
147
+ xxhash==3.5.0
wandb/run-20241105_160652-il3id7gi/files/wandb-metadata.json ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-162-generic-x86_64-with-glibc2.31",
3
+ "python": "3.9.19",
4
+ "startedAt": "2024-11-05T21:06:52.164993Z",
5
+ "args": [
6
+ "--perturbation",
7
+ "shuffle_deterministic21",
8
+ "--train_set",
9
+ "10M",
10
+ "--batch_size",
11
+ "3",
12
+ "--epoch",
13
+ "3",
14
+ "--seed",
15
+ "0"
16
+ ],
17
+ "program": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py",
18
+ "codePath": "train/train_deep_wandb.py",
19
+ "git": {
20
+ "remote": "git@hf.co:Yaning1001/Impossible_llm.git",
21
+ "commit": "ed716cdcfcdea02b67f7ed0f3504c2b1c8b737c4"
22
+ },
23
+ "email": "yaning1001@gmail.com",
24
+ "root": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train",
25
+ "host": "mms-large-2",
26
+ "username": "chunhui",
27
+ "executable": "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/bin/python",
28
+ "codePathLocal": "train_deep_wandb.py",
29
+ "cpu_count": 32,
30
+ "cpu_count_logical": 64,
31
+ "gpu": "NVIDIA RTX A6000",
32
+ "gpu_count": 8,
33
+ "disk": {
34
+ "/": {
35
+ "total": "1888559353856",
36
+ "used": "1792542826496"
37
+ }
38
+ },
39
+ "memory": {
40
+ "total": "202617098240"
41
+ },
42
+ "cpu": {
43
+ "count": 32,
44
+ "countLogical": 64
45
+ },
46
+ "gpu_nvidia": [
47
+ {
48
+ "name": "NVIDIA RTX A6000",
49
+ "memoryTotal": "51527024640",
50
+ "cudaCores": 10752,
51
+ "architecture": "Ampere"
52
+ },
53
+ {
54
+ "name": "NVIDIA RTX A6000",
55
+ "memoryTotal": "51527024640",
56
+ "cudaCores": 10752,
57
+ "architecture": "Ampere"
58
+ },
59
+ {
60
+ "name": "NVIDIA RTX A6000",
61
+ "memoryTotal": "51527024640",
62
+ "cudaCores": 10752,
63
+ "architecture": "Ampere"
64
+ },
65
+ {
66
+ "name": "NVIDIA RTX A6000",
67
+ "memoryTotal": "51527024640",
68
+ "cudaCores": 10752,
69
+ "architecture": "Ampere"
70
+ },
71
+ {
72
+ "name": "NVIDIA RTX A6000",
73
+ "memoryTotal": "51527024640",
74
+ "cudaCores": 10752,
75
+ "architecture": "Ampere"
76
+ },
77
+ {
78
+ "name": "NVIDIA RTX A6000",
79
+ "memoryTotal": "51527024640",
80
+ "cudaCores": 10752,
81
+ "architecture": "Ampere"
82
+ },
83
+ {
84
+ "name": "NVIDIA RTX A6000",
85
+ "memoryTotal": "51527024640",
86
+ "cudaCores": 10752,
87
+ "architecture": "Ampere"
88
+ },
89
+ {
90
+ "name": "NVIDIA RTX A6000",
91
+ "memoryTotal": "51527024640",
92
+ "cudaCores": 10752,
93
+ "architecture": "Ampere"
94
+ }
95
+ ],
96
+ "cudaVersion": "11.8"
97
+ }
wandb/run-20241105_160652-il3id7gi/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_wandb":{"runtime":5}}
wandb/run-20241105_160652-il3id7gi/logs/debug-internal.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2024-11-05T16:06:52.167714581-05:00","level":"INFO","msg":"using version","core version":"0.18.5"}
2
+ {"time":"2024-11-05T16:06:52.167734171-05:00","level":"INFO","msg":"created symlink","path":"/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241105_160652-il3id7gi/logs/debug-core.log"}
3
+ {"time":"2024-11-05T16:06:52.276915561-05:00","level":"INFO","msg":"created new stream","id":"il3id7gi"}
4
+ {"time":"2024-11-05T16:06:52.276986211-05:00","level":"INFO","msg":"stream: started","id":"il3id7gi"}
5
+ {"time":"2024-11-05T16:06:52.277107652-05:00","level":"INFO","msg":"sender: started","stream_id":"il3id7gi"}
6
+ {"time":"2024-11-05T16:06:52.277018542-05:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"il3id7gi"}}
7
+ {"time":"2024-11-05T16:06:52.277107022-05:00","level":"INFO","msg":"handler: started","stream_id":{"value":"il3id7gi"}}
8
+ {"time":"2024-11-05T16:06:52.50033616-05:00","level":"INFO","msg":"Starting system monitor"}
9
+ {"time":"2024-11-05T16:06:57.72841623-05:00","level":"INFO","msg":"Stopping system monitor"}
10
+ {"time":"2024-11-05T16:06:57.72836442-05:00","level":"INFO","msg":"stream: closing","id":"il3id7gi"}
11
+ {"time":"2024-11-05T16:06:57.729503476-05:00","level":"INFO","msg":"Stopped system monitor"}
12
+ {"time":"2024-11-05T16:06:57.814719044-05:00","level":"ERROR","msg":"sender: sendDefer: failed to build job artifact","error":"failed to write data to file: write /tmp/tmpfile-193938247: no space left on device"}
13
+ {"time":"2024-11-05T16:06:58.075824688-05:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
14
+ {"time":"2024-11-05T16:06:58.203385044-05:00","level":"INFO","msg":"handler: closed","stream_id":{"value":"il3id7gi"}}
15
+ {"time":"2024-11-05T16:06:58.203426045-05:00","level":"INFO","msg":"writer: Close: closed","stream_id":{"value":"il3id7gi"}}
16
+ {"time":"2024-11-05T16:06:58.203465995-05:00","level":"INFO","msg":"sender: closed","stream_id":"il3id7gi"}
17
+ {"time":"2024-11-05T16:06:58.203502235-05:00","level":"INFO","msg":"stream: closed","id":"il3id7gi"}
wandb/run-20241105_160652-il3id7gi/logs/debug.log ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-11-05 16:06:52,161 INFO MainThread:1771275 [wandb_setup.py:_flush():79] Current SDK version is 0.18.5
2
+ 2024-11-05 16:06:52,161 INFO MainThread:1771275 [wandb_setup.py:_flush():79] Configure stats pid to 1771275
3
+ 2024-11-05 16:06:52,161 INFO MainThread:1771275 [wandb_setup.py:_flush():79] Loading settings from /home/chunhui/.config/wandb/settings
4
+ 2024-11-05 16:06:52,161 INFO MainThread:1771275 [wandb_setup.py:_flush():79] Loading settings from /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/settings
5
+ 2024-11-05 16:06:52,161 INFO MainThread:1771275 [wandb_setup.py:_flush():79] Loading settings from environment variables: {}
6
+ 2024-11-05 16:06:52,161 INFO MainThread:1771275 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': None, '_disable_service': None}
7
+ 2024-11-05 16:06:52,161 INFO MainThread:1771275 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': 'train/train_deep_wandb.py', 'program_abspath': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py', 'program': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py'}
8
+ 2024-11-05 16:06:52,161 INFO MainThread:1771275 [wandb_setup.py:_flush():79] Applying login settings: {}
9
+ 2024-11-05 16:06:52,161 INFO MainThread:1771275 [wandb_init.py:_log_setup():534] Logging user logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241105_160652-il3id7gi/logs/debug.log
10
+ 2024-11-05 16:06:52,161 INFO MainThread:1771275 [wandb_init.py:_log_setup():535] Logging internal logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241105_160652-il3id7gi/logs/debug-internal.log
11
+ 2024-11-05 16:06:52,162 INFO MainThread:1771275 [wandb_init.py:init():621] calling init triggers
12
+ 2024-11-05 16:06:52,162 INFO MainThread:1771275 [wandb_init.py:init():628] wandb.init called with sweep_config: {}
13
+ config: {}
14
+ 2024-11-05 16:06:52,162 INFO MainThread:1771275 [wandb_init.py:init():671] starting backend
15
+ 2024-11-05 16:06:52,162 INFO MainThread:1771275 [wandb_init.py:init():675] sending inform_init request
16
+ 2024-11-05 16:06:52,164 INFO MainThread:1771275 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2024-11-05 16:06:52,164 INFO MainThread:1771275 [wandb_init.py:init():688] backend started and connected
18
+ 2024-11-05 16:06:52,167 INFO MainThread:1771275 [wandb_init.py:init():783] updated telemetry
19
+ 2024-11-05 16:06:52,198 INFO MainThread:1771275 [wandb_init.py:init():816] communicating run to backend with 90.0 second timeout
20
+ 2024-11-05 16:06:52,497 INFO MainThread:1771275 [wandb_init.py:init():867] starting run threads in backend
21
+ 2024-11-05 16:06:52,584 INFO MainThread:1771275 [wandb_run.py:_console_start():2463] atexit reg
22
+ 2024-11-05 16:06:52,584 INFO MainThread:1771275 [wandb_run.py:_redirect():2311] redirect: wrap_raw
23
+ 2024-11-05 16:06:52,584 INFO MainThread:1771275 [wandb_run.py:_redirect():2376] Wrapping output streams.
24
+ 2024-11-05 16:06:52,584 INFO MainThread:1771275 [wandb_run.py:_redirect():2401] Redirects installed.
25
+ 2024-11-05 16:06:52,586 INFO MainThread:1771275 [wandb_init.py:init():911] run started, returning control to user process
26
+ 2024-11-05 16:06:52,586 INFO MainThread:1771275 [wandb_run.py:_config_callback():1390] config_cb None None {'perturbation': 'shuffle_deterministic21', 'train_set': '10M', 'batch_size': 3, 'epoch': 3, 'seed': 0, 'lr': 5e-06}
27
+ 2024-11-05 16:06:57,728 WARNING MsgRouterThr:1771275 [router.py:message_loop():77] message_loop has been closed
wandb/run-20241105_160652-il3id7gi/run-il3id7gi.wandb ADDED
Binary file (2.29 kB). View file
 
wandb/run-20241105_163039-q4e8d8hm/files/config.yaml ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.18.5
4
+ m: []
5
+ python_version: 3.9.19
6
+ t:
7
+ "1":
8
+ - 1
9
+ - 5
10
+ - 11
11
+ - 49
12
+ - 51
13
+ - 53
14
+ - 55
15
+ - 71
16
+ - 98
17
+ "2":
18
+ - 1
19
+ - 5
20
+ - 11
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 98
27
+ "3":
28
+ - 13
29
+ - 23
30
+ - 55
31
+ "4": 3.9.19
32
+ "5": 0.18.5
33
+ "6": 4.45.1
34
+ "8":
35
+ - 5
36
+ "12": 0.18.5
37
+ "13": linux-x86_64
38
+ batch_size:
39
+ value: 3
40
+ epoch:
41
+ value: 3
42
+ lr:
43
+ value: 5e-06
44
+ perturbation:
45
+ value: shuffle_deterministic21
46
+ seed:
47
+ value: 0
48
+ train_set:
49
+ value: 10M
wandb/run-20241105_163039-q4e8d8hm/files/wandb-metadata.json ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-162-generic-x86_64-with-glibc2.31",
3
+ "python": "3.9.19",
4
+ "startedAt": "2024-11-05T21:30:39.177983Z",
5
+ "args": [
6
+ "--perturbation",
7
+ "shuffle_deterministic21",
8
+ "--train_set",
9
+ "10M",
10
+ "--batch_size",
11
+ "3",
12
+ "--epoch",
13
+ "3",
14
+ "--seed",
15
+ "0"
16
+ ],
17
+ "program": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py",
18
+ "codePath": "train/train_deep_wandb.py",
19
+ "git": {
20
+ "remote": "git@hf.co:Yaning1001/Impossible_llm.git",
21
+ "commit": "ed716cdcfcdea02b67f7ed0f3504c2b1c8b737c4"
22
+ },
23
+ "email": "yaning1001@gmail.com",
24
+ "root": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train",
25
+ "host": "mms-large-2",
26
+ "username": "chunhui",
27
+ "executable": "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/bin/python",
28
+ "codePathLocal": "train_deep_wandb.py",
29
+ "cpu_count": 32,
30
+ "cpu_count_logical": 64,
31
+ "gpu": "NVIDIA RTX A6000",
32
+ "gpu_count": 8,
33
+ "disk": {
34
+ "/": {
35
+ "total": "1888559353856",
36
+ "used": "1785984999424"
37
+ }
38
+ },
39
+ "memory": {
40
+ "total": "202617098240"
41
+ },
42
+ "cpu": {
43
+ "count": 32,
44
+ "countLogical": 64
45
+ },
46
+ "gpu_nvidia": [
47
+ {
48
+ "name": "NVIDIA RTX A6000",
49
+ "memoryTotal": "51527024640",
50
+ "cudaCores": 10752,
51
+ "architecture": "Ampere"
52
+ },
53
+ {
54
+ "name": "NVIDIA RTX A6000",
55
+ "memoryTotal": "51527024640",
56
+ "cudaCores": 10752,
57
+ "architecture": "Ampere"
58
+ },
59
+ {
60
+ "name": "NVIDIA RTX A6000",
61
+ "memoryTotal": "51527024640",
62
+ "cudaCores": 10752,
63
+ "architecture": "Ampere"
64
+ },
65
+ {
66
+ "name": "NVIDIA RTX A6000",
67
+ "memoryTotal": "51527024640",
68
+ "cudaCores": 10752,
69
+ "architecture": "Ampere"
70
+ },
71
+ {
72
+ "name": "NVIDIA RTX A6000",
73
+ "memoryTotal": "51527024640",
74
+ "cudaCores": 10752,
75
+ "architecture": "Ampere"
76
+ },
77
+ {
78
+ "name": "NVIDIA RTX A6000",
79
+ "memoryTotal": "51527024640",
80
+ "cudaCores": 10752,
81
+ "architecture": "Ampere"
82
+ },
83
+ {
84
+ "name": "NVIDIA RTX A6000",
85
+ "memoryTotal": "51527024640",
86
+ "cudaCores": 10752,
87
+ "architecture": "Ampere"
88
+ },
89
+ {
90
+ "name": "NVIDIA RTX A6000",
91
+ "memoryTotal": "51527024640",
92
+ "cudaCores": 10752,
93
+ "architecture": "Ampere"
94
+ }
95
+ ],
96
+ "cudaVersion": "11.8"
97
+ }
wandb/run-20241105_163039-q4e8d8hm/logs/debug.log ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-11-05 16:30:39,175 INFO MainThread:1780128 [wandb_setup.py:_flush():79] Current SDK version is 0.18.5
2
+ 2024-11-05 16:30:39,175 INFO MainThread:1780128 [wandb_setup.py:_flush():79] Configure stats pid to 1780128
3
+ 2024-11-05 16:30:39,175 INFO MainThread:1780128 [wandb_setup.py:_flush():79] Loading settings from /home/chunhui/.config/wandb/settings
4
+ 2024-11-05 16:30:39,175 INFO MainThread:1780128 [wandb_setup.py:_flush():79] Loading settings from /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/settings
5
+ 2024-11-05 16:30:39,175 INFO MainThread:1780128 [wandb_setup.py:_flush():79] Loading settings from environment variables: {}
6
+ 2024-11-05 16:30:39,175 INFO MainThread:1780128 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': None, '_disable_service': None}
7
+ 2024-11-05 16:30:39,175 INFO MainThread:1780128 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': 'train/train_deep_wandb.py', 'program_abspath': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py', 'program': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py'}
8
+ 2024-11-05 16:30:39,175 INFO MainThread:1780128 [wandb_setup.py:_flush():79] Applying login settings: {}
9
+ 2024-11-05 16:30:39,175 INFO MainThread:1780128 [wandb_init.py:_log_setup():534] Logging user logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241105_163039-q4e8d8hm/logs/debug.log
10
+ 2024-11-05 16:30:39,175 INFO MainThread:1780128 [wandb_init.py:_log_setup():535] Logging internal logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241105_163039-q4e8d8hm/logs/debug-internal.log
11
+ 2024-11-05 16:30:39,175 INFO MainThread:1780128 [wandb_init.py:init():621] calling init triggers
12
+ 2024-11-05 16:30:39,175 INFO MainThread:1780128 [wandb_init.py:init():628] wandb.init called with sweep_config: {}
13
+ config: {}
14
+ 2024-11-05 16:30:39,175 INFO MainThread:1780128 [wandb_init.py:init():671] starting backend
15
+ 2024-11-05 16:30:39,176 INFO MainThread:1780128 [wandb_init.py:init():675] sending inform_init request
16
+ 2024-11-05 16:30:39,177 INFO MainThread:1780128 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2024-11-05 16:30:39,177 INFO MainThread:1780128 [wandb_init.py:init():688] backend started and connected
18
+ 2024-11-05 16:30:39,181 INFO MainThread:1780128 [wandb_init.py:init():783] updated telemetry
19
+ 2024-11-05 16:30:39,205 INFO MainThread:1780128 [wandb_init.py:init():816] communicating run to backend with 90.0 second timeout
20
+ 2024-11-05 16:30:40,353 INFO MainThread:1780128 [wandb_init.py:init():867] starting run threads in backend
21
+ 2024-11-05 16:30:40,466 INFO MainThread:1780128 [wandb_run.py:_console_start():2463] atexit reg
22
+ 2024-11-05 16:30:40,466 INFO MainThread:1780128 [wandb_run.py:_redirect():2311] redirect: wrap_raw
23
+ 2024-11-05 16:30:40,466 INFO MainThread:1780128 [wandb_run.py:_redirect():2376] Wrapping output streams.
24
+ 2024-11-05 16:30:40,466 INFO MainThread:1780128 [wandb_run.py:_redirect():2401] Redirects installed.
25
+ 2024-11-05 16:30:40,467 INFO MainThread:1780128 [wandb_init.py:init():911] run started, returning control to user process
26
+ 2024-11-05 16:30:40,468 INFO MainThread:1780128 [wandb_run.py:_config_callback():1390] config_cb None None {'perturbation': 'shuffle_deterministic21', 'train_set': '10M', 'batch_size': 3, 'epoch': 3, 'seed': 0, 'lr': 5e-06}
27
+ 2024-11-05 16:32:00,266 WARNING MsgRouterThr:1780128 [router.py:message_loop():77] message_loop has been closed
wandb/run-20241105_163039-q4e8d8hm/run-q4e8d8hm.wandb ADDED
Binary file (32.8 kB). View file
 
wandb/run-20241105_163244-o1vw2gev/logs/debug-internal.log ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {"time":"2024-11-05T16:32:44.915617704-05:00","level":"INFO","msg":"using version","core version":"0.18.5"}
2
+ {"time":"2024-11-05T16:32:44.915629374-05:00","level":"INFO","msg":"created symlink","path":"/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241105_163244-o1vw2gev/logs/debug-core.log"}
3
+ {"time":"2024-11-05T16:32:45.021934355-05:00","level":"INFO","msg":"created new stream","id":"o1vw2gev"}
4
+ {"time":"2024-11-05T16:32:45.021967635-05:00","level":"INFO","msg":"stream: started","id":"o1vw2gev"}
5
+ {"time":"2024-11-05T16:32:45.022053626-05:00","level":"INFO","msg":"sender: started","stream_id":"o1vw2gev"}
6
+ {"time":"2024-11-05T16:32:45.022012265-05:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"o1vw2gev"}}
7
+ {"time":"2024-11-05T16:32:45.022137556-05:00","level":"INFO","msg":"handler: started","stream_id":{"value":"o1vw2gev"}}
8
+ {"time":"2024-11-05T16:32:45.194150228-05:00","level":"INFO","msg":"Starting system monitor"}
wandb/run-20241105_223736-2kagvv0a/files/config.yaml ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.18.5
4
+ m: []
5
+ python_version: 3.9.19
6
+ t:
7
+ "1":
8
+ - 1
9
+ - 5
10
+ - 11
11
+ - 49
12
+ - 51
13
+ - 53
14
+ - 55
15
+ - 71
16
+ - 98
17
+ "2":
18
+ - 1
19
+ - 5
20
+ - 11
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 98
27
+ "3":
28
+ - 2
29
+ - 13
30
+ - 23
31
+ - 55
32
+ "4": 3.9.19
33
+ "5": 0.18.5
34
+ "6": 4.45.1
35
+ "8":
36
+ - 5
37
+ "12": 0.18.5
38
+ "13": linux-x86_64
39
+ batch_size:
40
+ value: 3
41
+ epoch:
42
+ value: 3
43
+ lr:
44
+ value: 5e-06
45
+ perturbation:
46
+ value: shuffle_deterministic21
47
+ seed:
48
+ value: 0
49
+ train_set:
50
+ value: 10M
wandb/run-20241105_223736-2kagvv0a/files/output.log ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb: 500 encountered ({"errors":[{"message":"An internal error occurred. Please contact support.","path":["upsertBucket"]}],"data":{"upsertBucket":null}}), retrying request
2
+ Downloading shards: 100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 2/2 [02:32<00:00, 76.10s/it]
3
+ Loading checkpoint shards: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 2/2 [00:04<00:00, 2.45s/it]
4
+ tokenized_valid: Dataset({
5
+ features: ['input_ids', 'attention_mask'],
6
+ num_rows: 1000
7
+ })
8
+ /mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/training_args.py:1545: FutureWarning: `evaluation_strategy` is deprecated and will be removed in version 4.46 of 🤗 Transformers. Use `eval_strategy` instead
9
+ warnings.warn(
10
+ [2024-11-05 22:40:19,363] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect)
11
+ [2024-11-05 22:40:27,697] [INFO] [comm.py:652:init_distributed] cdb=None
12
+ Installed CUDA version 11.8 does not match the version torch was compiled with 11.7 but since the APIs are compatible, accepting this combination
13
+ Using /home/chunhui/.cache/torch_extensions/py39_cu117 as PyTorch extensions root...
14
+ Emitting ninja build file /home/chunhui/.cache/torch_extensions/py39_cu117/cpu_adam/build.ninja...
15
+ Building extension module cpu_adam...
16
+ Allowing ninja to set a default number of workers... (overridable by setting the environment variable MAX_JOBS=N)
17
+ Loading extension module cpu_adam...
18
+ Time to load cpu_adam op: 5.439234256744385 seconds
wandb/run-20241105_223736-2kagvv0a/files/wandb-metadata.json ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-162-generic-x86_64-with-glibc2.31",
3
+ "python": "3.9.19",
4
+ "startedAt": "2024-11-06T03:37:36.194798Z",
5
+ "args": [
6
+ "--perturbation",
7
+ "shuffle_deterministic21",
8
+ "--train_set",
9
+ "10M",
10
+ "--batch_size",
11
+ "3",
12
+ "--epoch",
13
+ "3",
14
+ "--seed",
15
+ "0"
16
+ ],
17
+ "program": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py",
18
+ "codePath": "train/train_deep_wandb.py",
19
+ "git": {
20
+ "remote": "git@hf.co:Yaning1001/Impossible_llm.git",
21
+ "commit": "ed716cdcfcdea02b67f7ed0f3504c2b1c8b737c4"
22
+ },
23
+ "email": "yaning1001@gmail.com",
24
+ "root": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train",
25
+ "host": "mms-large-2",
26
+ "username": "chunhui",
27
+ "executable": "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/bin/python",
28
+ "codePathLocal": "train_deep_wandb.py",
29
+ "cpu_count": 32,
30
+ "cpu_count_logical": 64,
31
+ "gpu": "NVIDIA RTX A6000",
32
+ "gpu_count": 8,
33
+ "disk": {
34
+ "/": {
35
+ "total": "1888559353856",
36
+ "used": "1785078390784"
37
+ }
38
+ },
39
+ "memory": {
40
+ "total": "202617098240"
41
+ },
42
+ "cpu": {
43
+ "count": 32,
44
+ "countLogical": 64
45
+ },
46
+ "gpu_nvidia": [
47
+ {
48
+ "name": "NVIDIA RTX A6000",
49
+ "memoryTotal": "51527024640",
50
+ "cudaCores": 10752,
51
+ "architecture": "Ampere"
52
+ },
53
+ {
54
+ "name": "NVIDIA RTX A6000",
55
+ "memoryTotal": "51527024640",
56
+ "cudaCores": 10752,
57
+ "architecture": "Ampere"
58
+ },
59
+ {
60
+ "name": "NVIDIA RTX A6000",
61
+ "memoryTotal": "51527024640",
62
+ "cudaCores": 10752,
63
+ "architecture": "Ampere"
64
+ },
65
+ {
66
+ "name": "NVIDIA RTX A6000",
67
+ "memoryTotal": "51527024640",
68
+ "cudaCores": 10752,
69
+ "architecture": "Ampere"
70
+ },
71
+ {
72
+ "name": "NVIDIA RTX A6000",
73
+ "memoryTotal": "51527024640",
74
+ "cudaCores": 10752,
75
+ "architecture": "Ampere"
76
+ },
77
+ {
78
+ "name": "NVIDIA RTX A6000",
79
+ "memoryTotal": "51527024640",
80
+ "cudaCores": 10752,
81
+ "architecture": "Ampere"
82
+ },
83
+ {
84
+ "name": "NVIDIA RTX A6000",
85
+ "memoryTotal": "51527024640",
86
+ "cudaCores": 10752,
87
+ "architecture": "Ampere"
88
+ },
89
+ {
90
+ "name": "NVIDIA RTX A6000",
91
+ "memoryTotal": "51527024640",
92
+ "cudaCores": 10752,
93
+ "architecture": "Ampere"
94
+ }
95
+ ],
96
+ "cudaVersion": "11.8"
97
+ }
wandb/run-20241105_223736-2kagvv0a/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_wandb":{"runtime":77797}}
wandb/run-20241105_223736-2kagvv0a/logs/debug-internal.log ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2024-11-05T22:37:36.202357237-05:00","level":"INFO","msg":"using version","core version":"0.18.5"}
2
+ {"time":"2024-11-05T22:37:36.202369547-05:00","level":"INFO","msg":"created symlink","path":"/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241105_223736-2kagvv0a/logs/debug-core.log"}
3
+ {"time":"2024-11-05T22:37:36.311435374-05:00","level":"INFO","msg":"created new stream","id":"2kagvv0a"}
4
+ {"time":"2024-11-05T22:37:36.311490154-05:00","level":"INFO","msg":"stream: started","id":"2kagvv0a"}
5
+ {"time":"2024-11-05T22:37:36.311518374-05:00","level":"INFO","msg":"sender: started","stream_id":"2kagvv0a"}
6
+ {"time":"2024-11-05T22:37:36.311506174-05:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"2kagvv0a"}}
7
+ {"time":"2024-11-05T22:37:36.311515264-05:00","level":"INFO","msg":"handler: started","stream_id":{"value":"2kagvv0a"}}
8
+ {"time":"2024-11-05T22:37:36.430260321-05:00","level":"INFO","msg":"api: retrying HTTP error","status":500,"url":"https://api.wandb.ai/graphql"}
9
+ {"time":"2024-11-05T22:37:38.657390136-05:00","level":"INFO","msg":"Starting system monitor"}
10
+ {"time":"2024-11-06T20:14:13.275515344-05:00","level":"INFO","msg":"Stopping system monitor"}
11
+ {"time":"2024-11-06T20:14:13.36533327-05:00","level":"INFO","msg":"Stopped system monitor"}
12
+ {"time":"2024-11-06T20:14:14.247035454-05:00","level":"INFO","msg":"handler: operation stats","stats":{"operations":[{"desc":"saving job artifact","runtime_seconds":0.766071365,"subtasks":[{"desc":"requirements.frozen.txt","runtime_seconds":0.267306342,"progress":"2.6KB/2.6KB"}]}],"total_operations":1}}
13
+ {"time":"2024-11-06T20:14:14.897915803-05:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
14
+ {"time":"2024-11-06T20:14:16.038235461-05:00","level":"INFO","msg":"stream: closing","id":"2kagvv0a"}
15
+ {"time":"2024-11-06T20:14:16.038268531-05:00","level":"INFO","msg":"handler: closed","stream_id":{"value":"2kagvv0a"}}
16
+ {"time":"2024-11-06T20:14:16.038289471-05:00","level":"INFO","msg":"writer: Close: closed","stream_id":{"value":"2kagvv0a"}}
17
+ {"time":"2024-11-06T20:14:16.038310401-05:00","level":"INFO","msg":"sender: closed","stream_id":"2kagvv0a"}
18
+ {"time":"2024-11-06T20:14:16.038388792-05:00","level":"INFO","msg":"stream: closed","id":"2kagvv0a"}
wandb/run-20241105_223736-2kagvv0a/logs/debug.log ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-11-05 22:37:36,192 INFO MainThread:1802405 [wandb_setup.py:_flush():79] Current SDK version is 0.18.5
2
+ 2024-11-05 22:37:36,193 INFO MainThread:1802405 [wandb_setup.py:_flush():79] Configure stats pid to 1802405
3
+ 2024-11-05 22:37:36,193 INFO MainThread:1802405 [wandb_setup.py:_flush():79] Loading settings from /home/chunhui/.config/wandb/settings
4
+ 2024-11-05 22:37:36,193 INFO MainThread:1802405 [wandb_setup.py:_flush():79] Loading settings from /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/settings
5
+ 2024-11-05 22:37:36,193 INFO MainThread:1802405 [wandb_setup.py:_flush():79] Loading settings from environment variables: {}
6
+ 2024-11-05 22:37:36,193 INFO MainThread:1802405 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': None, '_disable_service': None}
7
+ 2024-11-05 22:37:36,193 INFO MainThread:1802405 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': 'train/train_deep_wandb.py', 'program_abspath': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py', 'program': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py'}
8
+ 2024-11-05 22:37:36,193 INFO MainThread:1802405 [wandb_setup.py:_flush():79] Applying login settings: {}
9
+ 2024-11-05 22:37:36,193 INFO MainThread:1802405 [wandb_init.py:_log_setup():534] Logging user logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241105_223736-2kagvv0a/logs/debug.log
10
+ 2024-11-05 22:37:36,193 INFO MainThread:1802405 [wandb_init.py:_log_setup():535] Logging internal logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241105_223736-2kagvv0a/logs/debug-internal.log
11
+ 2024-11-05 22:37:36,193 INFO MainThread:1802405 [wandb_init.py:init():621] calling init triggers
12
+ 2024-11-05 22:37:36,193 INFO MainThread:1802405 [wandb_init.py:init():628] wandb.init called with sweep_config: {}
13
+ config: {}
14
+ 2024-11-05 22:37:36,193 INFO MainThread:1802405 [wandb_init.py:init():671] starting backend
15
+ 2024-11-05 22:37:36,193 INFO MainThread:1802405 [wandb_init.py:init():675] sending inform_init request
16
+ 2024-11-05 22:37:36,194 INFO MainThread:1802405 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2024-11-05 22:37:36,194 INFO MainThread:1802405 [wandb_init.py:init():688] backend started and connected
18
+ 2024-11-05 22:37:36,197 INFO MainThread:1802405 [wandb_init.py:init():783] updated telemetry
19
+ 2024-11-05 22:37:36,222 INFO MainThread:1802405 [wandb_init.py:init():816] communicating run to backend with 90.0 second timeout
20
+ 2024-11-05 22:37:38,650 INFO MainThread:1802405 [wandb_init.py:init():867] starting run threads in backend
21
+ 2024-11-05 22:37:38,746 INFO MainThread:1802405 [wandb_run.py:_console_start():2463] atexit reg
22
+ 2024-11-05 22:37:38,746 INFO MainThread:1802405 [wandb_run.py:_redirect():2311] redirect: wrap_raw
23
+ 2024-11-05 22:37:38,746 INFO MainThread:1802405 [wandb_run.py:_redirect():2376] Wrapping output streams.
24
+ 2024-11-05 22:37:38,746 INFO MainThread:1802405 [wandb_run.py:_redirect():2401] Redirects installed.
25
+ 2024-11-05 22:37:38,747 INFO MainThread:1802405 [wandb_init.py:init():911] run started, returning control to user process
26
+ 2024-11-05 22:37:38,747 INFO MainThread:1802405 [wandb_run.py:_config_callback():1390] config_cb None None {'perturbation': 'shuffle_deterministic21', 'train_set': '10M', 'batch_size': 3, 'epoch': 3, 'seed': 0, 'lr': 5e-06}
27
+ 2024-11-06 20:14:13,216 INFO MainThread:1802405 [wandb_run.py:_finish():2158] finishing run yaning1001-dartmouth-college/exp-impo-shuffle/2kagvv0a
28
+ 2024-11-06 20:14:13,227 INFO MainThread:1802405 [wandb_run.py:_atexit_cleanup():2426] got exitcode: 0
29
+ 2024-11-06 20:14:13,227 INFO MainThread:1802405 [wandb_run.py:_restore():2408] restore
30
+ 2024-11-06 20:14:13,246 INFO MainThread:1802405 [wandb_run.py:_restore():2414] restore done
31
+ 2024-11-06 20:14:16,032 INFO MainThread:1802405 [wandb_run.py:_footer_history_summary_info():3975] rendering history
32
+ 2024-11-06 20:14:16,032 INFO MainThread:1802405 [wandb_run.py:_footer_history_summary_info():4007] rendering summary
33
+ 2024-11-06 20:14:16,037 INFO MainThread:1802405 [wandb_run.py:_footer_sync_info():3934] logging synced files
wandb/run-20241106_203647-jauklpbe/files/config.yaml ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.18.5
4
+ m: []
5
+ python_version: 3.9.19
6
+ t:
7
+ "1":
8
+ - 1
9
+ - 5
10
+ - 11
11
+ - 49
12
+ - 51
13
+ - 53
14
+ - 55
15
+ - 71
16
+ - 98
17
+ "2":
18
+ - 1
19
+ - 5
20
+ - 11
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 98
27
+ "3":
28
+ - 13
29
+ - 23
30
+ - 55
31
+ "4": 3.9.19
32
+ "5": 0.18.5
33
+ "6": 4.45.1
34
+ "8":
35
+ - 5
36
+ "12": 0.18.5
37
+ "13": linux-x86_64
38
+ batch_size:
39
+ value: 3
40
+ epoch:
41
+ value: 3
42
+ lr:
43
+ value: 5e-06
44
+ perturbation:
45
+ value: shuffle_deterministic84
46
+ seed:
47
+ value: 0
48
+ train_set:
49
+ value: 10M
wandb/run-20241106_203647-jauklpbe/files/output.log ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Traceback (most recent call last):
2
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/utils/_http.py", line 406, in hf_raise_for_status
3
+ response.raise_for_status()
4
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/requests/models.py", line 1024, in raise_for_status
5
+ raise HTTPError(http_error_msg, response=self)
6
+ requests.exceptions.HTTPError: 401 Client Error: Unauthorized for url: https://huggingface.co/meta-llama/Llama-3.2-3B/resolve/main/config.json
7
+
8
+ The above exception was the direct cause of the following exception:
9
+
10
+ Traceback (most recent call last):
11
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/utils/hub.py", line 403, in cached_file
12
+ resolved_file = hf_hub_download(
13
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/utils/_deprecation.py", line 101, in inner_f
14
+ return f(*args, **kwargs)
15
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn
16
+ return fn(*args, **kwargs)
17
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/file_download.py", line 1232, in hf_hub_download
18
+ return _hf_hub_download_to_cache_dir(
19
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/file_download.py", line 1339, in _hf_hub_download_to_cache_dir
20
+ _raise_on_head_call_error(head_call_error, force_download, local_files_only)
21
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/file_download.py", line 1854, in _raise_on_head_call_error
22
+ raise head_call_error
23
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/file_download.py", line 1746, in _get_metadata_or_catch_error
24
+ metadata = get_hf_file_metadata(
25
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn
26
+ return fn(*args, **kwargs)
27
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/file_download.py", line 1666, in get_hf_file_metadata
28
+ r = _request_wrapper(
29
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/file_download.py", line 364, in _request_wrapper
30
+ response = _request_wrapper(
31
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/file_download.py", line 388, in _request_wrapper
32
+ hf_raise_for_status(response)
33
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/utils/_http.py", line 423, in hf_raise_for_status
34
+ raise _format(GatedRepoError, message, response) from e
35
+ huggingface_hub.errors.GatedRepoError: 401 Client Error. (Request ID: Root=1-672c19c6-4ab49c2772cb92404817a4a6;d7309389-7bc2-43ca-8268-d31dc94d9f27)
36
+
37
+ Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.2-3B/resolve/main/config.json.
38
+ Access to model meta-llama/Llama-3.2-3B is restricted. You must have access to it and be authenticated to access it. Please log in.
39
+
40
+ The above exception was the direct cause of the following exception:
41
+
42
+ Traceback (most recent call last):
43
+ File "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py", line 174, in <module>
44
+ model = AutoModelForCausalLM.from_pretrained(model_name,
45
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/models/auto/auto_factory.py", line 526, in from_pretrained
46
+ config, kwargs = AutoConfig.from_pretrained(
47
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/models/auto/configuration_auto.py", line 1006, in from_pretrained
48
+ config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs)
49
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/configuration_utils.py", line 567, in get_config_dict
50
+ config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
51
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/configuration_utils.py", line 626, in _get_config_dict
52
+ resolved_config_file = cached_file(
53
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/utils/hub.py", line 421, in cached_file
54
+ raise EnvironmentError(
55
+ OSError: You are trying to access a gated repo.
56
+ Make sure to have access to it at https://huggingface.co/meta-llama/Llama-3.2-3B.
57
+ 401 Client Error. (Request ID: Root=1-672c19c6-4ab49c2772cb92404817a4a6;d7309389-7bc2-43ca-8268-d31dc94d9f27)
58
+
59
+ Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.2-3B/resolve/main/config.json.
60
+ Access to model meta-llama/Llama-3.2-3B is restricted. You must have access to it and be authenticated to access it. Please log in.
wandb/run-20241106_203647-jauklpbe/files/wandb-metadata.json ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-162-generic-x86_64-with-glibc2.31",
3
+ "python": "3.9.19",
4
+ "startedAt": "2024-11-07T01:36:47.516115Z",
5
+ "args": [
6
+ "--perturbation",
7
+ "shuffle_deterministic84",
8
+ "--train_set",
9
+ "10M",
10
+ "--batch_size",
11
+ "3",
12
+ "--epoch",
13
+ "3",
14
+ "--seed",
15
+ "0"
16
+ ],
17
+ "program": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py",
18
+ "codePath": "train/train_deep_wandb.py",
19
+ "git": {
20
+ "remote": "git@hf.co:Yaning1001/Impossible_llm.git",
21
+ "commit": "ed716cdcfcdea02b67f7ed0f3504c2b1c8b737c4"
22
+ },
23
+ "email": "yaning1001@gmail.com",
24
+ "root": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train",
25
+ "host": "mms-large-2",
26
+ "username": "chunhui",
27
+ "executable": "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/bin/python",
28
+ "codePathLocal": "train_deep_wandb.py",
29
+ "cpu_count": 32,
30
+ "cpu_count_logical": 64,
31
+ "gpu": "NVIDIA RTX A6000",
32
+ "gpu_count": 8,
33
+ "disk": {
34
+ "/": {
35
+ "total": "1888559353856",
36
+ "used": "1791616151552"
37
+ }
38
+ },
39
+ "memory": {
40
+ "total": "202617098240"
41
+ },
42
+ "cpu": {
43
+ "count": 32,
44
+ "countLogical": 64
45
+ },
46
+ "gpu_nvidia": [
47
+ {
48
+ "name": "NVIDIA RTX A6000",
49
+ "memoryTotal": "51527024640",
50
+ "cudaCores": 10752,
51
+ "architecture": "Ampere"
52
+ },
53
+ {
54
+ "name": "NVIDIA RTX A6000",
55
+ "memoryTotal": "51527024640",
56
+ "cudaCores": 10752,
57
+ "architecture": "Ampere"
58
+ },
59
+ {
60
+ "name": "NVIDIA RTX A6000",
61
+ "memoryTotal": "51527024640",
62
+ "cudaCores": 10752,
63
+ "architecture": "Ampere"
64
+ },
65
+ {
66
+ "name": "NVIDIA RTX A6000",
67
+ "memoryTotal": "51527024640",
68
+ "cudaCores": 10752,
69
+ "architecture": "Ampere"
70
+ },
71
+ {
72
+ "name": "NVIDIA RTX A6000",
73
+ "memoryTotal": "51527024640",
74
+ "cudaCores": 10752,
75
+ "architecture": "Ampere"
76
+ },
77
+ {
78
+ "name": "NVIDIA RTX A6000",
79
+ "memoryTotal": "51527024640",
80
+ "cudaCores": 10752,
81
+ "architecture": "Ampere"
82
+ },
83
+ {
84
+ "name": "NVIDIA RTX A6000",
85
+ "memoryTotal": "51527024640",
86
+ "cudaCores": 10752,
87
+ "architecture": "Ampere"
88
+ },
89
+ {
90
+ "name": "NVIDIA RTX A6000",
91
+ "memoryTotal": "51527024640",
92
+ "cudaCores": 10752,
93
+ "architecture": "Ampere"
94
+ }
95
+ ],
96
+ "cudaVersion": "11.8"
97
+ }
wandb/run-20241106_203647-jauklpbe/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_wandb":{"runtime":23}}