Yaning1001 commited on
Commit
b87502c
·
verified ·
1 Parent(s): aa3912a

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +4 -0
  2. wandb/run-20241030_013141-l1dv6fk2/files/config.yaml +47 -0
  3. wandb/run-20241030_013141-l1dv6fk2/files/output.log +61 -0
  4. wandb/run-20241030_013141-l1dv6fk2/files/requirements.txt +147 -0
  5. wandb/run-20241030_013141-l1dv6fk2/files/wandb-metadata.json +97 -0
  6. wandb/run-20241030_013141-l1dv6fk2/files/wandb-summary.json +1 -0
  7. wandb/run-20241030_013141-l1dv6fk2/logs/debug-internal.log +11 -0
  8. wandb/run-20241030_013141-l1dv6fk2/logs/debug.log +27 -0
  9. wandb/run-20241030_013141-l1dv6fk2/run-l1dv6fk2.wandb +0 -0
  10. wandb/run-20241031_000839-xm9f2csf/files/output.log +16 -0
  11. wandb/run-20241031_000839-xm9f2csf/files/requirements.txt +147 -0
  12. wandb/run-20241031_000839-xm9f2csf/files/wandb-metadata.json +97 -0
  13. wandb/run-20241031_000839-xm9f2csf/logs/debug-internal.log +8 -0
  14. wandb/run-20241031_000839-xm9f2csf/logs/debug.log +26 -0
  15. wandb/run-20241031_000839-xm9f2csf/run-xm9f2csf.wandb +0 -0
  16. wandb/run-20241031_122005-oby2afce/files/config.yaml +49 -0
  17. wandb/run-20241031_122005-oby2afce/files/output.log +49 -0
  18. wandb/run-20241031_122005-oby2afce/files/wandb-metadata.json +97 -0
  19. wandb/run-20241031_122005-oby2afce/files/wandb-summary.json +1 -0
  20. wandb/run-20241031_122005-oby2afce/logs/debug-internal.log +11 -0
  21. wandb/run-20241031_122005-oby2afce/logs/debug.log +27 -0
  22. wandb/run-20241031_122005-oby2afce/run-oby2afce.wandb +0 -0
  23. wandb/run-20241101_012438-pm66le56/files/config.yaml +49 -0
  24. wandb/run-20241101_012438-pm66le56/files/output.log +12 -0
  25. wandb/run-20241101_012438-pm66le56/files/wandb-metadata.json +29 -0
  26. wandb/run-20241101_012438-pm66le56/files/wandb-summary.json +1 -0
  27. wandb/run-20241101_012438-pm66le56/logs/debug-internal.log +16 -0
  28. wandb/run-20241101_012438-pm66le56/logs/debug.log +27 -0
  29. wandb/run-20241101_012438-pm66le56/run-pm66le56.wandb +0 -0
  30. wandb/run-20241101_200502-ydj2ohih/files/output.log +1 -0
  31. wandb/run-20241101_200502-ydj2ohih/files/requirements.txt +147 -0
  32. wandb/run-20241101_200502-ydj2ohih/files/wandb-metadata.json +97 -0
  33. wandb/run-20241101_200502-ydj2ohih/logs/debug-internal.log +8 -0
  34. wandb/run-20241101_200502-ydj2ohih/logs/debug.log +26 -0
  35. wandb/run-20241101_200502-ydj2ohih/run-ydj2ohih.wandb +0 -0
  36. wandb/run-20241101_200517-7w8d5k1u/files/config.yaml +49 -0
  37. wandb/run-20241101_200517-7w8d5k1u/files/output.log +53 -0
  38. wandb/run-20241101_200517-7w8d5k1u/files/wandb-metadata.json +97 -0
  39. wandb/run-20241101_200517-7w8d5k1u/files/wandb-summary.json +1 -0
  40. wandb/run-20241101_200517-7w8d5k1u/logs/debug-internal.log +11 -0
  41. wandb/run-20241101_200517-7w8d5k1u/logs/debug.log +27 -0
  42. wandb/run-20241101_200517-7w8d5k1u/run-7w8d5k1u.wandb +0 -0
  43. wandb/run-20241101_200535-kh416n4m/files/output.log +14 -0
  44. wandb/run-20241101_200535-kh416n4m/files/requirements.txt +147 -0
  45. wandb/run-20241101_200535-kh416n4m/files/wandb-metadata.json +97 -0
  46. wandb/run-20241101_200535-kh416n4m/logs/debug-internal.log +8 -0
  47. wandb/run-20241101_200535-kh416n4m/logs/debug.log +26 -0
  48. wandb/run-20241101_201630-e5gt2fir/files/wandb-metadata.json +29 -0
  49. wandb/run-20241101_201630-e5gt2fir/run-e5gt2fir.wandb +0 -0
  50. wandb/run-20241101_202058-ptl7coag/run-ptl7coag.wandb +3 -0
.gitattributes CHANGED
@@ -88,3 +88,7 @@ wandb/run-20241113_180154-hcvkmgna/run-hcvkmgna.wandb filter=lfs diff=lfs merge=
88
  wandb/run-20241130_011200-7wusu3nn/run-7wusu3nn.wandb filter=lfs diff=lfs merge=lfs -text
89
  wandb/run-20241031_001056-a89didan/run-a89didan.wandb filter=lfs diff=lfs merge=lfs -text
90
  wandb/run-20241129_235322-uoj0g6xp/run-uoj0g6xp.wandb filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
88
  wandb/run-20241130_011200-7wusu3nn/run-7wusu3nn.wandb filter=lfs diff=lfs merge=lfs -text
89
  wandb/run-20241031_001056-a89didan/run-a89didan.wandb filter=lfs diff=lfs merge=lfs -text
90
  wandb/run-20241129_235322-uoj0g6xp/run-uoj0g6xp.wandb filter=lfs diff=lfs merge=lfs -text
91
+ wandb/run-20241106_232725-f16bcfrx/run-f16bcfrx.wandb filter=lfs diff=lfs merge=lfs -text
92
+ wandb/run-20241105_163244-59l4qxgx/run-59l4qxgx.wandb filter=lfs diff=lfs merge=lfs -text
93
+ wandb/run-20241101_202058-ptl7coag/run-ptl7coag.wandb filter=lfs diff=lfs merge=lfs -text
94
+ wandb/run-20241129_235322-bxqdruiw/run-bxqdruiw.wandb filter=lfs diff=lfs merge=lfs -text
wandb/run-20241030_013141-l1dv6fk2/files/config.yaml ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.18.5
4
+ m: []
5
+ python_version: 3.9.19
6
+ t:
7
+ "1":
8
+ - 1
9
+ - 5
10
+ - 11
11
+ - 49
12
+ - 51
13
+ - 53
14
+ - 55
15
+ - 71
16
+ - 98
17
+ "2":
18
+ - 1
19
+ - 5
20
+ - 11
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 98
27
+ "3":
28
+ - 13
29
+ - 23
30
+ - 55
31
+ "4": 3.9.19
32
+ "5": 0.18.5
33
+ "6": 4.45.1
34
+ "8":
35
+ - 5
36
+ "12": 0.18.5
37
+ "13": linux-x86_64
38
+ batch_size:
39
+ value: 3
40
+ epoch:
41
+ value: 7
42
+ perturbation:
43
+ value: reverse_full
44
+ seed:
45
+ value: 0
46
+ train_set:
47
+ value: 10M
wandb/run-20241030_013141-l1dv6fk2/files/output.log ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Downloading shards: 0%| | 0/2 [00:00<?, ?it/s]Exception ignored in: <generator object tqdm.__iter__ at 0x7f620061fdd0>
2
+ Traceback (most recent call last):
3
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/tqdm/std.py", line 1196, in __iter__
4
+ self.close()
5
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/tqdm/std.py", line 1302, in close
6
+ self.display(pos=0)
7
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/tqdm/std.py", line 1495, in display
8
+ self.sp(self.__str__() if msg is None else msg)
9
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/tqdm/std.py", line 459, in print_status
10
+ fp_write('\r' + s + (' ' * max(last_len[0] - len_s, 0)))
11
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/tqdm/std.py", line 452, in fp_write
12
+ fp.write(str(s))
13
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/tqdm/utils.py", line 196, in inner
14
+ return func(*args, **kwargs)
15
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/wandb/sdk/lib/redirect.py", line 648, in write
16
+ cb(data)
17
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/wandb/sdk/wandb_run.py", line 2386, in <lambda>
18
+ lambda data: self._console_raw_callback("stderr", data),
19
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/wandb/sdk/wandb_run.py", line 403, in wrapper_fn
20
+ return func(self, *args, **kwargs)
21
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/wandb/sdk/wandb_run.py", line 1547, in _console_raw_callback
22
+ self._backend.interface.publish_output_raw(name, data)
23
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/wandb/sdk/interface/interface.py", line 721, in publish_output_raw
24
+ self._publish_output_raw(o)
25
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/wandb/sdk/interface/interface_shared.py", line 79, in _publish_output_raw
26
+ self._publish(rec)
27
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/wandb/sdk/interface/interface_sock.py", line 51, in _publish
28
+ self._sock_client.send_record_publish(record)
29
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/wandb/sdk/lib/sock_client.py", line 225, in send_record_publish
30
+ self.send_server_request(server_req)
31
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/wandb/sdk/lib/sock_client.py", line 157, in send_server_request
32
+ self._send_message(msg)
33
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/wandb/sdk/lib/sock_client.py", line 149, in _send_message
34
+ raw_size = msg.ByteSize()
35
+ KeyboardInterrupt:
36
+ Traceback (most recent call last):
37
+ File "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py", line 172, in <module>
38
+ model = AutoModelForCausalLM.from_pretrained(model_name,
39
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/models/auto/auto_factory.py", line 564, in from_pretrained
40
+ return model_class.from_pretrained(
41
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/modeling_utils.py", line 3769, in from_pretrained
42
+ resolved_archive_file, sharded_metadata = get_checkpoint_shard_files(
43
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/utils/hub.py", line 1098, in get_checkpoint_shard_files
44
+ cached_filename = cached_file(
45
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/utils/hub.py", line 403, in cached_file
46
+ resolved_file = hf_hub_download(
47
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/utils/_deprecation.py", line 101, in inner_f
48
+ return f(*args, **kwargs)
49
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn
50
+ return fn(*args, **kwargs)
51
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/file_download.py", line 1232, in hf_hub_download
52
+ return _hf_hub_download_to_cache_dir(
53
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/file_download.py", line 1380, in _hf_hub_download_to_cache_dir
54
+ with WeakFileLock(lock_path):
55
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/contextlib.py", line 119, in __enter__
56
+ return next(self.gen)
57
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/utils/_fixes.py", line 98, in WeakFileLock
58
+ lock.acquire()
59
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/filelock/_api.py", line 225, in acquire
60
+ time.sleep(poll_interval)
61
+ KeyboardInterrupt
wandb/run-20241030_013141-l1dv6fk2/files/requirements.txt ADDED
@@ -0,0 +1,147 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ funcsigs==1.0.2
2
+ sentry-sdk==2.17.0
3
+ multiprocess==0.70.16
4
+ numpy==1.26.2
5
+ pluralizer==1.2.0
6
+ debugpy==1.6.7
7
+ nvidia-cudnn-cu11==8.5.0.96
8
+ deepspeed==0.15.2
9
+ data==0.4
10
+ pandas==2.1.3
11
+ tomli==2.0.1
12
+ charset-normalizer==3.3.2
13
+ attrs==24.2.0
14
+ aiosignal==1.3.1
15
+ fsspec==2023.10.0
16
+ nvidia-cusparse-cu11==11.7.4.91
17
+ zipp==3.12.0
18
+ mypy-extensions==1.0.0
19
+ datasets==3.0.1
20
+ joblib==1.3.2
21
+ hjson==3.1.0
22
+ traitlets==5.7.1
23
+ stack-data==0.6.0
24
+ transformers==4.45.1
25
+ sympy==1.11.1
26
+ Pygments==2.15.0
27
+ docker-pycreds==0.4.0
28
+ dill==0.3.8
29
+ wheel==0.44.0
30
+ prompt-toolkit==3.0.30
31
+ parso==0.8.3
32
+ ipykernel==6.23.1
33
+ pyarrow==17.0.0
34
+ certifi==2023.11.17
35
+ nvidia-cufft-cu11==10.9.0.58
36
+ six==1.16.0
37
+ pydantic==2.9.2
38
+ click==8.1.7
39
+ nest-asyncio==1.5.6
40
+ gmpy2==2.1.0
41
+ matplotlib==3.8.2
42
+ scipy==1.11.4
43
+ typing_extensions==4.12.2
44
+ statsmodels==0.14.0
45
+ huggingface-hub==0.25.0
46
+ frozenlist==1.4.1
47
+ gpustat==1.1.1
48
+ nvidia-nvtx-cu11==11.7.91
49
+ safetensors==0.4.5
50
+ stanza==1.9.2
51
+ decorator==5.1.1
52
+ seaborn==0.13.0
53
+ sentencepiece==0.2.0
54
+ PyYAML==6.0.1
55
+ black==24.8.0
56
+ protobuf==4.25.1
57
+ pickleshare==0.7.5
58
+ peft==0.13.0
59
+ triton==2.0.0
60
+ nvidia-cuda-runtime-cu11==11.7.99
61
+ Jinja2==3.1.2
62
+ nvidia-cusolver-cu11==11.4.0.1
63
+ executing==1.2.0
64
+ jupyter_client==8.1.0
65
+ pluggy==1.3.0
66
+ cmake==3.30.3
67
+ pytz==2023.3.post1
68
+ aiohappyeyeballs==2.4.2
69
+ kiwisolver==1.4.5
70
+ py-cpuinfo==9.0.0
71
+ Pillow==10.1.0
72
+ ptyprocess==0.7.0
73
+ importlib_resources==6.4.5
74
+ GitPython==3.1.43
75
+ importlib-metadata==6.0.0
76
+ iniconfig==2.0.0
77
+ scikit-learn==1.3.2
78
+ exceptiongroup==1.1.0
79
+ networkx==2.8.6
80
+ accelerate==1.0.0
81
+ nltk==3.8.1
82
+ shutilwhich==1.1.0
83
+ fonttools==4.45.1
84
+ future==0.18.3
85
+ aiohttp==3.10.6
86
+ wcwidth==0.2.5
87
+ idna==3.6
88
+ filelock==3.12.2
89
+ pathspec==0.12.1
90
+ jupyter_core==5.1.0
91
+ lit==18.1.8
92
+ nvidia-curand-cu11==10.2.10.91
93
+ nvidia-cublas-cu11==11.10.3.66
94
+ nvidia-ml-py==12.560.30
95
+ msgpack==1.1.0
96
+ python-dateutil==2.8.2
97
+ blessed==1.20.0
98
+ packaging==23.0
99
+ gitdb==4.0.11
100
+ yarl==1.13.0
101
+ emoji==2.8.0
102
+ tzdata==2023.3
103
+ cycler==0.12.1
104
+ tornado==6.2
105
+ backcall==0.2.0
106
+ plotnine==0.12.4
107
+ ninja==1.11.1.1
108
+ latex==0.7.0
109
+ wandb==0.18.5
110
+ setproctitle==1.3.3
111
+ threadpoolctl==3.2.0
112
+ requests==2.32.3
113
+ pyparsing==3.1.1
114
+ smmap==5.0.1
115
+ pyzmq==23.0.0
116
+ async-timeout==4.0.3
117
+ annotated-types==0.7.0
118
+ matplotlib-inline==0.1.6
119
+ latexcodec==1.0.0
120
+ ipython==8.0.0
121
+ patsy==0.5.3
122
+ contourpy==1.2.0
123
+ multidict==6.1.0
124
+ mizani==0.9.3
125
+ urllib3==2.1.0
126
+ tokenizers==0.20.0
127
+ MarkupSafe==2.1.2
128
+ pip==24.2
129
+ pexpect==4.8.0
130
+ tqdm==4.66.5
131
+ jedi==0.18.2
132
+ pydantic_core==2.23.4
133
+ tempdir==0.7.1
134
+ mpmath==1.2.1
135
+ setuptools==72.1.0
136
+ pytest==7.4.3
137
+ pure-eval==0.2.2
138
+ psutil==5.9.1
139
+ comm==0.1.2
140
+ nvidia-cuda-cupti-cu11==11.7.101
141
+ nvidia-cuda-nvrtc-cu11==11.7.99
142
+ regex==2023.10.3
143
+ platformdirs==2.5.2
144
+ asttokens==2.2.1
145
+ torch==2.0.0
146
+ nvidia-nccl-cu11==2.14.3
147
+ xxhash==3.5.0
wandb/run-20241030_013141-l1dv6fk2/files/wandb-metadata.json ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-162-generic-x86_64-with-glibc2.31",
3
+ "python": "3.9.19",
4
+ "startedAt": "2024-10-30T05:31:41.769137Z",
5
+ "args": [
6
+ "--perturbation",
7
+ "reverse_full",
8
+ "--train_set",
9
+ "10M",
10
+ "--batch_size",
11
+ "3",
12
+ "--epoch",
13
+ "7",
14
+ "--seed",
15
+ "0"
16
+ ],
17
+ "program": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py",
18
+ "codePath": "train/train_deep_wandb.py",
19
+ "git": {
20
+ "remote": "git@hf.co:Yaning1001/Impossible_llm.git",
21
+ "commit": "ed716cdcfcdea02b67f7ed0f3504c2b1c8b737c4"
22
+ },
23
+ "email": "yaning1001@gmail.com",
24
+ "root": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train",
25
+ "host": "mms-large-2",
26
+ "username": "chunhui",
27
+ "executable": "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/bin/python",
28
+ "codePathLocal": "train_deep_wandb.py",
29
+ "cpu_count": 32,
30
+ "cpu_count_logical": 64,
31
+ "gpu": "NVIDIA RTX A6000",
32
+ "gpu_count": 8,
33
+ "disk": {
34
+ "/": {
35
+ "total": "1888559353856",
36
+ "used": "1709824413696"
37
+ }
38
+ },
39
+ "memory": {
40
+ "total": "202617098240"
41
+ },
42
+ "cpu": {
43
+ "count": 32,
44
+ "countLogical": 64
45
+ },
46
+ "gpu_nvidia": [
47
+ {
48
+ "name": "NVIDIA RTX A6000",
49
+ "memoryTotal": "51527024640",
50
+ "cudaCores": 10752,
51
+ "architecture": "Ampere"
52
+ },
53
+ {
54
+ "name": "NVIDIA RTX A6000",
55
+ "memoryTotal": "51527024640",
56
+ "cudaCores": 10752,
57
+ "architecture": "Ampere"
58
+ },
59
+ {
60
+ "name": "NVIDIA RTX A6000",
61
+ "memoryTotal": "51527024640",
62
+ "cudaCores": 10752,
63
+ "architecture": "Ampere"
64
+ },
65
+ {
66
+ "name": "NVIDIA RTX A6000",
67
+ "memoryTotal": "51527024640",
68
+ "cudaCores": 10752,
69
+ "architecture": "Ampere"
70
+ },
71
+ {
72
+ "name": "NVIDIA RTX A6000",
73
+ "memoryTotal": "51527024640",
74
+ "cudaCores": 10752,
75
+ "architecture": "Ampere"
76
+ },
77
+ {
78
+ "name": "NVIDIA RTX A6000",
79
+ "memoryTotal": "51527024640",
80
+ "cudaCores": 10752,
81
+ "architecture": "Ampere"
82
+ },
83
+ {
84
+ "name": "NVIDIA RTX A6000",
85
+ "memoryTotal": "51527024640",
86
+ "cudaCores": 10752,
87
+ "architecture": "Ampere"
88
+ },
89
+ {
90
+ "name": "NVIDIA RTX A6000",
91
+ "memoryTotal": "51527024640",
92
+ "cudaCores": 10752,
93
+ "architecture": "Ampere"
94
+ }
95
+ ],
96
+ "cudaVersion": "11.8"
97
+ }
wandb/run-20241030_013141-l1dv6fk2/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_wandb":{"runtime":94}}
wandb/run-20241030_013141-l1dv6fk2/logs/debug-internal.log ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2024-10-30T01:31:41.770829124-04:00","level":"INFO","msg":"using version","core version":"0.18.5"}
2
+ {"time":"2024-10-30T01:31:41.770841454-04:00","level":"INFO","msg":"created symlink","path":"/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241030_013141-l1dv6fk2/logs/debug-core.log"}
3
+ {"time":"2024-10-30T01:31:41.876553546-04:00","level":"INFO","msg":"created new stream","id":"l1dv6fk2"}
4
+ {"time":"2024-10-30T01:31:41.876597536-04:00","level":"INFO","msg":"stream: started","id":"l1dv6fk2"}
5
+ {"time":"2024-10-30T01:31:41.876671397-04:00","level":"INFO","msg":"sender: started","stream_id":"l1dv6fk2"}
6
+ {"time":"2024-10-30T01:31:41.876656516-04:00","level":"INFO","msg":"handler: started","stream_id":{"value":"l1dv6fk2"}}
7
+ {"time":"2024-10-30T01:31:41.876621636-04:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"l1dv6fk2"}}
8
+ {"time":"2024-10-30T01:31:42.076183199-04:00","level":"INFO","msg":"Starting system monitor"}
9
+ {"time":"2024-10-30T01:33:16.601963699-04:00","level":"INFO","msg":"stream: closing","id":"l1dv6fk2"}
10
+ {"time":"2024-10-30T01:33:16.602019459-04:00","level":"INFO","msg":"Stopping system monitor"}
11
+ {"time":"2024-10-30T01:33:16.602687014-04:00","level":"INFO","msg":"Stopped system monitor"}
wandb/run-20241030_013141-l1dv6fk2/logs/debug.log ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-10-30 01:31:41,766 INFO MainThread:335753 [wandb_setup.py:_flush():79] Current SDK version is 0.18.5
2
+ 2024-10-30 01:31:41,766 INFO MainThread:335753 [wandb_setup.py:_flush():79] Configure stats pid to 335753
3
+ 2024-10-30 01:31:41,766 INFO MainThread:335753 [wandb_setup.py:_flush():79] Loading settings from /home/chunhui/.config/wandb/settings
4
+ 2024-10-30 01:31:41,767 INFO MainThread:335753 [wandb_setup.py:_flush():79] Loading settings from /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/settings
5
+ 2024-10-30 01:31:41,767 INFO MainThread:335753 [wandb_setup.py:_flush():79] Loading settings from environment variables: {}
6
+ 2024-10-30 01:31:41,767 INFO MainThread:335753 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': None, '_disable_service': None}
7
+ 2024-10-30 01:31:41,767 INFO MainThread:335753 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': 'train/train_deep_wandb.py', 'program_abspath': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py', 'program': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py'}
8
+ 2024-10-30 01:31:41,767 INFO MainThread:335753 [wandb_setup.py:_flush():79] Applying login settings: {}
9
+ 2024-10-30 01:31:41,767 INFO MainThread:335753 [wandb_init.py:_log_setup():534] Logging user logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241030_013141-l1dv6fk2/logs/debug.log
10
+ 2024-10-30 01:31:41,767 INFO MainThread:335753 [wandb_init.py:_log_setup():535] Logging internal logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241030_013141-l1dv6fk2/logs/debug-internal.log
11
+ 2024-10-30 01:31:41,767 INFO MainThread:335753 [wandb_init.py:init():621] calling init triggers
12
+ 2024-10-30 01:31:41,767 INFO MainThread:335753 [wandb_init.py:init():628] wandb.init called with sweep_config: {}
13
+ config: {}
14
+ 2024-10-30 01:31:41,767 INFO MainThread:335753 [wandb_init.py:init():671] starting backend
15
+ 2024-10-30 01:31:41,767 INFO MainThread:335753 [wandb_init.py:init():675] sending inform_init request
16
+ 2024-10-30 01:31:41,768 INFO MainThread:335753 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2024-10-30 01:31:41,768 INFO MainThread:335753 [wandb_init.py:init():688] backend started and connected
18
+ 2024-10-30 01:31:41,772 INFO MainThread:335753 [wandb_init.py:init():783] updated telemetry
19
+ 2024-10-30 01:31:41,821 INFO MainThread:335753 [wandb_init.py:init():816] communicating run to backend with 90.0 second timeout
20
+ 2024-10-30 01:31:42,070 INFO MainThread:335753 [wandb_init.py:init():867] starting run threads in backend
21
+ 2024-10-30 01:31:42,184 INFO MainThread:335753 [wandb_run.py:_console_start():2463] atexit reg
22
+ 2024-10-30 01:31:42,185 INFO MainThread:335753 [wandb_run.py:_redirect():2311] redirect: wrap_raw
23
+ 2024-10-30 01:31:42,185 INFO MainThread:335753 [wandb_run.py:_redirect():2376] Wrapping output streams.
24
+ 2024-10-30 01:31:42,185 INFO MainThread:335753 [wandb_run.py:_redirect():2401] Redirects installed.
25
+ 2024-10-30 01:31:42,186 INFO MainThread:335753 [wandb_init.py:init():911] run started, returning control to user process
26
+ 2024-10-30 01:31:42,186 INFO MainThread:335753 [wandb_run.py:_config_callback():1390] config_cb None None {'perturbation': 'reverse_full', 'train_set': '10M', 'batch_size': 3, 'epoch': 7, 'seed': 0}
27
+ 2024-10-30 01:33:16,602 WARNING MsgRouterThr:335753 [router.py:message_loop():77] message_loop has been closed
wandb/run-20241030_013141-l1dv6fk2/run-l1dv6fk2.wandb ADDED
Binary file (65.5 kB). View file
 
wandb/run-20241031_000839-xm9f2csf/files/output.log ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Loading checkpoint shards: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 2/2 [00:18<00:00, 9.33s/it]
2
+ tokenized_valid: Dataset({
3
+ features: ['input_ids', 'attention_mask'],
4
+ num_rows: 600
5
+ })
6
+ /mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/training_args.py:1545: FutureWarning: `evaluation_strategy` is deprecated and will be removed in version 4.46 of 🤗 Transformers. Use `eval_strategy` instead
7
+ warnings.warn(
8
+ [2024-10-31 00:09:00,201] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect)
9
+ [2024-10-31 00:09:10,050] [INFO] [comm.py:652:init_distributed] cdb=None
10
+ Installed CUDA version 11.8 does not match the version torch was compiled with 11.7 but since the APIs are compatible, accepting this combination
11
+ Using /home/chunhui/.cache/torch_extensions/py39_cu117 as PyTorch extensions root...
12
+ Emitting ninja build file /home/chunhui/.cache/torch_extensions/py39_cu117/cpu_adam/build.ninja...
13
+ Building extension module cpu_adam...
14
+ Allowing ninja to set a default number of workers... (overridable by setting the environment variable MAX_JOBS=N)
15
+ Loading extension module cpu_adam...
16
+ Time to load cpu_adam op: 5.461333274841309 seconds
wandb/run-20241031_000839-xm9f2csf/files/requirements.txt ADDED
@@ -0,0 +1,147 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ funcsigs==1.0.2
2
+ sentry-sdk==2.17.0
3
+ multiprocess==0.70.16
4
+ numpy==1.26.2
5
+ pluralizer==1.2.0
6
+ debugpy==1.6.7
7
+ nvidia-cudnn-cu11==8.5.0.96
8
+ deepspeed==0.15.2
9
+ data==0.4
10
+ pandas==2.1.3
11
+ tomli==2.0.1
12
+ charset-normalizer==3.3.2
13
+ attrs==24.2.0
14
+ aiosignal==1.3.1
15
+ fsspec==2023.10.0
16
+ nvidia-cusparse-cu11==11.7.4.91
17
+ zipp==3.12.0
18
+ mypy-extensions==1.0.0
19
+ datasets==3.0.1
20
+ joblib==1.3.2
21
+ hjson==3.1.0
22
+ traitlets==5.7.1
23
+ stack-data==0.6.0
24
+ transformers==4.45.1
25
+ sympy==1.11.1
26
+ Pygments==2.15.0
27
+ docker-pycreds==0.4.0
28
+ dill==0.3.8
29
+ wheel==0.44.0
30
+ prompt-toolkit==3.0.30
31
+ parso==0.8.3
32
+ ipykernel==6.23.1
33
+ pyarrow==17.0.0
34
+ certifi==2023.11.17
35
+ nvidia-cufft-cu11==10.9.0.58
36
+ six==1.16.0
37
+ pydantic==2.9.2
38
+ click==8.1.7
39
+ nest-asyncio==1.5.6
40
+ gmpy2==2.1.0
41
+ matplotlib==3.8.2
42
+ scipy==1.11.4
43
+ typing_extensions==4.12.2
44
+ statsmodels==0.14.0
45
+ huggingface-hub==0.25.0
46
+ frozenlist==1.4.1
47
+ gpustat==1.1.1
48
+ nvidia-nvtx-cu11==11.7.91
49
+ safetensors==0.4.5
50
+ stanza==1.9.2
51
+ decorator==5.1.1
52
+ seaborn==0.13.0
53
+ sentencepiece==0.2.0
54
+ PyYAML==6.0.1
55
+ black==24.8.0
56
+ protobuf==4.25.1
57
+ pickleshare==0.7.5
58
+ peft==0.13.0
59
+ triton==2.0.0
60
+ nvidia-cuda-runtime-cu11==11.7.99
61
+ Jinja2==3.1.2
62
+ nvidia-cusolver-cu11==11.4.0.1
63
+ executing==1.2.0
64
+ jupyter_client==8.1.0
65
+ pluggy==1.3.0
66
+ cmake==3.30.3
67
+ pytz==2023.3.post1
68
+ aiohappyeyeballs==2.4.2
69
+ kiwisolver==1.4.5
70
+ py-cpuinfo==9.0.0
71
+ Pillow==10.1.0
72
+ ptyprocess==0.7.0
73
+ importlib_resources==6.4.5
74
+ GitPython==3.1.43
75
+ importlib-metadata==6.0.0
76
+ iniconfig==2.0.0
77
+ scikit-learn==1.3.2
78
+ exceptiongroup==1.1.0
79
+ networkx==2.8.6
80
+ accelerate==1.0.0
81
+ nltk==3.8.1
82
+ shutilwhich==1.1.0
83
+ fonttools==4.45.1
84
+ future==0.18.3
85
+ aiohttp==3.10.6
86
+ wcwidth==0.2.5
87
+ idna==3.6
88
+ filelock==3.12.2
89
+ pathspec==0.12.1
90
+ jupyter_core==5.1.0
91
+ lit==18.1.8
92
+ nvidia-curand-cu11==10.2.10.91
93
+ nvidia-cublas-cu11==11.10.3.66
94
+ nvidia-ml-py==12.560.30
95
+ msgpack==1.1.0
96
+ python-dateutil==2.8.2
97
+ blessed==1.20.0
98
+ packaging==23.0
99
+ gitdb==4.0.11
100
+ yarl==1.13.0
101
+ emoji==2.8.0
102
+ tzdata==2023.3
103
+ cycler==0.12.1
104
+ tornado==6.2
105
+ backcall==0.2.0
106
+ plotnine==0.12.4
107
+ ninja==1.11.1.1
108
+ latex==0.7.0
109
+ wandb==0.18.5
110
+ setproctitle==1.3.3
111
+ threadpoolctl==3.2.0
112
+ requests==2.32.3
113
+ pyparsing==3.1.1
114
+ smmap==5.0.1
115
+ pyzmq==23.0.0
116
+ async-timeout==4.0.3
117
+ annotated-types==0.7.0
118
+ matplotlib-inline==0.1.6
119
+ latexcodec==1.0.0
120
+ ipython==8.0.0
121
+ patsy==0.5.3
122
+ contourpy==1.2.0
123
+ multidict==6.1.0
124
+ mizani==0.9.3
125
+ urllib3==2.1.0
126
+ tokenizers==0.20.0
127
+ MarkupSafe==2.1.2
128
+ pip==24.2
129
+ pexpect==4.8.0
130
+ tqdm==4.66.5
131
+ jedi==0.18.2
132
+ pydantic_core==2.23.4
133
+ tempdir==0.7.1
134
+ mpmath==1.2.1
135
+ setuptools==72.1.0
136
+ pytest==7.4.3
137
+ pure-eval==0.2.2
138
+ psutil==5.9.1
139
+ comm==0.1.2
140
+ nvidia-cuda-cupti-cu11==11.7.101
141
+ nvidia-cuda-nvrtc-cu11==11.7.99
142
+ regex==2023.10.3
143
+ platformdirs==2.5.2
144
+ asttokens==2.2.1
145
+ torch==2.0.0
146
+ nvidia-nccl-cu11==2.14.3
147
+ xxhash==3.5.0
wandb/run-20241031_000839-xm9f2csf/files/wandb-metadata.json ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-162-generic-x86_64-with-glibc2.31",
3
+ "python": "3.9.19",
4
+ "startedAt": "2024-10-31T04:08:39.128145Z",
5
+ "args": [
6
+ "--perturbation",
7
+ "reverse_full",
8
+ "--train_set",
9
+ "10M",
10
+ "--batch_size",
11
+ "3",
12
+ "--epoch",
13
+ "6",
14
+ "--seed",
15
+ "0"
16
+ ],
17
+ "program": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py",
18
+ "codePath": "train/train_deep_wandb.py",
19
+ "git": {
20
+ "remote": "git@hf.co:Yaning1001/Impossible_llm.git",
21
+ "commit": "ed716cdcfcdea02b67f7ed0f3504c2b1c8b737c4"
22
+ },
23
+ "email": "yaning1001@gmail.com",
24
+ "root": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train",
25
+ "host": "mms-large-2",
26
+ "username": "chunhui",
27
+ "executable": "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/bin/python",
28
+ "codePathLocal": "train_deep_wandb.py",
29
+ "cpu_count": 32,
30
+ "cpu_count_logical": 64,
31
+ "gpu": "NVIDIA RTX A6000",
32
+ "gpu_count": 8,
33
+ "disk": {
34
+ "/": {
35
+ "total": "1888559353856",
36
+ "used": "1727270539264"
37
+ }
38
+ },
39
+ "memory": {
40
+ "total": "202617098240"
41
+ },
42
+ "cpu": {
43
+ "count": 32,
44
+ "countLogical": 64
45
+ },
46
+ "gpu_nvidia": [
47
+ {
48
+ "name": "NVIDIA RTX A6000",
49
+ "memoryTotal": "51527024640",
50
+ "cudaCores": 10752,
51
+ "architecture": "Ampere"
52
+ },
53
+ {
54
+ "name": "NVIDIA RTX A6000",
55
+ "memoryTotal": "51527024640",
56
+ "cudaCores": 10752,
57
+ "architecture": "Ampere"
58
+ },
59
+ {
60
+ "name": "NVIDIA RTX A6000",
61
+ "memoryTotal": "51527024640",
62
+ "cudaCores": 10752,
63
+ "architecture": "Ampere"
64
+ },
65
+ {
66
+ "name": "NVIDIA RTX A6000",
67
+ "memoryTotal": "51527024640",
68
+ "cudaCores": 10752,
69
+ "architecture": "Ampere"
70
+ },
71
+ {
72
+ "name": "NVIDIA RTX A6000",
73
+ "memoryTotal": "51527024640",
74
+ "cudaCores": 10752,
75
+ "architecture": "Ampere"
76
+ },
77
+ {
78
+ "name": "NVIDIA RTX A6000",
79
+ "memoryTotal": "51527024640",
80
+ "cudaCores": 10752,
81
+ "architecture": "Ampere"
82
+ },
83
+ {
84
+ "name": "NVIDIA RTX A6000",
85
+ "memoryTotal": "51527024640",
86
+ "cudaCores": 10752,
87
+ "architecture": "Ampere"
88
+ },
89
+ {
90
+ "name": "NVIDIA RTX A6000",
91
+ "memoryTotal": "51527024640",
92
+ "cudaCores": 10752,
93
+ "architecture": "Ampere"
94
+ }
95
+ ],
96
+ "cudaVersion": "11.8"
97
+ }
wandb/run-20241031_000839-xm9f2csf/logs/debug-internal.log ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {"time":"2024-10-31T00:08:39.130243946-04:00","level":"INFO","msg":"using version","core version":"0.18.5"}
2
+ {"time":"2024-10-31T00:08:39.130254286-04:00","level":"INFO","msg":"created symlink","path":"/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241031_000839-xm9f2csf/logs/debug-core.log"}
3
+ {"time":"2024-10-31T00:08:39.236787826-04:00","level":"INFO","msg":"created new stream","id":"xm9f2csf"}
4
+ {"time":"2024-10-31T00:08:39.236865977-04:00","level":"INFO","msg":"stream: started","id":"xm9f2csf"}
5
+ {"time":"2024-10-31T00:08:39.236982138-04:00","level":"INFO","msg":"sender: started","stream_id":"xm9f2csf"}
6
+ {"time":"2024-10-31T00:08:39.236945847-04:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"xm9f2csf"}}
7
+ {"time":"2024-10-31T00:08:39.236987988-04:00","level":"INFO","msg":"handler: started","stream_id":{"value":"xm9f2csf"}}
8
+ {"time":"2024-10-31T00:08:39.449935127-04:00","level":"INFO","msg":"Starting system monitor"}
wandb/run-20241031_000839-xm9f2csf/logs/debug.log ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-10-31 00:08:39,126 INFO MainThread:477297 [wandb_setup.py:_flush():79] Current SDK version is 0.18.5
2
+ 2024-10-31 00:08:39,126 INFO MainThread:477297 [wandb_setup.py:_flush():79] Configure stats pid to 477297
3
+ 2024-10-31 00:08:39,126 INFO MainThread:477297 [wandb_setup.py:_flush():79] Loading settings from /home/chunhui/.config/wandb/settings
4
+ 2024-10-31 00:08:39,126 INFO MainThread:477297 [wandb_setup.py:_flush():79] Loading settings from /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/settings
5
+ 2024-10-31 00:08:39,126 INFO MainThread:477297 [wandb_setup.py:_flush():79] Loading settings from environment variables: {}
6
+ 2024-10-31 00:08:39,126 INFO MainThread:477297 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': None, '_disable_service': None}
7
+ 2024-10-31 00:08:39,126 INFO MainThread:477297 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': 'train/train_deep_wandb.py', 'program_abspath': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py', 'program': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py'}
8
+ 2024-10-31 00:08:39,126 INFO MainThread:477297 [wandb_setup.py:_flush():79] Applying login settings: {}
9
+ 2024-10-31 00:08:39,126 INFO MainThread:477297 [wandb_init.py:_log_setup():534] Logging user logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241031_000839-xm9f2csf/logs/debug.log
10
+ 2024-10-31 00:08:39,126 INFO MainThread:477297 [wandb_init.py:_log_setup():535] Logging internal logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241031_000839-xm9f2csf/logs/debug-internal.log
11
+ 2024-10-31 00:08:39,126 INFO MainThread:477297 [wandb_init.py:init():621] calling init triggers
12
+ 2024-10-31 00:08:39,126 INFO MainThread:477297 [wandb_init.py:init():628] wandb.init called with sweep_config: {}
13
+ config: {}
14
+ 2024-10-31 00:08:39,126 INFO MainThread:477297 [wandb_init.py:init():671] starting backend
15
+ 2024-10-31 00:08:39,126 INFO MainThread:477297 [wandb_init.py:init():675] sending inform_init request
16
+ 2024-10-31 00:08:39,127 INFO MainThread:477297 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2024-10-31 00:08:39,127 INFO MainThread:477297 [wandb_init.py:init():688] backend started and connected
18
+ 2024-10-31 00:08:39,131 INFO MainThread:477297 [wandb_init.py:init():783] updated telemetry
19
+ 2024-10-31 00:08:39,153 INFO MainThread:477297 [wandb_init.py:init():816] communicating run to backend with 90.0 second timeout
20
+ 2024-10-31 00:08:39,447 INFO MainThread:477297 [wandb_init.py:init():867] starting run threads in backend
21
+ 2024-10-31 00:08:39,540 INFO MainThread:477297 [wandb_run.py:_console_start():2463] atexit reg
22
+ 2024-10-31 00:08:39,541 INFO MainThread:477297 [wandb_run.py:_redirect():2311] redirect: wrap_raw
23
+ 2024-10-31 00:08:39,541 INFO MainThread:477297 [wandb_run.py:_redirect():2376] Wrapping output streams.
24
+ 2024-10-31 00:08:39,541 INFO MainThread:477297 [wandb_run.py:_redirect():2401] Redirects installed.
25
+ 2024-10-31 00:08:39,543 INFO MainThread:477297 [wandb_init.py:init():911] run started, returning control to user process
26
+ 2024-10-31 00:08:39,543 INFO MainThread:477297 [wandb_run.py:_config_callback():1390] config_cb None None {'perturbation': 'reverse_full', 'train_set': '10M', 'batch_size': 3, 'epoch': 6, 'seed': 0, 'lr': 1e-05}
wandb/run-20241031_000839-xm9f2csf/run-xm9f2csf.wandb ADDED
Binary file (65.5 kB). View file
 
wandb/run-20241031_122005-oby2afce/files/config.yaml ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.18.5
4
+ m: []
5
+ python_version: 3.9.19
6
+ t:
7
+ "1":
8
+ - 1
9
+ - 5
10
+ - 11
11
+ - 49
12
+ - 51
13
+ - 53
14
+ - 55
15
+ - 71
16
+ - 98
17
+ "2":
18
+ - 1
19
+ - 5
20
+ - 11
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 98
27
+ "3":
28
+ - 13
29
+ - 23
30
+ - 55
31
+ "4": 3.9.19
32
+ "5": 0.18.5
33
+ "6": 4.45.1
34
+ "8":
35
+ - 5
36
+ "12": 0.18.5
37
+ "13": linux-x86_64
38
+ batch_size:
39
+ value: 3
40
+ epoch:
41
+ value: 6
42
+ lr:
43
+ value: 5e-06
44
+ perturbation:
45
+ value: reverse_full
46
+ seed:
47
+ value: 0
48
+ train_set:
49
+ value: 10M
wandb/run-20241031_122005-oby2afce/files/output.log ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Downloading shards: 0%| | 0/2 [00:22<?, ?it/s]Exception ignored in: <generator object tqdm.__iter__ at 0x7f449c078d60>
2
+ Traceback (most recent call last):
3
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/tqdm/std.py", line 1196, in __iter__
4
+ self.close()
5
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/tqdm/std.py", line 1303, in close
6
+ fp_write('\n')
7
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/tqdm/std.py", line 1287, in fp_write
8
+ self.fp.write(str(s))
9
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/tqdm/utils.py", line 196, in inner
10
+ return func(*args, **kwargs)
11
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/wandb/sdk/lib/redirect.py", line 648, in write
12
+ cb(data)
13
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/wandb/sdk/wandb_run.py", line 2386, in <lambda>
14
+ lambda data: self._console_raw_callback("stderr", data),
15
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/wandb/sdk/wandb_run.py", line 403, in wrapper_fn
16
+ return func(self, *args, **kwargs)
17
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/wandb/sdk/wandb_run.py", line 1547, in _console_raw_callback
18
+ self._backend.interface.publish_output_raw(name, data)
19
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/wandb/sdk/interface/interface.py", line 715, in publish_output_raw
20
+ otype = pb.OutputRawRecord.OutputType.STDERR
21
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/google/protobuf/internal/enum_type_wrapper.py", line 94, in __getattr__
22
+ try:
23
+ KeyboardInterrupt:
24
+ Traceback (most recent call last):
25
+ File "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py", line 173, in <module>
26
+ model = AutoModelForCausalLM.from_pretrained(model_name,
27
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/models/auto/auto_factory.py", line 564, in from_pretrained
28
+ return model_class.from_pretrained(
29
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/modeling_utils.py", line 3769, in from_pretrained
30
+ resolved_archive_file, sharded_metadata = get_checkpoint_shard_files(
31
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/utils/hub.py", line 1098, in get_checkpoint_shard_files
32
+ cached_filename = cached_file(
33
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/utils/hub.py", line 403, in cached_file
34
+ resolved_file = hf_hub_download(
35
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/utils/_deprecation.py", line 101, in inner_f
36
+ return f(*args, **kwargs)
37
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn
38
+ return fn(*args, **kwargs)
39
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/file_download.py", line 1232, in hf_hub_download
40
+ return _hf_hub_download_to_cache_dir(
41
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/file_download.py", line 1380, in _hf_hub_download_to_cache_dir
42
+ with WeakFileLock(lock_path):
43
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/contextlib.py", line 119, in __enter__
44
+ return next(self.gen)
45
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/utils/_fixes.py", line 98, in WeakFileLock
46
+ lock.acquire()
47
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/filelock/_api.py", line 225, in acquire
48
+ time.sleep(poll_interval)
49
+ KeyboardInterrupt
wandb/run-20241031_122005-oby2afce/files/wandb-metadata.json ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-162-generic-x86_64-with-glibc2.31",
3
+ "python": "3.9.19",
4
+ "startedAt": "2024-10-31T16:20:05.857918Z",
5
+ "args": [
6
+ "--perturbation",
7
+ "reverse_full",
8
+ "--train_set",
9
+ "10M",
10
+ "--batch_size",
11
+ "3",
12
+ "--epoch",
13
+ "6",
14
+ "--seed",
15
+ "0"
16
+ ],
17
+ "program": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py",
18
+ "codePath": "train/train_deep_wandb.py",
19
+ "git": {
20
+ "remote": "git@hf.co:Yaning1001/Impossible_llm.git",
21
+ "commit": "ed716cdcfcdea02b67f7ed0f3504c2b1c8b737c4"
22
+ },
23
+ "email": "yaning1001@gmail.com",
24
+ "root": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train",
25
+ "host": "mms-large-2",
26
+ "username": "chunhui",
27
+ "executable": "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/bin/python",
28
+ "codePathLocal": "train_deep_wandb.py",
29
+ "cpu_count": 32,
30
+ "cpu_count_logical": 64,
31
+ "gpu": "NVIDIA RTX A6000",
32
+ "gpu_count": 8,
33
+ "disk": {
34
+ "/": {
35
+ "total": "1888559353856",
36
+ "used": "1753159847936"
37
+ }
38
+ },
39
+ "memory": {
40
+ "total": "202617098240"
41
+ },
42
+ "cpu": {
43
+ "count": 32,
44
+ "countLogical": 64
45
+ },
46
+ "gpu_nvidia": [
47
+ {
48
+ "name": "NVIDIA RTX A6000",
49
+ "memoryTotal": "51527024640",
50
+ "cudaCores": 10752,
51
+ "architecture": "Ampere"
52
+ },
53
+ {
54
+ "name": "NVIDIA RTX A6000",
55
+ "memoryTotal": "51527024640",
56
+ "cudaCores": 10752,
57
+ "architecture": "Ampere"
58
+ },
59
+ {
60
+ "name": "NVIDIA RTX A6000",
61
+ "memoryTotal": "51527024640",
62
+ "cudaCores": 10752,
63
+ "architecture": "Ampere"
64
+ },
65
+ {
66
+ "name": "NVIDIA RTX A6000",
67
+ "memoryTotal": "51527024640",
68
+ "cudaCores": 10752,
69
+ "architecture": "Ampere"
70
+ },
71
+ {
72
+ "name": "NVIDIA RTX A6000",
73
+ "memoryTotal": "51527024640",
74
+ "cudaCores": 10752,
75
+ "architecture": "Ampere"
76
+ },
77
+ {
78
+ "name": "NVIDIA RTX A6000",
79
+ "memoryTotal": "51527024640",
80
+ "cudaCores": 10752,
81
+ "architecture": "Ampere"
82
+ },
83
+ {
84
+ "name": "NVIDIA RTX A6000",
85
+ "memoryTotal": "51527024640",
86
+ "cudaCores": 10752,
87
+ "architecture": "Ampere"
88
+ },
89
+ {
90
+ "name": "NVIDIA RTX A6000",
91
+ "memoryTotal": "51527024640",
92
+ "cudaCores": 10752,
93
+ "architecture": "Ampere"
94
+ }
95
+ ],
96
+ "cudaVersion": "11.8"
97
+ }
wandb/run-20241031_122005-oby2afce/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_wandb":{"runtime":23}}
wandb/run-20241031_122005-oby2afce/logs/debug-internal.log ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2024-10-31T12:20:05.859652319-04:00","level":"INFO","msg":"using version","core version":"0.18.5"}
2
+ {"time":"2024-10-31T12:20:05.859662679-04:00","level":"INFO","msg":"created symlink","path":"/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241031_122005-oby2afce/logs/debug-core.log"}
3
+ {"time":"2024-10-31T12:20:05.96611756-04:00","level":"INFO","msg":"created new stream","id":"oby2afce"}
4
+ {"time":"2024-10-31T12:20:05.96615247-04:00","level":"INFO","msg":"stream: started","id":"oby2afce"}
5
+ {"time":"2024-10-31T12:20:05.96618829-04:00","level":"INFO","msg":"sender: started","stream_id":"oby2afce"}
6
+ {"time":"2024-10-31T12:20:05.96616822-04:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"oby2afce"}}
7
+ {"time":"2024-10-31T12:20:05.96618968-04:00","level":"INFO","msg":"handler: started","stream_id":{"value":"oby2afce"}}
8
+ {"time":"2024-10-31T12:20:06.184419076-04:00","level":"INFO","msg":"Starting system monitor"}
9
+ {"time":"2024-10-31T12:20:29.34968594-04:00","level":"INFO","msg":"stream: closing","id":"oby2afce"}
10
+ {"time":"2024-10-31T12:20:29.34972926-04:00","level":"INFO","msg":"Stopping system monitor"}
11
+ {"time":"2024-10-31T12:20:29.35043694-04:00","level":"INFO","msg":"Stopped system monitor"}
wandb/run-20241031_122005-oby2afce/logs/debug.log ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-10-31 12:20:05,856 INFO MainThread:557181 [wandb_setup.py:_flush():79] Current SDK version is 0.18.5
2
+ 2024-10-31 12:20:05,856 INFO MainThread:557181 [wandb_setup.py:_flush():79] Configure stats pid to 557181
3
+ 2024-10-31 12:20:05,856 INFO MainThread:557181 [wandb_setup.py:_flush():79] Loading settings from /home/chunhui/.config/wandb/settings
4
+ 2024-10-31 12:20:05,856 INFO MainThread:557181 [wandb_setup.py:_flush():79] Loading settings from /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/settings
5
+ 2024-10-31 12:20:05,856 INFO MainThread:557181 [wandb_setup.py:_flush():79] Loading settings from environment variables: {}
6
+ 2024-10-31 12:20:05,856 INFO MainThread:557181 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': None, '_disable_service': None}
7
+ 2024-10-31 12:20:05,856 INFO MainThread:557181 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': 'train/train_deep_wandb.py', 'program_abspath': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py', 'program': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py'}
8
+ 2024-10-31 12:20:05,856 INFO MainThread:557181 [wandb_setup.py:_flush():79] Applying login settings: {}
9
+ 2024-10-31 12:20:05,856 INFO MainThread:557181 [wandb_init.py:_log_setup():534] Logging user logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241031_122005-oby2afce/logs/debug.log
10
+ 2024-10-31 12:20:05,856 INFO MainThread:557181 [wandb_init.py:_log_setup():535] Logging internal logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241031_122005-oby2afce/logs/debug-internal.log
11
+ 2024-10-31 12:20:05,856 INFO MainThread:557181 [wandb_init.py:init():621] calling init triggers
12
+ 2024-10-31 12:20:05,856 INFO MainThread:557181 [wandb_init.py:init():628] wandb.init called with sweep_config: {}
13
+ config: {}
14
+ 2024-10-31 12:20:05,856 INFO MainThread:557181 [wandb_init.py:init():671] starting backend
15
+ 2024-10-31 12:20:05,856 INFO MainThread:557181 [wandb_init.py:init():675] sending inform_init request
16
+ 2024-10-31 12:20:05,857 INFO MainThread:557181 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2024-10-31 12:20:05,857 INFO MainThread:557181 [wandb_init.py:init():688] backend started and connected
18
+ 2024-10-31 12:20:05,860 INFO MainThread:557181 [wandb_init.py:init():783] updated telemetry
19
+ 2024-10-31 12:20:05,891 INFO MainThread:557181 [wandb_init.py:init():816] communicating run to backend with 90.0 second timeout
20
+ 2024-10-31 12:20:06,181 INFO MainThread:557181 [wandb_init.py:init():867] starting run threads in backend
21
+ 2024-10-31 12:20:06,275 INFO MainThread:557181 [wandb_run.py:_console_start():2463] atexit reg
22
+ 2024-10-31 12:20:06,276 INFO MainThread:557181 [wandb_run.py:_redirect():2311] redirect: wrap_raw
23
+ 2024-10-31 12:20:06,276 INFO MainThread:557181 [wandb_run.py:_redirect():2376] Wrapping output streams.
24
+ 2024-10-31 12:20:06,276 INFO MainThread:557181 [wandb_run.py:_redirect():2401] Redirects installed.
25
+ 2024-10-31 12:20:06,277 INFO MainThread:557181 [wandb_init.py:init():911] run started, returning control to user process
26
+ 2024-10-31 12:20:06,278 INFO MainThread:557181 [wandb_run.py:_config_callback():1390] config_cb None None {'perturbation': 'reverse_full', 'train_set': '10M', 'batch_size': 3, 'epoch': 6, 'seed': 0, 'lr': 5e-06}
27
+ 2024-10-31 12:20:29,349 WARNING MsgRouterThr:557181 [router.py:message_loop():77] message_loop has been closed
wandb/run-20241031_122005-oby2afce/run-oby2afce.wandb ADDED
File without changes
wandb/run-20241101_012438-pm66le56/files/config.yaml ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.18.5
4
+ m: []
5
+ python_version: 3.9.19
6
+ t:
7
+ "1":
8
+ - 1
9
+ - 5
10
+ - 11
11
+ - 49
12
+ - 51
13
+ - 53
14
+ - 55
15
+ - 71
16
+ - 98
17
+ "2":
18
+ - 1
19
+ - 5
20
+ - 11
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 98
27
+ "3":
28
+ - 13
29
+ - 23
30
+ - 55
31
+ "4": 3.9.19
32
+ "5": 0.18.5
33
+ "6": 4.45.1
34
+ "8":
35
+ - 5
36
+ "12": 0.18.5
37
+ "13": linux-x86_64
38
+ batch_size:
39
+ value: 3
40
+ epoch:
41
+ value: 6
42
+ lr:
43
+ value: 5e-06
44
+ perturbation:
45
+ value: shuffle_nodeterministic
46
+ seed:
47
+ value: 0
48
+ train_set:
49
+ value: 10M
wandb/run-20241101_012438-pm66le56/files/output.log ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Traceback (most recent call last):
2
+ File "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py", line 164, in <module>
3
+ dataset = load_dataset('babylm_dataset_test.py', name=dataset_name, trust_remote_code=True)
4
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/datasets/load.py", line 2074, in load_dataset
5
+ builder_instance = load_dataset_builder(
6
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/datasets/load.py", line 1832, in load_dataset_builder
7
+ builder_instance: DatasetBuilder = builder_cls(
8
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/datasets/builder.py", line 342, in __init__
9
+ self.config, self.config_id = self._create_builder_config(
10
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/datasets/builder.py", line 569, in _create_builder_config
11
+ raise ValueError(
12
+ ValueError: BuilderConfig 'babylm_shuffle_nodeterministic_10M_seed0' not found. Available: ['babylm_hop_control_10M_seed0', 'babylm_hop_tokens4_10M_seed0', 'babylm_hop_words4_10M_seed0', 'babylm_reverse_control_10M_seed0', 'babylm_reverse_partial_10M_seed0', 'babylm_reverse_full_10M_seed0', 'babylm_shuffle_control_10M_seed0', 'babylm_shuffle_nondeterministic_10M_seed0', 'babylm_shuffle_deterministic21_10M_seed0', 'babylm_shuffle_deterministic57_10M_seed0', 'babylm_shuffle_deterministic84_10M_seed0', 'babylm_shuffle_local3_10M_seed0', 'babylm_shuffle_local5_10M_seed0', 'babylm_shuffle_local10_10M_seed0', 'babylm_shuffle_even_odd_10M_seed0']
wandb/run-20241101_012438-pm66le56/files/wandb-metadata.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-162-generic-x86_64-with-glibc2.31",
3
+ "python": "3.9.19",
4
+ "startedAt": "2024-11-01T05:24:38.162026Z",
5
+ "args": [
6
+ "--perturbation",
7
+ "shuffle_nodeterministic",
8
+ "--train_set",
9
+ "10M",
10
+ "--batch_size",
11
+ "3",
12
+ "--epoch",
13
+ "6",
14
+ "--seed",
15
+ "0"
16
+ ],
17
+ "program": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py",
18
+ "codePath": "train/train_deep_wandb.py",
19
+ "git": {
20
+ "remote": "git@hf.co:Yaning1001/Impossible_llm.git",
21
+ "commit": "ed716cdcfcdea02b67f7ed0f3504c2b1c8b737c4"
22
+ },
23
+ "email": "yaning1001@gmail.com",
24
+ "root": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train",
25
+ "host": "mms-large-2",
26
+ "username": "chunhui",
27
+ "executable": "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/bin/python",
28
+ "codePathLocal": "train_deep_wandb.py"
29
+ }
wandb/run-20241101_012438-pm66le56/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_wandb":{"runtime":0}}
wandb/run-20241101_012438-pm66le56/logs/debug-internal.log ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2024-11-01T01:24:38.163852936-04:00","level":"INFO","msg":"using version","core version":"0.18.5"}
2
+ {"time":"2024-11-01T01:24:38.163865796-04:00","level":"INFO","msg":"created symlink","path":"/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241101_012438-pm66le56/logs/debug-core.log"}
3
+ {"time":"2024-11-01T01:24:38.271438014-04:00","level":"INFO","msg":"created new stream","id":"pm66le56"}
4
+ {"time":"2024-11-01T01:24:38.271465954-04:00","level":"INFO","msg":"stream: started","id":"pm66le56"}
5
+ {"time":"2024-11-01T01:24:38.271510694-04:00","level":"INFO","msg":"sender: started","stream_id":"pm66le56"}
6
+ {"time":"2024-11-01T01:24:38.271504754-04:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"pm66le56"}}
7
+ {"time":"2024-11-01T01:24:38.271523085-04:00","level":"INFO","msg":"handler: started","stream_id":{"value":"pm66le56"}}
8
+ {"time":"2024-11-01T01:24:38.47207326-04:00","level":"INFO","msg":"Starting system monitor"}
9
+ {"time":"2024-11-01T01:24:38.640865874-04:00","level":"INFO","msg":"stream: closing","id":"pm66le56"}
10
+ {"time":"2024-11-01T01:24:38.640889844-04:00","level":"INFO","msg":"Stopping system monitor"}
11
+ {"time":"2024-11-01T01:24:38.645833748-04:00","level":"INFO","msg":"Stopped system monitor"}
12
+ {"time":"2024-11-01T01:24:39.198611434-04:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
13
+ {"time":"2024-11-01T01:24:39.326514964-04:00","level":"INFO","msg":"handler: closed","stream_id":{"value":"pm66le56"}}
14
+ {"time":"2024-11-01T01:24:39.326560314-04:00","level":"INFO","msg":"sender: closed","stream_id":"pm66le56"}
15
+ {"time":"2024-11-01T01:24:39.326542564-04:00","level":"INFO","msg":"writer: Close: closed","stream_id":{"value":"pm66le56"}}
16
+ {"time":"2024-11-01T01:24:39.326608975-04:00","level":"INFO","msg":"stream: closed","id":"pm66le56"}
wandb/run-20241101_012438-pm66le56/logs/debug.log ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-11-01 01:24:38,159 INFO MainThread:676352 [wandb_setup.py:_flush():79] Current SDK version is 0.18.5
2
+ 2024-11-01 01:24:38,159 INFO MainThread:676352 [wandb_setup.py:_flush():79] Configure stats pid to 676352
3
+ 2024-11-01 01:24:38,159 INFO MainThread:676352 [wandb_setup.py:_flush():79] Loading settings from /home/chunhui/.config/wandb/settings
4
+ 2024-11-01 01:24:38,160 INFO MainThread:676352 [wandb_setup.py:_flush():79] Loading settings from /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/settings
5
+ 2024-11-01 01:24:38,160 INFO MainThread:676352 [wandb_setup.py:_flush():79] Loading settings from environment variables: {}
6
+ 2024-11-01 01:24:38,160 INFO MainThread:676352 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': None, '_disable_service': None}
7
+ 2024-11-01 01:24:38,160 INFO MainThread:676352 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': 'train/train_deep_wandb.py', 'program_abspath': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py', 'program': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py'}
8
+ 2024-11-01 01:24:38,160 INFO MainThread:676352 [wandb_setup.py:_flush():79] Applying login settings: {}
9
+ 2024-11-01 01:24:38,160 INFO MainThread:676352 [wandb_init.py:_log_setup():534] Logging user logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241101_012438-pm66le56/logs/debug.log
10
+ 2024-11-01 01:24:38,160 INFO MainThread:676352 [wandb_init.py:_log_setup():535] Logging internal logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241101_012438-pm66le56/logs/debug-internal.log
11
+ 2024-11-01 01:24:38,160 INFO MainThread:676352 [wandb_init.py:init():621] calling init triggers
12
+ 2024-11-01 01:24:38,160 INFO MainThread:676352 [wandb_init.py:init():628] wandb.init called with sweep_config: {}
13
+ config: {}
14
+ 2024-11-01 01:24:38,160 INFO MainThread:676352 [wandb_init.py:init():671] starting backend
15
+ 2024-11-01 01:24:38,160 INFO MainThread:676352 [wandb_init.py:init():675] sending inform_init request
16
+ 2024-11-01 01:24:38,161 INFO MainThread:676352 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2024-11-01 01:24:38,161 INFO MainThread:676352 [wandb_init.py:init():688] backend started and connected
18
+ 2024-11-01 01:24:38,164 INFO MainThread:676352 [wandb_init.py:init():783] updated telemetry
19
+ 2024-11-01 01:24:38,198 INFO MainThread:676352 [wandb_init.py:init():816] communicating run to backend with 90.0 second timeout
20
+ 2024-11-01 01:24:38,468 INFO MainThread:676352 [wandb_init.py:init():867] starting run threads in backend
21
+ 2024-11-01 01:24:38,576 INFO MainThread:676352 [wandb_run.py:_console_start():2463] atexit reg
22
+ 2024-11-01 01:24:38,577 INFO MainThread:676352 [wandb_run.py:_redirect():2311] redirect: wrap_raw
23
+ 2024-11-01 01:24:38,577 INFO MainThread:676352 [wandb_run.py:_redirect():2376] Wrapping output streams.
24
+ 2024-11-01 01:24:38,577 INFO MainThread:676352 [wandb_run.py:_redirect():2401] Redirects installed.
25
+ 2024-11-01 01:24:38,579 INFO MainThread:676352 [wandb_init.py:init():911] run started, returning control to user process
26
+ 2024-11-01 01:24:38,579 INFO MainThread:676352 [wandb_run.py:_config_callback():1390] config_cb None None {'perturbation': 'shuffle_nodeterministic', 'train_set': '10M', 'batch_size': 3, 'epoch': 6, 'seed': 0, 'lr': 5e-06}
27
+ 2024-11-01 01:24:38,640 WARNING MsgRouterThr:676352 [router.py:message_loop():77] message_loop has been closed
wandb/run-20241101_012438-pm66le56/run-pm66le56.wandb ADDED
Binary file (3.4 kB). View file
 
wandb/run-20241101_200502-ydj2ohih/files/output.log ADDED
@@ -0,0 +1 @@
 
 
1
+ Loading checkpoint shards: 0%| | 0/2 [00:00<?, ?it/s]
wandb/run-20241101_200502-ydj2ohih/files/requirements.txt ADDED
@@ -0,0 +1,147 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ funcsigs==1.0.2
2
+ sentry-sdk==2.17.0
3
+ multiprocess==0.70.16
4
+ numpy==1.26.2
5
+ pluralizer==1.2.0
6
+ debugpy==1.6.7
7
+ nvidia-cudnn-cu11==8.5.0.96
8
+ deepspeed==0.15.2
9
+ data==0.4
10
+ pandas==2.1.3
11
+ tomli==2.0.1
12
+ charset-normalizer==3.3.2
13
+ attrs==24.2.0
14
+ aiosignal==1.3.1
15
+ fsspec==2023.10.0
16
+ nvidia-cusparse-cu11==11.7.4.91
17
+ zipp==3.12.0
18
+ mypy-extensions==1.0.0
19
+ datasets==3.0.1
20
+ joblib==1.3.2
21
+ hjson==3.1.0
22
+ traitlets==5.7.1
23
+ stack-data==0.6.0
24
+ transformers==4.45.1
25
+ sympy==1.11.1
26
+ Pygments==2.15.0
27
+ docker-pycreds==0.4.0
28
+ dill==0.3.8
29
+ wheel==0.44.0
30
+ prompt-toolkit==3.0.30
31
+ parso==0.8.3
32
+ ipykernel==6.23.1
33
+ pyarrow==17.0.0
34
+ certifi==2023.11.17
35
+ nvidia-cufft-cu11==10.9.0.58
36
+ six==1.16.0
37
+ pydantic==2.9.2
38
+ click==8.1.7
39
+ nest-asyncio==1.5.6
40
+ gmpy2==2.1.0
41
+ matplotlib==3.8.2
42
+ scipy==1.11.4
43
+ typing_extensions==4.12.2
44
+ statsmodels==0.14.0
45
+ huggingface-hub==0.25.0
46
+ frozenlist==1.4.1
47
+ gpustat==1.1.1
48
+ nvidia-nvtx-cu11==11.7.91
49
+ safetensors==0.4.5
50
+ stanza==1.9.2
51
+ decorator==5.1.1
52
+ seaborn==0.13.0
53
+ sentencepiece==0.2.0
54
+ PyYAML==6.0.1
55
+ black==24.8.0
56
+ protobuf==4.25.1
57
+ pickleshare==0.7.5
58
+ peft==0.13.0
59
+ triton==2.0.0
60
+ nvidia-cuda-runtime-cu11==11.7.99
61
+ Jinja2==3.1.2
62
+ nvidia-cusolver-cu11==11.4.0.1
63
+ executing==1.2.0
64
+ jupyter_client==8.1.0
65
+ pluggy==1.3.0
66
+ cmake==3.30.3
67
+ pytz==2023.3.post1
68
+ aiohappyeyeballs==2.4.2
69
+ kiwisolver==1.4.5
70
+ py-cpuinfo==9.0.0
71
+ Pillow==10.1.0
72
+ ptyprocess==0.7.0
73
+ importlib_resources==6.4.5
74
+ GitPython==3.1.43
75
+ importlib-metadata==6.0.0
76
+ iniconfig==2.0.0
77
+ scikit-learn==1.3.2
78
+ exceptiongroup==1.1.0
79
+ networkx==2.8.6
80
+ accelerate==1.0.0
81
+ nltk==3.8.1
82
+ shutilwhich==1.1.0
83
+ fonttools==4.45.1
84
+ future==0.18.3
85
+ aiohttp==3.10.6
86
+ wcwidth==0.2.5
87
+ idna==3.6
88
+ filelock==3.12.2
89
+ pathspec==0.12.1
90
+ jupyter_core==5.1.0
91
+ lit==18.1.8
92
+ nvidia-curand-cu11==10.2.10.91
93
+ nvidia-cublas-cu11==11.10.3.66
94
+ nvidia-ml-py==12.560.30
95
+ msgpack==1.1.0
96
+ python-dateutil==2.8.2
97
+ blessed==1.20.0
98
+ packaging==23.0
99
+ gitdb==4.0.11
100
+ yarl==1.13.0
101
+ emoji==2.8.0
102
+ tzdata==2023.3
103
+ cycler==0.12.1
104
+ tornado==6.2
105
+ backcall==0.2.0
106
+ plotnine==0.12.4
107
+ ninja==1.11.1.1
108
+ latex==0.7.0
109
+ wandb==0.18.5
110
+ setproctitle==1.3.3
111
+ threadpoolctl==3.2.0
112
+ requests==2.32.3
113
+ pyparsing==3.1.1
114
+ smmap==5.0.1
115
+ pyzmq==23.0.0
116
+ async-timeout==4.0.3
117
+ annotated-types==0.7.0
118
+ matplotlib-inline==0.1.6
119
+ latexcodec==1.0.0
120
+ ipython==8.0.0
121
+ patsy==0.5.3
122
+ contourpy==1.2.0
123
+ multidict==6.1.0
124
+ mizani==0.9.3
125
+ urllib3==2.1.0
126
+ tokenizers==0.20.0
127
+ MarkupSafe==2.1.2
128
+ pip==24.2
129
+ pexpect==4.8.0
130
+ tqdm==4.66.5
131
+ jedi==0.18.2
132
+ pydantic_core==2.23.4
133
+ tempdir==0.7.1
134
+ mpmath==1.2.1
135
+ setuptools==72.1.0
136
+ pytest==7.4.3
137
+ pure-eval==0.2.2
138
+ psutil==5.9.1
139
+ comm==0.1.2
140
+ nvidia-cuda-cupti-cu11==11.7.101
141
+ nvidia-cuda-nvrtc-cu11==11.7.99
142
+ regex==2023.10.3
143
+ platformdirs==2.5.2
144
+ asttokens==2.2.1
145
+ torch==2.0.0
146
+ nvidia-nccl-cu11==2.14.3
147
+ xxhash==3.5.0
wandb/run-20241101_200502-ydj2ohih/files/wandb-metadata.json ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-162-generic-x86_64-with-glibc2.31",
3
+ "python": "3.9.19",
4
+ "startedAt": "2024-11-02T00:05:02.886763Z",
5
+ "args": [
6
+ "--perturbation",
7
+ "shuffle_nondeterministic",
8
+ "--train_set",
9
+ "10M",
10
+ "--batch_size",
11
+ "3",
12
+ "--epoch",
13
+ "3",
14
+ "--seed",
15
+ "0"
16
+ ],
17
+ "program": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py",
18
+ "codePath": "train/train_deep_wandb.py",
19
+ "git": {
20
+ "remote": "git@hf.co:Yaning1001/Impossible_llm.git",
21
+ "commit": "ed716cdcfcdea02b67f7ed0f3504c2b1c8b737c4"
22
+ },
23
+ "email": "yaning1001@gmail.com",
24
+ "root": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train",
25
+ "host": "mms-large-2",
26
+ "username": "chunhui",
27
+ "executable": "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/bin/python",
28
+ "codePathLocal": "train_deep_wandb.py",
29
+ "cpu_count": 32,
30
+ "cpu_count_logical": 64,
31
+ "gpu": "NVIDIA RTX A6000",
32
+ "gpu_count": 8,
33
+ "disk": {
34
+ "/": {
35
+ "total": "1888559353856",
36
+ "used": "1754801463296"
37
+ }
38
+ },
39
+ "memory": {
40
+ "total": "202617098240"
41
+ },
42
+ "cpu": {
43
+ "count": 32,
44
+ "countLogical": 64
45
+ },
46
+ "gpu_nvidia": [
47
+ {
48
+ "name": "NVIDIA RTX A6000",
49
+ "memoryTotal": "51527024640",
50
+ "cudaCores": 10752,
51
+ "architecture": "Ampere"
52
+ },
53
+ {
54
+ "name": "NVIDIA RTX A6000",
55
+ "memoryTotal": "51527024640",
56
+ "cudaCores": 10752,
57
+ "architecture": "Ampere"
58
+ },
59
+ {
60
+ "name": "NVIDIA RTX A6000",
61
+ "memoryTotal": "51527024640",
62
+ "cudaCores": 10752,
63
+ "architecture": "Ampere"
64
+ },
65
+ {
66
+ "name": "NVIDIA RTX A6000",
67
+ "memoryTotal": "51527024640",
68
+ "cudaCores": 10752,
69
+ "architecture": "Ampere"
70
+ },
71
+ {
72
+ "name": "NVIDIA RTX A6000",
73
+ "memoryTotal": "51527024640",
74
+ "cudaCores": 10752,
75
+ "architecture": "Ampere"
76
+ },
77
+ {
78
+ "name": "NVIDIA RTX A6000",
79
+ "memoryTotal": "51527024640",
80
+ "cudaCores": 10752,
81
+ "architecture": "Ampere"
82
+ },
83
+ {
84
+ "name": "NVIDIA RTX A6000",
85
+ "memoryTotal": "51527024640",
86
+ "cudaCores": 10752,
87
+ "architecture": "Ampere"
88
+ },
89
+ {
90
+ "name": "NVIDIA RTX A6000",
91
+ "memoryTotal": "51527024640",
92
+ "cudaCores": 10752,
93
+ "architecture": "Ampere"
94
+ }
95
+ ],
96
+ "cudaVersion": "11.8"
97
+ }
wandb/run-20241101_200502-ydj2ohih/logs/debug-internal.log ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {"time":"2024-11-01T20:05:02.888695149-04:00","level":"INFO","msg":"using version","core version":"0.18.5"}
2
+ {"time":"2024-11-01T20:05:02.888709799-04:00","level":"INFO","msg":"created symlink","path":"/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241101_200502-ydj2ohih/logs/debug-core.log"}
3
+ {"time":"2024-11-01T20:05:02.997414667-04:00","level":"INFO","msg":"created new stream","id":"ydj2ohih"}
4
+ {"time":"2024-11-01T20:05:02.997458267-04:00","level":"INFO","msg":"stream: started","id":"ydj2ohih"}
5
+ {"time":"2024-11-01T20:05:02.997548358-04:00","level":"INFO","msg":"sender: started","stream_id":"ydj2ohih"}
6
+ {"time":"2024-11-01T20:05:02.997488097-04:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"ydj2ohih"}}
7
+ {"time":"2024-11-01T20:05:02.997548098-04:00","level":"INFO","msg":"handler: started","stream_id":{"value":"ydj2ohih"}}
8
+ {"time":"2024-11-01T20:05:03.215129003-04:00","level":"INFO","msg":"Starting system monitor"}
wandb/run-20241101_200502-ydj2ohih/logs/debug.log ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-11-01 20:05:02,883 INFO MainThread:869510 [wandb_setup.py:_flush():79] Current SDK version is 0.18.5
2
+ 2024-11-01 20:05:02,883 INFO MainThread:869510 [wandb_setup.py:_flush():79] Configure stats pid to 869510
3
+ 2024-11-01 20:05:02,883 INFO MainThread:869510 [wandb_setup.py:_flush():79] Loading settings from /home/chunhui/.config/wandb/settings
4
+ 2024-11-01 20:05:02,883 INFO MainThread:869510 [wandb_setup.py:_flush():79] Loading settings from /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/settings
5
+ 2024-11-01 20:05:02,883 INFO MainThread:869510 [wandb_setup.py:_flush():79] Loading settings from environment variables: {}
6
+ 2024-11-01 20:05:02,883 INFO MainThread:869510 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': None, '_disable_service': None}
7
+ 2024-11-01 20:05:02,883 INFO MainThread:869510 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': 'train/train_deep_wandb.py', 'program_abspath': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py', 'program': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py'}
8
+ 2024-11-01 20:05:02,883 INFO MainThread:869510 [wandb_setup.py:_flush():79] Applying login settings: {}
9
+ 2024-11-01 20:05:02,883 INFO MainThread:869510 [wandb_init.py:_log_setup():534] Logging user logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241101_200502-ydj2ohih/logs/debug.log
10
+ 2024-11-01 20:05:02,883 INFO MainThread:869510 [wandb_init.py:_log_setup():535] Logging internal logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241101_200502-ydj2ohih/logs/debug-internal.log
11
+ 2024-11-01 20:05:02,884 INFO MainThread:869510 [wandb_init.py:init():621] calling init triggers
12
+ 2024-11-01 20:05:02,884 INFO MainThread:869510 [wandb_init.py:init():628] wandb.init called with sweep_config: {}
13
+ config: {}
14
+ 2024-11-01 20:05:02,884 INFO MainThread:869510 [wandb_init.py:init():671] starting backend
15
+ 2024-11-01 20:05:02,884 INFO MainThread:869510 [wandb_init.py:init():675] sending inform_init request
16
+ 2024-11-01 20:05:02,886 INFO MainThread:869510 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2024-11-01 20:05:02,886 INFO MainThread:869510 [wandb_init.py:init():688] backend started and connected
18
+ 2024-11-01 20:05:02,889 INFO MainThread:869510 [wandb_init.py:init():783] updated telemetry
19
+ 2024-11-01 20:05:02,921 INFO MainThread:869510 [wandb_init.py:init():816] communicating run to backend with 90.0 second timeout
20
+ 2024-11-01 20:05:03,212 INFO MainThread:869510 [wandb_init.py:init():867] starting run threads in backend
21
+ 2024-11-01 20:05:03,301 INFO MainThread:869510 [wandb_run.py:_console_start():2463] atexit reg
22
+ 2024-11-01 20:05:03,302 INFO MainThread:869510 [wandb_run.py:_redirect():2311] redirect: wrap_raw
23
+ 2024-11-01 20:05:03,302 INFO MainThread:869510 [wandb_run.py:_redirect():2376] Wrapping output streams.
24
+ 2024-11-01 20:05:03,302 INFO MainThread:869510 [wandb_run.py:_redirect():2401] Redirects installed.
25
+ 2024-11-01 20:05:03,303 INFO MainThread:869510 [wandb_init.py:init():911] run started, returning control to user process
26
+ 2024-11-01 20:05:03,304 INFO MainThread:869510 [wandb_run.py:_config_callback():1390] config_cb None None {'perturbation': 'shuffle_nondeterministic', 'train_set': '10M', 'batch_size': 3, 'epoch': 3, 'seed': 0, 'lr': 5e-06}
wandb/run-20241101_200502-ydj2ohih/run-ydj2ohih.wandb ADDED
File without changes
wandb/run-20241101_200517-7w8d5k1u/files/config.yaml ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.18.5
4
+ m: []
5
+ python_version: 3.9.19
6
+ t:
7
+ "1":
8
+ - 1
9
+ - 5
10
+ - 11
11
+ - 49
12
+ - 51
13
+ - 53
14
+ - 55
15
+ - 71
16
+ - 98
17
+ "2":
18
+ - 1
19
+ - 5
20
+ - 11
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 98
27
+ "3":
28
+ - 13
29
+ - 23
30
+ - 55
31
+ "4": 3.9.19
32
+ "5": 0.18.5
33
+ "6": 4.45.1
34
+ "8":
35
+ - 5
36
+ "12": 0.18.5
37
+ "13": linux-x86_64
38
+ batch_size:
39
+ value: 3
40
+ epoch:
41
+ value: 3
42
+ lr:
43
+ value: 5e-06
44
+ perturbation:
45
+ value: shuffle_nondeterministic
46
+ seed:
47
+ value: 0
48
+ train_set:
49
+ value: 10M
wandb/run-20241101_200517-7w8d5k1u/files/output.log ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ config.json: 100%|████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 844/844 [00:00<00:00, 423kB/s]
2
+ model.safetensors.index.json: 100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 20.9k/20.9k [00:00<00:00, 16.2MB/s]
3
+ model-00001-of-00002.safetensors: 6%|███████▋ | 294M/4.97G [00:06<01:50, 42.2MB/s]
4
+ Exception ignored in: <function BaseFileLock.__del__ at 0x7f9a29cd8af0> | 294M/4.97G [00:06<01:50, 42.4MB/s]
5
+ Traceback (most recent call last):
6
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/filelock/_api.py", line 273, in __del__
7
+ def __del__(self) -> None:
8
+ KeyboardInterrupt:
9
+ Downloading shards: 0%| | 0/2 [00:07<?, ?it/s]
10
+ Traceback (most recent call last):
11
+ File "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py", line 173, in <module>
12
+ model = AutoModelForCausalLM.from_pretrained(model_name,
13
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/models/auto/auto_factory.py", line 564, in from_pretrained
14
+ return model_class.from_pretrained(
15
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/modeling_utils.py", line 3769, in from_pretrained
16
+ resolved_archive_file, sharded_metadata = get_checkpoint_shard_files(
17
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/utils/hub.py", line 1098, in get_checkpoint_shard_files
18
+ cached_filename = cached_file(
19
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/utils/hub.py", line 403, in cached_file
20
+ resolved_file = hf_hub_download(
21
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/utils/_deprecation.py", line 101, in inner_f
22
+ return f(*args, **kwargs)
23
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn
24
+ return fn(*args, **kwargs)
25
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/file_download.py", line 1232, in hf_hub_download
26
+ return _hf_hub_download_to_cache_dir(
27
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/file_download.py", line 1381, in _hf_hub_download_to_cache_dir
28
+ _download_to_tmp_and_move(
29
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/file_download.py", line 1915, in _download_to_tmp_and_move
30
+ http_get(
31
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/huggingface_hub/file_download.py", line 541, in http_get
32
+ for chunk in r.iter_content(chunk_size=constants.DOWNLOAD_CHUNK_SIZE):
33
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/requests/models.py", line 820, in generate
34
+ yield from self.raw.stream(chunk_size, decode_content=True)
35
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/urllib3/response.py", line 934, in stream
36
+ data = self.read(amt=amt, decode_content=decode_content)
37
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/urllib3/response.py", line 877, in read
38
+ data = self._raw_read(amt)
39
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/urllib3/response.py", line 812, in _raw_read
40
+ data = self._fp_read(amt) if not fp_closed else b""
41
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/urllib3/response.py", line 789, in _fp_read
42
+ data = self._fp.read(chunk_amt)
43
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/http/client.py", line 463, in read
44
+ n = self.readinto(b)
45
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/http/client.py", line 507, in readinto
46
+ n = self.fp.readinto(b)
47
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/socket.py", line 704, in readinto
48
+ return self._sock.recv_into(b)
49
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/ssl.py", line 1275, in recv_into
50
+ return self.read(nbytes, buffer)
51
+ File "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/ssl.py", line 1133, in read
52
+ return self._sslobj.read(len, buffer)
53
+ KeyboardInterrupt
wandb/run-20241101_200517-7w8d5k1u/files/wandb-metadata.json ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-162-generic-x86_64-with-glibc2.31",
3
+ "python": "3.9.19",
4
+ "startedAt": "2024-11-02T00:05:17.122237Z",
5
+ "args": [
6
+ "--perturbation",
7
+ "shuffle_nondeterministic",
8
+ "--train_set",
9
+ "10M",
10
+ "--batch_size",
11
+ "3",
12
+ "--epoch",
13
+ "3",
14
+ "--seed",
15
+ "0"
16
+ ],
17
+ "program": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py",
18
+ "codePath": "train/train_deep_wandb.py",
19
+ "git": {
20
+ "remote": "git@hf.co:Yaning1001/Impossible_llm.git",
21
+ "commit": "ed716cdcfcdea02b67f7ed0f3504c2b1c8b737c4"
22
+ },
23
+ "email": "yaning1001@gmail.com",
24
+ "root": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train",
25
+ "host": "mms-large-2",
26
+ "username": "chunhui",
27
+ "executable": "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/bin/python",
28
+ "codePathLocal": "train_deep_wandb.py",
29
+ "cpu_count": 32,
30
+ "cpu_count_logical": 64,
31
+ "gpu": "NVIDIA RTX A6000",
32
+ "gpu_count": 8,
33
+ "disk": {
34
+ "/": {
35
+ "total": "1888559353856",
36
+ "used": "1754801557504"
37
+ }
38
+ },
39
+ "memory": {
40
+ "total": "202617098240"
41
+ },
42
+ "cpu": {
43
+ "count": 32,
44
+ "countLogical": 64
45
+ },
46
+ "gpu_nvidia": [
47
+ {
48
+ "name": "NVIDIA RTX A6000",
49
+ "memoryTotal": "51527024640",
50
+ "cudaCores": 10752,
51
+ "architecture": "Ampere"
52
+ },
53
+ {
54
+ "name": "NVIDIA RTX A6000",
55
+ "memoryTotal": "51527024640",
56
+ "cudaCores": 10752,
57
+ "architecture": "Ampere"
58
+ },
59
+ {
60
+ "name": "NVIDIA RTX A6000",
61
+ "memoryTotal": "51527024640",
62
+ "cudaCores": 10752,
63
+ "architecture": "Ampere"
64
+ },
65
+ {
66
+ "name": "NVIDIA RTX A6000",
67
+ "memoryTotal": "51527024640",
68
+ "cudaCores": 10752,
69
+ "architecture": "Ampere"
70
+ },
71
+ {
72
+ "name": "NVIDIA RTX A6000",
73
+ "memoryTotal": "51527024640",
74
+ "cudaCores": 10752,
75
+ "architecture": "Ampere"
76
+ },
77
+ {
78
+ "name": "NVIDIA RTX A6000",
79
+ "memoryTotal": "51527024640",
80
+ "cudaCores": 10752,
81
+ "architecture": "Ampere"
82
+ },
83
+ {
84
+ "name": "NVIDIA RTX A6000",
85
+ "memoryTotal": "51527024640",
86
+ "cudaCores": 10752,
87
+ "architecture": "Ampere"
88
+ },
89
+ {
90
+ "name": "NVIDIA RTX A6000",
91
+ "memoryTotal": "51527024640",
92
+ "cudaCores": 10752,
93
+ "architecture": "Ampere"
94
+ }
95
+ ],
96
+ "cudaVersion": "11.8"
97
+ }
wandb/run-20241101_200517-7w8d5k1u/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_wandb":{"runtime":8}}
wandb/run-20241101_200517-7w8d5k1u/logs/debug-internal.log ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2024-11-01T20:05:17.125935487-04:00","level":"INFO","msg":"using version","core version":"0.18.5"}
2
+ {"time":"2024-11-01T20:05:17.125950717-04:00","level":"INFO","msg":"created symlink","path":"/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241101_200517-7w8d5k1u/logs/debug-core.log"}
3
+ {"time":"2024-11-01T20:05:17.232850512-04:00","level":"INFO","msg":"created new stream","id":"7w8d5k1u"}
4
+ {"time":"2024-11-01T20:05:17.232884112-04:00","level":"INFO","msg":"stream: started","id":"7w8d5k1u"}
5
+ {"time":"2024-11-01T20:05:17.233000493-04:00","level":"INFO","msg":"sender: started","stream_id":"7w8d5k1u"}
6
+ {"time":"2024-11-01T20:05:17.232970413-04:00","level":"INFO","msg":"handler: started","stream_id":{"value":"7w8d5k1u"}}
7
+ {"time":"2024-11-01T20:05:17.232915643-04:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"7w8d5k1u"}}
8
+ {"time":"2024-11-01T20:05:17.489000404-04:00","level":"INFO","msg":"Starting system monitor"}
9
+ {"time":"2024-11-01T20:05:25.26447236-04:00","level":"INFO","msg":"stream: closing","id":"7w8d5k1u"}
10
+ {"time":"2024-11-01T20:05:25.26451279-04:00","level":"INFO","msg":"Stopping system monitor"}
11
+ {"time":"2024-11-01T20:05:25.265030874-04:00","level":"INFO","msg":"Stopped system monitor"}
wandb/run-20241101_200517-7w8d5k1u/logs/debug.log ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-11-01 20:05:17,118 INFO MainThread:870380 [wandb_setup.py:_flush():79] Current SDK version is 0.18.5
2
+ 2024-11-01 20:05:17,119 INFO MainThread:870380 [wandb_setup.py:_flush():79] Configure stats pid to 870380
3
+ 2024-11-01 20:05:17,119 INFO MainThread:870380 [wandb_setup.py:_flush():79] Loading settings from /home/chunhui/.config/wandb/settings
4
+ 2024-11-01 20:05:17,119 INFO MainThread:870380 [wandb_setup.py:_flush():79] Loading settings from /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/settings
5
+ 2024-11-01 20:05:17,119 INFO MainThread:870380 [wandb_setup.py:_flush():79] Loading settings from environment variables: {}
6
+ 2024-11-01 20:05:17,119 INFO MainThread:870380 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': None, '_disable_service': None}
7
+ 2024-11-01 20:05:17,119 INFO MainThread:870380 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': 'train/train_deep_wandb.py', 'program_abspath': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py', 'program': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py'}
8
+ 2024-11-01 20:05:17,119 INFO MainThread:870380 [wandb_setup.py:_flush():79] Applying login settings: {}
9
+ 2024-11-01 20:05:17,119 INFO MainThread:870380 [wandb_init.py:_log_setup():534] Logging user logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241101_200517-7w8d5k1u/logs/debug.log
10
+ 2024-11-01 20:05:17,119 INFO MainThread:870380 [wandb_init.py:_log_setup():535] Logging internal logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241101_200517-7w8d5k1u/logs/debug-internal.log
11
+ 2024-11-01 20:05:17,119 INFO MainThread:870380 [wandb_init.py:init():621] calling init triggers
12
+ 2024-11-01 20:05:17,119 INFO MainThread:870380 [wandb_init.py:init():628] wandb.init called with sweep_config: {}
13
+ config: {}
14
+ 2024-11-01 20:05:17,119 INFO MainThread:870380 [wandb_init.py:init():671] starting backend
15
+ 2024-11-01 20:05:17,119 INFO MainThread:870380 [wandb_init.py:init():675] sending inform_init request
16
+ 2024-11-01 20:05:17,121 INFO MainThread:870380 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2024-11-01 20:05:17,122 INFO MainThread:870380 [wandb_init.py:init():688] backend started and connected
18
+ 2024-11-01 20:05:17,126 INFO MainThread:870380 [wandb_init.py:init():783] updated telemetry
19
+ 2024-11-01 20:05:17,156 INFO MainThread:870380 [wandb_init.py:init():816] communicating run to backend with 90.0 second timeout
20
+ 2024-11-01 20:05:17,486 INFO MainThread:870380 [wandb_init.py:init():867] starting run threads in backend
21
+ 2024-11-01 20:05:17,572 INFO MainThread:870380 [wandb_run.py:_console_start():2463] atexit reg
22
+ 2024-11-01 20:05:17,572 INFO MainThread:870380 [wandb_run.py:_redirect():2311] redirect: wrap_raw
23
+ 2024-11-01 20:05:17,572 INFO MainThread:870380 [wandb_run.py:_redirect():2376] Wrapping output streams.
24
+ 2024-11-01 20:05:17,572 INFO MainThread:870380 [wandb_run.py:_redirect():2401] Redirects installed.
25
+ 2024-11-01 20:05:17,573 INFO MainThread:870380 [wandb_init.py:init():911] run started, returning control to user process
26
+ 2024-11-01 20:05:17,574 INFO MainThread:870380 [wandb_run.py:_config_callback():1390] config_cb None None {'perturbation': 'shuffle_nondeterministic', 'train_set': '10M', 'batch_size': 3, 'epoch': 3, 'seed': 0, 'lr': 5e-06}
27
+ 2024-11-01 20:05:25,264 WARNING MsgRouterThr:870380 [router.py:message_loop():77] message_loop has been closed
wandb/run-20241101_200517-7w8d5k1u/run-7w8d5k1u.wandb ADDED
File without changes
wandb/run-20241101_200535-kh416n4m/files/output.log ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Downloading shards: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 2/2 [02:25<00:00, 72.79s/it]
2
+ Loading checkpoint shards: 100%|████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 2/2 [00:05<00:00, 2.66s/it]
3
+ tokenized_valid: Dataset({
4
+ features: ['input_ids', 'attention_mask'],
5
+ num_rows: 600
6
+ })
7
+ /mnt/ssd3/chunhui/miniconda/envs/impossible_llm/lib/python3.9/site-packages/transformers/training_args.py:1545: FutureWarning: `evaluation_strategy` is deprecated and will be removed in version 4.46 of 🤗 Transformers. Use `eval_strategy` instead
8
+ warnings.warn(
9
+ [2024-11-01 20:08:09,698] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect)
10
+ [2024-11-01 20:08:20,613] [INFO] [comm.py:652:init_distributed] cdb=None
11
+ Installed CUDA version 11.8 does not match the version torch was compiled with 11.7 but since the APIs are compatible, accepting this combination
12
+ Using /home/chunhui/.cache/torch_extensions/py39_cu117 as PyTorch extensions root...
13
+ Loading extension module cpu_adam...
14
+ Time to load cpu_adam op: 5.421607255935669 seconds
wandb/run-20241101_200535-kh416n4m/files/requirements.txt ADDED
@@ -0,0 +1,147 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ funcsigs==1.0.2
2
+ sentry-sdk==2.17.0
3
+ multiprocess==0.70.16
4
+ numpy==1.26.2
5
+ pluralizer==1.2.0
6
+ debugpy==1.6.7
7
+ nvidia-cudnn-cu11==8.5.0.96
8
+ deepspeed==0.15.2
9
+ data==0.4
10
+ pandas==2.1.3
11
+ tomli==2.0.1
12
+ charset-normalizer==3.3.2
13
+ attrs==24.2.0
14
+ aiosignal==1.3.1
15
+ fsspec==2023.10.0
16
+ nvidia-cusparse-cu11==11.7.4.91
17
+ zipp==3.12.0
18
+ mypy-extensions==1.0.0
19
+ datasets==3.0.1
20
+ joblib==1.3.2
21
+ hjson==3.1.0
22
+ traitlets==5.7.1
23
+ stack-data==0.6.0
24
+ transformers==4.45.1
25
+ sympy==1.11.1
26
+ Pygments==2.15.0
27
+ docker-pycreds==0.4.0
28
+ dill==0.3.8
29
+ wheel==0.44.0
30
+ prompt-toolkit==3.0.30
31
+ parso==0.8.3
32
+ ipykernel==6.23.1
33
+ pyarrow==17.0.0
34
+ certifi==2023.11.17
35
+ nvidia-cufft-cu11==10.9.0.58
36
+ six==1.16.0
37
+ pydantic==2.9.2
38
+ click==8.1.7
39
+ nest-asyncio==1.5.6
40
+ gmpy2==2.1.0
41
+ matplotlib==3.8.2
42
+ scipy==1.11.4
43
+ typing_extensions==4.12.2
44
+ statsmodels==0.14.0
45
+ huggingface-hub==0.25.0
46
+ frozenlist==1.4.1
47
+ gpustat==1.1.1
48
+ nvidia-nvtx-cu11==11.7.91
49
+ safetensors==0.4.5
50
+ stanza==1.9.2
51
+ decorator==5.1.1
52
+ seaborn==0.13.0
53
+ sentencepiece==0.2.0
54
+ PyYAML==6.0.1
55
+ black==24.8.0
56
+ protobuf==4.25.1
57
+ pickleshare==0.7.5
58
+ peft==0.13.0
59
+ triton==2.0.0
60
+ nvidia-cuda-runtime-cu11==11.7.99
61
+ Jinja2==3.1.2
62
+ nvidia-cusolver-cu11==11.4.0.1
63
+ executing==1.2.0
64
+ jupyter_client==8.1.0
65
+ pluggy==1.3.0
66
+ cmake==3.30.3
67
+ pytz==2023.3.post1
68
+ aiohappyeyeballs==2.4.2
69
+ kiwisolver==1.4.5
70
+ py-cpuinfo==9.0.0
71
+ Pillow==10.1.0
72
+ ptyprocess==0.7.0
73
+ importlib_resources==6.4.5
74
+ GitPython==3.1.43
75
+ importlib-metadata==6.0.0
76
+ iniconfig==2.0.0
77
+ scikit-learn==1.3.2
78
+ exceptiongroup==1.1.0
79
+ networkx==2.8.6
80
+ accelerate==1.0.0
81
+ nltk==3.8.1
82
+ shutilwhich==1.1.0
83
+ fonttools==4.45.1
84
+ future==0.18.3
85
+ aiohttp==3.10.6
86
+ wcwidth==0.2.5
87
+ idna==3.6
88
+ filelock==3.12.2
89
+ pathspec==0.12.1
90
+ jupyter_core==5.1.0
91
+ lit==18.1.8
92
+ nvidia-curand-cu11==10.2.10.91
93
+ nvidia-cublas-cu11==11.10.3.66
94
+ nvidia-ml-py==12.560.30
95
+ msgpack==1.1.0
96
+ python-dateutil==2.8.2
97
+ blessed==1.20.0
98
+ packaging==23.0
99
+ gitdb==4.0.11
100
+ yarl==1.13.0
101
+ emoji==2.8.0
102
+ tzdata==2023.3
103
+ cycler==0.12.1
104
+ tornado==6.2
105
+ backcall==0.2.0
106
+ plotnine==0.12.4
107
+ ninja==1.11.1.1
108
+ latex==0.7.0
109
+ wandb==0.18.5
110
+ setproctitle==1.3.3
111
+ threadpoolctl==3.2.0
112
+ requests==2.32.3
113
+ pyparsing==3.1.1
114
+ smmap==5.0.1
115
+ pyzmq==23.0.0
116
+ async-timeout==4.0.3
117
+ annotated-types==0.7.0
118
+ matplotlib-inline==0.1.6
119
+ latexcodec==1.0.0
120
+ ipython==8.0.0
121
+ patsy==0.5.3
122
+ contourpy==1.2.0
123
+ multidict==6.1.0
124
+ mizani==0.9.3
125
+ urllib3==2.1.0
126
+ tokenizers==0.20.0
127
+ MarkupSafe==2.1.2
128
+ pip==24.2
129
+ pexpect==4.8.0
130
+ tqdm==4.66.5
131
+ jedi==0.18.2
132
+ pydantic_core==2.23.4
133
+ tempdir==0.7.1
134
+ mpmath==1.2.1
135
+ setuptools==72.1.0
136
+ pytest==7.4.3
137
+ pure-eval==0.2.2
138
+ psutil==5.9.1
139
+ comm==0.1.2
140
+ nvidia-cuda-cupti-cu11==11.7.101
141
+ nvidia-cuda-nvrtc-cu11==11.7.99
142
+ regex==2023.10.3
143
+ platformdirs==2.5.2
144
+ asttokens==2.2.1
145
+ torch==2.0.0
146
+ nvidia-nccl-cu11==2.14.3
147
+ xxhash==3.5.0
wandb/run-20241101_200535-kh416n4m/files/wandb-metadata.json ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-162-generic-x86_64-with-glibc2.31",
3
+ "python": "3.9.19",
4
+ "startedAt": "2024-11-02T00:05:35.952941Z",
5
+ "args": [
6
+ "--perturbation",
7
+ "shuffle_nondeterministic",
8
+ "--train_set",
9
+ "10M",
10
+ "--batch_size",
11
+ "3",
12
+ "--epoch",
13
+ "3",
14
+ "--seed",
15
+ "0"
16
+ ],
17
+ "program": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py",
18
+ "codePath": "train/train_deep_wandb.py",
19
+ "git": {
20
+ "remote": "git@hf.co:Yaning1001/Impossible_llm.git",
21
+ "commit": "ed716cdcfcdea02b67f7ed0f3504c2b1c8b737c4"
22
+ },
23
+ "email": "yaning1001@gmail.com",
24
+ "root": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train",
25
+ "host": "mms-large-2",
26
+ "username": "chunhui",
27
+ "executable": "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/bin/python",
28
+ "codePathLocal": "train_deep_wandb.py",
29
+ "cpu_count": 32,
30
+ "cpu_count_logical": 64,
31
+ "gpu": "NVIDIA RTX A6000",
32
+ "gpu_count": 8,
33
+ "disk": {
34
+ "/": {
35
+ "total": "1888559353856",
36
+ "used": "1754801680384"
37
+ }
38
+ },
39
+ "memory": {
40
+ "total": "202617098240"
41
+ },
42
+ "cpu": {
43
+ "count": 32,
44
+ "countLogical": 64
45
+ },
46
+ "gpu_nvidia": [
47
+ {
48
+ "name": "NVIDIA RTX A6000",
49
+ "memoryTotal": "51527024640",
50
+ "cudaCores": 10752,
51
+ "architecture": "Ampere"
52
+ },
53
+ {
54
+ "name": "NVIDIA RTX A6000",
55
+ "memoryTotal": "51527024640",
56
+ "cudaCores": 10752,
57
+ "architecture": "Ampere"
58
+ },
59
+ {
60
+ "name": "NVIDIA RTX A6000",
61
+ "memoryTotal": "51527024640",
62
+ "cudaCores": 10752,
63
+ "architecture": "Ampere"
64
+ },
65
+ {
66
+ "name": "NVIDIA RTX A6000",
67
+ "memoryTotal": "51527024640",
68
+ "cudaCores": 10752,
69
+ "architecture": "Ampere"
70
+ },
71
+ {
72
+ "name": "NVIDIA RTX A6000",
73
+ "memoryTotal": "51527024640",
74
+ "cudaCores": 10752,
75
+ "architecture": "Ampere"
76
+ },
77
+ {
78
+ "name": "NVIDIA RTX A6000",
79
+ "memoryTotal": "51527024640",
80
+ "cudaCores": 10752,
81
+ "architecture": "Ampere"
82
+ },
83
+ {
84
+ "name": "NVIDIA RTX A6000",
85
+ "memoryTotal": "51527024640",
86
+ "cudaCores": 10752,
87
+ "architecture": "Ampere"
88
+ },
89
+ {
90
+ "name": "NVIDIA RTX A6000",
91
+ "memoryTotal": "51527024640",
92
+ "cudaCores": 10752,
93
+ "architecture": "Ampere"
94
+ }
95
+ ],
96
+ "cudaVersion": "11.8"
97
+ }
wandb/run-20241101_200535-kh416n4m/logs/debug-internal.log ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {"time":"2024-11-01T20:05:35.954949925-04:00","level":"INFO","msg":"using version","core version":"0.18.5"}
2
+ {"time":"2024-11-01T20:05:35.954961945-04:00","level":"INFO","msg":"created symlink","path":"/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241101_200535-kh416n4m/logs/debug-core.log"}
3
+ {"time":"2024-11-01T20:05:36.060080068-04:00","level":"INFO","msg":"created new stream","id":"kh416n4m"}
4
+ {"time":"2024-11-01T20:05:36.060105218-04:00","level":"INFO","msg":"stream: started","id":"kh416n4m"}
5
+ {"time":"2024-11-01T20:05:36.060156568-04:00","level":"INFO","msg":"sender: started","stream_id":"kh416n4m"}
6
+ {"time":"2024-11-01T20:05:36.060165538-04:00","level":"INFO","msg":"handler: started","stream_id":{"value":"kh416n4m"}}
7
+ {"time":"2024-11-01T20:05:36.060141578-04:00","level":"INFO","msg":"writer: Do: started","stream_id":{"value":"kh416n4m"}}
8
+ {"time":"2024-11-01T20:05:36.221966887-04:00","level":"INFO","msg":"Starting system monitor"}
wandb/run-20241101_200535-kh416n4m/logs/debug.log ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-11-01 20:05:35,949 INFO MainThread:871228 [wandb_setup.py:_flush():79] Current SDK version is 0.18.5
2
+ 2024-11-01 20:05:35,949 INFO MainThread:871228 [wandb_setup.py:_flush():79] Configure stats pid to 871228
3
+ 2024-11-01 20:05:35,949 INFO MainThread:871228 [wandb_setup.py:_flush():79] Loading settings from /home/chunhui/.config/wandb/settings
4
+ 2024-11-01 20:05:35,950 INFO MainThread:871228 [wandb_setup.py:_flush():79] Loading settings from /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/settings
5
+ 2024-11-01 20:05:35,950 INFO MainThread:871228 [wandb_setup.py:_flush():79] Loading settings from environment variables: {}
6
+ 2024-11-01 20:05:35,950 INFO MainThread:871228 [wandb_setup.py:_flush():79] Applying setup settings: {'mode': None, '_disable_service': None}
7
+ 2024-11-01 20:05:35,950 INFO MainThread:871228 [wandb_setup.py:_flush():79] Inferring run settings from compute environment: {'program_relpath': 'train/train_deep_wandb.py', 'program_abspath': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py', 'program': '/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py'}
8
+ 2024-11-01 20:05:35,950 INFO MainThread:871228 [wandb_setup.py:_flush():79] Applying login settings: {}
9
+ 2024-11-01 20:05:35,950 INFO MainThread:871228 [wandb_init.py:_log_setup():534] Logging user logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241101_200535-kh416n4m/logs/debug.log
10
+ 2024-11-01 20:05:35,950 INFO MainThread:871228 [wandb_init.py:_log_setup():535] Logging internal logs to /mnt/ssd3/chunhui/yaning/project/impossible_llm/train/wandb/run-20241101_200535-kh416n4m/logs/debug-internal.log
11
+ 2024-11-01 20:05:35,950 INFO MainThread:871228 [wandb_init.py:init():621] calling init triggers
12
+ 2024-11-01 20:05:35,950 INFO MainThread:871228 [wandb_init.py:init():628] wandb.init called with sweep_config: {}
13
+ config: {}
14
+ 2024-11-01 20:05:35,950 INFO MainThread:871228 [wandb_init.py:init():671] starting backend
15
+ 2024-11-01 20:05:35,950 INFO MainThread:871228 [wandb_init.py:init():675] sending inform_init request
16
+ 2024-11-01 20:05:35,952 INFO MainThread:871228 [backend.py:_multiprocessing_setup():104] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
17
+ 2024-11-01 20:05:35,952 INFO MainThread:871228 [wandb_init.py:init():688] backend started and connected
18
+ 2024-11-01 20:05:35,955 INFO MainThread:871228 [wandb_init.py:init():783] updated telemetry
19
+ 2024-11-01 20:05:35,976 INFO MainThread:871228 [wandb_init.py:init():816] communicating run to backend with 90.0 second timeout
20
+ 2024-11-01 20:05:36,218 INFO MainThread:871228 [wandb_init.py:init():867] starting run threads in backend
21
+ 2024-11-01 20:05:36,308 INFO MainThread:871228 [wandb_run.py:_console_start():2463] atexit reg
22
+ 2024-11-01 20:05:36,308 INFO MainThread:871228 [wandb_run.py:_redirect():2311] redirect: wrap_raw
23
+ 2024-11-01 20:05:36,309 INFO MainThread:871228 [wandb_run.py:_redirect():2376] Wrapping output streams.
24
+ 2024-11-01 20:05:36,309 INFO MainThread:871228 [wandb_run.py:_redirect():2401] Redirects installed.
25
+ 2024-11-01 20:05:36,310 INFO MainThread:871228 [wandb_init.py:init():911] run started, returning control to user process
26
+ 2024-11-01 20:05:36,311 INFO MainThread:871228 [wandb_run.py:_config_callback():1390] config_cb None None {'perturbation': 'shuffle_nondeterministic', 'train_set': '10M', 'batch_size': 3, 'epoch': 3, 'seed': 0, 'lr': 5e-06}
wandb/run-20241101_201630-e5gt2fir/files/wandb-metadata.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-162-generic-x86_64-with-glibc2.31",
3
+ "python": "3.9.19",
4
+ "startedAt": "2024-11-02T00:16:30.559840Z",
5
+ "args": [
6
+ "--perturbation",
7
+ "shuffle_nodeterministic",
8
+ "--train_set",
9
+ "10M",
10
+ "--batch_size",
11
+ "3",
12
+ "--epoch",
13
+ "6",
14
+ "--seed",
15
+ "0"
16
+ ],
17
+ "program": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train/train_deep_wandb.py",
18
+ "codePath": "train/train_deep_wandb.py",
19
+ "git": {
20
+ "remote": "git@hf.co:Yaning1001/Impossible_llm.git",
21
+ "commit": "ed716cdcfcdea02b67f7ed0f3504c2b1c8b737c4"
22
+ },
23
+ "email": "yaning1001@gmail.com",
24
+ "root": "/mnt/ssd3/chunhui/yaning/project/impossible_llm/train",
25
+ "host": "mms-large-2",
26
+ "username": "chunhui",
27
+ "executable": "/mnt/ssd3/chunhui/miniconda/envs/impossible_llm/bin/python",
28
+ "codePathLocal": "train_deep_wandb.py"
29
+ }
wandb/run-20241101_201630-e5gt2fir/run-e5gt2fir.wandb ADDED
Binary file (3.4 kB). View file
 
wandb/run-20241101_202058-ptl7coag/run-ptl7coag.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f69d6a48f65835780c44feae49d31d9772a0b67ce04a0dd11866a80d8fc15eb
3
+ size 13298469