ace14459tv
codeT5ベースのエラー診断モデル(max_length=324)
4a56686
raw
history blame
8.68 kB
GPU available: True (cuda), used: True
TPU available: False, using: 0 TPU cores
IPU available: False, using: 0 IPUs
HPU available: False, using: 0 HPUs
Downloading and preparing dataset json/default to /home/ace14459tv/t5maru/cache/json/default-508436ce1b237537/0.0.0/e347ab1c932092252e717ff3f949105a4dd28b27e842dd53157d2f72e276c2e4...
Downloading data files: 0%| | 0/1 [00:00<?, ?it/s] Downloading data files: 100%|██████████| 1/1 [00:00<00:00, 8371.86it/s]
Extracting data files: 0%| | 0/1 [00:00<?, ?it/s] Extracting data files: 100%|██████████| 1/1 [00:00<00:00, 407.97it/s]
Generating train split: 0 examples [00:00, ? examples/s] Dataset json downloaded and prepared to /home/ace14459tv/t5maru/cache/json/default-508436ce1b237537/0.0.0/e347ab1c932092252e717ff3f949105a4dd28b27e842dd53157d2f72e276c2e4. Subsequent calls will reuse this data.
Map (num_proc=4): 0%| | 0/2880 [00:00<?, ? examples/s] Map (num_proc=4): 1%|▏ | 38/2880 [00:00<00:43, 64.61 examples/s] Map (num_proc=4): 7%|▋ | 215/2880 [00:00<00:07, 361.81 examples/s] Map (num_proc=4): 14%|█▍ | 396/2880 [00:00<00:04, 593.37 examples/s] Map (num_proc=4): 18%|█▊ | 511/2880 [00:00<00:03, 707.90 examples/s] Map (num_proc=4): 26%|██▌ | 744/2880 [00:01<00:02, 1067.25 examples/s] Map (num_proc=4): 33%|███▎ | 950/2880 [00:01<00:01, 1230.14 examples/s] Map (num_proc=4): 39%|███▉ | 1128/2880 [00:01<00:01, 1302.62 examples/s] Map (num_proc=4): 45%|████▌ | 1308/2880 [00:01<00:01, 1265.23 examples/s] Map (num_proc=4): 53%|█████▎ | 1536/2880 [00:01<00:00, 1508.47 examples/s] Map (num_proc=4): 60%|██████ | 1731/2880 [00:01<00:00, 1575.00 examples/s] Map (num_proc=4): 66%|██████▋ | 1912/2880 [00:01<00:00, 1463.51 examples/s] Map (num_proc=4): 74%|███████▍ | 2145/2880 [00:01<00:00, 1540.43 examples/s] Map (num_proc=4): 83%|████████▎ | 2386/2880 [00:02<00:00, 1751.55 examples/s] Map (num_proc=4): 90%|█████████ | 2601/2880 [00:02<00:00, 1469.03 examples/s] Map (num_proc=4): 97%|█████████▋| 2785/2880 [00:02<00:00, 1390.78 examples/s] Downloading and preparing dataset json/default to /home/ace14459tv/t5maru/cache/json/default-85f4bdc1e5b3a604/0.0.0/e347ab1c932092252e717ff3f949105a4dd28b27e842dd53157d2f72e276c2e4...
Downloading data files: 0%| | 0/1 [00:00<?, ?it/s] Downloading data files: 100%|██████████| 1/1 [00:00<00:00, 9962.72it/s]
Extracting data files: 0%| | 0/1 [00:00<?, ?it/s] Extracting data files: 100%|██████████| 1/1 [00:00<00:00, 483.05it/s]
Generating train split: 0 examples [00:00, ? examples/s] Dataset json downloaded and prepared to /home/ace14459tv/t5maru/cache/json/default-85f4bdc1e5b3a604/0.0.0/e347ab1c932092252e717ff3f949105a4dd28b27e842dd53157d2f72e276c2e4. Subsequent calls will reuse this data.
Map (num_proc=4): 0%| | 0/618 [00:00<?, ? examples/s] Map (num_proc=4): 8%|▊ | 49/618 [00:00<00:06, 85.68 examples/s] Map (num_proc=4): 25%|██▌ | 155/618 [00:00<00:01, 277.60 examples/s] Map (num_proc=4): 42%|████▏ | 262/618 [00:01<00:01, 274.08 examples/s] Map (num_proc=4): 58%|█████▊ | 359/618 [00:01<00:00, 261.13 examples/s] Map (num_proc=4): 75%|███████▌ | 464/618 [00:01<00:00, 356.96 examples/s] Map (num_proc=4): 92%|█████████▏| 571/618 [00:01<00:00, 329.51 examples/s] LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0]
| Name | Type | Params
------------------------------------------
0 | model | OptimizedModule | 222 M
------------------------------------------
222 M Trainable params
0 Non-trainable params
222 M Total params
891.528 Total estimated model params size (MB)
[2023-12-21 03:07:16,808] torch._inductor.utils: [WARNING] using triton random, expect difference from eager
Metric val_loss improved. New best score: 0.060
Metric val_loss improved by 0.034 >= min_delta = 0.0. New best score: 0.027
Metric val_loss improved by 0.004 >= min_delta = 0.0. New best score: 0.022
Metric val_loss improved by 0.002 >= min_delta = 0.0. New best score: 0.020
Metric val_loss improved by 0.001 >= min_delta = 0.0. New best score: 0.019
Monitored metric val_loss did not improve in the last 3 records. Best score: 0.019. Signaling Trainer to stop.
{"log": "trained", "date": "2023-12-21T03:06:16", "elapsed": "00:06:28", "model": "Salesforce/codet5-base", "max_length": 324, "target_max_length": 324, "batch_size": 32, "gradient_accumulation_steps": 1, "train_steps": 2700, "accelerator": "gpu", "devices": "auto", "precision": 32, "strategy": "auto", "gradient_clip_val": 1.0, "compile": true, "solver": "adamw", "lr": 0.0003, "warmup_steps": 1, "training_steps": 100000, "adam_epsilon": 1e-08, "weight_decay": 0.0, "epoch": 9, "step": 810, "saved": "error_codet5_base_324"}
Traceback (most recent call last):
File "/home/ace14459tv/cuda117/lib/python3.10/site-packages/huggingface_hub/utils/_errors.py", line 259, in hf_raise_for_status
response.raise_for_status()
File "/home/ace14459tv/cuda117/lib/python3.10/site-packages/requests/models.py", line 1021, in raise_for_status
raise HTTPError(http_error_msg, response=self)
requests.exceptions.HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/error_codet5_base_512/resolve/main/tokenizer_config.json
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/home/ace14459tv/cuda117/lib/python3.10/site-packages/transformers/utils/hub.py", line 409, in cached_file
resolved_file = hf_hub_download(
File "/home/ace14459tv/cuda117/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 120, in _inner_fn
return fn(*args, **kwargs)
File "/home/ace14459tv/cuda117/lib/python3.10/site-packages/huggingface_hub/file_download.py", line 1195, in hf_hub_download
metadata = get_hf_file_metadata(
File "/home/ace14459tv/cuda117/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 120, in _inner_fn
return fn(*args, **kwargs)
File "/home/ace14459tv/cuda117/lib/python3.10/site-packages/huggingface_hub/file_download.py", line 1541, in get_hf_file_metadata
hf_raise_for_status(r)
File "/home/ace14459tv/cuda117/lib/python3.10/site-packages/huggingface_hub/utils/_errors.py", line 291, in hf_raise_for_status
raise RepositoryNotFoundError(message, response) from e
huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-65832ea4-32e585824a1465906a86bc9c)
Repository Not Found for url: https://huggingface.co/error_codet5_base_512/resolve/main/tokenizer_config.json.
Please make sure you specified the correct `repo_id` and `repo_type`.
If you are trying to access a private or gated repo, make sure you are authenticated.
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/ace14459tv/cuda117/bin/t5test", line 8, in <module>
sys.exit(main_test())
File "/home/ace14459tv/cuda117/lib/python3.10/site-packages/t5maru/t5tune.py", line 612, in main_test
model = T5Model(
File "/home/ace14459tv/cuda117/lib/python3.10/site-packages/t5maru/t5tune.py", line 227, in __init__
self.tokenizer = AutoTokenizer.from_pretrained(
File "/home/ace14459tv/cuda117/lib/python3.10/site-packages/transformers/models/auto/tokenization_auto.py", line 642, in from_pretrained
tokenizer_config = get_tokenizer_config(pretrained_model_name_or_path, **kwargs)
File "/home/ace14459tv/cuda117/lib/python3.10/site-packages/transformers/models/auto/tokenization_auto.py", line 486, in get_tokenizer_config
resolved_config_file = cached_file(
File "/home/ace14459tv/cuda117/lib/python3.10/site-packages/transformers/utils/hub.py", line 424, in cached_file
raise EnvironmentError(
OSError: error_codet5_base_512 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models'
If this is a private repository, make sure to pass a token having permission to this repo with `use_auth_token` or log in with `huggingface-cli login` and pass `use_auth_token=True`.