runtime error
Exit code: 1. Reason: py", line 220, in setup_models tokenizer = AutoTokenizer.from_pretrained( "meta-llama/Llama-3.2-1B-Instruct", token=hf_token ) File "/usr/local/lib/python3.13/site-packages/transformers/models/auto/tokenization_auto.py", line 689, in from_pretrained config = PreTrainedConfig.from_pretrained(pretrained_model_name_or_path, **kwargs) File "/usr/local/lib/python3.13/site-packages/transformers/configuration_utils.py", line 626, in from_pretrained config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs) ~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.13/site-packages/transformers/configuration_utils.py", line 667, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) ~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.13/site-packages/transformers/configuration_utils.py", line 722, in _get_config_dict resolved_config_file = cached_file( pretrained_model_name_or_path, ...<9 lines>... _commit_hash=commit_hash, ) File "/usr/local/lib/python3.13/site-packages/transformers/utils/hub.py", line 278, in cached_file file = cached_files(path_or_repo_id=path_or_repo_id, filenames=[filename], **kwargs) File "/usr/local/lib/python3.13/site-packages/transformers/utils/hub.py", line 488, in cached_files raise OSError( ...<2 lines>... ) from e OSError: You are trying to access a gated repo. Make sure to have access to it at https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct. 401 Client Error. (Request ID: Root=1-69d5ab43-54aa05131c2048555ca8afc6;c64c69df-aac6-482c-a2dc-2d9acae8f5f7) Cannot access gated repo for url https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct/resolve/main/config.json. Access to model meta-llama/Llama-3.2-1B-Instruct is restricted. You must have access to it and be authenticated to access it. Please log in.
Container logs:
Fetching error logs...