Upload meta-llama_Llama-3.1-8B-Instruct_2.txt with huggingface_hub
Browse files
meta-llama_Llama-3.1-8B-Instruct_2.txt
CHANGED
|
@@ -49,7 +49,7 @@ Traceback (most recent call last):
|
|
| 49 |
~~~~~~~~~~~~~~~~~~~^^^^^^^^^^
|
| 50 |
File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/huggingface_hub/utils/_http.py", line 802, in hf_raise_for_status
|
| 51 |
raise _format(HfHubHTTPError, message, response) from e
|
| 52 |
-
huggingface_hub.errors.HfHubHTTPError: (Request ID: Root=1-
|
| 53 |
|
| 54 |
403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository..
|
| 55 |
Cannot access content at: https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json.
|
|
@@ -144,7 +144,7 @@ Traceback (most recent call last):
|
|
| 144 |
~~~~~~~~~~~~~~~~~~~^^^^^^^^^^
|
| 145 |
File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/huggingface_hub/utils/_http.py", line 802, in hf_raise_for_status
|
| 146 |
raise _format(HfHubHTTPError, message, response) from e
|
| 147 |
-
huggingface_hub.errors.HfHubHTTPError: (Request ID: Root=1-
|
| 148 |
|
| 149 |
403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository..
|
| 150 |
Cannot access content at: https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json.
|
|
@@ -183,7 +183,7 @@ huggingface_hub.errors.LocalEntryNotFoundError: An error happened while trying t
|
|
| 183 |
The above exception was the direct cause of the following exception:
|
| 184 |
|
| 185 |
Traceback (most recent call last):
|
| 186 |
-
File "/tmp/meta-llama_Llama-3.1-8B-
|
| 187 |
tokenizer = AutoTokenizer.from_pretrained("meta-llama/Llama-3.1-8B-Instruct")
|
| 188 |
File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/transformers/models/auto/tokenization_auto.py", line 658, in from_pretrained
|
| 189 |
config = PreTrainedConfig.from_pretrained(pretrained_model_name_or_path, **kwargs)
|
|
|
|
| 49 |
~~~~~~~~~~~~~~~~~~~^^^^^^^^^^
|
| 50 |
File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/huggingface_hub/utils/_http.py", line 802, in hf_raise_for_status
|
| 51 |
raise _format(HfHubHTTPError, message, response) from e
|
| 52 |
+
huggingface_hub.errors.HfHubHTTPError: (Request ID: Root=1-69aa8146-42d3c3697eda24dd75dcca8b;881239d1-057b-4c47-8c66-a1ed79ac2db1)
|
| 53 |
|
| 54 |
403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository..
|
| 55 |
Cannot access content at: https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json.
|
|
|
|
| 144 |
~~~~~~~~~~~~~~~~~~~^^^^^^^^^^
|
| 145 |
File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/huggingface_hub/utils/_http.py", line 802, in hf_raise_for_status
|
| 146 |
raise _format(HfHubHTTPError, message, response) from e
|
| 147 |
+
huggingface_hub.errors.HfHubHTTPError: (Request ID: Root=1-69aa8146-3c10f37735b2665c6e14e975;390b2f45-901b-4706-8ab9-c8ad4c67bdea)
|
| 148 |
|
| 149 |
403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository..
|
| 150 |
Cannot access content at: https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json.
|
|
|
|
| 183 |
The above exception was the direct cause of the following exception:
|
| 184 |
|
| 185 |
Traceback (most recent call last):
|
| 186 |
+
File "/tmp/meta-llama_Llama-3.1-8B-Instruct_2rBbsIp.py", line 26, in <module>
|
| 187 |
tokenizer = AutoTokenizer.from_pretrained("meta-llama/Llama-3.1-8B-Instruct")
|
| 188 |
File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/transformers/models/auto/tokenization_auto.py", line 658, in from_pretrained
|
| 189 |
config = PreTrainedConfig.from_pretrained(pretrained_model_name_or_path, **kwargs)
|