ariG23498 HF Staff commited on
Commit
171fcf9
·
verified ·
1 Parent(s): becb73e

Upload speakleash_Bielik-11B-v3.0-Instruct_2.txt with huggingface_hub

Browse files
speakleash_Bielik-11B-v3.0-Instruct_2.txt ADDED
@@ -0,0 +1,117 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ```CODE:
2
+ # Load model directly
3
+ from transformers import AutoTokenizer, AutoModelForCausalLM
4
+
5
+ tokenizer = AutoTokenizer.from_pretrained("speakleash/Bielik-11B-v3.0-Instruct")
6
+ model = AutoModelForCausalLM.from_pretrained("speakleash/Bielik-11B-v3.0-Instruct")
7
+ messages = [
8
+ {"role": "user", "content": "Who are you?"},
9
+ ]
10
+ inputs = tokenizer.apply_chat_template(
11
+ messages,
12
+ add_generation_prompt=True,
13
+ tokenize=True,
14
+ return_dict=True,
15
+ return_tensors="pt",
16
+ ).to(model.device)
17
+
18
+ outputs = model.generate(**inputs, max_new_tokens=40)
19
+ print(tokenizer.decode(outputs[0][inputs["input_ids"].shape[-1]:]))
20
+ ```
21
+
22
+ ERROR:
23
+ Traceback (most recent call last):
24
+ File "/tmp/.cache/uv/environments-v2/8567a30810114812/lib/python3.13/site-packages/huggingface_hub/utils/_http.py", line 402, in hf_raise_for_status
25
+ response.raise_for_status()
26
+ ~~~~~~~~~~~~~~~~~~~~~~~~~^^
27
+ File "/tmp/.cache/uv/environments-v2/8567a30810114812/lib/python3.13/site-packages/requests/models.py", line 1026, in raise_for_status
28
+ raise HTTPError(http_error_msg, response=self)
29
+ requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://huggingface.co/speakleash/Bielik-11B-v3.0-Instruct/resolve/main/config.json
30
+
31
+ The above exception was the direct cause of the following exception:
32
+
33
+ Traceback (most recent call last):
34
+ File "/tmp/.cache/uv/environments-v2/8567a30810114812/lib/python3.13/site-packages/transformers/utils/hub.py", line 479, in cached_files
35
+ hf_hub_download(
36
+ ~~~~~~~~~~~~~~~^
37
+ path_or_repo_id,
38
+ ^^^^^^^^^^^^^^^^
39
+ ...<10 lines>...
40
+ local_files_only=local_files_only,
41
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
42
+ )
43
+ ^
44
+ File "/tmp/.cache/uv/environments-v2/8567a30810114812/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn
45
+ return fn(*args, **kwargs)
46
+ File "/tmp/.cache/uv/environments-v2/8567a30810114812/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1007, in hf_hub_download
47
+ return _hf_hub_download_to_cache_dir(
48
+ # Destination
49
+ ...<14 lines>...
50
+ force_download=force_download,
51
+ )
52
+ File "/tmp/.cache/uv/environments-v2/8567a30810114812/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1114, in _hf_hub_download_to_cache_dir
53
+ _raise_on_head_call_error(head_call_error, force_download, local_files_only)
54
+ ~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
55
+ File "/tmp/.cache/uv/environments-v2/8567a30810114812/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1655, in _raise_on_head_call_error
56
+ raise head_call_error
57
+ File "/tmp/.cache/uv/environments-v2/8567a30810114812/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1543, in _get_metadata_or_catch_error
58
+ metadata = get_hf_file_metadata(
59
+ url=url, proxies=proxies, timeout=etag_timeout, headers=headers, token=token, endpoint=endpoint
60
+ )
61
+ File "/tmp/.cache/uv/environments-v2/8567a30810114812/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn
62
+ return fn(*args, **kwargs)
63
+ File "/tmp/.cache/uv/environments-v2/8567a30810114812/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1460, in get_hf_file_metadata
64
+ r = _request_wrapper(
65
+ method="HEAD",
66
+ ...<5 lines>...
67
+ timeout=timeout,
68
+ )
69
+ File "/tmp/.cache/uv/environments-v2/8567a30810114812/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 283, in _request_wrapper
70
+ response = _request_wrapper(
71
+ method=method,
72
+ ...<2 lines>...
73
+ **params,
74
+ )
75
+ File "/tmp/.cache/uv/environments-v2/8567a30810114812/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 307, in _request_wrapper
76
+ hf_raise_for_status(response)
77
+ ~~~~~~~~~~~~~~~~~~~^^^^^^^^^^
78
+ File "/tmp/.cache/uv/environments-v2/8567a30810114812/lib/python3.13/site-packages/huggingface_hub/utils/_http.py", line 419, in hf_raise_for_status
79
+ raise _format(GatedRepoError, message, response) from e
80
+ huggingface_hub.errors.GatedRepoError: 403 Client Error. (Request ID: Root=1-695cb73e-1ce1be6647f329ca68b8daa5;62d963e2-5ee3-4ad7-b4e2-d3051ccf841c)
81
+
82
+ Cannot access gated repo for url https://huggingface.co/speakleash/Bielik-11B-v3.0-Instruct/resolve/main/config.json.
83
+ Access to model speakleash/Bielik-11B-v3.0-Instruct is restricted and you are not in the authorized list. Visit https://huggingface.co/speakleash/Bielik-11B-v3.0-Instruct to ask for access.
84
+
85
+ The above exception was the direct cause of the following exception:
86
+
87
+ Traceback (most recent call last):
88
+ File "/tmp/speakleash_Bielik-11B-v3.0-Instruct_2jr7rw9.py", line 26, in <module>
89
+ tokenizer = AutoTokenizer.from_pretrained("speakleash/Bielik-11B-v3.0-Instruct")
90
+ File "/tmp/.cache/uv/environments-v2/8567a30810114812/lib/python3.13/site-packages/transformers/models/auto/tokenization_auto.py", line 1109, in from_pretrained
91
+ config = AutoConfig.from_pretrained(
92
+ pretrained_model_name_or_path, trust_remote_code=trust_remote_code, **kwargs
93
+ )
94
+ File "/tmp/.cache/uv/environments-v2/8567a30810114812/lib/python3.13/site-packages/transformers/models/auto/configuration_auto.py", line 1332, in from_pretrained
95
+ config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs)
96
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
97
+ File "/tmp/.cache/uv/environments-v2/8567a30810114812/lib/python3.13/site-packages/transformers/configuration_utils.py", line 662, in get_config_dict
98
+ config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
99
+ ~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
100
+ File "/tmp/.cache/uv/environments-v2/8567a30810114812/lib/python3.13/site-packages/transformers/configuration_utils.py", line 721, in _get_config_dict
101
+ resolved_config_file = cached_file(
102
+ pretrained_model_name_or_path,
103
+ ...<10 lines>...
104
+ _commit_hash=commit_hash,
105
+ )
106
+ File "/tmp/.cache/uv/environments-v2/8567a30810114812/lib/python3.13/site-packages/transformers/utils/hub.py", line 322, in cached_file
107
+ file = cached_files(path_or_repo_id=path_or_repo_id, filenames=[filename], **kwargs)
108
+ File "/tmp/.cache/uv/environments-v2/8567a30810114812/lib/python3.13/site-packages/transformers/utils/hub.py", line 543, in cached_files
109
+ raise OSError(
110
+ ...<2 lines>...
111
+ ) from e
112
+ OSError: You are trying to access a gated repo.
113
+ Make sure to have access to it at https://huggingface.co/speakleash/Bielik-11B-v3.0-Instruct.
114
+ 403 Client Error. (Request ID: Root=1-695cb73e-1ce1be6647f329ca68b8daa5;62d963e2-5ee3-4ad7-b4e2-d3051ccf841c)
115
+
116
+ Cannot access gated repo for url https://huggingface.co/speakleash/Bielik-11B-v3.0-Instruct/resolve/main/config.json.
117
+ Access to model speakleash/Bielik-11B-v3.0-Instruct is restricted and you are not in the authorized list. Visit https://huggingface.co/speakleash/Bielik-11B-v3.0-Instruct to ask for access.