ariG23498 HF Staff commited on
Commit
5738b72
·
verified ·
1 Parent(s): 4697bc5

Upload utter-project_EuroLLM-9B_2.txt with huggingface_hub

Browse files
Files changed (1) hide show
  1. utter-project_EuroLLM-9B_2.txt +35 -83
utter-project_EuroLLM-9B_2.txt CHANGED
@@ -8,97 +8,49 @@ model = AutoModelForCausalLM.from_pretrained("utter-project/EuroLLM-9B")
8
 
9
  ERROR:
10
  Traceback (most recent call last):
11
- File "/tmp/.cache/uv/environments-v2/31d9dd746308ee4e/lib/python3.13/site-packages/huggingface_hub/utils/_http.py", line 402, in hf_raise_for_status
12
- response.raise_for_status()
13
- ~~~~~~~~~~~~~~~~~~~~~~~~~^^
14
- File "/tmp/.cache/uv/environments-v2/31d9dd746308ee4e/lib/python3.13/site-packages/requests/models.py", line 1026, in raise_for_status
15
- raise HTTPError(http_error_msg, response=self)
16
- requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://huggingface.co/utter-project/EuroLLM-9B/resolve/main/config.json
17
-
18
- The above exception was the direct cause of the following exception:
19
-
20
- Traceback (most recent call last):
21
- File "/tmp/.cache/uv/environments-v2/31d9dd746308ee4e/lib/python3.13/site-packages/transformers/utils/hub.py", line 479, in cached_files
22
- hf_hub_download(
23
- ~~~~~~~~~~~~~~~^
24
- path_or_repo_id,
25
- ^^^^^^^^^^^^^^^^
26
  ...<10 lines>...
27
- local_files_only=local_files_only,
28
- ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
29
  )
30
  ^
31
- File "/tmp/.cache/uv/environments-v2/31d9dd746308ee4e/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn
32
- return fn(*args, **kwargs)
33
- File "/tmp/.cache/uv/environments-v2/31d9dd746308ee4e/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1007, in hf_hub_download
34
- return _hf_hub_download_to_cache_dir(
35
- # Destination
36
- ...<14 lines>...
37
- force_download=force_download,
38
- )
39
- File "/tmp/.cache/uv/environments-v2/31d9dd746308ee4e/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1114, in _hf_hub_download_to_cache_dir
40
- _raise_on_head_call_error(head_call_error, force_download, local_files_only)
41
- ~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
42
- File "/tmp/.cache/uv/environments-v2/31d9dd746308ee4e/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1655, in _raise_on_head_call_error
43
- raise head_call_error
44
- File "/tmp/.cache/uv/environments-v2/31d9dd746308ee4e/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1543, in _get_metadata_or_catch_error
45
- metadata = get_hf_file_metadata(
46
- url=url, proxies=proxies, timeout=etag_timeout, headers=headers, token=token, endpoint=endpoint
47
- )
48
- File "/tmp/.cache/uv/environments-v2/31d9dd746308ee4e/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn
49
- return fn(*args, **kwargs)
50
- File "/tmp/.cache/uv/environments-v2/31d9dd746308ee4e/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1460, in get_hf_file_metadata
51
- r = _request_wrapper(
52
- method="HEAD",
53
- ...<5 lines>...
54
- timeout=timeout,
55
- )
56
- File "/tmp/.cache/uv/environments-v2/31d9dd746308ee4e/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 283, in _request_wrapper
57
- response = _request_wrapper(
58
- method=method,
59
  ...<2 lines>...
60
- **params,
61
  )
62
- File "/tmp/.cache/uv/environments-v2/31d9dd746308ee4e/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 307, in _request_wrapper
63
- hf_raise_for_status(response)
64
- ~~~~~~~~~~~~~~~~~~~^^^^^^^^^^
65
- File "/tmp/.cache/uv/environments-v2/31d9dd746308ee4e/lib/python3.13/site-packages/huggingface_hub/utils/_http.py", line 419, in hf_raise_for_status
66
- raise _format(GatedRepoError, message, response) from e
67
- huggingface_hub.errors.GatedRepoError: 403 Client Error. (Request ID: Root=1-690316c5-3a9588ef231f119f6f90d113;0eae5f0f-7d79-44ee-97a1-8f06ef3c1289)
68
 
69
- Cannot access gated repo for url https://huggingface.co/utter-project/EuroLLM-9B/resolve/main/config.json.
70
- Access to model utter-project/EuroLLM-9B is restricted and you are not in the authorized list. Visit https://huggingface.co/utter-project/EuroLLM-9B to ask for access.
71
-
72
- The above exception was the direct cause of the following exception:
73
 
74
  Traceback (most recent call last):
75
- File "/tmp/utter-project_EuroLLM-9B_2TJlbw3.py", line 19, in <module>
76
  tokenizer = AutoTokenizer.from_pretrained("utter-project/EuroLLM-9B")
77
- File "/tmp/.cache/uv/environments-v2/31d9dd746308ee4e/lib/python3.13/site-packages/transformers/models/auto/tokenization_auto.py", line 1093, in from_pretrained
78
- config = AutoConfig.from_pretrained(
79
- pretrained_model_name_or_path, trust_remote_code=trust_remote_code, **kwargs
80
- )
81
- File "/tmp/.cache/uv/environments-v2/31d9dd746308ee4e/lib/python3.13/site-packages/transformers/models/auto/configuration_auto.py", line 1332, in from_pretrained
82
- config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs)
83
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
84
- File "/tmp/.cache/uv/environments-v2/31d9dd746308ee4e/lib/python3.13/site-packages/transformers/configuration_utils.py", line 662, in get_config_dict
85
- config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
86
- ~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
87
- File "/tmp/.cache/uv/environments-v2/31d9dd746308ee4e/lib/python3.13/site-packages/transformers/configuration_utils.py", line 721, in _get_config_dict
88
- resolved_config_file = cached_file(
89
- pretrained_model_name_or_path,
90
- ...<10 lines>...
91
- _commit_hash=commit_hash,
92
  )
93
- File "/tmp/.cache/uv/environments-v2/31d9dd746308ee4e/lib/python3.13/site-packages/transformers/utils/hub.py", line 322, in cached_file
94
- file = cached_files(path_or_repo_id=path_or_repo_id, filenames=[filename], **kwargs)
95
- File "/tmp/.cache/uv/environments-v2/31d9dd746308ee4e/lib/python3.13/site-packages/transformers/utils/hub.py", line 543, in cached_files
96
- raise OSError(
97
- ...<2 lines>...
98
- ) from e
99
- OSError: You are trying to access a gated repo.
100
- Make sure to have access to it at https://huggingface.co/utter-project/EuroLLM-9B.
101
- 403 Client Error. (Request ID: Root=1-690316c5-3a9588ef231f119f6f90d113;0eae5f0f-7d79-44ee-97a1-8f06ef3c1289)
 
102
 
103
- Cannot access gated repo for url https://huggingface.co/utter-project/EuroLLM-9B/resolve/main/config.json.
104
- Access to model utter-project/EuroLLM-9B is restricted and you are not in the authorized list. Visit https://huggingface.co/utter-project/EuroLLM-9B to ask for access.
 
8
 
9
  ERROR:
10
  Traceback (most recent call last):
11
+ File "/tmp/.cache/uv/environments-v2/31d9dd746308ee4e/lib/python3.13/site-packages/transformers/tokenization_utils_base.py", line 2343, in _from_pretrained
12
+ tokenizer = cls(*init_inputs, **init_kwargs)
13
+ File "/tmp/.cache/uv/environments-v2/31d9dd746308ee4e/lib/python3.13/site-packages/transformers/models/llama/tokenization_llama_fast.py", line 154, in __init__
14
+ super().__init__(
15
+ ~~~~~~~~~~~~~~~~^
16
+ vocab_file=vocab_file,
17
+ ^^^^^^^^^^^^^^^^^^^^^^
 
 
 
 
 
 
 
 
18
  ...<10 lines>...
19
+ **kwargs,
20
+ ^^^^^^^^^
21
  )
22
  ^
23
+ File "/tmp/.cache/uv/environments-v2/31d9dd746308ee4e/lib/python3.13/site-packages/transformers/tokenization_utils_fast.py", line 108, in __init__
24
+ raise ValueError(
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
25
  ...<2 lines>...
 
26
  )
27
+ ValueError: Cannot instantiate this tokenizer from a slow version. If it's based on sentencepiece, make sure you have sentencepiece installed.
 
 
 
 
 
28
 
29
+ During handling of the above exception, another exception occurred:
 
 
 
30
 
31
  Traceback (most recent call last):
32
+ File "/tmp/utter-project_EuroLLM-9B_2AxdS9w.py", line 19, in <module>
33
  tokenizer = AutoTokenizer.from_pretrained("utter-project/EuroLLM-9B")
34
+ File "/tmp/.cache/uv/environments-v2/31d9dd746308ee4e/lib/python3.13/site-packages/transformers/models/auto/tokenization_auto.py", line 1140, in from_pretrained
35
+ return tokenizer_class.from_pretrained(pretrained_model_name_or_path, *inputs, **kwargs)
36
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
37
+ File "/tmp/.cache/uv/environments-v2/31d9dd746308ee4e/lib/python3.13/site-packages/transformers/tokenization_utils_base.py", line 2097, in from_pretrained
38
+ return cls._from_pretrained(
39
+ ~~~~~~~~~~~~~~~~~~~~^
40
+ resolved_vocab_files,
41
+ ^^^^^^^^^^^^^^^^^^^^^
42
+ ...<9 lines>...
43
+ **kwargs,
44
+ ^^^^^^^^^
 
 
 
 
45
  )
46
+ ^
47
+ File "/tmp/.cache/uv/environments-v2/31d9dd746308ee4e/lib/python3.13/site-packages/transformers/tokenization_utils_base.py", line 2344, in _from_pretrained
48
+ except import_protobuf_decode_error():
49
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^
50
+ File "/tmp/.cache/uv/environments-v2/31d9dd746308ee4e/lib/python3.13/site-packages/transformers/tokenization_utils_base.py", line 87, in import_protobuf_decode_error
51
+ raise ImportError(PROTOBUF_IMPORT_ERROR.format(error_message))
52
+ ImportError:
53
+ requires the protobuf library but it was not found in your environment. Check out the instructions on the
54
+ installation page of its repo: https://github.com/protocolbuffers/protobuf/tree/master/python#installation and follow the ones
55
+ that match your environment. Please note that you may need to restart your runtime after installation.
56