ariG23498 HF Staff commited on
Commit
e573e51
·
verified ·
1 Parent(s): 4f58605

Upload utter-project_EuroLLM-9B_1.txt with huggingface_hub

Browse files
Files changed (1) hide show
  1. utter-project_EuroLLM-9B_1.txt +40 -82
utter-project_EuroLLM-9B_1.txt CHANGED
@@ -7,97 +7,55 @@ pipe = pipeline("text-generation", model="utter-project/EuroLLM-9B")
7
 
8
  ERROR:
9
  Traceback (most recent call last):
10
- File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/huggingface_hub/utils/_http.py", line 402, in hf_raise_for_status
11
- response.raise_for_status()
12
- ~~~~~~~~~~~~~~~~~~~~~~~~~^^
13
- File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/requests/models.py", line 1026, in raise_for_status
14
- raise HTTPError(http_error_msg, response=self)
15
- requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://huggingface.co/utter-project/EuroLLM-9B/resolve/main/config.json
16
-
17
- The above exception was the direct cause of the following exception:
18
-
19
- Traceback (most recent call last):
20
- File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/transformers/utils/hub.py", line 479, in cached_files
21
- hf_hub_download(
22
- ~~~~~~~~~~~~~~~^
23
- path_or_repo_id,
24
- ^^^^^^^^^^^^^^^^
25
  ...<10 lines>...
26
- local_files_only=local_files_only,
27
- ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
28
  )
29
  ^
30
- File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn
31
- return fn(*args, **kwargs)
32
- File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1007, in hf_hub_download
33
- return _hf_hub_download_to_cache_dir(
34
- # Destination
35
- ...<14 lines>...
36
- force_download=force_download,
37
- )
38
- File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1114, in _hf_hub_download_to_cache_dir
39
- _raise_on_head_call_error(head_call_error, force_download, local_files_only)
40
- ~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
41
- File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1655, in _raise_on_head_call_error
42
- raise head_call_error
43
- File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1543, in _get_metadata_or_catch_error
44
- metadata = get_hf_file_metadata(
45
- url=url, proxies=proxies, timeout=etag_timeout, headers=headers, token=token, endpoint=endpoint
46
- )
47
- File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn
48
- return fn(*args, **kwargs)
49
- File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1460, in get_hf_file_metadata
50
- r = _request_wrapper(
51
- method="HEAD",
52
- ...<5 lines>...
53
- timeout=timeout,
54
- )
55
- File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 283, in _request_wrapper
56
- response = _request_wrapper(
57
- method=method,
58
  ...<2 lines>...
59
- **params,
60
  )
61
- File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 307, in _request_wrapper
62
- hf_raise_for_status(response)
63
- ~~~~~~~~~~~~~~~~~~~^^^^^^^^^^
64
- File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/huggingface_hub/utils/_http.py", line 419, in hf_raise_for_status
65
- raise _format(GatedRepoError, message, response) from e
66
- huggingface_hub.errors.GatedRepoError: 403 Client Error. (Request ID: Root=1-690316c7-088fe153250430267ff1bd89;f18855bf-7918-4011-bebb-3333c9ac2ae0)
67
-
68
- Cannot access gated repo for url https://huggingface.co/utter-project/EuroLLM-9B/resolve/main/config.json.
69
- Access to model utter-project/EuroLLM-9B is restricted and you are not in the authorized list. Visit https://huggingface.co/utter-project/EuroLLM-9B to ask for access.
70
 
71
- The above exception was the direct cause of the following exception:
72
 
73
  Traceback (most recent call last):
74
- File "/tmp/utter-project_EuroLLM-9B_1ocI9QG.py", line 19, in <module>
75
  pipe = pipeline("text-generation", model="utter-project/EuroLLM-9B")
76
- File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/transformers/pipelines/__init__.py", line 922, in pipeline
77
- config = AutoConfig.from_pretrained(
78
- model, _from_pipeline=task, code_revision=code_revision, **hub_kwargs, **model_kwargs
 
 
79
  )
80
- File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/transformers/models/auto/configuration_auto.py", line 1332, in from_pretrained
81
- config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs)
82
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
83
- File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/transformers/configuration_utils.py", line 662, in get_config_dict
84
- config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
85
- ~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
86
- File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/transformers/configuration_utils.py", line 721, in _get_config_dict
87
- resolved_config_file = cached_file(
88
- pretrained_model_name_or_path,
89
- ...<10 lines>...
90
- _commit_hash=commit_hash,
91
  )
92
- File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/transformers/utils/hub.py", line 322, in cached_file
93
- file = cached_files(path_or_repo_id=path_or_repo_id, filenames=[filename], **kwargs)
94
- File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/transformers/utils/hub.py", line 543, in cached_files
95
- raise OSError(
96
- ...<2 lines>...
97
- ) from e
98
- OSError: You are trying to access a gated repo.
99
- Make sure to have access to it at https://huggingface.co/utter-project/EuroLLM-9B.
100
- 403 Client Error. (Request ID: Root=1-690316c7-088fe153250430267ff1bd89;f18855bf-7918-4011-bebb-3333c9ac2ae0)
 
101
 
102
- Cannot access gated repo for url https://huggingface.co/utter-project/EuroLLM-9B/resolve/main/config.json.
103
- Access to model utter-project/EuroLLM-9B is restricted and you are not in the authorized list. Visit https://huggingface.co/utter-project/EuroLLM-9B to ask for access.
 
7
 
8
  ERROR:
9
  Traceback (most recent call last):
10
+ File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/transformers/tokenization_utils_base.py", line 2343, in _from_pretrained
11
+ tokenizer = cls(*init_inputs, **init_kwargs)
12
+ File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/transformers/models/llama/tokenization_llama_fast.py", line 154, in __init__
13
+ super().__init__(
14
+ ~~~~~~~~~~~~~~~~^
15
+ vocab_file=vocab_file,
16
+ ^^^^^^^^^^^^^^^^^^^^^^
 
 
 
 
 
 
 
 
17
  ...<10 lines>...
18
+ **kwargs,
19
+ ^^^^^^^^^
20
  )
21
  ^
22
+ File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/transformers/tokenization_utils_fast.py", line 108, in __init__
23
+ raise ValueError(
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
24
  ...<2 lines>...
 
25
  )
26
+ ValueError: Cannot instantiate this tokenizer from a slow version. If it's based on sentencepiece, make sure you have sentencepiece installed.
 
 
 
 
 
 
 
 
27
 
28
+ During handling of the above exception, another exception occurred:
29
 
30
  Traceback (most recent call last):
31
+ File "/tmp/utter-project_EuroLLM-9B_1ELY12E.py", line 19, in <module>
32
  pipe = pipeline("text-generation", model="utter-project/EuroLLM-9B")
33
+ File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/transformers/pipelines/__init__.py", line 1078, in pipeline
34
+ raise e
35
+ File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/transformers/pipelines/__init__.py", line 1073, in pipeline
36
+ tokenizer = AutoTokenizer.from_pretrained(
37
+ tokenizer_identifier, use_fast=use_fast, _from_pipeline=task, **hub_kwargs, **tokenizer_kwargs
38
  )
39
+ File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/transformers/models/auto/tokenization_auto.py", line 1140, in from_pretrained
40
+ return tokenizer_class.from_pretrained(pretrained_model_name_or_path, *inputs, **kwargs)
41
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
42
+ File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/transformers/tokenization_utils_base.py", line 2097, in from_pretrained
43
+ return cls._from_pretrained(
44
+ ~~~~~~~~~~~~~~~~~~~~^
45
+ resolved_vocab_files,
46
+ ^^^^^^^^^^^^^^^^^^^^^
47
+ ...<9 lines>...
48
+ **kwargs,
49
+ ^^^^^^^^^
50
  )
51
+ ^
52
+ File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/transformers/tokenization_utils_base.py", line 2344, in _from_pretrained
53
+ except import_protobuf_decode_error():
54
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^
55
+ File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/transformers/tokenization_utils_base.py", line 87, in import_protobuf_decode_error
56
+ raise ImportError(PROTOBUF_IMPORT_ERROR.format(error_message))
57
+ ImportError:
58
+ requires the protobuf library but it was not found in your environment. Check out the instructions on the
59
+ installation page of its repo: https://github.com/protocolbuffers/protobuf/tree/master/python#installation and follow the ones
60
+ that match your environment. Please note that you may need to restart your runtime after installation.
61