ariG23498 HF Staff commited on
Commit
ff3372d
·
verified ·
1 Parent(s): 5c8447d

Upload utter-project_EuroLLM-9B_1.txt with huggingface_hub

Browse files
Files changed (1) hide show
  1. utter-project_EuroLLM-9B_1.txt +18 -32
utter-project_EuroLLM-9B_1.txt CHANGED
@@ -7,28 +7,7 @@ pipe = pipeline("text-generation", model="utter-project/EuroLLM-9B")
7
 
8
  ERROR:
9
  Traceback (most recent call last):
10
- File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/transformers/tokenization_utils_base.py", line 2343, in _from_pretrained
11
- tokenizer = cls(*init_inputs, **init_kwargs)
12
- File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/transformers/models/llama/tokenization_llama_fast.py", line 154, in __init__
13
- super().__init__(
14
- ~~~~~~~~~~~~~~~~^
15
- vocab_file=vocab_file,
16
- ^^^^^^^^^^^^^^^^^^^^^^
17
- ...<10 lines>...
18
- **kwargs,
19
- ^^^^^^^^^
20
- )
21
- ^
22
- File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/transformers/tokenization_utils_fast.py", line 108, in __init__
23
- raise ValueError(
24
- ...<2 lines>...
25
- )
26
- ValueError: Cannot instantiate this tokenizer from a slow version. If it's based on sentencepiece, make sure you have sentencepiece installed.
27
-
28
- During handling of the above exception, another exception occurred:
29
-
30
- Traceback (most recent call last):
31
- File "/tmp/utter-project_EuroLLM-9B_1DAHdsm.py", line 19, in <module>
32
  pipe = pipeline("text-generation", model="utter-project/EuroLLM-9B")
33
  File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/transformers/pipelines/__init__.py", line 1078, in pipeline
34
  raise e
@@ -49,13 +28,20 @@ Traceback (most recent call last):
49
  ^^^^^^^^^
50
  )
51
  ^
52
- File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/transformers/tokenization_utils_base.py", line 2344, in _from_pretrained
53
- except import_protobuf_decode_error():
54
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^
55
- File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/transformers/tokenization_utils_base.py", line 87, in import_protobuf_decode_error
56
- raise ImportError(PROTOBUF_IMPORT_ERROR.format(error_message))
57
- ImportError:
58
- requires the protobuf library but it was not found in your environment. Check out the instructions on the
59
- installation page of its repo: https://github.com/protocolbuffers/protobuf/tree/master/python#installation and follow the ones
60
- that match your environment. Please note that you may need to restart your runtime after installation.
61
-
 
 
 
 
 
 
 
 
7
 
8
  ERROR:
9
  Traceback (most recent call last):
10
+ File "/tmp/utter-project_EuroLLM-9B_1xfZBeJ.py", line 24, in <module>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  pipe = pipeline("text-generation", model="utter-project/EuroLLM-9B")
12
  File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/transformers/pipelines/__init__.py", line 1078, in pipeline
13
  raise e
 
28
  ^^^^^^^^^
29
  )
30
  ^
31
+ File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/transformers/tokenization_utils_base.py", line 2343, in _from_pretrained
32
+ tokenizer = cls(*init_inputs, **init_kwargs)
33
+ File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/transformers/models/llama/tokenization_llama_fast.py", line 154, in __init__
34
+ super().__init__(
35
+ ~~~~~~~~~~~~~~~~^
36
+ vocab_file=vocab_file,
37
+ ^^^^^^^^^^^^^^^^^^^^^^
38
+ ...<10 lines>...
39
+ **kwargs,
40
+ ^^^^^^^^^
41
+ )
42
+ ^
43
+ File "/tmp/.cache/uv/environments-v2/ab60391da2086f05/lib/python3.13/site-packages/transformers/tokenization_utils_fast.py", line 108, in __init__
44
+ raise ValueError(
45
+ ...<2 lines>...
46
+ )
47
+ ValueError: Cannot instantiate this tokenizer from a slow version. If it's based on sentencepiece, make sure you have sentencepiece installed.