runtime error
Exit code: 1. Reason: x/model.onnx: 100%|██████████| 88.5M/88.5M [00:01<00:00, 62.3MB/s] config.json: 0%| | 0.00/752 [00:00<?, ?B/s][A config.json: 100%|██████████| 752/752 [00:00<00:00, 4.92MB/s] You are using the default legacy behaviour of the <class 'transformers.models.llama.tokenization_llama_fast.LlamaTokenizerFast'>. This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 - if you loaded a llama tokenizer from a GGUF file you can ignore this message. Traceback (most recent call last): File "/app/app.py", line 85, in <module> tokenizer_ds = AutoTokenizer.from_pretrained( File "/usr/local/lib/python3.10/site-packages/transformers/models/auto/tokenization_auto.py", line 1175, in from_pretrained return tokenizer_class_fast.from_pretrained(pretrained_model_name_or_path, *inputs, **kwargs) File "/usr/local/lib/python3.10/site-packages/transformers/tokenization_utils_base.py", line 2113, in from_pretrained return cls._from_pretrained( File "/usr/local/lib/python3.10/site-packages/transformers/tokenization_utils_base.py", line 2359, in _from_pretrained tokenizer = cls(*init_inputs, **init_kwargs) File "/usr/local/lib/python3.10/site-packages/transformers/models/llama/tokenization_llama_fast.py", line 154, in __init__ super().__init__( File "/usr/local/lib/python3.10/site-packages/transformers/tokenization_utils_fast.py", line 139, in __init__ fast_tokenizer = convert_slow_tokenizer(self, from_tiktoken=True) File "/usr/local/lib/python3.10/site-packages/transformers/convert_slow_tokenizer.py", line 1857, in convert_slow_tokenizer elif transformer_tokenizer.vocab_file.endswith("tekken.json"): AttributeError: 'NoneType' object has no attribute 'endswith'
Container logs:
Fetching error logs...