runtime error

Exit code: 1. Reason: 00, 438kB/s] config.json: 0%| | 0.00/612 [00:00<?, ?B/s] config.json: 100%|██████████| 612/612 [00:00<00:00, 6.74MB/s] model.safetensors: 0%| | 0.00/90.9M [00:00<?, ?B/s] model.safetensors: 100%|██████████| 90.9M/90.9M [00:00<00:00, 127MB/s] tokenizer_config.json: 0%| | 0.00/350 [00:00<?, ?B/s] tokenizer_config.json: 100%|██████████| 350/350 [00:00<00:00, 2.77MB/s] vocab.txt: 0%| | 0.00/232k [00:00<?, ?B/s] vocab.txt: 100%|██████████| 232k/232k [00:00<00:00, 50.5MB/s] tokenizer.json: 0%| | 0.00/466k [00:00<?, ?B/s] tokenizer.json: 100%|██████████| 466k/466k [00:00<00:00, 101MB/s] special_tokens_map.json: 0%| | 0.00/112 [00:00<?, ?B/s] special_tokens_map.json: 100%|██████████| 112/112 [00:00<00:00, 868kB/s] /usr/local/lib/python3.10/site-packages/transformers/tokenization_utils_base.py:1617: FutureWarning: `clean_up_tokenization_spaces` was not set. It will be set to `True` by default. This behavior will be deprecated in transformers v4.45, and will be then set to `False` by default. For more details check this issue: https://github.com/huggingface/transformers/issues/31884 warnings.warn( config.json: 0%| | 0.00/190 [00:00<?, ?B/s] config.json: 100%|██████████| 190/190 [00:00<00:00, 1.06MB/s] Traceback (most recent call last): File "/app/app.py", line 35, in <module> chunks = json.loads(hf_download("chunks.json")) File "/usr/local/lib/python3.10/json/__init__.py", line 346, in loads return _default_decoder.decode(s) File "/usr/local/lib/python3.10/json/decoder.py", line 337, in decode obj, end = self.raw_decode(s, idx=_w(s, 0).end()) File "/usr/local/lib/python3.10/json/decoder.py", line 355, in raw_decode raise JSONDecodeError("Expecting value", s, err.value) from None json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0)

Container logs:

Fetching error logs...