runtime error

Exit code: 1. Reason: tokenizer_config.json: 0%| | 0.00/749 [00:00<?, ?B/s] tokenizer_config.json: 100%|██████████| 749/749 [00:00<00:00, 6.40MB/s] tokenizer.model: 0%| | 0.00/500k [00:00<?, ?B/s] tokenizer.model: 100%|██████████| 500k/500k [00:00<00:00, 63.2MB/s] tokenizer.json: 0%| | 0.00/1.84M [00:00<?, ?B/s] tokenizer.json: 100%|██████████| 1.84M/1.84M [00:00<00:00, 26.5MB/s] added_tokens.json: 0%| | 0.00/21.0 [00:00<?, ?B/s] added_tokens.json: 100%|██████████| 21.0/21.0 [00:00<00:00, 199kB/s] special_tokens_map.json: 0%| | 0.00/438 [00:00<?, ?B/s] special_tokens_map.json: 100%|██████████| 438/438 [00:00<00:00, 4.19MB/s] config.json: 0%| | 0.00/880 [00:00<?, ?B/s] config.json: 100%|██████████| 880/880 [00:00<00:00, 7.35MB/s] Traceback (most recent call last): File "/home/user/app/app.py", line 8, in <module> model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto", torch_dtype=torch.float16) File "/usr/local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 564, in from_pretrained return model_class.from_pretrained( File "/usr/local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 3611, in from_pretrained hf_quantizer = AutoHfQuantizer.from_config( File "/usr/local/lib/python3.10/site-packages/transformers/quantizers/auto.py", line 156, in from_config return target_cls(quantization_config, **kwargs) File "/usr/local/lib/python3.10/site-packages/transformers/quantizers/quantizer_gptq.py", line 49, in __init__ raise ImportError("Loading a GPTQ quantized model requires optimum (`pip install optimum`)") ImportError: Loading a GPTQ quantized model requires optimum (`pip install optimum`)

Container logs:

Fetching error logs...