runtime error

Exit code: 3. Reason: /usr/local/lib/python3.10/site-packages/transformers/utils/hub.py:128: FutureWarning: Using `TRANSFORMERS_CACHE` is deprecated and will be removed in v5 of Transformers. Use `HF_HOME` instead. warnings.warn( There was a problem when trying to write in your cache folder (/app/cache). You should set the environment variable TRANSFORMERS_CACHE to a writable directory. INFO: Started server process [1] INFO: Waiting for application startup. Loading checkpoint shards: 0%| | 0/2 [00:00<?, ?it/s] Loading checkpoint shards: 0%| | 0/2 [00:00<?, ?it/s] ERROR: Traceback (most recent call last): File "/usr/local/lib/python3.10/site-packages/starlette/routing.py", line 693, in lifespan async with self.lifespan_context(app) as maybe_state: File "/usr/local/lib/python3.10/site-packages/starlette/routing.py", line 569, in __aenter__ await self._router.startup() File "/usr/local/lib/python3.10/site-packages/starlette/routing.py", line 672, in startup handler() File "/app/main.py", line 23, in load_model model = AutoModelForCausalLM.from_pretrained(model_path, torch_dtype=torch.float16, cache_dir="./cache") File "/usr/local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 564, in from_pretrained return model_class.from_pretrained( File "/usr/local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 4225, in from_pretrained ) = cls._load_pretrained_model( File "/usr/local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 4706, in _load_pretrained_model state_dict = load_state_dict( File "/usr/local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 555, in load_state_dict with safe_open(checkpoint_file, framework="pt") as f: safetensors_rust.SafetensorError: Error while deserializing header: HeaderTooLarge ERROR: Application startup failed. Exiting.

Container logs:

Fetching error logs...