runtime error

9.self_attn.q_proj.weight', 'language_model.model.layers.9.self_attn.v_proj.weight', 'language_model.model.norm.weight'] You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference. You set `add_prefix_space`. The tokenizer needs to be converted from the slow tokenizers Use lora Lora Scaling: 0.03125 Finish use lora Loading weight from MODELS/pllava-7b <All keys matched successfully> Traceback (most recent call last): File "/home/user/app/app.py", line 18, in <module> import tasks.eval.demo.pllava_demo File "/home/user/app/tasks/eval/demo/pllava_demo.py", line 246, in <module> chat = init_model(args) File "/home/user/app/tasks/eval/demo/pllava_demo.py", line 36, in init_model model = model.to('cuda') File "/usr/local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 2576, in to return super().to(*args, **kwargs) File "/usr/local/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1152, in to return self._apply(convert) File "/usr/local/lib/python3.10/site-packages/torch/nn/modules/module.py", line 802, in _apply module._apply(fn) File "/usr/local/lib/python3.10/site-packages/torch/nn/modules/module.py", line 802, in _apply module._apply(fn) File "/usr/local/lib/python3.10/site-packages/torch/nn/modules/module.py", line 802, in _apply module._apply(fn) [Previous line repeated 1 more time] File "/usr/local/lib/python3.10/site-packages/torch/nn/modules/module.py", line 825, in _apply param_applied = fn(param) File "/usr/local/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1150, in convert return t.to(device, dtype if t.is_floating_point() or t.is_complex() else None, non_blocking) File "/usr/local/lib/python3.10/site-packages/torch/cuda/__init__.py", line 302, in _lazy_init torch._C._cuda_init() RuntimeError: Found no NVIDIA driver on your system. Please check that you have an NVIDIA GPU and installed a driver from http://www.nvidia.com/Download/index.aspx

Container logs:

Fetching error logs...