runtime error

n fn_recursive_set_mem_eff module.set_use_memory_efficient_attention_xformers(valid, attention_op) File "/usr/local/lib/python3.10/site-packages/diffusers/models/attention_processor.py", line 274, in set_use_memory_efficient_attention_xformers raise e File "/usr/local/lib/python3.10/site-packages/diffusers/models/attention_processor.py", line 268, in set_use_memory_efficient_attention_xformers _ = xformers.ops.memory_efficient_attention( File "/usr/local/lib/python3.10/site-packages/xformers/ops/fmha/__init__.py", line 247, in memory_efficient_attention return _memory_efficient_attention( File "/usr/local/lib/python3.10/site-packages/xformers/ops/fmha/__init__.py", line 365, in _memory_efficient_attention return _memory_efficient_attention_forward( File "/usr/local/lib/python3.10/site-packages/xformers/ops/fmha/__init__.py", line 381, in _memory_efficient_attention_forward op = _dispatch_fw(inp, False) File "/usr/local/lib/python3.10/site-packages/xformers/ops/fmha/dispatch.py", line 125, in _dispatch_fw return _run_priority_list( File "/usr/local/lib/python3.10/site-packages/xformers/ops/fmha/dispatch.py", line 65, in _run_priority_list raise NotImplementedError(msg) NotImplementedError: No operator found for `memory_efficient_attention_forward` with inputs: query : shape=(1, 2, 1, 40) (torch.float32) key : shape=(1, 2, 1, 40) (torch.float32) value : shape=(1, 2, 1, 40) (torch.float32) attn_bias : <class 'NoneType'> p : 0.0 `decoderF` is not supported because: device=cpu (supported: {'cuda'}) attn_bias type is <class 'NoneType'> `flshattF@v2.5.6` is not supported because: device=cpu (supported: {'cuda'}) dtype=torch.float32 (supported: {torch.bfloat16, torch.float16}) `cutlassF` is not supported because: device=cpu (supported: {'cuda'}) `smallkF` is not supported because: max(query.shape[-1] != value.shape[-1]) > 32 device=cpu (supported: {'cuda'}) unsupported embed per head: 40

Container logs:

Fetching error logs...