runtime error
Exit code: 1. Reason: File "/usr/local/lib/python3.10/site-packages/fsspec/asyn.py", line 103, in sync raise return_result File "/usr/local/lib/python3.10/site-packages/fsspec/asyn.py", line 56, in _runner result[0] = await coro File "/usr/local/lib/python3.10/site-packages/gradio/helpers.py", line 361, in cache prediction = await Context.root_block.process_api( File "/usr/local/lib/python3.10/site-packages/gradio/blocks.py", line 1710, in process_api result = await self.call_function( File "/usr/local/lib/python3.10/site-packages/gradio/blocks.py", line 1250, in call_function prediction = await anyio.to_thread.run_sync( File "/usr/local/lib/python3.10/site-packages/anyio/to_thread.py", line 56, in run_sync return await get_async_backend().run_sync_in_worker_thread( File "/usr/local/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 2441, in run_sync_in_worker_thread return await future File "/usr/local/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 943, in run result = context.run(func, *args) File "/usr/local/lib/python3.10/site-packages/gradio/utils.py", line 693, in wrapper response = f(*args, **kwargs) File "/usr/local/lib/python3.10/site-packages/gradio/external.py", line 357, in query_huggingface_inference_endpoints data = fn(*data) # type: ignore File "/usr/local/lib/python3.10/site-packages/huggingface_hub/inference/_client.py", line 2218, in text_to_image response = self.post(json=payload, model=model, task="text-to-image") File "/usr/local/lib/python3.10/site-packages/huggingface_hub/inference/_client.py", line 305, in post hf_raise_for_status(response) File "/usr/local/lib/python3.10/site-packages/huggingface_hub/utils/_http.py", line 477, in hf_raise_for_status raise _format(HfHubHTTPError, str(e), response) from e huggingface_hub.errors.HfHubHTTPError: 500 Server Error: Internal Server Error for url: https://api-inference.huggingface.co/models/Dipl0/pepe-diffuser (Request ID: YWqyTRUJPGhNJmSNFmsHl) unknown error
Container logs:
Fetching error logs...