runtime error

/site-packages/gradio/helpers.py", line 277, in create await self.cache() File "/home/user/.local/lib/python3.10/site-packages/gradio/helpers.py", line 338, in cache prediction = await Context.root_block.process_api( File "/home/user/.local/lib/python3.10/site-packages/gradio/blocks.py", line 1554, in process_api result = await self.call_function( File "/home/user/.local/lib/python3.10/site-packages/gradio/blocks.py", line 1192, in call_function prediction = await anyio.to_thread.run_sync( File "/home/user/.local/lib/python3.10/site-packages/anyio/to_thread.py", line 56, in run_sync return await get_async_backend().run_sync_in_worker_thread( File "/home/user/.local/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 2134, in run_sync_in_worker_thread return await future File "/home/user/.local/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 851, in run result = context.run(func, *args) File "/home/user/.local/lib/python3.10/site-packages/gradio/utils.py", line 659, in wrapper response = f(*args, **kwargs) File "/home/user/.local/lib/python3.10/site-packages/gradio/external.py", line 421, in query_huggingface_api raise Error( gradio.exceptions.Error: "Could not complete request to HuggingFace API, Status Code: 500, Error: unknown error, Warnings: ['CUDA out of memory. Tried to allocate 128.00 MiB (GPU 0; 14.75 GiB total capacity; 1.72 GiB already allocated; 121.06 MiB free; 1.82 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF', 'There was an inference error: CUDA out of memory. Tried to allocate 128.00 MiB (GPU 0; 14.75 GiB total capacity; 1.72 GiB already allocated; 121.06 MiB free; 1.82 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF']"

Container logs:

Fetching error logs...