runtime error

Exit code: 1. Reason: The cache for model files in Transformers v4.22.0 has been updated. Migrating your old cache. This is a one-time only operation. You can interrupt this and resume the migration later on by calling `transformers.utils.move_cache()`. 0it [00:00, ?it/s] 0it [00:00, ?it/s] /usr/local/lib/python3.10/site-packages/xformers/ops/fmha/flash.py:211: FutureWarning: `torch.library.impl_abstract` was renamed to `torch.library.register_fake`. Please use that instead; we will remove `torch.library.impl_abstract` in a future version of PyTorch. @torch.library.impl_abstract("xformers_flash::flash_fwd") /usr/local/lib/python3.10/site-packages/xformers/ops/fmha/flash.py:344: FutureWarning: `torch.library.impl_abstract` was renamed to `torch.library.register_fake`. Please use that instead; we will remove `torch.library.impl_abstract` in a future version of PyTorch. @torch.library.impl_abstract("xformers_flash::flash_bwd") /usr/local/lib/python3.10/site-packages/torch/cuda/__init__.py:654: UserWarning: Can't initialize NVML warnings.warn("Can't initialize NVML") The installed version of bitsandbytes was compiled without GPU support. 8-bit optimizers, 8-bit multiplication, and GPU quantization are unavailable. ↑ Those bitsandbytes warnings are expected on ZeroGPU ↑ Traceback (most recent call last): File "/home/user/app/app.py", line 27, in <module> import spaces File "/usr/local/lib/python3.10/site-packages/spaces/__init__.py", line 19, in <module> from .zero.decorator import GPU File "/usr/local/lib/python3.10/site-packages/spaces/zero/__init__.py", line 15, in <module> raise RuntimeError( RuntimeError: CUDA has been initialized before importing the `spaces` package

Container logs:

Fetching error logs...