runtime error

Exit code: 1. Reason: /usr/local/lib/python3.10/site-packages/auto_gptq/nn_modules/triton_utils/kernels.py:411: FutureWarning: `torch.cuda.amp.custom_fwd(args...)` is deprecated. Please use `torch.amp.custom_fwd(args..., device_type='cuda')` instead. def forward(ctx, input, qweight, scales, qzeros, g_idx, bits, maxq): /usr/local/lib/python3.10/site-packages/auto_gptq/nn_modules/triton_utils/kernels.py:419: FutureWarning: `torch.cuda.amp.custom_bwd(args...)` is deprecated. Please use `torch.amp.custom_bwd(args..., device_type='cuda')` instead. def backward(ctx, grad_output): /usr/local/lib/python3.10/site-packages/auto_gptq/nn_modules/triton_utils/kernels.py:461: FutureWarning: `torch.cuda.amp.custom_fwd(args...)` is deprecated. Please use `torch.amp.custom_fwd(args..., device_type='cuda')` instead. @custom_fwd(cast_inputs=torch.float16) CUDA extension not installed. CUDA extension not installed. Traceback (most recent call last): File "/home/user/app/app.py", line 13, in <module> model = AutoGPTQForCausalLM.from_pretrained( File "/usr/local/lib/python3.10/site-packages/auto_gptq/modeling/auto.py", line 76, in from_pretrained return GPTQ_CAUSAL_LM_MODEL_MAP[model_type].from_pretrained( File "/usr/local/lib/python3.10/site-packages/auto_gptq/modeling/_base.py", line 716, in from_pretrained raise EnvironmentError("Load pretrained model to do quantization requires CUDA available.") OSError: Load pretrained model to do quantization requires CUDA available.

Container logs:

Fetching error logs...