Model Loading Error

#40
by kcsham - opened

I have the following code

from transformers import pipeline
pipe = pipeline("feature-extraction", model="nvidia/NV-Embed-v1", trust_remote_code=True)

I got this error seeming loading the model to local cache.

Loading checkpoint shards:  25%|██▌       | 1/4 [00:01<00:03,  1.31s/it]/tmp/tmpwsi3vxvc: line 3: 12435 Killed                  python /mnt/c/Users/kcsha/Projects/learn/nvidia-embedding/embed2.py
ERROR conda.cli.main_run:execute(125): `conda run python /mnt/c/Users/kcsha/Projects/learn/nvidia-embedding/embed2.py` failed. (See above for error)

Any suggestion what could be the issue? Thanks.

Hi, @kcsham . Thanks for reporting the issue. However, we can execute the below code snippets from our side. Can you check your system has enough storage to load the model?

from transformers import pipeline
pipe = pipeline("feature-extraction", model="nvidia/NV-Embed-v1", trust_remote_code=True)

Hi, @kcsham . Thanks for reporting the issue. However, we can execute the below code snippets from our side. Can you check your system has enough storage to load the model?

from transformers import pipeline
pipe = pipeline("feature-extraction", model="nvidia/NV-Embed-v1", trust_remote_code=True)

I was trying to load a locally downloaded version of the model. I got the same error too. I'm not out of space in disk and memory when that happened.

However, I'm trying to load the above in Ubuntu 20.04 in WSL of Windows 11.

Sign up or log in to comment