runtime error

.py", line 284, in cached_path output_path = get_from_cache( File "/home/user/.local/lib/python3.10/site-packages/transformers/utils/hub.py", line 495, in get_from_cache _raise_for_status(r) File "/home/user/.local/lib/python3.10/site-packages/transformers/utils/hub.py", line 417, in _raise_for_status raise RepositoryNotFoundError( transformers.utils.hub.RepositoryNotFoundError: 401 Client Error: Repository not found for url: https://huggingface.co/Kurkur99/modeling/resolve/main/config.json. If the repo is private, make sure you are authenticated. During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/user/app/app.py", line 53, in <module> main() File "/home/user/app/app.py", line 40, in main model, tokenizer = load_model(model_directory) File "/home/user/app/app.py", line 7, in load_model model = AutoModelForSequenceClassification.from_pretrained(model_directory) File "/home/user/.local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 423, in from_pretrained config, kwargs = AutoConfig.from_pretrained( File "/home/user/.local/lib/python3.10/site-packages/transformers/models/auto/configuration_auto.py", line 705, in from_pretrained config_dict, _ = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) File "/home/user/.local/lib/python3.10/site-packages/transformers/configuration_utils.py", line 553, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File "/home/user/.local/lib/python3.10/site-packages/transformers/configuration_utils.py", line 613, in _get_config_dict raise EnvironmentError( OSError: Kurkur99/modeling is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo with `use_auth_token` or log in with `huggingface-cli login` and pass `use_auth_token=True`.

Container logs:

Fetching error logs...