runtime error

= fn(*data) # type: ignore File "/home/user/.local/lib/python3.10/site-packages/gradio/external_utils.py", line 115, in text_generation_inner return input + client.text_generation(input) File "/home/user/.local/lib/python3.10/site-packages/huggingface_hub/inference/_client.py", line 1582, in text_generation raise_text_generation_error(e) File "/home/user/.local/lib/python3.10/site-packages/huggingface_hub/inference/_text_generation.py", line 539, in raise_text_generation_error raise http_error File "/home/user/.local/lib/python3.10/site-packages/huggingface_hub/inference/_client.py", line 1558, in text_generation bytes_output = self.post(json=payload, model=model, task="text-generation", stream=stream) # type: ignore File "/home/user/.local/lib/python3.10/site-packages/huggingface_hub/inference/_client.py", line 242, in post hf_raise_for_status(response) File "/home/user/.local/lib/python3.10/site-packages/huggingface_hub/utils/_errors.py", line 362, in hf_raise_for_status raise HfHubHTTPError(str(e), response=response) from e huggingface_hub.utils._errors.HfHubHTTPError: 500 Server Error: Internal Server Error for url: https://api-inference.huggingface.co/models/krishnavpai/Autism-Counsellor (Request ID: y8BbK_e7i0aCccdC0Me5J) Could not load model krishnavpai/Autism-Counsellor with any of the following classes: (<class 'transformers.models.falcon.modeling_falcon.FalconForCausalLM'>,). See the original errors: while loading with FalconForCausalLM, an error is thrown: Traceback (most recent call last): File "/src/transformers/src/transformers/pipelines/base.py", line 279, in infer_framework_load_model model = model_class.from_pretrained(model, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/src/transformers/src/transformers/modeling_utils.py", line 3236, in from_pretrained raise EnvironmentError( OSError: krishnavpai/Autism-Counsellor does not appear to have a file named pytorch_model.bin, tf_model.h5, model.ckpt or flax_model.msgpack.

Container logs:

Fetching error logs...