Could not locate the configuration_RW.py inside tiiuae/falcon-40b-instruct.

#114
by cosmino - opened

I am trying to use a different version of falcon-40b

this is the code:

 model_config = AutoConfig.from_pretrained('ichitaka/falcon-40b-instruct-8bit', trust_remote_code=True)
 model_config.max_seq_len = self.model_max_length

model = AutoModelForCausalLM.from_pretrained(
    ichitaka/falcon-40b-instruct-8bit,
    trust_remote_code=True,
    config=model_config,
    quantization_config=self.bnb_config,
    torch_dtype=torch.bfloat16,
    device_map='auto'

)
model.eval() # Set the model to evaluation mode

tokenizer = AutoTokenizer.from_pretrained(
    'ichitaka/falcon-40b-instruct-8bit', padding_side="left", truncation_side="left",
    model_max_length=self.model_max_length
)
# Most LLMs don't have a pad token by default
tokenizer.pad_token = tokenizer.eos_token  

And the fist line returned -> tiiuae/falcon-40b-instruct does not appear to have a file named configuration_RW.py. Checkout 'https://huggingface.co/tiiuae/falcon-40b-instruct/main' for available files.
Output is truncated. View as a scrollable element or open in a text editor. Adjust cell output settings

I also checked this thread -> https://huggingface.co/tiiuae/falcon-7b/discussions/60 but it didn't work

Thanks for your help

Sign up or log in to comment