模型的model.py文件是否可以提供?

#3
by lance729 - opened

无论是Github还是HF,好像都没有提供模型的model文件,是否可以提供呢?

https://github.com/huggingface/transformers/blob/main/src/transformers/models/qwen2/modeling_qwen2.py

您好,请问是否有更详细的文档提供模型的细节,还有tokenizer的设置?

在这个文件中,我没有找到对 generation() 方法,我需要知道这个模型的输入形式,以便更好地使用

因为我在这个文件中找到了一个例子:

        >>> from transformers import AutoTokenizer, Qwen2ForCausalLM

        >>> model = Qwen2ForCausalLM.from_pretrained(PATH_TO_CONVERTED_WEIGHTS)
        >>> tokenizer = AutoTokenizer.from_pretrained(PATH_TO_CONVERTED_TOKENIZER)

        >>> prompt = "Hey, are you conscious? Can you talk to me?"
        >>> inputs = tokenizer(prompt, return_tensors="pt")

        >>> # Generate
        >>> generate_ids = model.generate(inputs.input_ids, max_length=30)

却有报错:The attention mask is not set and cannot be inferred from input because pad token is same as eos token. As a consequence, you may observe unexpected behavior. Please pass your input's attention_mask to obtain reliable results.

我可能需要更详细的模型各层的设置,才能使用它

Sign up or log in to comment