t5m_pocet / 1_setting.py
makprgmax
2 files+
47b1c1c
raw
history blame contribute delete
No virus
1.47 kB
# from transformers import MT5Tokenizer, MT5ForConditionalGeneration
# local_model_path = "./t5m_pocet"
# tokenizer = MT5Tokenizer.from_pretrained(local_model_path, legacy=False)
# model = MT5ForConditionalGeneration.from_pretrained(local_model_path)
model_directory = './t5m_pocet/model_directory'
from transformers import MT5ForConditionalGeneration, MT5Tokenizer
import torch
tokenizer = MT5Tokenizer(sp_model_kwargs={"model_file": f"{model_directory}/spiece.model"})
# local_model_path = "./t5m_pocet"
# tokenizer = MT5Tokenizer.from_pretrained(local_model_path, legacy=False)
# model = MT5ForConditionalGeneration.from_pretrained(local_model_path)
# Загрузка токенизатора и модели mT5
# model_name = "google/mt5-small"
# tokenizer = MT5Tokenizer.from_pretrained(model_name)
# model = MT5ForConditionalGeneration.from_pretrained(model_name)
# # Пример входных данных
# context = "Контекст, на основе которого нужно ответить на вопрос."
# question = "Какой вопрос нужно задать?"
# # Форматирование входных данных для модели
# input_text = f"question: {question} context: {context}"
# input_ids = tokenizer(input_text, return_tensors="pt").input_ids
# # Генерация ответа
# outputs = model.generate(input_ids)
# answer = tokenizer.decode(outputs[0], skip_special_tokens=True)
# print(answer)