KvrParaskevi commited on
Commit
3c353fd
1 Parent(s): 9775596

Update chatbot_bedrock.py

Browse files

add Hugging Face runnable model wrapper

Files changed (1) hide show
  1. chatbot_bedrock.py +17 -21
chatbot_bedrock.py CHANGED
@@ -6,32 +6,28 @@ import langchain.globals
6
  from transformers import AutoModelForCausalLM, AutoTokenizer
7
  import streamlit as st
8
 
9
- @st.cache_resource
10
- def load_model():
11
- tokenizer = AutoTokenizer.from_pretrained("KvrParaskevi/Hotel-Assistant-Attempt4-Llama-2-7b")
12
- model = AutoModelForCausalLM.from_pretrained("KvrParaskevi/Hotel-Assistant-Attempt4-Llama-2-7b")
13
- return tokenizer,model
14
-
15
 
16
- #Write function to connect to Bedrock
17
- # def demo_chatbot():
18
- # # client = boto3.client('bedrock-runtime')
19
-
20
- # template = """Question: {question}
21
 
22
- # Answer: Let's think step by step."""
23
- # prompt = PromptTemplate(template=template, input_variables=["question"])
24
- # llm=HuggingFaceHub(repo_id="google/flan-t5-xl", model_kwargs={"temperature":1e-10})
25
 
26
- # question = "When was Google founded?"
 
 
27
 
28
- # print(llm_chain.run(question))
29
- # return demo_llm
30
 
31
- #test out the code with the Predicgt method
32
- #return demo_llm.predict(input)
33
- # = demo_chatbot('What is the temperature in Nuremberg today?')
34
- #print(response)
 
35
 
36
  def demo_miny_memory(model):
37
  # llm_data = get_Model(hugging_face_key)
 
6
  from transformers import AutoModelForCausalLM, AutoTokenizer
7
  import streamlit as st
8
 
9
+ class HuggingFaceModelWrapper(Runnable): # Assuming Runnable is the required interface
10
+ def __init__(self, model, tokenizer):
11
+ self.model = model
12
+ self.tokenizer = tokenizer
 
 
13
 
14
+ def run(self, input_text):
15
+ # Convert the input text to tokens
16
+ input_ids = self.tokenizer.encode(input_text, return_tensors="pt")
 
 
17
 
18
+ # Generate a response from the model
19
+ output = self.model.generate(input_ids, max_length=100, num_return_sequences=1)
 
20
 
21
+ # Decode the generated tokens to a string
22
+ response_text = self.tokenizer.decode(output[0], skip_special_tokens=True)
23
+ return response_text
24
 
 
 
25
 
26
+ @st.cache_resource
27
+ def load_model():
28
+ tokenizer = AutoTokenizer.from_pretrained("KvrParaskevi/Hotel-Assistant-Attempt4-Llama-2-7b")
29
+ model = AutoModelForCausalLM.from_pretrained("KvrParaskevi/Hotel-Assistant-Attempt4-Llama-2-7b")
30
+ return tokenizer,model
31
 
32
  def demo_miny_memory(model):
33
  # llm_data = get_Model(hugging_face_key)