helloworld53 commited on
Commit
d7b6305
1 Parent(s): f1220a5
Files changed (1) hide show
  1. app.py +35 -9
app.py CHANGED
@@ -25,13 +25,25 @@ def load_model():
25
 
26
  # pc = Pinecone(api_key=api_key)
27
  # index = pc.Index("law")
28
- # model_2_name = "TheBloke/zephyr-7B-beta-GGUF"
29
- # model_2base_name = "zephyr-7b-beta.Q4_K_M.gguf"
30
- # model_path_model = hf_hub_download(
31
- # repo_id=model_2_name,
32
- # filename=model_2base_name,
33
- # cache_dir= '/content/models' # Directory for the model
34
- # )
 
 
 
 
 
 
 
 
 
 
 
 
35
  # prompt_template = "<|system|>\
36
  # </s>\
37
  # <|user|>\
@@ -50,11 +62,25 @@ def load_model():
50
  # n_ctx=2048,
51
  # n_threads = 2# Verbose is required to pass to the callback manager
52
  # )
53
- return model
54
 
55
  st.title("Please ask your question on Lithuanian rules for foreigners.")
56
- a = load_model()
 
 
57
  question = st.text_input("Enter your question:")
 
 
 
 
 
 
 
 
 
 
 
 
58
  # if question:
59
  # # Perform Question Answering
60
  # answer = qa_chain(context=context, question=question)
 
25
 
26
  # pc = Pinecone(api_key=api_key)
27
  # index = pc.Index("law")
28
+ model_2_name = "TheBloke/zephyr-7B-beta-GGUF"
29
+ model_2base_name = "zephyr-7b-beta.Q4_K_M.gguf"
30
+ model_path_model = hf_hub_download(
31
+ repo_id=model_2_name,
32
+ filename=model_2base_name,
33
+ cache_dir= '/content/models' # Directory for the model
34
+ )
35
+ callback_manager = CallbackManager([StreamingStdOutCallbackHandler()])
36
+ llm = LlamaCpp(
37
+ model_path=model_path_model,
38
+ temperature=0.75,
39
+ max_tokens=2500,
40
+ top_p=1,
41
+ callback_manager=callback_manager,
42
+ verbose=True,
43
+ n_ctx=2048,
44
+ n_threads = 2# Verbose is required to pass to the callback manager
45
+ )
46
+ st.success("loaded the second NLP model from Hugging Face!")
47
  # prompt_template = "<|system|>\
48
  # </s>\
49
  # <|user|>\
 
62
  # n_ctx=2048,
63
  # n_threads = 2# Verbose is required to pass to the callback manager
64
  # )
65
+ return model, llm
66
 
67
  st.title("Please ask your question on Lithuanian rules for foreigners.")
68
+ model,llm = load_model()
69
+ pc = Pinecone(api_key="003117b0-6caf-4de4-adf9-cc49da6587e6")
70
+ index = pc.Index("law")
71
  question = st.text_input("Enter your question:")
72
+ query = model.create_embedding(question)
73
+ q = query['data'][0]['embedding']
74
+ response = index.query(
75
+ vector=q,
76
+ top_k=1,
77
+ include_metadata = True,
78
+ namespace = "ns1"
79
+ )
80
+ response_t = response['matches'][0]['metadata']['text']
81
+ st.header("Answer:")
82
+ st.write(response_t)
83
+
84
  # if question:
85
  # # Perform Question Answering
86
  # answer = qa_chain(context=context, question=question)