kk53 commited on
Commit
0e7a465
β€’
1 Parent(s): f1220a5

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +36 -31
app.py CHANGED
@@ -21,17 +21,26 @@ def load_model():
21
  model = Llama(model_path, embedding=True)
22
 
23
  st.success("Loaded NLP model from Hugging Face!") # πŸ‘ˆ Show a success message
24
-
25
-
26
- # pc = Pinecone(api_key=api_key)
27
- # index = pc.Index("law")
28
- # model_2_name = "TheBloke/zephyr-7B-beta-GGUF"
29
- # model_2base_name = "zephyr-7b-beta.Q4_K_M.gguf"
30
- # model_path_model = hf_hub_download(
31
- # repo_id=model_2_name,
32
- # filename=model_2base_name,
33
- # cache_dir= '/content/models' # Directory for the model
34
- # )
 
 
 
 
 
 
 
 
 
35
  # prompt_template = "<|system|>\
36
  # </s>\
37
  # <|user|>\
@@ -39,26 +48,22 @@ def load_model():
39
  # <|assistant|>"
40
  # template = prompt_template
41
  # prompt = PromptTemplate.from_template(template)
42
- # callback_manager = CallbackManager([StreamingStdOutCallbackHandler()])
43
- # llm = LlamaCpp(
44
- # model_path=model_path_model,
45
- # temperature=0.75,
46
- # max_tokens=2500,
47
- # top_p=1,
48
- # callback_manager=callback_manager,
49
- # verbose=True,
50
- # n_ctx=2048,
51
- # n_threads = 2# Verbose is required to pass to the callback manager
52
- # )
53
- return model
54
 
55
  st.title("Please ask your question on Lithuanian rules for foreigners.")
56
- a = load_model()
 
 
57
  question = st.text_input("Enter your question:")
58
- # if question:
59
- # # Perform Question Answering
60
- # answer = qa_chain(context=context, question=question)
61
-
62
- # # Display the answer
63
- # st.header("Answer:")
64
- # st.write(answer)
 
 
 
 
 
21
  model = Llama(model_path, embedding=True)
22
 
23
  st.success("Loaded NLP model from Hugging Face!") # πŸ‘ˆ Show a success message
24
+
25
+ model_2_name = "TheBloke/zephyr-7B-beta-GGUF"
26
+ model_2base_name = "zephyr-7b-beta.Q4_K_M.gguf"
27
+ model_path_model = hf_hub_download(
28
+ repo_id=model_2_name,
29
+ filename=model_2base_name,
30
+ cache_dir= '/content/models' # Directory for the model
31
+ )
32
+ callback_manager = CallbackManager([StreamingStdOutCallbackHandler()])
33
+ llm = LlamaCpp(
34
+ model_path=model_path_model,
35
+ temperature=0.75,
36
+ max_tokens=2500,
37
+ top_p=1,
38
+ callback_manager=callback_manager,
39
+ verbose=True,
40
+ n_ctx=2048,
41
+ n_threads = 2# Verbose is required to pass to the callback manager
42
+ )
43
+ st.success("loaded the second NLP model from Hugging Face!")
44
  # prompt_template = "<|system|>\
45
  # </s>\
46
  # <|user|>\
 
48
  # <|assistant|>"
49
  # template = prompt_template
50
  # prompt = PromptTemplate.from_template(template)
51
+
52
+ return model, llm
 
 
 
 
 
 
 
 
 
 
53
 
54
  st.title("Please ask your question on Lithuanian rules for foreigners.")
55
+ model,llm = load_model()
56
+ pc = Pinecone(api_key=apikeys)
57
+ index = pc.Index("law")
58
  question = st.text_input("Enter your question:")
59
+ query = model.create_embedding(question)
60
+ q = query['data'][0]['embedding']
61
+ response = index.query(
62
+ vector=q,
63
+ top_k=1,
64
+ include_metadata = True,
65
+ namespace = "ns1"
66
+ )
67
+ response_t = response['matches'][0]['metadata']['text']
68
+ st.header("Answer:")
69
+ st.write(response_t)