ArturG9's picture
Upload 2 files
86010a6 verified
raw
history blame
No virus
3.8 kB
import os
import sys
from langchain.text_splitter import TokenTextSplitter,RecursiveCharacterTextSplitter
from langchain_community.embeddings import HuggingFaceEmbeddings
import torch
from transformers import AutoTokenizer
from langchain.retrievers.document_compressors import LLMChainExtractor
from langchain.text_splitter import RecursiveCharacterTextSplitter
from langchain.document_loaders import TextLoader
from langchain.chains import RetrievalQA, ConversationalRetrievalChain
from langchain.memory import ConversationBufferMemory
from langchain.chains import create_history_aware_retriever, create_retrieval_chain
from langchain.chains.combine_documents import create_stuff_documents_chain
from langchain_core.runnables.history import RunnableWithMessageHistory
from langchain_core.chat_history import BaseChatMessageHistory
from langchain_community.chat_message_histories import ChatMessageHistory
from langchain_core.prompts import ChatPromptTemplate, MessagesPlaceholder
from typing import Callable, Dict, List, Optional, Union
from langchain.vectorstores import Chroma
from langchain_core.callbacks import CallbackManager, StreamingStdOutCallbackHandler
from langchain.document_loaders import PyPDFLoader
from langchain.document_loaders.pdf import PyPDFDirectoryLoader
from langchain_community.llms import llamacpp
store = {}
def get_session_history(session_id: str):
if session_id not in store:
store[session_id] = ChatMessageHistory()
return store[session_id]
def load_documents(data_path):
try:
document_loader = PyPDFDirectoryLoader(data_path)
return document_loader.load()
except Exception as e:
print(f"Error loading documents from {data_path}: {e}")
return None # or handle the error in an appropriate manner
def split_docs(documents, chunk_size, chunk_overlap):
try:
text_splitter = RecursiveCharacterTextSplitter.from_tiktoken_encoder(
chunk_size=chunk_size, chunk_overlap=chunk_overlap,
separators=["\n \n \n", "\n \n", "\n1", "(?<=\. )", " ", ""]
)
docs = text_splitter.split_documents(documents)
return docs
except Exception as e:
print(f"Error splitting documents: {e}")
return [] # or handle the error in an appropriate manner
def chroma_db(docs, embeddings):
try:
vectordb = Chroma.from_documents(
documents=docs, embedding=embeddings, persist_directory="docs/chroma/"
)
return vectordb
except Exception as e:
print(f"Error creating Chroma vector database: {e}")
return None # or handle the error in an appropriate manner
def retriever_from_chroma(vectordb, search_type, k):
retriever = vectordb.as_retriever(search_type=search_type, search_kwargs={"k": k})
return retriever
def history_aware_retriever(llm, retriever, contextualize_q_system_prompt):
try:
contextualize_q_prompt = ChatPromptTemplate.from_messages(
[
("system", contextualize_q_system_prompt),
MessagesPlaceholder("chat_history"),
("human", "{input}"),
]
)
history_aware_retriever = create_history_aware_retriever(llm, retriever, contextualize_q_prompt)
return history_aware_retriever
except Exception as e:
print(f"Error creating history-aware retriever: {e}")
return None # or handle the error in an appropriate manner
def echo(question, history):
ai_message = rag_chain.invoke({"input": question, "chat_history": chat_history})
chat_history.extend([HumanMessage(content=question), ai_message["answer"]])
return ai_message['answer']