buster / data /tmp.py
Louis-François Bouchard
Advanced rag course update (#44)
0f06abd unverified
raw
history blame
No virus
2.91 kB
import pandas as pd
import time
import os
from deeplake.core.vectorstore import VectorStore
from langchain.embeddings.openai import OpenAIEmbeddings
import logging
from buster.documents_manager import DeepLakeDocumentsManager
from buster.llm_utils import get_openai_embedding_constructor
# Set the logging level of `httpx` to WARNING or higher to suppress annoying INFO logs
logging.getLogger("httpx").setLevel(logging.WARNING)
openai_embedding_fn = get_openai_embedding_constructor(
client_kwargs={"max_retries": 10}
)
# from openai import OpenAI
DEEPLAKE_DATASET = os.getenv("DEEPLAKE_DATASET", "ai-tutor-dataset")
DEEPLAKE_ORG = os.getenv("DEEPLAKE_ORG", "towards_ai")
df1 = pd.read_csv("./data/langchain.csv") # or 'latin1' or 'cp1252'
df2 = pd.read_csv("./data/hf_transformers.csv")
df3 = pd.read_csv("./data/langchain_course.csv")
df4 = pd.read_csv("./data/filtered_tai_v2.csv")
df5 = pd.read_csv("./data/wiki.csv") # , encoding="ISO-8859-1")
df6 = pd.read_csv("./data/openai.csv")
df7 = pd.read_csv("./data/activeloop.csv")
df8 = pd.read_csv("./data/llm_course.csv")
print(
f"Number of samples: {len(df1)},{len(df2)},{len(df3)},{len(df4)},{len(df5)},{len(df6)},{len(df7)},{len(df8)}"
)
dataset_path = f"hub://{DEEPLAKE_ORG}/{DEEPLAKE_DATASET}"
dm = DeepLakeDocumentsManager(
vector_store_path=dataset_path,
overwrite=True,
)
dm.batch_add(
df=df1,
batch_size=3000,
min_time_interval=60,
num_workers=32,
embedding_fn=openai_embedding_fn,
csv_filename="embeddings.csv",
csv_overwrite=False,
)
dm.batch_add(
df=df2,
batch_size=3000,
min_time_interval=60,
num_workers=32,
embedding_fn=openai_embedding_fn,
csv_filename="embeddings.csv",
csv_overwrite=False,
)
dm.batch_add(
df=df3,
batch_size=3000,
min_time_interval=60,
num_workers=32,
embedding_fn=openai_embedding_fn,
csv_filename="embeddings.csv",
csv_overwrite=False,
)
dm.batch_add(
df=df4,
batch_size=3000,
min_time_interval=60,
num_workers=32,
embedding_fn=openai_embedding_fn,
csv_filename="embeddings.csv",
csv_overwrite=False,
)
dm.batch_add(
df=df5,
batch_size=3000,
min_time_interval=60,
num_workers=32,
embedding_fn=openai_embedding_fn,
csv_filename="embeddings.csv",
csv_overwrite=False,
)
dm.batch_add(
df=df6,
batch_size=3000,
min_time_interval=60,
num_workers=32,
embedding_fn=openai_embedding_fn,
csv_filename="embeddings.csv",
csv_overwrite=False,
)
dm.batch_add(
df=df7,
batch_size=3000,
min_time_interval=60,
num_workers=32,
embedding_fn=openai_embedding_fn,
csv_filename="embeddings.csv",
csv_overwrite=False,
)
dm.batch_add(
df=df8,
batch_size=3000,
min_time_interval=60,
num_workers=32,
embedding_fn=openai_embedding_fn,
csv_filename="embeddings.csv",
csv_overwrite=False,
)