Kaveen Kumarasinghe 1 year ago
commit 5db5c84bc4

@ -169,13 +169,6 @@ class Index_handler:
def __init__(self, bot, usage_service): def __init__(self, bot, usage_service):
self.bot = bot self.bot = bot
self.openai_key = os.getenv("OPENAI_TOKEN") self.openai_key = os.getenv("OPENAI_TOKEN")
self.llm_predictor = LLMPredictor(
llm=OpenAIChat(
temperature=0,
model_name="gpt-3.5-turbo",
openai_api_key=self.openai_key,
)
)
self.index_storage = defaultdict(IndexData) self.index_storage = defaultdict(IndexData)
self.loop = asyncio.get_running_loop() self.loop = asyncio.get_running_loop()
self.usage_service = usage_service self.usage_service = usage_service
@ -755,6 +748,10 @@ class Index_handler:
) )
index_objects.append(index) index_objects.append(index)
llm_predictor = LLMPredictor(
llm=OpenAIChat(temperature=0, model_name="gpt-3.5-turbo")
)
# For each index object, add its documents to a GPTTreeIndex # For each index object, add its documents to a GPTTreeIndex
if deep_compose: if deep_compose:
documents = [] documents = []
@ -793,14 +790,14 @@ class Index_handler:
partial( partial(
GPTTreeIndex, GPTTreeIndex,
documents=documents, documents=documents,
llm_predictor=self.llm_predictor, llm_predictor=llm_predictor,
embed_model=embedding_model, embed_model=embedding_model,
use_async=True, use_async=True,
), ),
) )
await self.usage_service.update_usage( await self.usage_service.update_usage(
self.llm_predictor.last_token_usage, chatgpt=True llm_predictor.last_token_usage, chatgpt=True
) )
await self.usage_service.update_usage( await self.usage_service.update_usage(
embedding_model.last_token_usage, embeddings=True embedding_model.last_token_usage, embeddings=True
@ -917,6 +914,10 @@ class Index_handler:
else: else:
os.environ["OPENAI_API_KEY"] = user_api_key os.environ["OPENAI_API_KEY"] = user_api_key
llm_predictor = LLMPredictor(
llm=OpenAIChat(temperature=0, model_name="gpt-3.5-turbo")
)
ctx_response = await ctx.respond( ctx_response = await ctx.respond(
embed=EmbedStatics.build_index_query_progress_embed(query) embed=EmbedStatics.build_index_query_progress_embed(query)
) )
@ -924,16 +925,6 @@ class Index_handler:
try: try:
embedding_model = OpenAIEmbedding() embedding_model = OpenAIEmbedding()
embedding_model.last_token_usage = 0 embedding_model.last_token_usage = 0
# response = await get_and_query(
# ctx.user.id,
# self.index_storage,
# query,
# response_mode,
# nodes,
# self.llm_predictor,
# embedding_model,
# child_branch_factor,
# )
response = await self.loop.run_in_executor( response = await self.loop.run_in_executor(
None, None,
partial( partial(
@ -943,14 +934,15 @@ class Index_handler:
query, query,
response_mode, response_mode,
nodes, nodes,
self.llm_predictor, llm_predictor,
embedding_model, embedding_model,
child_branch_factor, child_branch_factor,
), ),
) )
print("The last token usage was ", self.llm_predictor.last_token_usage) print("The last token usage was ", llm_predictor.last_token_usage)
await self.usage_service.update_usage( await self.usage_service.update_usage(
self.llm_predictor.last_token_usage, chatgpt=True llm_predictor.last_token_usage, chatgpt=True
) )
await self.usage_service.update_usage( await self.usage_service.update_usage(
embedding_model.last_token_usage, embeddings=True embedding_model.last_token_usage, embeddings=True
@ -959,7 +951,7 @@ class Index_handler:
try: try:
total_price = round( total_price = round(
await self.usage_service.get_price( await self.usage_service.get_price(
self.llm_predictor.last_token_usage, chatgpt=True llm_predictor.last_token_usage, chatgpt=True
) )
+ await self.usage_service.get_price( + await self.usage_service.get_price(
embedding_model.last_token_usage, embeddings=True embedding_model.last_token_usage, embeddings=True

@ -39,7 +39,8 @@ dependencies = [
"protobuf==3.20.2", "protobuf==3.20.2",
"python-pptx==0.6.21", "python-pptx==0.6.21",
"langchain==0.0.105", "langchain==0.0.105",
"unidecode==1.3.6" "unidecode==1.3.6",
"tqdm==4.64.1"
] ]
dynamic = ["version"] dynamic = ["version"]
@ -58,7 +59,6 @@ full = [
"torch==1.9.1", "torch==1.9.1",
"torchvision==1.10.1", "torchvision==1.10.1",
"tokenizers==0.10.3", "tokenizers==0.10.3",
"tqdm==4.64.1",
"numpy==1.24.2", "numpy==1.24.2",
"scipy==1.10.1", "scipy==1.10.1",
"nltk==3.8.1", "nltk==3.8.1",

@ -22,3 +22,4 @@ sentence-transformers==2.2.2
langchain==0.0.105 langchain==0.0.105
openai-whisper openai-whisper
unidecode==1.3.6 unidecode==1.3.6
tqdm==4.64.1

@ -19,4 +19,5 @@ sentencepiece==0.1.97
protobuf==3.20.2 protobuf==3.20.2
python-pptx==0.6.21 python-pptx==0.6.21
langchain==0.0.105 langchain==0.0.105
unidecode==1.3.6 unidecode==1.3.6
tqdm==4.64.1

@ -1,5 +1,4 @@
tokenizers==0.13.2 tokenizers==0.13.2
tqdm==4.64.1
numpy==1.24.2 numpy==1.24.2
scipy==1.10.1 scipy==1.10.1
nltk==3.8.1 nltk==3.8.1

Loading…
Cancel
Save