From cee34dcb0c4e9fe24fdf86fd870a4f70cd67573f Mon Sep 17 00:00:00 2001 From: github-actions <${GITHUB_ACTOR}@users.noreply.github.com> Date: Thu, 2 Mar 2023 02:17:46 +0000 Subject: [PATCH] Format Python code with psf/black push --- models/index_model.py | 16 +++++++++++----- models/search_model.py | 15 ++++++++++----- 2 files changed, 21 insertions(+), 10 deletions(-) diff --git a/models/index_model.py b/models/index_model.py index d1ce25d..9e82f81 100644 --- a/models/index_model.py +++ b/models/index_model.py @@ -615,7 +615,8 @@ class Index_handler: ), ) total_usage_price = await self.usage_service.get_price( - llm_predictor_mock.last_token_usage, chatgpt=False, # TODO Enable again when tree indexes are fixed + llm_predictor_mock.last_token_usage, + chatgpt=False, # TODO Enable again when tree indexes are fixed ) + await self.usage_service.get_price( embedding_model_mock.last_token_usage, embeddings=True ) @@ -625,7 +626,9 @@ class Index_handler: "Doing this deep search would be prohibitively expensive. Please try a narrower search scope." ) - llm_predictor_temp_non_cgpt = LLMPredictor(llm=OpenAI(model_name="text-davinci-003")) # TODO Get rid of this + llm_predictor_temp_non_cgpt = LLMPredictor( + llm=OpenAI(model_name="text-davinci-003") + ) # TODO Get rid of this tree_index = await self.loop.run_in_executor( None, @@ -638,7 +641,9 @@ class Index_handler: ), ) - await self.usage_service.update_usage(llm_predictor_temp_non_cgpt.last_token_usage, chatgpt=False) # Todo set to false + await self.usage_service.update_usage( + llm_predictor_temp_non_cgpt.last_token_usage, chatgpt=False + ) # Todo set to false await self.usage_service.update_usage( embedding_model.last_token_usage, embeddings=True ) @@ -748,7 +753,6 @@ class Index_handler: ) try: - embedding_model = OpenAIEmbedding() embedding_model.last_token_usage = 0 response = await self.loop.run_in_executor( @@ -766,7 +770,9 @@ class Index_handler: ), ) print("The last token usage was ", llm_predictor.last_token_usage) - await self.usage_service.update_usage(llm_predictor.last_token_usage, chatgpt=True) + await self.usage_service.update_usage( + llm_predictor.last_token_usage, chatgpt=True + ) await self.usage_service.update_usage( embedding_model.last_token_usage, embeddings=True ) diff --git a/models/search_model.py b/models/search_model.py index 5003bc5..ad29a0d 100644 --- a/models/search_model.py +++ b/models/search_model.py @@ -400,15 +400,18 @@ class Search: ) total_usage_price = await self.usage_service.get_price( - llm_predictor_deep.last_token_usage, chatgpt=True, + llm_predictor_deep.last_token_usage, + chatgpt=True, ) + await self.usage_service.get_price( - embedding_model.last_token_usage, embeddings=True) + embedding_model.last_token_usage, embeddings=True + ) await self.usage_service.update_usage( embedding_model.last_token_usage, embeddings=True ) await self.usage_service.update_usage( - llm_predictor_deep.last_token_usage, chatgpt=True, + llm_predictor_deep.last_token_usage, + chatgpt=True, ) price += total_usage_price @@ -451,7 +454,7 @@ class Search: partial( index.query, query, - embedding_mode='hybrid', + embedding_mode="hybrid", llm_predictor=llm_predictor, include_text=True, embed_model=embedding_model, @@ -461,7 +464,9 @@ class Search: ), ) - await self.usage_service.update_usage(llm_predictor.last_token_usage, chatgpt=True) + await self.usage_service.update_usage( + llm_predictor.last_token_usage, chatgpt=True + ) await self.usage_service.update_usage( embedding_model.last_token_usage, embeddings=True )