From cd88cc163bb6ae0f2a185e26a7a1140bfebff2ac Mon Sep 17 00:00:00 2001 From: Rene Teigen Date: Thu, 9 Mar 2023 07:23:12 +0000 Subject: [PATCH 1/3] Fix broken per user api key on query and compose --- models/index_model.py | 33 ++++++++++++++++++++------------- 1 file changed, 20 insertions(+), 13 deletions(-) diff --git a/models/index_model.py b/models/index_model.py index c1fcbe7..63b855d 100644 --- a/models/index_model.py +++ b/models/index_model.py @@ -166,13 +166,6 @@ class Index_handler: def __init__(self, bot, usage_service): self.bot = bot self.openai_key = os.getenv("OPENAI_TOKEN") - self.llm_predictor = LLMPredictor( - llm=OpenAIChat( - temperature=0, - model_name="gpt-3.5-turbo", - openai_api_key=self.openai_key, - ) - ) self.index_storage = defaultdict(IndexData) self.loop = asyncio.get_running_loop() self.usage_service = usage_service @@ -752,6 +745,13 @@ class Index_handler: ) index_objects.append(index) + llm_predictor = LLMPredictor( + llm=OpenAIChat( + temperature=0, + model_name="gpt-3.5-turbo" + ) + ) + # For each index object, add its documents to a GPTTreeIndex if deep_compose: documents = [] @@ -790,14 +790,14 @@ class Index_handler: partial( GPTTreeIndex, documents=documents, - llm_predictor=self.llm_predictor, + llm_predictor=llm_predictor, embed_model=embedding_model, use_async=True, ), ) await self.usage_service.update_usage( - self.llm_predictor.last_token_usage, chatgpt=True + llm_predictor.last_token_usage, chatgpt=True ) await self.usage_service.update_usage( embedding_model.last_token_usage, embeddings=True @@ -914,6 +914,13 @@ class Index_handler: else: os.environ["OPENAI_API_KEY"] = user_api_key + llm_predictor = LLMPredictor( + llm=OpenAIChat( + temperature=0, + model_name="gpt-3.5-turbo" + ) + ) + ctx_response = await ctx.respond( embed=EmbedStatics.build_index_query_progress_embed(query) ) @@ -930,14 +937,14 @@ class Index_handler: query, response_mode, nodes, - self.llm_predictor, + llm_predictor, embedding_model, child_branch_factor, ), ) - print("The last token usage was ", self.llm_predictor.last_token_usage) + print("The last token usage was ", llm_predictor.last_token_usage) await self.usage_service.update_usage( - self.llm_predictor.last_token_usage, chatgpt=True + llm_predictor.last_token_usage, chatgpt=True ) await self.usage_service.update_usage( embedding_model.last_token_usage, embeddings=True @@ -946,7 +953,7 @@ class Index_handler: try: total_price = round( await self.usage_service.get_price( - self.llm_predictor.last_token_usage, chatgpt=True + llm_predictor.last_token_usage, chatgpt=True ) + await self.usage_service.get_price( embedding_model.last_token_usage, embeddings=True From 7bac650914f06777b743f47e8fef46c03bcbea6b Mon Sep 17 00:00:00 2001 From: Rene Teigen Date: Thu, 9 Mar 2023 07:26:08 +0000 Subject: [PATCH 2/3] Move tqdm from optional to required Add tqdm to requirements.txt --- pyproject.toml | 4 ++-- requirements.txt | 1 + requirements_base.txt | 3 ++- requirements_full.txt | 1 - 4 files changed, 5 insertions(+), 4 deletions(-) diff --git a/pyproject.toml b/pyproject.toml index 7b40f81..897d4fc 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -39,7 +39,8 @@ dependencies = [ "protobuf==3.20.2", "python-pptx==0.6.21", "langchain==0.0.105", -"unidecode==1.3.6" +"unidecode==1.3.6", +"tqdm==4.64.1" ] dynamic = ["version"] @@ -58,7 +59,6 @@ full = [ "torch==1.9.1", "torchvision==1.10.1", "tokenizers==0.10.3", - "tqdm==4.64.1", "numpy==1.24.2", "scipy==1.10.1", "nltk==3.8.1", diff --git a/requirements.txt b/requirements.txt index ade1dee..40901fc 100644 --- a/requirements.txt +++ b/requirements.txt @@ -22,3 +22,4 @@ sentence-transformers==2.2.2 langchain==0.0.105 openai-whisper unidecode==1.3.6 +tqdm==4.64.1 diff --git a/requirements_base.txt b/requirements_base.txt index 69539da..c98ce4e 100644 --- a/requirements_base.txt +++ b/requirements_base.txt @@ -19,4 +19,5 @@ sentencepiece==0.1.97 protobuf==3.20.2 python-pptx==0.6.21 langchain==0.0.105 -unidecode==1.3.6 \ No newline at end of file +unidecode==1.3.6 +tqdm==4.64.1 \ No newline at end of file diff --git a/requirements_full.txt b/requirements_full.txt index c6dfb55..34b7adc 100644 --- a/requirements_full.txt +++ b/requirements_full.txt @@ -1,5 +1,4 @@ tokenizers==0.13.2 -tqdm==4.64.1 numpy==1.24.2 scipy==1.10.1 nltk==3.8.1 From f059b323d285b749d89985acfca8fa3f2fdd8c0c Mon Sep 17 00:00:00 2001 From: github-actions <${GITHUB_ACTOR}@users.noreply.github.com> Date: Sun, 12 Mar 2023 02:54:06 +0000 Subject: [PATCH 3/3] Format Python code with psf/black push --- models/index_model.py | 10 ++-------- 1 file changed, 2 insertions(+), 8 deletions(-) diff --git a/models/index_model.py b/models/index_model.py index 1975263..fc59f0c 100644 --- a/models/index_model.py +++ b/models/index_model.py @@ -747,10 +747,7 @@ class Index_handler: index_objects.append(index) llm_predictor = LLMPredictor( - llm=OpenAIChat( - temperature=0, - model_name="gpt-3.5-turbo" - ) + llm=OpenAIChat(temperature=0, model_name="gpt-3.5-turbo") ) # For each index object, add its documents to a GPTTreeIndex @@ -916,10 +913,7 @@ class Index_handler: os.environ["OPENAI_API_KEY"] = user_api_key llm_predictor = LLMPredictor( - llm=OpenAIChat( - temperature=0, - model_name="gpt-3.5-turbo" - ) + llm=OpenAIChat(temperature=0, model_name="gpt-3.5-turbo") ) ctx_response = await ctx.respond(