From 4aa20f832f1e1d70a17a9142558fc93b7d2546c8 Mon Sep 17 00:00:00 2001 From: github-actions <${GITHUB_ACTOR}@users.noreply.github.com> Date: Thu, 12 Jan 2023 00:13:20 +0000 Subject: [PATCH] Format Python code with psf/black push --- cogs/gpt_3_commands_and_converser.py | 50 +++++++++++++++++----------- models/user_model.py | 8 ++++- 2 files changed, 38 insertions(+), 20 deletions(-) diff --git a/cogs/gpt_3_commands_and_converser.py b/cogs/gpt_3_commands_and_converser.py index dfc85ea..dd64c59 100644 --- a/cogs/gpt_3_commands_and_converser.py +++ b/cogs/gpt_3_commands_and_converser.py @@ -805,9 +805,9 @@ class GPT3ComCon(discord.Cog, name="GPT3ComCon"): ] ) - #set conversation overrides + # set conversation overrides overrides = self.conversation_threads[message.channel.id].get_overrides() - + await self.encapsulated_send( message.channel.id, primary_prompt, @@ -1340,36 +1340,48 @@ class GPT3ComCon(discord.Cog, name="GPT3ComCon"): user_id_normalized = ctx.author.id if not opener_file: pass - else: + else: if not opener_file.endswith((".txt", ".json")): - opener_file = None # Just start a regular thread if the file fails to load + opener_file = ( + None # Just start a regular thread if the file fails to load + ) else: # Load the file and read it into opener try: - opener_file = re.sub(".+(?=[\\//])", "", opener_file) # remove paths from the opener file + opener_file = re.sub( + ".+(?=[\\//])", "", opener_file + ) # remove paths from the opener file opener_file = EnvService.find_shared_file( f"openers{separator}{opener_file}" ) opener_file = await self.load_file(opener_file, ctx) - try: # Try opening as json, if it fails it'll just pass the whole txt or json to the opener + try: # Try opening as json, if it fails it'll just pass the whole txt or json to the opener opener_file = json.loads(opener_file) - temperature=opener_file.get("temperature", None) - top_p=opener_file.get("top_p", None) - frequency_penalty=opener_file.get("frequency_penalty", None) - presence_penalty=opener_file.get("presence_penalty", None) - self.conversation_threads[thread.id].set_overrides(temperature, top_p, frequency_penalty, presence_penalty) - if not opener: # if we only use opener_file then only pass on opener_file for the opening prompt - opener = opener_file.get('text', "error getting text") + temperature = opener_file.get("temperature", None) + top_p = opener_file.get("top_p", None) + frequency_penalty = opener_file.get( + "frequency_penalty", None + ) + presence_penalty = opener_file.get("presence_penalty", None) + self.conversation_threads[thread.id].set_overrides( + temperature, top_p, frequency_penalty, presence_penalty + ) + if ( + not opener + ): # if we only use opener_file then only pass on opener_file for the opening prompt + opener = opener_file.get("text", "error getting text") else: - opener = opener_file.get('text', "error getting text") + opener - except: # Parse as just regular text - if not opener: + opener = ( + opener_file.get("text", "error getting text") + + opener + ) + except: # Parse as just regular text + if not opener: opener = opener_file else: - opener = opener_file + opener + opener = opener_file + opener except: - opener_file = None # Just start a regular thread if the file fails to load - + opener_file = None # Just start a regular thread if the file fails to load # Append the starter text for gpt3 to the user's history so it gets concatenated with the prompt later if minimal or opener_file: diff --git a/models/user_model.py b/models/user_model.py index baf5736..cefb86e 100644 --- a/models/user_model.py +++ b/models/user_model.py @@ -62,7 +62,13 @@ class Thread: self.frequency_penalty = None self.presence_penalty = None - def set_overrides(self, temperature=None,top_p=None,frequency_penalty=None,presence_penalty=None): + def set_overrides( + self, + temperature=None, + top_p=None, + frequency_penalty=None, + presence_penalty=None, + ): self.temperature = temperature self.top_p = top_p self.frequency_penalty = frequency_penalty