Format Python code with psf/black push

github-actions 1 year ago
parent fc1a5c5c32
commit 5dcd493c67

@ -64,7 +64,9 @@ class DrawDallEService(discord.Cog, name="DrawDallEService"):
# Error catching for API errors # Error catching for API errors
except aiohttp.ClientResponseError as e: except aiohttp.ClientResponseError as e:
message = f"The API returned an invalid response: **{e.status}: {e.message}**" message = (
f"The API returned an invalid response: **{e.status}: {e.message}**"
)
await ctx.channel.send(message) if not from_context else await ctx.respond( await ctx.channel.send(message) if not from_context else await ctx.respond(
message message
) )
@ -72,7 +74,9 @@ class DrawDallEService(discord.Cog, name="DrawDallEService"):
except ValueError as e: except ValueError as e:
message = f"Error: {e}. Please try again with a different prompt." message = f"Error: {e}. Please try again with a different prompt."
await ctx.channel.send( message )if not from_context else await ctx.respond( message ) await ctx.channel.send(message) if not from_context else await ctx.respond(
message
)
return return

@ -1159,14 +1159,15 @@ class GPT3ComCon(discord.Cog, name="GPT3ComCon"):
# Error catching for AIOHTTP Errors # Error catching for AIOHTTP Errors
except aiohttp.ClientResponseError as e: except aiohttp.ClientResponseError as e:
message = f"The API returned an invalid response: **{e.status}: {e.message}**" message = (
f"The API returned an invalid response: **{e.status}: {e.message}**"
)
if from_context: if from_context:
await ctx.send_followup(message) await ctx.send_followup(message)
else: else:
await ctx.reply(message) await ctx.reply(message)
self.remove_awaiting(ctx.author.id, ctx.channel.id, from_g_command) self.remove_awaiting(ctx.author.id, ctx.channel.id, from_g_command)
# Error catching for OpenAI model value errors # Error catching for OpenAI model value errors
except ValueError as e: except ValueError as e:
if from_context: if from_context:
@ -1175,7 +1176,6 @@ class GPT3ComCon(discord.Cog, name="GPT3ComCon"):
await ctx.reply(e) await ctx.reply(e)
self.remove_awaiting(ctx.author.id, ctx.channel.id, from_g_command) self.remove_awaiting(ctx.author.id, ctx.channel.id, from_g_command)
# General catch case for everything # General catch case for everything
except Exception: except Exception:

@ -344,8 +344,10 @@ class Model:
self._prompt_min_length = value self._prompt_min_length = value
def backoff_handler(details): def backoff_handler(details):
print (f"Backing off {details['wait']:0.1f} seconds after {details['tries']} tries calling function {details['target']} | " print(
f"{details['exception'].status}: {details['exception'].message}") f"Backing off {details['wait']:0.1f} seconds after {details['tries']} tries calling function {details['target']} | "
f"{details['exception'].status}: {details['exception'].message}"
)
async def valid_text_request(self, response): async def valid_text_request(self, response):
try: try:
@ -357,7 +359,14 @@ class Model:
+ str(response["error"]["message"]) + str(response["error"]["message"])
) )
@backoff.on_exception(backoff.expo, aiohttp.ClientResponseError, factor=3, base=5, max_tries=4, on_backoff=backoff_handler) @backoff.on_exception(
backoff.expo,
aiohttp.ClientResponseError,
factor=3,
base=5,
max_tries=4,
on_backoff=backoff_handler,
)
async def send_embedding_request(self, text, custom_api_key=None): async def send_embedding_request(self, text, custom_api_key=None):
async with aiohttp.ClientSession(raise_for_status=True) as session: async with aiohttp.ClientSession(raise_for_status=True) as session:
payload = { payload = {
@ -380,7 +389,14 @@ class Model:
traceback.print_exc() traceback.print_exc()
return return
@backoff.on_exception(backoff.expo, aiohttp.ClientResponseError, factor=3, base=5, max_tries=6, on_backoff=backoff_handler) @backoff.on_exception(
backoff.expo,
aiohttp.ClientResponseError,
factor=3,
base=5,
max_tries=6,
on_backoff=backoff_handler,
)
async def send_moderations_request(self, text): async def send_moderations_request(self, text):
# Use aiohttp to send the above request: # Use aiohttp to send the above request:
async with aiohttp.ClientSession(raise_for_status=True) as session: async with aiohttp.ClientSession(raise_for_status=True) as session:
@ -396,7 +412,14 @@ class Model:
) as response: ) as response:
return await response.json() return await response.json()
@backoff.on_exception(backoff.expo, aiohttp.ClientResponseError, factor=3, base=5, max_tries=4, on_backoff=backoff_handler) @backoff.on_exception(
backoff.expo,
aiohttp.ClientResponseError,
factor=3,
base=5,
max_tries=4,
on_backoff=backoff_handler,
)
async def send_summary_request(self, prompt, custom_api_key=None): async def send_summary_request(self, prompt, custom_api_key=None):
""" """
Sends a summary request to the OpenAI API Sends a summary request to the OpenAI API
@ -438,7 +461,14 @@ class Model:
return response return response
@backoff.on_exception(backoff.expo, aiohttp.ClientResponseError, factor=3, base=5, max_tries=4, on_backoff=backoff_handler) @backoff.on_exception(
backoff.expo,
aiohttp.ClientResponseError,
factor=3,
base=5,
max_tries=4,
on_backoff=backoff_handler,
)
async def send_request( async def send_request(
self, self,
prompt, prompt,
@ -521,7 +551,14 @@ class Model:
return response return response
@backoff.on_exception(backoff.expo, aiohttp.ClientResponseError, factor=3, base=5, max_tries=4, on_backoff=backoff_handler) @backoff.on_exception(
backoff.expo,
aiohttp.ClientResponseError,
factor=3,
base=5,
max_tries=4,
on_backoff=backoff_handler,
)
async def send_image_request( async def send_image_request(
self, ctx, prompt, vary=None, custom_api_key=None self, ctx, prompt, vary=None, custom_api_key=None
) -> tuple[File, list[Any]]: ) -> tuple[File, list[Any]]:

Loading…
Cancel
Save