import asyncio import sys import traceback from pathlib import Path import discord import pinecone from pycord.multicog import apply_multicog import os from models.pinecone_service_model import PineconeService if sys.platform == "win32": separator = "\\" else: separator = "/" from cogs.draw_image_generation import DrawDallEService from cogs.gpt_3_commands_and_converser import GPT3ComCon from cogs.image_prompt_optimizer import ImgPromptOptimizer from models.deletion_service_model import Deletion from models.message_model import Message from models.openai_model import Model from models.usage_service_model import UsageService from models.env_service_model import EnvService __version__ = "5.1.1" """ The pinecone service is used to store and retrieve conversation embeddings. """ try: PINECONE_TOKEN = os.getenv("PINECONE_TOKEN") except: PINECONE_TOKEN = None pinecone_service = None if PINECONE_TOKEN: pinecone.init(api_key=PINECONE_TOKEN, environment="us-west1-gcp") PINECONE_INDEX = "conversation-embeddings" if PINECONE_INDEX not in pinecone.list_indexes(): print("Creating pinecone index. Please wait...") pinecone.create_index( "conversation-embeddings", dimension=1536, metric="dotproduct", pod_type="s1", ) pinecone_service = PineconeService(pinecone.Index(PINECONE_INDEX)) print("Got the pinecone service") """ Message queueing for the debug service, defer debug messages to be sent later so we don't hit rate limits. """ message_queue = asyncio.Queue() deletion_queue = asyncio.Queue() asyncio.ensure_future(Message.process_message_queue(message_queue, 1.5, 5)) asyncio.ensure_future(Deletion.process_deletion_queue(deletion_queue, 1, 1)) """ Settings for the bot """ activity = discord.Activity( type=discord.ActivityType.watching, name="for /help /gpt, and more!" ) bot = discord.Bot(intents=discord.Intents.all(), command_prefix="!", activity=activity) usage_service = UsageService(Path(os.environ.get("DATA_DIR", os.getcwd()))) model = Model(usage_service) """ An encapsulating wrapper for the discord.py client. This uses the old re-write without cogs, but it gets the job done! """ @bot.event # Using self gives u async def on_ready(): # I can make self optional by print("We have logged in as {0.user}".format(bot)) @bot.event async def on_application_command_error( ctx: discord.ApplicationContext, error: discord.DiscordException ): if isinstance(error, discord.CheckFailure): pass else: raise error async def main(): data_path = EnvService.environment_path_with_fallback("DATA_DIR") debug_guild = int(os.getenv("DEBUG_GUILD")) debug_channel = int(os.getenv("DEBUG_CHANNEL")) if not data_path.exists(): raise OSError(f"Data path: {data_path} does not exist ... create it?") # Load the main GPT3 Bot service bot.add_cog( GPT3ComCon( bot, usage_service, model, message_queue, deletion_queue, debug_guild, debug_channel, data_path, pinecone_service=pinecone_service, ) ) bot.add_cog( DrawDallEService( bot, usage_service, model, message_queue, deletion_queue, bot.get_cog("GPT3ComCon"), ) ) bot.add_cog( ImgPromptOptimizer( bot, usage_service, model, message_queue, deletion_queue, bot.get_cog("GPT3ComCon"), bot.get_cog("DrawDallEService"), ) ) apply_multicog(bot) await bot.start(os.getenv("DISCORD_TOKEN")) # Run the bot with a token taken from an environment file. def init(): PID_FILE = "bot.pid" if os.path.exists(PID_FILE): print("Process ID file already exists") sys.exit(1) else: with open(PID_FILE, "w") as f: f.write(str(os.getpid())) print("" "Wrote PID to f" "ile the file " + PID_FILE) f.close() try: asyncio.get_event_loop().run_until_complete(main()) except KeyboardInterrupt: print("Caught keyboard interrupt, killing and removing PID") os.remove(PID_FILE) except Exception as e: traceback.print_exc() print(str(e)) print("Removing PID file") os.remove(PID_FILE) finally: sys.exit(0) if __name__ == "__main__": sys.exit(init())