Kaveen Kumarasinghe
749e660e37
|
2 years ago | |
---|---|---|
.github | ||
cogs | 2 years ago | |
detailed_guides | ||
models | 2 years ago | |
openers | ||
services | 2 years ago | |
.gitignore | ||
CODE_OF_CONDUCT.md | ||
CONTRIBUTING.md | ||
Dockerfile | ||
LICENSE | ||
README.md | 2 years ago | |
conversation_starter_pretext.txt | ||
conversation_starter_pretext_minimal.txt | ||
gpt3discord.py | 2 years ago | |
image_optimizer_pretext.txt | 2 years ago | |
pyproject.toml | 2 years ago | |
requirements.txt | 2 years ago | |
sample.env |
README.md
Overview
A robust, all-in-one GPT3 interface for Discord. Chat just like ChatGPT right inside Discord! Generate beautiful AI art using DALL-E 2! Automatically moderate your server using AI! Upload documents, videos, and files to get AI-assisted insights! A thorough integration with permanent conversation memory, automatic request retry, fault tolerance and reliability for servers of any scale, and much more.
SUPPORT SERVER FOR BOT SETUP: https://discord.gg/WvAHXDMS7Q (You can try out the bot here also in a limited fashion)
Screenshots
Recent Notable Updates
- CUSTOM INDEXES - This is a huge update. You can now upload files to your server and use them as custom context when asking GPT3 questions. You can also use webpage links as context, images, full documents, csvs, powerpoints, audio files, and even youtube videos! Read more in the 'Custom Indexes' section below.
Features
-
Directly prompt GPT3 with
/gpt ask <prompt>
-
Have long term, permanent conversations with the bot, just like chatgpt, with
/gpt converse
- Conversations happen in threads that get automatically cleaned up! -
Custom Indexes - Use your own files, pdfs, txt files, websites, discord channel content as context when asking GPT3 questions!
-
DALL-E Image Generation - Generate DALL-E AI images right in discord with
/dalle draw <prompt>
! It even supports multiple image qualities, multiple images, creating image variants, retrying, and saving images. -
DALL-E Image Prompt Optimization - Given some text that you're trying to generate an image for, the bot will automatically optimize the text to be more DALL-E friendly!
/dalle optimize <prompt>
-
Edit Requests - Ask GPT to edit a piece of text or code with a given instruction.
/gpt edit <instruction> <text>
-
DeepL Translations - Translate text with DeepL.
/translate <text>
-
Redo Requests - A simple button after the GPT3 response or DALL-E generation allows you to redo the initial prompt you asked. You can also redo conversation messages by just editing your message!
-
Automatic AI-Based Server Moderation - Moderate your server automatically with AI!
-
Auto-retry on API errors - Automatically resend failed requests to OpenAI's APIs!
-
Automatically re-send your prompt and update the response in place if you edit your original prompt!
-
Async and fault tolerant, can handle hundreds of users at once, if the upstream API permits!
-
Change and view model parameters such as temp, top_p, and etc directly within discord.
-
Tracks token usage automatically
-
Automatic pagination and discord support, the bot will automatically send very long message as multiple messages, and is able to send discord code blocks and emoji, gifs, etc.
-
A low usage mode, use a command to automatically switch to a cheaper and faster model to conserve your tokens during times of peak usage.
-
Prints debug to a channel of your choice, so you can view the raw response JSON
-
Ability to specify a limit to how long a conversation can be with the bot, to conserve your tokens.
Commands
These commands are grouped, so each group has a prefix but you can easily tab complete the command without the prefix. For example, for /gpt ask
, if you type /ask
and press tab, it'll show up too.
/help
- Display help text for the bot
(Chat)GPT3 Commands
/gpt ask <prompt> <temp> <top_p> <frequency penalty> <presence penalty>
Ask the GPT3 Davinci 003 model a question. Optional overrides available
/gpt edit <instruction> <input> <temp> <top_p> <codex>
Use the bot to edit text using the given instructions for how to do it, currently an alpha openai feature so results might vary. Codex uses a model trained on code. Editing is currently free
/gpt converse <opener> <opener_file> <private> <minimal>
- Start a conversation with the bot, like ChatGPT
-
opener:<opener text>
- Start a conversation with the bot, with a custom opener text (this is useful if you want it to take on a custom personality from the start). -
opener_file:<opener file name>.txt|.json
- Starts a conversation with the bot, using a custom file.-
Loads files from the
/openers
folder, has autocomplete support so files in the folder will show up. Added before theopener
as both can be used at the same time -
Custom openers need to be placed as a .txt file in the
openers
directory, in the same directory asgpt3discord.py
-
Enables minimal
-
Can use .json files in the
{"text": "your prompt", "temp":0, "top_p":0,"frequency_penalty":0,"presence_penalty":0}
format to include permanent overrides
-
-
private
- Start a private conversation with the bot, like ChatGPT -
minimal
- Start a conversation with the bot, like ChatGPT, with minimal context (saves tokens)
/gpt end
- End a conversation with the bot.
DALL-E2 Commands
/dalle draw <prompt>
- Have DALL-E generate images based on a prompt
/dalle optimize <image prompt text>
Optimize a given prompt text for DALL-E image generation.
Custom Indexes Commands
This bot supports per-user custom indexes. This means that users can upload files of their choosing, such as PDFs and ask GPT to answer questions based on those files.
/index add file:<file> or link:<link>
- Use a document or use a link to create/add to your indexes. If you provide a youtube link, the transcript of the video will be used. If you provide a web url, the contents of the webpage will be used, if you provide an image, the image text will be extracted and used!
/index query query:<prompt>
- Query your current index for a given prompt. GPT will answer based on your current document/indedx
/index load index:<index>
- Load a previously created index to query
/index compose
- Combine multiple saved indexes into one, or upgrade existing indexes into Deep Compositions.
/index reset
- Reset and delete all of your saved indexes
/index add_discord channel:<discord channel>
- Create an add an index based on a discord channel
/index discord_backup
- Use the last 3000 messages of every channel on your discord server as an index
System and Settings
/system settings
- Display settings for the model (temperature, top_p, etc)
/system settings <setting> <value>
- Change a model setting to a new value. Has autocomplete support, certain settings will have autocompleted values too.
/system usage
Estimate current usage details (based on davinci)
/system settings low_usage_mode True/False
Turn low usage mode on and off. If on, it will use the curie-001 model, and if off, it will use the davinci-003 model.
/system delete-conversation-threads
- Delete all threads related to this bot across all servers.
/system local-size
- Get the size of the local dalleimages folder
/system clear-local
- Clear all the local dalleimages.
Automatic AI Moderation
/mod set status:on
- Turn on automatic chat moderations.
/mod set status:off
- Turn off automatic chat moderations
/mod set status:off alert_channel_id:<CHANNEL ID>
- Turn on moderations and set the alert channel to the channel ID you specify in the command.
/mod config type:<warn/delete> hate:# hate_threatening:# self_harm:# sexual:# sexual_minors:# violence:# violence_graphic:#
- Set the moderation thresholds of the bot for the specific type of moderation (
warn
ordelete
). You can view the thresholds by typing just/mod config type:<warn/delete>
without any other parameters. You don't have to set all of them, you can just set one or two items if you want. For example, to set the hate threshold for warns, you can type/mod config type:warn hate:0.2
- Lower values are more strict, higher values are more lenient. There are default values that I've fine tuned the service with for a general server.
The bot needs Administrative permissions for this, and you need to set MODERATIONS_ALERT_CHANNEL
to the channel ID of a desired channel in your .env file if you want to receive alerts about moderated messages.
This uses the OpenAI Moderations endpoint to check for messages, requests are only sent to the moderations endpoint at a MINIMUM request gap of 0.5 seconds, to ensure you don't get blocked and to ensure reliability.
The bot uses numerical thresholds to determine whether a message is toxic or not, and I have manually tested and fine tuned these thresholds to a point that I think is good, please open an issue if you have any suggestions for the thresholds!
There are two thresholds for the bot, there are instances in which the bot will outright delete a message and an instance where the bot will send a message to the alert channel notifying admins and giving them quick options to delete and timeout the user (check out the screenshots at the beginning of the README to see this).
If you'd like to help us test and fine tune our thresholds for the moderation service, please join this test server: https://discord.gg/CWhsSgNdrP. You can let off some steam in a controlled environment ;)
To set a certain role immune to moderations, add the line CHAT_BYPASS_ROLES="Role1,Role2,etc"
to your `.env file.
The above server is NOT for support or discussions about GPT3Discord
Permanent Memory and Conversations
Permanent memory has now been implemented into the bot, using the OpenAI Ada embeddings endpoint, and Pinecone.
Pinecone is a vector database. The OpenAI Ada embeddings endpoint turns pieces of text into embeddings. The way that this feature works is by embedding the user prompts and the GPT3 responses, storing them in a pinecone index, and then retrieving the most relevant bits of conversation whenever a new user prompt is given in a conversation.
You do NOT need to use pinecone, if you do not define a PINECONE_TOKEN
in your .env
file, the bot will default to not using pinecone, and will use conversation summarization as the long term conversation method instead.
To enable permanent memory with pinecone, you must define a PINECONE_TOKEN
in your .env
file as follows (along with the other variables too):
PINECONE_TOKEN="87juwi58-1jk9-9182-9b3c-f84d90e8bshq"
To get a pinecone token, you can sign up for a free pinecone account here: https://app.pinecone.io/ and click the "API Keys" section on the left navbar to find the key. (I am not affiliated with pinecone).
After signing up for a free pinecone account, you need to create an index in pinecone. To do this, go to the pinecone dashboard and click "Create Index" on the top right.
Then, name the index conversation-embeddings
, set the dimensions to 1536
, and set the metric to DotProduct
:
Permanent memory using pinecone is still in alpha, I will be working on cleaning up this work, adding auto-clearing, and optimizing for stability and reliability, any help and feedback is appreciated (add me on Discord Kaveen#0001 for pinecone help)! If at any time you're having too many issues with pinecone, simply remove the PINECONE_TOKEN
line in your .env
file and the bot will revert to using conversation summarizations.
Custom Indexes / Knowledgebase
This bot supports per-user custom indexes. This means that users can upload files of their choosing, such as PDFs and ask GPT to answer questions based on those files. We also support using URLs for indexes.
This feature uses a large amount of tokens and money, and you should restrict it to trusted users.
Supported filetypes:
- All text and data based files (PDF, TXT, DOCX, PPTX, CSV etc)
- Images (JPG, PNG, etc) (Note: The bot will do OCR on the images to extract the text, this requires a lot of processing power sometimes)
- Videos/Audio (MP4, MP3, etc) (Note: The bot will use OpenAI on the audio to extract the text, this requires a lot of processing power sometimes)
- Youtube Videos - For all youtube videos that are transcribable, the bot will index the entire transcription of the given youtube video URL!
Index Compositions:
Indexes can be combined with other indexes through a composition. To combine indexes, you can run the /index compose
command, and select the indexes that you want to combine together. You should only combine relevant indexes together, combining irrelevant indexes together will result in poor results (for example, don't upload a math textbook and then upload a large set of poems and combine them together). When creating a composition, you will be given the option to do a "Deep" composition, deep compositions are more detailed and will give you better results, but are incredibly costly and will sometimes take multiple minutes to compose.
You can also compose a singular index with itself with "Deep Compose", this will give you a more detailed version of the index, but will be costly and will sometimes take multiple minutes to compose. Deep compositions are useless for very short documents!
Translations with DeepL
This bot supports and uses DeepL for translations (optionally). If you want to enable the translations service, you can add a line in your .env
file as follows:
DEEPL_TOKEN="your deepl token"
You can get a DeepL token by signing up at https://www.deepl.com/pro-api?cta=header-pro-api/ and clicking on the free plan to start. The DeepL translation service unlocks some new commands for your bot:
/translate <text> <language>
- Translate any given piece of text into the language that you provide
/languages
- See a list of all supported languages
Using DeepL also adds a new app menu button (when you right click a message) to the bot which allows you to quickly translate any message in a channel into any language you want:
User-Input API Keys (Multi-key tenancy)
This bot supports multi-user tenancy in regards to API keys. This means that, if you wanted, you could make it such that each user needs to enter their own API key in order to use commands that use GPT3 and DALLE.
To enable this, add the following line to the end of your .env
file:
USER_INPUT_API_KEYS="True"
Then, restart the bot, and it will set up the system for everyone to input their own API keys.
The bot will use SQLite to store API keys for the users, each user's key will be saved with a USER_ID <> API_KEY mapping in SQLite, and will be persistent across restarts. All the data will be saved in a file called user_key_db.sqlite
in the current working directory of the bot.
With this feature enabled, any attempt to use a GPT3 or DALL-E command without a valid API key set for the user will pop up the following modal for them to enter their API key:
Once the user enters their key, the bot will send a small test request to OpenAI to validate that the key indeed works, if not, it will tell the user to try again and tell them why it did not work.
After the user's key is validated, they will be able to use GPT3 and DALLE commands.
The Moderations service still uses the main API key defined in the .env
file. Pinecone and discord-tokens are also per-host tokens, not per-user.
Step-by-Step Guides for GPT3Discord
If you follow the link above, you will now get to detailed step-by-step guides that will help you to install and set up your GPT3Discord bot quickly and easily. If you still run into problems or have suggestions for improving the guides, you can join the Discord-Server and we will try to help you. Keep in mind that the maintainers are volunteers and will try to help you on their schedule.
The number and content of the guides is constantly adapted to current requirements.
Configuration
All the model parameters are configurable inside discord. Type /system settings
to view all the configurable parameters, and use /system settings <param> <value>
to set parameters.
For example, if I wanted to change the number of images generated by DALL-E by default to 4, I can type the following command in discord: /system settings num_images 4
Requirements
For OCR, and document functionalities:
pip3 install torch==1.9.1+cpu torchvision==0.10.1+cpu -f https://download.pytorch.org/whl/torch_stable.html
or
python3.9 -m pip install torch==1.9.1+cpu torchvision==0.10.1+cpu -f https://download.pytorch.org/whl/torch_stable.html
For audio extraction for indexing from .mp3 and .mp4 files:
python3.9 -m pip install git+https://github.com/openai/whisper.git
All other dependencies:
python3.9 -m pip install -r requirements.txt
I recommend using python 3.9!
OpenAI API Key (https://beta.openai.com/docs/api-reference/introduction)
Discord Bot Token (https://discord.com/developers/applications)
You can learn how to add the discord bot to your server via https://www.ionos.co.uk/digitalguide/server/know-how/creating-discord-bot/
Both the OpenAI API key and the Discord bot token needed to be loaded into a .env file in the same local directory as the bot file.
You also need to add a DEBUG_GUILD id and a DEBUG_CHANNEL id, the debug guild id is a server id, and the debug channel id is a text-channel id in Discord. Your final .env file should look like the following:
You also need to add the allowed guilds that the bot can operate on, this is the ALLOWED_GUILDS
field, to get a guild ID, right click a server and click "Copy ID".
You also need to add the roles that can use the bots various features, scroll down a bit to "Permissions", and check out the sample environment file below.
OPENAI_TOKEN = "<openai_api_token>"
DISCORD_TOKEN = "<discord_bot_token>"
#PINECONE_TOKEN = "<pinecone_token>" # pinecone token if you have it enabled. See readme
DEBUG_GUILD = "974519864045756446" # discord_server_id
DEBUG_CHANNEL = "977697652147892304" # discord_chanel_id
ALLOWED_GUILDS = "971268468148166697,971268468148166697"
# People with the roles in ADMIN_ROLES can use admin commands like /clear-local, and etc
ADMIN_ROLES = "Admin,Owner"
# People with the roles in DALLE_ROLES can use commands like /dalle draw or /dalle imgoptimize
DALLE_ROLES = "Admin,Openai,Dalle,gpt"
# People with the roles in GPT_ROLES can use commands like /gpt ask or /gpt converse
GPT_ROLES = "openai,gpt"
WELCOME_MESSAGE = "Hi There! Welcome to our Discord server. We hope you'll enjoy our server and we look forward to engaging with you!" # This is a fallback message if gpt3 fails to generate a welcome message.
USER_INPUT_API_KEYS="False" # If True, users must use their own API keys for OpenAI. If False, the bot will use the API key in the .env file.
# Moderations Service alert channel, this is where moderation alerts will be sent as a default if enabled
MODERATIONS_ALERT_CHANNEL = "977697652147892304"
# User API key db path configuration. This is where the user API keys will be stored.
USER_KEY_DB_PATH = "user_key_db.sqlite"
Permissions
As mentioned in the comments of the sample environment file, there are three permission groups that you can edit in the environment (.env
) file. ADMIN_ROLES
are roles that allow users to use /system
and /mod
commands. GPT_ROLES
are roles that allow users to use /gpt
commands, and DALLE_ROLES
are roles that allow users to use /dalle
commands.
Optionally, you can include your own conversation starter text for the bot that's used with /gpt converse
, with CONVERSATION_STARTER_TEXT
If for a command group you want everybody to be able to use those commands, just don't include the relevant line in the .env
file. For example, if you want everyone to be able to use GPT3 commands, you can just omit the GPT_ROLES="...."
line.
Server Installation
First, you want to get a server, for this guide, I will be using DigitalOcean as the host.
For instructions on how to get a server from start to finish, they are available on DigitalOcean's website directly from the community, available here: https://www.digitalocean.com/community/tutorials/how-to-set-up-an-ubuntu-20-04-server-on-a-digitalocean-droplet. Ignore the part about setting up an "ssh key", and just use a password instead.
Please sign up for a DigitalOcean account using my referral link if you'd like to support me https://m.do.co/c/e31eff1231a4
After you set up the server, the DigitalOcean GUI will give you an IP address, copy this IP address. Afterwards, you will need to SSH into the server. This can be done using a program such as "PuTTy", or by using your commandline, if it's supported. To login to the server, your username will be "root", your password will be the password that you defined earlier when setting up the droplet, and the IP address will be the IP address you copied after the droplet was finished creation.
To connect with ssh, run the following command in terminal:
ssh root@{IP ADDRESS}
It will then prompt you for your password, which you should enter, and then you will be logged in.
After login, we need to install the various dependencies that the bot needs. To do this, we will run the following commands:
Download the source code.
git clone https://github.com/Kav-K/GPT3Discord.git
cd GPT3Discord/
# Install system packages (python)
sudo apt-get update
sudo apt install software-properties-common
sudo add-apt-repository ppa:deadsnakes/ppa
sudo apt install python3.9
sudo apt install python3.9-distutils # If this doesn't work, try sudo apt install python3-distutils
# Install Pip for python3.9
curl https://bootstrap.pypa.io/get-pip.py -o get-pip.py
python3.9 get-pip.py
# Install project dependencies
python3.9 -m pip install --ignore-installed PyYAML
python3.9 -m pip install torch==1.9.1+cpu torchvision==0.10.1+cpu -f https://download.pytorch.org/whl/torch_stable.html
python3.9 -m pip install -r requirements.txt
python3.9 -m pip install .
# Copy the sample.env file into a regular .env file. `DEBUG_GUILD` and the ID for `ALLOWED_GUILDS` can be found by right-clicking your server and choosing "Copy ID". Similarly, `DEBUG_CHANNEL` can be found by right-clicking your debug channel.
cp sample.env .env
# The command below is used to edit the .env file and to put in your API keys. You can right click within the
# editor after running this command to paste. When you are done editing, press CTRL + X, and then type Y, to save.
nano .env
# Run the bot using [screen](https://www.gnu.org/software/screen/manual/screen.html) to keep it running after you disconnect from your SSH session:
screen gpt3discord
# Hit `Ctrl+a` then `d` to detach from the running bot.
# The bot's screen session can be reattached:
screen -r
If the last few commands don't allow the bot to run screen gpt3discord
, you can attempt to run the bot another way:
{Navigate to the folder where the project files are}
screen -dmS GPTBot bash -c 'python3.9 gpt3discord.py'
# Reattach to screen session
screen -x # will reattach if this is the only screen session, if there are multiple, it will show IDs
# If there are multiple IDs returned by screen -x:
screen -d -r {ID} # replace {ID} with the ID of the screen session you want to reattach to
As a last resort, you can try to run the bot using python in a basic way, with simply
cd (the folder where the files for GPT3Discord are located/cloned)
python3.9 gpt3discord.py
Updating
To update the bot,
run (when working in the directory of GPT3discord):
#To get the latest branch:
git pull
#Install the latest modules so the bot keeps working.
python3.9 -m pip install -r requirements.txt
python3.9 -m pip install .
Permanent overrides in threads
This bot now supports having overrides be permanent in an entire conversation if you use an opener file which includes them. The new opener files should be .json files formatted like this. text
corresponds to what you want the conversational opener to be and the rest map 1:1 to the appropriate model settings. An example .json file is included by the name of english_translator.json
in the openers
folder
{
"text": "your prompt",
"temp":0,
"top_p":0,
"frequency_penalty":0,
"presence_penalty":0
}
Non-Server, Non-Docker usage
With python3.9 installed and the requirements installed, you can run this bot anywhere.
Install the dependencies with:
pip3 install torch==1.9.1+cpu torchvision==0.10.1+cpu -f https://download.pytorch.org/whl/torch_stable.html
python3.9 -m pip install -r requirements.txt
Then, run the bot with:
python3.9 gpt3discord.py
Docker Installation
We now have a Dockerfile
in the repository. This will build / install all dependencies and put a gpt3discord
binary (main.py) into path.
To build:
- Install docker
- Clone repository and build (hopefully eventually we'll add CI to automatically build + push to docker hub)
docker build -t gpt3discord .
- From repository root or supply path to repository
- Make a .env file to bind mount to
/opt/gpt3discord/etc/environment
- Optional: Make a data directory + bind mount it
- Add
DATA_DIR=/data
to env file ->usage.txt
is made here - Add
SHARE_DIR=/data/share
to env file -> this is whereconversation starters, optimizer pretext and the 'openers' folder
is loaded from - If
SHARE_DIR
is not included it'll load from the files added during the docker image build
- Add
- Run via docker:
docker run [-d] --name gpt3discord -v env_file:/opt/gpt3discord/etc/environment [-v /containers/gpt3discord:/data] [-v /containers/gpt3discord/share:/data/share] gpt3discord
- You can also mount extra volumes and set
DATA_DIR
andSHARE_DIR
in the env file to keep persistent data env_file
should be replaced with where you have your .env file stored on your machine
This can also be run via screen/tmux or detached like a daemon.
Bot on discord:
- Create a new Bot on Discord Developer Portal:
- Applications -> New Application
- Generate Token for the app (discord_bot_token)
- Select App (Bot) -> Bot -> Reset Token
- Toogle PRESENCE INTENT:
- Select App (Bot) -> Bot -> PRESENCE INTENT, SERVER MEMBERS INTENT, MESSAGES INTENT, (basically turn on all intents)
- Add Bot the the server.
- Select App (Bot) -> OAuth2 -> URL Generator -> Select Scope: Bot, application.commands
- Bot Permissions will appear, select the desired permissions
- Copy the link generated below and paste it on the browser
- On add to server select the desired server to add the bot
- Make sure you have updated your .env file with valid values for
DEBUG_GUILD
,DEBUG_CHANNEL
andALLOWED_GUILDS
, otherwise the bot will not work. Guild IDs can be found by right clicking a server and clickingCopy ID
, similarly, channel IDs can be found by right clicking a channel and clickingCopy ID
.
Health Check Service
The bot has the ability to launch a HTTP endpoint at <host>:8181/
that will return a json response of the bot's status and uptime. This is especially useful if you want to run this bot on cloud application containers, like Azure App Service.
To enable this, add HEALTH_SERVICE_ENABLED="True"
to your .env
file.
The health check endpoint will then be present in your bot's console when it is starting up, it will look like this, the possible HTTP urls for your health endpoint will be visible near the bottom:
Custom Bot Name
Add a line CUSTOM_BOT_NAME=<NAME>
to your .env
to give your bot a custom name in conversations.