1
0
Fork 0
mirror of https://gitlab.com/niansa/discord_llama.git synced 2025-03-06 20:48:25 +01:00
No description
Find a file
2024-03-27 18:42:06 +01:00
cosched2@d3d2b42f6e Updated cosched2 to fix D++ crash 2024-03-25 21:42:28 +01:00
DPP@46044c8f80 Updated various submodules 2024-03-25 01:43:53 +01:00
example_models Updated example configs 2024-03-25 01:56:21 +01:00
fmt Added fmt 2023-04-24 22:19:45 +02:00
libjustlm@449a25c360 Updated various submodules 2024-03-25 01:43:53 +01:00
sqlite_modern_cpp Added tons of features including thread creation 2023-04-25 15:55:46 +02:00
.gitignore Initial commit 2023-03-19 21:09:15 +01:00
.gitmodules Updated various submodules 2024-03-25 01:43:53 +01:00
CMakeLists.txt Updated various submodules 2024-03-25 01:43:53 +01:00
config.cpp Updated various submodules 2024-03-25 01:43:53 +01:00
config.hpp Updated various submodules 2024-03-25 01:43:53 +01:00
Dockerfile Updated Dockerfile to use Debian Bookworm 2023-05-07 12:59:08 +02:00
example_config.txt Updated example configs 2024-03-25 01:56:21 +01:00
example_instruct_prompt.txt Suggest use of Markdown in example_instruct_prompt.txt 2023-05-26 17:24:43 +00:00
example_prompt.txt Added example prompt file 2023-04-24 22:36:42 +02:00
example_texts.txt Made status texts configurable 2023-04-30 00:52:12 +02:00
explained_config.txt Updated example configs 2024-03-25 01:56:21 +01:00
LICENSE Add LICENSE 2023-04-20 06:21:05 +00:00
main.cpp Lock thread_embeds_mutex 2024-03-27 18:42:06 +01:00
README.md Updated various submodules 2024-03-25 01:43:53 +01:00
utils.cpp Fixed utils::max_words cutting off the string too early 2023-04-29 21:52:55 +02:00
utils.hpp Reorganization and added experimental thread embeds 2023-04-29 21:26:13 +02:00

Discord LLaMa

Overview

This Discord bot lets you use GPT-J and LLaMa models to both participate in your conversations and answer your questions in thread channels, in a ChatGPT-like fashion!

Setup

On a fresh Debian 11 install, you'll need the following packages: git g++ make cmake zlib1g-dev libssl-dev libsqlite3-dev. For better performance with large batch sizes, libopenblas-dev is recommended.

Then, you can clone this repository including all its submodules:

git clone https://gitlab.com/niansa/discord_llama.git --recursive

Once the command has completed, create a build directory, configure the project and compile it:

cd discord_llama
mkdir build
cd build
cmake ..
make -j$(nproc)

We should now download a model. Let's start with gpt4all-j-v1.3-groovy! I've already written an example model config for it, so let's just grab that and download the model:

mkdir models
cd models
cp ../../example_models/gpt4all-j-v1.3-groovy.txt .
wget https://gpt4all.io/models/ggml-gpt4all-j-v1.3-groovy.bin
cd ..

Next, we need a configuration file. Let's just grab the example config:

cp ../example_config.txt ./config.txt

and edit it. For reference, you can read explained_config.txt.

Please note that gpt4all-j-v1.3-groovy is only capable of instruct mode, so you'll have to set threads_only to true, if you've downloaded just that model.

Finally, start the Discord Bot by passing the config file to the executable:

./discord_llama config.txt

And that's it! Feel free to play around, try different models, tweak the config file, etc... It's really easy! If you still have any questions, please write an issue or contact me on Discord: Tuxifan#0981.

Credits

A huge thank you to Nomic AI for helping me get this project to where it is now! Also, this project wouldn't have been possible without the effort of Georgi Gerganov (ggerganov), who has writte the core of it all, the inference code itself.