1
0
Fork 0
mirror of https://gitlab.com/niansa/discord_llama.git synced 2025-03-06 20:48:25 +01:00
discord_llama/explained_config.txt
2024-03-25 01:56:21 +01:00

57 lines
1.7 KiB
Text

token MTA0MDYxMTQzNjUwNzk1OTMyNw.Gl_iMU.jVVM3bRqBJVi8ORVpWHquOivlASGJpRySt8qFg
# The following parameters are set to their defaults here and can be ommited
# Directory the models are located in. For example, see example_models/
models_dir models
# File containing status texts. For example, see example_texts.txt
texts_file none
# Weather the bot should respond to pings outside threads. Disabling this may increase load by a LOT
threads_only true
# Chance for bot to respond at random when allowed to talk outside threads (see option above). Chance in percent is 100 divided by given number (Example: 2 = 50%). 0 implies no random responses
random_response_chance 0
# Weather the bot should update messages periodically while writing them. Incompatible with translation
live_edit false
# Model to use outside threads
default_inference_model 13b-vanilla
# Few-shot prompt for non-instruct-mode. See example_prompt.txt
prompt_file none
# Prompt for instruct-mode. See example_instruct_prompt.txt
instruct_prompt_file none
# Amount of shards ("instances") of this bot. This is NOT Discord sharding
shard_count 1
# Number of this shard. Must be unique in the entire bot
shard_id 0
# Weather context ("chat histories") should persist restarts
persistance true
# Weather swapping should be prevented using mlock
mlock false
# Amount of contexts to keep in RAM at a time
pool_size 2
# Amount of CPU threads to use
threads 4
# Response/Evaluation timeout in seconds; after which generation will be depriorized
timeout 120
# Max. context size
ctx_size 1012
# Max. context age in seconds; 0 to disable
max_context_age 0
# Percentage of context below prompt to be kept when scrolling. 0 means no context will be kept when scolling (not recommended!!!)
scroll_keep 20