token MTA0MDYxMTQzNjUwNzk1OTMyNw.Gl_iMU.jVVM3bRqBJVi8ORVpWHquOivlASGJpRySt8qFg # The following parameters are set to their defaults here and can be ommited # Directory the models are located in. For example, see example_models/ models_dir models # File containing status texts. For example, see example_texts.txt texts_file none # Weather the bot should respond to pings outside threads. Disabling this may increase load by a LOT threads_only true # Chance for bot to respond at random when allowed to talk outside threads (see option above). Chance in percent is 100 divided by given number (Example: 2 = 50%). 0 implies no random responses random_response_chance 0 # Weather the bot should update messages periodically while writing them. Incompatible with translation live_edit false # Model to use outside threads default_inference_model 13b-vanilla # Few-shot prompt for non-instruct-mode. See example_prompt.txt prompt_file none # Prompt for instruct-mode. See example_instruct_prompt.txt instruct_prompt_file none # Amount of shards ("instances") of this bot. This is NOT Discord sharding shard_count 1 # Number of this shard. Must be unique in the entire bot shard_id 0 # Weather context ("chat histories") should persist restarts persistance true # Weather swapping should be prevented using mlock mlock false # Amount of contexts to keep in RAM at a time pool_size 2 # Amount of CPU threads to use threads 4 # Response/Evaluation timeout in seconds; after which generation will be depriorized timeout 120 # Max. context size ctx_size 1012 # Max. context age in seconds; 0 to disable max_context_age 0 # Percentage of context below prompt to be kept when scrolling. 0 means no context will be kept when scolling (not recommended!!!) scroll_keep 20