From 6a75bda4194fe16ddce804bd1a5af7602573a13d Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Mon, 5 Jun 2023 12:07:52 -0300 Subject: [PATCH] Assign some 4096 seq lengths --- models/config.yaml | 2 ++ server.py | 1 + 2 files changed, 3 insertions(+) diff --git a/models/config.yaml b/models/config.yaml index 59a2afb..2c09267 100644 --- a/models/config.yaml +++ b/models/config.yaml @@ -180,9 +180,11 @@ llama-65b-gptq-3bit: .*bluemoonrp-(30|13)b: mode: 'instruct' instruction_template: 'Bluemoon' + truncation_length: 4096 .*Nous-Hermes-13b: mode: 'instruct' instruction_template: 'Alpaca' .*airoboros-13b-gpt4: mode: 'instruct' instruction_template: 'Vicuna-v1.1' + truncation_length: 4096 diff --git a/server.py b/server.py index 1c8a5fe..8bfb45a 100644 --- a/server.py +++ b/server.py @@ -1049,6 +1049,7 @@ if __name__ == "__main__": 'mode': shared.settings['mode'], 'skip_special_tokens': shared.settings['skip_special_tokens'], 'custom_stopping_strings': shared.settings['custom_stopping_strings'], + 'truncation_length': shared.settings['truncation_length'], } shared.model_config.move_to_end('.*', last=False) # Move to the beginning