From a7670c31cad79f5758cc7e0d69844a9157b3901e Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Tue, 28 Nov 2023 18:43:33 -0800 Subject: [PATCH] Sort --- modules/loaders.py | 18 +++++++++--------- 1 file changed, 9 insertions(+), 9 deletions(-) diff --git a/modules/loaders.py b/modules/loaders.py index 12c30e7..545c0e0 100644 --- a/modules/loaders.py +++ b/modules/loaders.py @@ -85,6 +85,15 @@ loaders_and_params = OrderedDict({ 'no_use_fast', 'autogptq_info', ], + 'AutoAWQ': [ + 'cpu_memory', + 'gpu_memory', + 'auto_devices', + 'max_seq_len', + 'no_inject_fused_attention', + 'trust_remote_code', + 'no_use_fast', + ], 'GPTQ-for-LLaMa': [ 'wbits', 'groupsize', @@ -125,15 +134,6 @@ loaders_and_params = OrderedDict({ 'alpha_value', 'compress_pos_emb', ], - 'AutoAWQ': [ - 'cpu_memory', - 'gpu_memory', - 'auto_devices', - 'max_seq_len', - 'no_inject_fused_attention', - 'trust_remote_code', - 'no_use_fast', - ], 'ctransformers': [ 'n_ctx', 'n_gpu_layers',