Add AutoAWQ to README

This commit is contained in:
oobabooga 2023-10-05 09:22:37 -07:00
parent cc632c3f33
commit 7ffb424c7b
2 changed files with 2 additions and 4 deletions

View file

@ -107,14 +107,12 @@ def infer_loader(model_name, model_settings):
loader = None
elif (path_to_model / 'quantize_config.json').exists() or ('wbits' in model_settings and type(model_settings['wbits']) is int and model_settings['wbits'] > 0):
loader = 'AutoGPTQ'
elif (path_to_model / 'quant_config.json').exists():
elif (path_to_model / 'quant_config.json').exists() or re.match(r'.*-awq', model_name.lower()):
loader = 'AutoAWQ'
elif len(list(path_to_model.glob('*.gguf'))) > 0:
loader = 'llama.cpp'
elif re.match(r'.*\.gguf', model_name.lower()):
loader = 'llama.cpp'
elif re.match(r'.*-awq', model_name.lower()):
loader = 'AutoAWQ'
elif re.match(r'.*rwkv.*\.pth', model_name.lower()):
loader = 'RWKV'
elif re.match(r'.*exl2', model_name.lower()):