Rename --llama-bits to --gptq-bits
This commit is contained in:
parent
fed3617f07
commit
65dda28c9d
3 changed files with 4 additions and 4 deletions
|
@ -16,7 +16,7 @@ def load_quantized_LLaMA(model_name):
|
|||
if shared.args.load_in_4bit:
|
||||
bits = 4
|
||||
else:
|
||||
bits = shared.args.llama_bits
|
||||
bits = shared.args.gptq_bits
|
||||
|
||||
path_to_model = Path(f'models/{model_name}')
|
||||
pt_model = ''
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue