Merge pull request #263 from HideLord/main
Fixing compatibility with GPTQ repository
This commit is contained in:
commit
901dcba9b4
1 changed files with 1 additions and 1 deletions
|
@ -41,7 +41,7 @@ def load_quantized_LLaMA(model_name):
|
|||
print(f"Could not find {pt_model}, exiting...")
|
||||
exit()
|
||||
|
||||
model = load_quant(path_to_model, pt_path, bits)
|
||||
model = load_quant(path_to_model, os.path.abspath(pt_path), bits)
|
||||
|
||||
# Multi-GPU setup
|
||||
if shared.args.gpu_memory:
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue