Add --no_use_cuda_fp16 param for AutoGPTQ
This commit is contained in:
parent
5646690769
commit
3ae9af01aa
5 changed files with 6 additions and 2 deletions
|
@ -9,6 +9,7 @@ loaders_and_params = {
|
|||
'triton',
|
||||
'no_inject_fused_attention',
|
||||
'no_inject_fused_mlp',
|
||||
'no_use_cuda_fp16',
|
||||
'wbits',
|
||||
'groupsize',
|
||||
'desc_act',
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue