Add low vram mode on llama cpp (#3076)
This commit is contained in:
parent
180420d2c9
commit
eedb3bf023
5 changed files with 5 additions and 0 deletions
|
@ -49,6 +49,7 @@ class LlamaCppModel:
|
|||
'n_batch': shared.args.n_batch,
|
||||
'use_mmap': not shared.args.no_mmap,
|
||||
'use_mlock': shared.args.mlock,
|
||||
'low_vram': shared.args.low_vram,
|
||||
'n_gpu_layers': shared.args.n_gpu_layers
|
||||
}
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue