Add low vram mode on llama cpp (#3076)
This commit is contained in:
parent
180420d2c9
commit
eedb3bf023
5 changed files with 5 additions and 0 deletions
|
|
@ -57,6 +57,7 @@ def list_model_elements():
|
|||
'threads',
|
||||
'n_batch',
|
||||
'no_mmap',
|
||||
'low_vram',
|
||||
'mlock',
|
||||
'n_gpu_layers',
|
||||
'n_ctx',
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue