Add llama.cpp GPU offload option (#2060)
This commit is contained in:
parent
eee986348c
commit
071f0776ad
4 changed files with 23 additions and 5 deletions
|
@ -27,7 +27,8 @@ class LlamaCppModel:
|
|||
'n_threads': shared.args.threads or None,
|
||||
'n_batch': shared.args.n_batch,
|
||||
'use_mmap': not shared.args.no_mmap,
|
||||
'use_mlock': shared.args.mlock
|
||||
'use_mlock': shared.args.mlock,
|
||||
'n_gpu_layers': shared.args.n_gpu_layers
|
||||
}
|
||||
self.model = Llama(**params)
|
||||
self.model.set_cache(LlamaCache)
|
||||
|
|
|
@ -123,6 +123,7 @@ parser.add_argument('--threads', type=int, default=0, help='Number of threads to
|
|||
parser.add_argument('--n_batch', type=int, default=512, help='Maximum number of prompt tokens to batch together when calling llama_eval.')
|
||||
parser.add_argument('--no-mmap', action='store_true', help='Prevent mmap from being used.')
|
||||
parser.add_argument('--mlock', action='store_true', help='Force the system to keep the model in RAM.')
|
||||
parser.add_argument('--n-gpu-layers', type=int, default=0, help='Number of layers to offload to the GPU.')
|
||||
|
||||
# GPTQ
|
||||
parser.add_argument('--wbits', type=int, default=0, help='Load a pre-quantized model with specified precision in bits. 2, 3, 4 and 8 are supported.')
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue