Make llama.cpp read prompt size and seed from settings (#2299)
This commit is contained in:
parent
ee674afa50
commit
cf088566f8
5 changed files with 9 additions and 3 deletions
|
@ -39,8 +39,8 @@ class LlamaCppModel:
|
|||
|
||||
params = {
|
||||
'model_path': str(path),
|
||||
'n_ctx': 2048,
|
||||
'seed': 0,
|
||||
'n_ctx': shared.args.n_ctx,
|
||||
'seed': int(shared.args.llama_cpp_seed),
|
||||
'n_threads': shared.args.threads or None,
|
||||
'n_batch': shared.args.n_batch,
|
||||
'use_mmap': not shared.args.no_mmap,
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue