Enable NUMA feature for llama_cpp_python (#4040)
This commit is contained in:
parent
87ea2d96fd
commit
7e6ff8d1f0
7 changed files with 21 additions and 12 deletions
|
@ -100,6 +100,7 @@ loaders_and_params = OrderedDict({
|
|||
'rope_freq_base',
|
||||
'compress_pos_emb',
|
||||
'cpu',
|
||||
'numa',
|
||||
],
|
||||
'llamacpp_HF': [
|
||||
'n_ctx',
|
||||
|
@ -115,6 +116,7 @@ loaders_and_params = OrderedDict({
|
|||
'rope_freq_base',
|
||||
'compress_pos_emb',
|
||||
'cpu',
|
||||
'numa',
|
||||
'cfg_cache',
|
||||
'use_fast',
|
||||
'llamacpp_HF_info',
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue