Add the --cpu option for llama.cpp to prevent CUDA from being used (#3432)
This commit is contained in:
parent
3e70bce576
commit
87dab03dc0
5 changed files with 40 additions and 10 deletions
|
@ -41,6 +41,7 @@ loaders_and_params = {
|
|||
'llama_cpp_seed',
|
||||
'compress_pos_emb',
|
||||
'alpha_value',
|
||||
'cpu',
|
||||
],
|
||||
'llamacpp_HF': [
|
||||
'n_ctx',
|
||||
|
@ -55,6 +56,7 @@ loaders_and_params = {
|
|||
'llama_cpp_seed',
|
||||
'compress_pos_emb',
|
||||
'alpha_value',
|
||||
'cpu',
|
||||
'llamacpp_HF_info',
|
||||
],
|
||||
'Transformers': [
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue