Add the --cpu option for llama.cpp to prevent CUDA from being used (#3432)

This commit is contained in:
oobabooga 2023-08-03 11:00:36 -03:00 committed by GitHub
parent 3e70bce576
commit 87dab03dc0
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23
5 changed files with 40 additions and 10 deletions

View file

@ -41,6 +41,7 @@ loaders_and_params = {
'llama_cpp_seed',
'compress_pos_emb',
'alpha_value',
'cpu',
],
'llamacpp_HF': [
'n_ctx',
@ -55,6 +56,7 @@ loaders_and_params = {
'llama_cpp_seed',
'compress_pos_emb',
'alpha_value',
'cpu',
'llamacpp_HF_info',
],
'Transformers': [