Add llama-2-70b GGML support (#3285)
This commit is contained in:
parent
6f4830b4d3
commit
a07d070b6c
8 changed files with 20 additions and 4 deletions
|
@ -30,6 +30,8 @@ loaders_and_params = {
|
|||
],
|
||||
'llama.cpp': [
|
||||
'n_ctx',
|
||||
'n_gqa',
|
||||
'rms_norm_eps',
|
||||
'n_gpu_layers',
|
||||
'n_batch',
|
||||
'threads',
|
||||
|
@ -42,6 +44,8 @@ loaders_and_params = {
|
|||
],
|
||||
'llamacpp_HF': [
|
||||
'n_ctx',
|
||||
'n_gqa',
|
||||
'rms_norm_eps',
|
||||
'n_gpu_layers',
|
||||
'n_batch',
|
||||
'threads',
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue