Implement auto_max_new_tokens for ExLlama

This commit is contained in:
oobabooga 2023-08-02 11:01:29 -07:00
parent e931844fe2
commit 32a2bbee4a
2 changed files with 6 additions and 1 deletions

View file

@ -151,6 +151,7 @@ loaders_samplers = {
'repetition_penalty_range',
'seed',
'ban_eos_token',
'auto_max_new_tokens',
},
'AutoGPTQ': {
'temperature',