ExLlama with long context (#2875)
This commit is contained in:
parent
9290c6236f
commit
c52290de50
14 changed files with 22 additions and 25 deletions
|
@ -91,7 +91,8 @@ class ExllamaHF(PreTrainedModel):
|
|||
assert weight_path is not None, f'could not find weight in "{pretrained_model_name_or_path}"'
|
||||
|
||||
config.model_path = str(weight_path)
|
||||
|
||||
config.max_seq_len = shared.args.max_seq_len
|
||||
config.compress_pos_emb = shared.args.compress_pos_emb
|
||||
if shared.args.gpu_split:
|
||||
config.set_auto_map(shared.args.gpu_split)
|
||||
config.gpu_peer_fix = True
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue