Make universal tokenizer, xformers, sdp-attention apply to monkey patch
This commit is contained in:
parent
f39c99fa14
commit
f642135517
2 changed files with 2 additions and 11 deletions
|
@ -149,8 +149,7 @@ def load_model(model_name):
|
|||
print("Warning: applying the monkey patch for using LoRAs in 4-bit mode.\nIt may cause undefined behavior outside its intended scope.")
|
||||
from modules.monkey_patch_gptq_lora import load_model_llama
|
||||
|
||||
model, tokenizer = load_model_llama(model_name)
|
||||
return model, tokenizer
|
||||
model, _ = load_model_llama(model_name)
|
||||
|
||||
# No monkey patch
|
||||
else:
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue