Add CPU LoRA training (#938)

(It's very slow)
This commit is contained in:
IggoOnCode 2023-04-10 22:29:00 +02:00 committed by GitHub
parent 0caf718a21
commit 09d8119e3c
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23
3 changed files with 5 additions and 4 deletions

View file

@ -238,7 +238,7 @@ def do_train(lora_name: str, micro_batch_size: int, batch_size: int, epochs: int
warmup_steps=100,
num_train_epochs=epochs,
learning_rate=actual_lr,
fp16=True,
fp16=False if shared.args.cpu else True,
logging_steps=20,
evaluation_strategy="steps" if eval_data is not None else "no",
save_strategy="steps",
@ -248,7 +248,8 @@ def do_train(lora_name: str, micro_batch_size: int, batch_size: int, epochs: int
save_total_limit=3,
load_best_model_at_end=True if eval_data is not None else False,
# TODO: Enable multi-device support
ddp_find_unused_parameters=None
ddp_find_unused_parameters=None,
no_cuda=shared.args.cpu
),
data_collator=transformers.DataCollatorForLanguageModeling(shared.tokenizer, mlm=False),
callbacks=list([Callbacks()])