Bump peft and transformers
This commit is contained in:
parent
727fd229f4
commit
335c49cc7e
2 changed files with 6 additions and 7 deletions
|
@ -21,7 +21,7 @@ from datasets import Dataset, load_dataset
|
||||||
from peft import (
|
from peft import (
|
||||||
LoraConfig,
|
LoraConfig,
|
||||||
get_peft_model,
|
get_peft_model,
|
||||||
prepare_model_for_int8_training,
|
prepare_model_for_kbit_training,
|
||||||
set_peft_model_state_dict
|
set_peft_model_state_dict
|
||||||
)
|
)
|
||||||
from peft.utils.other import \
|
from peft.utils.other import \
|
||||||
|
@ -483,7 +483,7 @@ def do_train(lora_name: str, always_override: bool, save_steps: int, micro_batch
|
||||||
# == Start prepping the model itself ==
|
# == Start prepping the model itself ==
|
||||||
if not hasattr(shared.model, 'lm_head') or hasattr(shared.model.lm_head, 'weight'):
|
if not hasattr(shared.model, 'lm_head') or hasattr(shared.model.lm_head, 'weight'):
|
||||||
logger.info("Getting model ready...")
|
logger.info("Getting model ready...")
|
||||||
prepare_model_for_int8_training(shared.model)
|
prepare_model_for_kbit_training(shared.model)
|
||||||
|
|
||||||
# base model is now frozen and should not be reused for any other LoRA training than this one
|
# base model is now frozen and should not be reused for any other LoRA training than this one
|
||||||
shared.model_dirty_from_training = True
|
shared.model_dirty_from_training = True
|
||||||
|
|
|
@ -3,26 +3,25 @@ fastapi==0.95.2
|
||||||
gradio_client==0.2.5
|
gradio_client==0.2.5
|
||||||
gradio==3.33.1
|
gradio==3.33.1
|
||||||
|
|
||||||
accelerate==0.21.0
|
accelerate==0.21.*
|
||||||
colorama
|
colorama
|
||||||
datasets
|
datasets
|
||||||
einops
|
einops
|
||||||
markdown
|
markdown
|
||||||
numpy==1.24
|
numpy==1.24
|
||||||
pandas
|
pandas
|
||||||
|
peft==0.5.*
|
||||||
Pillow>=9.5.0
|
Pillow>=9.5.0
|
||||||
pyyaml
|
pyyaml
|
||||||
requests
|
requests
|
||||||
safetensors==0.3.1
|
safetensors==0.3.*
|
||||||
|
transformers==4.32.*
|
||||||
scipy
|
scipy
|
||||||
sentencepiece
|
sentencepiece
|
||||||
tensorboard
|
tensorboard
|
||||||
tqdm
|
tqdm
|
||||||
wandb
|
wandb
|
||||||
|
|
||||||
git+https://github.com/huggingface/peft@4b371b489b9850fd583f204cdf8b5471e098d4e4
|
|
||||||
git+https://github.com/huggingface/transformers@baf1daa58eb2960248fd9f7c3af0ed245b8ce4af
|
|
||||||
|
|
||||||
bitsandbytes==0.41.1; platform_system != "Windows"
|
bitsandbytes==0.41.1; platform_system != "Windows"
|
||||||
https://github.com/jllllll/bitsandbytes-windows-webui/releases/download/wheels/bitsandbytes-0.41.1-py3-none-win_amd64.whl; platform_system == "Windows"
|
https://github.com/jllllll/bitsandbytes-windows-webui/releases/download/wheels/bitsandbytes-0.41.1-py3-none-win_amd64.whl; platform_system == "Windows"
|
||||||
https://github.com/PanQiWei/AutoGPTQ/releases/download/v0.4.1/auto_gptq-0.4.1+cu117-cp310-cp310-win_amd64.whl; platform_system == "Windows"
|
https://github.com/PanQiWei/AutoGPTQ/releases/download/v0.4.1/auto_gptq-0.4.1+cu117-cp310-cp310-win_amd64.whl; platform_system == "Windows"
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue