Create alternative requirements.txt with AMD and Metal wheels (#4052)
This commit is contained in:
parent
9de2dfa887
commit
2e7b6b0014
13 changed files with 336 additions and 66 deletions
34
requirements_cpu_only_noavx2.txt
Normal file
34
requirements_cpu_only_noavx2.txt
Normal file
|
@ -0,0 +1,34 @@
|
|||
aiofiles==23.1.0
|
||||
fastapi==0.95.2
|
||||
gradio_client==0.2.5
|
||||
gradio==3.33.1
|
||||
pydantic==1.10.12
|
||||
|
||||
accelerate==0.23.*
|
||||
colorama
|
||||
datasets
|
||||
einops
|
||||
exllamav2==0.0.3
|
||||
markdown
|
||||
numpy==1.24
|
||||
optimum==1.13.1
|
||||
pandas
|
||||
peft==0.5.*
|
||||
Pillow>=9.5.0
|
||||
pyyaml
|
||||
requests
|
||||
safetensors==0.3.2
|
||||
transformers==4.33.*
|
||||
scipy
|
||||
sentencepiece
|
||||
tensorboard
|
||||
tqdm
|
||||
wandb
|
||||
|
||||
# bitsandbytes
|
||||
bitsandbytes==0.41.1; platform_system != "Windows"
|
||||
https://github.com/jllllll/bitsandbytes-windows-webui/releases/download/wheels/bitsandbytes-0.41.1-py3-none-win_amd64.whl; platform_system == "Windows"
|
||||
|
||||
# llama-cpp-python (CPU only)
|
||||
https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/basic/llama_cpp_python-0.2.6+cu117-cp310-cp310-manylinux_2_31_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
||||
https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels/releases/download/basic/llama_cpp_python-0.2.6+cu117-cp310-cp310-win_amd64.whl; platform_system == "Windows"
|
Loading…
Add table
Add a link
Reference in a new issue