Bump llama-cpp-python, remove python 3.8/3.9, cuda 11.7 (#5397)
This commit is contained in:
parent
528318b700
commit
89f6036e98
10 changed files with 60 additions and 162 deletions
|
@ -65,12 +65,9 @@
|
|||
" torver = torch.__version__\n",
|
||||
" print(f\"TORCH: {torver}\")\n",
|
||||
" is_cuda118 = '+cu118' in torver # 2.1.0+cu118\n",
|
||||
" is_cuda117 = '+cu117' in torver # 2.0.1+cu117\n",
|
||||
"\n",
|
||||
" textgen_requirements = open('requirements.txt').read().splitlines()\n",
|
||||
" if is_cuda117:\n",
|
||||
" textgen_requirements = [req.replace('+cu121', '+cu117').replace('+cu122', '+cu117').replace('torch2.1', 'torch2.0') for req in textgen_requirements]\n",
|
||||
" elif is_cuda118:\n",
|
||||
" if is_cuda118:\n",
|
||||
" textgen_requirements = [req.replace('+cu121', '+cu118').replace('+cu122', '+cu118') for req in textgen_requirements]\n",
|
||||
" with open('temp_requirements.txt', 'w') as file:\n",
|
||||
" file.write('\\n'.join(textgen_requirements))\n",
|
||||
|
@ -130,4 +127,4 @@
|
|||
"outputs": []
|
||||
}
|
||||
]
|
||||
}
|
||||
}
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue