Bump llama-cpp-python to 0.2.18 (#4611)
This commit is contained in:
parent
61f429563e
commit
923c8e25fb
17 changed files with 92 additions and 174 deletions
|
@ -316,6 +316,9 @@ def update_requirements(initial_installation=False):
|
|||
run_cmd("python -m pip uninstall -y " + package_name, environment=True)
|
||||
print(f"Uninstalled {package_name}")
|
||||
|
||||
# Uninstall previous llama-cpp-python versions
|
||||
run_cmd("python -m pip uninstall -y llama-cpp-python-cuda" + package_name, environment=True)
|
||||
|
||||
# Make sure that API requirements are installed (temporary)
|
||||
extension_req_path = os.path.join("extensions", "openai", "requirements.txt")
|
||||
if os.path.exists(extension_req_path):
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue