From 2c4699a7e9a1e611052f6e5635ddb9942b26524a Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Mon, 13 Mar 2023 00:20:02 -0300 Subject: [PATCH] Change a comment --- modules/quantized_LLaMA.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/modules/quantized_LLaMA.py b/modules/quantized_LLaMA.py index fa7f15c..e9352f9 100644 --- a/modules/quantized_LLaMA.py +++ b/modules/quantized_LLaMA.py @@ -42,7 +42,7 @@ def load_quantized_LLaMA(model_name): model = load_quant(path_to_model, str(pt_path), bits) - # Multi-GPU setup + # Multiple GPUs or GPU+CPU if shared.args.gpu_memory: max_memory = {} for i in range(len(shared.args.gpu_memory)):