Replace ggml occurences with gguf

This commit is contained in:
oobabooga 2023-08-26 01:06:59 -07:00
parent 1a642c12b5
commit 83640d6f43
5 changed files with 12 additions and 12 deletions

View file

@ -156,7 +156,7 @@ text-generation-webui
In the "Model" tab of the UI, those models can be automatically downloaded from Hugging Face. You can also download them via the command-line with `python download-model.py organization/model`.
* GGML models are a single file and should be placed directly into `models`. Example:
* GGUF models are a single file and should be placed directly into `models`. Example:
```
text-generation-webui
@ -258,7 +258,7 @@ Optionally, you can use the following command-line flags:
| `--quant_type QUANT_TYPE` | quant_type for 4-bit. Valid options: nf4, fp4. |
| `--use_double_quant` | use_double_quant for 4-bit. |
#### GGML (for llama.cpp and ctransformers)
#### GGUF (for llama.cpp and ctransformers)
| Flag | Description |
|-------------|-------------|