add n_batch support for llama.cpp (#1115)
This commit is contained in:
parent
2f6e2ddeac
commit
78d1977ebf
3 changed files with 4 additions and 1 deletions
|
@ -220,6 +220,7 @@ Optionally, you can use the following command-line flags:
|
|||
| Flag | Description |
|
||||
|-------------|-------------|
|
||||
| `--threads` | Number of threads to use in llama.cpp. |
|
||||
| `--n_batch` | Processing batch size for llama.cpp. |
|
||||
|
||||
#### GPTQ
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue