add n_batch support for llama.cpp (#1115)

This commit is contained in:
eiery 2023-04-24 02:46:18 -04:00 committed by GitHub
parent 2f6e2ddeac
commit 78d1977ebf
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23
3 changed files with 4 additions and 1 deletions

View file

@ -220,6 +220,7 @@ Optionally, you can use the following command-line flags:
| Flag | Description |
|-------------|-------------|
| `--threads` | Number of threads to use in llama.cpp. |
| `--n_batch` | Processing batch size for llama.cpp. |
#### GPTQ