Add files via upload
This commit is contained in:
parent
25b433990a
commit
80ef7c7bcb
15 changed files with 911 additions and 0 deletions
111
docs/Spell-book.md
Normal file
111
docs/Spell-book.md
Normal file
|
@ -0,0 +1,111 @@
|
|||
You have now entered a hidden corner of the internet.
|
||||
|
||||
A confusing yet intriguing realm of paradoxes and contradictions.
|
||||
|
||||
A place where you will find out that what you thought you knew, you in fact didn't know, and what you didn't know was in front of you all along.
|
||||
|
||||

|
||||
|
||||
*In other words, here I will document little-known facts about this web UI that I could not find another place for in the wiki.*
|
||||
|
||||
#### You can train LoRAs in CPU mode
|
||||
|
||||
Load the web UI with
|
||||
|
||||
```
|
||||
python server.py --cpu
|
||||
```
|
||||
|
||||
and start training the LoRA from the training tab as usual.
|
||||
|
||||
#### 8-bit mode works with CPU offloading
|
||||
|
||||
```
|
||||
python server.py --load-in-8bit --gpu-memory 4000MiB
|
||||
```
|
||||
|
||||
#### `--pre_layer`, and not `--gpu-memory`, is the right way to do CPU offloading with 4-bit models
|
||||
|
||||
```
|
||||
python server.py --wbits 4 --groupsize 128 --pre_layer 20
|
||||
```
|
||||
|
||||
#### Models can be loaded in 32-bit, 16-bit, 8-bit, and 4-bit modes
|
||||
|
||||
```
|
||||
python server.py --cpu
|
||||
python server.py
|
||||
python server.py --load-in-8bit
|
||||
python server.py --wbits 4
|
||||
```
|
||||
|
||||
#### The web UI works with any version of GPTQ-for-LLaMa
|
||||
|
||||
Including the up to date triton and cuda branches. But you have to delete the `repositories/GPTQ-for-LLaMa` folder and reinstall the new one every time:
|
||||
|
||||
```
|
||||
cd text-generation-webui/repositories
|
||||
rm -r GPTQ-for-LLaMa
|
||||
pip uninstall quant-cuda
|
||||
git clone https://github.com/oobabooga/GPTQ-for-LLaMa -b cuda # or any other repository and branch
|
||||
cd GPTQ-for-LLaMa
|
||||
python setup_cuda.py install
|
||||
```
|
||||
|
||||
#### Instruction-following templates are represented as chat characters
|
||||
|
||||
https://github.com/oobabooga/text-generation-webui/tree/main/characters/instruction-following
|
||||
|
||||
#### The right way to run Alpaca, Open Assistant, Vicuna, etc is Instruct mode, not normal chat mode
|
||||
|
||||
Otherwise the prompt will not be formatted correctly.
|
||||
|
||||
1. Start the web UI with
|
||||
|
||||
```
|
||||
python server.py --chat
|
||||
```
|
||||
|
||||
2. Click on the "instruct" option under "Chat modes"
|
||||
|
||||
3. Select the correct template in the hidden dropdown menu that will become visible.
|
||||
|
||||
#### Notebook mode is best mode
|
||||
|
||||
Ascended individuals have realized that notebook mode is the superset of chat mode and can do chats with ultimate flexibility, including group chats, editing replies, starting a new bot reply in a given way, and impersonating.
|
||||
|
||||
#### RWKV is a RNN
|
||||
|
||||
Most models are transformers, but not RWKV, which is a RNN. It's a great model.
|
||||
|
||||
#### `--gpu-memory` is not a hard limit on the GPU memory
|
||||
|
||||
It is simply a parameter that is passed to the `accelerate` library while loading the model. More memory will be allocated during generation. That's why this parameter has to be set to less than your total GPU memory.
|
||||
|
||||
#### Contrastive search perhaps the best preset
|
||||
|
||||
But it uses a ton of VRAM.
|
||||
|
||||
#### You can check the sha256sum of downloaded models with the download script
|
||||
|
||||
```
|
||||
python download-model.py facebook/galactica-125m --check
|
||||
```
|
||||
|
||||
#### The download script continues interrupted downloads by default
|
||||
|
||||
It doesn't start over.
|
||||
|
||||
#### You can download models with multiple threads
|
||||
|
||||
```
|
||||
python download-model.py facebook/galactica-125m --threads 8
|
||||
```
|
||||
|
||||
#### LoRAs work in 4-bit mode
|
||||
|
||||
You need to follow these instructions
|
||||
|
||||
https://github.com/oobabooga/text-generation-webui/wiki/GPTQ-models-(4-bit-mode)#using-loras-in-4-bit-mode
|
||||
|
||||
and then start the web UI with the `--monkey-patch` flag.
|
Loading…
Add table
Add a link
Reference in a new issue