| .. |
|
AutoGPTQ_loader.py
|
Add --no_use_cuda_fp16 param for AutoGPTQ
|
2023-06-23 12:22:56 -03:00 |
|
block_requests.py
|
Add modules/block_requests.py
|
2023-06-18 16:31:14 -03:00 |
|
callbacks.py
|
Implement stopping string search in string space (#2847)
|
2023-06-24 09:43:00 -03:00 |
|
chat.py
|
Implement stopping string search in string space (#2847)
|
2023-06-24 09:43:00 -03:00 |
|
deepspeed_parameters.py
|
Style improvements (#1957)
|
2023-05-09 22:49:39 -03:00 |
|
evaluate.py
|
Reorganize model loading UI completely (#2720)
|
2023-06-16 19:00:37 -03:00 |
|
exllama.py
|
Add ExLlama+LoRA support (#2756)
|
2023-06-19 12:31:24 -03:00 |
|
exllama_hf.py
|
Fix Multi-GPU not working on exllama_hf (#2803)
|
2023-06-22 16:05:25 -03:00 |
|
extensions.py
|
Prevent unwanted log messages from modules
|
2023-05-21 22:42:34 -03:00 |
|
github.py
|
Reorganize model loading UI completely (#2720)
|
2023-06-16 19:00:37 -03:00 |
|
GPTQ_loader.py
|
Prevent unwanted log messages from modules
|
2023-05-21 22:42:34 -03:00 |
|
html_generator.py
|
Add markdown table rendering
|
2023-05-10 13:41:23 -03:00 |
|
llama_attn_hijack.py
|
Prevent unwanted log messages from modules
|
2023-05-21 22:42:34 -03:00 |
|
llamacpp_model.py
|
fix usage of self in classmethod (#2781)
|
2023-06-20 16:18:42 -03:00 |
|
loaders.py
|
Add --no_use_cuda_fp16 param for AutoGPTQ
|
2023-06-23 12:22:56 -03:00 |
|
logging_colors.py
|
Add menus for saving presets/characters/instruction templates/prompts (#2621)
|
2023-06-11 12:19:18 -03:00 |
|
LoRA.py
|
Add ExLlama+LoRA support (#2756)
|
2023-06-19 12:31:24 -03:00 |
|
models.py
|
Fix some models not loading on exllama_hf (#2835)
|
2023-06-23 11:31:02 -03:00 |
|
models_settings.py
|
Add ExLlama support (#2444)
|
2023-06-16 20:35:38 -03:00 |
|
monkey_patch_gptq_lora.py
|
Better warning messages
|
2023-05-03 21:43:17 -03:00 |
|
presets.py
|
Reorganize model loading UI completely (#2720)
|
2023-06-16 19:00:37 -03:00 |
|
relative_imports.py
|
Add ExLlama+LoRA support (#2756)
|
2023-06-19 12:31:24 -03:00 |
|
RWKV.py
|
Add ExLlama support (#2444)
|
2023-06-16 20:35:38 -03:00 |
|
sampler_hijack.py
|
Add Mirostat v2 sampling to transformer models (#2571)
|
2023-06-09 21:26:31 -03:00 |
|
shared.py
|
Add --no_use_cuda_fp16 param for AutoGPTQ
|
2023-06-23 12:22:56 -03:00 |
|
text_generation.py
|
Implement stopping string search in string space (#2847)
|
2023-06-24 09:43:00 -03:00 |
|
training.py
|
Added Falcon LoRA training support (#2684)
|
2023-06-20 01:03:44 -03:00 |
|
ui.py
|
Add --no_use_cuda_fp16 param for AutoGPTQ
|
2023-06-23 12:22:56 -03:00 |
|
utils.py
|
Allow API requests to use parameter presets
|
2023-06-14 11:32:20 -03:00 |