Support for MPT, INCITE, WizardLM, StableLM, Galactica, Vicuna, Guanaco, and Baize instruction following (#1596)
This commit is contained in:
parent
06c7db017d
commit
814f754451
51 changed files with 352 additions and 28 deletions
|
@ -145,12 +145,12 @@ def load_quantized(model_name):
|
|||
# Find the model type
|
||||
if not shared.args.model_type:
|
||||
name = model_name.lower()
|
||||
if any((k in name for k in ['llama', 'alpaca', 'vicuna', 'llava'])):
|
||||
model_type = 'llama'
|
||||
elif any((k in name for k in ['opt-', 'galactica'])):
|
||||
if any((k in name for k in ['opt-', 'opt_', 'opt1', 'opt3', 'optfor', 'galactica', 'galpaca', 'pygmalion-350m'])):
|
||||
model_type = 'opt'
|
||||
elif any((k in name for k in ['gpt-j', 'pygmalion-6b'])):
|
||||
elif any((k in name for k in ['gpt-j', 'gptj', 'gpt4all-j', 'malion-6b', 'pygway'])):
|
||||
model_type = 'gptj'
|
||||
elif any((k in name for k in ['llama', 'alpac', 'vicuna', 'guanaco', 'koala', 'llava', 'wizardlm'])):
|
||||
model_type = 'llama'
|
||||
else:
|
||||
logging.error("Can't determine model type from model name. Please specify it manually using --model_type argument")
|
||||
exit()
|
||||
|
|
|
@ -71,12 +71,31 @@ settings = {
|
|||
'prompts': {
|
||||
'default': 'QA',
|
||||
'.*(gpt4chan|gpt-4chan|4chan)': 'GPT-4chan',
|
||||
'.*oasst': 'Open Assistant',
|
||||
'.*alpaca': "Alpaca",
|
||||
'.*(oasst|stablelm-7b-sft-v7-epoch-3)': 'Open Assistant',
|
||||
'.*(alpac|dolly)': "Alpaca",
|
||||
'.*mpt-.*instruct': "Alpaca",
|
||||
"(?!.*v0)(?!.*1.1)(?!.*1_1)(?!.*stable).*vicuna": "Vicuna v0",
|
||||
".*vicuna.*v0": "Vicuna v0",
|
||||
".*vicuna.*(1.1|1_1)": "Vicuna v1.1",
|
||||
".*stable.*vicuna": "StableVicuna",
|
||||
".*guanaco": "Guanaco-Chat",
|
||||
".*koala": "Koala",
|
||||
".*stablelm-tuned": "StableLM",
|
||||
".*wizardlm": "WizardLM",
|
||||
".*galactica.*finetuned": "Galactica Finetuned",
|
||||
".*galactica.*-v2": "Galactica v2",
|
||||
"(?!.*finetuned)(?!.*-v2).*galactica": "Galactica",
|
||||
".*baize": "Baize",
|
||||
".*mpt-.*instruct": "Alpaca",
|
||||
".*mpt-.*chat": "MPT-Chat",
|
||||
"(?!.*-flan-)(?!.*-t5-).*lamini-": "Alpaca",
|
||||
".*incite.*chat": "INCITE-Chat",
|
||||
".*incite.*instruct": "INCITE-Instruct",
|
||||
},
|
||||
'lora_prompts': {
|
||||
'default': 'QA',
|
||||
'.*alpaca': "Alpaca",
|
||||
'.*baize': "Baize",
|
||||
}
|
||||
}
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue