Remove obsolete models/config.yaml and related code

This commit is contained in:
oobabooga 2026-03-24 18:48:50 -07:00
parent f48a2b79d0
commit 807be11832
7 changed files with 4 additions and 228 deletions

View file

@ -112,7 +112,7 @@ Used for talking to an instruction-following model using the prompt format defin
The prompt format is defined by the **Instruction template** parameter in "Parameters" > "Instruction template", which represents a Jinja2 template.
Note that when you load a model in the "Model" tab, the web UI will try to automatically detect its instruction template (if any), and will update the values under "Parameters" > "Instruction template" accordingly. This is done using a set of regular expressions defined in `user_data/models/config.yaml`. This detection is not guaranteed to be accurate. You should check the model card on Hugging Face to see if you are using the correct prompt format.
Note that when you load a model in the "Model" tab, the web UI will try to automatically detect its instruction template (if any) from the model metadata (e.g. `tokenizer_config.json` or GGUF metadata), and will update the values under "Parameters" > "Instruction template" accordingly. You should check the model card on Hugging Face to see if you are using the correct prompt format.
### Chat-instruct

View file

@ -39,7 +39,7 @@ curl http://127.0.0.1:5000/v1/completions \
#### Chat completions
Works best with instruction-following models. If the "instruction_template" variable is not provided, it will be guessed automatically based on the model name using the regex patterns in `user_data/models/config.yaml`.
Works best with instruction-following models. If the "instruction_template" variable is not provided, it will be detected automatically from the model metadata.
```shell
curl http://127.0.0.1:5000/v1/chat/completions \

View file

@ -67,7 +67,6 @@ def load_model(model_name, loader=None):
logger.info(f"Loaded \"{model_name}\" in {(time.time()-t0):.2f} seconds.")
logger.info(f"LOADER: \"{loader}\"")
logger.info(f"TRUNCATION LENGTH: {shared.settings['truncation_length']}")
logger.info(f"INSTRUCTION TEMPLATE: \"{metadata['instruction_template']}\"")
return model, tokenizer

View file

@ -23,14 +23,9 @@ def get_fallback_settings():
def get_model_metadata(model):
model_path = resolve_model_path(model)
model_settings = {}
# Get settings from user_data/models/config.yaml and user_data/models/config-user.yaml
settings = shared.model_config
for pat in settings:
if re.match(pat.lower(), Path(model).name.lower()):
for k in settings[pat]:
model_settings[k] = settings[pat][k]
# Fallback settings
model_settings = get_fallback_settings()
path = model_path / 'config.json'
if path.exists():

View file

@ -454,17 +454,7 @@ def load_user_config():
args.loader = fix_loader_name(args.loader)
# Load model-specific settings
p = Path(f'{args.model_dir}/config.yaml')
if p.exists():
model_config = yaml.safe_load(open(p, 'r').read())
else:
model_config = {}
del p
# Load custom model-specific settings
user_config = load_user_config()
model_config = OrderedDict(model_config)
user_config = OrderedDict(user_config)

View file

@ -18,7 +18,6 @@ import modules.extensions as extensions_module
from modules.LoRA import add_lora_to_model
from modules.models import load_model, unload_model_if_idle
from modules.models_settings import (
get_fallback_settings,
get_model_metadata,
update_model_parameters
)
@ -271,10 +270,6 @@ if __name__ == "__main__":
# Apply CLI overrides for image model settings (CLI flags take precedence over saved settings)
shared.apply_image_model_cli_overrides()
# Fallback settings for models
shared.model_config['.*'] = get_fallback_settings()
shared.model_config.move_to_end('.*', last=False) # Move to the beginning
# Activate the extensions listed on settings.yaml
extensions_module.available_extensions = utils.get_available_extensions()
for extension in shared.settings['default_extensions']:

View file

@ -1,203 +0,0 @@
.*(llama|alpac|vicuna|guanaco|koala|llava|wizardlm|metharme|pygmalion-7b|pygmalion-2|mythalion|wizard-mega|openbuddy|vigogne|h2ogpt-research|manticore):
model_type: 'llama'
.*(opt-|opt_|opt1|opt3|optfor|galactica|galpaca|pygmalion-350m):
model_type: 'opt'
.*(gpt-j|gptj|gpt4all-j|malion-6b|pygway|pygmalion-6b|dolly-v1):
model_type: 'gptj'
.*(gpt-neox|koalpaca-polyglot|polyglot.*koalpaca|polyglot-ko|polyglot_ko|pythia|stablelm|incite|dolly-v2|polycoder|h2ogpt-oig|h2ogpt-oasst1|h2ogpt-gm):
model_type: 'gptneox'
.*bloom:
model_type: 'bloom'
.*gpt2:
model_type: 'gpt2'
.*falcon:
model_type: 'falcon'
.*mpt:
model_type: 'mpt'
.*(starcoder|starchat):
model_type: 'starcoder'
.*dolly-v2:
model_type: 'dollyv2'
.*replit:
model_type: 'replit'
.*(oasst|openassistant-|stablelm-7b-sft-v7-epoch-3):
instruction_template: 'Open Assistant'
skip_special_tokens: false
(?!.*galactica)(?!.*reward).*openassistant:
instruction_template: 'Open Assistant'
skip_special_tokens: false
.*galactica:
skip_special_tokens: false
.*dolly-v[0-9]-[0-9]*b:
instruction_template: 'Alpaca'
skip_special_tokens: false
.*alpaca-native-4bit:
instruction_template: 'Alpaca'
.*llava:
instruction_template: 'LLaVA'
.*llava.*1.5:
instruction_template: 'Vicuna-v1.1'
.*wizard.*mega:
instruction_template: 'Wizard-Mega'
.*starchat-beta:
instruction_template: 'Starchat-Beta'
(?!.*v0)(?!.*1.1)(?!.*1_1)(?!.*stable)(?!.*chinese).*vicuna:
instruction_template: 'Vicuna-v0'
.*vicuna.*v0:
instruction_template: 'Vicuna-v0'
.*vicuna.*(1.1|1_1|1.3|1_3):
instruction_template: 'Vicuna-v1.1'
.*vicuna.*(1.5|1_5):
instruction_template: 'Vicuna-v1.1'
.*stable.*vicuna:
instruction_template: 'StableVicuna'
(?!.*chat).*chinese-vicuna:
instruction_template: 'Alpaca'
.*chinese-vicuna.*chat:
instruction_template: 'Chinese-Vicuna-Chat'
.*alpaca:
instruction_template: 'Alpaca'
.*koala:
instruction_template: 'Koala'
.*chatglm:
instruction_template: 'ChatGLM'
.*(metharme|pygmalion|mythalion):
instruction_template: 'Metharme'
.*raven:
instruction_template: 'RWKV-Raven'
.*moss-moon.*sft:
instruction_template: 'MOSS'
.*stablelm-tuned:
instruction_template: 'StableLM'
.*galactica.*finetuned:
instruction_template: 'Galactica Finetuned'
.*galactica.*-v2:
instruction_template: 'Galactica v2'
(?!.*finetuned)(?!.*-v2).*galactica:
instruction_template: 'Galactica'
.*guanaco:
instruction_template: 'Guanaco non-chat'
.*baize:
instruction_template: 'Baize'
.*mpt-.*instruct:
instruction_template: 'Alpaca'
.*mpt-.*chat:
instruction_template: 'ChatML'
(?!.*-flan-)(?!.*-t5-).*lamini-:
instruction_template: 'Alpaca'
.*incite.*chat:
instruction_template: 'INCITE-Chat'
.*incite.*instruct:
instruction_template: 'INCITE-Instruct'
.*ziya-:
instruction_template: 'Ziya'
.*koalpaca:
instruction_template: 'KoAlpaca'
.*openbuddy:
instruction_template: 'OpenBuddy'
(?!.*chat).*vigogne:
instruction_template: 'Vigogne-Instruct'
.*vigogne.*chat:
instruction_template: 'Vigogne-Chat'
.*(llama-deus|supercot|llama-natural-instructions|open-llama-0.3t-7b-instruct-dolly-hhrlhf|open-llama-0.3t-7b-open-instruct):
instruction_template: 'Alpaca'
.*bactrian:
instruction_template: 'Bactrian'
.*(h2ogpt-oig-|h2ogpt-oasst1-|h2ogpt-research-oasst1-):
instruction_template: 'INCITE-Chat'
.*h2ogpt-gm-:
instruction_template: 'H2O-prompt_answer'
.*manticore:
instruction_template: 'Manticore Chat'
.*bluemoonrp-(30|13)b:
instruction_template: 'Bluemoon'
.*Nous-Hermes-13b:
instruction_template: 'Alpaca'
.*airoboros:
instruction_template: 'Vicuna-v1.1'
.*airoboros.*1.2:
instruction_template: 'Airoboros-v1.2'
.*alpa(cino|sta):
instruction_template: 'Alpaca'
.*hippogriff:
instruction_template: 'Hippogriff'
.*lazarus:
instruction_template: 'Alpaca'
.*guanaco-.*(7|13|33|65)b:
instruction_template: 'Vicuna-v0'
.*hypermantis:
instruction_template: 'Alpaca'
.*open-llama-.*-open-instruct:
instruction_template: 'Alpaca'
.*starcoder-gpteacher-code-instruct:
instruction_template: 'Alpaca'
.*tulu:
instruction_template: 'Tulu'
.*chronos:
instruction_template: 'Alpaca'
.*samantha:
instruction_template: 'Samantha'
.*wizardcoder:
instruction_template: 'Alpaca'
.*minotaur:
instruction_template: 'Manticore Chat'
.*orca_mini:
instruction_template: 'Orca Mini'
.*(platypus|gplatty|superplatty):
instruction_template: 'Alpaca'
.*(openorca-platypus2):
instruction_template: 'OpenOrca-Platypus2'
.*longchat:
instruction_template: 'Vicuna-v1.1'
.*vicuna-33b:
instruction_template: 'Vicuna-v1.1'
.*redmond-hermes-coder:
instruction_template: 'Alpaca'
.*wizardcoder-15b:
instruction_template: 'Alpaca'
.*wizardlm:
instruction_template: 'Vicuna-v1.1'
.*godzilla:
instruction_template: 'Alpaca'
.*llama(-?)(2|v2).*chat:
instruction_template: 'Llama-v2'
.*newhope:
instruction_template: 'NewHope'
.*stablebeluga2:
instruction_template: 'StableBeluga2'
.*openchat:
instruction_template: 'OpenChat'
.*codellama.*instruct:
instruction_template: 'Llama-v2'
.*(mistral|mixtral).*instruct:
instruction_template: 'Mistral'
.*mistral.*openorca:
instruction_template: 'ChatML'
.*(WizardCoder-Python-34B-V1.0|Phind-CodeLlama-34B-v2|CodeBooga-34B-v0.1):
instruction_template: 'Alpaca'
.*orca-2-(13|7)b:
instruction_template: 'ChatML'
.*openhermes.*mistral:
instruction_template: 'ChatML'
.*Yi-34B-Chat:
instruction_template: 'ChatML'
(dolphin).*:
instruction_template: 'ChatML'
.*synthia:
instruction_template: 'Synthia'
.*(hercules|hyperion):
instruction_template: 'ChatML'
.*command-r:
instruction_template: 'Command-R'
.*xwin-lm-70b-v0.1:
instruction_template: 'Vicuna-v1.1'
.*platypus-yi-34b:
instruction_template: 'Vicuna-v1.1'
.*CausalLM-RP-34B:
instruction_template: 'ChatML'
34b-beta:
instruction_template: 'ChatML'
.*airoboros-3_1-yi-34b-200k:
instruction_template: 'Llama-v2'
.*chatqa:
instruction_template: 'NVIDIA-ChatQA'