Slightly more robust model loading

This commit is contained in:
oobabooga 2025-09-02 10:16:26 -07:00
parent d843afcf66
commit 00ed878b05

View file

@ -45,12 +45,13 @@ def load_model(model_name, loader=None):
model, tokenizer = output model, tokenizer = output
else: else:
model = output model = output
if model is None: if model is not None:
return None, None
else:
from modules.transformers_loader import load_tokenizer from modules.transformers_loader import load_tokenizer
tokenizer = load_tokenizer(model_name) tokenizer = load_tokenizer(model_name)
if model is None:
return None, None
shared.settings.update({k: v for k, v in metadata.items() if k in shared.settings}) shared.settings.update({k: v for k, v in metadata.items() if k in shared.settings})
if loader.lower().startswith('exllama') or loader.lower().startswith('tensorrt') or loader == 'llama.cpp': if loader.lower().startswith('exllama') or loader.lower().startswith('tensorrt') or loader == 'llama.cpp':
shared.settings['truncation_length'] = shared.args.ctx_size shared.settings['truncation_length'] = shared.args.ctx_size