mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2025-12-06 07:12:10 +01:00
Slightly more robust model loading
This commit is contained in:
parent
d843afcf66
commit
00ed878b05
|
|
@ -45,12 +45,13 @@ def load_model(model_name, loader=None):
|
|||
model, tokenizer = output
|
||||
else:
|
||||
model = output
|
||||
if model is None:
|
||||
return None, None
|
||||
else:
|
||||
if model is not None:
|
||||
from modules.transformers_loader import load_tokenizer
|
||||
tokenizer = load_tokenizer(model_name)
|
||||
|
||||
if model is None:
|
||||
return None, None
|
||||
|
||||
shared.settings.update({k: v for k, v in metadata.items() if k in shared.settings})
|
||||
if loader.lower().startswith('exllama') or loader.lower().startswith('tensorrt') or loader == 'llama.cpp':
|
||||
shared.settings['truncation_length'] = shared.args.ctx_size
|
||||
|
|
|
|||
Loading…
Reference in a new issue