mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2025-12-06 07:12:10 +01:00
Slightly more robust model loading
This commit is contained in:
parent
d843afcf66
commit
00ed878b05
|
|
@ -45,12 +45,13 @@ def load_model(model_name, loader=None):
|
||||||
model, tokenizer = output
|
model, tokenizer = output
|
||||||
else:
|
else:
|
||||||
model = output
|
model = output
|
||||||
if model is None:
|
if model is not None:
|
||||||
return None, None
|
|
||||||
else:
|
|
||||||
from modules.transformers_loader import load_tokenizer
|
from modules.transformers_loader import load_tokenizer
|
||||||
tokenizer = load_tokenizer(model_name)
|
tokenizer = load_tokenizer(model_name)
|
||||||
|
|
||||||
|
if model is None:
|
||||||
|
return None, None
|
||||||
|
|
||||||
shared.settings.update({k: v for k, v in metadata.items() if k in shared.settings})
|
shared.settings.update({k: v for k, v in metadata.items() if k in shared.settings})
|
||||||
if loader.lower().startswith('exllama') or loader.lower().startswith('tensorrt') or loader == 'llama.cpp':
|
if loader.lower().startswith('exllama') or loader.lower().startswith('tensorrt') or loader == 'llama.cpp':
|
||||||
shared.settings['truncation_length'] = shared.args.ctx_size
|
shared.settings['truncation_length'] = shared.args.ctx_size
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue