mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2025-12-06 07:12:10 +01:00
Fix model unloading when switching loaders (closes #7203)
This commit is contained in:
parent
08594e5263
commit
7d23a55901
|
|
@ -128,10 +128,12 @@ def unload_model(keep_model_name=False):
|
||||||
if shared.model is None:
|
if shared.model is None:
|
||||||
return
|
return
|
||||||
|
|
||||||
is_llamacpp = (shared.model.__class__.__name__ == 'LlamaServer')
|
model_class_name = shared.model.__class__.__name__
|
||||||
if shared.args.loader in ['ExLlamav3_HF', 'ExLlamav3']:
|
is_llamacpp = (model_class_name == 'LlamaServer')
|
||||||
|
|
||||||
|
if model_class_name in ['Exllamav3Model', 'Exllamav3HF']:
|
||||||
shared.model.unload()
|
shared.model.unload()
|
||||||
elif shared.args.loader in ['ExLlamav2_HF', 'ExLlamav2'] and hasattr(shared.model, 'unload'):
|
elif model_class_name in ['Exllamav2Model', 'Exllamav2HF'] and hasattr(shared.model, 'unload'):
|
||||||
shared.model.unload()
|
shared.model.unload()
|
||||||
|
|
||||||
shared.model = shared.tokenizer = None
|
shared.model = shared.tokenizer = None
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue