Fix exllamav3_hf models failing to unload (closes #7031)

This commit is contained in:
oobabooga 2025-05-30 12:05:49 -07:00
parent 298d4719c6
commit 219f0a7731
2 changed files with 20 additions and 0 deletions

View file

@ -116,10 +116,13 @@ def unload_model(keep_model_name=False):
return
is_llamacpp = (shared.model.__class__.__name__ == 'LlamaServer')
if shared.args.loader == 'ExLlamav3_HF':
shared.model.unload()
shared.model = shared.tokenizer = None
shared.lora_names = []
shared.model_dirty_from_training = False
if not is_llamacpp:
from modules.torch_utils import clear_torch_cache
clear_torch_cache()