diff --git a/modules/transformers_loader.py b/modules/transformers_loader.py index ef524b57..e4163b6d 100644 --- a/modules/transformers_loader.py +++ b/modules/transformers_loader.py @@ -131,6 +131,8 @@ def load_tokenizer(model_name, tokenizer_dir=None): def load_model_HF(model_name): + torch._dynamo.config.disable = True + path_to_model = Path(f'{shared.args.model_dir}/{model_name}') params = { 'low_cpu_mem_usage': True,