From abc699db9b3c370c6be977d59a23a0e02987dd79 Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Fri, 6 Mar 2026 19:03:38 -0800 Subject: [PATCH] Minor UI change --- modules/models_settings.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/modules/models_settings.py b/modules/models_settings.py index 104635b9..472871ce 100644 --- a/modules/models_settings.py +++ b/modules/models_settings.py @@ -415,7 +415,7 @@ def update_gpu_layers_and_vram(loader, model, gpu_layers, ctx_size, cache_type): an HTML string for the UI display. """ if loader != 'llama.cpp' or model in ["None", None] or not model.endswith(".gguf") or gpu_layers < 0 or ctx_size == 0: - return "