From 7bf15ad933a17e81d46a8d3d43a6965638343377 Mon Sep 17 00:00:00 2001 From: Sense_wang <167664334+haosenwang1018@users.noreply.github.com> Date: Thu, 5 Mar 2026 05:06:17 +0800 Subject: [PATCH] fix: replace bare except clauses with except Exception (#7400) --- extensions/openai/script.py | 4 ++-- extensions/sd_api_pictures/script.py | 8 ++++---- modules/callbacks.py | 2 +- modules/chat.py | 6 +++--- modules/evaluate.py | 2 +- modules/llama_cpp_server.py | 4 ++-- modules/logits.py | 2 +- modules/metadata_gguf.py | 2 +- modules/prompts.py | 2 +- modules/training.py | 4 ++-- modules/ui_model_menu.py | 2 +- one_click.py | 2 +- server.py | 2 +- 13 files changed, 21 insertions(+), 21 deletions(-) diff --git a/extensions/openai/script.py b/extensions/openai/script.py index 12f99ba4..edb22c22 100644 --- a/extensions/openai/script.py +++ b/extensions/openai/script.py @@ -357,7 +357,7 @@ async def handle_load_model(request_data: LoadModelRequest): try: OAImodels._load_model(to_dict(request_data)) return JSONResponse(content="OK") - except: + except Exception: traceback.print_exc() return HTTPException(status_code=400, detail="Failed to load the model.") @@ -378,7 +378,7 @@ async def handle_load_loras(request_data: LoadLorasRequest): try: OAImodels.load_loras(request_data.lora_names) return JSONResponse(content="OK") - except: + except Exception: traceback.print_exc() return HTTPException(status_code=400, detail="Failed to apply the LoRA(s).") diff --git a/extensions/sd_api_pictures/script.py b/extensions/sd_api_pictures/script.py index f216da38..6674d43d 100644 --- a/extensions/sd_api_pictures/script.py +++ b/extensions/sd_api_pictures/script.py @@ -264,7 +264,7 @@ def SD_api_address_update(address): response = requests.get(url=f'{params["address"]}/sdapi/v1/sd-models') response.raise_for_status() # r = response.json() - except: + except Exception: msg = "❌ No SD API endpoint on:" return gr.Textbox.update(label=msg) @@ -284,7 +284,7 @@ def get_checkpoints(): options_json = options.json() params['sd_checkpoint'] = options_json['sd_model_checkpoint'] params['checkpoint_list'] = [result["title"] for result in models.json()] - except: + except Exception: params['sd_checkpoint'] = "" params['checkpoint_list'] = [] @@ -298,7 +298,7 @@ def load_checkpoint(checkpoint): try: requests.post(url=f'{params["address"]}/sdapi/v1/options', json=payload) - except: + except Exception: pass @@ -307,7 +307,7 @@ def get_samplers(): response = requests.get(url=f'{params["address"]}/sdapi/v1/samplers') response.raise_for_status() samplers = [x["name"] for x in response.json()] - except: + except Exception: samplers = [] return samplers diff --git a/modules/callbacks.py b/modules/callbacks.py index 082365f0..afddf92d 100644 --- a/modules/callbacks.py +++ b/modules/callbacks.py @@ -37,7 +37,7 @@ class Iteratorize: ret = self.mfunc(callback=_callback, *args, **self.kwargs) except StopNowException: pass - except: + except Exception: traceback.print_exc() pass diff --git a/modules/chat.py b/modules/chat.py index 9ef8f5e9..21a2104a 100644 --- a/modules/chat.py +++ b/modules/chat.py @@ -1300,7 +1300,7 @@ def load_last_chat_state(): try: with open(state_file, 'r', encoding='utf-8') as f: return json.loads(f.read()) - except: + except Exception: pass return {"last_chats": {}} @@ -1372,7 +1372,7 @@ def load_history_json(file, history): update_message_metadata(history['metadata'], "assistant", i, timestamp="") return history - except: + except Exception: return history @@ -1526,7 +1526,7 @@ def upload_character(file, img_path, tavern=False): decoded_file = file if isinstance(file, str) else file.decode('utf-8') try: data = json.loads(decoded_file) - except: + except Exception: data = yaml.safe_load(decoded_file) if 'char_name' in data: diff --git a/modules/evaluate.py b/modules/evaluate.py index 4f41c1fc..ccf72ec6 100644 --- a/modules/evaluate.py +++ b/modules/evaluate.py @@ -86,7 +86,7 @@ def calculate_perplexity(models, input_dataset, stride, _max_length): update_model_parameters(model_settings) # hijacking the command-line arguments unload_model() shared.model, shared.tokenizer = load_model(model) - except: + except Exception: cumulative_log += f"Failed to load `{model}`. Moving on.\n\n" yield cumulative_log continue diff --git a/modules/llama_cpp_server.py b/modules/llama_cpp_server.py index a3463233..a90a192b 100644 --- a/modules/llama_cpp_server.py +++ b/modules/llama_cpp_server.py @@ -466,7 +466,7 @@ class LlamaServer: response = self.session.get(health_url) if response.status_code == 200: break - except: + except Exception: pass time.sleep(1) @@ -559,5 +559,5 @@ def filter_stderr_with_progress(process_stderr): finally: try: process_stderr.close() - except: + except Exception: pass diff --git a/modules/logits.py b/modules/logits.py index 7a5f98ae..d9e1ab48 100644 --- a/modules/logits.py +++ b/modules/logits.py @@ -121,7 +121,7 @@ def _get_next_logits(prompt, state, use_samplers, previous, top_logits=25, retur if isinstance(key, bytes): try: key = key.decode() - except: + except Exception: key = key.decode('latin') output[key] = row[0] diff --git a/modules/metadata_gguf.py b/modules/metadata_gguf.py index 70ad41dc..849f4485 100644 --- a/modules/metadata_gguf.py +++ b/modules/metadata_gguf.py @@ -53,7 +53,7 @@ def get_single(value_type, file): value = file.read(value_length) try: value = value.decode('utf-8') - except: + except Exception: pass else: type_str = _simple_value_packing.get(value_type) diff --git a/modules/prompts.py b/modules/prompts.py index b800af91..91f5812a 100644 --- a/modules/prompts.py +++ b/modules/prompts.py @@ -33,5 +33,5 @@ def count_tokens(text): try: tokens = get_encoded_length(text) return str(tokens) - except: + except Exception: return '0' diff --git a/modules/training.py b/modules/training.py index 1de0b702..41b2028e 100644 --- a/modules/training.py +++ b/modules/training.py @@ -503,7 +503,7 @@ def do_train(lora_name: str, always_override: bool, q_proj_en: bool, v_proj_en: print("Model reloaded OK, continue with training.") else: return f"Failed to load {selected_model}." - except: + except Exception: exc = traceback.format_exc() logger.error('Failed to reload the model.') print(exc) @@ -542,7 +542,7 @@ def do_train(lora_name: str, always_override: bool, q_proj_en: bool, v_proj_en: logger.info("Loading existing LoRA data") state_dict_peft = torch.load(f"{lora_file_path}/adapter_model.bin", weights_only=True) set_peft_model_state_dict(lora_model, state_dict_peft) - except: + except Exception: yield traceback.format_exc().replace('\n', '\n\n') return diff --git a/modules/ui_model_menu.py b/modules/ui_model_menu.py index a81a5474..c65c7a1b 100644 --- a/modules/ui_model_menu.py +++ b/modules/ui_model_menu.py @@ -224,7 +224,7 @@ def load_model_wrapper(selected_model, loader, autoload=False): yield f"Successfully loaded `{selected_model}`." else: yield f"Failed to load `{selected_model}`." - except: + except Exception: exc = traceback.format_exc() logger.error('Failed to load the model.') print(exc) diff --git a/one_click.py b/one_click.py index dad9e807..1af1a686 100644 --- a/one_click.py +++ b/one_click.py @@ -71,7 +71,7 @@ def load_state(): try: with open(state_file, 'r') as f: return json.load(f) - except: + except Exception: return {} return {} diff --git a/server.py b/server.py index ccfda5ac..56526666 100644 --- a/server.py +++ b/server.py @@ -75,7 +75,7 @@ def signal_handler(sig, frame): if shared.model and shared.model.__class__.__name__ == 'LlamaServer': try: shared.model.stop() - except: + except Exception: pass sys.exit(0)