From e0f5905a97bd40a343003b4626e08b3fec9416de Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Tue, 19 Aug 2025 06:34:05 -0700 Subject: [PATCH] Code formatting --- modules/exllamav3.py | 3 ++- one_click.py | 2 +- 2 files changed, 3 insertions(+), 2 deletions(-) diff --git a/modules/exllamav3.py b/modules/exllamav3.py index 73962977..fd676a00 100644 --- a/modules/exllamav3.py +++ b/modules/exllamav3.py @@ -236,11 +236,12 @@ class Exllamav3Model: """ Generate text with streaming using native ExLlamaV3 API """ - image_embeddings = [] if shared.is_multimodal: # Process images and modify prompt (ExLlamaV3-specific) prompt, image_embeddings = self._process_images_for_generation(prompt, state) + else: + image_embeddings = [] # Greedy decoding is a special case if state['temperature'] == 0: diff --git a/one_click.py b/one_click.py index 58c0e8ff..881d7489 100644 --- a/one_click.py +++ b/one_click.py @@ -155,7 +155,7 @@ def get_pytorch_update_command(gpu_choice): base_cmd = f"python -m pip install --upgrade torch=={TORCH_VERSION} " if gpu_choice == "NVIDIA_CUDA128": - return "python -m pip install --upgrade torch==2.7.1 --index-url https://download.pytorch.org/whl/cu128" + return f"{base_cmd} --index-url https://download.pytorch.org/whl/cu128" elif gpu_choice == "AMD": return f"{base_cmd} --index-url https://download.pytorch.org/whl/rocm6.2.4" elif gpu_choice in ["APPLE", "NONE"]: