mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2025-12-06 07:12:10 +01:00
Move the new image_utils.py file to modules/
This commit is contained in:
parent
c6b4d1e87f
commit
2f90ac9880
|
|
@ -7,7 +7,6 @@ import tiktoken
|
||||||
from pydantic import ValidationError
|
from pydantic import ValidationError
|
||||||
|
|
||||||
from extensions.openai.errors import InvalidRequestError
|
from extensions.openai.errors import InvalidRequestError
|
||||||
from extensions.openai.image_utils import convert_openai_messages_to_images
|
|
||||||
from extensions.openai.typing import ToolDefinition
|
from extensions.openai.typing import ToolDefinition
|
||||||
from extensions.openai.utils import debug_msg, getToolCallId, parseToolCall
|
from extensions.openai.utils import debug_msg, getToolCallId, parseToolCall
|
||||||
from modules import shared
|
from modules import shared
|
||||||
|
|
@ -17,6 +16,7 @@ from modules.chat import (
|
||||||
load_character_memoized,
|
load_character_memoized,
|
||||||
load_instruction_template_memoized
|
load_instruction_template_memoized
|
||||||
)
|
)
|
||||||
|
from modules.image_utils import convert_openai_messages_to_images
|
||||||
from modules.logging_colors import logger
|
from modules.logging_colors import logger
|
||||||
from modules.presets import load_preset_memoized
|
from modules.presets import load_preset_memoized
|
||||||
from modules.text_generation import decode, encode, generate_reply
|
from modules.text_generation import decode, encode, generate_reply
|
||||||
|
|
|
||||||
|
|
@ -18,11 +18,11 @@ from exllamav3.generator.sampler import (
|
||||||
SS_TopP
|
SS_TopP
|
||||||
)
|
)
|
||||||
|
|
||||||
from extensions.openai.image_utils import (
|
from modules import shared
|
||||||
|
from modules.image_utils import (
|
||||||
convert_image_attachments_to_pil,
|
convert_image_attachments_to_pil,
|
||||||
convert_openai_messages_to_images
|
convert_openai_messages_to_images
|
||||||
)
|
)
|
||||||
from modules import shared
|
|
||||||
from modules.logging_colors import logger
|
from modules.logging_colors import logger
|
||||||
from modules.text_generation import get_max_prompt_length
|
from modules.text_generation import get_max_prompt_length
|
||||||
from modules.torch_utils import clear_torch_cache
|
from modules.torch_utils import clear_torch_cache
|
||||||
|
|
|
||||||
|
|
@ -12,11 +12,11 @@ from pathlib import Path
|
||||||
import llama_cpp_binaries
|
import llama_cpp_binaries
|
||||||
import requests
|
import requests
|
||||||
|
|
||||||
from extensions.openai.image_utils import (
|
from modules import shared
|
||||||
|
from modules.image_utils import (
|
||||||
convert_image_attachments_to_pil,
|
convert_image_attachments_to_pil,
|
||||||
convert_pil_to_base64
|
convert_pil_to_base64
|
||||||
)
|
)
|
||||||
from modules import shared
|
|
||||||
from modules.logging_colors import logger
|
from modules.logging_colors import logger
|
||||||
|
|
||||||
llamacpp_valid_cache_types = {"fp16", "q8_0", "q4_0"}
|
llamacpp_valid_cache_types = {"fp16", "q8_0", "q4_0"}
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue