mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2025-12-06 07:12:10 +01:00
mtmd: Update the llama-server request
This commit is contained in:
parent
c47e6deda2
commit
e6447cd24a
|
|
@ -149,11 +149,10 @@ class LlamaServer:
|
||||||
IMAGE_TOKEN_COST_ESTIMATE = 600 # A safe, conservative estimate per image
|
IMAGE_TOKEN_COST_ESTIMATE = 600 # A safe, conservative estimate per image
|
||||||
|
|
||||||
base64_images = [convert_pil_to_base64(img) for img in pil_images]
|
base64_images = [convert_pil_to_base64(img) for img in pil_images]
|
||||||
multimodal_prompt_object = {
|
payload["prompt"] = {
|
||||||
"prompt": prompt,
|
"prompt_string": prompt,
|
||||||
"multimodal_data": base64_images
|
"multimodal_data": base64_images
|
||||||
}
|
}
|
||||||
payload["prompt"] = multimodal_prompt_object
|
|
||||||
|
|
||||||
# Calculate an estimated token count
|
# Calculate an estimated token count
|
||||||
text_tokens = self.encode(prompt, add_bos_token=state["add_bos_token"])
|
text_tokens = self.encode(prompt, add_bos_token=state["add_bos_token"])
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue