From 95d6c53e13673defecff6def4aead4c7ea157911 Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Fri, 3 Apr 2026 07:30:48 -0700 Subject: [PATCH] Revert "API: Add warning about vanilla llama-server not supporting prompt logprobs + instructions" This reverts commit 42dfcdfc5b50333c40a6adda0f4c8672508212cb. --- modules/llama_cpp_server.py | 6 ------ 1 file changed, 6 deletions(-) diff --git a/modules/llama_cpp_server.py b/modules/llama_cpp_server.py index 2d873f00..34080466 100644 --- a/modules/llama_cpp_server.py +++ b/modules/llama_cpp_server.py @@ -333,12 +333,6 @@ class LlamaServer: prompt_probs = result.get("prompt_probabilities", []) if not prompt_probs: - logger.warning( - "The llama.cpp server did not return prompt probabilities. " - "This feature requires a custom build with prompt_logprobs support. " - "See: https://github.com/oobabooga/llama.cpp/tree/prompt-logprobs " - "or https://github.com/oobabooga/ik_llama.cpp/tree/prompt-logprobs" - ) return [] # Null first token (no conditioning context); use empty string for BOS