API: Add warning about vanilla llama-server not supporting prompt logprobs + instructions

This commit is contained in:
oobabooga 2026-04-02 20:46:27 -07:00
parent 6e2b70bde6
commit 42dfcdfc5b

View file

@ -333,6 +333,12 @@ class LlamaServer:
prompt_probs = result.get("prompt_probabilities", [])
if not prompt_probs:
logger.warning(
"The llama.cpp server did not return prompt probabilities. "
"This feature requires a custom build with prompt_logprobs support. "
"See: https://github.com/oobabooga/llama.cpp/tree/prompt-logprobs "
"or https://github.com/oobabooga/ik_llama.cpp/tree/prompt-logprobs"
)
return []
# Null first token (no conditioning context); use empty string for BOS