mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2026-03-18 03:14:39 +01:00
API: Minor logprobs fixes
This commit is contained in:
parent
2466305f76
commit
d45c9b3c59
|
|
@ -98,7 +98,7 @@ def format_chat_logprobs(entries):
|
|||
"top_logprobs": top_list
|
||||
})
|
||||
|
||||
return {"content": content} if content else None
|
||||
return {"content": content, "refusal": None} if content else None
|
||||
|
||||
|
||||
def format_completion_logprobs(entries):
|
||||
|
|
@ -174,7 +174,7 @@ def process_parameters(body, is_legacy=False):
|
|||
logprobs = body.get('logprobs', None)
|
||||
top_logprobs = body.get('top_logprobs', None)
|
||||
if logprobs is True:
|
||||
logprobs = top_logprobs if top_logprobs and top_logprobs > 0 else 5
|
||||
logprobs = max(top_logprobs, 1) if top_logprobs is not None else 5
|
||||
generate_params['logprobs'] = logprobs
|
||||
|
||||
# For llama.cpp and ExLlamav3 native, logit_bias and logprobs are forwarded natively
|
||||
|
|
@ -677,6 +677,9 @@ def completions_common(body: dict, is_legacy: bool = False, stream=False, stop_e
|
|||
if original_seed >= 0:
|
||||
generate_params['seed'] = original_seed + _n
|
||||
|
||||
if logprob_proc:
|
||||
logprob_proc.token_alternatives_history.clear()
|
||||
|
||||
# generate reply #######################################
|
||||
debug_msg({'prompt': prompt, 'generate_params': generate_params})
|
||||
generator = generate_reply(prompt, generate_params, is_chat=False)
|
||||
|
|
|
|||
Loading…
Reference in a new issue