From 7170a16b91dab7e5de5cd01e01d5239050995474 Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Sun, 8 Mar 2026 04:09:18 -0700 Subject: [PATCH] Fix passing adaptive-p to llama-server --- modules/llama_cpp_server.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/modules/llama_cpp_server.py b/modules/llama_cpp_server.py index 12ff173e..6f7cbd20 100644 --- a/modules/llama_cpp_server.py +++ b/modules/llama_cpp_server.py @@ -129,7 +129,7 @@ class LlamaServer: # places it at the end of the chain regardless of position, so we # activate it based on the parameter value rather than sampler order. if state.get("adaptive_target", 0) > 0: - filtered_samplers.append("adaptive-p") + filtered_samplers.append("adaptive_p") payload["samplers"] = filtered_samplers