From 9488df3e489c97cc26018d9ae1dc6a4bc0384f1b Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Sat, 21 Mar 2026 20:47:26 -0700 Subject: [PATCH] llama.cpp: Don't suppress llama-server logs --- modules/llama_cpp_server.py | 7 +++++-- 1 file changed, 5 insertions(+), 2 deletions(-) diff --git a/modules/llama_cpp_server.py b/modules/llama_cpp_server.py index 2ae01ddc..b77a8605 100644 --- a/modules/llama_cpp_server.py +++ b/modules/llama_cpp_server.py @@ -588,8 +588,11 @@ def filter_stderr_with_progress(process_stderr): print(display_line, end=end_char, file=sys.stderr, flush=True) last_was_progress = (progress < 1.0) - # skip noise lines - elif not (line.startswith(('srv ', 'slot ')) or 'log_server_r: request: GET /health' in line or 'No parser definition detected' in line): + # skip health check polling and parser warnings + elif 'log_server_r: request: GET /health' in line or 'No parser definition detected' in line: + continue + + else: # if we were in progress, finish that line first if last_was_progress: print(file=sys.stderr)