llama.cpp: Disable jinja by default (we use Python jinja, not cpp jinja)

This was causing template compilation issues with qwen models.
This commit is contained in:
oobabooga 2026-04-02 21:56:06 -07:00
parent 42dfcdfc5b
commit a1cb5b5dc0

View file

@ -418,6 +418,7 @@ class LlamaServer:
"--ubatch-size", str(shared.args.ubatch_size),
"--port", str(self.port),
"--no-webui",
"--no-jinja",
"--flash-attn", "on",
]