From 6ec4ca8b102725d2ebb810c07ad668179835a567 Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Tue, 10 Mar 2026 09:56:07 -0300 Subject: [PATCH] Add missing custom_token_bans to llama.cpp and reasoning_effort to ExLlamav3 --- modules/loaders.py | 2 ++ 1 file changed, 2 insertions(+) diff --git a/modules/loaders.py b/modules/loaders.py index 22ee5ed9..d2ebdbc3 100644 --- a/modules/loaders.py +++ b/modules/loaders.py @@ -205,6 +205,7 @@ loaders_samplers = { 'ban_eos_token', 'add_bos_token', 'enable_thinking', + 'reasoning_effort', 'seed', 'skip_special_tokens', }, @@ -241,6 +242,7 @@ loaders_samplers = { 'reasoning_effort', 'seed', 'sampler_priority', + 'custom_token_bans', 'dry_sequence_breakers', 'grammar_string', 'grammar_file_row',