From 9e3867dc8358baf153d6f7c182496dad158696a4 Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Fri, 2 May 2025 09:36:15 -0700 Subject: [PATCH] llama.cpp: Fix manual random seeds --- modules/text_generation.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/modules/text_generation.py b/modules/text_generation.py index 8d091868..b9bf9b16 100644 --- a/modules/text_generation.py +++ b/modules/text_generation.py @@ -480,7 +480,7 @@ def generate_reply_custom(question, original_question, state, stopping_strings=N For models that do not use the transformers library for sampling """ - seed = set_manual_seed(state['seed']) + state['seed'] = set_manual_seed(state['seed']) t0 = time.time() reply = '' try: