mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2026-03-15 18:04:41 +01:00
llama.cpp: Fix manual random seeds
This commit is contained in:
parent
d5c407cf35
commit
9e3867dc83
|
|
@ -480,7 +480,7 @@ def generate_reply_custom(question, original_question, state, stopping_strings=N
|
|||
For models that do not use the transformers library for sampling
|
||||
"""
|
||||
|
||||
seed = set_manual_seed(state['seed'])
|
||||
state['seed'] = set_manual_seed(state['seed'])
|
||||
t0 = time.time()
|
||||
reply = ''
|
||||
try:
|
||||
|
|
|
|||
Loading…
Reference in a new issue