mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2025-12-06 07:12:10 +01:00
Attempt at fixing new exllamav3 loader undefined behavior when switching conversations
This commit is contained in:
parent
a6d6bee88c
commit
d489eb589a
|
|
@ -304,9 +304,7 @@ class Exllamav3Model:
|
|||
response_text += chunk
|
||||
yield response_text
|
||||
finally:
|
||||
# No cleanup needed. MMEmbedding lifetime is managed by Python.
|
||||
# Cache and page table resets are unnecessary and can cause token ID conflicts.
|
||||
pass
|
||||
self.generator.clear_queue()
|
||||
|
||||
def generate(self, prompt, state):
|
||||
output = ""
|
||||
|
|
|
|||
Loading…
Reference in a new issue