mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2026-04-06 07:03:37 +00:00
llama.cpp: Remove the --flash-attn flag (it's always on now)
This commit is contained in:
parent
7b80e9a2ad
commit
13876a1ee8
5 changed files with 0 additions and 9 deletions
|
|
@ -318,8 +318,6 @@ class LlamaServer:
|
|||
"--no-webui",
|
||||
]
|
||||
|
||||
if shared.args.flash_attn:
|
||||
cmd.append("--flash-attn")
|
||||
if shared.args.threads > 0:
|
||||
cmd += ["--threads", str(shared.args.threads)]
|
||||
if shared.args.threads_batch > 0:
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue