mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2026-04-20 22:13:43 +00:00
Estimate the VRAM for GGUF models + autoset gpu-layers (#6980)
This commit is contained in:
parent
c4a715fd1e
commit
5534d01da0
6 changed files with 193 additions and 4 deletions
|
|
@ -71,7 +71,6 @@ def llama_cpp_server_loader(model_name):
|
|||
else:
|
||||
model_file = sorted(Path(f'{shared.args.model_dir}/{model_name}').glob('*.gguf'))[0]
|
||||
|
||||
logger.info(f"llama.cpp weights detected: \"{model_file}\"")
|
||||
try:
|
||||
model = LlamaServer(model_file)
|
||||
return model, model
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue