mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2025-12-06 07:12:10 +01:00
Changing transformers_loader.py to Match User Expectations for --bf16 and Flash Attention 2 (#7217)
This commit is contained in:
parent
9e9ab39892
commit
dd6d2223a5
|
|
@ -137,6 +137,7 @@ def load_model_HF(model_name):
|
|||
params = {
|
||||
'low_cpu_mem_usage': True,
|
||||
'attn_implementation': shared.args.attn_implementation,
|
||||
'torch_dtype': torch.bfloat16 if shared.args.bf16 else torch.float16,
|
||||
}
|
||||
|
||||
if shared.args.trust_remote_code:
|
||||
|
|
|
|||
Loading…
Reference in a new issue