mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2025-12-06 07:12:10 +01:00
Changing transformers_loader.py to Match User Expectations for --bf16 and Flash Attention 2 (#7217)
This commit is contained in:
parent
9e9ab39892
commit
dd6d2223a5
|
|
@ -137,6 +137,7 @@ def load_model_HF(model_name):
|
||||||
params = {
|
params = {
|
||||||
'low_cpu_mem_usage': True,
|
'low_cpu_mem_usage': True,
|
||||||
'attn_implementation': shared.args.attn_implementation,
|
'attn_implementation': shared.args.attn_implementation,
|
||||||
|
'torch_dtype': torch.bfloat16 if shared.args.bf16 else torch.float16,
|
||||||
}
|
}
|
||||||
|
|
||||||
if shared.args.trust_remote_code:
|
if shared.args.trust_remote_code:
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue