Update loaders.py

This commit is contained in:
Hermann Hans Klie 2025-10-24 09:06:29 +03:00 committed by GitHub
parent 779795266f
commit eb6c5a171e
No known key found for this signature in database
GPG key ID: B5690EEEBB952194

View file

@ -46,6 +46,17 @@ loaders_and_params = OrderedDict({
'use_double_quant',
'bf16',
'no_use_fast',
],
'ktransformers': [
'ctx_size',
'gpu_split',
'cache_type',
'cpu', # CPU-Offload (HF accelerate)
'disk', # Disk-Offload (HF accelerate)
'cpu_memory', # z.B. "48GiB" (String)
'quant_type', # falls du 4/8-bit via bitsandbytes/awq testen willst
'compute_dtype', # bf16/fp16 perhaps torch_dtype would be better here ???
'attn_implementation', # sdpa/flash_attention_2 (je nach Build)
],
'ExLlamav3_HF': [
'ctx_size',
@ -163,6 +174,15 @@ def transformers_samplers():
loaders_samplers = {
'Transformers': transformers_samplers(),
'ktransformers': {
'temperature',
'top_p',
'top_k',
'repetition_penalty',
'presence_penalty',
'frequency_penalty',
'seed',
},
'ExLlamav3_HF': {
'temperature',
'dynatemp_low',