mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2025-12-06 07:12:10 +01:00
llama.cpp: Fix obtaining the maximum sequence length for GPT-OSS
This commit is contained in:
parent
030ba7bfeb
commit
a92758a144
|
|
@ -68,7 +68,7 @@ def get_model_metadata(model):
|
|||
metadata = load_gguf_metadata_with_cache(model_file)
|
||||
|
||||
for k in metadata:
|
||||
if k.endswith('context_length'):
|
||||
if k.endswith('.context_length'):
|
||||
model_settings['ctx_size'] = min(metadata[k], 8192)
|
||||
model_settings['truncation_length_info'] = metadata[k]
|
||||
elif k.endswith('rope.freq_base'):
|
||||
|
|
|
|||
Loading…
Reference in a new issue