llama.cpp: Fix obtaining the maximum sequence length for GPT-OSS

This commit is contained in:
oobabooga 2025-08-27 16:15:20 -07:00
parent 030ba7bfeb
commit a92758a144

View file

@ -68,7 +68,7 @@ def get_model_metadata(model):
metadata = load_gguf_metadata_with_cache(model_file)
for k in metadata:
if k.endswith('context_length'):
if k.endswith('.context_length'):
model_settings['ctx_size'] = min(metadata[k], 8192)
model_settings['truncation_length_info'] = metadata[k]
elif k.endswith('rope.freq_base'):