Add StreamingLLM for llamacpp & llamacpp_HF (2nd attempt) (#5669)

This commit is contained in:
oobabooga 2024-03-09 00:25:33 -03:00 committed by GitHub
parent 9271e80914
commit afb51bd5d6
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
7 changed files with 147 additions and 0 deletions

View file

@ -97,6 +97,8 @@ def list_model_elements():
'no_offload_kqv',
'row_split',
'tensorcores',
'streaming_llm',
'attention_sink_size',
'hqq_backend',
]
if is_torch_xpu_available():