Commit graph

39 commits

Author SHA1 Message Date
oobabooga 552cb09f09 Do not bump Transformers to 4.52 on CUDA 12.8
Performance is slow, and the older version works fine with torch 2.7.
2025-06-10 18:45:42 -07:00
oobabooga 2dabdbc7da Update llama.cpp 2025-06-10 05:25:23 -07:00
oobabooga 263b5d5557 Use html2text to extract the text of web searches without losing formatting 2025-06-09 17:55:26 -07:00
rakha abadi susilo db847eed4c
Add RTX 50XX Nvidia blackwell support (ExLlamaV2/V3 and Transformers) (#7011)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2025-06-07 21:44:15 -03:00
oobabooga b38ec0ec38 Update llama.cpp 2025-06-02 11:33:17 -07:00
oobabooga 45c9ae312c Use the flash-attention wheels in https://github.com/kingbri1/flash-attention 2025-06-01 22:17:22 -07:00
oobabooga 1d88456659 Add support for .docx attachments 2025-05-31 20:15:07 -07:00
oobabooga dc8ed6dbe7 Bump exllamav3 to 0.0.3 2025-05-31 14:27:33 -07:00
oobabooga c55d3c61c6 Bump exllamav2 to 0.3.1 2025-05-31 14:21:42 -07:00
oobabooga 28e6bd4fcd Revert "Update transformers requirement in /requirements/full (#7017)"
This reverts commit cc9b7253c1.
2025-05-29 14:49:07 -07:00
oobabooga 8078c41ec6 Revert "Bump llama.cpp"
This reverts commit a8d02dec8f.
2025-05-29 13:32:19 -07:00
oobabooga a8d02dec8f Bump llama.cpp 2025-05-29 04:24:21 -07:00
oobabooga 077bbc6b10
Add web search support (#7023) 2025-05-28 04:27:28 -03:00
dependabot[bot] cc9b7253c1
Update transformers requirement in /requirements/full (#7017) 2025-05-26 23:13:10 -03:00
oobabooga 409a48d6bd
Add attachments support (text files, PDF documents) (#7005) 2025-05-21 00:36:20 -03:00
oobabooga 51c50b265d Update llama.cpp to b7a17463ec 2025-05-20 11:16:12 -07:00
oobabooga 0c1bc6d1d0 Bump llama.cpp 2025-05-18 14:08:54 -07:00
oobabooga 8cb73b78e1 Update ExLlamaV3 2025-05-15 20:10:34 -07:00
oobabooga 041248cc9f Update llama.cpp 2025-05-15 20:10:02 -07:00
oobabooga a1b3307b66 Bump llama.cpp 2025-05-08 08:58:43 -07:00
oobabooga 13a434f351 Bump exllamav3 2025-05-08 08:06:07 -07:00
oobabooga 605cc9ab14 Update exllamav3 2025-05-06 06:43:35 -07:00
oobabooga 89590adc14 Update llama.cpp 2025-05-06 06:41:17 -07:00
oobabooga d5c407cf35 Use Vulkan instead of ROCm for llama.cpp on AMD 2025-05-01 20:05:36 -07:00
oobabooga c12a53c998 Use turboderp's exllamav2 wheels 2025-05-01 19:46:56 -07:00
oobabooga a4bf339724 Bump llama.cpp 2025-04-30 11:13:14 -07:00
oobabooga e9569c3984 Fixes after c5fe92d152 2025-04-30 06:57:23 -07:00
oobabooga 7f49e3c3ce Bump ExLlamaV3 2025-04-30 05:25:09 -07:00
oobabooga c5fe92d152 Bump llama.cpp 2025-04-30 05:24:58 -07:00
oobabooga fa861de05b Fix portable builds with Python 3.12 2025-04-26 18:52:44 -07:00
oobabooga bf2aa19b21 Bump llama.cpp 2025-04-26 16:39:22 -07:00
oobabooga 2c7ff86015 Bump exllamav3 to de83084184 2025-04-25 05:28:22 -07:00
oobabooga 5993ebeb1b Bump exllamav2 to 0.2.9 2025-04-25 05:27:59 -07:00
oobabooga 8ebe868916 Fix typos in b313adf653 2025-04-24 09:32:17 -07:00
oobabooga b313adf653 Bump llama.cpp, make the wheels work with any Python >= 3.7 2025-04-24 08:26:12 -07:00
oobabooga 06619e5f03 Add vulkan requirements.txt files 2025-04-22 17:46:54 -07:00
oobabooga ee09e44c85
Portable version (#6868) 2025-04-22 09:25:57 -03:00
oobabooga c178ea02fe Revert "Move the requirements*.txt to a requirements folder"
This reverts commit 6117ef7d64.
2025-04-20 19:27:38 -07:00
oobabooga 6117ef7d64 Move the requirements*.txt to a requirements folder 2025-04-20 19:12:04 -07:00