Commit graph

39 commits

Author SHA1 Message Date
oobabooga
552cb09f09 Do not bump Transformers to 4.52 on CUDA 12.8
Performance is slow, and the older version works fine with torch 2.7.
2025-06-10 18:45:42 -07:00
oobabooga
2dabdbc7da Update llama.cpp 2025-06-10 05:25:23 -07:00
oobabooga
263b5d5557 Use html2text to extract the text of web searches without losing formatting 2025-06-09 17:55:26 -07:00
rakha abadi susilo
db847eed4c
Add RTX 50XX Nvidia blackwell support (ExLlamaV2/V3 and Transformers) (#7011)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2025-06-07 21:44:15 -03:00
oobabooga
b38ec0ec38 Update llama.cpp 2025-06-02 11:33:17 -07:00
oobabooga
45c9ae312c Use the flash-attention wheels in https://github.com/kingbri1/flash-attention 2025-06-01 22:17:22 -07:00
oobabooga
1d88456659 Add support for .docx attachments 2025-05-31 20:15:07 -07:00
oobabooga
dc8ed6dbe7 Bump exllamav3 to 0.0.3 2025-05-31 14:27:33 -07:00
oobabooga
c55d3c61c6 Bump exllamav2 to 0.3.1 2025-05-31 14:21:42 -07:00
oobabooga
28e6bd4fcd Revert "Update transformers requirement in /requirements/full (#7017)"
This reverts commit cc9b7253c1.
2025-05-29 14:49:07 -07:00
oobabooga
8078c41ec6 Revert "Bump llama.cpp"
This reverts commit a8d02dec8f.
2025-05-29 13:32:19 -07:00
oobabooga
a8d02dec8f Bump llama.cpp 2025-05-29 04:24:21 -07:00
oobabooga
077bbc6b10
Add web search support (#7023) 2025-05-28 04:27:28 -03:00
dependabot[bot]
cc9b7253c1
Update transformers requirement in /requirements/full (#7017) 2025-05-26 23:13:10 -03:00
oobabooga
409a48d6bd
Add attachments support (text files, PDF documents) (#7005) 2025-05-21 00:36:20 -03:00
oobabooga
51c50b265d Update llama.cpp to b7a17463ec 2025-05-20 11:16:12 -07:00
oobabooga
0c1bc6d1d0 Bump llama.cpp 2025-05-18 14:08:54 -07:00
oobabooga
8cb73b78e1 Update ExLlamaV3 2025-05-15 20:10:34 -07:00
oobabooga
041248cc9f Update llama.cpp 2025-05-15 20:10:02 -07:00
oobabooga
a1b3307b66 Bump llama.cpp 2025-05-08 08:58:43 -07:00
oobabooga
13a434f351 Bump exllamav3 2025-05-08 08:06:07 -07:00
oobabooga
605cc9ab14 Update exllamav3 2025-05-06 06:43:35 -07:00
oobabooga
89590adc14 Update llama.cpp 2025-05-06 06:41:17 -07:00
oobabooga
d5c407cf35 Use Vulkan instead of ROCm for llama.cpp on AMD 2025-05-01 20:05:36 -07:00
oobabooga
c12a53c998 Use turboderp's exllamav2 wheels 2025-05-01 19:46:56 -07:00
oobabooga
a4bf339724 Bump llama.cpp 2025-04-30 11:13:14 -07:00
oobabooga
e9569c3984 Fixes after c5fe92d152 2025-04-30 06:57:23 -07:00
oobabooga
7f49e3c3ce Bump ExLlamaV3 2025-04-30 05:25:09 -07:00
oobabooga
c5fe92d152 Bump llama.cpp 2025-04-30 05:24:58 -07:00
oobabooga
fa861de05b Fix portable builds with Python 3.12 2025-04-26 18:52:44 -07:00
oobabooga
bf2aa19b21 Bump llama.cpp 2025-04-26 16:39:22 -07:00
oobabooga
2c7ff86015 Bump exllamav3 to de83084184 2025-04-25 05:28:22 -07:00
oobabooga
5993ebeb1b Bump exllamav2 to 0.2.9 2025-04-25 05:27:59 -07:00
oobabooga
8ebe868916 Fix typos in b313adf653 2025-04-24 09:32:17 -07:00
oobabooga
b313adf653 Bump llama.cpp, make the wheels work with any Python >= 3.7 2025-04-24 08:26:12 -07:00
oobabooga
06619e5f03 Add vulkan requirements.txt files 2025-04-22 17:46:54 -07:00
oobabooga
ee09e44c85
Portable version (#6868) 2025-04-22 09:25:57 -03:00
oobabooga
c178ea02fe Revert "Move the requirements*.txt to a requirements folder"
This reverts commit 6117ef7d64.
2025-04-20 19:27:38 -07:00
oobabooga
6117ef7d64 Move the requirements*.txt to a requirements folder 2025-04-20 19:12:04 -07:00