oobabooga
|
e8dc7b0ee9
|
Bump exllamav3 to 0.0.4
|
2025-06-15 08:15:29 -07:00 |
|
oobabooga
|
ad0be25c46
|
Update llama.cpp
|
2025-06-14 15:00:14 -07:00 |
|
oobabooga
|
552cb09f09
|
Do not bump Transformers to 4.52 on CUDA 12.8
Performance is slow, and the older version works fine with torch 2.7.
|
2025-06-10 18:45:42 -07:00 |
|
oobabooga
|
2dabdbc7da
|
Update llama.cpp
|
2025-06-10 05:25:23 -07:00 |
|
oobabooga
|
263b5d5557
|
Use html2text to extract the text of web searches without losing formatting
|
2025-06-09 17:55:26 -07:00 |
|
rakha abadi susilo
|
db847eed4c
|
Add RTX 50XX Nvidia blackwell support (ExLlamaV2/V3 and Transformers) (#7011)
---------
Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
|
2025-06-07 21:44:15 -03:00 |
|
oobabooga
|
b38ec0ec38
|
Update llama.cpp
|
2025-06-02 11:33:17 -07:00 |
|
oobabooga
|
45c9ae312c
|
Use the flash-attention wheels in https://github.com/kingbri1/flash-attention
|
2025-06-01 22:17:22 -07:00 |
|
oobabooga
|
1d88456659
|
Add support for .docx attachments
|
2025-05-31 20:15:07 -07:00 |
|
oobabooga
|
dc8ed6dbe7
|
Bump exllamav3 to 0.0.3
|
2025-05-31 14:27:33 -07:00 |
|
oobabooga
|
c55d3c61c6
|
Bump exllamav2 to 0.3.1
|
2025-05-31 14:21:42 -07:00 |
|
oobabooga
|
28e6bd4fcd
|
Revert "Update transformers requirement in /requirements/full (#7017)"
This reverts commit cc9b7253c1.
|
2025-05-29 14:49:07 -07:00 |
|
oobabooga
|
8078c41ec6
|
Revert "Bump llama.cpp"
This reverts commit a8d02dec8f.
|
2025-05-29 13:32:19 -07:00 |
|
oobabooga
|
a8d02dec8f
|
Bump llama.cpp
|
2025-05-29 04:24:21 -07:00 |
|
oobabooga
|
077bbc6b10
|
Add web search support (#7023)
|
2025-05-28 04:27:28 -03:00 |
|
dependabot[bot]
|
cc9b7253c1
|
Update transformers requirement in /requirements/full (#7017)
|
2025-05-26 23:13:10 -03:00 |
|
oobabooga
|
409a48d6bd
|
Add attachments support (text files, PDF documents) (#7005)
|
2025-05-21 00:36:20 -03:00 |
|
oobabooga
|
51c50b265d
|
Update llama.cpp to b7a17463ec
|
2025-05-20 11:16:12 -07:00 |
|
oobabooga
|
0c1bc6d1d0
|
Bump llama.cpp
|
2025-05-18 14:08:54 -07:00 |
|
oobabooga
|
8cb73b78e1
|
Update ExLlamaV3
|
2025-05-15 20:10:34 -07:00 |
|
oobabooga
|
041248cc9f
|
Update llama.cpp
|
2025-05-15 20:10:02 -07:00 |
|
oobabooga
|
a1b3307b66
|
Bump llama.cpp
|
2025-05-08 08:58:43 -07:00 |
|
oobabooga
|
13a434f351
|
Bump exllamav3
|
2025-05-08 08:06:07 -07:00 |
|
oobabooga
|
605cc9ab14
|
Update exllamav3
|
2025-05-06 06:43:35 -07:00 |
|
oobabooga
|
89590adc14
|
Update llama.cpp
|
2025-05-06 06:41:17 -07:00 |
|
oobabooga
|
d5c407cf35
|
Use Vulkan instead of ROCm for llama.cpp on AMD
|
2025-05-01 20:05:36 -07:00 |
|
oobabooga
|
c12a53c998
|
Use turboderp's exllamav2 wheels
|
2025-05-01 19:46:56 -07:00 |
|
oobabooga
|
a4bf339724
|
Bump llama.cpp
|
2025-04-30 11:13:14 -07:00 |
|
oobabooga
|
e9569c3984
|
Fixes after c5fe92d152
|
2025-04-30 06:57:23 -07:00 |
|
oobabooga
|
7f49e3c3ce
|
Bump ExLlamaV3
|
2025-04-30 05:25:09 -07:00 |
|
oobabooga
|
c5fe92d152
|
Bump llama.cpp
|
2025-04-30 05:24:58 -07:00 |
|
oobabooga
|
fa861de05b
|
Fix portable builds with Python 3.12
|
2025-04-26 18:52:44 -07:00 |
|
oobabooga
|
bf2aa19b21
|
Bump llama.cpp
|
2025-04-26 16:39:22 -07:00 |
|
oobabooga
|
2c7ff86015
|
Bump exllamav3 to de83084184
|
2025-04-25 05:28:22 -07:00 |
|
oobabooga
|
5993ebeb1b
|
Bump exllamav2 to 0.2.9
|
2025-04-25 05:27:59 -07:00 |
|
oobabooga
|
8ebe868916
|
Fix typos in b313adf653
|
2025-04-24 09:32:17 -07:00 |
|
oobabooga
|
b313adf653
|
Bump llama.cpp, make the wheels work with any Python >= 3.7
|
2025-04-24 08:26:12 -07:00 |
|
oobabooga
|
06619e5f03
|
Add vulkan requirements.txt files
|
2025-04-22 17:46:54 -07:00 |
|
oobabooga
|
ee09e44c85
|
Portable version (#6868)
|
2025-04-22 09:25:57 -03:00 |
|
oobabooga
|
c178ea02fe
|
Revert "Move the requirements*.txt to a requirements folder"
This reverts commit 6117ef7d64.
|
2025-04-20 19:27:38 -07:00 |
|
oobabooga
|
6117ef7d64
|
Move the requirements*.txt to a requirements folder
|
2025-04-20 19:12:04 -07:00 |
|