oobabooga
|
0f3793d608
|
Update llama.cpp
|
2025-10-09 09:38:22 -07:00 |
|
Ionoclast Laboratories
|
d229dfe991
|
Fix portable apple intel requirement for llama binaries (issue #7238) (#7239)
|
2025-10-08 12:40:53 -03:00 |
|
oobabooga
|
292c91abbb
|
Update llama.cpp
|
2025-10-08 08:31:34 -07:00 |
|
oobabooga
|
64829071e0
|
Update llama.cpp
|
2025-10-05 07:32:41 -07:00 |
|
oobabooga
|
0eb8543d74
|
Update transformers
|
2025-10-05 07:30:33 -07:00 |
|
oobabooga
|
b7effb22e0
|
Update exllamav3
|
2025-10-05 07:29:57 -07:00 |
|
oobabooga
|
8c9df34696
|
Update llama.cpp
|
2025-09-20 20:57:15 -07:00 |
|
oobabooga
|
9c0a833a0a
|
Revert "Update bitsandbytes requirement in /requirements/full (#7193)"
This reverts commit fe15b67160.
|
2025-09-17 11:58:54 -07:00 |
|
oobabooga
|
8087a57fd8
|
Bump transformers to 4.56
|
2025-09-17 08:19:18 -07:00 |
|
dependabot[bot]
|
7131a478b9
|
Update safetensors requirement in /requirements/full (#7192)
|
2025-09-17 12:18:13 -03:00 |
|
dependabot[bot]
|
fe15b67160
|
Update bitsandbytes requirement in /requirements/full (#7193)
|
2025-09-17 12:17:58 -03:00 |
|
dependabot[bot]
|
8f731a566c
|
Update peft requirement from ==0.16.* to ==0.17.* in /requirements/full (#7172)
|
2025-09-17 12:17:16 -03:00 |
|
oobabooga
|
483927a5be
|
Update llama.cpp
|
2025-09-17 05:09:12 -07:00 |
|
oobabooga
|
557b78d31e
|
Update llama.cpp
|
2025-09-03 16:50:03 -07:00 |
|
oobabooga
|
d843afcf66
|
Update llama.cpp
|
2025-09-02 05:43:33 -07:00 |
|
oobabooga
|
00ebb295d3
|
Update llama.cpp
|
2025-08-31 16:27:23 -07:00 |
|
oobabooga
|
7b80e9a2ad
|
Update llama.cpp
|
2025-08-30 20:22:11 -07:00 |
|
oobabooga
|
8042f76399
|
Make portable installs functional with Python 3.13
|
2025-08-27 05:37:01 -07:00 |
|
oobabooga
|
1f77427088
|
Update llama.cpp
|
2025-08-24 19:56:22 -07:00 |
|
oobabooga
|
fd41f2fafc
|
Update llama.cpp
|
2025-08-22 11:18:56 -07:00 |
|
oobabooga
|
8805a50d24
|
Update llama.cpp
|
2025-08-18 15:31:01 -07:00 |
|
oobabooga
|
3a91ca2dd1
|
Update flash attention
|
2025-08-17 13:57:23 -07:00 |
|
oobabooga
|
9651b5c873
|
Make CUDA 12.8 the default CUDA option, remove the CUDA 12.4 option
Exllamav3 doesn't compile with torch 2.6 anymore, and torch 2.7
requires newer CUDA.
|
2025-08-17 13:26:09 -07:00 |
|
oobabooga
|
a633793a00
|
Bump exllamav3 to 0.0.6
|
2025-08-17 13:19:42 -07:00 |
|
oobabooga
|
0882970a94
|
Update llama.cpp
|
2025-08-12 07:00:24 -07:00 |
|
oobabooga
|
1e3c4e8bdb
|
Update llama.cpp
|
2025-08-11 14:40:59 -07:00 |
|
oobabooga
|
eb16f64017
|
Update llama.cpp
|
2025-08-09 17:12:16 -07:00 |
|
oobabooga
|
f1147c9926
|
Update llama.cpp
|
2025-08-06 19:32:36 -07:00 |
|
oobabooga
|
42e3a7a5ae
|
Update llama.cpp
|
2025-08-05 14:56:12 -07:00 |
|
oobabooga
|
02a3420a50
|
Bump transformers to 4.55 (adds gpt-oss support)
|
2025-08-05 10:09:30 -07:00 |
|
oobabooga
|
74230f559a
|
Bump transformers to 4.54
|
2025-08-01 11:03:15 -07:00 |
|
oobabooga
|
ccf5e3e3a7
|
Update exllamav3
|
2025-07-19 12:07:38 -07:00 |
|
oobabooga
|
a00983b2ba
|
Update llama.cpp
|
2025-07-19 12:07:20 -07:00 |
|
oobabooga
|
9371867238
|
Update exllamav2
|
2025-07-15 07:38:03 -07:00 |
|
oobabooga
|
03fb85e49a
|
Update llama.cpp
|
2025-07-15 07:37:13 -07:00 |
|
oobabooga
|
5a8a9c22e8
|
Update llama.cpp
|
2025-07-11 09:20:27 -07:00 |
|
oobabooga
|
0f3a88057c
|
Don't downgrade triton-windows on CUDA 12.8
|
2025-07-10 05:39:04 -07:00 |
|
oobabooga
|
e523f25b9f
|
Downgrade triton-windows to 3.2.0.post19
https://github.com/oobabooga/text-generation-webui/issues/7107#issuecomment-3057250374
|
2025-07-10 05:35:57 -07:00 |
|
oobabooga
|
a7a3a0c700
|
Merge remote-tracking branch 'refs/remotes/origin/dev' into dev
|
2025-07-09 21:07:42 -07:00 |
|
oobabooga
|
21e0e9f32b
|
Add the triton-windows requirement on Windows to make transformers functional
|
2025-07-09 21:05:17 -07:00 |
|
dependabot[bot]
|
d1f4622a96
|
Update peft requirement from ==0.15.* to ==0.16.* in /requirements/full (#7127)
|
2025-07-10 00:15:50 -03:00 |
|
oobabooga
|
8b3c7aa795
|
Bump bitsandbytes to 0.46
|
2025-07-09 19:46:55 -07:00 |
|
oobabooga
|
f045b72826
|
Bump accelerate to 1.8
|
2025-07-09 19:46:26 -07:00 |
|
oobabooga
|
c357601c01
|
Bump transformers to 4.53
|
2025-07-09 18:48:04 -07:00 |
|
oobabooga
|
e52bc0acb2
|
Update llama.cpp
|
2025-07-06 20:28:35 -07:00 |
|
oobabooga
|
de4ccffff8
|
Fix the duckduckgo search
|
2025-07-06 16:24:57 -07:00 |
|
oobabooga
|
c4d738f39f
|
Update llama.cpp
|
2025-07-05 14:09:29 -07:00 |
|
oobabooga
|
23bb94a5fb
|
Update llama.cpp
|
2025-07-03 20:36:54 -07:00 |
|
oobabooga
|
93cd47c948
|
Bump numpy to 2.2 (loses #7090)
|
2025-06-19 08:00:30 -07:00 |
|
oobabooga
|
315e06f695
|
Update llama.cpp
|
2025-06-17 07:51:16 -07:00 |
|