oobabooga
|
bf5d85c922
|
Revert "Downloader: Gracefully handle '416 Range Not Satisfiable' when continuing downloads"
This reverts commit 1aa2b924d2.
|
2025-10-09 17:22:41 -07:00 |
|
oobabooga
|
0d03813e98
|
Update modules/chat.py
Co-authored-by: Copilot <175728472+Copilot@users.noreply.github.com>
|
2025-10-09 21:01:13 -03:00 |
|
oobabooga
|
748007f6ee
|
Revert "Update exllamav3 to 0.0.8"
This reverts commit 977ffbaa04.
|
2025-10-09 16:50:00 -07:00 |
|
dependabot[bot]
|
af3c70651c
|
Update bitsandbytes requirement in /requirements/full (#7255)
|
2025-10-09 19:53:34 -03:00 |
|
oobabooga
|
977ffbaa04
|
Update exllamav3 to 0.0.8
|
2025-10-09 15:53:14 -07:00 |
|
oobabooga
|
e0f0fae59d
|
Exllamav3: Add fla to requirements for qwen3-next
|
2025-10-09 13:03:48 -07:00 |
|
oobabooga
|
deb37b821b
|
Same as 7f06aec3a1 but for exllamav3_hf
|
2025-10-09 13:02:38 -07:00 |
|
oobabooga
|
7f06aec3a1
|
exllamav3: Implement the logits function for /v1/internal/logits
|
2025-10-09 11:24:25 -07:00 |
|
oobabooga
|
218dc01b51
|
Add fallbacks after 93aa7b3ed3
|
2025-10-09 10:59:34 -07:00 |
|
oobabooga
|
1aa2b924d2
|
Downloader: Gracefully handle '416 Range Not Satisfiable' when continuing downloads
|
2025-10-09 10:52:31 -07:00 |
|
oobabooga
|
0f3793d608
|
Update llama.cpp
|
2025-10-09 09:38:22 -07:00 |
|
oobabooga
|
282aa19189
|
Safer profile picture uploading
|
2025-10-09 09:26:35 -07:00 |
|
oobabooga
|
93aa7b3ed3
|
Better handle multigpu setups with transformers + bitsandbytes
|
2025-10-09 08:49:44 -07:00 |
|
Ionoclast Laboratories
|
d229dfe991
|
Fix portable apple intel requirement for llama binaries (issue #7238) (#7239)
|
2025-10-08 12:40:53 -03:00 |
|
oobabooga
|
292c91abbb
|
Update llama.cpp
|
2025-10-08 08:31:34 -07:00 |
|
oobabooga
|
f660e0836b
|
Merge branch 'main' into dev
|
2025-10-08 05:38:33 -07:00 |
|
oobabooga
|
898a3ed2fe
|
Add sponsor (Warp) to README <3
|
2025-10-07 18:33:28 -03:00 |
|
oobabooga
|
22997c134e
|
Merge remote-tracking branch 'refs/remotes/origin/dev' into dev
|
2025-10-05 20:34:49 -07:00 |
|
Remowylliams
|
38a7fd685d
|
chat.py fixes Instruct mode History
|
2025-10-05 11:34:47 -03:00 |
|
oobabooga
|
64829071e0
|
Update llama.cpp
|
2025-10-05 07:32:41 -07:00 |
|
oobabooga
|
0eb8543d74
|
Update transformers
|
2025-10-05 07:30:33 -07:00 |
|
oobabooga
|
b7effb22e0
|
Update exllamav3
|
2025-10-05 07:29:57 -07:00 |
|
oobabooga
|
042b828c73
|
Merge pull request #7231 from oobabooga/dev
Merge dev branch
|
2025-09-21 01:18:56 -03:00 |
|
oobabooga
|
8c9df34696
|
Update llama.cpp
|
2025-09-20 20:57:15 -07:00 |
|
oobabooga
|
1e863a7113
|
Fix exllamav3 ignoring the stop button
|
2025-09-19 16:12:50 -07:00 |
|
oobabooga
|
005fcf3f98
|
Formatting
|
2025-09-17 21:58:37 -07:00 |
|
oobabooga
|
e4412f0634
|
Slightly more robust syntax highlighting
|
2025-09-17 21:57:17 -07:00 |
|
stevenxdavis
|
dd6d2223a5
|
Changing transformers_loader.py to Match User Expectations for --bf16 and Flash Attention 2 (#7217)
|
2025-09-17 16:39:04 -03:00 |
|
oobabooga
|
9e9ab39892
|
Make exllamav3_hf and exllamav2_hf functional again
|
2025-09-17 12:29:22 -07:00 |
|
oobabooga
|
9c0a833a0a
|
Revert "Update bitsandbytes requirement in /requirements/full (#7193)"
This reverts commit fe15b67160.
|
2025-09-17 11:58:54 -07:00 |
|
oobabooga
|
8087a57fd8
|
Bump transformers to 4.56
|
2025-09-17 08:19:18 -07:00 |
|
dependabot[bot]
|
7131a478b9
|
Update safetensors requirement in /requirements/full (#7192)
|
2025-09-17 12:18:13 -03:00 |
|
dependabot[bot]
|
fe15b67160
|
Update bitsandbytes requirement in /requirements/full (#7193)
|
2025-09-17 12:17:58 -03:00 |
|
dependabot[bot]
|
8f731a566c
|
Update peft requirement from ==0.16.* to ==0.17.* in /requirements/full (#7172)
|
2025-09-17 12:17:16 -03:00 |
|
oobabooga
|
483927a5be
|
Update llama.cpp
|
2025-09-17 05:09:12 -07:00 |
|
oobabooga
|
557b78d31e
|
Update llama.cpp
|
2025-09-03 16:50:03 -07:00 |
|
oobabooga
|
ba62783b72
|
UI: Don't use $ $ for LaTeX, only $$ $$
|
2025-09-02 14:22:22 -07:00 |
|
oobabooga
|
d3a7710c62
|
Merge pull request #7215 from oobabooga/dev
Merge dev branch
|
2025-09-02 16:51:50 -03:00 |
|
oobabooga
|
f3829b268a
|
llama.cpp: Always pass --flash-attn on
|
2025-09-02 12:12:17 -07:00 |
|
oobabooga
|
2395c647d4
|
Fix the instruct message height on mobile
|
2025-09-02 12:11:15 -07:00 |
|
oobabooga
|
c6ea67bbdb
|
Lint
|
2025-09-02 10:22:03 -07:00 |
|
oobabooga
|
00ed878b05
|
Slightly more robust model loading
|
2025-09-02 10:16:26 -07:00 |
|
oobabooga
|
d843afcf66
|
Update llama.cpp
|
2025-09-02 05:43:33 -07:00 |
|
oobabooga
|
00ebb295d3
|
Update llama.cpp
|
2025-08-31 16:27:23 -07:00 |
|
oobabooga
|
387e249dec
|
Change an info message
|
2025-08-31 16:27:10 -07:00 |
|
oobabooga
|
8028d88541
|
Lint
|
2025-08-30 21:29:20 -07:00 |
|
oobabooga
|
13876a1ee8
|
llama.cpp: Remove the --flash-attn flag (it's always on now)
|
2025-08-30 20:28:26 -07:00 |
|
oobabooga
|
7b80e9a2ad
|
Update llama.cpp
|
2025-08-30 20:22:11 -07:00 |
|
oobabooga
|
5631d4e3d6
|
Minor change after 21d790f87e
|
2025-08-30 15:34:49 -07:00 |
|
oobabooga
|
5920ad8834
|
UI: Give streaming instruct messages more vertical space
|
2025-08-30 15:22:50 -07:00 |
|