Commit graph

5082 commits

Author SHA1 Message Date
oobabooga 1831b3fb51 Use my custom gradio_client build (small changes to work with pydantic 2.11) 2025-10-10 18:01:21 -07:00
oobabooga dd0b003493 Bump pydantic to 2.11.0 2025-10-10 17:52:16 -07:00
oobabooga a74596374d Reapply "Update exllamav3 to 0.0.8"
This reverts commit 748007f6ee.
2025-10-10 17:51:31 -07:00
oobabooga 78ff21d512 Organize the --help message 2025-10-10 15:21:08 -07:00
oobabooga 5d734cc7ca Remove unused CSS 2025-10-10 12:54:54 -07:00
oobabooga 25360387ec Downloader: Fix resuming downloads after HF moved to Xet 2025-10-10 08:27:40 -07:00
oobabooga 7833650aa1
Merge pull request #7260 from oobabooga/dev
Merge dev branch
2025-10-10 10:46:34 -03:00
oobabooga bf5d85c922 Revert "Downloader: Gracefully handle '416 Range Not Satisfiable' when continuing downloads"
This reverts commit 1aa2b924d2.
2025-10-09 17:22:41 -07:00
oobabooga 0d03813e98
Update modules/chat.py
Co-authored-by: Copilot <175728472+Copilot@users.noreply.github.com>
2025-10-09 21:01:13 -03:00
oobabooga 748007f6ee Revert "Update exllamav3 to 0.0.8"
This reverts commit 977ffbaa04.
2025-10-09 16:50:00 -07:00
dependabot[bot] af3c70651c
Update bitsandbytes requirement in /requirements/full (#7255) 2025-10-09 19:53:34 -03:00
oobabooga 977ffbaa04 Update exllamav3 to 0.0.8 2025-10-09 15:53:14 -07:00
oobabooga e0f0fae59d Exllamav3: Add fla to requirements for qwen3-next 2025-10-09 13:03:48 -07:00
oobabooga deb37b821b Same as 7f06aec3a1 but for exllamav3_hf 2025-10-09 13:02:38 -07:00
oobabooga 7f06aec3a1 exllamav3: Implement the logits function for /v1/internal/logits 2025-10-09 11:24:25 -07:00
oobabooga 218dc01b51 Add fallbacks after 93aa7b3ed3 2025-10-09 10:59:34 -07:00
oobabooga 1aa2b924d2 Downloader: Gracefully handle '416 Range Not Satisfiable' when continuing downloads 2025-10-09 10:52:31 -07:00
oobabooga 0f3793d608 Update llama.cpp 2025-10-09 09:38:22 -07:00
oobabooga 282aa19189 Safer profile picture uploading 2025-10-09 09:26:35 -07:00
oobabooga 93aa7b3ed3 Better handle multigpu setups with transformers + bitsandbytes 2025-10-09 08:49:44 -07:00
Ionoclast Laboratories d229dfe991
Fix portable apple intel requirement for llama binaries (issue #7238) (#7239) 2025-10-08 12:40:53 -03:00
oobabooga 292c91abbb Update llama.cpp 2025-10-08 08:31:34 -07:00
oobabooga f660e0836b Merge branch 'main' into dev 2025-10-08 05:38:33 -07:00
oobabooga 898a3ed2fe
Add sponsor (Warp) to README <3 2025-10-07 18:33:28 -03:00
oobabooga 22997c134e Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2025-10-05 20:34:49 -07:00
Remowylliams 38a7fd685d
chat.py fixes Instruct mode History 2025-10-05 11:34:47 -03:00
oobabooga 64829071e0 Update llama.cpp 2025-10-05 07:32:41 -07:00
oobabooga 0eb8543d74 Update transformers 2025-10-05 07:30:33 -07:00
oobabooga b7effb22e0 Update exllamav3 2025-10-05 07:29:57 -07:00
oobabooga 042b828c73
Merge pull request #7231 from oobabooga/dev
Merge dev branch
2025-09-21 01:18:56 -03:00
oobabooga 8c9df34696 Update llama.cpp 2025-09-20 20:57:15 -07:00
oobabooga 1e863a7113 Fix exllamav3 ignoring the stop button 2025-09-19 16:12:50 -07:00
oobabooga 005fcf3f98 Formatting 2025-09-17 21:58:37 -07:00
oobabooga e4412f0634 Slightly more robust syntax highlighting 2025-09-17 21:57:17 -07:00
stevenxdavis dd6d2223a5
Changing transformers_loader.py to Match User Expectations for --bf16 and Flash Attention 2 (#7217) 2025-09-17 16:39:04 -03:00
oobabooga 9e9ab39892 Make exllamav3_hf and exllamav2_hf functional again 2025-09-17 12:29:22 -07:00
oobabooga 9c0a833a0a Revert "Update bitsandbytes requirement in /requirements/full (#7193)"
This reverts commit fe15b67160.
2025-09-17 11:58:54 -07:00
oobabooga 8087a57fd8 Bump transformers to 4.56 2025-09-17 08:19:18 -07:00
dependabot[bot] 7131a478b9
Update safetensors requirement in /requirements/full (#7192) 2025-09-17 12:18:13 -03:00
dependabot[bot] fe15b67160
Update bitsandbytes requirement in /requirements/full (#7193) 2025-09-17 12:17:58 -03:00
dependabot[bot] 8f731a566c
Update peft requirement from ==0.16.* to ==0.17.* in /requirements/full (#7172) 2025-09-17 12:17:16 -03:00
oobabooga 483927a5be Update llama.cpp 2025-09-17 05:09:12 -07:00
oobabooga 557b78d31e Update llama.cpp 2025-09-03 16:50:03 -07:00
oobabooga ba62783b72 UI: Don't use $ $ for LaTeX, only $$ $$ 2025-09-02 14:22:22 -07:00
oobabooga d3a7710c62
Merge pull request #7215 from oobabooga/dev
Merge dev branch
2025-09-02 16:51:50 -03:00
oobabooga f3829b268a llama.cpp: Always pass --flash-attn on 2025-09-02 12:12:17 -07:00
oobabooga 2395c647d4 Fix the instruct message height on mobile 2025-09-02 12:11:15 -07:00
oobabooga c6ea67bbdb Lint 2025-09-02 10:22:03 -07:00
oobabooga 00ed878b05 Slightly more robust model loading 2025-09-02 10:16:26 -07:00
oobabooga d843afcf66 Update llama.cpp 2025-09-02 05:43:33 -07:00