oobabooga
|
16f77b74c4
|
Merge remote-tracking branch 'refs/remotes/origin/dev' into dev
|
2025-11-01 19:58:53 -07:00 |
|
oobabooga
|
cd645f80f8
|
Update exllamav3 to 0.0.12
|
2025-11-01 19:58:18 -07:00 |
|
Trenten Miller
|
6871484398
|
fix: Rename 'evaluation_strategy' to 'eval_strategy' in training
|
2025-10-28 16:48:04 -03:00 |
|
oobabooga
|
338ae36f73
|
Add weights_only=True to torch.load in Training_PRO
|
2025-10-28 12:43:16 -07:00 |
|
dependabot[bot]
|
c8cd840b24
|
Bump flash-linear-attention from 0.3.2 to 0.4.0 in /requirements/full (#7285)
Bumps [flash-linear-attention](https://github.com/fla-org/flash-linear-attention) from 0.3.2 to 0.4.0.
- [Release notes](https://github.com/fla-org/flash-linear-attention/releases)
- [Commits](https://github.com/fla-org/flash-linear-attention/compare/v0.3.2...v0.4.0)
---
updated-dependencies:
- dependency-name: flash-linear-attention
dependency-version: 0.4.0
dependency-type: direct:production
update-type: version-update:semver-minor
...
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
|
2025-10-28 10:07:03 -03:00 |
|
oobabooga
|
f4c9e67155
|
Update llama.cpp
|
2025-10-23 08:19:32 -07:00 |
|
Immanuel
|
9a84a828fc
|
Fixed python requirements for apple devices with macos tahoe (#7273)
|
2025-10-22 14:59:27 -03:00 |
|
reksarka
|
138cc654c4
|
Make it possible to run a portable Web UI build via a symlink (#7277)
|
2025-10-22 14:55:17 -03:00 |
|
oobabooga
|
24fd2b4dec
|
Update exllamav3 to 0.0.11
|
2025-10-21 07:26:38 -07:00 |
|
oobabooga
|
be81f050a7
|
Merge remote-tracking branch 'refs/remotes/origin/dev' into dev
|
2025-10-20 19:43:36 -07:00 |
|
oobabooga
|
9476123ee6
|
Update llama.cpp
|
2025-10-20 19:43:26 -07:00 |
|
dependabot[bot]
|
0d85744205
|
Bump triton-windows in /requirements/full (#7274)
|
2025-10-20 20:36:55 -03:00 |
|
oobabooga
|
a156ebbf76
|
Lint
|
2025-10-15 13:15:01 -07:00 |
|
oobabooga
|
c871d9cdbd
|
Revert "Same as 7f06aec3a1 but for exllamav3_hf"
This reverts commit deb37b821b.
|
2025-10-15 13:05:41 -07:00 |
|
oobabooga
|
163d863443
|
Update llama.cpp
|
2025-10-15 11:23:10 -07:00 |
|
oobabooga
|
c93d567f97
|
Update exllamav3 to 0.0.10
|
2025-10-15 06:41:09 -07:00 |
|
oobabooga
|
b5a6904c4a
|
Make --trust-remote-code immutable from the UI/API
|
2025-10-14 20:47:01 -07:00 |
|
oobabooga
|
efaf2aef3d
|
Update exllamav3 to 0.0.9
|
2025-10-13 15:32:25 -07:00 |
|
oobabooga
|
047855c591
|
Update llama.cpp
|
2025-10-13 15:32:03 -07:00 |
|
mamei16
|
308e726e11
|
log error when llama-server request exceeds context size (#7263)
|
2025-10-12 23:00:11 -03:00 |
|
oobabooga
|
611399e089
|
Update README
|
2025-10-11 17:22:48 -07:00 |
|
oobabooga
|
968c79db06
|
Minor README fix (closes #7251)
|
2025-10-11 17:20:49 -07:00 |
|
oobabooga
|
655c3e86e3
|
Fix "continue" missing an initial space in chat-instruct/chat modes
|
2025-10-11 17:00:25 -07:00 |
|
oobabooga
|
c7dd920dc8
|
Fix metadata leaking into branched chats
|
2025-10-11 14:12:05 -07:00 |
|
oobabooga
|
1831b3fb51
|
Use my custom gradio_client build (small changes to work with pydantic 2.11)
|
2025-10-10 18:01:21 -07:00 |
|
oobabooga
|
dd0b003493
|
Bump pydantic to 2.11.0
|
2025-10-10 17:52:16 -07:00 |
|
oobabooga
|
a74596374d
|
Reapply "Update exllamav3 to 0.0.8"
This reverts commit 748007f6ee.
|
2025-10-10 17:51:31 -07:00 |
|
oobabooga
|
78ff21d512
|
Organize the --help message
|
2025-10-10 15:21:08 -07:00 |
|
oobabooga
|
5d734cc7ca
|
Remove unused CSS
|
2025-10-10 12:54:54 -07:00 |
|
oobabooga
|
25360387ec
|
Downloader: Fix resuming downloads after HF moved to Xet
|
2025-10-10 08:27:40 -07:00 |
|
oobabooga
|
bf5d85c922
|
Revert "Downloader: Gracefully handle '416 Range Not Satisfiable' when continuing downloads"
This reverts commit 1aa2b924d2.
|
2025-10-09 17:22:41 -07:00 |
|
oobabooga
|
0d03813e98
|
Update modules/chat.py
Co-authored-by: Copilot <175728472+Copilot@users.noreply.github.com>
|
2025-10-09 21:01:13 -03:00 |
|
oobabooga
|
748007f6ee
|
Revert "Update exllamav3 to 0.0.8"
This reverts commit 977ffbaa04.
|
2025-10-09 16:50:00 -07:00 |
|
dependabot[bot]
|
af3c70651c
|
Update bitsandbytes requirement in /requirements/full (#7255)
|
2025-10-09 19:53:34 -03:00 |
|
oobabooga
|
977ffbaa04
|
Update exllamav3 to 0.0.8
|
2025-10-09 15:53:14 -07:00 |
|
oobabooga
|
e0f0fae59d
|
Exllamav3: Add fla to requirements for qwen3-next
|
2025-10-09 13:03:48 -07:00 |
|
oobabooga
|
deb37b821b
|
Same as 7f06aec3a1 but for exllamav3_hf
|
2025-10-09 13:02:38 -07:00 |
|
oobabooga
|
7f06aec3a1
|
exllamav3: Implement the logits function for /v1/internal/logits
|
2025-10-09 11:24:25 -07:00 |
|
oobabooga
|
218dc01b51
|
Add fallbacks after 93aa7b3ed3
|
2025-10-09 10:59:34 -07:00 |
|
oobabooga
|
1aa2b924d2
|
Downloader: Gracefully handle '416 Range Not Satisfiable' when continuing downloads
|
2025-10-09 10:52:31 -07:00 |
|
oobabooga
|
0f3793d608
|
Update llama.cpp
|
2025-10-09 09:38:22 -07:00 |
|
oobabooga
|
282aa19189
|
Safer profile picture uploading
|
2025-10-09 09:26:35 -07:00 |
|
oobabooga
|
93aa7b3ed3
|
Better handle multigpu setups with transformers + bitsandbytes
|
2025-10-09 08:49:44 -07:00 |
|
Ionoclast Laboratories
|
d229dfe991
|
Fix portable apple intel requirement for llama binaries (issue #7238) (#7239)
|
2025-10-08 12:40:53 -03:00 |
|
oobabooga
|
292c91abbb
|
Update llama.cpp
|
2025-10-08 08:31:34 -07:00 |
|
oobabooga
|
f660e0836b
|
Merge branch 'main' into dev
|
2025-10-08 05:38:33 -07:00 |
|
oobabooga
|
898a3ed2fe
|
Add sponsor (Warp) to README <3
|
2025-10-07 18:33:28 -03:00 |
|
oobabooga
|
22997c134e
|
Merge remote-tracking branch 'refs/remotes/origin/dev' into dev
|
2025-10-05 20:34:49 -07:00 |
|
Remowylliams
|
38a7fd685d
|
chat.py fixes Instruct mode History
|
2025-10-05 11:34:47 -03:00 |
|
oobabooga
|
64829071e0
|
Update llama.cpp
|
2025-10-05 07:32:41 -07:00 |
|