oobabooga
|
544fcb0b7f
|
Simplify modules/image_models.py
|
2026-04-04 23:29:57 -07:00 |
|
oobabooga
|
c63a79ee48
|
Image generation: Embed generation metadata in API image responses
|
2026-04-04 23:15:14 -07:00 |
|
oobabooga
|
9805ddcde9
|
Update the custom gradio wheels
|
2026-04-04 21:34:09 -07:00 |
|
oobabooga
|
91f9b01516
|
UI: Minor change
|
2026-04-04 21:13:20 -07:00 |
|
oobabooga
|
1f49a64e1a
|
UI: Improve blockquote border width and color
|
2026-04-04 20:44:37 -07:00 |
|
oobabooga
|
e8b31c063a
|
UI: Soften message action icons in light mode
|
2026-04-04 20:38:31 -07:00 |
|
oobabooga
|
ee917cd5ed
|
UI: Make table and hr borders more subtle
|
2026-04-04 20:35:27 -07:00 |
|
oobabooga
|
dfd8ec9c49
|
UI: Make accordion outline styling global
|
2026-04-04 20:13:20 -07:00 |
|
oobabooga
|
0c033caf0e
|
UI: Reduce spacing above chat input
|
2026-04-04 20:09:28 -07:00 |
|
oobabooga
|
1b403a4ffa
|
UI: Fix inline LaTeX rendering by protecting $...$ from markdown (closes #7423)
|
2026-04-04 19:33:05 -07:00 |
|
oobabooga
|
8cb7fe9c47
|
UI: Improve message action icon visibility in light mode
|
2026-04-04 19:14:17 -07:00 |
|
oobabooga
|
41bce3f4de
|
UI: Improve scrollbars style
|
2026-04-04 19:07:36 -07:00 |
|
oobabooga
|
ffea8f282e
|
UI: Improve message text contrast
|
2026-04-04 18:53:13 -07:00 |
|
oobabooga
|
7fed60f90a
|
UI: Improve the hover menu looks
|
2026-04-04 18:29:36 -07:00 |
|
oobabooga
|
2eef90a323
|
API: Remove deprecated "settings" parameter from model load endpoint
|
2026-04-04 11:00:14 -07:00 |
|
oobabooga
|
9183dc444e
|
API: Fix loader args leaking between sequential model loads
|
2026-04-04 10:48:53 -07:00 |
|
oobabooga
|
e0ad4e60df
|
UI: Fix tool buffer check truncating visible text at end of generation
|
2026-04-04 09:57:07 -07:00 |
|
oobabooga
|
16af11f868
|
Update README
|
2026-04-04 04:22:37 -07:00 |
|
oobabooga
|
54b2f39c78
|
Cleanup modules/chat.py
|
2026-04-03 22:07:21 -07:00 |
|
oobabooga
|
2fbaee58cd
|
Add Windows + ROCm portable builds
|
2026-04-03 20:54:28 -07:00 |
|
oobabooga
|
fc35acab9b
|
API: Fix tool call parser crash on non-dict JSON output
|
2026-04-03 16:56:15 -07:00 |
|
oobabooga
|
8ecdb41078
|
fix(security): sanitize filenames in all prompt file operations (CWE-22) (#7462)
---------
Co-authored-by: Alex Chen <ffulbtech@gmail.com>
|
2026-04-03 19:36:50 -03:00 |
|
oobabooga
|
5fb8c4fbd6
|
Update the custom gradio wheels
|
2026-04-03 11:02:00 -07:00 |
|
oobabooga
|
6b66da84d2
|
Update the custom gradio wheels
|
2026-04-03 10:01:51 -07:00 |
|
oobabooga
|
8e8e1ba898
|
Update the custom gradio wheels
|
2026-04-03 09:50:15 -07:00 |
|
oobabooga
|
131a9a0140
|
Update llama.cpp
|
2026-04-03 09:15:03 -07:00 |
|
oobabooga
|
95d6c53e13
|
Revert "API: Add warning about vanilla llama-server not supporting prompt logprobs + instructions"
This reverts commit 42dfcdfc5b.
|
2026-04-03 07:30:48 -07:00 |
|
oobabooga
|
8bba9ecc3f
|
Update the custom gradio wheels
|
2026-04-03 05:58:05 -07:00 |
|
oobabooga
|
66d1a22c73
|
Fix crash when no model is selected (None passed to resolve_model_path)
|
2026-04-03 05:56:36 -07:00 |
|
oobabooga
|
000d776967
|
Revert "llama.cpp: Disable jinja by default (we use Python jinja, not cpp jinja)"
This reverts commit a1cb5b5dc0.
|
2026-04-03 05:49:03 -07:00 |
|
oobabooga
|
a1cb5b5dc0
|
llama.cpp: Disable jinja by default (we use Python jinja, not cpp jinja)
This was causing template compilation issues with qwen models.
|
2026-04-02 21:56:40 -07:00 |
|
oobabooga
|
42dfcdfc5b
|
API: Add warning about vanilla llama-server not supporting prompt logprobs + instructions
|
2026-04-02 20:46:27 -07:00 |
|
oobabooga
|
6e2b70bde6
|
Add Gemma 4 tool-calling support
|
2026-04-02 20:26:27 -07:00 |
|
oobabooga
|
b108c55353
|
Fix portable builds not starting due to missing ik element
|
2026-04-02 19:14:50 -07:00 |
|
oobabooga
|
7aab2fdf9a
|
API: Improve cache clearing in logprobs
|
2026-04-02 17:50:42 -07:00 |
|
oobabooga
|
d84157403a
|
Update the custom gradio wheels
|
2026-04-02 17:31:44 -07:00 |
|
oobabooga
|
a61bde509f
|
Update llama.cpp
|
2026-04-02 17:30:02 -07:00 |
|
oobabooga
|
091037ec20
|
Fix top_logprobs_ids missing for llama.cpp loader
|
2026-04-02 16:13:45 -03:00 |
|
oobabooga
|
f6f8f14c8d
|
Security: Fix SSRF in superbooga extensions
|
2026-04-02 16:13:39 -03:00 |
|
oobabooga
|
80e81a54ca
|
Remove ik macOS wheels from full requirements
|
2026-04-02 11:11:44 -07:00 |
|
oobabooga
|
468cb5cb87
|
Update accelerate
|
2026-04-02 10:59:28 -07:00 |
|
oobabooga
|
6a1f720c7b
|
Update transformers
|
2026-04-02 10:58:20 -07:00 |
|
oobabooga
|
8f8b57a029
|
Update exllamav3
|
2026-04-02 10:54:20 -07:00 |
|
oobabooga
|
c50e17bdbe
|
Add dedicated ik portable requirements files and remove macOS ik builds
|
2026-04-02 14:50:52 -03:00 |
|
oobabooga
|
ea1f8c71f2
|
API: Optimize prompt logprobs and refactor ExLlamav3 forward pass
|
2026-04-02 14:31:11 -03:00 |
|
oobabooga
|
c10c6e87ae
|
API: Add token ids to logprobs output
|
2026-04-02 07:17:27 -07:00 |
|
oobabooga
|
a32ce254f2
|
Don't pass torch_dtype to transformers, autodetect from model config
|
2026-04-02 00:44:14 -03:00 |
|
oobabooga
|
4073164be0
|
Fix ExLlamav3 OOM on prompt logprobs and qwen3_5_moe HF compat
|
2026-04-01 19:44:55 -07:00 |
|
oobabooga
|
328534b762
|
Update llama.cpp
|
2026-04-01 12:51:07 -07:00 |
|
oobabooga
|
71c1a52afe
|
API: Implement echo + logprobs for /v1/completions endpoint
|
2026-03-31 07:43:11 -07:00 |
|