Commit graph

2209 commits

Author SHA1 Message Date
Th-Underscore
b6460908de
Add model download branch handling in download_model_wrapper
Works with `*/tree/<branch>` URL or `*:<branch>` ID
2026-04-17 03:50:36 -04:00
oobabooga
78a178de9e UI: Past chats menu improvement 2026-04-16 03:32:44 -07:00
oobabooga
57be34d7a6 Fix bos/eos tokens being overwritten after GGUF metadata sets them (closes #7496) 2026-04-15 13:14:27 -07:00
oobabooga
a56e281204 Clean up unused function 2026-04-15 11:34:29 -07:00
oobabooga
dd77820a3a UI: Fix consecutive thinking blocks rendering with Gemma 4 2026-04-15 11:24:25 -07:00
oobabooga
e4168ee8b8 Fix tool responses with Gemma 4 template (closes #7498) 2026-04-15 08:56:57 -07:00
oobabooga
5992fa9044 Fix "Start reply with" crash (closes #7497) 2026-04-15 08:12:35 -07:00
oobabooga
f66a01bd06 UI: Add an error message when no model is loaded 2026-04-14 21:05:56 -07:00
oobabooga
734d3c38a0 Reset bos/eos tokens to defaults before reading model metadata 2026-04-14 20:30:29 -07:00
oobabooga
b06b36a6a6 Lint 2026-04-14 20:28:55 -07:00
oobabooga
9e323425d8 Fix row_split not working with ik_llama.cpp (closes #7489) 2026-04-14 23:18:36 -03:00
oobabooga
a8a0f8dc82 Fix version metadata not syncing on Continue (closes #7492) 2026-04-14 23:13:58 -03:00
wiger3
75bf2feb59
Logits display improvements (#7486) 2026-04-14 22:29:19 -03:00
mamei16
fbd95bd5e6
Handle Double Quotes and Newline Chars in Gemma-4 Tool Call Arguments (#7477) 2026-04-14 22:25:10 -03:00
oobabooga
65cef2c731 Rename project from text-generation-webui to textgen 2026-04-13 17:03:19 -03:00
oobabooga
add91613db Fix bos/eos tokens not being set for models without a chat template 2026-04-12 16:01:00 -07:00
oobabooga
326867e799 Reduce VRAM peak in prompt logprobs forward pass 2026-04-10 19:21:57 -07:00
oobabooga
e18f32cba7 Remove hardcoded trust_remote_code=True in embedding loader 2026-04-06 17:47:50 -07:00
oobabooga
775c913de2 Fix crash when truncating prompts with tool call messages 2026-04-06 14:13:01 -07:00
oobabooga
cb511928e2 Fix GPT-OSS tag leak during streaming between thinking and tool calls 2026-04-06 12:06:28 -07:00
oobabooga
193424cc93 API: Fix IPv6 address formatting 2026-04-06 10:07:52 -07:00
oobabooga
c26ffdd24c API: add instruction_template support to the model load endpoint 2026-04-06 07:02:53 -07:00
oobabooga
4d6230a944 Follow-up to d78fc46114 2026-04-06 06:48:48 -07:00
oobabooga
05e4842033 Fix image generation: default to SDPA attention backend 2026-04-05 20:03:06 -07:00
oobabooga
b1d06dcf96 UI: Add MCP server support 2026-04-05 19:46:01 -07:00
oobabooga
abc3487f4d UI: Move cpu-moe checkbox to extra flags (no longer useful now that --fit exists) 2026-04-05 18:24:26 -07:00
oobabooga
d78fc46114 Fix "address already in use" on server restart (Linux/macOS) 2026-04-05 16:42:27 -07:00
oobabooga
422f42ca7f Pre-compile LaTeX regex in html_generator.py 2026-04-04 23:51:15 -07:00
oobabooga
544fcb0b7f Simplify modules/image_models.py 2026-04-04 23:29:57 -07:00
oobabooga
c63a79ee48 Image generation: Embed generation metadata in API image responses 2026-04-04 23:15:14 -07:00
oobabooga
dfd8ec9c49 UI: Make accordion outline styling global 2026-04-04 20:13:20 -07:00
oobabooga
1b403a4ffa UI: Fix inline LaTeX rendering by protecting $...$ from markdown (closes #7423) 2026-04-04 19:33:05 -07:00
oobabooga
ffea8f282e UI: Improve message text contrast 2026-04-04 18:53:13 -07:00
oobabooga
7fed60f90a UI: Improve the hover menu looks 2026-04-04 18:29:36 -07:00
oobabooga
2eef90a323 API: Remove deprecated "settings" parameter from model load endpoint 2026-04-04 11:00:14 -07:00
oobabooga
9183dc444e API: Fix loader args leaking between sequential model loads 2026-04-04 10:48:53 -07:00
oobabooga
e0ad4e60df UI: Fix tool buffer check truncating visible text at end of generation 2026-04-04 09:57:07 -07:00
oobabooga
54b2f39c78 Cleanup modules/chat.py 2026-04-03 22:07:21 -07:00
oobabooga
fc35acab9b API: Fix tool call parser crash on non-dict JSON output 2026-04-03 16:56:15 -07:00
oobabooga
8ecdb41078
fix(security): sanitize filenames in all prompt file operations (CWE-22) (#7462)
---------

Co-authored-by: Alex Chen <ffulbtech@gmail.com>
2026-04-03 19:36:50 -03:00
oobabooga
95d6c53e13 Revert "API: Add warning about vanilla llama-server not supporting prompt logprobs + instructions"
This reverts commit 42dfcdfc5b.
2026-04-03 07:30:48 -07:00
oobabooga
66d1a22c73 Fix crash when no model is selected (None passed to resolve_model_path) 2026-04-03 05:56:36 -07:00
oobabooga
000d776967 Revert "llama.cpp: Disable jinja by default (we use Python jinja, not cpp jinja)"
This reverts commit a1cb5b5dc0.
2026-04-03 05:49:03 -07:00
oobabooga
a1cb5b5dc0 llama.cpp: Disable jinja by default (we use Python jinja, not cpp jinja)
This was causing template compilation issues with qwen models.
2026-04-02 21:56:40 -07:00
oobabooga
42dfcdfc5b API: Add warning about vanilla llama-server not supporting prompt logprobs + instructions 2026-04-02 20:46:27 -07:00
oobabooga
6e2b70bde6 Add Gemma 4 tool-calling support 2026-04-02 20:26:27 -07:00
oobabooga
b108c55353 Fix portable builds not starting due to missing ik element 2026-04-02 19:14:50 -07:00
oobabooga
7aab2fdf9a API: Improve cache clearing in logprobs 2026-04-02 17:50:42 -07:00
oobabooga
091037ec20 Fix top_logprobs_ids missing for llama.cpp loader 2026-04-02 16:13:45 -03:00
oobabooga
ea1f8c71f2 API: Optimize prompt logprobs and refactor ExLlamav3 forward pass 2026-04-02 14:31:11 -03:00