Commit graph

5179 commits

Author SHA1 Message Date
oobabooga 2260e530c9 Remove gradio monkey-patches (moved to gradio fork) 2026-03-03 17:17:36 -08:00
oobabooga e9f22813e4 Replace gradio with my gradio 4.37.2 fork 2026-03-03 16:51:27 -08:00
dependabot[bot] 3519890c8e
Bump flask-cloudflared from 0.0.14 to 0.0.15 in /requirements/full (#7380) 2026-03-03 21:41:51 -03:00
dependabot[bot] 9c604628a0
Bump flask-cloudflared from 0.0.14 to 0.0.15 in /requirements/portable (#7382) 2026-03-03 21:41:46 -03:00
oobabooga fbd2acfa19 Remove triton-windows from non-CUDA requirements 2026-03-03 16:16:55 -08:00
oobabooga 5fd79b23d1 Add CUDA 13.1 portable builds 2026-03-03 15:36:41 -08:00
oobabooga b8fcc8ea32 Update llama.cpp, remove noavx2 builds, add ROCm Windows portable builds 2026-03-03 15:27:19 -08:00
Pádraic Slattery d7dd533b99
chore: Update outdated GitHub Actions versions (#7384) 2026-03-03 17:54:12 -03:00
oobabooga 9576c5a5f4 Update Miniforge from 25.3.0 to 26.1.0 2026-03-03 12:33:20 -08:00
oobabooga 9814d3d0ae Patch gradio 4.x for huggingface-hub 1.x compatibility 2026-03-03 12:20:37 -08:00
oobabooga 38d0eeefc0 Update dependencies: torch 2.9.1, transformers 5.2, exllamav3 0.0.22, accelerate 1.12, huggingface-hub 1.5 2026-03-03 12:01:02 -08:00
oobabooga ddd74324fe Update PyTorch to 2.9.1 and ROCm to 6.4 2026-03-03 11:38:52 -08:00
oobabooga efc72d5c32 Update Python from 3.11 to 3.13 2026-03-03 11:03:26 -08:00
oobabooga aecbc5a8ac Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2026-01-28 08:30:28 -08:00
oobabooga c54e8a2b3d Try to spawn llama.cpp on port 5001 instead of random port 2026-01-28 08:23:55 -08:00
oobabooga dc2bbf1861 Refactor thinking block detection and add Solar Open support 2026-01-28 08:21:34 -08:00
dependabot[bot] cae1fef42d
Bump triton-windows in /requirements/full (#7368) 2026-01-14 21:30:59 -03:00
q5sys (JT) 7493fe7841
feat: Add a dropdown to save/load user personas (#7367) 2026-01-14 20:35:08 -03:00
jakubartur 21b979c02a
Fix code block copy button on HTTP (Clipboard API fallback) (#7358) 2026-01-14 19:34:21 -03:00
oobabooga a731861127 Update README 2026-01-13 15:38:32 -08:00
oobabooga d79cdc614c Update llama.cpp 2026-01-08 11:24:15 -08:00
oobabooga 332fd40653 Update llama.cpp 2026-01-07 19:06:23 -08:00
dependabot[bot] 50a35b483c
Update bitsandbytes requirement in /requirements/full (#7353) 2026-01-06 15:27:23 -03:00
dependabot[bot] 45fbec0320
Update torchao requirement in /requirements/full (#7356) 2026-01-06 15:27:10 -03:00
oobabooga b0968ed8b4 Update flash-linear-attention 2026-01-06 10:26:43 -08:00
oobabooga 36747cf99c Lint 2026-01-06 10:24:34 -08:00
oobabooga 2fcbadec67 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2026-01-06 10:24:07 -08:00
oobabooga bb3b7bc197 Update llama.cpp 2026-01-06 10:23:58 -08:00
Sergey 'Jin' Bostandzhyan 6e2c4e9c23
Fix loading models which have their eos token disabled (#7363) 2026-01-06 11:31:10 -03:00
oobabooga a2ed640aa6
UI: Improved border color for tables + hr 2025-12-21 15:38:48 -03:00
oobabooga 1066fe8c21
UI: Improve table styles (more minimalistic) 2025-12-21 15:32:02 -03:00
oobabooga 9530d3a6d8
UI: Improve hr (horizontal separator) style 2025-12-21 15:30:54 -03:00
oobabooga 09d88f91e8 Update llama.cpp 2025-12-19 21:00:13 -08:00
oobabooga 6e8fb0e7b1 Update llama.cpp 2025-12-14 13:32:14 -08:00
oobabooga 9fe40ff90f Update exllamav3 to 0.0.18 2025-12-10 05:37:33 -08:00
oobabooga 8e762e04b4 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2025-12-09 05:27:43 -08:00
oobabooga aa16266c38 Update llama.cpp 2025-12-09 03:19:23 -08:00
dependabot[bot] 85269d7fbb
Update safetensors requirement in /requirements/full (#7323) 2025-12-08 17:58:27 -03:00
dependabot[bot] c4ebab9b29
Bump triton-windows in /requirements/full (#7346) 2025-12-08 17:56:07 -03:00
oobabooga 502f59d39b Update diffusers to 0.36 2025-12-08 05:08:54 -08:00
oobabooga e7c8b51fec Revert "Use flash_attention_2 by default for Transformers models"
This reverts commit 85f2df92e9.
2025-12-07 18:48:41 -08:00
oobabooga b758059e95 Revert "Clear the torch cache between sequential image generations"
This reverts commit 1ec9f708e5.
2025-12-07 12:23:19 -08:00
oobabooga 1ec9f708e5 Clear the torch cache between sequential image generations 2025-12-07 11:49:22 -08:00
oobabooga 3b8369a679 Update llama.cpp 2025-12-07 11:18:36 -08:00
oobabooga 058e78411d docs: Small changes 2025-12-07 10:16:08 -08:00
oobabooga 17bd8d10f0 Update exllamav3 to 0.0.17 2025-12-07 09:37:18 -08:00
oobabooga 85f2df92e9 Use flash_attention_2 by default for Transformers models 2025-12-07 06:56:58 -08:00
oobabooga 1762312fb4 Use random instead of np.random for image seeds (makes it work on Windows) 2025-12-06 20:10:32 -08:00
oobabooga 160a25165a docs: Small change 2025-12-06 08:41:12 -08:00
oobabooga f93cc4b5c3 Add an API example to the image generation tutorial 2025-12-06 08:33:06 -08:00