Sense_wang
7bf15ad933
fix: replace bare except clauses with except Exception ( #7400 )
2026-03-04 18:06:17 -03:00
mamei16
1d1f4dfc88
Disable uncommonly used indented codeblocks ( #7401 )
2026-03-04 17:51:00 -03:00
mamei16
68109bc5da
Improve process_markdown_content ( #7403 )
2026-03-04 17:26:13 -03:00
oobabooga
cdf0e392e6
llama.cpp: Reorganize speculative decoding UI and use recommended ngram-mod defaults
2026-03-04 12:05:08 -08:00
oobabooga
eb90daf098
ExLlamaV2: Don't expose unused seed parameter
2026-03-04 11:14:50 -08:00
oobabooga
d8af0505a8
ExLlamav3_HF: Optimize prefill and fix CFG cache initialization
2026-03-04 11:09:58 -08:00
oobabooga
9b916f02cd
ExLlamaV3: Attach AdaptiveP, fix speculative decoding parameter, add seed
2026-03-04 10:51:15 -08:00
oobabooga
5d93f4e800
Fix requires_grad warning in logits API
2026-03-04 10:43:23 -08:00
oobabooga
64eb77e782
Fix the logits API endpoint with transformers
2026-03-04 10:41:47 -08:00
oobabooga
65de4c30c8
Add adaptive-p sampler and n-gram speculative decoding support
2026-03-04 09:41:29 -08:00
oobabooga
f010aa1612
Replace PyPDF2 with pymupdf for PDF text extraction
...
pymupdf produces cleaner text (e.g. no concatenated words in headers),
handles encrypted and malformed PDFs that PyPDF2 failed on, and
supports non-Latin scripts.
2026-03-04 06:43:37 -08:00
oobabooga
f4d787ab8d
Delegate GPU layer allocation to llama.cpp's --fit
2026-03-04 06:37:50 -08:00
oobabooga
8a3d866401
Fix temperature_last having no effect in llama.cpp server sampler order
2026-03-04 06:10:51 -08:00
oobabooga
b3fd0d16e0
Use a new gr.Headless component for efficient chat streaming
2026-03-03 18:12:03 -08:00
oobabooga
2260e530c9
Remove gradio monkey-patches (moved to gradio fork)
2026-03-03 17:17:36 -08:00
oobabooga
c54e8a2b3d
Try to spawn llama.cpp on port 5001 instead of random port
2026-01-28 08:23:55 -08:00
oobabooga
dc2bbf1861
Refactor thinking block detection and add Solar Open support
2026-01-28 08:21:34 -08:00
q5sys (JT)
7493fe7841
feat: Add a dropdown to save/load user personas ( #7367 )
2026-01-14 20:35:08 -03:00
Sergey 'Jin' Bostandzhyan
6e2c4e9c23
Fix loading models which have their eos token disabled ( #7363 )
2026-01-06 11:31:10 -03:00
oobabooga
e7c8b51fec
Revert "Use flash_attention_2 by default for Transformers models"
...
This reverts commit 85f2df92e9 .
2025-12-07 18:48:41 -08:00
oobabooga
b758059e95
Revert "Clear the torch cache between sequential image generations"
...
This reverts commit 1ec9f708e5 .
2025-12-07 12:23:19 -08:00
oobabooga
1ec9f708e5
Clear the torch cache between sequential image generations
2025-12-07 11:49:22 -08:00
oobabooga
85f2df92e9
Use flash_attention_2 by default for Transformers models
2025-12-07 06:56:58 -08:00
oobabooga
1762312fb4
Use random instead of np.random for image seeds (makes it work on Windows)
2025-12-06 20:10:32 -08:00
oobabooga
02518a96a9
Lint
2025-12-06 06:55:06 -08:00
oobabooga
455dc06db0
Serve the original PNG images in the UI instead of webp
2025-12-06 05:43:00 -08:00
oobabooga
6ca99910ba
Image: Quantize the text encoder for lower VRAM
2025-12-05 13:08:46 -08:00
oobabooga
11937de517
Use flash attention for image generation by default
2025-12-05 12:13:24 -08:00
oobabooga
c11c14590a
Image: Better LLM variation default prompt
2025-12-05 08:08:11 -08:00
oobabooga
0dd468245c
Image: Add back the gallery cache (for performance)
2025-12-05 07:11:38 -08:00
oobabooga
b63d57158d
Image: Add TGW as a prefix to output images
2025-12-05 05:59:54 -08:00
oobabooga
afa29b9554
Image: Several fixes
2025-12-05 05:58:57 -08:00
oobabooga
8eac99599a
Image: Better LLM variation default prompt
2025-12-04 19:58:06 -08:00
oobabooga
b4f06a50b0
fix: Pass bos_token and eos_token from metadata to jinja2
...
Fixes loading Seed-Instruct-36B
2025-12-04 19:11:31 -08:00
oobabooga
56f2a9512f
Revert "Image: Add the LLM-generated prompt to the API result"
...
This reverts commit c7ad28a4cd .
2025-12-04 17:34:27 -08:00
oobabooga
c7ad28a4cd
Image: Add the LLM-generated prompt to the API result
2025-12-04 17:22:08 -08:00
oobabooga
b451bac082
Image: Improve a log message
2025-12-04 16:33:46 -08:00
oobabooga
47a0fcd614
Image: PNG metadata improvements
2025-12-04 16:25:48 -08:00
oobabooga
ac31a7c008
Image: Organize the UI
2025-12-04 15:45:04 -08:00
oobabooga
a90739f498
Image: Better LLM variation default prompt
2025-12-04 10:50:40 -08:00
oobabooga
ffef3c7b1d
Image: Make the LLM Variations prompt configurable
2025-12-04 10:44:35 -08:00
oobabooga
5763947c37
Image: Simplify the API code, add the llm_variations option
2025-12-04 10:23:00 -08:00
oobabooga
2793153717
Image: Add LLM-generated prompt variations
2025-12-04 08:10:24 -08:00
oobabooga
7fb9f19bd8
Progress bar style improvements
2025-12-04 06:20:45 -08:00
oobabooga
a838223d18
Image: Add a progress bar during generation
2025-12-04 05:49:57 -08:00
oobabooga
14dbc3488e
Image: Clear the torch cache after generation, not before
2025-12-04 05:32:58 -08:00
oobabooga
c357eed4c7
Image: Remove the flash_attention_3 option (no idea how to get it working)
2025-12-03 18:40:34 -08:00
oobabooga
fbca54957e
Image generation: Yield partial results for batch count > 1
2025-12-03 16:13:07 -08:00
oobabooga
49c60882bf
Image generation: Safer image uploading
2025-12-03 16:07:51 -08:00
oobabooga
59285d501d
Image generation: Small UI improvements
2025-12-03 16:03:31 -08:00