Compare commits

...

4789 commits

Author SHA1 Message Date
oobabooga 910456ba31
Merge pull request #7366 from oobabooga/dev
Merge dev branch
2026-01-08 17:54:12 -03:00
oobabooga d79cdc614c Update llama.cpp 2026-01-08 11:24:15 -08:00
oobabooga 332fd40653 Update llama.cpp 2026-01-07 19:06:23 -08:00
dependabot[bot] 50a35b483c
Update bitsandbytes requirement in /requirements/full (#7353) 2026-01-06 15:27:23 -03:00
dependabot[bot] 45fbec0320
Update torchao requirement in /requirements/full (#7356) 2026-01-06 15:27:10 -03:00
oobabooga b0968ed8b4 Update flash-linear-attention 2026-01-06 10:26:43 -08:00
oobabooga 36747cf99c Lint 2026-01-06 10:24:34 -08:00
oobabooga 2fcbadec67 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2026-01-06 10:24:07 -08:00
oobabooga bb3b7bc197 Update llama.cpp 2026-01-06 10:23:58 -08:00
Sergey 'Jin' Bostandzhyan 6e2c4e9c23
Fix loading models which have their eos token disabled (#7363) 2026-01-06 11:31:10 -03:00
oobabooga a2ed640aa6
UI: Improved border color for tables + hr 2025-12-21 15:38:48 -03:00
oobabooga 1066fe8c21
UI: Improve table styles (more minimalistic) 2025-12-21 15:32:02 -03:00
oobabooga 9530d3a6d8
UI: Improve hr (horizontal separator) style 2025-12-21 15:30:54 -03:00
oobabooga a0b5599e9b
Merge pull request #7355 from oobabooga/dev
Merge dev branch
2025-12-20 02:18:31 -03:00
oobabooga 09d88f91e8 Update llama.cpp 2025-12-19 21:00:13 -08:00
oobabooga 34804f9354
Merge pull request #7352 from oobabooga/dev
Merge dev branch
2025-12-14 22:59:34 -03:00
oobabooga 6e8fb0e7b1 Update llama.cpp 2025-12-14 13:32:14 -08:00
oobabooga 9fe40ff90f Update exllamav3 to 0.0.18 2025-12-10 05:37:33 -08:00
oobabooga 8e762e04b4 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2025-12-09 05:27:43 -08:00
oobabooga aa16266c38 Update llama.cpp 2025-12-09 03:19:23 -08:00
dependabot[bot] 85269d7fbb
Update safetensors requirement in /requirements/full (#7323) 2025-12-08 17:58:27 -03:00
dependabot[bot] c4ebab9b29
Bump triton-windows in /requirements/full (#7346) 2025-12-08 17:56:07 -03:00
oobabooga bb004bacb1
Merge pull request #7345 from oobabooga/dev
Merge dev branch
2025-12-08 10:14:49 -03:00
oobabooga 502f59d39b Update diffusers to 0.36 2025-12-08 05:08:54 -08:00
oobabooga 4d94f66832
Merge pull request #7343 from oobabooga/dev
Merge dev branch
2025-12-07 23:49:19 -03:00
oobabooga e7c8b51fec Revert "Use flash_attention_2 by default for Transformers models"
This reverts commit 85f2df92e9.
2025-12-07 18:48:41 -08:00
oobabooga 652d13c003
Merge pull request #7339 from oobabooga/dev
Merge dev branch
2025-12-07 17:58:00 -03:00
oobabooga b758059e95 Revert "Clear the torch cache between sequential image generations"
This reverts commit 1ec9f708e5.
2025-12-07 12:23:19 -08:00
oobabooga 1ec9f708e5 Clear the torch cache between sequential image generations 2025-12-07 11:49:22 -08:00
oobabooga 3b8369a679 Update llama.cpp 2025-12-07 11:18:36 -08:00
oobabooga 058e78411d docs: Small changes 2025-12-07 10:16:08 -08:00
oobabooga 17bd8d10f0 Update exllamav3 to 0.0.17 2025-12-07 09:37:18 -08:00
oobabooga 85f2df92e9 Use flash_attention_2 by default for Transformers models 2025-12-07 06:56:58 -08:00
oobabooga 1762312fb4 Use random instead of np.random for image seeds (makes it work on Windows) 2025-12-06 20:10:32 -08:00
oobabooga 160a25165a docs: Small change 2025-12-06 08:41:12 -08:00
oobabooga f93cc4b5c3 Add an API example to the image generation tutorial 2025-12-06 08:33:06 -08:00
oobabooga c026dbaf64 Fix API requests always returning the same 'created' time 2025-12-06 08:23:21 -08:00
oobabooga 194e4c285f Update llama.cpp 2025-12-06 08:14:48 -08:00
oobabooga 1c36559e2b Add a News section to the README 2025-12-06 07:05:00 -08:00
oobabooga 02518a96a9 Lint 2025-12-06 06:55:06 -08:00
oobabooga 0100ad1bd7 Add user_data/image_outputs to the Gradio allowed paths 2025-12-06 06:39:30 -08:00
oobabooga 6411142111 docs: Small changes 2025-12-06 06:36:16 -08:00
oobabooga 455dc06db0 Serve the original PNG images in the UI instead of webp 2025-12-06 05:43:00 -08:00
oobabooga 1a9ed1fe98 Fix the height of the image output gallery 2025-12-06 05:21:26 -08:00
oobabooga 17b12567d8 docs: Small changes 2025-12-05 14:15:15 -08:00
oobabooga e20b2d38ff docs: Add VRAM measurements for Z-Image-Turbo 2025-12-05 14:12:08 -08:00
oobabooga 6ca99910ba Image: Quantize the text encoder for lower VRAM 2025-12-05 13:08:46 -08:00
oobabooga 11937de517 Use flash attention for image generation by default 2025-12-05 12:13:24 -08:00
oobabooga eba8a59466 docs: Improve the image generation tutorial 2025-12-05 12:10:41 -08:00
oobabooga 5848c7884d Increase the height of the image output gallery 2025-12-05 10:24:51 -08:00
oobabooga c11c14590a Image: Better LLM variation default prompt 2025-12-05 08:08:11 -08:00
oobabooga 0dd468245c Image: Add back the gallery cache (for performance) 2025-12-05 07:11:38 -08:00
oobabooga b63d57158d Image: Add TGW as a prefix to output images 2025-12-05 05:59:54 -08:00
oobabooga afa29b9554 Image: Several fixes 2025-12-05 05:58:57 -08:00
oobabooga 8eac99599a Image: Better LLM variation default prompt 2025-12-04 19:58:06 -08:00
oobabooga b4f06a50b0 fix: Pass bos_token and eos_token from metadata to jinja2
Fixes loading Seed-Instruct-36B
2025-12-04 19:11:31 -08:00
oobabooga 15c6e43597 Image: Add a revised_prompt field to API results for OpenAI compatibility 2025-12-04 17:41:09 -08:00
oobabooga 56f2a9512f Revert "Image: Add the LLM-generated prompt to the API result"
This reverts commit c7ad28a4cd.
2025-12-04 17:34:27 -08:00
oobabooga 3ef428efaa Image: Remove llm_variations from the API 2025-12-04 17:34:17 -08:00
oobabooga c7ad28a4cd Image: Add the LLM-generated prompt to the API result 2025-12-04 17:22:08 -08:00
oobabooga b451bac082 Image: Improve a log message 2025-12-04 16:33:46 -08:00
oobabooga 47a0fcd614 Image: PNG metadata improvements 2025-12-04 16:25:48 -08:00
oobabooga ac31a7c008 Image: Organize the UI 2025-12-04 15:45:04 -08:00
oobabooga a90739f498 Image: Better LLM variation default prompt 2025-12-04 10:50:40 -08:00
oobabooga ffef3c7b1d Image: Make the LLM Variations prompt configurable 2025-12-04 10:44:35 -08:00
oobabooga 5763947c37 Image: Simplify the API code, add the llm_variations option 2025-12-04 10:23:00 -08:00
oobabooga 2793153717 Image: Add LLM-generated prompt variations 2025-12-04 08:10:24 -08:00
oobabooga 7fb9f19bd8 Progress bar style improvements 2025-12-04 06:20:45 -08:00
oobabooga a838223d18 Image: Add a progress bar during generation 2025-12-04 05:49:57 -08:00
oobabooga 14dbc3488e Image: Clear the torch cache after generation, not before 2025-12-04 05:32:58 -08:00
oobabooga 235b94f097 Image: Add placeholder file for user_data/image_models 2025-12-03 18:43:30 -08:00
oobabooga c357eed4c7 Image: Remove the flash_attention_3 option (no idea how to get it working) 2025-12-03 18:40:34 -08:00
oobabooga c93d27add3 Update llama.cpp 2025-12-03 18:29:43 -08:00
oobabooga fbca54957e Image generation: Yield partial results for batch count > 1 2025-12-03 16:13:07 -08:00
oobabooga 49c60882bf Image generation: Safer image uploading 2025-12-03 16:07:51 -08:00
oobabooga 59285d501d Image generation: Small UI improvements 2025-12-03 16:03:31 -08:00
oobabooga 373baa5c9c UI: Minor image gallery improvements 2025-12-03 14:45:02 -08:00
oobabooga 906dc54969 Load --image-model before --model 2025-12-03 12:15:38 -08:00
oobabooga 4468c49439 Add semaphore to image generation API endpoint 2025-12-03 12:02:47 -08:00
oobabooga 5ad174fad2 docs: Add an image generation API example 2025-12-03 11:58:54 -08:00
oobabooga 5433ef3333 Add an API endpoint for generating images 2025-12-03 11:50:56 -08:00
oobabooga 9448bf1caa Image generation: add torchao quantization (supports torch.compile) 2025-12-02 14:22:51 -08:00
oobabooga 97281ff831 UI: Fix an index error in the new image gallery 2025-12-02 11:20:52 -08:00
oobabooga 9d07d3a229 Make portable builds functional again after b3666e140d 2025-12-02 10:06:57 -08:00
oobabooga 6291e72129 Remove quanto for now (requires messy compilation) 2025-12-02 09:57:18 -08:00
oobabooga b3666e140d
Add image generation support (#7328) 2025-12-02 14:55:38 -03:00
oobabooga a83821e941 Revert "UI: Optimize typing in all textareas"
This reverts commit e24ba92ef2.
2025-12-01 10:34:23 -08:00
oobabooga 24fd963c38 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2025-12-01 08:06:08 -08:00
oobabooga e24ba92ef2 UI: Optimize typing in all textareas 2025-12-01 08:05:21 -08:00
oobabooga bd9f2de73a
Merge pull request #7331 from oobabooga/dev
Merge dev branch
2025-11-28 23:00:01 -03:00
aidevtime 661e42d2b7
fix(deps): upgrade coqui-tts to >=0.27.0 for transformers 4.55 compatibility (#7329) 2025-11-28 22:59:36 -03:00
oobabooga 5327bc9397
Update modules/shared.py
Co-authored-by: Copilot <175728472+Copilot@users.noreply.github.com>
2025-11-28 22:48:05 -03:00
oobabooga 78b315344a Update exllamav3 2025-11-28 06:45:05 -08:00
oobabooga 3cad0cd4c1 Update llama.cpp 2025-11-28 03:52:37 -08:00
GodEmperor785 400bb0694b
Add slider for --ubatch-size for llama.cpp loader, change defaults for better MoE performance (#7316) 2025-11-21 16:56:02 -03:00
oobabooga 8f0048663d More modular HTML generator 2025-11-21 07:09:16 -08:00
oobabooga b0baf7518b Remove macos x86-64 portable builds (macos-13 runner deprecated by GitHub) 2025-11-19 06:07:15 -08:00
oobabooga 1afe0827ba
Merge pull request #7317 from oobabooga/dev
Merge dev branch
2025-11-19 11:04:02 -03:00
oobabooga 0d4eff284c Add a --cpu-moe model for llama.cpp 2025-11-19 05:23:43 -08:00
oobabooga d6f39e1fef Add ROCm portable builds 2025-11-18 16:32:20 -08:00
oobabooga 327a234d23 Add ROCm requirements.txt files 2025-11-18 16:24:56 -08:00
oobabooga 4e4abd0841 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2025-11-18 14:07:05 -08:00
oobabooga c45f35ccc2 Remove the macos 13 wheels (deprecated by GitHub) 2025-11-18 14:06:42 -08:00
oobabooga d85b95bb15 Update llama.cpp 2025-11-18 14:06:04 -08:00
dependabot[bot] 4a36b7be5b
Bump triton-windows in /requirements/full (#7311) 2025-11-18 18:51:26 -03:00
dependabot[bot] 3d7e9856a2
Update peft requirement from ==0.17.* to ==0.18.* in /requirements/full (#7310) 2025-11-18 18:51:15 -03:00
oobabooga a26e28bdea Update exllamav3 to 0.0.15 2025-11-18 11:24:16 -08:00
oobabooga 6a3bf1de92 Update exllamav3 to 0.0.14 2025-11-09 19:43:53 -08:00
oobabooga 9ad9afad7d
Merge pull request #7296 from oobabooga/dev
Merge dev branch
2025-11-06 00:38:25 -03:00
oobabooga e7534a90d8 Update llama.cpp 2025-11-05 18:46:01 -08:00
oobabooga 6be1bfcc87 Remove the CUDA 11.7 portable builds 2025-11-05 05:45:10 -08:00
oobabooga 92d9cd36a6 Update llama.cpp 2025-11-05 05:43:34 -08:00
oobabooga 67f9288891 Pin huggingface-hub to 0.36.0 (solves #7284 and #7289) 2025-11-02 14:01:00 -08:00
oobabooga 16f77b74c4 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2025-11-01 19:58:53 -07:00
oobabooga cd645f80f8 Update exllamav3 to 0.0.12 2025-11-01 19:58:18 -07:00
Trenten Miller 6871484398
fix: Rename 'evaluation_strategy' to 'eval_strategy' in training 2025-10-28 16:48:04 -03:00
oobabooga 338ae36f73 Add weights_only=True to torch.load in Training_PRO 2025-10-28 12:43:16 -07:00
dependabot[bot] c8cd840b24
Bump flash-linear-attention from 0.3.2 to 0.4.0 in /requirements/full (#7285)
Bumps [flash-linear-attention](https://github.com/fla-org/flash-linear-attention) from 0.3.2 to 0.4.0.
- [Release notes](https://github.com/fla-org/flash-linear-attention/releases)
- [Commits](https://github.com/fla-org/flash-linear-attention/compare/v0.3.2...v0.4.0)

---
updated-dependencies:
- dependency-name: flash-linear-attention
  dependency-version: 0.4.0
  dependency-type: direct:production
  update-type: version-update:semver-minor
...

Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2025-10-28 10:07:03 -03:00
oobabooga fc67e5e692
Merge pull request #7279 from oobabooga/dev
Merge dev branch
2025-10-23 12:50:31 -03:00
oobabooga f4c9e67155 Update llama.cpp 2025-10-23 08:19:32 -07:00
Immanuel 9a84a828fc
Fixed python requirements for apple devices with macos tahoe (#7273) 2025-10-22 14:59:27 -03:00
reksarka 138cc654c4
Make it possible to run a portable Web UI build via a symlink (#7277) 2025-10-22 14:55:17 -03:00
oobabooga 24fd2b4dec Update exllamav3 to 0.0.11 2025-10-21 07:26:38 -07:00
oobabooga be81f050a7 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2025-10-20 19:43:36 -07:00
oobabooga 9476123ee6 Update llama.cpp 2025-10-20 19:43:26 -07:00
dependabot[bot] 0d85744205
Bump triton-windows in /requirements/full (#7274) 2025-10-20 20:36:55 -03:00
oobabooga 771130532c
Merge pull request #7267 from oobabooga/dev
Merge dev branch
2025-10-15 17:15:28 -03:00
oobabooga a156ebbf76 Lint 2025-10-15 13:15:01 -07:00
oobabooga c871d9cdbd Revert "Same as 7f06aec3a1 but for exllamav3_hf"
This reverts commit deb37b821b.
2025-10-15 13:05:41 -07:00
oobabooga 163d863443 Update llama.cpp 2025-10-15 11:23:10 -07:00
oobabooga c93d567f97 Update exllamav3 to 0.0.10 2025-10-15 06:41:09 -07:00
oobabooga b5a6904c4a Make --trust-remote-code immutable from the UI/API 2025-10-14 20:47:01 -07:00
oobabooga efaf2aef3d Update exllamav3 to 0.0.9 2025-10-13 15:32:25 -07:00
oobabooga 047855c591 Update llama.cpp 2025-10-13 15:32:03 -07:00
mamei16 308e726e11
log error when llama-server request exceeds context size (#7263) 2025-10-12 23:00:11 -03:00
oobabooga 611399e089 Update README 2025-10-11 17:22:48 -07:00
oobabooga 968c79db06 Minor README fix (closes #7251) 2025-10-11 17:20:49 -07:00
oobabooga 655c3e86e3 Fix "continue" missing an initial space in chat-instruct/chat modes 2025-10-11 17:00:25 -07:00
oobabooga c7dd920dc8 Fix metadata leaking into branched chats 2025-10-11 14:12:05 -07:00
oobabooga 1831b3fb51 Use my custom gradio_client build (small changes to work with pydantic 2.11) 2025-10-10 18:01:21 -07:00
oobabooga dd0b003493 Bump pydantic to 2.11.0 2025-10-10 17:52:16 -07:00
oobabooga a74596374d Reapply "Update exllamav3 to 0.0.8"
This reverts commit 748007f6ee.
2025-10-10 17:51:31 -07:00
oobabooga 78ff21d512 Organize the --help message 2025-10-10 15:21:08 -07:00
oobabooga 5d734cc7ca Remove unused CSS 2025-10-10 12:54:54 -07:00
oobabooga 25360387ec Downloader: Fix resuming downloads after HF moved to Xet 2025-10-10 08:27:40 -07:00
oobabooga 7833650aa1
Merge pull request #7260 from oobabooga/dev
Merge dev branch
2025-10-10 10:46:34 -03:00
oobabooga bf5d85c922 Revert "Downloader: Gracefully handle '416 Range Not Satisfiable' when continuing downloads"
This reverts commit 1aa2b924d2.
2025-10-09 17:22:41 -07:00
oobabooga 0d03813e98
Update modules/chat.py
Co-authored-by: Copilot <175728472+Copilot@users.noreply.github.com>
2025-10-09 21:01:13 -03:00
oobabooga 748007f6ee Revert "Update exllamav3 to 0.0.8"
This reverts commit 977ffbaa04.
2025-10-09 16:50:00 -07:00
dependabot[bot] af3c70651c
Update bitsandbytes requirement in /requirements/full (#7255) 2025-10-09 19:53:34 -03:00
oobabooga 977ffbaa04 Update exllamav3 to 0.0.8 2025-10-09 15:53:14 -07:00
oobabooga e0f0fae59d Exllamav3: Add fla to requirements for qwen3-next 2025-10-09 13:03:48 -07:00
oobabooga deb37b821b Same as 7f06aec3a1 but for exllamav3_hf 2025-10-09 13:02:38 -07:00
oobabooga 7f06aec3a1 exllamav3: Implement the logits function for /v1/internal/logits 2025-10-09 11:24:25 -07:00
oobabooga 218dc01b51 Add fallbacks after 93aa7b3ed3 2025-10-09 10:59:34 -07:00
oobabooga 1aa2b924d2 Downloader: Gracefully handle '416 Range Not Satisfiable' when continuing downloads 2025-10-09 10:52:31 -07:00
oobabooga 0f3793d608 Update llama.cpp 2025-10-09 09:38:22 -07:00
oobabooga 282aa19189 Safer profile picture uploading 2025-10-09 09:26:35 -07:00
oobabooga 93aa7b3ed3 Better handle multigpu setups with transformers + bitsandbytes 2025-10-09 08:49:44 -07:00
Ionoclast Laboratories d229dfe991
Fix portable apple intel requirement for llama binaries (issue #7238) (#7239) 2025-10-08 12:40:53 -03:00
oobabooga 292c91abbb Update llama.cpp 2025-10-08 08:31:34 -07:00
oobabooga f660e0836b Merge branch 'main' into dev 2025-10-08 05:38:33 -07:00
oobabooga 898a3ed2fe
Add sponsor (Warp) to README <3 2025-10-07 18:33:28 -03:00
oobabooga 22997c134e Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2025-10-05 20:34:49 -07:00
Remowylliams 38a7fd685d
chat.py fixes Instruct mode History 2025-10-05 11:34:47 -03:00
oobabooga 64829071e0 Update llama.cpp 2025-10-05 07:32:41 -07:00
oobabooga 0eb8543d74 Update transformers 2025-10-05 07:30:33 -07:00
oobabooga b7effb22e0 Update exllamav3 2025-10-05 07:29:57 -07:00
oobabooga 042b828c73
Merge pull request #7231 from oobabooga/dev
Merge dev branch
2025-09-21 01:18:56 -03:00
oobabooga 8c9df34696 Update llama.cpp 2025-09-20 20:57:15 -07:00
oobabooga 1e863a7113 Fix exllamav3 ignoring the stop button 2025-09-19 16:12:50 -07:00
oobabooga 005fcf3f98 Formatting 2025-09-17 21:58:37 -07:00
oobabooga e4412f0634 Slightly more robust syntax highlighting 2025-09-17 21:57:17 -07:00
stevenxdavis dd6d2223a5
Changing transformers_loader.py to Match User Expectations for --bf16 and Flash Attention 2 (#7217) 2025-09-17 16:39:04 -03:00
oobabooga 9e9ab39892 Make exllamav3_hf and exllamav2_hf functional again 2025-09-17 12:29:22 -07:00
oobabooga 9c0a833a0a Revert "Update bitsandbytes requirement in /requirements/full (#7193)"
This reverts commit fe15b67160.
2025-09-17 11:58:54 -07:00
oobabooga 8087a57fd8 Bump transformers to 4.56 2025-09-17 08:19:18 -07:00
dependabot[bot] 7131a478b9
Update safetensors requirement in /requirements/full (#7192) 2025-09-17 12:18:13 -03:00
dependabot[bot] fe15b67160
Update bitsandbytes requirement in /requirements/full (#7193) 2025-09-17 12:17:58 -03:00
dependabot[bot] 8f731a566c
Update peft requirement from ==0.16.* to ==0.17.* in /requirements/full (#7172) 2025-09-17 12:17:16 -03:00
oobabooga 483927a5be Update llama.cpp 2025-09-17 05:09:12 -07:00
oobabooga 557b78d31e Update llama.cpp 2025-09-03 16:50:03 -07:00
oobabooga ba62783b72 UI: Don't use $ $ for LaTeX, only $$ $$ 2025-09-02 14:22:22 -07:00
oobabooga d3a7710c62
Merge pull request #7215 from oobabooga/dev
Merge dev branch
2025-09-02 16:51:50 -03:00
oobabooga f3829b268a llama.cpp: Always pass --flash-attn on 2025-09-02 12:12:17 -07:00
oobabooga 2395c647d4 Fix the instruct message height on mobile 2025-09-02 12:11:15 -07:00
oobabooga c6ea67bbdb Lint 2025-09-02 10:22:03 -07:00
oobabooga 00ed878b05 Slightly more robust model loading 2025-09-02 10:16:26 -07:00
oobabooga d843afcf66 Update llama.cpp 2025-09-02 05:43:33 -07:00
oobabooga 00ebb295d3 Update llama.cpp 2025-08-31 16:27:23 -07:00
oobabooga 387e249dec Change an info message 2025-08-31 16:27:10 -07:00
oobabooga 8028d88541 Lint 2025-08-30 21:29:20 -07:00
oobabooga 13876a1ee8 llama.cpp: Remove the --flash-attn flag (it's always on now) 2025-08-30 20:28:26 -07:00
oobabooga 7b80e9a2ad Update llama.cpp 2025-08-30 20:22:11 -07:00
oobabooga 5631d4e3d6 Minor change after 21d790f87e 2025-08-30 15:34:49 -07:00
oobabooga 5920ad8834 UI: Give streaming instruct messages more vertical space 2025-08-30 15:22:50 -07:00
oobabooga 21d790f87e Optimize LaTeX rendering during streaming for long replies 2025-08-30 14:52:22 -07:00
oobabooga 3a3e247f3c Even better way to handle continue for thinking blocks 2025-08-30 12:36:35 -07:00
oobabooga cf1aad2a68 Fix "continue" for Byte-OSS for partial thinking blocks 2025-08-30 12:16:45 -07:00
oobabooga 96136ea760 Fix LaTeX rendering for equations with asterisks 2025-08-30 10:13:32 -07:00
oobabooga a3eb67e466 Fix the UI failing to launch if the Notebook prompt is too long 2025-08-30 08:42:26 -07:00
oobabooga 08f90f4b64 Lint 2025-08-29 14:09:04 -07:00
oobabooga 07a2e226c1 UI: Minor font color fixes in instruct mode 2025-08-29 14:08:38 -07:00
oobabooga a2b37adb26 UI: Preload the correct fonts for chat mode 2025-08-29 09:25:44 -07:00
oobabooga 084675cf75 UI: Improve thinking blocks in chat-instruct mode 2025-08-29 09:11:10 -07:00
oobabooga d78b7d0fad Lint 2025-08-28 20:22:07 -07:00
oobabooga fc2eb48664 Style fixes after 73442a2b6d 2025-08-28 20:21:55 -07:00
oobabooga 2720955478 Fix a bug after d9eec31886 2025-08-28 19:48:16 -07:00
oobabooga d9eec31886 UI: Suppress "Attempted to select a non-interactive or hidden tab" warnings 2025-08-28 17:46:29 -07:00
oobabooga cb8780a4ce Safer check for is_multimodal when loading models
Avoids unrelated multimodal error when a model fails to load due
to lack of memory.
2025-08-28 11:13:19 -07:00
oobabooga cfc83745ec UI: Improve right sidebar borders in light mode 2025-08-28 08:34:48 -07:00
oobabooga a336a8bbeb UI: Fix italic and quote color in headings 2025-08-28 08:26:40 -07:00
oobabooga ba6041251d UI: Minor change 2025-08-28 06:20:00 -07:00
oobabooga a92758a144 llama.cpp: Fix obtaining the maximum sequence length for GPT-OSS 2025-08-27 16:15:40 -07:00
oobabooga 030ba7bfeb UI: Mention that Seed-OSS uses enable_thinking 2025-08-27 07:44:35 -07:00
oobabooga 0b4518e61c "Text generation web UI" -> "Text Generation Web UI" 2025-08-27 05:53:09 -07:00
oobabooga 73442a2b6d UI: Better handle the chat input position with CSS
This also solves scrolling issues with the main chat content
when the height of the textarea increases.
2025-08-27 05:43:13 -07:00
oobabooga 8042f76399 Make portable installs functional with Python 3.13 2025-08-27 05:37:01 -07:00
oobabooga ccc8a2229d Revert "UI: Preserve chat scroll position on textarea resize"
This reverts commit 750adf793d.
2025-08-26 13:59:54 -07:00
oobabooga 750adf793d UI: Preserve chat scroll position on textarea resize 2025-08-26 12:19:23 -07:00
oobabooga 02ca96fa44 Multiple fixes 2025-08-25 22:17:22 -07:00
oobabooga 6a7166fffa Add support for the Seed-OSS template 2025-08-25 19:46:48 -07:00
oobabooga 8fcb4b3102 Make bot_prefix extensions functional again 2025-08-25 19:10:46 -07:00
oobabooga 8f660aefe3 Fix chat-instruct replies leaking the bot name sometimes 2025-08-25 18:50:16 -07:00
oobabooga a531328f7e Fix the GPT-OSS stopping string 2025-08-25 18:41:58 -07:00
oobabooga 6c165d2e55 Fix the chat template 2025-08-25 18:28:43 -07:00
oobabooga b657be7381 Obtain stopping strings in chat mode 2025-08-25 18:22:08 -07:00
oobabooga ded6c41cf8 Fix impersonate for chat-instruct 2025-08-25 18:16:17 -07:00
oobabooga c1aa4590ea Code simplifications, fix impersonate 2025-08-25 18:05:40 -07:00
oobabooga b330ec3517 Simplifications 2025-08-25 17:54:15 -07:00
oobabooga 3ad5970374 Make the llama.cpp --verbose output less verbose 2025-08-25 17:43:21 -07:00
oobabooga adeca8a658 Remove changes to the jinja2 templates 2025-08-25 17:36:01 -07:00
oobabooga aad0104c1b Remove a function 2025-08-25 17:33:13 -07:00
oobabooga f919cdf881 chat.py code simplifications 2025-08-25 17:20:51 -07:00
oobabooga d08800c359 chat.py improvements 2025-08-25 17:03:37 -07:00
oobabooga 3bc48014a5 chat.py code simplifications 2025-08-25 16:48:21 -07:00
oobabooga 1f77427088 Update llama.cpp 2025-08-24 19:56:22 -07:00
oobabooga 2478294c06 UI: Preload the instruct and chat fonts 2025-08-24 12:37:41 -07:00
oobabooga 8be798e15f llama.cpp: Fix stderr deadlock while loading some multimodal models 2025-08-24 12:20:05 -07:00
oobabooga 7fe8da8944 Minor simplification after f247c2ae62 2025-08-22 14:42:56 -07:00
oobabooga f247c2ae62 Make --model work with absolute paths, eg --model /tmp/gemma-3-270m-it-IQ4_NL.gguf 2025-08-22 11:47:33 -07:00
oobabooga fd41f2fafc Update llama.cpp 2025-08-22 11:18:56 -07:00
oobabooga cb00db15c9
Merge pull request #7205 from oobabooga/dev
Merge dev branch
2025-08-19 11:51:06 -03:00
oobabooga 9e7b326e34 Lint 2025-08-19 06:50:40 -07:00
oobabooga 1972479610 Add the TP option to exllamav3_HF 2025-08-19 06:48:22 -07:00
oobabooga e0f5905a97 Code formatting 2025-08-19 06:34:05 -07:00
oobabooga 5b06284a8a UI: Keep ExLlamav3_HF selected if already selected for EXL3 models 2025-08-19 06:23:21 -07:00
oobabooga cbba58bef9 UI: Fix code blocks having an extra empty line 2025-08-18 15:50:09 -07:00
oobabooga 8805a50d24 Update llama.cpp 2025-08-18 15:31:01 -07:00
oobabooga 7d23a55901 Fix model unloading when switching loaders (closes #7203) 2025-08-18 09:05:47 -07:00
oobabooga 08594e5263 Installer: Slight improvement 2025-08-18 05:59:46 -07:00
oobabooga 15f99b1b71 Installer: Fix a requirement file 2025-08-18 05:51:46 -07:00
oobabooga 6b1b2e2373 Update README 2025-08-17 22:19:20 -07:00
oobabooga 8a14aa62ff Update README 2025-08-17 22:06:59 -07:00
oobabooga 8cdb911a6e Update README 2025-08-17 22:06:12 -07:00
oobabooga 6bf31479d9 Update README 2025-08-17 22:00:21 -07:00
oobabooga 320f7339cd Update README 2025-08-17 21:56:35 -07:00
oobabooga 3dec47eaf8 Small one-click installer changes 2025-08-17 21:43:46 -07:00
oobabooga 35707c2dd8 Update README 2025-08-17 21:39:57 -07:00
oobabooga 58797a9eb5 Minor change after 9651b5c873 2025-08-17 14:18:23 -07:00
oobabooga 64eba9576c mtmd: Fix a bug when "include past attachments" is unchecked 2025-08-17 14:08:40 -07:00
oobabooga 3a91ca2dd1 Update flash attention 2025-08-17 13:57:23 -07:00
oobabooga 9651b5c873 Make CUDA 12.8 the default CUDA option, remove the CUDA 12.4 option
Exllamav3 doesn't compile with torch 2.6 anymore, and torch 2.7
requires newer CUDA.
2025-08-17 13:26:09 -07:00
oobabooga a633793a00 Bump exllamav3 to 0.0.6 2025-08-17 13:19:42 -07:00
oobabooga dbabe67e77 ExLlamaV3: Enable the --enable-tp option, add a --tp-backend option 2025-08-17 13:19:11 -07:00
oobabooga d771ca4a13 Fix web search (attempt) 2025-08-14 12:05:14 -07:00
oobabooga 73a8a737b2 docs: Improve the multimodal examples slightly 2025-08-13 18:23:18 -07:00
altoiddealer 57f6e9af5a
Set multimodal status during Model Loading (#7199) 2025-08-13 16:47:27 -03:00
oobabooga 45e2935e87
Merge pull request #7198 from oobabooga/dev
Merge dev branch
2025-08-13 10:50:09 -03:00
oobabooga 725a8bcf60 Small docs change 2025-08-13 06:49:28 -07:00
oobabooga 331eab81f7 mtmd: Explain base64 inputs in the API docs 2025-08-13 06:46:10 -07:00
oobabooga 8c9a7e1334
Merge pull request #7195 from oobabooga/dev
Merge dev branch
2025-08-12 18:20:24 -03:00
oobabooga bd05fb899e Update README 2025-08-12 14:19:18 -07:00
oobabooga 6c2fdfdbda
Merge pull request #7190 from oobabooga/dev
Merge dev branch
2025-08-12 18:14:53 -03:00
oobabooga 41b95e9ec3 Lint 2025-08-12 13:37:37 -07:00
oobabooga 2f979ce294 docs: Add a multimodal tutorial 2025-08-12 13:33:49 -07:00
oobabooga 7301452b41 UI: Minor info message change 2025-08-12 13:23:24 -07:00
oobabooga 8d7b88106a Revert "mtmd: Fail early if images are provided but the model doesn't support them (llama.cpp)"
This reverts commit d8fcc71616.
2025-08-12 13:20:16 -07:00
oobabooga 2f6a629393 UI: Minor improvement after 0e88a621fd 2025-08-12 08:51:01 -07:00
oobabooga 2238302b49 ExLlamaV3: Add speculative decoding 2025-08-12 08:50:45 -07:00
oobabooga 0882970a94 Update llama.cpp 2025-08-12 07:00:24 -07:00
oobabooga d8fcc71616 mtmd: Fail early if images are provided but the model doesn't support them (llama.cpp) 2025-08-11 18:02:33 -07:00
oobabooga e6447cd24a mtmd: Update the llama-server request 2025-08-11 17:42:35 -07:00
oobabooga c47e6deda2 Update README 2025-08-11 16:20:20 -07:00
oobabooga 0e3def449a llama.cpp: --swa-full to llama-server when streaming-llm is checked 2025-08-11 15:17:25 -07:00
oobabooga 0e88a621fd UI: Better organize the right sidebar 2025-08-11 15:16:03 -07:00
oobabooga 1e3c4e8bdb Update llama.cpp 2025-08-11 14:40:59 -07:00
oobabooga 765af1ba17 API: Improve a validation 2025-08-11 12:39:48 -07:00
oobabooga a78ca6ffcd Remove a comment 2025-08-11 12:33:38 -07:00
oobabooga dfd9c60d80 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2025-08-11 12:33:27 -07:00
oobabooga 999471256c Lint 2025-08-11 12:32:17 -07:00
Mykeehu 1ba1211ca0
Fix edit window and buttons in Messenger theme (#7100) 2025-08-11 16:13:56 -03:00
oobabooga b10d525bf7 UI: Update a tooltip 2025-08-11 12:05:22 -07:00
oobabooga b62c8845f3 mtmd: Fix /chat/completions for llama.cpp 2025-08-11 12:01:59 -07:00
oobabooga 38c0b4a1ad Default ctx-size to 8192 when not found in the metadata 2025-08-11 07:39:53 -07:00
oobabooga 52d1cbbbe9 Fix an import 2025-08-11 07:38:39 -07:00
oobabooga 1cb800d392 Docs: small change 2025-08-11 07:37:10 -07:00
oobabooga 4809ddfeb8 Exllamav3: small sampler fixes 2025-08-11 07:35:22 -07:00
oobabooga 4d8dbbab64 API: Fix sampler_priority usage for ExLlamaV3 2025-08-11 07:26:11 -07:00
oobabooga c5340533c0 mtmd: Add another API example 2025-08-10 20:39:04 -07:00
oobabooga 9ec310d858 UI: Fix the color of italic text 2025-08-10 07:54:21 -07:00
oobabooga cc964ee579 mtmd: Increase the size of the UI image preview 2025-08-10 07:44:38 -07:00
oobabooga 6fbf162d71 Default max_tokens to 512 in the API instead of 16 2025-08-10 07:21:55 -07:00
oobabooga 1fb5807859 mtmd: Fix API text completion when no images are sent 2025-08-10 06:54:44 -07:00
oobabooga 0ea62d88f6 mtmd: Fix "continue" when an image is present 2025-08-09 21:47:02 -07:00
oobabooga 4663b1a56e Update docs 2025-08-09 21:45:50 -07:00
oobabooga 2f90ac9880 Move the new image_utils.py file to modules/ 2025-08-09 21:41:38 -07:00
oobabooga c6b4d1e87f Fix the exllamav2 loader ignoring add_bos 2025-08-09 21:34:35 -07:00
oobabooga d86b0ec010
Add multimodal support (llama.cpp) (#7027) 2025-08-10 01:27:25 -03:00
oobabooga eb16f64017 Update llama.cpp 2025-08-09 17:12:16 -07:00
oobabooga a289a92b94 Fix exllamav3 token count 2025-08-09 17:10:58 -07:00
oobabooga d489eb589a Attempt at fixing new exllamav3 loader undefined behavior when switching conversations 2025-08-09 14:11:31 -07:00
oobabooga a6d6bee88c Change a comment 2025-08-09 07:51:03 -07:00
oobabooga 2fe79a93cc mtmd: Handle another case after 3f5ec9644f 2025-08-09 07:50:24 -07:00
oobabooga 59c6138e98 Remove a log message 2025-08-09 07:32:15 -07:00
oobabooga f396b82a4f mtmd: Better way to detect if an EXL3 model is multimodal 2025-08-09 07:31:36 -07:00
oobabooga fa9be444fa Use ExLlamav3 instead of ExLlamav3_HF by default for EXL3 models 2025-08-09 07:26:59 -07:00
oobabooga d9db8f63a7 mtmd: Simplifications 2025-08-09 07:25:42 -07:00
oobabooga 3f5ec9644f mtmd: Place the image <__media__> at the top of the prompt 2025-08-09 07:06:07 -07:00
oobabooga 1168004067 Minor change 2025-08-09 07:01:55 -07:00
oobabooga 9e260332cc Remove some unnecessary code 2025-08-08 21:22:47 -07:00
oobabooga 544c3a7c9f Polish the new exllamav3 loader 2025-08-08 21:15:53 -07:00
oobabooga 8fcadff8d3 mtmd: Use the base64 attachment for the UI preview instead of the file 2025-08-08 20:13:54 -07:00
oobabooga 6e9de75727 Support loading chat templates from chat_template.json files 2025-08-08 19:35:09 -07:00
Katehuuh 88127f46c1
Add multimodal support (ExLlamaV3) (#7174) 2025-08-08 23:31:16 -03:00
oobabooga b391ac8eb1 Fix getting the ctx-size for EXL3/EXL2/Transformers models 2025-08-08 18:11:45 -07:00
oobabooga 88ba4b1ebf
Merge pull request #7181 from oobabooga/dev
Merge dev branch
2025-08-07 00:30:46 -03:00
oobabooga f1147c9926 Update llama.cpp 2025-08-06 19:32:36 -07:00
oobabooga 3e24f455c8 Fix continue for GPT-OSS (hopefully the final fix) 2025-08-06 10:18:42 -07:00
oobabooga 0c1403f2c7 Handle GPT-OSS as a special case when continuing 2025-08-06 08:05:37 -07:00
oobabooga 6ce4b353c4 Fix the GPT-OSS template 2025-08-06 07:12:39 -07:00
oobabooga fefdb20f69
Merge pull request #7180 from oobabooga/dev
Merge dev branch
2025-08-05 23:54:32 -03:00
oobabooga 7c82d65a9d Handle GPT-OSS as a special template case 2025-08-05 18:05:09 -07:00
oobabooga fbea21a1f1 Only use enable_thinking if the template supports it 2025-08-05 17:33:27 -07:00
oobabooga bfbbfc2361 Ignore add_generation_prompt in GPT-OSS 2025-08-05 17:33:01 -07:00
oobabooga 20adc3c967 Start over new template handling (to avoid overcomplicating) 2025-08-05 16:58:45 -07:00
oobabooga 80f6abb07e Begin fixing 'Continue' with GPT-OSS 2025-08-05 16:01:19 -07:00
oobabooga e5b8d4d072 Fix a typo 2025-08-05 15:52:56 -07:00
oobabooga 701048cf33 Try to avoid breaking jinja2 parsing for older models 2025-08-05 15:51:24 -07:00
oobabooga 7d98ca6195 Make web search functional with thinking models 2025-08-05 15:44:33 -07:00
oobabooga 0e42575c57 Fix thinking block parsing for GPT-OSS under llama.cpp 2025-08-05 15:36:20 -07:00
oobabooga 498778b8ac Add a new 'Reasoning effort' UI element 2025-08-05 15:19:11 -07:00
oobabooga 6bb8212731 Fix thinking block rendering for GPT-OSS 2025-08-05 15:06:22 -07:00
oobabooga 42e3a7a5ae Update llama.cpp 2025-08-05 14:56:12 -07:00
oobabooga 5c5a4dfc14 Fix impersonate 2025-08-05 13:04:10 -07:00
oobabooga ecd16d6bf9 Automatically set skip_special_tokens to False for channel-based templates 2025-08-05 12:57:49 -07:00
oobabooga 178c3e75cc Handle templates with channels separately 2025-08-05 12:52:17 -07:00
oobabooga 9f28f53cfc Better parsing of the gpt-oss template 2025-08-05 11:56:00 -07:00
oobabooga 3b28dc1821 Don't pass torch_dtype to transformers loader, let it be autodetected 2025-08-05 11:35:53 -07:00
oobabooga 3039aeffeb Fix parsing the gpt-oss-20b template 2025-08-05 11:35:17 -07:00
oobabooga 5989043537 Transformers: Support standalone .jinja chat templates (for GPT-OSS) 2025-08-05 11:22:18 -07:00
oobabooga 02a3420a50 Bump transformers to 4.55 (adds gpt-oss support) 2025-08-05 10:09:30 -07:00
oobabooga 74230f559a Bump transformers to 4.54 2025-08-01 11:03:15 -07:00
oobabooga f08bb9a201 Handle edge case in chat history loading (closes #7155) 2025-07-24 10:34:59 -07:00
oobabooga d746484521 Handle both int and str types in grammar char processing 2025-07-23 11:52:51 -07:00
oobabooga 714f745713
Merge pull request #7141 from oobabooga/dev
Merge dev branch
2025-07-19 17:54:06 -03:00
oobabooga 0c667de7a7 UI: Add a None option for the speculative decoding model (closes #7145) 2025-07-19 12:14:41 -07:00
oobabooga ccf5e3e3a7 Update exllamav3 2025-07-19 12:07:38 -07:00
oobabooga a00983b2ba Update llama.cpp 2025-07-19 12:07:20 -07:00
oobabooga 9371867238 Update exllamav2 2025-07-15 07:38:03 -07:00
oobabooga 03fb85e49a Update llama.cpp 2025-07-15 07:37:13 -07:00
oobabooga 845432b9b4 Remove the obsolete modules/relative_imports.py file 2025-07-14 21:03:18 -07:00
oobabooga 1d1b20bd77 Remove the --torch-compile option (it doesn't do anything currently) 2025-07-11 10:51:23 -07:00
oobabooga 5a8a9c22e8 Update llama.cpp 2025-07-11 09:20:27 -07:00
oobabooga 273888f218 Revert "Use eager attention by default instead of sdpa"
This reverts commit bd4881c4dc.
2025-07-10 18:56:46 -07:00
oobabooga caf69d871a Revert "Standardize margins and paddings across all chat styles"
This reverts commit 86cb5e0587.
2025-07-10 18:43:01 -07:00
oobabooga 188c7c8f2b Revert "CSS simplifications"
This reverts commit c6c1b725e9.
2025-07-10 18:42:52 -07:00
oobabooga 635e6efd18 Ignore add_bos_token in instruct prompts, let the jinja2 template decide 2025-07-10 07:14:01 -07:00
oobabooga 0f3a88057c Don't downgrade triton-windows on CUDA 12.8 2025-07-10 05:39:04 -07:00
oobabooga e523f25b9f Downgrade triton-windows to 3.2.0.post19
https://github.com/oobabooga/text-generation-webui/issues/7107#issuecomment-3057250374
2025-07-10 05:35:57 -07:00
oobabooga a7a3a0c700 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2025-07-09 21:07:42 -07:00
oobabooga 21e0e9f32b Add the triton-windows requirement on Windows to make transformers functional 2025-07-09 21:05:17 -07:00
dependabot[bot] d1f4622a96
Update peft requirement from ==0.15.* to ==0.16.* in /requirements/full (#7127) 2025-07-10 00:15:50 -03:00
oobabooga e015355e4a Update README 2025-07-09 20:03:53 -07:00
oobabooga bd4881c4dc Use eager attention by default instead of sdpa 2025-07-09 19:57:37 -07:00
oobabooga b69f435311 Fix latest transformers being super slow 2025-07-09 19:56:50 -07:00
oobabooga 8b3c7aa795 Bump bitsandbytes to 0.46 2025-07-09 19:46:55 -07:00
oobabooga f045b72826 Bump accelerate to 1.8 2025-07-09 19:46:26 -07:00
oobabooga c357601c01 Bump transformers to 4.53 2025-07-09 18:48:04 -07:00
oobabooga 6c2bdda0f0 Transformers loader: replace use_flash_attention_2/use_eager_attention with a unified attn_implementation
Closes #7107
2025-07-09 18:39:37 -07:00
oobabooga 6338dc0051
Merge pull request #7129 from oobabooga/dev
Merge dev branch
2025-07-09 00:10:16 -03:00
oobabooga 511bb31646 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2025-07-08 20:04:37 -07:00
oobabooga d1e9301a43 Remove fragile js from 9a58964834 2025-07-08 19:57:46 -07:00
Cats cd5d867b62
docs: Add Mirostat Explanation (#7128) 2025-07-08 17:54:38 -03:00
oobabooga 3e24a127c7 Remove more unnecessary files from portable builds 2025-07-08 09:13:11 -07:00
oobabooga 2f544fe199 Update the keyboard shortcuts documentation 2025-07-08 09:02:42 -07:00
oobabooga 93e08c0d4a Update README 2025-07-08 08:59:29 -07:00
oobabooga 42191a36ab Keep navigation icons visible when switching versions 2025-07-08 07:10:04 -07:00
oobabooga c6c1b725e9 CSS simplifications 2025-07-07 21:11:13 -07:00
oobabooga 86cb5e0587 Standardize margins and paddings across all chat styles 2025-07-07 21:02:19 -07:00
oobabooga b7d5982944
Merge pull request #7125 from oobabooga/dev
Merge dev branch
2025-07-07 18:19:58 -03:00
oobabooga e8266b0356 Use windows-2022 in workflows 2025-07-07 14:19:20 -07:00
oobabooga e1034fc79e
Merge pull request #7124 from oobabooga/dev
Merge dev branch
2025-07-07 18:13:30 -03:00
oobabooga 74d98186fc Slightly more robust autoscroll 2025-07-07 13:23:23 -07:00
oobabooga ca226a54c6 Disable the message version navigation hover effects during streaming 2025-07-07 11:29:37 -07:00
oobabooga 07e6f004c5 Rename a button in the Session tab for clarity 2025-07-07 11:28:47 -07:00
oobabooga 426e7a4cec Update the extensions documentation 2025-07-07 08:43:01 -07:00
oobabooga e52bc0acb2 Update llama.cpp 2025-07-06 20:28:35 -07:00
oobabooga cbef2720ce Revert "Fix: use embedded Python in start_windows.bat to avoid system interpreter conflicts (#7120)"
This reverts commit 8df1127ce2.
2025-07-06 20:14:02 -07:00
Alidr79 e5767d4fc5
Update ui_model_menu.py blocking the --multi-user access in backend (#7098) 2025-07-06 21:48:53 -03:00
oobabooga 60123a67ac Better log message when extension requirements are not found 2025-07-06 17:44:41 -07:00
oobabooga e6bc7742fb Support installing user extensions in user_data/extensions/ 2025-07-06 17:30:23 -07:00
Philipp Claßen 959d4ddb91
Fix for chat sidebars toggle buttons disappearing (#7106) 2025-07-06 20:51:42 -03:00
Ali 8df1127ce2
Fix: use embedded Python in start_windows.bat to avoid system interpreter conflicts (#7120) 2025-07-06 20:42:34 -03:00
oobabooga de4ccffff8 Fix the duckduckgo search 2025-07-06 16:24:57 -07:00
oobabooga 0f258774d3 Minor README changes 2025-07-05 14:25:59 -07:00
oobabooga 4583924ce7 Remove torchvision/torchaudio mentions from the README 2025-07-05 14:24:15 -07:00
oobabooga c4d738f39f Update llama.cpp 2025-07-05 14:09:29 -07:00
oobabooga c4d5331c03 Fix autoscroll after fonts load 2025-07-04 13:21:52 -07:00
oobabooga 92ec8dda03 Fix chat history getting lost if the UI is inactive for a long time (closes #7109) 2025-07-04 06:04:04 -07:00
oobabooga 23bb94a5fb Update llama.cpp 2025-07-03 20:36:54 -07:00
zombiegreedo 877c651c04
Handle either missing <think> start or </think> end tags (#7102) 2025-07-03 23:05:46 -03:00
oobabooga cbba88f565 Fix scrolling during streaming when thinking blocks are present 2025-07-03 18:16:29 -07:00
oobabooga 13373391df Rename miniconda -> miniforge everywhere 2025-07-03 14:13:22 -07:00
oobabooga ab162f976c Use miniforge instead of miniconda to avoid anaconda licensing issues 2025-07-03 11:31:52 -07:00
oobabooga 9a58964834 Keep the last message visible when the input height changes 2025-06-22 20:44:04 -07:00
oobabooga c3faecfd27 Minor change 2025-06-22 17:51:09 -07:00
oobabooga 1b19dd77a4 Move 'Enable thinking' to the Chat tab 2025-06-22 17:29:17 -07:00
oobabooga 02f604479d Remove the pre-jinja2 custom stopping string handling (closes #7094) 2025-06-21 14:03:35 -07:00
oobabooga 58282f7107 Replace 'Generate' with 'Send' in the Chat tab 2025-06-20 06:59:48 -07:00
oobabooga bb97ca1b22 Fix a small issue with the chat input 2025-06-19 21:41:41 -07:00
oobabooga f154aeafea Optimize chat scrolling for the 40th time, hopefully the last one 2025-06-19 21:23:10 -07:00
oobabooga 17f9c188bd
Merge pull request #7092 from oobabooga/dev
Merge dev branch
2025-06-19 19:42:16 -03:00
oobabooga acd57b6a85 Minor UI change 2025-06-19 15:39:43 -07:00
oobabooga f08db63fbc Change some comments 2025-06-19 15:26:45 -07:00
oobabooga 2517ea9c9e Lint 2025-06-19 15:23:06 -07:00
oobabooga 90f42f311a Update README 2025-06-19 12:43:05 -07:00
oobabooga ee945517ff Update README 2025-06-19 12:39:53 -07:00
oobabooga a1b606a6ac Fix obtaining the maximum number of GPU layers for DeepSeek-R1-0528-GGUF 2025-06-19 12:30:57 -07:00
oobabooga 3344510553 Force dark theme on the Gradio login page 2025-06-19 12:11:34 -07:00
oobabooga 645463b9f0 Add fallback values for theme colors 2025-06-19 11:28:12 -07:00
oobabooga 09cd1cb4e2 Update README 2025-06-19 10:51:45 -07:00
oobabooga c4029914e8 Update README 2025-06-19 10:48:33 -07:00
oobabooga 84617abdeb Properly fix the /v1/models endpoint 2025-06-19 10:25:55 -07:00
oobabooga 93cd47c948 Bump numpy to 2.2 (loses #7090) 2025-06-19 08:00:30 -07:00
oobabooga dcdc42fa06 Fix the /v1/models output format (closes #7089) 2025-06-19 07:57:17 -07:00
oobabooga 9c6913ad61 Show file sizes on "Get file list" 2025-06-18 21:35:07 -07:00
oobabooga 9bd114b5d7 Merge branch 'main' into dev 2025-06-18 21:03:52 -07:00
oobabooga 76a722dc90 Remove .github and .gitignore folders from portable builds 2025-06-18 21:03:45 -07:00
oobabooga 4e0dfbdde3 Remove .github and .gitignore folders from portable builds 2025-06-18 21:02:57 -07:00
oobabooga 92547becff
Merge pull request #7085 from oobabooga/dev
Merge dev branch
2025-06-18 22:43:07 -03:00
oobabooga 0cb82483ef Lint 2025-06-18 18:26:59 -07:00
oobabooga e33921a629 Fix jittering while typing on Firefox (closes #7086) 2025-06-18 17:54:34 -07:00
oobabooga 6af3598cfa API: Remove obsolete list_dummy_models function 2025-06-18 16:15:42 -07:00
NoxWorld2660 0b26650f47
Expose real model list via /v1/models endpoint (#7088) 2025-06-18 20:14:24 -03:00
oobabooga 6cc7bbf009 Better autosave behavior for notebook tab when there are 2 columns 2025-06-18 15:54:32 -07:00
oobabooga 197b327374 Minor log message change 2025-06-18 13:36:54 -07:00
oobabooga 2f45d75309 Increase the area of the notebook textbox 2025-06-18 13:22:06 -07:00
oobabooga 7cb2b1bfdb Fix some events 2025-06-18 10:27:38 -07:00
oobabooga 8b7eb5c87c Code simplification 2025-06-18 10:22:36 -07:00
oobabooga 22cc9e0115 Remove 'Send to Default' 2025-06-18 10:21:48 -07:00
oobabooga 678f40297b Clear the default tab output when switching prompts 2025-06-17 17:40:48 -07:00
oobabooga a2cdd06afc Revert "Workaround for jittering while typing on firefox"
This reverts commit b4edfce993.
2025-06-17 15:29:40 -07:00
oobabooga 2d37602382 Small improvements to wpp style 2025-06-17 15:26:59 -07:00
oobabooga da148232eb Better filenames for new prompts in the Notebook tab 2025-06-17 15:10:44 -07:00
oobabooga fc23345c6d Send the default input to the notebook textbox when switching 2 columns to 1 (instead of the output) 2025-06-17 15:03:14 -07:00
oobabooga 75217d3713 Change issue template 2025-06-17 09:37:24 -07:00
oobabooga b4edfce993 Workaround for jittering while typing on firefox 2025-06-17 09:30:03 -07:00
oobabooga 01ef4c61bd Only open/close both sidebars at the same time on desktop 2025-06-17 08:45:11 -07:00
oobabooga 315e06f695 Update llama.cpp 2025-06-17 07:51:16 -07:00
oobabooga 73138a29fa Small change 2025-06-17 07:49:24 -07:00
oobabooga 87ae09ecd6 Improve the basic API examples 2025-06-17 07:46:58 -07:00
oobabooga aa44e542cb Revert "Safer usage of mkdir across the project"
This reverts commit 0d1597616f.
2025-06-17 07:11:59 -07:00
oobabooga 0d1597616f Safer usage of mkdir across the project 2025-06-17 07:09:33 -07:00
oobabooga 8689d7ecea Update README 2025-06-16 21:21:39 -07:00
oobabooga 8f49e6144e Update README 2025-06-16 21:09:45 -07:00
oobabooga 66e991841a Fix the character pfp not appearing when switching from instruct to chat modes 2025-06-16 18:45:44 -07:00
oobabooga be3d371290 Close the big profile picture when switching to instruct mode 2025-06-16 18:42:17 -07:00
oobabooga 26eda537f0 Add auto-save for notebook textbox while typing 2025-06-16 17:48:23 -07:00
oobabooga 88c0204357 Disable start_with when generating the websearch query 2025-06-16 14:53:05 -07:00
oobabooga 97a539cab6 Minor style change 2025-06-16 13:55:45 -07:00
oobabooga faae4dc1b0
Autosave generated text in the Notebook tab (#7079) 2025-06-16 17:36:05 -03:00
oobabooga d0befe0729 Add a comment 2025-06-16 09:22:22 -07:00
oobabooga de24b3bb31
Merge the Default and Notebook tabs into a single Notebook tab (#7078) 2025-06-16 13:19:29 -03:00
oobabooga db67d69ddc Lint 2025-06-16 07:28:14 -07:00
oobabooga cac225b589 Small style improvements 2025-06-16 07:26:39 -07:00
oobabooga 7ba3d4425f Remove the 'Send to negative prompt' button 2025-06-16 07:23:09 -07:00
oobabooga 34bf93ef47 Move 'Custom system message' to the Parameters tab 2025-06-16 07:22:14 -07:00
oobabooga c9c3b716fb Move character settings to a new 'Character' main tab 2025-06-16 07:21:25 -07:00
oobabooga f77f1504f5 Improve the style of the Character and User tabs 2025-06-16 06:12:37 -07:00
oobabooga 949b7ec9cf Further optimize scrolling in the chat tab 2025-06-15 18:50:21 -07:00
oobabooga d347b056e3 Always close/open the two sidebars at the same time 2025-06-15 18:12:11 -07:00
oobabooga 9bcef8a648 Fix "show controls" conflicting with manually hiding the sidebars 2025-06-15 17:57:41 -07:00
oobabooga bc2b0f54e9 Only save extensions settings on manual save 2025-06-15 15:53:16 -07:00
oobabooga cc757f6226 Small style improvements to the chat tab 2025-06-15 08:32:06 -07:00
oobabooga b279460a81 Improve the wpp style 2025-06-15 08:25:07 -07:00
oobabooga e8dc7b0ee9 Bump exllamav3 to 0.0.4 2025-06-15 08:15:29 -07:00
oobabooga 4fc254c1dd Optimize syntax highlighting on long conversations 2025-06-15 08:13:13 -07:00
oobabooga 609c3ac893 Optimize the end of generation with llama.cpp 2025-06-15 08:03:27 -07:00
oobabooga db7d717df7 Remove images and links from websearch results
This reduces noise a lot
2025-06-14 20:00:25 -07:00
oobabooga e263dbf852 Improve user input truncation 2025-06-14 19:43:51 -07:00
oobabooga 09606a38d3 Truncate web search results to at most 8192 tokens 2025-06-14 19:37:32 -07:00
oobabooga ad0be25c46 Update llama.cpp 2025-06-14 15:00:14 -07:00
oobabooga 7c0225931a Merge branch 'main' into dev 2025-06-14 14:59:37 -07:00
oobabooga 1c1cf09a59 Update workflows 2025-06-14 14:52:49 -07:00
oobabooga 58c3b549ba Merge branch 'main' into dev 2025-06-14 10:16:13 -07:00
oobabooga 8e9c0287aa UI: Fix edge case where gpu-layers slider maximum is incorrectly limited 2025-06-14 10:12:11 -07:00
oobabooga 8e0ef5b419 Hide the header bar on Ctrl+S 2025-06-14 09:09:46 -07:00
oobabooga 1d23159837 Increase the size of the enlarged character profile picture 2025-06-14 08:45:59 -07:00
oobabooga d2da40b0e4 Remember the last selected chat for each mode/character 2025-06-14 08:25:00 -07:00
oobabooga 879fa3d8c4 Improve the wpp style & simplify the code 2025-06-14 07:14:22 -07:00
oobabooga 09eb326486 Merge README.md changes from dev branch 2025-06-13 07:46:43 -07:00
oobabooga dfab11f0b5 Update README 2025-06-13 07:45:42 -07:00
oobabooga 9a2353f97b Better log message when the user input gets truncated 2025-06-13 05:44:02 -07:00
oobabooga 322cd28e24 Update README 2025-06-13 01:27:33 -07:00
oobabooga 7cb650237c Update the README 2025-06-13 01:12:52 -07:00
oobabooga aab28398ef Update README 2025-06-13 01:06:44 -07:00
oobabooga 5ba52967ac Update README 2025-06-13 01:04:41 -07:00
oobabooga b58e80cb99 Update README 2025-06-13 01:02:11 -07:00
Miriam f4f621b215
ensure estimated vram is updated when switching between different models (#7071) 2025-06-13 02:56:33 -03:00
oobabooga f337767f36 Add error handling for non-llama.cpp models in portable mode 2025-06-12 22:17:39 -07:00
oobabooga a25a1fc8d0 Disable message action icons during streaming for better performance 2025-06-12 22:01:02 -07:00
oobabooga 2dee3a66ff Add an option to include/exclude attachments from previous messages in the chat prompt 2025-06-12 21:37:18 -07:00
oobabooga 2cfb77d16f
Merge pull request #7070 from oobabooga/dev
Merge dev branch
2025-06-12 12:38:47 -03:00
oobabooga b4d2a00e20 Update README 2025-06-12 08:35:33 -07:00
oobabooga 9ff5961853
Merge pull request #7067 from oobabooga/dev
Merge dev branch
2025-06-11 11:58:52 -03:00
oobabooga 9d6a7f1bcf Minor changes 2025-06-11 07:55:35 -07:00
oobabooga 004fd8316c Minor changes 2025-06-11 07:49:51 -07:00
oobabooga 570d5b8936 Only save extensions on manual save 2025-06-11 07:39:49 -07:00
oobabooga 27140f3563 Revert "Don't save active extensions through the UI"
This reverts commit df98f4b331.
2025-06-11 07:25:27 -07:00
oobabooga 2ebc8ff252
Merge pull request #7065 from oobabooga/dev
Merge dev branch
2025-06-11 01:09:06 -03:00
oobabooga 13a5288d01 Fix an error when upgrading from cuda 12.4 to cuda 12.8 2025-06-10 21:08:18 -07:00
oobabooga 801db438b0 Undo changes to portable builds 2025-06-10 19:55:40 -07:00
oobabooga 00fbbd6f57 Undo changes to portable builds 2025-06-10 19:54:42 -07:00
oobabooga e8041069e2
Merge pull request #7064 from oobabooga/dev
Merge dev branch
2025-06-10 23:43:10 -03:00
oobabooga fe0685a742 New attempt 2025-06-10 19:42:22 -07:00
oobabooga 036976aeb8
Merge pull request #7063 from oobabooga/dev
Merge dev branch
2025-06-10 23:35:22 -03:00
oobabooga 43fc170224 Fix the Windows workflow 2025-06-10 19:34:41 -07:00
oobabooga e9a433832e
Merge pull request #7062 from oobabooga/dev
Merge dev branch
2025-06-10 23:26:21 -03:00
oobabooga a86a5a026e Fix the GitHub Actions workflows 2025-06-10 19:25:22 -07:00
oobabooga 1e96dcf369
Merge pull request #7057 from oobabooga/dev
Merge dev branch
2025-06-10 23:08:44 -03:00
oobabooga 552cb09f09 Do not bump Transformers to 4.52 on CUDA 12.8
Performance is slow, and the older version works fine with torch 2.7.
2025-06-10 18:45:42 -07:00
LawnMauer bc921c66e5
Load js and css sources in UTF-8 (#7059) 2025-06-10 22:16:50 -03:00
oobabooga 4cf39120fc Fix chat area sometimes not scrolling up to edit message 2025-06-10 18:03:00 -07:00
oobabooga 75da90190f Fix character dropdown sometimes disappearing in the Parameters tab 2025-06-10 17:34:54 -07:00
oobabooga 1c1fd3be46 Remove some log messages 2025-06-10 14:29:28 -07:00
oobabooga 3f9eb3aad1 Fix the preset dropdown when the default preset file is not present 2025-06-10 14:22:37 -07:00
oobabooga 18bd78f1f0 Make the llama.cpp prompt processing messages shorter 2025-06-10 14:03:25 -07:00
oobabooga 889153952f Lint 2025-06-10 09:02:52 -07:00
oobabooga 2dabdbc7da Update llama.cpp 2025-06-10 05:25:23 -07:00
oobabooga c92eba0b0a Reorganize the Parameters tab (left: preset parameters, right: everything else) 2025-06-09 22:05:20 -07:00
oobabooga efd9c9707b Fix random seeds being saved to settings.yaml 2025-06-09 20:57:25 -07:00
oobabooga df98f4b331 Don't save active extensions through the UI
Prevents command-line activated extensions from becoming permanently active due to autosave.
2025-06-09 20:28:16 -07:00
Mykeehu ec73121020
Fix continue/start reply with when using translation extensions (#6944)
---------

Co-authored-by: oobabooga <oobabooga4@gmail.com>
2025-06-10 00:17:05 -03:00
Miriam 331d03c33f
fix failure when --nowebui called without --api (#7055) 2025-06-09 23:25:39 -03:00
Miriam 1443612e72
check .attention.head_count if .attention.head_count_kv doesn't exist (#7048) 2025-06-09 23:22:01 -03:00
oobabooga d085dc6a93 Minor optimization after e976a5ddc7 2025-06-09 18:40:54 -07:00
oobabooga 263b5d5557 Use html2text to extract the text of web searches without losing formatting 2025-06-09 17:55:26 -07:00
oobabooga f5a5d0c0cb Add the URL of web attachments to the prompt 2025-06-09 17:32:25 -07:00
oobabooga 747a4a0e56 Reposition the ... typing dots 2025-06-09 13:41:29 -07:00
oobabooga 14efd42084 Improve scroll performance by disabling hover effects during scroll 2025-06-09 11:43:15 -07:00
oobabooga 1602ac1c8f Improve the style of thinking blocks in dark mode 2025-06-09 09:03:39 -07:00
oobabooga eefbf96f6a Don't save truncation_length to user_data/settings.yaml 2025-06-08 22:14:56 -07:00
oobabooga 80637cae28 Add version to portable build folder names 2025-06-08 21:55:49 -07:00
oobabooga f9a007c6a8 Properly filter out failed web search downloads from attachments 2025-06-08 19:25:23 -07:00
oobabooga f3388c2ab4 Fix selecting next chat when deleting with active search 2025-06-08 18:53:04 -07:00
oobabooga 4a369e070a Add buttons for easily deleting past chats 2025-06-08 18:47:48 -07:00
oobabooga 0b8d2d65a2 Minor style improvement 2025-06-08 18:11:27 -07:00
oobabooga 06dfb7e772 Improve the style of the hover menu 2025-06-08 18:03:07 -07:00
oobabooga b5e021fc49 Make the dark theme darker 2025-06-08 17:44:04 -07:00
oobabooga e976a5ddc7 Re-highlight code blocks when switching light/dark themes 2025-06-08 17:35:36 -07:00
oobabooga 7ed1926ce7 Small change after previous commit 2025-06-08 15:38:40 -07:00
oobabooga ff01bcb870 Use user_data/cache/gradio for Gradio temp files 2025-06-08 15:33:05 -07:00
oobabooga f81b1540ca Small style improvements 2025-06-08 15:19:25 -07:00
oobabooga eb0ab9db1d Fix light/dark theme persistence across page reloads 2025-06-08 15:04:05 -07:00
oobabooga 78899244d5 Remove settings-template.yaml 2025-06-08 09:40:09 -07:00
oobabooga 1f1435997a Don't show the new 'Restore character' button in the Chat tab 2025-06-08 09:37:54 -07:00
oobabooga 84f66484c5 Make it optional to paste long pasted content to an attachment 2025-06-08 09:31:38 -07:00
oobabooga 42e7864d62 Reorganize the Session tab 2025-06-08 09:21:23 -07:00
oobabooga af6bb7513a Add back the "Save UI defaults" button
It's useful for saving extensions settings.
2025-06-08 09:09:36 -07:00
oobabooga 1cab149c1a Remove the contrastive search preset 2025-06-07 22:26:13 -07:00
oobabooga ae150fa24f Remove the null preset 2025-06-07 22:25:46 -07:00
oobabooga 1bdf11b511 Use the Qwen3 - Thinking preset by default 2025-06-07 22:23:09 -07:00
oobabooga 0dbc4cbc71 Add Qwen3 presets 2025-06-07 22:20:58 -07:00
oobabooga fe955cac1f Small UI changes 2025-06-07 22:15:19 -07:00
oobabooga caf9fca5f3 Avoid some code repetition 2025-06-07 22:11:35 -07:00
oobabooga 3650a6fd1f Small UI changes 2025-06-07 22:02:34 -07:00
oobabooga 6436bf1920
More UI persistence: presets and characters (#7051) 2025-06-08 01:58:02 -03:00
oobabooga 35ed55d18f
UI persistence (#7050) 2025-06-07 22:46:52 -03:00
rakha abadi susilo db847eed4c
Add RTX 50XX Nvidia blackwell support (ExLlamaV2/V3 and Transformers) (#7011)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2025-06-07 21:44:15 -03:00
oobabooga 2d263f227d Fix the chat input reappearing when the page is reloaded 2025-06-06 22:38:20 -07:00
oobabooga 379dd01ca7 Filter out failed web search downloads from attachments 2025-06-06 22:32:07 -07:00
oobabooga f8f23b5489 Simplify the llama.cpp stderr filter code 2025-06-06 22:25:13 -07:00
oobabooga 45f823ddf6 Print \n after the llama.cpp progress bar reaches 1.0 2025-06-06 22:23:34 -07:00
oobabooga d47c8eb956 Remove quotes from LLM-generated websearch query (closes #7045).
Fix by @Quiet-Joker
2025-06-05 06:57:59 -07:00
oobabooga 977ec801b7 Improve table colors in instruct mode 2025-06-05 06:33:45 -07:00
Hanusz Leszek 3829507d0f
Stop model during graceful shutdown (#7042) 2025-06-04 15:13:36 -03:00
oobabooga 3d676cd50f Optimize syntax highlighting 2025-06-04 11:02:04 -07:00
oobabooga 66a75c899a Improve the scrollbars in code blocks 2025-06-04 10:59:43 -07:00
oobabooga 9bd7359ffa Scroll the textarea into view when editing a message 2025-06-04 10:47:14 -07:00
oobabooga 93b3752cdf Revert "Remove the "Is typing..." yield by default"
This reverts commit b30a73016d.
2025-06-04 09:40:30 -07:00
oobabooga b38ec0ec38 Update llama.cpp 2025-06-02 11:33:17 -07:00
oobabooga b30a73016d Remove the "Is typing..." yield by default 2025-06-02 07:49:22 -07:00
oobabooga 7278548cd1
Simplify the one-click installer (#7039) 2025-06-02 09:57:55 -03:00
oobabooga bb409c926e
Update only the last message during streaming + add back dynamic UI update speed (#7038) 2025-06-02 09:50:17 -03:00
oobabooga 45c9ae312c Use the flash-attention wheels in https://github.com/kingbri1/flash-attention 2025-06-01 22:17:22 -07:00
oobabooga 2db7745cbd Show llama.cpp prompt processing on one line instead of many lines 2025-06-01 22:12:24 -07:00
oobabooga ad6d0218ae Fix after 219f0a7731 2025-06-01 19:27:14 -07:00
oobabooga 92adceb7b5 UI: Fix the model downloader progress bar 2025-06-01 19:22:21 -07:00
oobabooga 7a81beb0c1 Turn long pasted text into an attachment automatically 2025-06-01 18:26:14 -07:00
oobabooga bf42b2c3a1 Fix thinking blocks sometimes showing a white outline 2025-06-01 11:02:04 -07:00
oobabooga 83849336d8 Improve how Show controls looks in the hover menu 2025-06-01 10:58:49 -07:00
oobabooga 3e3746283c Improve the typing dots position 2025-06-01 10:55:31 -07:00
oobabooga 88ff3e6ad8 CSS fixes after 98a7508a99 2025-06-01 08:04:35 -07:00
oobabooga 9e80193008 Add the model name to each message's metadata 2025-05-31 22:41:35 -07:00
oobabooga 0816ecedb7 Lint 2025-05-31 22:25:09 -07:00
oobabooga 98a7508a99 UI: Move 'Show controls' inside the hover menu 2025-05-31 22:22:13 -07:00
oobabooga 85f2f01a3a UI: Fix extra gaps on the right sidebar 2025-05-31 21:29:57 -07:00
oobabooga f8d220c1e6 Add a tooltip to the web search checkbox 2025-05-31 21:22:36 -07:00
oobabooga 4a2727b71d Add a tooltip to the file upload button 2025-05-31 20:24:31 -07:00
oobabooga 1d88456659 Add support for .docx attachments 2025-05-31 20:15:07 -07:00
oobabooga dc8ed6dbe7 Bump exllamav3 to 0.0.3 2025-05-31 14:27:33 -07:00
oobabooga c55d3c61c6 Bump exllamav2 to 0.3.1 2025-05-31 14:21:42 -07:00
oobabooga ae61c1a0f4
Merge pull request #7034 from oobabooga/dev
Merge dev branch
2025-05-30 23:07:56 -03:00
oobabooga 15f466ca3f Update README 2025-05-30 15:49:57 -07:00
oobabooga 219f0a7731 Fix exllamav3_hf models failing to unload (closes #7031) 2025-05-30 12:05:49 -07:00
oobabooga 298d4719c6 Multiple small style improvements 2025-05-30 11:32:24 -07:00
oobabooga 7c29879e79 Fix 'Start reply with' (closes #7033) 2025-05-30 11:17:47 -07:00
oobabooga af1eef1b08
Merge pull request #7028 from oobabooga/dev
Merge dev branch
2025-05-29 19:07:56 -03:00
oobabooga 28e6bd4fcd Revert "Update transformers requirement in /requirements/full (#7017)"
This reverts commit cc9b7253c1.
2025-05-29 14:49:07 -07:00
oobabooga d1bfb08e8d Improve the style of message editing 2025-05-29 14:27:47 -07:00
oobabooga acbcc12e7b Clean up 2025-05-29 14:11:21 -07:00
oobabooga dce02732a4 Fix timestamp issues when editing/swiping messages 2025-05-29 14:08:48 -07:00
oobabooga 8078c41ec6 Revert "Bump llama.cpp"
This reverts commit a8d02dec8f.
2025-05-29 13:32:19 -07:00
oobabooga a45a652130 CSS fix 2025-05-29 13:28:51 -07:00
oobabooga f59998d268 Don't limit the number of prompt characters printed with --verbose 2025-05-29 13:08:48 -07:00
oobabooga aff41f3482 Update README 2025-05-29 12:53:41 -07:00
oobabooga e7129f9dbe Prevent footer buttons below last assistant message from always appearing 2025-05-29 12:47:07 -07:00
oobabooga 724147ffab Better detect when no model is available 2025-05-29 10:49:29 -07:00
oobabooga faa5c82c64 Fix message version count not updating during regeneration streaming 2025-05-29 09:16:26 -07:00
oobabooga 3f37a2e915 Update README 2025-05-29 08:49:31 -07:00
oobabooga c970c5f166 Make scrollbars darker in dark theme 2025-05-29 08:15:13 -07:00
oobabooga 81794692ab UI: Make the dark theme darker 2025-05-29 08:07:14 -07:00
oobabooga 36bc276005 Update README 2025-05-29 05:39:26 -07:00
oobabooga 0986d075fb Update README 2025-05-29 05:03:59 -07:00
oobabooga 9a94d7b4f6 Update README 2025-05-29 05:02:52 -07:00
oobabooga 2a9699033d Update README 2025-05-29 04:55:59 -07:00
oobabooga f2ee917d4f Update README 2025-05-29 04:55:05 -07:00
oobabooga 685cfe2540 Lint 2025-05-29 04:26:43 -07:00
oobabooga a8d02dec8f Bump llama.cpp 2025-05-29 04:24:21 -07:00
Underscore 63234b9b6f
UI: Fix impersonate (#7025) 2025-05-29 08:22:03 -03:00
oobabooga 75d6cfd14d Download fetched web search results in parallel 2025-05-28 20:36:24 -07:00
oobabooga 7080a02252 Reduce the timeout for downloading web pages 2025-05-28 18:15:21 -07:00
oobabooga 3eb0b77427 Improve the web search query generation 2025-05-28 18:14:51 -07:00
oobabooga 27641ac182 UI: Make message editing work the same for user and assistant messages 2025-05-28 17:23:46 -07:00
oobabooga 6c3590ba9a Make web search attachments clickable 2025-05-28 05:28:15 -07:00
oobabooga 0aedb89921 UI: Small style improvement to attachments 2025-05-28 00:35:20 -07:00
oobabooga 75c6ae8502 UI: Don't edit messages on double click 2025-05-28 00:29:17 -07:00
oobabooga 077bbc6b10
Add web search support (#7023) 2025-05-28 04:27:28 -03:00
oobabooga 1b0e2d8750 UI: Add a token counter to the chat tab (counts input + history) 2025-05-27 22:36:24 -07:00
oobabooga f6ca0ee072 Fix regenerate sometimes not creating a new message version 2025-05-27 21:20:51 -07:00
oobabooga 2db36da979 UI: Make scrollbars more discrete in dark mode 2025-05-27 21:00:11 -07:00
Underscore 5028480eba
UI: Add footer buttons for editing messages (#7019)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2025-05-28 00:55:27 -03:00
Underscore 355b5f6c8b
UI: Add message version navigation (#6947)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2025-05-27 22:54:18 -03:00
dependabot[bot] cc9b7253c1
Update transformers requirement in /requirements/full (#7017) 2025-05-26 23:13:10 -03:00
Underscore 8531100109
Fix textbox text usage in methods (#7009) 2025-05-26 22:40:09 -03:00
djholtby 73bfc936a0
Close response generator when stopping API generation (#7014) 2025-05-26 22:39:03 -03:00
oobabooga bae1aa34aa Fix loading Llama-3_3-Nemotron-Super-49B-v1 and similar models (closes #7012) 2025-05-25 17:19:26 -07:00
oobabooga 7f6579ab20 Minor style change 2025-05-20 21:49:44 -07:00
oobabooga 0d3f854778 Improve the style of thinking blocks 2025-05-20 21:40:42 -07:00
oobabooga 8620d6ffe7 Make it possible to upload multiple text files/pdfs at once 2025-05-20 21:34:07 -07:00
oobabooga cc8a4fdcb1 Minor improvement to attachments prompt format 2025-05-20 21:31:18 -07:00
oobabooga 409a48d6bd
Add attachments support (text files, PDF documents) (#7005) 2025-05-21 00:36:20 -03:00
oobabooga 5d00574a56 Minor UI fixes 2025-05-20 16:20:49 -07:00
oobabooga 51c50b265d Update llama.cpp to b7a17463ec 2025-05-20 11:16:12 -07:00
oobabooga 616ea6966d
Store previous reply versions on regenerate (#7004) 2025-05-20 12:51:28 -03:00
Daniel Dengler c25a381540
Add a "Branch here" footer button to chat messages (#6967) 2025-05-20 11:07:40 -03:00
oobabooga 8e10f9894a
Add a metadata field to the chat history & add date/time to chat messages (#7003) 2025-05-20 10:48:46 -03:00
oobabooga 9ec46b8c44 Remove the HQQ loader (HQQ models can be loaded through Transformers) 2025-05-19 09:23:24 -07:00
oobabooga 0c7237e4b7 Update README 2025-05-18 20:01:29 -07:00
oobabooga bad1da99db Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2025-05-18 14:09:08 -07:00
oobabooga 0c1bc6d1d0 Bump llama.cpp 2025-05-18 14:08:54 -07:00
Tiago Silva 9cd6ea6c0b
Fix Dockerfile in AMD and Intel (#6995) 2025-05-18 18:07:16 -03:00
oobabooga 83bfd5c64b Fix API issues 2025-05-18 12:45:01 -07:00
oobabooga 126b3a768f Revert "Dynamic Chat Message UI Update Speed (#6952)" (for now)
This reverts commit 8137eb8ef4.
2025-05-18 12:38:36 -07:00
oobabooga 9d7a36356d Remove unnecessary js that was causing scrolling issues 2025-05-18 10:56:16 -07:00
oobabooga 2faaf18f1f Add back the "Common values" to the ctx-size slider 2025-05-18 09:06:20 -07:00
oobabooga f1ec6c8662 Minor label changes 2025-05-18 09:04:51 -07:00
oobabooga bd13a8f255 UI: Light theme improvement 2025-05-17 22:31:55 -07:00
oobabooga 076aa67963 Fix API issues 2025-05-17 22:22:18 -07:00
oobabooga 366de4b561 UI: Fix the chat area height when "Show controls" is unchecked 2025-05-17 17:11:38 -07:00
oobabooga e8595730b4
Merge pull request #6992 from oobabooga/dev
Merge dev branch
2025-05-17 11:58:46 -03:00
oobabooga 61276f6a37 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2025-05-17 07:22:51 -07:00
oobabooga 4800d1d522 More robust VRAM calculation 2025-05-17 07:20:38 -07:00
mamei16 052c82b664
Fix KeyError: 'gpu_layers' when loading existing model settings (#6991) 2025-05-17 11:19:13 -03:00
oobabooga 0f77ff9670 UI: Use total VRAM (not free) for layers calculation when a model is loaded 2025-05-16 19:19:22 -07:00
oobabooga 17c29fa0a2
Merge pull request #6987 from oobabooga/dev
Merge dev branch
2025-05-16 22:23:59 -03:00
oobabooga 4bf763e1d9 Multiple small CSS fixes 2025-05-16 18:22:43 -07:00
oobabooga c0e295dd1d Remove the 'None' option from the model menu 2025-05-16 17:53:20 -07:00
oobabooga e3bba510d4 UI: Only add a blank space to streaming messages in instruct mode 2025-05-16 17:49:17 -07:00
oobabooga 71fa046c17 Minor changes after 1c549d176b 2025-05-16 17:38:08 -07:00
oobabooga d99fb0a22a Add backward compatibility with saved n_gpu_layers values 2025-05-16 17:29:18 -07:00
oobabooga 1c549d176b Fix GPU layers slider: honor saved settings and show true maximum 2025-05-16 17:26:13 -07:00
oobabooga dc3094549e
Merge pull request #6984 from oobabooga/dev
Merge dev branch
2025-05-16 17:13:26 -03:00
oobabooga e4d3f4449d API: Fix a regression 2025-05-16 13:02:27 -07:00
oobabooga 470c822f44 API: Hide the uvicorn access logs from the terminal 2025-05-16 12:54:39 -07:00
oobabooga adb975a380 Prevent fractional gpu-layers in the UI 2025-05-16 12:52:43 -07:00
oobabooga fc483650b5 Set the maximum gpu_layers value automatically when the model is loaded with --model 2025-05-16 11:58:17 -07:00
oobabooga 38c50087fe Prevent a crash on systems without an NVIDIA GPU 2025-05-16 11:55:30 -07:00
oobabooga 253e85a519 Only compute VRAM/GPU layers for llama.cpp models 2025-05-16 10:02:30 -07:00
oobabooga 9ec9b1bf83 Auto-adjust GPU layers after model unload to utilize freed VRAM 2025-05-16 09:56:23 -07:00
oobabooga ee7b3028ac Always cache GGUF metadata calls 2025-05-16 09:12:36 -07:00
oobabooga 4925c307cf Auto-adjust GPU layers on context size and cache type changes + many fixes 2025-05-16 09:07:38 -07:00
oobabooga 93e1850a2c Only show the VRAM info for llama.cpp 2025-05-15 21:42:15 -07:00
oobabooga cbf4daf1c8 Hide the LoRA menu in portable mode 2025-05-15 21:21:54 -07:00
oobabooga fd61297933 Lint 2025-05-15 21:19:19 -07:00
oobabooga 8cb73b78e1 Update ExLlamaV3 2025-05-15 20:10:34 -07:00
oobabooga 041248cc9f Update llama.cpp 2025-05-15 20:10:02 -07:00
oobabooga 5534d01da0
Estimate the VRAM for GGUF models + autoset gpu-layers (#6980) 2025-05-16 00:07:37 -03:00
oobabooga c4a715fd1e UI: Move the LoRA menu under "Other options" 2025-05-13 20:14:09 -07:00
oobabooga 035cd3e2a9 UI: Hide the extension install menu in portable builds 2025-05-13 20:09:22 -07:00
oobabooga 2826c60044 Use logger for "Output generated in ..." messages 2025-05-13 14:45:46 -07:00
oobabooga 3fa1a899ae UI: Fix gpu-layers being ignored (closes #6973) 2025-05-13 12:07:59 -07:00
oobabooga c375b69413 API: Fix llama.cpp generating after disconnect, improve disconnect detection, fix deadlock on simultaneous requests 2025-05-13 11:23:33 -07:00
oobabooga 62c774bf24 Revert "New attempt"
This reverts commit e7ac06c169.
2025-05-13 06:42:25 -07:00
oobabooga e7ac06c169 New attempt 2025-05-10 19:20:04 -07:00
oobabooga 0c5fa3728e Revert "Fix API failing to cancel streams (attempt), closes #6966"
This reverts commit 006a866079.
2025-05-10 19:12:40 -07:00
oobabooga 006a866079 Fix API failing to cancel streams (attempt), closes #6966 2025-05-10 17:55:48 -07:00
oobabooga 47d4758509 Fix #6970 2025-05-10 17:46:00 -07:00
oobabooga 4920981b14 UI: Remove the typing cursor 2025-05-09 20:35:38 -07:00
oobabooga 8984e95c67 UI: More friendly message when no model is loaded 2025-05-09 07:21:05 -07:00
oobabooga 2bde625d57 Update README 2025-05-09 00:19:25 -07:00
oobabooga 512bc2d0e0 UI: Update some labels 2025-05-08 23:43:55 -07:00
oobabooga f8ef6e09af UI: Make ctx-size a slider 2025-05-08 18:19:04 -07:00
oobabooga bf7e4a4597 Docs: Add a tool/function calling example (from https://github.com/oobabooga/text-generation-webui/pull/6827#issuecomment-2854716960) 2025-05-08 16:12:07 -07:00
oobabooga 9ea2a69210 llama.cpp: Add --no-webui to the llama-server command 2025-05-08 10:41:25 -07:00
oobabooga 3bc2ec2b11 Fix #6965 2025-05-08 10:34:09 -07:00
oobabooga 1c7209a725 Save the chat history periodically during streaming 2025-05-08 09:46:43 -07:00
oobabooga a1b3307b66 Bump llama.cpp 2025-05-08 08:58:43 -07:00
Jonas fa960496d5
Tools support for OpenAI compatible API (#6827) 2025-05-08 12:30:27 -03:00
Scott Z ed6e16191d
Docker fix for NVIDIA (#6964) 2025-05-08 12:21:52 -03:00
oobabooga 13a434f351 Bump exllamav3 2025-05-08 08:06:07 -07:00
oobabooga a2ab42d390 UI: Remove the exllamav2 info message 2025-05-08 08:00:38 -07:00
oobabooga 348d4860c2 UI: Create a "Main options" section in the Model tab 2025-05-08 07:58:59 -07:00
oobabooga d2bae7694c UI: Change the ctx-size description 2025-05-08 07:26:23 -07:00
oobabooga b28fa86db6 Default --gpu-layers to 256 2025-05-06 17:51:55 -07:00
oobabooga 760b4dd115 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2025-05-06 14:02:57 -07:00
oobabooga e4fb2475d2 UI: Multiple small style improvements (light/dark themes) 2025-05-06 14:02:15 -07:00
Downtown-Case 5ef564a22e
Fix model config loading in shared.py for Python 3.13 (#6961) 2025-05-06 17:03:33 -03:00
oobabooga c4f36db0d8 llama.cpp: remove tfs (it doesn't get used) 2025-05-06 08:41:13 -07:00
oobabooga 05115e42ee Set top_n_sigma before temperature by default 2025-05-06 08:27:21 -07:00
oobabooga 1927afe894 Fix top_n_sigma not showing for llama.cpp 2025-05-06 08:18:49 -07:00
oobabooga 605cc9ab14 Update exllamav3 2025-05-06 06:43:35 -07:00
oobabooga 89590adc14 Update llama.cpp 2025-05-06 06:41:17 -07:00
oobabooga d1c0154d66 llama.cpp: Add top_n_sigma, fix typical_p in sampler priority 2025-05-06 06:38:39 -07:00
oobabooga cbef35054c UI: CSS fix 2025-05-05 17:46:09 -07:00
Evgenii Novikov 4e8f628d3c
docker: App uid typo in other docker composes (#6958) 2025-05-05 20:05:15 -03:00
oobabooga 530223bf0b UI: Fix the hover menu colors 2025-05-05 16:03:43 -07:00
oobabooga 76f947e3cf UI: Minor style change 2025-05-05 15:58:29 -07:00
Alireza Ghasemi 99bd66445f
SuperboogaV2: minor update to avoid json serialization errors #6945 2025-05-05 19:04:06 -03:00
Evgenii Novikov 987505ead3
docker: Fix app uid typo in cpu docker compose (#6957) 2025-05-05 19:03:33 -03:00
oobabooga 941e0663da Update README 2025-05-05 14:18:16 -07:00
oobabooga f82667f0b4 Remove more multimodal extension references 2025-05-05 14:17:00 -07:00
oobabooga 85bf2e15b9 API: Remove obsolete multimodal extension handling
Multimodal support will be added back once it's implemented in llama-server.
2025-05-05 14:14:48 -07:00
mamei16 8137eb8ef4
Dynamic Chat Message UI Update Speed (#6952) 2025-05-05 18:05:23 -03:00
oobabooga 53d8e46502 Ensure environment isolation in portable installs 2025-05-05 12:28:17 -07:00
oobabooga bf5290bc0f Fix the hover menu in light theme 2025-05-05 08:04:12 -07:00
oobabooga 967b70327e Light theme improvement 2025-05-05 07:59:02 -07:00
oobabooga 6001d279c6 Light theme improvement 2025-05-05 07:42:13 -07:00
oobabooga 475e012ee8 UI: Improve the light theme colors 2025-05-05 06:16:29 -07:00
oobabooga b817bb33fd Minor fix after df7bb0db1f 2025-05-05 05:00:20 -07:00
oobabooga f3da45f65d ExLlamaV3_HF: Change max_chunk_size to 256 2025-05-04 20:37:15 -07:00
oobabooga df7bb0db1f Rename --n-gpu-layers to --gpu-layers 2025-05-04 20:03:55 -07:00
oobabooga d0211afb3c Save the chat history right after sending a message 2025-05-04 18:52:01 -07:00
oobabooga 2da197bba4 Refinement after previous commit 2025-05-04 18:29:05 -07:00
oobabooga 690d693913 UI: Add padding to only show the last message/reply after sending a message
To avoid scrolling
2025-05-04 18:13:29 -07:00
oobabooga d9da16edba UI: Remove the chat input textarea border 2025-05-04 16:53:52 -07:00
oobabooga 84ab1f95be UI: Increase the chat area a bit 2025-05-04 15:21:52 -07:00
oobabooga d186621926 UI: Fixes after previous commit 2025-05-04 15:19:46 -07:00
oobabooga 7853fb1c8d
Optimize the Chat tab (#6948) 2025-05-04 18:58:37 -03:00
oobabooga b7a5c7db8d llama.cpp: Handle short arguments in --extra-flags 2025-05-04 07:14:42 -07:00
oobabooga 5f5569e9ac Update README 2025-05-04 06:20:36 -07:00
oobabooga 4c2e3b168b llama.cpp: Add a retry mechanism when getting the logits (sometimes it fails) 2025-05-03 06:51:20 -07:00
oobabooga ea60f14674 UI: Show the list of files if the user tries to download a GGUF repository 2025-05-03 06:06:50 -07:00
oobabooga b71ef50e9d UI: Add a min-height to prevent constant scrolling during chat streaming 2025-05-02 23:45:58 -07:00
oobabooga b21bd8bb1e UI: Invert user/assistant message colors in instruct mode
The goal is to make assistant messages more readable.
2025-05-02 22:43:33 -07:00
oobabooga d08acb4af9 UI: Rename enable_thinking -> Enable thinking 2025-05-02 20:50:52 -07:00
oobabooga 3526b7923c Remove extensions with requirements from portable builds 2025-05-02 17:40:53 -07:00
oobabooga 4cea720da8 UI: Remove the "Autoload the model" feature 2025-05-02 16:38:28 -07:00
oobabooga 905afced1c Add a --portable flag to hide things in portable mode 2025-05-02 16:34:29 -07:00
oobabooga 3f26b0408b Fix after 9e3867dc83 2025-05-02 16:17:22 -07:00
oobabooga 9e3867dc83 llama.cpp: Fix manual random seeds 2025-05-02 09:36:15 -07:00
oobabooga d5c407cf35 Use Vulkan instead of ROCm for llama.cpp on AMD 2025-05-01 20:05:36 -07:00
oobabooga f8aaf3c23a Use ROCm 6.2.4 on AMD 2025-05-01 19:50:46 -07:00
oobabooga c12a53c998 Use turboderp's exllamav2 wheels 2025-05-01 19:46:56 -07:00
oobabooga ace8afb825
Merge dev branch 2025-05-01 12:25:04 -03:00
oobabooga 89090d9a61 Update README 2025-05-01 08:22:54 -07:00
oobabooga a41da1ec95
Merge pull request #6939 from oobabooga/dev
Merge dev branch
2025-05-01 00:15:11 -03:00
oobabooga b950a0c6db Lint 2025-04-30 20:02:10 -07:00
oobabooga 307d13b540 UI: Minor label change 2025-04-30 18:58:14 -07:00
oobabooga 55283bb8f1 Fix CFG with ExLlamaV2_HF (closes #6937) 2025-04-30 18:43:45 -07:00
oobabooga ec2e641749 Update settings-template.yaml 2025-04-30 15:25:26 -07:00
oobabooga a6c3ec2299 llama.cpp: Explicitly send cache_prompt = True 2025-04-30 15:24:07 -07:00
oobabooga 195a45c6e1 UI: Make thinking blocks closed by default 2025-04-30 15:12:46 -07:00
oobabooga cd5c32dc19 UI: Fix max_updates_second not working 2025-04-30 14:54:05 -07:00
oobabooga b46ca01340 UI: Set max_updates_second to 12 by default
When the tokens/second at at ~50 and the model is a thinking model,
the markdown rendering for the streaming message becomes a CPU
bottleneck.
2025-04-30 14:53:15 -07:00
oobabooga a4bf339724 Bump llama.cpp 2025-04-30 11:13:14 -07:00
oobabooga e9569c3984 Fixes after c5fe92d152 2025-04-30 06:57:23 -07:00
oobabooga 771d3d8ed6 Fix getting the llama.cpp logprobs for Qwen3-30B-A3B 2025-04-30 06:48:32 -07:00
oobabooga 7f49e3c3ce Bump ExLlamaV3 2025-04-30 05:25:09 -07:00
oobabooga c5fe92d152 Bump llama.cpp 2025-04-30 05:24:58 -07:00
oobabooga 1dd4aedbe1 Fix the streaming_llm UI checkbox not being interactive 2025-04-29 05:28:46 -07:00
oobabooga c5fb51e5d1 Update README 2025-04-28 22:40:26 -07:00
oobabooga d10bded7f8 UI: Add an enable_thinking option to enable/disable Qwen3 thinking 2025-04-28 22:37:01 -07:00
oobabooga 1ee0acc852 llama.cpp: Make --verbose print the llama-server command 2025-04-28 15:56:25 -07:00
oobabooga 15a29e99f8 Lint 2025-04-27 21:41:34 -07:00
oobabooga be13f5199b UI: Add an info message about how to use Speculative Decoding 2025-04-27 21:40:38 -07:00
oobabooga c6c2855c80 llama.cpp: Remove the timeout while loading models (closes #6907) 2025-04-27 21:22:21 -07:00
oobabooga bbcaec75b4 API: Find a new port if the default one is taken (closes #6918) 2025-04-27 21:13:16 -07:00
oobabooga ee0592473c Fix ExLlamaV3_HF leaking memory (attempt) 2025-04-27 21:04:02 -07:00
oobabooga 6e6f9971a2
Merge pull request #6919 from oobabooga/dev
Merge dev branch
2025-04-27 11:35:19 -03:00
oobabooga 965ca7948f Update README 2025-04-27 07:33:08 -07:00
oobabooga 1180bb0d80
Merge pull request #6913 from oobabooga/dev
Merge dev branch
2025-04-27 00:12:16 -03:00
oobabooga f5b59d2b0b Fix the vulkan workflow 2025-04-26 20:11:24 -07:00
oobabooga 9bb9ce079e
Merge pull request #6912 from oobabooga/dev
Merge dev branch
2025-04-27 00:03:16 -03:00
oobabooga 765fea5e36 UI: minor style change 2025-04-26 19:33:46 -07:00
oobabooga 70952553c7 Lint 2025-04-26 19:29:08 -07:00
oobabooga 363b632a0d Lint 2025-04-26 19:22:36 -07:00
oobabooga fa861de05b Fix portable builds with Python 3.12 2025-04-26 18:52:44 -07:00
oobabooga 7b80acd524 Fix parsing --extra-flags 2025-04-26 18:40:03 -07:00
oobabooga 943451284f Fix the Notebook tab not loading its default prompt 2025-04-26 18:25:06 -07:00
oobabooga 511eb6aa94 Fix saving settings to settings.yaml 2025-04-26 18:20:00 -07:00
oobabooga 8b83e6f843 Prevent Gradio from saying 'Thank you for being a Gradio user!' 2025-04-26 18:14:57 -07:00
oobabooga 4a32e1f80c UI: show draft_max for ExLlamaV2 2025-04-26 18:01:44 -07:00
oobabooga 0fe3b033d0 Fix parsing of --n_ctx and --max_seq_len (2nd attempt) 2025-04-26 17:52:21 -07:00
oobabooga c4afc0421d Fix parsing of --n_ctx and --max_seq_len 2025-04-26 17:43:53 -07:00
oobabooga 234aba1c50 llama.cpp: Simplify the prompt processing progress indicator
The progress bar was unreliable
2025-04-26 17:33:47 -07:00
oobabooga 4ff91b6588 Better default settings for Speculative Decoding 2025-04-26 17:24:40 -07:00
oobabooga bf2aa19b21 Bump llama.cpp 2025-04-26 16:39:22 -07:00
oobabooga 029aab6404 Revert "Add -noavx2 portable builds"
This reverts commit 0dd71e78c9.
2025-04-26 16:38:13 -07:00
oobabooga 35717a088c API: Add an /v1/internal/health endpoint 2025-04-26 15:42:27 -07:00
oobabooga bc55feaf3e Improve host header validation in local mode 2025-04-26 15:42:17 -07:00
oobabooga a317450dfa Update README 2025-04-26 14:59:29 -07:00
oobabooga d1e7d9c5d5 Update CMD_FLAGS.txt 2025-04-26 09:00:56 -07:00
oobabooga 3a207e7a57 Improve the --help formatting a bit 2025-04-26 07:31:04 -07:00
oobabooga 6acb0e1bee Change a UI description 2025-04-26 05:13:08 -07:00
oobabooga cbd4d967cc Update a --help message 2025-04-26 05:09:52 -07:00
oobabooga 19c8dced67 Move settings-template.yaml into user_data 2025-04-26 05:03:23 -07:00
oobabooga b976112539 Remove the WSL installation scripts
They were useful in 2023 but now everything runs natively on Windows.
2025-04-26 05:02:17 -07:00
oobabooga 763a7011c0 Remove an ancient/obsolete migration check 2025-04-26 04:59:05 -07:00
oobabooga d9de14d1f7
Restructure the repository (#6904) 2025-04-26 08:56:54 -03:00
oobabooga d4017fbb6d
ExLlamaV3: Add kv cache quantization (#6903) 2025-04-25 21:32:00 -03:00
oobabooga d4b1e31c49 Use --ctx-size to specify the context size for all loaders
Old flags are still recognized as alternatives.
2025-04-25 16:59:03 -07:00
oobabooga faababc4ea llama.cpp: Add a prompt processing progress bar 2025-04-25 16:42:30 -07:00
oobabooga 877cf44c08 llama.cpp: Add StreamingLLM (--streaming-llm) 2025-04-25 16:21:41 -07:00
oobabooga d35818f4e1
UI: Add a collapsible thinking block to messages with <think> steps (#6902) 2025-04-25 18:02:02 -03:00
oobabooga 0dd71e78c9 Add -noavx2 portable builds 2025-04-25 09:07:14 -07:00
oobabooga 98f4c694b9 llama.cpp: Add --extra-flags parameter for passing additional flags to llama-server 2025-04-25 07:32:51 -07:00
oobabooga b6fffbd216 UI: minor style change 2025-04-25 05:37:44 -07:00
oobabooga 2c7ff86015 Bump exllamav3 to de83084184 2025-04-25 05:28:22 -07:00
oobabooga 5993ebeb1b Bump exllamav2 to 0.2.9 2025-04-25 05:27:59 -07:00
oobabooga 23399aff3c UI: minor style change 2025-04-24 20:39:00 -07:00
oobabooga 5861013e68 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2025-04-24 20:36:20 -07:00
oobabooga a90df27ff5 UI: Add a greeting when the chat history is empty 2025-04-24 20:33:40 -07:00
oobabooga ae1fe87365
ExLlamaV2: Add speculative decoding (#6899) 2025-04-25 00:11:04 -03:00
Matthew Jenkins 8f2493cc60
Prevent llamacpp defaults from locking up consumer hardware (#6870) 2025-04-24 23:38:57 -03:00
oobabooga 370fe7b7cf Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2025-04-24 09:33:17 -07:00
oobabooga 8ebe868916 Fix typos in b313adf653 2025-04-24 09:32:17 -07:00
oobabooga 93fd4ad25d llama.cpp: Document the --device-draft syntax 2025-04-24 09:20:11 -07:00
oobabooga f1b64df8dd EXL2: add another torch.cuda.synchronize() call to prevent errors 2025-04-24 09:03:49 -07:00
Ziya 60ac495d59
extensions/superboogav2: existing embedding check bug fix (#6898) 2025-04-24 12:42:05 -03:00
oobabooga b313adf653 Bump llama.cpp, make the wheels work with any Python >= 3.7 2025-04-24 08:26:12 -07:00
oobabooga c71a2af5ab Handle CMD_FLAGS.txt in the main code (closes #6896) 2025-04-24 08:21:06 -07:00
oobabooga bfbde73409 Make 'instruct' the default chat mode 2025-04-24 07:08:49 -07:00
oobabooga e99c20bcb0
llama.cpp: Add speculative decoding (#6891) 2025-04-23 20:10:16 -03:00
oobabooga 9424ba17c8 UI: show only part 00001 of multipart GGUF models in the model menu 2025-04-22 19:56:42 -07:00
oobabooga 1aa76b3beb
Merge pull request #6885 from oobabooga/dev
Merge dev branch
2025-04-22 22:38:24 -03:00
oobabooga bce1b68ca9 Minor fix after previous commit 2025-04-22 18:37:36 -07:00
oobabooga 812d878812 Make the dependabot less spammy 2025-04-22 18:35:22 -07:00
oobabooga 1df2b0d3ae
Merge pull request #6884 from oobabooga/dev
Merge dev branch
2025-04-22 22:02:30 -03:00
oobabooga 8228822a6c Revert "Temporary change"
This reverts commit 765de6f678.
2025-04-22 18:01:47 -07:00
oobabooga 62455b415c
Merge pull request #6883 from oobabooga/dev
Merge dev branch
2025-04-22 21:54:34 -03:00
oobabooga 765de6f678 Temporary change 2025-04-22 17:53:56 -07:00
oobabooga 89ec4c9ba6 Add vulkan workflow 2025-04-22 17:51:08 -07:00
oobabooga 06619e5f03 Add vulkan requirements.txt files 2025-04-22 17:46:54 -07:00
oobabooga 022664f2bd
Merge pull request #6881 from oobabooga/dev
Merge dev branch
2025-04-22 12:15:34 -03:00
oobabooga 4335a24ff8 Fix the workflow 2025-04-22 08:14:13 -07:00
oobabooga a778270536
Merge pull request #6869 from oobabooga/dev
Merge dev branch
2025-04-22 12:09:20 -03:00
oobabooga 25cf3600aa Lint 2025-04-22 08:04:02 -07:00
oobabooga 39cbb5fee0 Lint 2025-04-22 08:03:25 -07:00
oobabooga da1919baae Update the README 2025-04-22 08:03:22 -07:00
oobabooga a3031795a3 Update the zip filename 2025-04-22 08:03:16 -07:00
oobabooga 008c6dd682 Lint 2025-04-22 08:02:37 -07:00
oobabooga ee09e44c85
Portable version (#6868) 2025-04-22 09:25:57 -03:00
oobabooga 78aeabca89 Fix the transformers loader 2025-04-21 18:33:14 -07:00
oobabooga 8320190184 Fix the exllamav2_HF and exllamav3_HF loaders 2025-04-21 18:32:23 -07:00
oobabooga 15989c2ed8 Make llama.cpp the default loader 2025-04-21 16:36:35 -07:00
oobabooga 86c3ed3218 Small change to the unload_model() function 2025-04-20 20:00:56 -07:00
oobabooga c178ea02fe Revert "Move the requirements*.txt to a requirements folder"
This reverts commit 6117ef7d64.
2025-04-20 19:27:38 -07:00
oobabooga 6117ef7d64 Move the requirements*.txt to a requirements folder 2025-04-20 19:12:04 -07:00
oobabooga fe8e80e04a Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2025-04-20 19:09:27 -07:00
oobabooga ff1c00bdd9 llama.cpp: set the random seed manually 2025-04-20 19:08:44 -07:00
Matthew Jenkins d3e7c655e5
Add support for llama-cpp builds from https://github.com/ggml-org/llama.cpp (#6862) 2025-04-20 23:06:24 -03:00
oobabooga 99588be576 Organize one_click.py 2025-04-20 18:57:26 -07:00
oobabooga e243424ba1 Fix an import 2025-04-20 17:51:28 -07:00
oobabooga 8cfd7f976b Revert "Remove the old --model-menu flag"
This reverts commit 109de34e3b.
2025-04-20 13:35:42 -07:00
oobabooga d5e1bccef9 Remove the SpeechRecognition requirement 2025-04-20 11:47:28 -07:00
oobabooga b3bf7a885d Fix ExLlamaV2_HF and ExLlamaV3_HF after ae02ffc605 2025-04-20 11:32:48 -07:00
oobabooga 9c59acf820 Remove the numba requirement (it's no longer used) 2025-04-20 10:02:40 -07:00
oobabooga ae02ffc605
Refactor the transformers loader (#6859) 2025-04-20 13:33:47 -03:00
oobabooga c19b995b8e
Merge pull request #6857 from oobabooga/dev
Merge dev branch
2025-04-19 21:45:55 -03:00
oobabooga 6ba0164c70 Lint 2025-04-19 17:45:21 -07:00
oobabooga 5ab069786b llama.cpp: add back the two encode calls (they are harmless now) 2025-04-19 17:38:36 -07:00
oobabooga b9da5c7e3a Use 127.0.0.1 instead of localhost for faster llama.cpp on Windows 2025-04-19 17:36:04 -07:00
oobabooga 9c9df2063f llama.cpp: fix unicode decoding (closes #6856) 2025-04-19 16:38:15 -07:00
oobabooga ba976d1390 llama.cpp: avoid two 'encode' calls 2025-04-19 16:35:01 -07:00
oobabooga ed42154c78 Revert "llama.cpp: close the connection immediately on 'Stop'"
This reverts commit 5fdebc554b.
2025-04-19 05:32:36 -07:00
oobabooga 5fdebc554b llama.cpp: close the connection immediately on 'Stop' 2025-04-19 04:59:24 -07:00
oobabooga b1495d52e5
Merge pull request #6855 from oobabooga/dev
Merge dev branch
2025-04-19 01:53:11 -03:00
oobabooga 6589ebeca8 Revert "llama.cpp: new optimization attempt"
This reverts commit e2e73ed22f.
2025-04-18 21:16:21 -07:00
oobabooga e2e73ed22f llama.cpp: new optimization attempt 2025-04-18 21:05:08 -07:00
oobabooga e2e90af6cd llama.cpp: don't include --rope-freq-base in the launch command if null 2025-04-18 20:51:18 -07:00
oobabooga 44a6d8a761
Merge pull request #6854 from oobabooga/dev
Merge dev branch
2025-04-18 23:41:56 -03:00
oobabooga 9f07a1f5d7 llama.cpp: new attempt at optimizing the llama-server connection 2025-04-18 19:30:53 -07:00
oobabooga f727b4a2cc llama.cpp: close the connection properly when generation is cancelled 2025-04-18 19:01:39 -07:00
oobabooga b3342b8dd8 llama.cpp: optimize the llama-server connection 2025-04-18 18:46:36 -07:00
oobabooga 4fa52a1302
Merge pull request #6852 from oobabooga/dev
Merge dev branch
2025-04-18 22:15:40 -03:00
oobabooga 2002590536 Revert "Attempt at making the llama-server streaming more efficient."
This reverts commit 5ad080ff25.
2025-04-18 18:13:54 -07:00
oobabooga 71ae05e0a4 llama.cpp: Fix the sampler priority handling 2025-04-18 18:06:36 -07:00
oobabooga 5ad080ff25 Attempt at making the llama-server streaming more efficient. 2025-04-18 18:04:49 -07:00
oobabooga 4fabd729c9 Fix the API without streaming or without 'sampler_priority' (closes #6851) 2025-04-18 17:25:22 -07:00
oobabooga 5135523429 Fix the new llama.cpp loader failing to unload models 2025-04-18 17:10:26 -07:00
oobabooga 4eecb6611f
Merge pull request #6850 from oobabooga/dev
Merge dev branch
2025-04-18 15:33:32 -03:00
oobabooga 8d481ef9d5 Update README 2025-04-18 11:31:22 -07:00
oobabooga caa6afc88b Only show 'GENERATE_PARAMS=...' in the logits endpoint if use_logits is True 2025-04-18 09:57:57 -07:00
oobabooga c5e54c0b37
Merge pull request #6848 from oobabooga/dev
Merge dev branch
2025-04-18 13:36:06 -03:00
oobabooga e52f62d3ff Update README 2025-04-18 09:29:57 -07:00
oobabooga 85c4486d4a Update the colab notebook 2025-04-18 08:53:44 -07:00
oobabooga d00d713ace Rename get_max_context_length to get_vocabulary_size in the new llama.cpp loader 2025-04-18 08:14:15 -07:00
oobabooga c1cc65e82e Lint 2025-04-18 08:06:51 -07:00
oobabooga d68f0fbdf7 Remove obsolete references to llamacpp_HF 2025-04-18 07:46:04 -07:00
oobabooga a0abf93425 Connect --rope-freq-base to the new llama.cpp loader 2025-04-18 06:53:51 -07:00
oobabooga ef9910c767 Fix a bug after c6901aba9f 2025-04-18 06:51:28 -07:00
oobabooga 1c4a2c9a71 Make exllamav3 safer as well 2025-04-18 06:17:58 -07:00
oobabooga 03544d4fb6 Bump llama.cpp and exllamav3 to the latest commits 2025-04-18 06:14:13 -07:00
oobabooga c6901aba9f Remove deprecation warning code 2025-04-18 06:05:47 -07:00
oobabooga 170ad3d3ec Update the README 2025-04-18 06:03:35 -07:00
oobabooga 8144e1031e Remove deprecated command-line flags 2025-04-18 06:02:28 -07:00
oobabooga ae54d8faaa
New llama.cpp loader (#6846) 2025-04-18 09:59:37 -03:00
oobabooga 5c2f8d828e Fix exllamav2 generating eos randomly after previous fix 2025-04-18 05:42:38 -07:00
oobabooga 2fc58ad935 Consider files with .pt extension in the new model menu function 2025-04-17 23:10:43 -07:00
Googolplexed d78abe480b
Allow for model subfolder organization for GGUF files (#6686)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2025-04-18 02:53:59 -03:00
oobabooga ce9e2d94b1 Revert "Attempt at solving the ExLlamaV2 issue"
This reverts commit c9b3c9dfbf.
2025-04-17 22:03:21 -07:00
oobabooga 5dfab7d363 New attempt at solving the exl2 issue 2025-04-17 22:03:11 -07:00
oobabooga c9b3c9dfbf Attempt at solving the ExLlamaV2 issue 2025-04-17 21:45:15 -07:00
oobabooga 2c2d453c8c Revert "Use ExLlamaV2 (instead of the HF one) for EXL2 models for now"
This reverts commit 0ef1b8f8b4.
2025-04-17 21:31:32 -07:00
oobabooga 0ef1b8f8b4 Use ExLlamaV2 (instead of the HF one) for EXL2 models for now
It doesn't seem to have the "OverflowError" bug
2025-04-17 05:47:40 -07:00
oobabooga 38dc09dca5 Bump exllamav3 to the latest commit 2025-04-15 09:50:36 -07:00
oobabooga 038a012581 Installer: Remove .installer_state.json on reinstalling 2025-04-11 21:12:32 -07:00
oobabooga 682c78ea42 Add back detection of GPTQ models (closes #6841) 2025-04-11 21:00:42 -07:00
oobabooga 454366f93e Change the ExLlamaV3 wheel version to 0.0.1a1 2025-04-10 18:33:29 -07:00
oobabooga d7b336d37e Update the README 2025-04-09 20:12:14 -07:00
oobabooga 4ed0da74a8 Remove the obsolete 'multimodal' extension 2025-04-09 20:09:48 -07:00
oobabooga 598568b1ed Revert "UI: remove the streaming cursor"
This reverts commit 6ea0206207.
2025-04-09 16:03:14 -07:00
oobabooga 297a406e05 UI: smoother chat streaming
This removes the throttling associated to gr.Textbox that made words appears in chunks rather than one at a time
2025-04-09 16:02:37 -07:00
oobabooga 6ea0206207 UI: remove the streaming cursor 2025-04-09 14:59:34 -07:00
oobabooga 14e6baeb48
Merge pull request #6838 from oobabooga/dev
Merge dev branch
2025-04-09 14:48:37 -03:00
oobabooga 9025848df5 Small change to installer 2025-04-09 10:25:47 -07:00
oobabooga d337ea31fa Revert "Reapply "Update transformers requirement from ==4.50.* to ==4.51.* (#6834)""
This reverts commit 8229736ec4.
2025-04-09 10:16:47 -07:00
oobabooga 8229736ec4 Reapply "Update transformers requirement from ==4.50.* to ==4.51.* (#6834)"
This reverts commit 0b3503c91f.
2025-04-09 08:38:06 -07:00
oobabooga 89f40cdcf7 Update libstdcxx-ng for GLIBCXX_3.4.30 support on Linux 2025-04-09 08:28:44 -07:00
oobabooga ad1ada6574 Change one message in the installer 2025-04-09 05:17:10 -07:00
oobabooga d8aad6da94 Fix an update bug 2025-04-08 20:20:24 -07:00
oobabooga 8b8d39ec4e
Add ExLlamaV3 support (#6832) 2025-04-09 00:07:08 -03:00
oobabooga 0b3503c91f Revert "Update transformers requirement from ==4.50.* to ==4.51.* (#6834)"
This reverts commit f1f32386b4.
2025-04-08 12:26:03 -07:00
oobabooga 649ee729c1 Remove Python 3.10 support 2025-04-08 09:22:06 -07:00
oobabooga bf48ec8c44 Remove an unnecessary UI message 2025-04-07 17:43:41 -07:00
oobabooga a5855c345c
Set context lengths to at most 8192 by default (to prevent out of memory errors) (#6835) 2025-04-07 21:42:33 -03:00
dependabot[bot] f1f32386b4
Update transformers requirement from ==4.50.* to ==4.51.* (#6834) 2025-04-07 19:29:39 -03:00
oobabooga 204db28362 Update the dockerfiles 2025-04-06 18:48:31 -07:00
oobabooga eef90a4964 Update some intel arc installation commands 2025-04-06 17:44:07 -07:00
oobabooga a8a64b6c1c Update the README 2025-04-06 17:40:18 -07:00
oobabooga c010cea7be Remove CUDA 11.8 support 2025-04-06 17:17:25 -07:00
Shixian Sheng cbffcf67ef
Fix links in the ngrok extension README (#6826) 2025-04-02 14:28:29 -03:00
dependabot[bot] 77a73cc561
Update peft requirement from ==0.12.* to ==0.15.* (#6820) 2025-03-31 21:01:27 -03:00
oobabooga 109de34e3b Remove the old --model-menu flag 2025-03-31 09:24:03 -07:00
oobabooga bb1905ebc5 Fix the colab notebook 2025-03-29 19:17:36 -07:00
oobabooga 1981327285 Fix the colab notebook 2025-03-29 19:17:14 -07:00
oobabooga 79a26d7a5c Lint 2025-03-29 18:49:48 -07:00
oobabooga 1bd208c219
Add a new chat style: Dark (#6817) 2025-03-29 22:47:10 -03:00
oobabooga 9b80d1d6c2 Remove the stalebot 2025-03-29 13:44:37 -07:00
oobabooga 525b1e0207 Remove the stalebot 2025-03-29 13:43:16 -07:00
dependabot[bot] 2bfaf44df0
Update accelerate requirement from ==1.4.* to ==1.5.* (#6802) 2025-03-26 10:03:21 -03:00
oobabooga 01e42a00ff Bump transformers to 4.50 2025-03-26 06:01:57 -07:00
oobabooga 80cdbe4e09
Merge pull request #6797 from oobabooga/dev
Merge dev branch
2025-03-15 00:11:25 -03:00
oobabooga 758c3f15a5 Lint 2025-03-14 20:04:43 -07:00
SeanScripts 60d67994d9
Perplexity colors extension updates (#6764) 2025-03-14 16:45:53 -03:00
oobabooga 5bcd2d7ad0
Add the top N-sigma sampler (#6796) 2025-03-14 16:45:11 -03:00
oobabooga 677d74a6a0 Revert "UI: improved scrollbar styles", add just a small change instead 2025-03-14 12:10:48 -07:00
oobabooga 6ab04698f6 UI: improve the light mode left sidebar color 2025-03-14 12:03:49 -07:00
oobabooga 26317a4c7e Fix jinja2 error while loading c4ai-command-a-03-2025 2025-03-14 10:59:05 -07:00
oobabooga f04a37adc2 UI: improved scrollbar styles 2025-03-14 05:20:15 -07:00
oobabooga 0261338910 Bump llama-cpp-python to 0.3.8 2025-03-12 17:55:25 -07:00
oobabooga 39fded487a Bump ExllamaV2 to 0.2.8 2025-03-12 17:54:30 -07:00
dependabot[bot] a12e05d9c0
Bump jinja2 from 3.1.5 to 3.1.6 (#6786) 2025-03-12 16:11:03 -03:00
Kelvie Wong 769eee1ff3 Fix OpenAI API with new param (show_after), closes #6747 (#6749)
---------

Co-authored-by: oobabooga <oobabooga4@gmail.com>
2025-02-18 07:02:19 -08:00
Kelvie Wong 16fa9215c4
Fix OpenAI API with new param (show_after), closes #6747 (#6749)
---------

Co-authored-by: oobabooga <oobabooga4@gmail.com>
2025-02-18 12:01:30 -03:00
SeanScripts b131f86584
Perplexity colors extension v2 (#6756) 2025-02-18 11:56:28 -03:00
Alireza Ghasemi 01f20d2d9f
Improve SuperboogaV2 with Date/Time Embeddings, GPU Support, and Multiple File Formats (#6748) 2025-02-17 22:38:15 -03:00
dependabot[bot] 12f6f7ba9f
Update accelerate requirement from ==1.3.* to ==1.4.* (#6753) 2025-02-17 22:35:38 -03:00
oobabooga dba17c40fc Make transformers 4.49 functional 2025-02-17 17:31:11 -08:00
oobabooga 16f4f1a1c3 Bump transformers to 4.49 2025-02-17 17:20:10 -08:00
oobabooga 7c883ef2f0
Merge pull request #6746 from oobabooga/dev
Merge dev branch
2025-02-14 23:25:31 -03:00
oobabooga cf9676c4d5 Update README 2025-02-14 18:05:36 -08:00
Manuel Schmid b54bf359bf
sd_api_pictures model reload fix (#6720) 2025-02-03 00:11:49 -03:00
oobabooga edbe0af647 Minor fixes after 0360f54ae8 2025-02-02 17:04:56 -08:00
oobabooga 6724d2bfa4 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2025-02-02 16:59:03 -08:00
oobabooga 44e569c3a2 Remove obsolete convert-to-safetensors.py from the repository 2025-02-02 16:15:33 -08:00
SamAcctX f28f39792d
update deprecated deepspeed import for transformers 4.46+ (#6725) 2025-02-02 20:41:36 -03:00
oobabooga f074ffc31b UI: minor light theme improvement 2025-02-02 15:39:36 -08:00
oobabooga c6f2c2fd7e UI: style improvements 2025-02-02 15:34:03 -08:00
oobabooga 0360f54ae8 UI: add a "Show after" parameter (to use with DeepSeek </think>) 2025-02-02 15:30:09 -08:00
oobabooga 01c46f8b56 Merge branch 'main' into dev 2025-01-30 09:49:30 -08:00
oobabooga 32cdaa540f Update README 2025-01-30 09:49:25 -08:00
oobabooga 461d1fdb76 Update README 2025-01-30 09:48:52 -08:00
SpyTech Labs fea98f82c5
DOCS FIX: WSL Port Forwarding Loop. (#6519) 2025-01-30 14:34:23 -03:00
oobabooga 9ac4d81c8b
Merge pull request #6713 from oobabooga/dev
Merge dev branch
2025-01-29 19:12:56 -03:00
oobabooga b614ea6596 Installer: small fixes 2025-01-29 14:05:39 -08:00
oobabooga f01cc079b9 Lint 2025-01-29 14:00:59 -08:00
oobabooga b7c17727b0 Update .gitignore 2025-01-29 13:57:56 -08:00
oobabooga 9ddcc91a91 Bump llama-cpp-python to 0.3.7 2025-01-29 13:56:46 -08:00
oobabooga e3fd4a0ea7 Merge branch 'main' into dev 2025-01-28 12:54:57 -08:00
oobabooga a1c353a4b3 Update README 2025-01-28 12:54:25 -08:00
oobabooga 3936589755 Update README 2025-01-28 12:53:55 -08:00
oobabooga 0b9ab1438d Clean up 2025-01-27 10:28:59 -08:00
oobabooga bac652bb1d Another fix 2025-01-27 10:25:26 -08:00
oobabooga 340022d4b0 Fix after previous commit 2025-01-27 10:02:21 -08:00
oobabooga 053911b629 Installer: don't ignore .whl requirements if the commit has changed
By the user manually switching branches or calling git pull.
2025-01-27 09:24:44 -08:00
oobabooga 1c9dfa871b Revert "Installer: change a message"
This reverts commit c49251e95d.
2025-01-26 18:17:31 -08:00
oobabooga 87de91dd65 Docs: fix an API example 2025-01-25 18:29:11 -08:00
oobabooga c49251e95d Installer: change a message 2025-01-25 15:03:09 -08:00
oobabooga 75ff3f3815 UI: Mention common context length values 2025-01-25 08:22:23 -08:00
oobabooga 3d4f3e423c Downloader: Make progress bars not jump around
Adapted from: https://gist.github.com/NiklasBeierl/13096bfdd8b2084da8c1163dd06f91d3
2025-01-25 07:44:24 -08:00
FP HAM 71a551a622
Add strftime_now to JINJA to sattisfy LLAMA 3.1 and 3.2 (and granite) (#6692) 2025-01-24 11:37:20 -03:00
FP HAM 5d6f3e6f92
Training pro- removed monkeypatch references (#6695) 2025-01-24 11:23:44 -03:00
oobabooga 0485ff20e8 Workaround for convert_to_markdown bug 2025-01-23 06:21:40 -08:00
oobabooga 7f8c1c1f07 Docs: update the API examples 2025-01-22 08:48:02 -08:00
Shay Molcho b76b7f6bf5
Minor README change (#6687) 2025-01-22 12:02:43 -03:00
FP HAM 4bd260c60d
Give SillyTavern a bit of leaway the way the do OpenAI (#6685) 2025-01-22 12:01:44 -03:00
oobabooga b56eb0b9cd Merge branch 'main' into dev 2025-01-22 06:44:22 -08:00
oobabooga 39799adc47 Add a helpful error message when llama.cpp fails to load the model 2025-01-21 12:49:12 -08:00
oobabooga 079ace63ec Installer: minor change 2025-01-21 10:14:05 -08:00
oobabooga 41f4fee085 Lint 2025-01-21 10:01:52 -08:00
oobabooga ff250dd800 Installer: simplify the script 2025-01-21 09:58:13 -08:00
oobabooga 2bf8788c30 Installer: Fix a bug after ecb5d3c485 2025-01-21 09:35:22 -08:00
oobabooga 5e99dded4e UI: add "Continue" and "Remove" buttons below the last chat message 2025-01-21 09:05:44 -08:00
oobabooga ecb5d3c485 Installer: do not redownload wheels for each update 2025-01-21 08:45:13 -08:00
dependabot[bot] f8a5b0bc43
Update accelerate requirement from ==1.2.* to ==1.3.* (#6683) 2025-01-20 17:41:03 -03:00
oobabooga 096272f49e Update README 2025-01-17 09:47:45 -08:00
oobabooga c32f06d62f Update README 2025-01-17 07:03:22 -08:00
oobabooga 878f378e9f
Merge pull request #6670 from oobabooga/dev
Merge dev branch
2025-01-16 10:22:49 -03:00
oobabooga 0258a6f877 Fix the Google Colab notebook 2025-01-16 05:21:18 -08:00
oobabooga fe96678692 Update some comments in the requirements 2025-01-14 19:28:48 -08:00
oobabooga ddb0f71741
Merge pull request #6666 from oobabooga/dev
Merge dev branch
2025-01-14 22:24:39 -03:00
oobabooga 2344366c9b Remove a debug message 2025-01-14 17:23:44 -08:00
oobabooga 7e80266ae9
Merge pull request #6665 from oobabooga/dev
Merge dev branch
2025-01-14 22:01:08 -03:00
oobabooga 5d25739767 Make the update wizards nice 2025-01-14 16:59:36 -08:00
oobabooga 1ef748fb20 Lint 2025-01-14 16:44:15 -08:00
oobabooga f843cb475b UI: update a help message 2025-01-14 08:12:51 -08:00
oobabooga c832953ff7 UI: Activate auto_max_new_tokens by default 2025-01-14 05:59:55 -08:00
Underscore 53b838d6c5
HTML: Fix quote pair RegEx matching for all quote types (#6661) 2025-01-13 18:01:50 -03:00
oobabooga c85e5e58d0 UI: move the new morphdom code to a .js file 2025-01-13 06:20:42 -08:00
oobabooga facb4155d4 Fix morphdom leaving ghost elements behind 2025-01-11 20:57:28 -08:00
Lounger ed16374ece
Fix the gallery extension (#6656) 2025-01-11 23:35:22 -03:00
oobabooga a0492ce325
Optimize syntax highlighting during chat streaming (#6655) 2025-01-11 21:14:10 -03:00
mamei16 f1797f4323
Unescape backslashes in html_output (#6648) 2025-01-11 18:39:44 -03:00
oobabooga 1b9121e5b8 Add a "refresh" button below the last message, add a missing file 2025-01-11 12:42:25 -08:00
oobabooga a5d64b586d
Add a "copy" button below each message (#6654) 2025-01-11 16:59:21 -03:00
oobabooga 58342740a5 Bump flash-attn to 2.7.3 2025-01-11 07:59:49 -08:00
oobabooga 3a722a36c8
Use morphdom to make chat streaming 1902381098231% faster (#6653) 2025-01-11 12:55:19 -03:00
oobabooga 02db4b0d06 Bump transformers to 4.48 2025-01-10 15:05:08 -08:00
oobabooga d2f6c0f65f Update README 2025-01-10 13:25:40 -08:00
oobabooga c393f7650d Update settings-template.yaml, organize modules/shared.py 2025-01-10 13:22:18 -08:00
oobabooga 83c426e96b
Organize internals (#6646) 2025-01-10 18:04:32 -03:00
oobabooga 17aa97248f Installer: make the hashsum verification more robust on Windows 2025-01-10 07:22:25 -08:00
oobabooga 7fe46764fb Improve the --help message about --tensorcores as well 2025-01-10 07:07:41 -08:00
oobabooga da6d868f58 Remove old deprecated flags (~6 months or more) 2025-01-09 16:11:46 -08:00
oobabooga 15bfe36619 Installer: update miniconda to 24.11.1 (experimental) 2025-01-09 15:58:14 -08:00
oobabooga e6eda6a3bb
Merge pull request #6645 from oobabooga/dev
Merge dev branch
2025-01-09 18:46:28 -03:00
oobabooga f3c0f964a2 Lint 2025-01-09 13:18:23 -08:00
oobabooga 0e94d7075e UI: minor style fix on Windows 2025-01-09 13:12:30 -08:00
oobabooga 3020f2e5ec UI: improve the info message about --tensorcores 2025-01-09 12:44:03 -08:00
oobabooga c08d87b78d Make the huggingface loader more readable 2025-01-09 12:23:38 -08:00
oobabooga 03b4067f31 Installer: ask 1 question for NVIDIA users instead of 2 2025-01-09 12:03:49 -08:00
BPplays 619265b32c
add ipv6 support to the API (#6559) 2025-01-09 10:23:44 -03:00
oobabooga 5c89068168 UI: add an info message for the new Static KV cache option 2025-01-08 17:36:30 -08:00
oobabooga 4ffc9ffc7a UI: fix a list style 2025-01-08 17:24:38 -08:00
oobabooga e6796c3859 Bump llama-cpp-python to 0.3.6, add macOS 14 and 15 wheels 2025-01-08 17:24:21 -08:00
nclok1405 b9e2ded6d4
Added UnicodeDecodeError workaround for modules/llamacpp_model.py (#6040)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2025-01-08 21:17:31 -03:00
oobabooga 91a8a87887 Remove obsolete code 2025-01-08 15:07:21 -08:00
oobabooga ad118056b8 Update README 2025-01-08 14:29:46 -08:00
oobabooga 7157257c3f
Remove the AutoGPTQ loader (#6641) 2025-01-08 19:28:56 -03:00
Jack Cloudman d3adcbf64b
Add --exclude-pattern flag to download-model.py script (#6542) 2025-01-08 17:30:21 -03:00
dependabot[bot] 1f86722977
Update safetensors requirement from ==0.4.* to ==0.5.* (#6634) 2025-01-08 16:56:55 -03:00
FP HAM 03a0f236a4
Training_PRO fix: add if 'quantization_config' in shared.model.config.to_dict() 2025-01-08 16:54:09 -03:00
oobabooga c0f600c887 Add a --torch-compile flag for transformers 2025-01-05 05:47:00 -08:00
oobabooga 11af199aff Add a "Static KV cache" option for transformers 2025-01-04 17:52:57 -08:00
oobabooga 3967520e71 Connect XTC, DRY, smoothing_factor, and dynatemp to ExLlamaV2 loader (non-HF) 2025-01-04 16:25:06 -08:00
oobabooga d56b500568 UI: add padding to file saving dialog 2025-01-04 16:22:40 -08:00
oobabooga 049297fa66 UI: reduce the size of CSS sent to the UI during streaming 2025-01-04 14:09:36 -08:00
oobabooga 0e673a7a42 UI: reduce the size of HTML sent to the UI during streaming 2025-01-04 11:40:24 -08:00
mamei16 9f24885bd2
Sane handling of markdown lists (#6626) 2025-01-04 15:41:31 -03:00
oobabooga 3815f46838 UI: minor style improvements to chat tab 2025-01-03 04:35:29 -08:00
oobabooga e2702200e1 UI: fix the font size of lists in chat mode 2025-01-02 19:26:50 -08:00
oobabooga 4b3e1b3757 UI: add a "Search chats" input field 2025-01-02 18:46:40 -08:00
oobabooga b8fc9010fa UI: fix orjson.JSONDecodeError error on page reload 2025-01-02 16:57:04 -08:00
oobabooga 973255cb0b UI: fix codeblocks overflowing on mobile 2025-01-02 16:48:49 -08:00
oobabooga 75f1b5ccde UI: add a "Branch chat" button 2025-01-02 16:24:18 -08:00
Petr Korolev 13c033c745
Fix CUDA error on MPS backend during API request (#6572)
---------

Co-authored-by: oobabooga <oobabooga4@gmail.com>
2025-01-02 00:06:11 -03:00
oobabooga 979e1f1bd6 Fix a bug after 9163951f3a 2025-01-01 17:57:09 -08:00
oobabooga f011787a83 UI: make codeblocks scroll horizontally on overflow 2025-01-01 17:55:18 -08:00
oobabooga 9163951f3a UI: reduce the CPU usage during text streaming 2025-01-01 17:49:57 -08:00
oobabooga 725639118a UI: Use a tab length of 2 for lists (rather than 4) 2025-01-01 13:53:50 -08:00
oobabooga 7b88724711
Make responses start faster by removing unnecessary cleanup calls (#6625) 2025-01-01 18:33:38 -03:00
oobabooga 88a6331abf
Merge pull request #6623 from oobabooga/dev
Merge dev branch
2024-12-31 20:47:48 -03:00
oobabooga 64853f8509 Reapply a necessary change that I removed from #6599 (thanks @mamei16!) 2024-12-31 14:43:22 -08:00
mamei16 e953af85cd
Fix newlines in the markdown renderer (#6599)
---------

Co-authored-by: oobabooga <oobabooga4@gmail.com>
2024-12-31 01:04:02 -03:00
dependabot[bot] d24b83132b
Bump jinja2 from 3.1.4 to 3.1.5 (#6601) 2024-12-30 09:35:20 -03:00
mamei16 cca4ac56fa
Fix interface loading with dark theme even when 'dark_theme' is set to false (#6614) 2024-12-30 09:34:19 -03:00
oobabooga 292cd489e9 Bump ExLlamaV2 to 0.2.7 2024-12-30 04:31:10 -08:00
oobabooga 4ce9d13dbe
Preset cleanup (#6619) 2024-12-29 12:25:26 -03:00
oobabooga 39a5c9a49c
UI organization (#6618) 2024-12-29 11:16:17 -03:00
oobabooga 0490ee620a UI: increase the threshold for a <li> to be considered long (some more) 2024-12-19 16:51:34 -08:00
oobabooga ee3a533e5c UI: improve the message width in instruct mode 2024-12-19 16:11:29 -08:00
oobabooga 89888bef56 UI: increase the threshold for a <li> to be considered long 2024-12-19 14:38:36 -08:00
oobabooga 2acec386fc UI: improve the streaming cursor 2024-12-19 14:08:56 -08:00
oobabooga e2fb86e5df UI: further improve the style of lists and headings 2024-12-19 13:59:24 -08:00
oobabooga c8ddb86c22 UI: improve some light mode colors 2024-12-19 12:24:04 -08:00
oobabooga 24a4c98d42 UI: improve the style of links in messages 2024-12-19 12:23:03 -08:00
oobabooga 836a868abc UI: improve the heading fonts 2024-12-19 12:21:28 -08:00
oobabooga 4d466d5c80
Merge pull request #6585 from oobabooga/dev
Merge dev branch
2024-12-18 23:24:55 -03:00
oobabooga fee23df1a5 Update README.md 2024-12-18 18:13:01 -08:00
oobabooga 9fd12605ac Update README.md 2024-12-18 17:58:53 -08:00
oobabooga 228caf0f3c UI: add a scrollbar to the right sidebar 2024-12-18 15:33:05 -08:00
oobabooga d01dd2e1c8 UI: fix a margin 2024-12-18 13:35:40 -08:00
Aluísio Pires 2bea4dfa96
Fix an issue caused during the installation of tts (#6496) 2024-12-18 18:16:56 -03:00
oobabooga 0a15cff6a0 UI: close sidebars by clicking outside their areas on mobile 2024-12-18 12:27:06 -08:00
oobabooga 636a6621cc UI: fix sidebars closing when typing on mobile 2024-12-18 12:16:59 -08:00
oobabooga 0c069e5b3f UI: remove obsolete js event 2024-12-18 12:16:26 -08:00
oobabooga c48e4622e8 UI: update a link 2024-12-18 06:28:14 -08:00
oobabooga b27f6f8915 Lint 2024-12-17 20:13:32 -08:00
oobabooga e83235a0cc UI: fix a font color 2024-12-17 20:11:51 -08:00
oobabooga ac0f60eb1a UI: make dropdown menus more readable 2024-12-17 20:02:04 -08:00
oobabooga b051e2c161 UI: improve a margin for readability 2024-12-17 19:58:21 -08:00
oobabooga 60c93e0c66 UI: Set cache_type to fp16 by default 2024-12-17 19:44:20 -08:00
oobabooga ddccc0d657 UI: minor change to log messages 2024-12-17 19:39:00 -08:00
oobabooga 3030c79e8c UI: show progress while loading a model 2024-12-17 19:37:43 -08:00
Diner Burger addad3c63e
Allow more granular KV cache settings (#6561) 2024-12-17 17:43:48 -03:00
oobabooga c43ee5db11 UI: very minor color change 2024-12-17 07:59:55 -08:00
oobabooga 517fcc1f23 Better centralize the chat tab 2024-12-16 20:12:16 -08:00
oobabooga d769618591
Improved UI (#6575) 2024-12-17 00:47:41 -03:00
dependabot[bot] dc56fcff12
Update bitsandbytes requirement from ==0.44.* to ==0.45.* (#6584) 2024-12-16 19:48:51 -03:00
dependabot[bot] 25c640ec0c
Update accelerate requirement from ==1.1.* to ==1.2.* (#6583) 2024-12-16 18:58:50 -03:00
oobabooga 97f5615661 Bump llama-cpp-python to 0.3.5, remove macos 12 wheels (workflow is failing) 2024-12-11 07:14:59 -08:00
oobabooga 27398428f6 Bump flash-attention to v2.7.2.post1 2024-12-09 10:17:17 -08:00
oobabooga baa566b0c6 Bump exllamav2 to 0.2.6 2024-12-09 10:16:33 -08:00
oobabooga f7836c4bd8 Bump transformers to 4.47 2024-12-09 07:00:15 -08:00
oobabooga aa629e2809 Bump exllamav2 to 0.2.5 2024-12-01 12:00:28 -08:00
oobabooga 350758f81c UI: Fix the history upload event 2024-11-19 20:34:53 -08:00
oobabooga d01293861b Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2024-11-18 10:15:36 -08:00
oobabooga 3d19746a5d UI: improve HTML rendering for lists with sub-lists 2024-11-18 10:14:09 -08:00
mefich 1c937dad72
Filter whitespaces in downloader fields in model tab (#6518) 2024-11-18 12:01:40 -03:00
dependabot[bot] f93196e306
Update accelerate requirement from ==1.0.* to ==1.1.* (#6515) 2024-11-18 12:00:24 -03:00
hronoas 9b3a3d8f12
openai extension fix: Handle Multiple Content Items in Messages (#6528) 2024-11-18 11:59:52 -03:00
oobabooga 5fa9336dab Bump flash-attention to 2.7.0.post2 2024-11-18 06:55:29 -08:00
oobabooga 0c48ecf359 Bump exllamav2 to 0.2.4 2024-11-18 06:51:56 -08:00
oobabooga 8d5cf7b134 Bump llama-cpp-python to 0.3.2 2024-11-18 06:51:06 -08:00
oobabooga cc8c7ed209
Merge pull request #6491 from oobabooga/dev
Merge dev branch
2024-10-25 01:10:23 -03:00
oobabooga 3a92fa517b Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2024-10-24 11:26:21 -07:00
oobabooga 8deea2936d Remove lm_eval from requirements 2024-10-24 11:25:42 -07:00
PIRI e1061ba7e3
Make token bans work again on HF loaders (#6488) 2024-10-24 15:24:02 -03:00
oobabooga b50dc3bf57 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2024-10-24 11:22:54 -07:00
oobabooga 386c0d8289 Bump transformers to 4.46 2024-10-24 11:09:09 -07:00
Paul Richardson 6a0837451e
Minor Documentation update - query cuda compute for docker .env (#6469) 2024-10-15 10:39:00 -03:00
Molly Sophia 18f836b280
Add RWKV-World instruction template (#6456) 2024-10-14 17:51:20 -03:00
dependabot[bot] e784938654
Update accelerate requirement from ==0.33.* to ==1.0.* (#6441) 2024-10-14 17:32:53 -03:00
oobabooga f1a8eae04d Remove optimum from requirements 2024-10-14 13:30:45 -07:00
oobabooga 2468cfd8bb Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2024-10-14 13:25:27 -07:00
oobabooga bb62e796eb Fix locally compiled llama-cpp-python failing to import 2024-10-14 13:24:13 -07:00
oobabooga c9a9f63d1b Fix llama.cpp loader not being random (thanks @reydeljuego12345) 2024-10-14 13:07:07 -07:00
PIRI 03a2e70054
Fix temperature_last when temperature not in sampler priority (#6439) 2024-10-09 11:25:14 -03:00
Grzegorz Lippe 9d8b1c5fd9
Fix intel bug described in #6253 (#6433) 2024-10-05 11:58:17 -03:00
Luana 22baa5378f
Fix for systems that have bash in a non-standard directory (#6428) 2024-10-03 00:35:13 -03:00
SeanScripts e1338a1804
Add whisper turbo (#6423) 2024-10-01 17:49:35 -03:00
oobabooga d1af7a41ad
Merge pull request #6422 from oobabooga/dev
Merge dev branch
2024-10-01 15:21:53 -03:00
oobabooga 49dfa0adaf Fix the "save preset" event 2024-10-01 11:20:48 -07:00
oobabooga 93c250b9b6 Add a UI element for enable_tp 2024-10-01 11:16:15 -07:00
oobabooga 3b06cb4523
Merge pull request #6421 from oobabooga/dev
Merge dev branch
2024-10-01 14:48:41 -03:00
oobabooga d364aa0a3c Lint 2024-10-01 10:22:57 -07:00
oobabooga cca9d6e22d Lint 2024-10-01 10:21:06 -07:00
oobabooga c6b50f88da Lint 2024-10-01 10:19:28 -07:00
oobabooga 7cb98351da
Merge branch 'main' into dev 2024-10-01 14:18:32 -03:00
oobabooga 617cd7b705 Revert "Update accelerate requirement from ==0.33.* to ==0.34.* (#6416)"
This reverts commit 6063a66414.
2024-10-01 09:06:25 -07:00
dependabot[bot] 6063a66414
Update accelerate requirement from ==0.33.* to ==0.34.* (#6416) 2024-09-30 18:50:38 -03:00
oobabooga 4d9ce586d3 Update llama_cpp_python_hijack.py, fix llamacpp_hf 2024-09-30 14:49:21 -07:00
oobabooga 9ca0cd7749 Bump llama-cpp-python to 0.3.1 2024-09-29 20:47:04 -07:00
oobabooga bbdeed3cf4 Make sampler priority high if unspecified 2024-09-29 20:45:27 -07:00
oobabooga 01362681f2 Bump exllamav2 to 0.2.4 2024-09-29 07:42:44 -07:00
Hanusz Leszek e4b0467f9f
Add beforeunload event to add confirmation dialog when leaving page (#6279) 2024-09-29 01:14:19 -03:00
Manuel Schmid 0f90a1b50f
Do not set value for histories in chat when --multi-user is used (#6317) 2024-09-29 01:08:55 -03:00
oobabooga 055f3f5632 Fix after #6386 (thanks @Touch-Night) 2024-09-28 20:55:26 -07:00
oobabooga 57160cd6fa Update README 2024-09-28 20:50:41 -07:00
oobabooga 3f0571b62b Update README 2024-09-28 20:48:30 -07:00
oobabooga 3fb02f43f6 Update README 2024-09-28 20:38:43 -07:00
oobabooga 3b99532e02 Remove HQQ and AQLM from requirements 2024-09-28 20:34:59 -07:00
oobabooga c61b29b9ce Simplify the warning when flash-attn fails to import 2024-09-28 20:33:17 -07:00
oobabooga b92d7fd43e Add warnings for when AutoGPTQ, TensorRT-LLM, or HQQ are missing 2024-09-28 20:30:24 -07:00
oobabooga 65e5864084 Update README 2024-09-28 20:25:26 -07:00
oobabooga 1a870b3ea7 Remove AutoAWQ and AutoGPTQ from requirements (no wheels available) 2024-09-28 19:38:56 -07:00
oobabooga 85994e3ef0 Bump pytorch to 2.4.1 2024-09-28 09:44:08 -07:00
oobabooga ca5a2dba72 Bump rocm to 6.1.2 2024-09-28 09:39:53 -07:00
oobabooga 7276dca933 Fix a typo 2024-09-27 20:28:17 -07:00
RandoInternetPreson 46996f6519
ExllamaV2 tensor parallelism to increase multi gpu inference speeds (#6356) 2024-09-28 00:26:03 -03:00
Philipp Emanuel Weidmann 301375834e
Exclude Top Choices (XTC): A sampler that boosts creativity, breaks writing clichés, and inhibits non-verbatim repetition (#6335) 2024-09-27 22:50:12 -03:00
oobabooga 3492e33fd5 Bump bitsandbytes to 0.44 2024-09-27 16:59:30 -07:00
Thireus ☠ 626b0a0437
Force /bin/bash shell for conda (#6386) 2024-09-27 19:47:04 -03:00
oobabooga 5c918c5b2d Make it possible to sort DRY 2024-09-27 15:40:48 -07:00
oobabooga 78b8705400 Bump llama-cpp-python to 0.3.0 (except for AMD) 2024-09-27 15:06:31 -07:00
oobabooga c5f048e912 Bump ExLlamaV2 to 0.2.2 2024-09-27 15:04:08 -07:00
oobabooga 7424f789bf
Fix the sampling monkey patch (and add more options to sampler_priority) (#6411) 2024-09-27 19:03:25 -03:00
oobabooga c497a32372 Bump transformers to 4.45 2024-09-26 11:55:51 -07:00
oobabooga f98431c744 Apply the change to all requirements (oops) 2024-09-06 18:48:13 -07:00
oobabooga a50477ec85 Apply the change to all requirements (oops) 2024-09-06 18:47:25 -07:00
oobabooga ac30b004ef Pin fastapi/pydantic requirement versions 2024-09-06 18:45:15 -07:00
oobabooga e86ab37aaf Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2024-09-06 18:44:43 -07:00
oobabooga 27797a92d0 Pin fastapi/pydantic requirement versions 2024-09-06 18:38:57 -07:00
Jean-Sylvain Boige 4924ee2901
typo in OpenAI response format (#6365) 2024-09-05 21:42:23 -03:00
oobabooga bba5b36d33 Don't import PEFT unless necessary 2024-09-03 19:40:53 -07:00
oobabooga c5b40eb555 llama.cpp: prevent prompt evaluation progress bar with just 1 step 2024-09-03 17:37:06 -07:00
oobabooga 2cb8d4c96e Bump llama-cpp-python to 0.2.90 2024-09-03 05:53:18 -07:00
oobabooga 64919e0d69 Bump flash-attention to 2.6.3 2024-09-03 05:51:46 -07:00
oobabooga 68d52c60f3 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2024-09-02 21:16:39 -07:00
oobabooga d1168afa76 Bump ExLlamaV2 to 0.2.0 2024-09-02 21:15:51 -07:00
Stefan Merettig 9a150c3368
API: Relax multimodal format, fixes HuggingFace Chat UI (#6353) 2024-09-02 23:03:15 -03:00
GralchemOz 4c74c7a116
Fix UnicodeDecodeError for BPE-based Models (especially GLM-4) (#6357) 2024-09-02 23:00:59 -03:00
FartyPants (FP HAM) 41a8eb4eeb
Training pro update script.py (#6359) 2024-09-02 23:00:15 -03:00
oobabooga 1f288b4072 Bump ExLlamaV2 to 0.1.9 2024-08-22 12:40:15 -07:00
joachimchauvet c24966c591
update API documentation with examples to list/load models (#5902) 2024-08-21 15:33:45 -03:00
oobabooga 5522584992
Merge pull request #6339 from oobabooga/dev
Merge dev branch
2024-08-20 11:20:52 -03:00
oobabooga 1124f71cf3
Update README.md 2024-08-20 11:19:46 -03:00
oobabooga 1b62cd8508
Merge pull request #6337 from oobabooga/dev
Merge dev branch
2024-08-20 01:54:47 -03:00
oobabooga d9a031fcad
Update README.md 2024-08-20 01:52:30 -03:00
oobabooga 073694bf15
Merge pull request #6336 from oobabooga/dev
Merge dev branch
2024-08-20 01:27:58 -03:00
oobabooga 9d99156ca3
Update README.md 2024-08-20 01:27:02 -03:00
oobabooga 406995f722 Update README 2024-08-19 21:24:01 -07:00
oobabooga 1b1518aa6a
Update README.md 2024-08-20 00:36:18 -03:00
oobabooga 5058269143 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2024-08-19 19:55:45 -07:00
oobabooga fd9cb26619 UI: update the DRY parameters descriptions/order 2024-08-19 19:40:17 -07:00
dependabot[bot] 64e16e9a46
Update accelerate requirement from ==0.32.* to ==0.33.* (#6291) 2024-08-19 23:34:10 -03:00
dependabot[bot] 68f928b5e0
Update peft requirement from ==0.8.* to ==0.12.* (#6292) 2024-08-19 23:33:56 -03:00
oobabooga 8bac1a9382
Update README.md 2024-08-19 23:10:04 -03:00
oobabooga bb987ffe66
Update README.md 2024-08-19 23:06:52 -03:00
oobabooga 4d8c1801c2 Bump llama-cpp-python to 0.2.89 2024-08-19 17:45:01 -07:00
oobabooga bf8187124d Bump llama-cpp-python to 0.2.88 2024-08-13 12:40:18 -07:00
oobabooga 089d5a9415 Bump llama-cpp-python to 0.2.87 2024-08-07 20:36:28 -07:00
oobabooga 81773f7f36 Bump transformers to 4.44 2024-08-06 20:07:05 -07:00
oobabooga e926c03b3d Add a --tokenizer-dir command-line flag for llamacpp_HF 2024-08-06 19:41:18 -07:00
oobabooga f106e780ba downloader: use 1 session for all files for better speed 2024-08-06 19:41:12 -07:00
oobabooga d011040f43
Merge pull request #6300 from oobabooga/dev
Merge dev branch
2024-08-01 02:26:12 -03:00
oobabooga 608545d282 Bump llama-cpp-python to 0.2.85 2024-07-31 18:44:46 -07:00
oobabooga 30b4d8c8b2 Fix Llama 3.1 template including lengthy "tools" headers 2024-07-29 11:52:17 -07:00
oobabooga f4d95f33b8 downloader: better progress bar 2024-07-28 22:21:56 -07:00
oobabooga 9dcff21da9 Remove unnecessary shared.previous_model_name variable 2024-07-28 18:35:11 -07:00
oobabooga addcb52c56 Make --idle-timeout work for API requests 2024-07-28 18:31:40 -07:00
oobabooga 514fb2e451 Fix UI error caused by --idle-timeout 2024-07-28 18:30:06 -07:00
oobabooga 3aa646c1d0 UI: improve the style of headers in chat messages 2024-07-28 15:26:15 -07:00
oobabooga 92ab3a9a6a Bump llama-cpp-python to 0.2.84 2024-07-28 15:13:06 -07:00
oobabooga 5223c009fe Minor change after previous commit 2024-07-27 23:13:34 -07:00
oobabooga 7050bb880e UI: make n_ctx/max_seq_len/truncation_length numbers rather than sliders 2024-07-27 23:11:53 -07:00
Harry 078e8c8969
Make compress_pos_emb float (#6276) 2024-07-28 03:03:19 -03:00
oobabooga ffc713f72b UI: fix multiline LaTeX equations 2024-07-27 15:36:10 -07:00
oobabooga 493f8c3242 UI: remove animation after clicking on "Stop" in the Chat tab 2024-07-27 15:22:34 -07:00
oobabooga e4d411b841 UI: fix rendering LaTeX enclosed between \[ and \] 2024-07-27 15:21:44 -07:00
oobabooga 6bab4c2faa UI: add back single $ for equations 2024-07-26 23:03:53 -07:00
oobabooga f32d26240d UI: Fix the chat "stop" event 2024-07-26 23:03:05 -07:00
oobabooga 9e82f8c394 UI: Fix chat sometimes not scrolling down after sending a message 2024-07-26 22:35:30 -07:00
oobabooga c5814db173 UI: fix double quotes in instruct mode 2024-07-25 20:22:07 -07:00
oobabooga 498fec2c7c UI: fix saving characters 2024-07-25 15:11:27 -07:00
oobabooga b80d5906c2 UI: fix saving characters 2024-07-25 15:09:31 -07:00
oobabooga dd97a83534
Merge pull request #6271 from oobabooga/dev
Merge dev branch
2024-07-25 12:12:04 -03:00
oobabooga e4624fbc68
Merge branch 'main' into dev 2024-07-25 12:03:45 -03:00
oobabooga 42e80108f5 UI: clear the markdown LRU cache when using the default/notebook tabs 2024-07-25 08:01:42 -07:00
oobabooga a34273755b Revert "Updater: don't reinstall requirements if no updates after git pull"
This reverts commit ac30e7fe9c.
2024-07-25 07:34:01 -07:00
oobabooga d581334a41 Don't install AutoAWQ on CUDA 11.8 2024-07-25 05:38:52 -07:00
oobabooga 14584fda36 UI: don't change the color of italics in instruct mode 2024-07-24 20:55:18 -07:00
oobabooga b85ae6bc96 Fix after previous commit 2024-07-24 19:10:17 -07:00
oobabooga b6830bcdae Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2024-07-24 19:04:38 -07:00
oobabooga ac30e7fe9c Updater: don't reinstall requirements if no updates after git pull 2024-07-24 19:03:34 -07:00
oobabooga 1f101ee3e5 UI: improve the quote colors 2024-07-24 18:56:54 -07:00
Luana 3170b6efc9
Fixes Linux shebangs (#6110) 2024-07-24 22:23:29 -03:00
oobabooga 7e2851e505 UI: fix "Command for chat-instruct mode" not appearing by default 2024-07-24 15:04:12 -07:00
oobabooga 947016d010 UI: make the markdown LRU cache infinite (for really long conversations) 2024-07-24 11:54:26 -07:00
oobabooga 3b2c23dfb5 Add AutoAWQ 0.2.6 wheels for PyTorch 2.2.2 2024-07-24 11:15:00 -07:00
oobabooga 8a5f110c14 Bump ExLlamaV2 to 0.1.8 2024-07-24 09:22:48 -07:00
oobabooga e637b702ff UI: make text between quotes colored in chat mode 2024-07-23 21:30:32 -07:00
oobabooga 98ed6d3a66 Don't use flash attention on Google Colab 2024-07-23 19:50:56 -07:00
oobabooga af839d20ac Remove the AutoAWQ requirement 2024-07-23 19:38:39 -07:00
oobabooga 9d5513fda0 Remove the AutoAWQ requirement 2024-07-23 19:38:04 -07:00
oobabooga 8b52b93e85 Make the Google Colab notebook functional again (attempt) 2024-07-23 19:35:00 -07:00
oobabooga e777b73349 UI: prevent LaTeX from being rendered for inline "$" 2024-07-23 19:04:19 -07:00
oobabooga 1815877061 UI: fix the default character not loading correctly on startup 2024-07-23 18:48:10 -07:00
oobabooga e6181e834a Remove AutoAWQ as a standalone loader
(it works better through transformers)
2024-07-23 15:31:17 -07:00
oobabooga f66ab63d64 Bump transformers to 4.43 2024-07-23 14:06:34 -07:00
oobabooga 6b4d762120
Merge pull request #6261 from oobabooga/dev
Merge dev branch
2024-07-23 03:11:02 -03:00
oobabooga 95b3e98c36 UI: Fix code syntax highlighting 2024-07-22 23:08:48 -07:00
oobabooga d1115f18b9
Merge pull request #6260 from oobabooga/dev
Merge dev branch
2024-07-23 02:30:35 -03:00
oobabooga 3ee682208c Revert "Bump hqq from 0.1.7.post3 to 0.1.8 (#6238)"
This reverts commit 1c3671699c.
2024-07-22 19:53:56 -07:00
oobabooga 5e7f4ee88a UI: simplify the interface load events 2024-07-22 19:11:55 -07:00
oobabooga 5c5e7264ec Update README 2024-07-22 18:20:01 -07:00
oobabooga 7e73058943 UI: fix h1/h2/h3/h4 color in light mode 2024-07-22 18:18:02 -07:00
oobabooga f18c947a86 Update the tensorcores description 2024-07-22 18:06:41 -07:00
oobabooga aa809e420e Bump llama-cpp-python to 0.2.83, add back tensorcore wheels
Also add back the progress bar patch
2024-07-22 18:05:11 -07:00
oobabooga 11bbf71aa5
Bump back llama-cpp-python (#6257) 2024-07-22 16:19:41 -03:00
oobabooga 0f53a736c1 Revert the llama-cpp-python update 2024-07-22 12:02:25 -07:00
oobabooga a687f950ba Remove the tensorcores llama.cpp wheels
They are not faster than the default wheels anymore and they use a lot of space.
2024-07-22 11:54:35 -07:00
oobabooga 017d2332ea Remove no longer necessary llama-cpp-python patch 2024-07-22 11:50:36 -07:00
oobabooga 7d2449f8b0 Bump llama-cpp-python to 0.2.82.3 (unofficial build) 2024-07-22 11:49:20 -07:00
oobabooga f2d802e707 UI: make Default/Notebook contents persist on page reload 2024-07-22 11:07:10 -07:00
oobabooga 8768b69a2d Lint 2024-07-21 22:08:14 -07:00
oobabooga 79e8dbe45f UI: minor optimization 2024-07-21 22:06:49 -07:00
oobabooga e1085180cf UI: better handle scrolling when the input area grows 2024-07-21 21:20:22 -07:00
oobabooga 7ef2414357 UI: Make the file saving dialogs more robust 2024-07-21 15:38:20 -07:00
oobabooga 423372d6e7 Organize ui_file_saving.py 2024-07-21 13:23:18 -07:00
oobabooga af99e0697e UI: increase the font weight of chat messages 2024-07-21 10:45:27 -07:00
oobabooga 17df2d7bdf UI: don't export the instruction template on "Save UI defaults to settings.yaml" 2024-07-21 10:45:01 -07:00
oobabooga d05846eae5 UI: refresh the pfp cache on handle_your_picture_change 2024-07-21 10:17:22 -07:00
oobabooga 58a1581b96 Add missing dark_theme.js (oops) 2024-07-21 09:47:55 -07:00
oobabooga e9d4bff7d0 Update the --tensor_split description 2024-07-20 22:04:48 -07:00
oobabooga 916d1d8283 UI: improve the style of code blocks in light theme 2024-07-20 20:32:57 -07:00
Patrick Leiser 9b205f94a4
Fix for issue #6024, don't auto-hide the chat contents (#6247) 2024-07-21 00:05:28 -03:00
oobabooga 564d8c8c0d Make alpha_value a float number 2024-07-20 20:02:54 -07:00
oobabooga 79c4d3da3d
Optimize the UI (#6251) 2024-07-21 00:01:42 -03:00
Alberto Cano a14c510afb
Customize the subpath for gradio, use with reverse proxy (#5106) 2024-07-20 19:10:39 -03:00
FartyPants (FP HAM) 6ab477f375
training: Added ChatML-format.json format example (#5899) 2024-07-20 19:05:09 -03:00
Vhallo a9a6d72d8c
Use gr.Number for RoPE scaling parameters (#6233)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2024-07-20 18:57:09 -03:00
dependabot[bot] 1c3671699c
Bump hqq from 0.1.7.post3 to 0.1.8 (#6238) 2024-07-20 18:20:26 -03:00
oobabooga aa7c14a463 Use chat-instruct mode by default 2024-07-19 21:43:52 -07:00
oobabooga 0315122cf0
Merge pull request #6232 from oobabooga/dev
Merge dev branch
2024-07-13 14:52:34 -03:00
oobabooga b19d239a60 Bump flash-attention to 2.6.1 2024-07-12 20:16:11 -07:00
InvectorGator 4148a9201f
Fix for MacOS users encountering model load errors (#6227)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
Co-authored-by: Invectorgator <Kudzu12gaming@outlook.com>
2024-07-13 00:04:19 -03:00
oobabooga d01c68f2a3
Merge pull request #6224 from oobabooga/dev
Merge dev branch
2024-07-11 20:42:46 -03:00
oobabooga 05676caf70 Update README 2024-07-11 16:25:52 -07:00
oobabooga f5599656b4 Update README 2024-07-11 16:22:00 -07:00
oobabooga d4eac58f2d Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2024-07-11 16:21:16 -07:00
oobabooga a30ec2e7db Update README 2024-07-11 16:20:44 -07:00
dependabot[bot] 063d2047dd
Update accelerate requirement from ==0.31.* to ==0.32.* (#6217) 2024-07-11 19:56:42 -03:00
oobabooga e436d69e2b Add --no_xformers and --no_sdpa flags for ExllamaV2 2024-07-11 15:47:37 -07:00
oobabooga 512b311137 Improve the llama-cpp-python exception messages 2024-07-11 13:00:29 -07:00
oobabooga 01e4721da7 Bump ExLlamaV2 to 0.1.7 2024-07-11 12:33:46 -07:00
oobabooga fa075e41f4 Bump llama-cpp-python to 0.2.82 2024-07-10 06:03:24 -07:00
oobabooga f957b17d18 UI: update an obsolete message 2024-07-10 06:01:36 -07:00
oobabooga c176244327 UI: Move cache_8bit/cache_4bit further up 2024-07-05 12:16:21 -07:00
oobabooga e813b322cf
Merge pull request #6203 from oobabooga/dev
Merge dev branch
2024-07-05 07:37:19 -03:00
oobabooga aa653e3b5a Prevent llama.cpp from being monkey patched more than once (closes #6201) 2024-07-05 03:34:15 -07:00
oobabooga a210e61df1 UI: Fix broken chat histories not showing (closes #6196) 2024-07-04 20:31:25 -07:00
oobabooga 3315d00651
Merge pull request #6200 from oobabooga/dev
Merge dev branch
2024-07-05 00:22:24 -03:00
oobabooga e79e7b90dc UI: Move the cache_8bit and cache_4bit elements up 2024-07-04 20:21:28 -07:00
oobabooga 363efe54f4
Merge pull request #6199 from oobabooga/dev
Merge dev branch
2024-07-05 00:17:14 -03:00
oobabooga 8b44d7b12a Lint 2024-07-04 20:16:44 -07:00
oobabooga a47de06088 Force only 1 llama-cpp-python version at a time for now 2024-07-04 19:43:34 -07:00
oobabooga f243b4ca9c Make llama-cpp-python not crash immediately 2024-07-04 19:16:00 -07:00
oobabooga f77cf159ba UI: fix a glitch when switching tabs with "show controls" unchecked 2024-07-02 20:57:03 -07:00
oobabooga 7e22eaa36c Bump llama-cpp-python to 0.2.81 2024-07-02 20:29:35 -07:00
oobabooga 907137a13d Automatically set bf16 & use_eager_attention for Gemma-2 2024-07-01 21:46:35 -07:00
TimStrauven 8074fba18d
Whisper stt overhaul js (#6194)
---------

Co-authored-by: RandoInternetPreson <aaronalai1@gmail.com>
2024-07-01 23:27:18 -03:00
GralchemOz 8a39f579d8
transformers: Add eager attention option to make Gemma-2 work properly (#6188) 2024-07-01 12:08:08 -03:00
oobabooga 19a56dd538 UI: Minor CSS improvement to chat mode 2024-06-30 21:09:54 -07:00
oobabooga 1ea3826333 UI: improve the chat area width on mobile devices 2024-06-30 17:08:23 -07:00
oobabooga ed01322763 Obtain the EOT token from the jinja template (attempt)
To use as a stopping string.
2024-06-30 15:09:22 -07:00
oobabooga 3e3f8637d6 Fix the AUTOMATIC1111 request in sd-api-pictures (closes #5993) 2024-06-29 11:43:57 -07:00
oobabooga 4ea260098f llama.cpp: add 4-bit/8-bit kv cache options 2024-06-29 09:10:33 -07:00
oobabooga 220c1797fc UI: do not show the "save character" button in the Chat tab 2024-06-28 22:11:31 -07:00
oobabooga f62aad3d59 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2024-06-28 21:42:03 -07:00
oobabooga 8803ae1845 UI: decrease the number of lines for "Command for chat-instruct mode" 2024-06-28 21:41:30 -07:00
mamei16 cc825dd1f4
Addressing Whisper STT issues (#5929) 2024-06-29 01:32:54 -03:00
oobabooga 5c6b9c610d
UI: allow the character dropdown to coexist in the Chat tab and the Parameters tab (#6177) 2024-06-29 01:20:27 -03:00
oobabooga de69a62004 Revert "UI: move "Character" dropdown to the main Chat tab"
This reverts commit 83534798b2.
2024-06-28 15:38:11 -07:00
oobabooga 38d58764db UI: remove unused gr.State variable from the Default tab 2024-06-28 15:17:44 -07:00
oobabooga 04cb197ed6 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2024-06-27 21:25:23 -07:00
oobabooga da196707cf UI: improve the light theme a bit 2024-06-27 21:05:38 -07:00
dependabot[bot] 9660f6f10e
Bump aqlm[cpu,gpu] from 1.1.5 to 1.1.6 (#6157) 2024-06-27 21:13:02 -03:00
dependabot[bot] a5df8f4e3c
Bump jinja2 from 3.1.2 to 3.1.4 (#6172) 2024-06-27 21:12:39 -03:00
dependabot[bot] c6cec0588c
Update accelerate requirement from ==0.30.* to ==0.31.* (#6156) 2024-06-27 21:12:02 -03:00
oobabooga 2f71515cb0 Make dependabot target the dev branch 2024-06-27 17:08:59 -07:00
oobabooga 1da47f2ae6 Make dependabot target the dev branch 2024-06-27 17:07:04 -07:00
oobabooga 9dbcb1aeea Small fix to make transformers 4.42 functional 2024-06-27 17:05:29 -07:00
oobabooga 66090758df Bump transformers to 4.42 (for gemma support) 2024-06-27 11:26:02 -07:00
oobabooga 6915c5077a
Merge pull request #6166 from oobabooga/dev
Merge dev branch
2024-06-26 23:33:09 -03:00
oobabooga 8ec8bc0b85 UI: handle another edge case while streaming lists 2024-06-26 18:40:43 -07:00
oobabooga 0e138e4be1 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2024-06-26 18:30:08 -07:00
mefich a85749dcbe
Update models_settings.py: add default alpha_value, add proper compress_pos_emb for newer GGUFs (#6111) 2024-06-26 22:17:56 -03:00
oobabooga 5fe532a5ce UI: remove DRY info text
It was visible for loaders without DRY.
2024-06-26 15:33:11 -07:00
oobabooga b1187fc9a5 UI: prevent flickering while streaming lists / bullet points 2024-06-25 19:19:45 -07:00
oobabooga 3691451d00
Add back the "Rename chat" feature (#6161) 2024-06-25 22:28:58 -03:00
oobabooga 53fbd2f245 Add TensorRT-LLM to the README 2024-06-25 14:45:37 -07:00
oobabooga ac3f92d36a UI: store chat history in the browser 2024-06-25 14:18:07 -07:00
oobabooga 46ca15cb79 Minor bug fixes after e7e1f5901e 2024-06-25 11:49:33 -07:00
oobabooga 83534798b2 UI: move "Character" dropdown to the main Chat tab 2024-06-25 11:25:57 -07:00
oobabooga 279cba607f UI: don't show an animation when updating the "past chats" menu 2024-06-25 11:10:17 -07:00
oobabooga 3290edfad9 Bug fix: force chat history to be loaded on launch 2024-06-25 11:06:05 -07:00
oobabooga e7e1f5901e
Prompts in the "past chats" menu (#6160) 2024-06-25 15:01:43 -03:00
oobabooga 602b455507 Bump llama-cpp-python to 0.2.79 2024-06-24 20:26:38 -07:00
oobabooga a43c210617
Improved past chats menu (#6158) 2024-06-25 00:07:22 -03:00
oobabooga 96ba53d916 Handle another fix after 57119c1b30 2024-06-24 15:51:12 -07:00
oobabooga 7db8b3b532 Bump ExLlamaV2 to 0.1.6 2024-06-24 05:38:11 -07:00
oobabooga 35f32d08bc GitHub: Increase the stalebot time to 6 months 2024-06-23 22:34:18 -07:00
oobabooga 564a3e1553 Remove the awkward "Tab" keyboard shortcut 2024-06-23 22:31:07 -07:00
oobabooga 577a8cd3ee
Add TensorRT-LLM support (#5715) 2024-06-24 02:30:03 -03:00
oobabooga 536f8d58d4 Do not expose alpha_value to llama.cpp & rope_freq_base to transformers
To avoid confusion
2024-06-23 22:09:24 -07:00
oobabooga b48ab482f8 Remove obsolete "gptq_for_llama_info" message 2024-06-23 22:05:19 -07:00
oobabooga 5e8dc56f8a Fix after previous commit 2024-06-23 21:58:28 -07:00
Louis Del Valle 57119c1b30
Update block_requests.py to resolve unexpected type error (500 error) (#5976) 2024-06-24 01:56:51 -03:00
oobabooga 125bb7b03b Revert "Bump llama-cpp-python to 0.2.78"
This reverts commit b6eaf7923e.
2024-06-23 19:54:28 -07:00
CharlesCNorton 5993904acf
Fix several typos in the codebase (#6151) 2024-06-22 21:40:25 -03:00
GodEmperor785 2c5a9eb597
Change limits of RoPE scaling sliders in UI (#6142) 2024-06-19 21:42:17 -03:00
oobabooga 5904142777 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2024-06-19 17:41:09 -07:00
oobabooga b10d735176 Minor CSS linting 2024-06-19 17:40:33 -07:00
Guanghua Lu 229d89ccfb
Make logs more readable, no more \u7f16\u7801 (#6127) 2024-06-15 23:00:13 -03:00
oobabooga fd7c3c5bb0 Don't git pull on installation (to make past releases installable) 2024-06-15 06:38:05 -07:00
oobabooga b6eaf7923e Bump llama-cpp-python to 0.2.78 2024-06-14 21:22:09 -07:00
oobabooga 9420973b62
Downgrade PyTorch to 2.2.2 (#6124) 2024-06-14 16:42:03 -03:00
Forkoz 1576227f16
Fix GGUFs with no BOS token present, mainly qwen2 models. (#6119)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2024-06-14 13:51:01 -03:00
dependabot[bot] fdd8fab9cf
Bump hqq from 0.1.7.post2 to 0.1.7.post3 (#6090) 2024-06-14 13:46:35 -03:00
oobabooga 10601850d9 Fix after previous commit 2024-06-13 19:54:12 -07:00
oobabooga 0f3a423de1 Alternative solution to "get next logits" deadlock (#6106) 2024-06-13 19:34:16 -07:00
oobabooga 9aef01551d Revert "Use reentrant generation lock (#6107)"
This reverts commit b675151f25.
2024-06-13 17:53:07 -07:00
oobabooga 8930bfc5f4
Bump PyTorch, ExLlamaV2, flash-attention (#6122) 2024-06-13 20:38:31 -03:00
oobabooga 386500aa37 Avoid unnecessary calls UI -> backend, to make it faster 2024-06-12 20:52:42 -07:00
oobabooga 4820ae9aef
Merge pull request #6118 from oobabooga/dev
Merge dev branch
2024-06-13 00:38:03 -03:00
Forkoz 1d79aa67cf
Fix flash-attn UI parameter to actually store true. (#6076) 2024-06-13 00:34:54 -03:00
Belladore 3abafee696
DRY sampler improvements (#6053) 2024-06-12 23:39:11 -03:00
theo77186 b675151f25
Use reentrant generation lock (#6107) 2024-06-12 23:25:05 -03:00
oobabooga a36fa73071 Lint 2024-06-12 19:00:21 -07:00
oobabooga 2d196ed2fe Remove obsolete pre_layer parameter 2024-06-12 18:56:44 -07:00
Belladore 46174a2d33
Fix error when bos_token_id is None. (#6061) 2024-06-12 22:52:27 -03:00
Belladore a363cdfca1
Fix missing bos token for some models (including Llama-3) (#6050) 2024-05-27 09:21:30 -03:00
oobabooga 8df68b05e9 Remove MinPLogitsWarper (it's now a transformers built-in) 2024-05-27 05:03:30 -07:00
oobabooga 4f1e96b9e3 Downloader: Add --model-dir argument, respect --model-dir in the UI 2024-05-23 20:42:46 -07:00
oobabooga ad54d524f7 Revert "Fix stopping strings for llama-3 and phi (#6043)"
This reverts commit 5499bc9bc8.
2024-05-22 17:18:08 -07:00
oobabooga 5499bc9bc8
Fix stopping strings for llama-3 and phi (#6043) 2024-05-22 13:53:59 -03:00
rohitanshu 8aaa0a6f4e
Fixed minor typo in docs - Training Tab.md (#6038) 2024-05-21 14:52:22 -03:00
oobabooga 9e189947d1 Minor fix after bd7cc4234d (thanks @belladoreai) 2024-05-21 10:37:30 -07:00
oobabooga ae86292159 Fix getting Phi-3-small-128k-instruct logits 2024-05-21 10:35:00 -07:00
oobabooga bd7cc4234d
Backend cleanup (#6025) 2024-05-21 13:32:02 -03:00
oobabooga 6a1682aa95 README: update command-line flags with raw --help output
This helps me keep this up-to-date more easily.
2024-05-19 20:28:46 -07:00
Philipp Emanuel Weidmann 852c943769
DRY: A modern repetition penalty that reliably prevents looping (#5677) 2024-05-19 23:53:47 -03:00
oobabooga 9f77ed1b98
--idle-timeout flag to unload the model if unused for N minutes (#6026) 2024-05-19 23:29:39 -03:00
altoiddealer 818b4e0354
Let grammar escape backslashes (#5865) 2024-05-19 20:26:09 -03:00
Tisjwlf 907702c204
Fix gguf multipart file loading (#5857) 2024-05-19 20:22:09 -03:00
Guanghua Lu d7bd3da35e
Add Llama 3 instruction template (#5891) 2024-05-19 20:17:26 -03:00
A0nameless0man 5cb59707f3
fix: grammar not support utf-8 (#5900) 2024-05-19 20:10:39 -03:00
Jari Van Melckebeke 8456d13349
[docs] small docker changes (#5917) 2024-05-19 20:09:37 -03:00
Samuel Wein b63dc4e325
UI: Warn user if they are trying to load a model from no path (#6006) 2024-05-19 20:05:17 -03:00
dependabot[bot] 2de586f586
Update accelerate requirement from ==0.27.* to ==0.30.* (#5989) 2024-05-19 20:03:18 -03:00
chr 6b546a2c8b
llama.cpp: increase the max threads from 32 to 256 (#5889) 2024-05-19 20:02:19 -03:00
oobabooga abe5ddc883
Merge pull request #6027 from oobabooga/dev
Merge dev branch
2024-05-19 19:01:11 -03:00
oobabooga a38a37b3b3 llama.cpp: default n_gpu_layers to the maximum value for the model automatically 2024-05-19 10:57:42 -07:00
oobabooga a4611232b7 Make --verbose output less spammy 2024-05-18 09:57:00 -07:00
oobabooga 0d90b3a25c Bump llama-cpp-python to 0.2.75 2024-05-18 05:26:26 -07:00
oobabooga e225b0b995 downloader: fix downloading 01-ai/Yi-1.5-34B-Chat 2024-05-12 10:43:50 -07:00
oobabooga 9557f49f2f Bump llama-cpp-python to 0.2.73 2024-05-11 10:53:19 -07:00
oobabooga 9ac528715c
Merge pull request #5996 from oobabooga/dev
Merge dev branch
2024-05-08 16:37:26 -03:00
oobabooga 7a728a38eb Update README 2024-05-07 02:59:36 -07:00
oobabooga d5bde7babc UI: improve the performance of code syntax highlighting 2024-05-06 17:45:03 -07:00
oobabooga 0b193b8553 Downloader: handle one more retry case after 5770e06c48 2024-05-04 19:25:22 -07:00
oobabooga cb31998605 Add a template for NVIDIA ChatQA models 2024-05-03 08:19:04 -07:00
oobabooga e9c9483171 Improve the logging messages while loading models 2024-05-03 08:10:44 -07:00
oobabooga e61055253c Bump llama-cpp-python to 0.2.69, add --flash-attn option 2024-05-03 04:31:22 -07:00
oobabooga 0476f9fe70 Bump ExLlamaV2 to 0.0.20 2024-05-01 16:20:50 -07:00
oobabooga ae0f28530c Bump llama-cpp-python to 0.2.68 2024-05-01 08:40:50 -07:00
oobabooga 8f12fb028d
Merge pull request #5970 from oobabooga/dev
Merge dev branch
2024-05-01 09:56:23 -03:00
oobabooga 1eba888af6 Update FUNDING.yml 2024-05-01 05:54:21 -07:00
oobabooga 51fb766bea
Add back my llama-cpp-python wheels, bump to 0.2.65 (#5964) 2024-04-30 09:11:31 -03:00
oobabooga 81f603d09f
Merge pull request #5959 from oobabooga/dev
Merge dev branch
2024-04-29 15:45:48 -03:00
oobabooga 5770e06c48
Add a retry mechanism to the model downloader (#5943) 2024-04-27 12:25:28 -03:00
oobabooga dfdb6fee22 Set llm_int8_enable_fp32_cpu_offload=True for --load-in-4bit
To allow for 32-bit CPU offloading (it's very slow).
2024-04-26 09:39:27 -07:00
oobabooga 70845c76fb
Add back the max_updates_second parameter (#5937) 2024-04-26 10:14:51 -03:00
oobabooga 6761b5e7c6
Improved instruct style (with syntax highlighting & LaTeX rendering) (#5936) 2024-04-26 10:13:11 -03:00
oobabooga 9c04365f54 Detect the airoboros-3_1-yi-34b-200k template 2024-04-25 16:50:54 -07:00
oobabooga 8b1dee3ec8 Detect platypus-yi-34b, CausalLM-RP-34B, 34b-beta instruction templates 2024-04-24 21:47:43 -07:00
oobabooga 4aa481282b Detect the xwin-lm-70b-v0.1 instruction template 2024-04-24 17:02:20 -07:00
oobabooga ad122361ea
Merge pull request #5927 from oobabooga/dev
Merge dev branch
2024-04-24 13:58:53 -03:00
oobabooga c9b0df16ee Lint 2024-04-24 09:55:00 -07:00
oobabooga 4094813f8d Lint 2024-04-24 09:53:41 -07:00
oobabooga 64e2a9a0a7 Fix the Phi-3 template when used in the UI 2024-04-24 01:34:11 -07:00
oobabooga f0538efb99 Remove obsolete --tensorcores references 2024-04-24 00:31:28 -07:00
Colin f3c9103e04
Revert walrus operator for params['max_memory'] (#5878) 2024-04-24 01:09:14 -03:00
Jari Van Melckebeke c725d97368
nvidia docker: make sure gradio listens on 0.0.0.0 (#5918) 2024-04-23 23:17:55 -03:00
oobabooga 9b623b8a78
Bump llama-cpp-python to 0.2.64, use official wheels (#5921) 2024-04-23 23:17:05 -03:00
Ashley Kleynhans 0877741b03
Bumped ExLlamaV2 to version 0.0.19 to resolve #5851 (#5880) 2024-04-19 19:04:40 -03:00
oobabooga a4b732c30b
Merge pull request #5887 from oobabooga/dev
Merge dev branch
2024-04-19 12:34:50 -03:00
oobabooga f27e1ba302
Add a /v1/internal/chat-prompt endpoint (#5879) 2024-04-19 00:24:46 -03:00
oobabooga b30bce3b2f Bump transformers to 4.40 2024-04-18 16:19:31 -07:00
Philipp Emanuel Weidmann a0c69749e6
Revert sse-starlette version bump because it breaks API request cancellation (#5873) 2024-04-18 15:05:00 -03:00
mamei16 8985a8538b
Fix whisper STT (#5856) 2024-04-14 10:55:58 -03:00
oobabooga 26d822f64f
Merge pull request #5848 from oobabooga/dev
Merge dev branch
2024-04-12 12:46:25 -03:00
dependabot[bot] 597556cb77
Bump sse-starlette from 1.6.5 to 2.1.0 (#5831) 2024-04-11 18:54:05 -03:00
oobabooga e158299fb4 Fix loading sharted GGUF models through llamacpp_HF 2024-04-11 14:50:05 -07:00
wangshuai09 fd4e46bce2
Add Ascend NPU support (basic) (#5541) 2024-04-11 18:42:20 -03:00
zaypen a90509d82e
Model downloader: Take HF_ENDPOINT in consideration (#5571) 2024-04-11 18:28:10 -03:00
Ashley Kleynhans 70c637bf90
Fix saving of UI defaults to settings.yaml - Fixes #5592 (#5794) 2024-04-11 18:19:16 -03:00
oobabooga 3e3a7c4250 Bump llama-cpp-python to 0.2.61 & fix the crash 2024-04-11 14:15:34 -07:00
oobabooga 5f5ceaf025 Revert "Bump llama-cpp-python to 0.2.61"
This reverts commit 3ae61c0338.
2024-04-11 13:24:57 -07:00
dependabot[bot] bd71a504b8
Update gradio requirement from ==4.25.* to ==4.26.* (#5832) 2024-04-11 02:24:53 -03:00
Victorivus c423d51a83
Fix issue #5783 for character images with transparency (#5827) 2024-04-11 02:23:43 -03:00
Alex O'Connell b94cd6754e
UI: Respect model and lora directory settings when downloading files (#5842) 2024-04-11 01:55:02 -03:00
oobabooga 17c4319e2d Fix loading command-r context length metadata 2024-04-10 21:39:59 -07:00
oobabooga 3ae61c0338 Bump llama-cpp-python to 0.2.61 2024-04-10 21:39:46 -07:00
oobabooga cbd65ba767
Add a simple min_p preset, make it the default (#5836) 2024-04-09 12:50:16 -03:00
oobabooga ed4001e324 Bump ExLlamaV2 to 0.0.18 2024-04-08 18:05:16 -07:00
oobabooga 91a7370a65
Merge pull request #5823 from oobabooga/dev
Merge dev branch
2024-04-07 11:01:08 -03:00
oobabooga f6828de3f2 Downgrade llama-cpp-python to 0.2.56 2024-04-07 07:00:12 -07:00
Jared Van Bortel 39ff9c9dcf
requirements: add psutil (#5819) 2024-04-06 23:02:20 -03:00
oobabooga 65099dc192
Merge pull request #5822 from oobabooga/dev
Merge dev branch
2024-04-06 22:58:06 -03:00
oobabooga d02744282b Minor logging change 2024-04-06 18:56:58 -07:00
oobabooga dfb01f9a63 Bump llama-cpp-python to 0.2.60 2024-04-06 18:32:36 -07:00
oobabooga 096f75a432 Documentation: remove obsolete RWKV docs 2024-04-06 14:06:39 -07:00
oobabooga dd6e4ac55f Prevent double <BOS_TOKEN> with Command R+ 2024-04-06 13:14:32 -07:00
oobabooga 1bdceea2d4 UI: Focus on the chat input after starting a new chat 2024-04-06 12:57:57 -07:00
oobabooga 168a0f4f67 UI: do not load the "gallery" extension by default 2024-04-06 12:43:21 -07:00
oobabooga 64a76856bd Metadata: Fix loading Command R+ template with multiple options 2024-04-06 07:32:17 -07:00
oobabooga 1b87844928 Minor fix 2024-04-05 18:43:43 -07:00
oobabooga 6b7f7555fc Logging message to make transformers loader a bit more transparent 2024-04-05 18:40:02 -07:00
oobabooga 4e739dc211 Add an instruction template for Command R 2024-04-05 18:22:25 -07:00
oobabooga 8a8dbf2f16 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2024-04-05 12:42:23 -07:00
oobabooga 0f536dd97d UI: Fix the "Show controls" action 2024-04-05 12:18:33 -07:00
dependabot[bot] a4c67e1974
Bump aqlm[cpu,gpu] from 1.1.2 to 1.1.3 (#5790) 2024-04-05 13:26:49 -03:00
oobabooga 14f6194211 Bump Gradio to 4.25 2024-04-05 09:22:44 -07:00
oobabooga 5b91dbb73b
Merge pull request #5810 from oobabooga/dev
Merge dev branch
2024-04-05 10:55:16 -03:00
oobabooga 308452b783 Bitsandbytes: load preconverted 4bit models without additional flags 2024-04-04 18:10:24 -07:00
oobabooga d423021a48
Remove CTransformers support (#5807) 2024-04-04 20:23:58 -03:00
oobabooga 13fe38eb27 Remove specialized code for gpt-4chan 2024-04-04 16:11:47 -07:00
oobabooga 3952560da8 Bump llama-cpp-python to 0.2.59 2024-04-04 11:20:48 -07:00
oobabooga 9ab7365b56 Read rope_theta for DBRX model (thanks turboderp) 2024-04-01 20:25:31 -07:00
oobabooga db5f6cd1d8 Fix ExLlamaV2 loaders using unnecessary "bits" metadata 2024-03-30 21:51:39 -07:00
oobabooga 624faa1438 Fix ExLlamaV2 context length setting (closes #5750) 2024-03-30 21:33:16 -07:00
oobabooga 70c58b5fc2 Bump ExLlamaV2 to 0.0.17 2024-03-30 21:08:26 -07:00
oobabooga 1a7c027386
Merge pull request #5772 from oobabooga/dev
Merge dev branch
2024-03-29 15:09:53 -03:00
oobabooga c37f792afa Better way to handle user_bio default in the API (alternative to bdcf31035f) 2024-03-29 10:54:01 -07:00
oobabooga 9653a9176c Minor improvements to Parameters tab 2024-03-29 10:41:24 -07:00
oobabooga 3ce0d9221b Bump transformers to 4.39 2024-03-28 19:40:31 -07:00
oobabooga e0e28ecb0b Set the gradio 4 allowed_paths 2024-03-28 15:10:54 -07:00
oobabooga 723f912c16 Fix the "typing dots" position in latest Gradio version 2024-03-28 12:57:35 -07:00
oobabooga 35da6b989d
Organize the parameters tab (#5767) 2024-03-28 16:45:03 -03:00
dependabot[bot] 3609ea69e4
Bump aqlm[cpu,gpu] from 1.1.0 to 1.1.2 (#5728) 2024-03-26 16:36:16 -03:00
Bartowski 9ad116a6e2
Add config for hyperion and hercules models to use chatml (#5742) 2024-03-26 16:35:29 -03:00
wldhx 7cbafc0540
docker: Remove obsolete CLI_ARGS variable (#5726) 2024-03-26 16:34:53 -03:00
Yiximail bdcf31035f
Set a default empty string for user_bio to fix #5717 issue (#5722) 2024-03-26 16:34:03 -03:00
Yiximail 8c9aca239a
Fix prompt incorrectly set to empty when suffix is empty string (#5757) 2024-03-26 16:33:09 -03:00
oobabooga 2a92a842ce
Bump gradio to 4.23 (#5758) 2024-03-26 16:32:20 -03:00
oobabooga 7cf1402bde
Merge pull request #5716 from oobabooga/dev
Merge dev branch
2024-03-17 12:34:53 -03:00
oobabooga 49b111e2dd Lint 2024-03-17 08:33:23 -07:00
oobabooga d890c99b53 Fix StreamingLLM when content is removed from the beginning of the prompt 2024-03-14 09:18:54 -07:00
oobabooga d828844a6f Small fix: don't save truncation_length to settings.yaml
It should derive from model metadata or from a command-line flag.
2024-03-14 08:56:28 -07:00
oobabooga 2ef5490a36 UI: make light theme less blinding 2024-03-13 08:23:16 -07:00
oobabooga 40a60e0297 Convert attention_sink_size to int (closes #5696) 2024-03-13 08:15:49 -07:00
oobabooga edec3bf3b0 UI: avoid caching convert_to_markdown calls during streaming 2024-03-13 08:14:34 -07:00
oobabooga 8152152dd6 Small fix after 28076928ac 2024-03-11 19:56:35 -07:00
oobabooga 28076928ac
UI: Add a new "User description" field for user personality/biography (#5691) 2024-03-11 23:41:57 -03:00
oobabooga 63701f59cf UI: mention that n_gpu_layers > 0 is necessary for the GPU to be used 2024-03-11 18:54:15 -07:00
oobabooga 46031407b5 Increase the cache size of convert_to_markdown to 4096 2024-03-11 18:43:04 -07:00
oobabooga 9eca197409 Minor logging change 2024-03-11 16:31:13 -07:00
oobabooga afadc787d7 Optimize the UI by caching convert_to_markdown calls 2024-03-10 20:10:07 -07:00
oobabooga 1934cb61ef
Merge pull request #5680 from oobabooga/dev
Merge dev branch
2024-03-10 23:39:20 -03:00
oobabooga 056717923f Document StreamingLLM 2024-03-10 19:15:23 -07:00
oobabooga 15d90d9bd5 Minor logging change 2024-03-10 18:20:50 -07:00
oobabooga abcdd0ad5b API: don't use settings.yaml for default values 2024-03-10 16:15:52 -07:00
oobabooga a102c704f5 Add numba to requirements.txt 2024-03-10 16:13:29 -07:00
oobabooga b3ade5832b Keep AQLM only for Linux (fails to install on Windows) 2024-03-10 09:41:17 -07:00
oobabooga 67b24b0b88 Bump llama-cpp-python to 0.2.56 2024-03-10 09:07:27 -07:00
oobabooga 763f9beb7e Bump bitsandbytes to 0.43, add official Windows wheel 2024-03-10 08:30:53 -07:00
oobabooga 52a34921ef Installer: validate the checksum for the miniconda installer on Windows 2024-03-09 16:33:12 -08:00
oobabooga cf0697936a Optimize StreamingLLM by over 10x 2024-03-08 21:48:28 -08:00
oobabooga afb51bd5d6
Add StreamingLLM for llamacpp & llamacpp_HF (2nd attempt) (#5669) 2024-03-09 00:25:33 -03:00
oobabooga 9271e80914 Add back AutoAWQ for Windows
https://github.com/casper-hansen/AutoAWQ/issues/377#issuecomment-1986440695
2024-03-08 14:54:56 -08:00
oobabooga 549bb88975 Increase height of "Custom stopping strings" UI field 2024-03-08 12:54:30 -08:00
oobabooga 238f69accc Move "Command for chat-instruct mode" to the main chat tab (closes #5634) 2024-03-08 12:52:52 -08:00
oobabooga d0663bae31
Bump AutoAWQ to 0.2.3 (Linux only) (#5658) 2024-03-08 17:36:28 -03:00
oobabooga 0e6eb7c27a
Add AQLM support (transformers loader) (#5466) 2024-03-08 17:30:36 -03:00
oobabooga 2681f6f640
Make superbooga & superboogav2 functional again (#5656) 2024-03-07 15:03:18 -03:00
oobabooga bae14c8f13 Right-truncate long chat completion prompts instead of left-truncating
Instructions are usually at the beginning of the prompt.
2024-03-07 08:50:24 -08:00
oobabooga aa0da07af0
Merge pull request #5655 from oobabooga/dev
Merge dev branch
2024-03-07 13:13:10 -03:00
Bartowski 104573f7d4
Update cache_4bit documentation (#5649)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2024-03-07 13:08:21 -03:00
oobabooga bef08129bc Small fix for cuda 11.8 in the one-click installer 2024-03-06 21:43:36 -08:00
oobabooga 303433001f Fix a check in the installer 2024-03-06 21:13:54 -08:00
oobabooga bde7f00cae Change the exllamav2 version number 2024-03-06 21:08:29 -08:00
oobabooga 2ec1d96c91
Add cache_4bit option for ExLlamaV2 (#5645) 2024-03-06 23:02:25 -03:00
oobabooga fa0e68cefd Installer: add back INSTALL_EXTENSIONS environment variable (for docker) 2024-03-06 11:31:06 -08:00
oobabooga 992affefef
Merge pull request #5641 from oobabooga/dev
Merge dev branch
2024-03-06 12:40:10 -03:00
oobabooga fcc92caa30 Installer: add option to install requirements for just one extension 2024-03-06 07:36:23 -08:00
oobabooga 2174958362
Revert gradio to 3.50.2 (#5640) 2024-03-06 11:52:46 -03:00
oobabooga 7eee9e9470 Add -k to curl command to download miniconda on windows (closes #5628) 2024-03-06 06:46:50 -08:00
oobabooga 03f03af535 Revert "Update peft requirement from ==0.8.* to ==0.9.* (#5626)"
This reverts commit 72a498ddd4.
2024-03-05 02:56:37 -08:00
oobabooga d61e31e182
Save the extensions after Gradio 4 (#5632) 2024-03-05 07:54:34 -03:00
oobabooga ae12d045ea Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2024-03-05 02:35:04 -08:00
dependabot[bot] 72a498ddd4
Update peft requirement from ==0.8.* to ==0.9.* (#5626) 2024-03-05 07:34:32 -03:00
oobabooga 1437f757a1 Bump HQQ to 0.1.5 2024-03-05 02:33:51 -08:00
oobabooga 63a1d4afc8
Bump gradio to 4.19 (#5522) 2024-03-05 07:32:28 -03:00
oobabooga 164ff2440d Use the correct PyTorch in the Colab notebook 2024-03-05 01:05:19 -08:00
oobabooga 3cfcab63a5 Update an installation message 2024-03-04 20:37:44 -08:00
oobabooga 907bda0d56 Move update_wizard_wsl.sh to update_wizard_wsl.bat 2024-03-04 19:57:49 -08:00
oobabooga f697cb4609 Move update_wizard_windows.sh to update_wizard_windows.bat (oops) 2024-03-04 19:26:24 -08:00
oobabooga 2d74660733 Don't git pull on "Install/update extensions requirements" 2024-03-04 12:37:10 -08:00
oobabooga fbe83854ca Minor message change 2024-03-04 11:10:37 -08:00
oobabooga 90ab022856 Minor message change 2024-03-04 10:54:16 -08:00
oobabooga 97dc3602fc
Create an update wizard (#5623) 2024-03-04 15:52:24 -03:00
oobabooga 6adf222599 One-click installer: change an info message 2024-03-04 08:20:04 -08:00
oobabooga 4bb79c57ac One-click installer: change an info message 2024-03-04 08:11:55 -08:00
oobabooga 74564fe8d0 One-click installer: delete the Miniconda installer after completion 2024-03-04 08:11:03 -08:00
oobabooga dc2dd5b9d8 One-click installer: add an info message before git pull 2024-03-04 08:00:39 -08:00
oobabooga 527ba98105
Do not install extensions requirements by default (#5621) 2024-03-04 04:46:39 -03:00
oobabooga fa4ce0eee8 One-click installer: minor change to CMD_FLAGS.txt in CPU mode 2024-03-03 17:42:59 -08:00
oobabooga 8bd4960d05
Update PyTorch to 2.2 (also update flash-attn to 2.5.6) (#5618) 2024-03-03 19:40:32 -03:00
oobabooga 70047a5c57 Bump bitsandytes to 0.42.0 on Windows 2024-03-03 13:19:27 -08:00
oobabooga 24e86bb21b Bump llama-cpp-python to 0.2.55 2024-03-03 12:14:48 -08:00
oobabooga 60f3d87309
Merge pull request #5617 from oobabooga/dev
Merge dev branch
2024-03-03 15:50:40 -03:00
oobabooga 314e42fd98 Fix transformers requirement 2024-03-03 10:49:28 -08:00
oobabooga 71b1617c1b Remove bitsandbytes from incompatible requirements.txt files 2024-03-03 08:24:54 -08:00
kalomaze cfb25c9b3f
Cubic sampling w/ curve param (#5551)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2024-03-03 13:22:21 -03:00
jeffbiocode 3168644152
Training: Update llama2-chat-format.json (#5593) 2024-03-03 12:42:14 -03:00
oobabooga 71dc5b4dee Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2024-02-28 19:59:20 -08:00
oobabooga 09b13acfb2 Perplexity evaluation: print to terminal after calculation is finished 2024-02-28 19:58:21 -08:00
dependabot[bot] dfdf6eb5b4
Bump hqq from 0.1.3 to 0.1.3.post1 (#5582) 2024-02-26 20:51:39 -03:00
oobabooga 332957ffec Bump llama-cpp-python to 0.2.52 2024-02-26 15:05:53 -08:00
oobabooga b64770805b Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2024-02-26 08:51:31 -08:00
oobabooga 830168d3d4 Revert "Replace hashlib.sha256 with hashlib.file_digest so we don't need to load entire files into ram before hashing them. (#4383)"
This reverts commit 0ced78fdfa.
2024-02-26 05:54:33 -08:00
Bartowski 21acf504ce
Bump transformers to 4.38 for gemma compatibility (#5575) 2024-02-25 20:15:13 -03:00
oobabooga 4164e29416 Block the "To create a public link, set share=True" gradio message 2024-02-25 15:06:08 -08:00
oobabooga ba852716fd
Merge pull request #5574 from oobabooga/dev
Merge dev branch
2024-02-25 14:29:35 -03:00
oobabooga d34126255d Fix loading extensions with "-" in the name (closes #5557) 2024-02-25 09:24:52 -08:00
Lounger 0f68c6fb5b
Big picture fixes (#5565) 2024-02-25 14:10:16 -03:00
jeffbiocode 45c4cd01c5
Add llama 2 chat format for lora training (#5553) 2024-02-25 02:36:36 -03:00
Devin Roark e0fc808980
fix: ngrok logging does not use the shared logger module (#5570) 2024-02-25 02:35:59 -03:00
oobabooga 32ee5504ed
Remove -k from curl command to download miniconda (#5535) 2024-02-25 02:35:23 -03:00
oobabooga c07dc56736 Bump llama-cpp-python to 0.2.50 2024-02-24 21:34:11 -08:00
oobabooga 98580cad8e Bump exllamav2 to 0.0.14 2024-02-24 18:35:42 -08:00
oobabooga 527f2652af Bump llama-cpp-python to 0.2.47 2024-02-22 19:48:49 -08:00
oobabooga 3f42e3292a Revert "Bump autoawq from 0.1.8 to 0.2.2 (#5547)"
This reverts commit d04fef6a07.
2024-02-22 19:48:04 -08:00
oobabooga 10aedc329f Logging: more readable messages when renaming chat histories 2024-02-22 07:57:06 -08:00
oobabooga faf3bf2503 Perplexity evaluation: make UI events more robust (attempt) 2024-02-22 07:13:22 -08:00
oobabooga ac5a7a26ea Perplexity evaluation: add some informative error messages 2024-02-21 20:20:52 -08:00
oobabooga 59032140b5 Fix CFG with llamacpp_HF (2nd attempt) 2024-02-19 18:35:42 -08:00
oobabooga c203c57c18 Fix CFG with llamacpp_HF 2024-02-19 18:09:49 -08:00
dependabot[bot] 5f7dbf454a
Update optimum requirement from ==1.16.* to ==1.17.* (#5548) 2024-02-19 19:15:21 -03:00
dependabot[bot] d04fef6a07
Bump autoawq from 0.1.8 to 0.2.2 (#5547) 2024-02-19 19:14:55 -03:00
dependabot[bot] ed6ff49431
Update accelerate requirement from ==0.25.* to ==0.27.* (#5546) 2024-02-19 19:14:04 -03:00
oobabooga d6bb6e7390
Merge pull request #5549 from oobabooga/dev
Merge dev branch
2024-02-19 18:53:25 -03:00
Kevin Pham 10df23efb7
Remove message.content from openai streaming API (#5503) 2024-02-19 18:50:27 -03:00
oobabooga 0b2279d031 Bump llama-cpp-python to 0.2.44 2024-02-19 13:42:31 -08:00
oobabooga ae05d9830f Replace {{char}}, {{user}} in the chat template itself 2024-02-18 19:57:54 -08:00
oobabooga 717c3494e8 Minor width change after daa140447e 2024-02-18 15:23:45 -08:00
oobabooga 1f27bef71b
Move chat UI elements to the right on desktop (#5538) 2024-02-18 14:32:05 -03:00
oobabooga d8064c00e8 UI: hide chat scrollbar on desktop when not hovered 2024-02-17 20:47:14 -08:00
oobabooga 36c29084bb UI: fix instruct style background for multiline inputs 2024-02-17 20:09:47 -08:00
oobabooga 904867a139 UI: fix scroll down after sending a multiline message 2024-02-17 19:27:13 -08:00
oobabooga 7838075990
Merge pull request #5534 from oobabooga/dev
Merge dev branch
2024-02-17 18:09:40 -03:00
oobabooga d6bd71db7f ExLlamaV2: fix loading when autosplit is not set 2024-02-17 12:54:37 -08:00
oobabooga dd46229487
Merge pull request #5530 from oobabooga/dev
Merge dev branch
2024-02-17 14:02:39 -03:00
oobabooga af0bbf5b13 Lint 2024-02-17 09:01:04 -08:00
fschuh fa1019e8fe
Removed extra spaces from Mistral instruction template that were causing Mistral to misbehave (#5517) 2024-02-16 21:40:51 -03:00
oobabooga c375c753d6 Bump bitsandbytes to 0.42 (Linux only) 2024-02-16 10:47:57 -08:00
oobabooga a6730f88f7
Add --autosplit flag for ExLlamaV2 (#5524) 2024-02-16 15:26:10 -03:00
oobabooga 4039999be5 Autodetect llamacpp_HF loader when tokenizer exists 2024-02-16 09:29:26 -08:00
oobabooga 76d28eaa9e
Add a menu for customizing the instruction template for the model (#5521) 2024-02-16 14:21:17 -03:00
oobabooga 0e1d8d5601 Instruction template: make "Send to default/notebook" work without a tokenizer 2024-02-16 08:01:07 -08:00
oobabooga f465b7b486
Downloader: start one session per file (#5520) 2024-02-16 12:55:27 -03:00
oobabooga 44018c2f69
Add a "llamacpp_HF creator" menu (#5519) 2024-02-16 12:43:24 -03:00
oobabooga b2b74c83a6 Fix Qwen1.5 in llamacpp_HF 2024-02-15 19:04:19 -08:00
oobabooga 080f7132c0
Revert gradio to 3.50.2 (#5513) 2024-02-15 20:40:23 -03:00
oobabooga ea0e1feee7 Bump llama-cpp-python to 0.2.43 2024-02-14 21:58:24 -08:00
oobabooga 549f106879 Bump ExLlamaV2 to v0.0.13.2 2024-02-14 21:57:48 -08:00
oobabooga 7123ac3f77
Remove "Maximum UI updates/second" parameter (#5507) 2024-02-14 23:34:30 -03:00
DominikKowalczyk 33c4ce0720
Bump gradio to 4.19 (#5419)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2024-02-14 23:28:26 -03:00
oobabooga 771c59290a
Merge pull request #5502 from oobabooga/dev
Merge dev branch
2024-02-14 11:32:58 -03:00
oobabooga 04d8bdf929 Fix ExLlamaV2 requirement on Windows 2024-02-14 06:31:20 -08:00
oobabooga b16958575f Minor bug fix 2024-02-13 19:48:32 -08:00
oobabooga d47182d9d1
llamacpp_HF: do not use oobabooga/llama-tokenizer (#5499) 2024-02-14 00:28:51 -03:00
oobabooga 3a9ce3cfa6 Update stalebot message 2024-02-13 19:06:32 -08:00
oobabooga 93dd31fc0f Increase stalebot timeout 2024-02-13 16:07:33 -08:00
oobabooga dc6adefd87
Merge pull request #5496 from oobabooga/dev
Merge dev branch
2024-02-13 21:06:16 -03:00
oobabooga 069ed7c6ef Lint 2024-02-13 16:05:41 -08:00
oobabooga 193548edce Minor fix to ExLlamaV2 requirements 2024-02-13 16:00:06 -08:00
oobabooga 25b655faeb Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2024-02-13 15:49:53 -08:00
oobabooga f99f1fc68e Bump llama-cpp-python to 0.2.42 2024-02-13 15:49:20 -08:00
dependabot[bot] d8081e85ec
Update peft requirement from ==0.7.* to ==0.8.* (#5446) 2024-02-13 16:27:18 -03:00
dependabot[bot] 653b195b1e
Update numpy requirement from ==1.24.* to ==1.26.* (#5490) 2024-02-13 16:26:35 -03:00
dependabot[bot] 147b4cf3e0
Bump hqq from 0.1.2.post1 to 0.1.3 (#5489) 2024-02-13 16:25:02 -03:00
Steven K 512933fa44
Update main.css to allow scrolling in code blocks (#5495) 2024-02-13 16:24:30 -03:00
oobabooga e9fea353c5 Bump llama-cpp-python to 0.2.40 2024-02-13 11:22:34 -08:00
oobabooga 7342afaf19 Update the PyTorch installation instructions 2024-02-08 20:36:11 -08:00
oobabooga 86c320ab5a llama.cpp: add a progress bar for prompt evaluation 2024-02-07 21:56:10 -08:00
oobabooga acea6a6669 Add more exllamav2 wheels 2024-02-07 08:24:29 -08:00
oobabooga 35537ad3d1
Bump exllamav2 to 0.0.13.1 (#5463) 2024-02-07 13:17:04 -03:00
oobabooga b8e25e8678 Bump llama-cpp-python to 0.2.39 2024-02-07 06:50:47 -08:00
oobabooga c55b8ce932 Improved random preset generation 2024-02-06 08:51:52 -08:00
oobabooga 4e34ae0587 Minor logging improvements 2024-02-06 08:22:08 -08:00
oobabooga 3add2376cd Better warpers logging 2024-02-06 07:09:21 -08:00
oobabooga 494cc3c5b0 Handle empty sampler priority field, use default values 2024-02-06 07:05:32 -08:00
oobabooga 0f134bf744
Merge pull request #5453 from oobabooga/dev
Merge dev branch
2024-02-06 11:50:21 -03:00
oobabooga 775902c1f2 Sampler priority: better logging, always save to presets 2024-02-06 06:49:22 -08:00
oobabooga a329db062e
Merge pull request #5452 from oobabooga/dev
Merge dev branch
2024-02-06 11:36:00 -03:00
oobabooga acfbe6b3b3 Minor doc changes 2024-02-06 06:35:01 -08:00
oobabooga 8ee3cea7cb Improve some log messages 2024-02-06 06:31:27 -08:00
oobabooga 8a6d9abb41 Small fixes 2024-02-06 06:26:27 -08:00
oobabooga 2a1063eff5 Revert "Remove non-HF ExLlamaV2 loader (#5431)"
This reverts commit cde000d478.
2024-02-06 06:21:36 -08:00
oobabooga 8c35fefb3b
Add custom sampler order support (#5443) 2024-02-06 11:20:10 -03:00
oobabooga 7301c7618f Minor change to Models tab 2024-02-04 21:49:58 -08:00
oobabooga f234fbe83f Improve a log message after previous commit 2024-02-04 21:44:53 -08:00
oobabooga 7073665a10
Truncate long chat completions inputs (#5439) 2024-02-05 02:31:24 -03:00
oobabooga 9033fa5eee Organize the Model tab 2024-02-04 19:30:22 -08:00
oobabooga cd4ffd3dd4 Update docs 2024-02-04 18:48:04 -08:00
oobabooga 92d0617bce Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2024-02-04 18:40:46 -08:00
oobabooga a210999255 Bump safetensors version 2024-02-04 18:40:25 -08:00
Badis Ghoubali 9fdee65cf5
Improve ChatML template (#5411) 2024-02-04 23:39:15 -03:00
Forkoz 2a45620c85
Split by rows instead of layers for llama.cpp multi-gpu (#5435) 2024-02-04 23:36:40 -03:00
Badis Ghoubali 3df7e151f7
fix the n_batch slider (#5436) 2024-02-04 18:15:30 -03:00
oobabooga 4e188eeb80 Lint 2024-02-03 20:40:10 -08:00
oobabooga cde000d478
Remove non-HF ExLlamaV2 loader (#5431) 2024-02-04 01:15:51 -03:00
kalomaze b6077b02e4
Quadratic sampling (#5403)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2024-02-04 00:20:02 -03:00
oobabooga e98d1086f5
Bump llama-cpp-python to 0.2.38 (#5420) 2024-02-01 20:09:30 -03:00
oobabooga 4f3fdf1b5f
Merge pull request #5404 from oobabooga/dev
Merge dev branch
2024-01-30 14:17:08 -03:00
oobabooga 167ee72d4e Lint 2024-01-30 09:16:23 -08:00
oobabooga ee65f4f014 Downloader: don't assume that huggingface_hub is installed 2024-01-30 09:14:11 -08:00
oobabooga 89f6036e98
Bump llama-cpp-python, remove python 3.8/3.9, cuda 11.7 (#5397) 2024-01-30 13:19:20 -03:00
Forkoz 528318b700
API: Remove tiktoken from logit bias (#5391) 2024-01-28 21:42:03 -03:00
Badis Ghoubali 40c7977f9b
Add roleplay.gbnf grammar (#5368) 2024-01-28 21:41:28 -03:00
smCloudInTheSky b1463df0a1
docker: add options for CPU only, Intel GPU, AMD GPU (#5380) 2024-01-28 11:18:14 -03:00
oobabooga d921f80322 one-click: minor fix after 5e87678fea 2024-01-28 06:14:15 -08:00
Evgenii 26c3ab367e
one-click: use f-strings to improve readability and unify with the rest code (#5068) 2024-01-27 17:31:22 -03:00
Andrew C. Dvorak 5e87678fea
Support running as a git submodule. (#5227) 2024-01-27 17:18:50 -03:00
Hubert Kasperek 69622930c7
Ability to run the Coqui TTS extension on the CPU (#5365) 2024-01-27 17:15:34 -03:00
Anthony Guijarro 828be63f2c
Downloader: use HF get_token function (#5381) 2024-01-27 17:13:09 -03:00
oobabooga e7a760e6b3
Merge pull request #5379 from oobabooga/dev
Merge dev branch
2024-01-26 11:18:45 -03:00
oobabooga de387069da Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2024-01-26 06:12:19 -08:00
sam-ngu c0bdcee646
added trust_remote_code to deepspeed init loaderClass (#5237) 2024-01-26 11:10:57 -03:00
dependabot[bot] bfe2326a24
Bump hqq from 0.1.2 to 0.1.2.post1 (#5349) 2024-01-26 11:10:18 -03:00
oobabooga 70648e75e6 Docs: minor change 2024-01-26 06:00:26 -08:00
oobabooga c1470870bb Update README 2024-01-26 05:58:40 -08:00
oobabooga 87dc421ee8
Bump exllamav2 to 0.0.12 (#5352) 2024-01-22 22:40:12 -03:00
oobabooga 837bd888e4
Merge pull request #5348 from oobabooga/dev
Merge dev branch
2024-01-22 11:18:46 -03:00
oobabooga 1343aa3d33
Merge pull request #5347 from oobabooga/dev
Merge dev branch
2024-01-22 09:44:53 -03:00
oobabooga aa575119e6 API: minor fix 2024-01-22 04:38:43 -08:00
oobabooga 821dd65fb3 API: add a comment 2024-01-22 04:15:51 -08:00
oobabooga 6247eafcc5 API: better handle temperature = 0 2024-01-22 04:12:23 -08:00
oobabooga 817866c9cf Lint 2024-01-22 04:07:25 -08:00
oobabooga b9d1873301 Bump transformers to 4.37 2024-01-22 04:07:12 -08:00
oobabooga aad73667af Lint 2024-01-22 03:25:55 -08:00
oobabooga 6ada77cf5a Update README.md 2024-01-22 03:17:15 -08:00
oobabooga 8b5495ebf8 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2024-01-22 03:15:29 -08:00
oobabooga cc6505df14 Update README.md 2024-01-22 03:14:56 -08:00
Cohee fbf8ae39f8
API: Allow content arrays for multimodal OpenAI requests (#5277) 2024-01-22 08:10:26 -03:00
Ercan 166fdf09f3
API: Properly handle Images with RGBA color format (#5332) 2024-01-22 08:08:51 -03:00
lmg-anon db1da9f98d
Fix logprobs tokens in OpenAI API (#5339) 2024-01-22 08:07:42 -03:00
oobabooga b5cabb6e9d
Bump llama-cpp-python to 0.2.31 (#5345) 2024-01-22 08:05:59 -03:00
oobabooga 8962bb173e
Bump llama-cpp-python to 0.2.29 (#5307) 2024-01-18 14:24:17 -03:00
Stefan Daniel Schwarz 232c07bf1f
API: set do_sample=false when temperature=0 (#5275) 2024-01-17 23:58:11 -03:00
Yiximail 3fef37cda8
UI: Update position of show-controls label to avoid line breaks due to font size (#5256) 2024-01-17 23:56:48 -03:00
oobabooga 7916cf863b Bump transformers (necesary for e055967974) 2024-01-17 12:37:31 -08:00
Forkoz 5c5ef4cef7
UI: change n_gpu_layers maximum to 256 for larger models. (#5262) 2024-01-17 17:13:16 -03:00
ilya sheprut 4d14eb8b82
LoRA: Fix error "Attempting to unscale FP16 gradients" when training (#5268) 2024-01-17 17:11:49 -03:00
Katehuuh 535ea9928a
Fixed whisper README Typo Hyperlinks (#5281) 2024-01-17 17:10:45 -03:00
oobabooga e055967974
Add prompt_lookup_num_tokens parameter (#5296) 2024-01-17 17:09:36 -03:00
oobabooga d8c3a5bee8
Merge pull request #5266 from oobabooga/dev
Merge dev branch (#5257)
2024-01-14 13:31:40 -03:00
Samuel Weinhardt 952a05a7c8
Correct field alias types for OpenAI extension (#5257) 2024-01-14 13:30:36 -03:00
oobabooga 61e4bfe305
Merge pull request #5253 from oobabooga/dev
Merge dev branch
2024-01-13 21:49:32 -03:00
Rimmy J d80b191b1c
Add requirement jinja2==3.1.* to fix error as described in issue #5240 (#5249)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
Co-authored-by: Rim <anonymous@mail.com>
2024-01-13 21:47:13 -03:00
oobabooga e1dd5ee2de UI: focus on the chat input when switching to the chat tab 2024-01-10 05:12:49 -08:00
oobabooga ec2da5adef Docs: document keyboard shortcuts 2024-01-10 03:58:39 -08:00
oobabooga b3fc2cd887 UI: Do not save unchanged extension settings to settings.yaml 2024-01-10 03:48:30 -08:00
oobabooga bb2c4707c4 API: fix bug after previous commit 2024-01-09 19:08:02 -08:00
oobabooga 4332e24740 API: Make user_name/bot_name the official and name1/name2 the alias 2024-01-09 19:06:11 -08:00
oobabooga a4c51b5a05 API: add "user_name" and "bot_name" aliases for name1 and name2 2024-01-09 19:02:45 -08:00
oobabooga 53dc1d8197 UI: Do not save unchanged settings to settings.yaml 2024-01-09 18:59:04 -08:00
oobabooga 2dc8db8aa4
Merge pull request #5220 from oobabooga/dev
Merge dev branch
2024-01-09 21:38:35 -03:00
oobabooga 038b4fc8af Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2024-01-09 16:28:14 -08:00
oobabooga 89e7e107fc Lint 2024-01-09 16:27:50 -08:00
Badis Ghoubali c44836c4d7
Fix spaces in Mistral/Mixtral instruct prompt (#5214) 2024-01-09 21:12:54 -03:00
mamei16 bec4e0a1ce
Fix update event in refresh buttons (#5197) 2024-01-09 14:49:37 -03:00
oobabooga 4333d82b9d Minor bug fix 2024-01-09 06:55:18 -08:00
oobabooga fbce30b09f
Reduce the number of built-in presets (#5217) 2024-01-09 11:50:10 -03:00
oobabooga 953343cced Improve the file saving/deletion menus 2024-01-09 06:33:47 -08:00
oobabooga 123f27a3c5 Load the nearest character after deleting a character
Instead of the first.
2024-01-09 06:24:27 -08:00
oobabooga ba87b9993d Change a label in the gallery extension 2024-01-09 06:06:57 -08:00
oobabooga b908ed318d Revert "Rename past chats -> chat history"
This reverts commit aac93a1fd6.
2024-01-09 05:26:07 -08:00
oobabooga 4ca82a4df9 Save light/dark theme on "Save UI defaults to settings.yaml" 2024-01-09 04:20:10 -08:00
oobabooga 7af50ede94 Reorder some buttons 2024-01-09 04:11:50 -08:00
oobabooga a9f49a7574 Confirm the chat history rename with enter 2024-01-09 04:00:53 -08:00
oobabooga 4d730a759a Focus on the rename text area when it becomes visible 2024-01-09 04:00:47 -08:00
oobabooga 6e9d814095 Change a padding after 4f7e1eeafd 2024-01-09 03:41:31 -08:00
oobabooga 7bdd2118a2 Change some log messages when deleting files 2024-01-09 03:32:01 -08:00
oobabooga aac93a1fd6 Rename past chats -> chat history 2024-01-09 03:14:30 -08:00
oobabooga 615fa11af8 Move new chat button, improve history deletion handling 2024-01-08 21:22:37 -08:00
oobabooga 4f7e1eeafd
Past chat histories in a side bar on desktop (#5098)
Lots of room for improvement, but that's a start.
2024-01-09 01:57:29 -03:00
oobabooga 372ef5e2d8 Fix dynatemp parameters always visible 2024-01-08 19:42:31 -08:00
oobabooga 29c2693ea0
dynatemp_low, dynatemp_high, dynatemp_exponent parameters (#5209) 2024-01-08 23:28:35 -03:00
oobabooga dc1df22a2b
Press Tab to switch between current tab and Parameters tab (#5210) 2024-01-08 23:23:55 -03:00
dependabot[bot] 32cdc66cf1
Bump hqq from 0.1.1.post1 to 0.1.2 (#5204) 2024-01-08 22:51:44 -03:00
oobabooga c4e005efec Fix dropdown menus sometimes failing to refresh 2024-01-08 17:49:54 -08:00
oobabooga 9cd2106303 Revert "Add dynamic temperature to the random preset button"
This reverts commit 4365fb890f.
2024-01-08 16:46:24 -08:00
oobabooga 4365fb890f Add dynamic temperature to the random preset button 2024-01-07 13:08:15 -08:00
oobabooga ad1ff53034
Merge pull request #5199 from oobabooga/dev
Merge dev branch
2024-01-07 17:06:02 -03:00
oobabooga 0d07b3a6a1
Add dynamic_temperature_low parameter (#5198) 2024-01-07 17:03:47 -03:00
oobabooga b8a0b3f925 Don't print torch tensors with --verbose 2024-01-07 10:35:55 -08:00
oobabooga e169993b7a
Merge pull request #5195 from oobabooga/dev
Merge dev branch
2024-01-07 15:12:27 -03:00
oobabooga cf820c69c5 Print generation parameters with --verbose (HF only) 2024-01-07 10:06:23 -08:00
oobabooga c4c7fc4ab3 Lint 2024-01-07 09:36:56 -08:00
Yilong Guo d93db3b486
Refine ipex setup (#5191) 2024-01-07 10:40:30 -03:00
kalomaze 48327cc5c4
Dynamic Temperature HF loader support (#5174)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2024-01-07 10:36:26 -03:00
Philipp Claßen 3eca20c015
Typo fixed in variable names (#5184) 2024-01-06 03:05:03 -03:00
oobabooga 8ea3f31601
Merge pull request #5181 from oobabooga/dev
Merge dev branch
2024-01-05 18:42:30 -03:00
oobabooga 91c2b8e11c Improvements to character_bias extension 2024-01-04 20:48:26 -08:00
oobabooga 248742df1c Save extension fields to settings.yaml on "Save UI defaults" 2024-01-04 20:33:42 -08:00
oobabooga 9e86bea8e9 Use requirements_cpu.txt for intel 2024-01-04 18:52:14 -08:00
oobabooga 3d854ee516
Pin PyTorch version to 2.1 (#5056) 2024-01-04 23:50:23 -03:00
Matthew Raaff c9c31f71b8
Various one-click installer improvements (#4994)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2024-01-04 23:41:54 -03:00
oobabooga c9d814592e Increase maximum temperature value to 5 2024-01-04 17:28:15 -08:00
Guanghua Lu 3bb4b0504e
Close the menu on second click. (#5110) 2024-01-04 13:52:11 -03:00
oobabooga e4d724eb3f Fix cache_folder bug introduced in 37eff915d6 2024-01-04 07:49:40 -08:00
Alberto Cano 37eff915d6
Use --disk-cache-dir for all caches 2024-01-04 00:27:26 -03:00
oobabooga c54d1daaaa
Merge pull request #5163 from oobabooga/dev
Merge dev branch
2024-01-03 22:57:00 -03:00
Lounger 7965f6045e
Fix loading latest history for file names with dots (#5162) 2024-01-03 22:39:41 -03:00
Adam Florizone 894e1a0700
Docker: added build args for non AVX2 CPU (#5154) 2024-01-03 20:43:02 -03:00
AstrisCantCode b80e6365d0
Fix various bugs for LoRA training (#5161) 2024-01-03 20:42:20 -03:00
oobabooga f6a204d7c9 Bump llama-cpp-python to 0.2.26 2024-01-03 11:06:36 -08:00
oobabooga 3a6cba9021 Add top_k=1 to Debug-deterministic preset
Makes it work with llama.cpp
2024-01-02 15:54:56 -08:00
oobabooga 3f28925a8d
Merge pull request #5152 from oobabooga/dev
Merge dev branch
2024-01-02 13:22:14 -03:00
oobabooga 7cce88c403 Rmove an unncecessary exception 2024-01-02 07:20:59 -08:00
oobabooga 90c7e84b01 UI: improve chat style margin for last bot message 2024-01-01 19:50:13 -08:00
oobabooga a4b4708560 Decrease "Show controls" button opacity 2024-01-01 19:08:30 -08:00
oobabooga 94afa0f9cf Minor style changes 2024-01-01 16:00:22 -08:00
oobabooga 3e3a66e721
Merge pull request #5132 from oobabooga/dev
Merge dev branch
2023-12-31 02:32:25 -03:00
oobabooga cbf6f9e695 Update some UI messages 2023-12-30 21:31:17 -08:00
oobabooga 2aad91f3c9
Remove deprecated command-line flags (#5131) 2023-12-31 02:07:48 -03:00
TheInvisibleMage 485b85ee76
Superboogav2 Quick Fixes (#5089) 2023-12-31 02:03:23 -03:00
oobabooga 2734ce3e4c
Remove RWKV loader (#5130) 2023-12-31 02:01:40 -03:00
oobabooga 0e54a09bcb
Remove exllamav1 loaders (#5128) 2023-12-31 01:57:06 -03:00
oobabooga 8e397915c9
Remove --sdp-attention, --xformers flags (#5126) 2023-12-31 01:36:51 -03:00
B611 b7dd1f9542
Specify utf-8 encoding for model metadata file open (#5125) 2023-12-31 01:34:32 -03:00
oobabooga 20a2eaaf95 Add .vs to .gitignore 2023-12-27 12:58:07 -08:00
oobabooga a4079e879e CSS: don't change --chat-height when outside the chat tab 2023-12-27 11:51:55 -08:00
oobabooga c419206ce1 Lint the JS/CSS 2023-12-27 09:59:23 -08:00
oobabooga 3fd7073808
Merge pull request #5100 from oobabooga/dev
Merge dev branch
2023-12-27 13:23:28 -03:00
oobabooga 648c2d1cc2 Update settings-template.yaml 2023-12-25 15:25:16 -08:00
oobabooga c21e3d6300
Merge pull request #5044 from TheLounger/style_improvements
Improve chat styles
2023-12-25 20:00:50 -03:00
oobabooga 2ad6c526b8 Check if extensions block exists before changing it 2023-12-25 14:43:12 -08:00
oobabooga 63553b41ed Improve some paddings 2023-12-25 14:25:31 -08:00
oobabooga abd227594c Fix a border radius 2023-12-25 14:17:00 -08:00
oobabooga 8d0359a6d8 Rename some CSS variables 2023-12-25 14:10:07 -08:00
oobabooga 5466ae59a7 Prevent input/chat area overlap with new --my-delta variable 2023-12-25 14:07:31 -08:00
oobabooga 19d13743a6
Merge pull request #5078 from oobabooga/dev
Merge dev branch
2023-12-25 17:23:01 -03:00
oobabooga 02d063fb9f Fix extra space after 18ca35faaa 2023-12-25 08:38:17 -08:00
oobabooga ae927950a8 Remove instruct style border radius 2023-12-25 08:35:33 -08:00
oobabooga 18ca35faaa Space between chat tab and extensions block 2023-12-25 08:34:02 -08:00
oobabooga 73ba7a8921 Change height -> min-height for .chat 2023-12-25 08:32:02 -08:00
oobabooga 29b0f14d5a
Bump llama-cpp-python to 0.2.25 (#5077) 2023-12-25 12:36:32 -03:00
oobabooga af876095e2
Merge pull request #5073 from oobabooga/dev
Merge dev branch
2023-12-25 02:58:45 -03:00
oobabooga c06f630bcc Increase max_updates_second maximum value 2023-12-24 13:29:47 -08:00
Casper 92d5e64a82
Bump AutoAWQ to 0.1.8 (#5061) 2023-12-24 14:27:34 -03:00
oobabooga 4aeebfc571 Merge branch 'dev' into TheLounger-style_improvements 2023-12-24 09:24:55 -08:00
oobabooga d76b00c211 Pin lm_eval package version 2023-12-24 09:22:31 -08:00
oobabooga 8c60495878 UI: add "Maximum UI updates/second" parameter 2023-12-24 09:17:40 -08:00
zhangningboo 1b8b61b928
Fix output_ids decoding for Qwen/Qwen-7B-Chat (#5045) 2023-12-22 23:11:02 -03:00
kabachuha dbe438564e
Support for sending images into OpenAI chat API (#4827) 2023-12-22 22:45:53 -03:00
Stefan Daniel Schwarz 8956f3ebe2
Synthia instruction templates (#5041) 2023-12-22 22:19:43 -03:00
Yiximail afc91edcb2
Reset the model_name after unloading the model (#5051) 2023-12-22 22:18:24 -03:00
Lounger 554a8f910b Attempt at shrinking chat area when input box grows 2023-12-22 04:51:20 +01:00
oobabooga 4b25acf58f
Merge pull request #5039 from oobabooga/dev
Merge dev branch
2023-12-21 20:22:48 -03:00
Lounger 588b37c032 Add slight padding to top of message container 2023-12-21 22:04:41 +01:00
Lounger 568541aa31 Remove bottom padding on chat tab 2023-12-21 21:48:34 +01:00
oobabooga c1b99f45cb Make --help output instant 2023-12-21 09:32:20 -08:00
Lounger 0dd759c44f Claim more vertical space 2023-12-21 05:42:06 +01:00
Lounger 6fbd64db72 Set borders for all chat styles 2023-12-21 05:00:56 +01:00
oobabooga 2706149c65
Organize the CMD arguments by group (#5027) 2023-12-21 00:33:55 -03:00
oobabooga c727a70572 Remove redundancy from modules/loaders.py 2023-12-20 19:18:07 -08:00
Lounger e3e053ab99 UI: Expand chat vertically and handle header wrapping 2023-12-21 03:42:23 +01:00
Lounger a098c7eee3 Merge branch 'dev' into style_improvements 2023-12-20 23:09:15 +01:00
oobabooga 11288d11d4
Merge pull request #5022 from oobabooga/dev
Merge dev branch
2023-12-20 15:56:04 -03:00
luna 6efbe3009f
let exllama v1 models load safetensor loras (#4854) 2023-12-20 13:29:19 -03:00
oobabooga bcba200790 Fix EOS being ignored in ExLlamav2 after previous commit 2023-12-20 07:54:06 -08:00
oobabooga f0f6d9bdf9 Add HQQ back & update version
This reverts commit 2289e9031e.
2023-12-20 07:46:09 -08:00
oobabooga b15f510154 Optimize ExLlamav2 (non-HF) loader 2023-12-20 07:31:42 -08:00
oobabooga 489f4a23bf
Merge pull request #5012 from oobabooga/dev
Merge dev branch
2023-12-20 02:59:30 -03:00
oobabooga 258c695ead Add rich requirement 2023-12-19 21:58:36 -08:00
oobabooga c1f78dbd0f
Merge pull request #5011 from oobabooga/dev
Merge dev branch
2023-12-20 02:38:25 -03:00
oobabooga fadb295d4d Lint 2023-12-19 21:36:57 -08:00
oobabooga 2289e9031e Remove HQQ from requirements (after https://github.com/oobabooga/text-generation-webui/issues/4993) 2023-12-19 21:33:49 -08:00
oobabooga fb8ee9f7ff Add a specific error if HQQ is missing 2023-12-19 21:32:58 -08:00
oobabooga 366c93a008 Hide a warning 2023-12-19 21:03:20 -08:00
oobabooga 9992f7d8c0 Improve several log messages 2023-12-19 20:54:32 -08:00
oobabooga 23818dc098 Better logger
Credits: vladmandic/automatic
2023-12-19 20:38:33 -08:00
oobabooga 95600073bc Add an informative error when extension requirements are missing 2023-12-19 20:20:45 -08:00
Lounger f9accd38e0 UI: Update chat instruct styles 2023-12-20 02:54:08 +01:00
oobabooga d8279dc710 Replace character name placeholders in chat context (closes #5007) 2023-12-19 17:31:46 -08:00
Lounger ff3e845b04 UI: Header boy is dropping shadows 2023-12-20 01:24:34 +01:00
Lounger 40d5bf6c35 Set margin on other tabs too 2023-12-19 23:42:13 +01:00
Lounger f42074b6c1 UI: Remove header margin on chat tab 2023-12-19 23:27:11 +01:00
oobabooga 5b791cae4a
Merge pull request #5005 from oobabooga/dev
Merge dev branch
2023-12-19 18:21:09 -03:00
oobabooga e83e6cedbe Organize the model menu 2023-12-19 13:18:26 -08:00
oobabooga f4ae0075e8 Fix conversion from old template format to jinja2 2023-12-19 13:16:52 -08:00
oobabooga de138b8ba6
Add llama-cpp-python wheels with tensor cores support (#5003) 2023-12-19 17:30:53 -03:00
oobabooga 71eb744b1c
Merge pull request #5002 from oobabooga/dev
Merge dev branch
2023-12-19 15:24:40 -03:00
oobabooga 0a299d5959
Bump llama-cpp-python to 0.2.24 (#5001) 2023-12-19 15:22:21 -03:00
oobabooga 83cf1a6b67 Fix Yi space issue (closes #4996) 2023-12-19 07:54:19 -08:00
oobabooga 781367bdc3
Merge pull request #4988 from oobabooga/dev
Merge dev branch
2023-12-18 23:42:16 -03:00
oobabooga 9847809a7a Add a warning about ppl evaluation without --no_use_fast 2023-12-18 18:09:24 -08:00
oobabooga f6d701624c UI: mention that QuIP# does not work on Windows 2023-12-18 18:05:02 -08:00
oobabooga a23a004434 Update the example template 2023-12-18 17:47:35 -08:00
oobabooga 3d10c574e7 Fix custom system messages in instruction templates 2023-12-18 17:45:06 -08:00
dependabot[bot] 9e48e50428
Update optimum requirement from ==1.15.* to ==1.16.* (#4986) 2023-12-18 21:43:29 -03:00
俞航 9fa3883630
Add ROCm wheels for exllamav2 (#4973)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-12-18 21:40:38 -03:00
Water 674be9a09a
Add HQQ quant loader (#4888)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-12-18 21:23:16 -03:00
oobabooga b28020a9e4
Merge pull request #4980 from oobabooga/dev
Merge dev branch
2023-12-18 10:11:32 -03:00
oobabooga 64a57d9dc2 Remove duplicate instruction templates 2023-12-17 21:39:47 -08:00
oobabooga 1f9e25e76a UI: update "Saved instruction templates" dropdown after loading template 2023-12-17 21:19:06 -08:00
oobabooga da1c8d77ea Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2023-12-17 21:05:10 -08:00
oobabooga cac89df97b Instruction templates: better handle unwanted bos tokens 2023-12-17 21:04:30 -08:00
oobabooga f0d6ead877
llama.cpp: read instruction template from GGUF metadata (#4975) 2023-12-18 01:51:58 -03:00
oobabooga 3f3cd4fbe4 UI: improve list style in chat modes 2023-12-17 20:26:57 -08:00
oobabooga 306c479d3a Minor fix to Vigogne-Chat template 2023-12-17 19:15:54 -08:00
Hirose 3f973e1fbf
Add detection for Eric Hartford's Dolphin models in models/config.yaml (#4966) 2023-12-17 23:56:34 -03:00
Eve 7c6f39382b
Add Orca-Vicuna instruction template (#4971) 2023-12-17 23:55:23 -03:00
FartyPants (FP HAM) 59da429cbd
Update Training PRO (#4972)
- rolling back safetensors to bi, until it is fixed correctly
- removing the ugly checkpoint detour
2023-12-17 23:54:06 -03:00
oobabooga 7be09836fc
Merge pull request #4961 from oobabooga/dev
Merge dev branch
2023-12-17 12:11:13 -03:00
oobabooga f1f2c4c3f4
Add --num_experts_per_token parameter (ExLlamav2) (#4955) 2023-12-17 12:08:33 -03:00
oobabooga 12690d3ffc
Better HF grammar implementation (#4953) 2023-12-17 02:01:23 -03:00
oobabooga aa200f8723 UI: remove no longer necessary js in Default/Notebook tabs 2023-12-16 19:39:00 -08:00
oobabooga 7a84d7b2da
Instruct style improvements (#4951) 2023-12-16 22:16:26 -03:00
oobabooga 41424907b1 Update README 2023-12-16 16:35:36 -08:00
oobabooga d2ed0a06bf Bump ExLlamav2 to 0.0.11 (adds Mixtral support) 2023-12-16 16:34:15 -08:00
oobabooga 0087dca286 Update README 2023-12-16 12:28:51 -08:00
oobabooga f8079d067d UI: save the sent chat message on "no model is loaded" error 2023-12-16 10:52:41 -08:00
oobabooga 443be391f2
Merge pull request #4937 from oobabooga/dev
Merge dev branch
2023-12-15 12:03:22 -03:00
oobabooga a060908d6c Mixtral Instruct: detect prompt format for llama.cpp loader
Workaround until the tokenizer.chat_template kv field gets implemented
2023-12-15 06:59:15 -08:00
oobabooga 3bbf6c601d AutoGPTQ: Add --disable_exllamav2 flag (Mixtral CPU offloading needs this) 2023-12-15 06:46:13 -08:00
oobabooga 7de10f4c8e Bump AutoGPTQ to 0.6.0 (adds Mixtral support) 2023-12-15 06:18:49 -08:00
oobabooga d0677caf2c Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2023-12-15 04:51:41 -08:00
oobabooga 69ba3cb0d9 Bump openai-whisper requirement (closes #4848) 2023-12-15 04:48:04 -08:00
Song Fuchang 127c71a22a
Update IPEX to 2.1.10+xpu (#4931)
* This will require Intel oneAPI Toolkit 2024.0
2023-12-15 03:19:01 -03:00
oobabooga 85816898f9
Bump llama-cpp-python to 0.2.23 (including Linux ROCm and MacOS >= 12) (#4930) 2023-12-15 01:58:08 -03:00
oobabooga 2cb5b68ad9
Bug fix: when generation fails, save the sent message (#4915) 2023-12-15 01:01:45 -03:00
Felipe Ferreira 11f082e417
[OpenAI Extension] Add more types to Embeddings Endpoint (#4895) 2023-12-15 00:26:16 -03:00
Kim Jaewon e53f99faa0
[OpenAI Extension] Add 'max_logits' parameter in logits endpoint (#4916) 2023-12-15 00:22:43 -03:00
oobabooga eaa1fe67f3
Remove elevenlabs extension (#4928) 2023-12-15 00:00:07 -03:00
oobabooga c3e0fcfc52
Merge pull request #4927 from oobabooga/dev
Merge dev branch
2023-12-14 22:39:08 -03:00
oobabooga f336f8a811 Merge branch 'main' into dev 2023-12-14 17:38:16 -08:00
oobabooga dde7921057 One-click installer: minor message change 2023-12-14 17:27:32 -08:00
oobabooga fd1449de20 One-click installer: fix minor bug introduced in previous commit 2023-12-14 16:52:44 -08:00
oobabooga 4ae2dcebf5 One-click installer: more friendly progress messages 2023-12-14 16:48:00 -08:00
oobabooga 8acecf3aee Bump llama-cpp-python to 0.2.23 (NVIDIA & CPU-only, no AMD, no Metal) (#4924) 2023-12-14 09:41:36 -08:00
oobabooga 8835ea3704
Bump llama-cpp-python to 0.2.23 (NVIDIA & CPU-only, no AMD, no Metal) (#4924) 2023-12-14 14:39:43 -03:00
oobabooga bf68d4499e
Merge pull request #4923 from oobabooga/dev
Merge dev branch
2023-12-14 13:01:05 -03:00
oobabooga 623c92792a Update README 2023-12-14 07:56:48 -08:00
oobabooga 3580bed041 Update README 2023-12-14 07:54:20 -08:00
oobabooga e91c09b8af
Merge pull request #4920 from oobabooga/dev
Merge dev branch
2023-12-14 11:24:00 -03:00
oobabooga d5ec3c3444 Update README 2023-12-14 06:20:52 -08:00
oobabooga 5b283fff22 Update README 2023-12-14 06:15:14 -08:00
oobabooga 958799221f Update README 2023-12-14 06:09:03 -08:00
oobabooga e7fa17740a Update README 2023-12-13 22:49:42 -08:00
oobabooga 03babe7d81 Update README 2023-12-13 22:47:08 -08:00
oobabooga aad14174e4 Update README 2023-12-13 22:46:18 -08:00
oobabooga 783947a2aa Update README 2023-12-13 22:44:25 -08:00
oobabooga 7fef16950f Update README 2023-12-13 22:42:54 -08:00
oobabooga d36e7f1762 Update README 2023-12-13 22:35:22 -08:00
oobabooga 9695db0ee4 Update README 2023-12-13 22:30:31 -08:00
oobabooga d354f5009c Update README 2023-12-13 22:21:29 -08:00
oobabooga 0a4fad2d46 Update README 2023-12-13 22:20:37 -08:00
oobabooga fade6abfe9 Update README 2023-12-13 22:18:40 -08:00
oobabooga aafd15109d Update README 2023-12-13 22:15:58 -08:00
oobabooga 634518a412 Update README 2023-12-13 22:08:41 -08:00
oobabooga 0d5ca05ab9 Update README 2023-12-13 22:06:04 -08:00
oobabooga d241de86c4 Update README 2023-12-13 22:02:26 -08:00
Lounger 5754f0c357
Fix deleting chat logs (#4914) 2023-12-13 21:54:43 -03:00
Bartowski f51156705d
Allow symlinked folder within root directory (#4863) 2023-12-13 18:08:21 -03:00
oobabooga 36e850fe89
Update README.md 2023-12-13 17:55:41 -03:00
oobabooga 3e0c11a758
Merge pull request #4912 from oobabooga/dev
Merge dev branch
2023-12-13 15:49:36 -03:00
oobabooga 1bfee1d12e Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2023-12-13 10:48:34 -08:00
oobabooga d14d4cad4a Lint 2023-12-13 10:48:15 -08:00
Ixion 3f3960dbfb
Fixed invalid Jinja2 syntax in instruction templates (#4911) 2023-12-13 15:46:23 -03:00
oobabooga 4eeac70af7 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2023-12-13 10:40:41 -08:00
oobabooga fcf5512364 Jinja templates: fix a potential small bug 2023-12-13 10:19:39 -08:00
missionfloyd bdcc769e6f
Bypass coqui TTS EULA check (#4905) 2023-12-13 02:26:46 -03:00
oobabooga 7f1a6a70e3 Update the llamacpp_HF comment 2023-12-12 21:04:20 -08:00
oobabooga 314a095c74
Merge pull request #4903 from oobabooga/dev
Merge dev branch
2023-12-12 23:10:45 -03:00
oobabooga c2802bc3ac Lint 2023-12-12 18:05:10 -08:00
oobabooga b2cae6cac6 Docs: minor update 2023-12-12 14:11:13 -08:00
oobabooga 21a5bfc67f Relax optimum requirement 2023-12-12 14:05:58 -08:00
oobabooga 12f58e2cac Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2023-12-12 13:28:24 -08:00
oobabooga 1c531a3713 Minor cleanup 2023-12-12 13:25:21 -08:00
Penagwin 85a1d8965c
Updated Docker Docs (#4900) 2023-12-12 18:03:50 -03:00
oobabooga 8513028968 Fix lag in the chat tab during streaming 2023-12-12 13:01:25 -08:00
oobabooga 736fe4aa3e Fix server refusing to close on Ctrl+C 2023-12-12 12:27:40 -08:00
oobabooga 39d2fe1ed9
Jinja templates for Instruct and Chat (#4874) 2023-12-12 17:23:14 -03:00
oobabooga aab0dd962d Revert "Update callbacks.py to show tracebacks on ValueError (#4892)"
This reverts commit 993ca51a65.
2023-12-12 11:47:11 -08:00
dependabot[bot] 7a987417bb
Bump optimum from 1.14.0 to 1.15.0 (#4885) 2023-12-12 02:32:19 -03:00
dependabot[bot] a17750db91
Update peft requirement from ==0.6.* to ==0.7.* (#4886) 2023-12-12 02:31:30 -03:00
dependabot[bot] a8a92c6c87
Update transformers requirement from ==4.35.* to ==4.36.* (#4882) 2023-12-12 02:30:25 -03:00
Nehereus 993ca51a65
Update callbacks.py to show tracebacks on ValueError (#4892) 2023-12-12 02:29:27 -03:00
Morgan Schweers 602b8c6210
Make new browser reloads recognize current model. (#4865) 2023-12-11 02:51:01 -03:00
oobabooga 8c8825b777 Add QuIP# to README 2023-12-08 08:40:42 -08:00
oobabooga 2a335b8aa7 Cleanup: set shared.model_name only once 2023-12-08 06:35:23 -08:00
oobabooga 62d59a516f Add trust_remote_code to all HF loaders 2023-12-08 06:29:26 -08:00
oobabooga 705f04a0c9
Merge pull request #4851 from oobabooga/dev
Merge dev branch
2023-12-08 10:25:57 -03:00
oobabooga 181743fd97 Fix missing spaces tokenizer issue (closes #4834) 2023-12-08 05:16:46 -08:00
oobabooga 884871c107
Merge pull request #4849 from oobabooga/dev
Merge dev branch
2023-12-08 10:05:02 -03:00
oobabooga 00aedf9209 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2023-12-08 05:02:25 -08:00
oobabooga 7bbe7e803a Minor fix 2023-12-08 05:01:25 -08:00
Yiximail 1c74b3ab45
Fix partial unicode characters issue (#4837) 2023-12-08 09:50:53 -03:00
oobabooga 2c5a1e67f9
Parameters: change max_new_tokens & repetition_penalty_range defaults (#4842) 2023-12-07 20:04:52 -03:00
Song Fuchang e16e5997ef
Update IPEX install URL. (#4825)
* Old pip url no longer works. Use the latest url from
  * https://intel.github.io/intel-extension-for-pytorch/index.html#installation
2023-12-06 21:07:01 -03:00
oobabooga d516815c9c Model downloader: download only fp16 if both fp16 and GGUF are present 2023-12-05 21:09:12 -08:00
oobabooga 98361af4d5
Add QuIP# support (#4803)
It has to be installed manually for now.
2023-12-06 00:01:01 -03:00
oobabooga 6430acadde Minor bug fix after https://github.com/oobabooga/text-generation-webui/pull/4814 2023-12-05 10:08:11 -08:00
oobabooga c21a9668a5 Lint 2023-12-04 21:17:05 -08:00
erew123 f786aa3caa
Clean-up Ctrl+C Shutdown (#4802) 2023-12-05 02:16:16 -03:00
oobabooga 2694ef45a3 Do not limit API updates/second 2023-12-04 20:46:18 -08:00
oobabooga 0f828ea441 Do not limit API updates/second 2023-12-04 20:45:43 -08:00
oobabooga af261e5dd4
Merge pull request #4815 from oobabooga/dev
Merge dev branch
2023-12-05 01:30:57 -03:00
oobabooga 9edb193def
Optimize HF text generation (#4814) 2023-12-05 00:00:40 -03:00
oobabooga 1ccbcb967e
Merge pull request #4811 from oobabooga/dev
Merge dev branch
2023-12-04 21:29:45 -03:00
俞航 ac9f154bcc
Bump exllamav2 from 0.0.8 to 0.0.10 & Fix code change (#4782) 2023-12-04 21:15:05 -03:00
oobabooga 131a5212ce UI: update context upper limit to 200000 2023-12-04 15:48:34 -08:00
oobabooga f7145544f9 Update README 2023-12-04 15:44:44 -08:00
oobabooga 8e1f86a866 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2023-12-04 15:41:56 -08:00
oobabooga be88b072e9 Update --loader flag description 2023-12-04 15:41:25 -08:00
dependabot[bot] 801ba87c68
Update accelerate requirement from ==0.24.* to ==0.25.* (#4810) 2023-12-04 20:36:01 -03:00
oobabooga 7fc9033b2e Recommend ExLlama_HF and ExLlamav2_HF 2023-12-04 15:28:46 -08:00
oobabooga e4e35f357b
Merge pull request #4807 from oobabooga/dev
Merge dev branch
2023-12-04 12:28:34 -03:00
oobabooga 3f993280e4 Minor changes 2023-12-04 07:27:44 -08:00
oobabooga 0931ed501b Minor changes 2023-12-04 07:25:18 -08:00
oobabooga 427a165597 Bump TTS version in coqui_tts 2023-12-04 07:21:56 -08:00
Song Fuchang 0bfd5090be
Import accelerate very early to make Intel GPU happy (#4704) 2023-12-03 22:51:18 -03:00
dependabot[bot] 2e83844f35
Bump safetensors from 0.4.0 to 0.4.1 (#4750) 2023-12-03 22:50:10 -03:00
Ikko Eltociear Ashimine 06cc9a85f7
README: minor typo fix (#4793) 2023-12-03 22:46:34 -03:00
Lounger 7c0a17962d
Gallery improvements (#4789) 2023-12-03 22:45:50 -03:00
oobabooga 96df4f10b9
Merge pull request #4777 from oobabooga/dev
Merge dev branch
2023-12-01 00:00:17 -03:00
oobabooga 77d6ccf12b Add a LOADER debug message while loading models 2023-11-30 12:00:32 -08:00
oobabooga 1c90e02243 Update Colab-TextGen-GPU.ipynb 2023-11-30 11:55:18 -08:00
oobabooga 092a2c3516 Fix a bug in llama.cpp get_logits() function 2023-11-30 11:21:40 -08:00
oobabooga 6d3a9b8689
Merge pull request #4773 from oobabooga/dev
Merge dev branch
2023-11-30 02:31:37 -03:00
oobabooga 000b77a17d Minor docker changes 2023-11-29 21:27:23 -08:00
Callum 88620c6b39
feature/docker_improvements (#4768) 2023-11-30 02:20:23 -03:00
oobabooga 2698d7c9fd Fix llama.cpp model unloading 2023-11-29 15:19:48 -08:00
oobabooga fa89d305e3 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2023-11-29 15:13:17 -08:00
oobabooga 9940ed9c77 Sort the loaders 2023-11-29 15:13:03 -08:00
Manu Kashyap 78fd7f6aa8
Fixed naming for sentence-transformers library (#4764) 2023-11-29 12:15:03 -03:00
oobabooga a7670c31ca Sort 2023-11-28 18:43:33 -08:00
oobabooga 6e51bae2e0 Sort the loaders menu 2023-11-28 18:41:11 -08:00
oobabooga f4b956b47c Detect yi instruction template 2023-11-27 10:45:47 -08:00
oobabooga 68059d7c23 llama.cpp: minor log change & lint 2023-11-27 10:44:55 -08:00
Denis Iskandarov 1b05832f9a
Add direnv artifacts to gitignore (#4737) 2023-11-27 15:43:42 -03:00
xr4dsh b5b3d18773
resonable cli args for docker container (#4727) 2023-11-27 15:43:01 -03:00
tsukanov-as 9f7ae6bb2e
fix detection of stopping strings when HTML escaping is used (#4728) 2023-11-27 15:42:08 -03:00
Eve d06ce7b75c
add openhermes mistral support (#4730) 2023-11-27 15:41:06 -03:00
oobabooga b6d16a35b1 Minor API fix 2023-11-21 17:56:28 -08:00
oobabooga 51add248c8
Merge pull request #4702 from oobabooga/dev
Merge dev branch
2023-11-21 21:18:27 -03:00
oobabooga cb0dbffccc Merge branch 'main' into dev 2023-11-21 16:12:45 -08:00
oobabooga 8d811a4d58 one-click: move on instead of crashing if extension fails to install 2023-11-21 16:09:44 -08:00
oobabooga 0589ff5b12
Bump llama-cpp-python to 0.2.19 & add min_p and typical_p parameters to llama.cpp loader (#4701) 2023-11-21 20:59:39 -03:00
oobabooga 2769a1fa25 Hide deprecated args from Session tab 2023-11-21 15:15:16 -08:00
oobabooga 0047d9f5e0 Do not install coqui_tts requirements by default
It breaks the one-click installer on Windows.
2023-11-21 15:13:42 -08:00
oobabooga fb124ab6e2 Bump to flash-attention 2.3.4 + switch to Github Actions wheels on Windows (#4700) 2023-11-21 15:07:17 -08:00
oobabooga e9cdaa2ada
Bump to flash-attention 2.3.4 + switch to Github Actions wheels on Windows (#4700) 2023-11-21 20:06:56 -03:00
oobabooga b81d6ad8a4
Detect Orca 2 template (#4697) 2023-11-21 15:26:42 -03:00
oobabooga 360eeb9ff1
Merge pull request #4686 from oobabooga/dev
Merge dev branch
2023-11-21 08:38:50 -03:00
oobabooga 54a4eb60a3
Remove --no-dependencies from TTS installation command 2023-11-21 08:30:50 -03:00
oobabooga efdd99623c
Merge pull request #4683 from oobabooga/dev
Merge dev branch
2023-11-21 00:36:58 -03:00
oobabooga b02dc4dc0d Add --no-dependencies to TTS installation command 2023-11-20 19:02:12 -08:00
oobabooga 55f2a3643b Update multimodal API example 2023-11-20 18:41:09 -08:00
oobabooga 829c6d4f78 Add "remove_trailing_dots" option to XTTSv2 2023-11-20 18:33:29 -08:00
kanttouchthis 8dc9ec3491
add XTTSv2 (coqui_tts extension) (#4673)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-11-20 22:37:52 -03:00
oobabooga ff24648510 Credit llama-cpp-python in the README 2023-11-20 12:13:15 -08:00
oobabooga be78d79811 Revert accidental noavx2 changes 2023-11-20 11:48:04 -08:00
oobabooga 4b84e45116 Use +cpuavx2 instead of +cpuavx 2023-11-20 11:46:38 -08:00
oobabooga d7f1bc102b
Fix "Illegal instruction" bug in llama.cpp CPU only version (#4677) 2023-11-20 16:36:38 -03:00
drew9781 5e70263e25
docker: install xformers with sepcific cuda version, matching the docker image. (#4670) 2023-11-19 21:43:15 -03:00
oobabooga f11092ac2a
Merge pull request #4664 from oobabooga/dev
Merge dev branch
2023-11-19 15:12:55 -03:00
oobabooga f0d66cf817 Add missing file 2023-11-19 10:12:13 -08:00
oobabooga 22e7a22d1e
Merge pull request #4662 from oobabooga/dev
Merge dev branch
2023-11-19 14:23:19 -03:00
oobabooga a2e6d00128 Use convert_ids_to_tokens instead of decode in logits endpoint
This preserves the llama tokenizer spaces.
2023-11-19 09:22:08 -08:00
oobabooga d1bba48a83
Merge pull request #4660 from oobabooga/dev
Merge dev branch
2023-11-19 13:32:08 -03:00
oobabooga 8cf05c1b31 Fix disappearing character gallery 2023-11-19 08:31:01 -08:00
oobabooga 9da7bb203d Minor LoRA bug fix 2023-11-19 07:59:29 -08:00
oobabooga 78af3b0a00 Update docs/What Works.md 2023-11-19 07:57:16 -08:00
oobabooga a6f1e1bcc5 Fix PEFT LoRA unloading 2023-11-19 07:55:25 -08:00
oobabooga a290d17386 Add hover cursor to bot pfp 2023-11-19 06:56:42 -08:00
oobabooga ab94f0d9bf Minor style change 2023-11-18 21:11:04 -08:00
oobabooga 5fcee696ea
New feature: enlarge character pictures on click (#4654) 2023-11-19 02:05:17 -03:00
Jordan Tucker cb836dd49c
fix: use shared chat-instruct_command with api (#4653) 2023-11-19 01:19:10 -03:00
oobabooga 771e62e476
Add /v1/internal/lora endpoints (#4652) 2023-11-19 00:35:22 -03:00
oobabooga ef6feedeb2
Add --nowebui flag for pure API mode (#4651) 2023-11-18 23:38:39 -03:00
oobabooga 0fa1af296c
Add /v1/internal/logits endpoint (#4650) 2023-11-18 23:19:31 -03:00
oobabooga 8f4f4daf8b
Add --admin-key flag for API (#4649) 2023-11-18 22:33:27 -03:00
wizd af76fbedb8
Openai embedding fix to support jina-embeddings-v2 (#4642) 2023-11-18 20:24:29 -03:00
Jordan Tucker baab894759
fix: use system message in chat-instruct mode (#4648) 2023-11-18 20:20:13 -03:00
oobabooga 47d9e2618b Refresh the Preset menu after saving a preset 2023-11-18 14:03:42 -08:00
oobabooga 83b64e7fc1
New feature: "random preset" button (#4647) 2023-11-18 18:31:41 -03:00
oobabooga d1a58da52f Update ancient Docker instructions 2023-11-17 19:52:53 -08:00
oobabooga e0ca49ed9c
Bump llama-cpp-python to 0.2.18 (2nd attempt) (#4637)
* Update requirements*.txt

* Add back seed
2023-11-18 00:31:27 -03:00
oobabooga 3146124ec0
Merge pull request #4632 from oobabooga/dev
Merge dev branch
2023-11-17 10:18:31 -03:00
oobabooga 9d6f79db74 Revert "Bump llama-cpp-python to 0.2.18 (#4611)"
This reverts commit 923c8e25fb.
2023-11-17 05:14:25 -08:00
oobabooga e0a7cc5e0f Simplify CORS code 2023-11-16 20:11:55 -08:00
oobabooga 13dc3b61da Update README 2023-11-16 19:57:55 -08:00
oobabooga 8b66d83aa9 Set use_fast=True by default, create --no_use_fast flag
This increases tokens/second for HF loaders.
2023-11-16 19:55:28 -08:00
oobabooga f889302d24
Merge pull request #4628 from oobabooga/dev
Merge dev branch
2023-11-16 23:47:07 -03:00
oobabooga b2ce8dc7ee Update a message 2023-11-16 18:46:26 -08:00
oobabooga 0ee8d2b66b
Merge pull request #4627 from oobabooga/dev
Merge dev branch
2023-11-16 23:41:18 -03:00
oobabooga 780b00e1cf Minor bug fix 2023-11-16 18:39:39 -08:00
oobabooga c0233bb9d3 Minor message change 2023-11-16 18:36:57 -08:00
oobabooga 94b7177174 Update docs/07 - Extensions 2023-11-16 18:24:46 -08:00
oobabooga 6525707a7f Fix "send instruction template to..." buttons (closes #4625) 2023-11-16 18:16:42 -08:00
oobabooga 510a01ef46 Lint 2023-11-16 18:03:06 -08:00
oobabooga 923c8e25fb
Bump llama-cpp-python to 0.2.18 (#4611) 2023-11-16 22:55:14 -03:00
Casper 61f429563e
Bump AutoAWQ to 0.1.7 (#4620) 2023-11-16 17:08:08 -03:00
oobabooga e7d460d932 Make sure that API requirements are installed 2023-11-16 10:08:41 -08:00
oobabooga cbf2b47476 Strip trailing "\" characters in CMD_FLAGS.txt 2023-11-16 09:33:36 -08:00
oobabooga 58c6001be9 Add missing exllamav2 samplers 2023-11-16 07:09:40 -08:00
oobabooga cd41f8912b Warn users about n_ctx / max_seq_len 2023-11-15 18:56:42 -08:00
oobabooga a475aa7816 Improve API documentation 2023-11-15 18:39:08 -08:00
oobabooga 9be48e83a9 Start API when "api" checkbox is checked 2023-11-15 16:35:47 -08:00
oobabooga a85ce5f055 Add more info messages for truncation / instruction template 2023-11-15 16:20:31 -08:00
oobabooga 883701bc40 Alternative solution to 025da386a0
Fixes an error.
2023-11-15 16:04:02 -08:00
oobabooga 8ac942813c Revert "Fix CPU memory limit error (issue #3763) (#4597)"
This reverts commit 025da386a0.
2023-11-15 16:01:54 -08:00
oobabooga e6f44d6d19 Print context length / instruction template to terminal when loading models 2023-11-15 16:00:51 -08:00
oobabooga e05d8fd441 Style changes 2023-11-15 15:51:37 -08:00
oobabooga be125e2708 Add /v1/internal/model/unload endpoint 2023-11-15 15:48:33 -08:00
David Nielson 564d0cde82
Use standard hyphens in filenames (#4576) 2023-11-15 20:29:00 -03:00
Andy Bao 025da386a0
Fix CPU memory limit error (issue #3763) (#4597)
get_max_memory_dict() was not properly formatting shared.args.cpu_memory

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-11-15 20:27:20 -03:00
Anton Rogozin 8a9d5a0cea
update AutoGPTQ to higher version for lora applying error fixing (#4604) 2023-11-15 20:23:22 -03:00
oobabooga 8a2af87d3a
Merge pull request #4608 from oobabooga/dev
Merge dev branch
2023-11-15 13:19:15 -03:00
oobabooga 072cfe19e9 Minor Colab fix 2023-11-15 08:18:32 -08:00
oobabooga 2337aebe4d
Merge pull request #4606 from oobabooga/dev
Merge dev branch
2023-11-15 13:16:44 -03:00
oobabooga 3d861a459d Minor Colab fix 2023-11-15 08:15:43 -08:00
oobabooga dea90c7b67 Bump exllamav2 to 0.0.8 2023-11-13 10:34:10 -08:00
oobabooga 454fcf39a9
Merge pull request #4579 from oobabooga/dev
Merge dev branch
2023-11-13 11:39:08 -03:00
oobabooga 4f9bc63edf Installer: update a message for clarity 2023-11-10 09:43:02 -08:00
oobabooga 74fee4f312 Update Colab-TextGen-GPU.ipynb 2023-11-10 09:18:25 -08:00
oobabooga 52758f15da Remove sentence-transformers requirement (for #1575) 2023-11-10 07:35:29 -08:00
oobabooga c5be3f7acb Make /v1/embeddings functional, add request/response types 2023-11-10 07:34:27 -08:00
oobabooga 7ed2143cd6
Update 12 - OpenAI API.md 2023-11-10 11:56:04 -03:00
oobabooga 0777b0d3c7 Add system_message parameter, document model (unused) parameter 2023-11-10 06:47:18 -08:00
oobabooga 4aabff3728 Remove old API, launch OpenAI API with --api 2023-11-10 06:39:08 -08:00
GuizzyQC 6a7cd01ebf
Fix bug with /internal/model/load (#4549)
Update shared.model_name after loading model through API call
2023-11-10 00:16:38 -03:00
oobabooga 2af7e382b1 Revert "Bump llama-cpp-python to 0.2.14"
This reverts commit 5c3eb22ce6.

The new version has issues:

https://github.com/oobabooga/text-generation-webui/issues/4540
https://github.com/abetlen/llama-cpp-python/issues/893
2023-11-09 10:02:13 -08:00
oobabooga 07d66e45b4
Merge pull request #4541 from oobabooga/dev
Merge dev branch
2023-11-09 14:53:34 -03:00
Ashley Kleynhans 372d712921
Fix deprecated API (#4539) 2023-11-09 14:51:50 -03:00
oobabooga d86f1fd2c3 OpenAI API: stop streaming on client disconnect (closes #4521) 2023-11-09 06:37:32 -08:00
oobabooga f7534b2f4b
Merge pull request #4532 from oobabooga/dev
Merge dev branch
2023-11-09 09:33:55 -03:00
oobabooga effb3aef42 Prevent deadlocks in OpenAI API with simultaneous requests 2023-11-08 20:55:39 -08:00
oobabooga 4da00b6032
Merge pull request #4522 from oobabooga/dev
Merge dev branch
2023-11-08 22:57:08 -03:00
oobabooga 21ed9a260e Document the new "Custom system message" field 2023-11-08 17:54:10 -08:00
oobabooga 678fd73aef Document /v1/internal/model/load and fix a bug 2023-11-08 17:41:12 -08:00
MrMojoR 1754a3761b
Include trust remote code usage in openai api's embedder (#4513) 2023-11-08 11:25:43 -03:00
hronoas 6c7aad11f3
openai extension: wrong frequency_penalty type (#4512) 2023-11-08 11:23:51 -03:00
oobabooga 881e8a6e70
Small bug fix in /v1/internal/model/load 2023-11-08 02:34:13 -03:00
oobabooga 050ff36bd6 Revert "Add a comment to /v1/models"
This reverts commit 38b07493a0.
2023-11-07 21:09:47 -08:00
oobabooga 38b07493a0 Add a comment to /v1/models 2023-11-07 21:07:12 -08:00
oobabooga 2358706453 Add /v1/internal/model/load endpoint (tentative) 2023-11-07 20:58:06 -08:00
oobabooga 43c53a7820 Refactor the /v1/models endpoint 2023-11-07 19:59:27 -08:00
oobabooga 1b69694fe9 Add types to the encode/decode/token-count endpoints 2023-11-07 19:32:14 -08:00
oobabooga f6ca9cfcdc Add /v1/internal/model-info endpoint 2023-11-07 18:59:02 -08:00
oobabooga 6e2e0317af
Separate context and system message in instruction formats (#4499) 2023-11-07 20:02:58 -03:00
oobabooga 322c170566 Document logits_all 2023-11-07 14:45:11 -08:00
oobabooga 5c0559da69 Training: fix .txt files now showing in dropdowns 2023-11-07 14:41:11 -08:00
oobabooga af3d25a503 Disable logits_all in llamacpp_HF (makes processing 3x faster) 2023-11-07 14:35:48 -08:00
oobabooga 5c3eb22ce6 Bump llama-cpp-python to 0.2.14 2023-11-07 14:20:43 -08:00
oobabooga 3fc505dc0f Document unused parameters 2023-11-07 08:56:09 -08:00
oobabooga 3d59346871 Implement echo/suffix parameters 2023-11-07 08:43:45 -08:00
oobabooga cee099f131 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2023-11-07 08:25:22 -08:00
oobabooga 48c9c31440 Document the "preset" option in the API 2023-11-07 08:23:17 -08:00
oobabooga d59f1ad89a
Update README.md 2023-11-07 13:05:06 -03:00
oobabooga 0c440877de
Update 12 - OpenAI API.md 2023-11-07 12:59:40 -03:00
oobabooga 55dc9845cb
Update 12 - OpenAI API.md 2023-11-07 12:51:41 -03:00
oobabooga b0b999dd68 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2023-11-07 07:46:08 -08:00
oobabooga 2bda1a9c9b Mention --api-key 2023-11-07 07:45:55 -08:00
oobabooga cc04abda49
Update 12 - OpenAI API.md 2023-11-07 12:40:52 -03:00
oobabooga ddca6948b2
Update 12 - OpenAI API.md 2023-11-07 12:39:59 -03:00
oobabooga 40e73aafce
Update 12 - OpenAI API.md 2023-11-07 12:38:39 -03:00
oobabooga 6ec997f195
Update 12 - OpenAI API.md 2023-11-07 12:36:52 -03:00
oobabooga 15d4ea180d Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2023-11-07 07:35:36 -08:00
oobabooga b2afdda4e8 Add more API examples 2023-11-07 07:35:04 -08:00
Morgan Cheng 349604458b
Update 12 - OpenAI API.md (#4501)
Fix the typo in argument. It should be `--api-port` instead of `--port`.

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-11-07 11:22:17 -03:00
dependabot[bot] fd893baba1
Bump optimum from 1.13.1 to 1.14.0 (#4492) 2023-11-07 00:13:41 -03:00
dependabot[bot] 18739c8b3a
Update peft requirement from ==0.5.* to ==0.6.* (#4494) 2023-11-07 00:12:59 -03:00
oobabooga 79b3f5a546
Add /v1/internal/stop-generation to OpenAI API (#4498) 2023-11-07 00:10:42 -03:00
oobabooga 97c21e5667 Don't strip leading spaces in OpenAI API 2023-11-06 19:09:41 -08:00
oobabooga 4a45dc4041 Reorder the parameters in the FastAPI documentation 2023-11-06 09:55:36 -08:00
oobabooga 1fba6db69f
Merge pull request #4488 from oobabooga/dev
Merge dev branch
2023-11-06 12:18:55 -03:00
oobabooga 0ed6a17ed4 Update warning 2023-11-06 07:17:49 -08:00
oobabooga 0db81355bc Reorder a parameter 2023-11-06 07:11:49 -08:00
oobabooga b87c6213ae Remove obsolete endpoint 2023-11-06 05:45:45 -08:00
oobabooga fcc9114b58 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2023-11-06 05:38:47 -08:00
oobabooga ceb8c92dfc
Update 12 - OpenAI API.md 2023-11-06 10:38:22 -03:00
oobabooga 28fd535f9c Make chat API more robust 2023-11-06 05:22:01 -08:00
oobabooga 5b5ef57049 Remove file 2023-11-05 21:39:59 -08:00
oobabooga ec17a5d2b7
Make OpenAI API the default API (#4430) 2023-11-06 02:38:29 -03:00
俞航 84d957ba62
[Fix] fix openai embedding_model loading as str (#4147) 2023-11-05 20:42:45 -03:00
kabachuha e18a0460d4
fix openai extension not working because of absent new defaults (#4477) 2023-11-04 16:12:51 -03:00
oobabooga b7a409ef57
Merge pull request #4476 from oobabooga/dev
Merge dev branch
2023-11-04 15:04:43 -03:00
oobabooga fb3bd0203d Update docs 2023-11-04 11:02:24 -07:00
oobabooga 1d8c7c1fc4 Update docs 2023-11-04 11:01:15 -07:00
oobabooga b5c53041b8
Merge pull request #4475 from oobabooga/dev
Merge dev branch
2023-11-04 14:19:55 -03:00
oobabooga 40f7f37009 Update requirements 2023-11-04 10:12:06 -07:00
Orang 2081f43ac2
Bump transformers to 4.35.* (#4474) 2023-11-04 14:00:24 -03:00
feng lui 4766a57352
transformers: add use_flash_attention_2 option (#4373) 2023-11-04 13:59:33 -03:00
wouter van der plas add359379e
fixed two links in the ui (#4452) 2023-11-04 13:41:42 -03:00
Casper cfbd108826
Bump AWQ to 0.1.6 (#4470) 2023-11-04 13:09:41 -03:00
oobabooga aa5d671579
Add temperature_last parameter (#4472) 2023-11-04 13:09:07 -03:00
oobabooga 1ab8700d94 Change frequency/presence penalty ranges 2023-11-03 17:38:19 -07:00
oobabooga 45fcb60e7a Make truncation_length_max apply to max_seq_len/n_ctx 2023-11-03 11:29:31 -07:00
oobabooga 7f9c1cbb30 Change min_p default to 0.0 2023-11-03 08:25:22 -07:00
oobabooga 4537853e2c Change min_p default to 1.0 2023-11-03 08:13:50 -07:00
kalomaze 367e5e6e43
Implement Min P as a sampler option in HF loaders (#4449) 2023-11-02 16:32:51 -03:00
oobabooga fcb7017b7a Remove a checkbox 2023-11-02 12:24:09 -07:00
Julien Chaumond fdcaa955e3
transformers: Add a flag to force load from safetensors (#4450) 2023-11-02 16:20:54 -03:00
oobabooga c0655475ae Add cache_8bit option 2023-11-02 11:23:04 -07:00
oobabooga 42f816312d Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2023-11-02 11:09:26 -07:00
oobabooga 77abd9b69b Add no_flash_attn option 2023-11-02 11:08:53 -07:00
Julien Chaumond a56ef2a942
make torch.load a bit safer (#4448) 2023-11-02 14:07:08 -03:00
deevis deba039c03
(fix): OpenOrca-Platypus2 models should use correct instruction_template and custom_stopping_strings (#4435) 2023-11-01 01:51:00 -03:00
Mehran Ziadloo aaf726dbfb
Updating the shared settings object when loading a model (#4425) 2023-11-01 01:29:57 -03:00
oobabooga 9bd0724d85 Change frequency/presence penalty ranges 2023-10-31 20:57:56 -07:00
Orang 6b7fa45cc3
Update exllamav2 version (#4417) 2023-10-31 19:12:14 -03:00
Casper 41e159e88f
Bump AutoAWQ to v0.1.5 (#4410) 2023-10-31 19:11:22 -03:00
Meheret 0707ed7677
updated wiki link (#4415) 2023-10-31 19:09:05 -03:00
oobabooga 262f8ae5bb Use default gr.Dataframe for evaluation table 2023-10-27 06:49:14 -07:00
James Braza f481ce3dd8
Adding platform_system to autoawq (#4390) 2023-10-27 01:02:28 -03:00
dependabot[bot] af98587580
Update accelerate requirement from ==0.23.* to ==0.24.* (#4400) 2023-10-27 00:46:16 -03:00
oobabooga 839a87bac8 Fix is_ccl_available & is_xpu_available imports 2023-10-26 20:27:04 -07:00
Abhilash Majumder 778a010df8
Intel Gpu support initialization (#4340) 2023-10-26 23:39:51 -03:00
GuizzyQC 317e2c857e
sd_api_pictures: fix Gradio warning message regarding custom value (#4391) 2023-10-26 23:03:21 -03:00
oobabooga 92b2f57095 Minor metadata bug fix (second attempt) 2023-10-26 18:57:32 -07:00
oobabooga 2d97897a25 Don't install flash-attention on windows + cuda 11 2023-10-25 11:21:18 -07:00
LightningDragon 0ced78fdfa
Replace hashlib.sha256 with hashlib.file_digest so we don't need to load entire files into ram before hashing them. (#4383) 2023-10-25 12:15:34 -03:00
tdrussell 72f6fc6923
Rename additive_repetition_penalty to presence_penalty, add frequency_penalty (#4376) 2023-10-25 12:10:28 -03:00
oobabooga ef1489cd4d Remove unused parameter in AutoAWQ 2023-10-23 20:45:43 -07:00
oobabooga 1edf321362 Lint 2023-10-23 13:09:03 -07:00
oobabooga 280ae720d7 Organize 2023-10-23 13:07:17 -07:00
oobabooga 49e5eecce4 Merge remote-tracking branch 'refs/remotes/origin/main' 2023-10-23 12:54:05 -07:00
oobabooga 82c11be067 Update 04 - Model Tab.md 2023-10-23 12:49:07 -07:00
oobabooga 306d764ff6 Minor metadata bug fix 2023-10-23 12:46:24 -07:00
adrianfiedler 4bc411332f
Fix broken links (#4367)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-10-23 14:09:57 -03:00
oobabooga 92691ee626 Disable trust_remote_code by default 2023-10-23 09:57:44 -07:00
tdrussell 4440f87722
Add additive_repetition_penalty sampler setting. (#3627) 2023-10-23 02:28:07 -03:00
oobabooga 6086768309 Bump gradio to 3.50.* 2023-10-22 21:21:26 -07:00
oobabooga b8183148cf
Update 04 ‐ Model Tab.md 2023-10-22 17:15:55 -03:00
oobabooga cea7fc2435 Update html_instruct_style.css 2023-10-22 12:28:23 -07:00
oobabooga df90d03e0b Replace --mul_mat_q with --no_mul_mat_q 2023-10-22 12:23:03 -07:00
Googulator d0c3b407b3
transformers loader: multi-LoRAs support (#3120) 2023-10-22 16:06:22 -03:00
omo 4405513ca5
Option to select/target additional linear modules/layers in LORA training (#4178) 2023-10-22 15:57:19 -03:00
oobabooga 7a3f885ea8
Update 03 ‐ Parameters Tab.md 2023-10-22 14:52:23 -03:00
oobabooga 63688004dc Add default cmd flags to colab 2023-10-22 09:56:43 -07:00
oobabooga 613feca23b Make colab functional for llama.cpp
- Download only Q4_K_M for GGUF repositories by default
- Use maximum n-gpu-layers by default
2023-10-22 09:08:25 -07:00
oobabooga 994502d41b Colab fixes 2023-10-22 08:57:16 -07:00
Jiashu Xu c544f5cc51
Support LLaVA v1.5 7B (#4348) 2023-10-22 12:49:04 -03:00
oobabooga 05741821a5 Minor colab changes 2023-10-22 08:44:35 -07:00
FartyPants (FP HAM) 6a61158adf
Training PRO a month worth of updates (#4345) 2023-10-22 12:38:09 -03:00
mongolu c18504f369
USE_CUDA118 from ENV remains null one_click.py + cuda-toolkit (#4352) 2023-10-22 12:37:24 -03:00
oobabooga cd45635f53 tqdm improvement for colab 2023-10-21 22:00:29 -07:00
oobabooga ae79c510cc Merge remote-tracking branch 'refs/remotes/origin/main' 2023-10-21 21:46:15 -07:00
oobabooga 2d1b3332e4 Ignore warnings on Colab 2023-10-21 21:45:25 -07:00
oobabooga caf6db07ad
Update README.md 2023-10-22 01:22:17 -03:00
oobabooga 1a34927314 Make API URLs more visible 2023-10-21 21:11:07 -07:00
oobabooga 09f807af83 Use ExLlama_HF for GPTQ models by default 2023-10-21 20:45:38 -07:00
oobabooga 619093483e Add Colab notebook 2023-10-21 20:27:52 -07:00
oobabooga 506d05aede Organize command-line arguments 2023-10-21 18:52:59 -07:00
oobabooga b1f33b55fd
Update 01 ‐ Chat Tab.md 2023-10-21 20:17:56 -03:00
oobabooga ac6d5d50b7
Update README.md 2023-10-21 20:03:43 -03:00
oobabooga 6efb990b60
Add a proper documentation (#3885) 2023-10-21 19:15:54 -03:00
Adam White 5a5bc135e9
Docker: Remove explicit CUDA 11.8 Reference (#4343) 2023-10-21 15:09:34 -03:00
oobabooga b98fbe0afc Add download link 2023-10-20 23:58:05 -07:00
oobabooga fbac6d21ca Add missing exception 2023-10-20 23:53:24 -07:00
Brian Dashore 3345da2ea4
Add flash-attention 2 for windows (#4235) 2023-10-21 03:46:23 -03:00
oobabooga 258d046218 More robust way of initializing empty .git folder 2023-10-20 23:13:09 -07:00
Johan 1d5a015ce7
Enable special token support for exllamav2 (#4314) 2023-10-21 01:54:06 -03:00
mjbogusz 8f6405d2fa
Python 3.11, 3.9, 3.8 support (#4233)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-10-20 21:13:33 -03:00
oobabooga 9be74fb57c Change 2 margins 2023-10-20 14:04:14 -07:00
oobabooga e208128d68 Lint the CSS files 2023-10-20 13:02:18 -07:00
oobabooga dedbdb46c2 Chat CSS improvements 2023-10-20 12:49:36 -07:00
Haotian Liu 32984ea2f0
Support LLaVA v1.5 (#4305) 2023-10-20 02:28:14 -03:00
oobabooga bb71272903 Detect WizardCoder-Python-34B & Phind-CodeLlama-34B 2023-10-19 14:35:56 -07:00
oobabooga eda7126b25 Organize the .gitignore 2023-10-19 14:33:44 -07:00
turboderp ae8cd449ae
ExLlamav2_HF: Convert logits to FP32 (#4310) 2023-10-18 23:16:05 -03:00
missionfloyd c0ffb77fd8
More silero languages (#3950) 2023-10-16 17:12:32 -03:00
hronoas db7ecdd274
openai: fix empty models list on query present in url (#4139) 2023-10-16 17:02:47 -03:00
oobabooga f17f7a6913 Increase the evaluation table height 2023-10-16 12:55:35 -07:00
oobabooga 8ea554bc19 Check for torch.xpu.is_available() 2023-10-16 12:53:40 -07:00
oobabooga 188d20e9e5 Reduce the evaluation table height 2023-10-16 10:53:42 -07:00
oobabooga 2d44adbb76 Clear the torch cache while evaluating 2023-10-16 10:52:50 -07:00
oobabooga 388d1864a6 Merge remote-tracking branch 'refs/remotes/origin/main' 2023-10-15 21:58:16 -07:00
oobabooga 71cac7a1b2 Increase the height of the evaluation table 2023-10-15 21:56:40 -07:00
oobabooga e14bde4946 Minor improvements to evaluation logs 2023-10-15 20:51:43 -07:00
oobabooga b88b2b74a6 Experimental Intel Arc transformers support (untested) 2023-10-15 20:51:11 -07:00
Sam d331501ebc
Fix for using Torch with CUDA 11.8 (#4298) 2023-10-15 19:27:19 -03:00
oobabooga 3bb4046fad
Update auto-release.yml 2023-10-15 17:27:16 -03:00
oobabooga 45fa803943
Create auto-release.yml 2023-10-15 17:25:29 -03:00
Johan 2706394bfe
Relax numpy version requirements (#4291) 2023-10-15 12:05:06 -03:00
Forkoz 8cce1f1126
Exllamav2 lora support (#4229)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-10-14 16:12:41 -03:00
jllllll 1f5a2c5597
Use Pytorch 2.1 exllama wheels (#4285) 2023-10-14 15:27:59 -03:00
oobabooga cd1cad1b47 Bump exllamav2 2023-10-14 11:23:07 -07:00
Eve 6e2dec82f1
add chatml support + mistral-openorca (#4275) 2023-10-13 11:49:17 -03:00
Jesus Alvarez ed66ca3cdf
Add HTTPS support to APIs (openai and default) (#4270)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-10-13 01:31:13 -03:00
oobabooga 43be1be598 Manually install CUDA runtime libraries 2023-10-12 21:02:44 -07:00
oobabooga faf5c4dd58 Fix code blocks in instruct mode 2023-10-11 12:18:46 -07:00
oobabooga 773c17faec Fix a warning 2023-10-10 20:53:38 -07:00
oobabooga f63361568c Fix safetensors kwarg usage in AutoAWQ 2023-10-10 19:03:09 -07:00
oobabooga 39f16ff83d Fix default/notebook tabs css 2023-10-10 18:45:12 -07:00
oobabooga fae8062d39
Bump to latest gradio (3.47) (#4258) 2023-10-10 22:20:49 -03:00
Haotian Liu 2b75d725e6
Initial support for LLaVA-LLaMA-2. (#3377) 2023-10-10 18:40:52 -03:00
oobabooga 9fab9a1ca6 Minor fix 2023-10-10 14:08:11 -07:00
oobabooga a49cc69a4a Ignore rope_freq_base if value is 10000 2023-10-10 13:57:40 -07:00
oobabooga 3a9d90c3a1 Download models with 4 threads by default 2023-10-10 13:52:10 -07:00
dependabot[bot] 520cbb2ab1
Bump safetensors from 0.3.2 to 0.4.0 (#4249) 2023-10-10 17:41:09 -03:00
Forkoz 35695e18c7
Remove import. (#4247)
For real this time.
2023-10-09 18:06:11 -03:00
Forkoz 2e471071af
Update llama_attn_hijack.py (#4231) 2023-10-08 15:16:48 -03:00
oobabooga 2e8b5f7c80
Update ROCm command 2023-10-08 10:12:13 -03:00
oobabooga 00187d641a
Note about pytorch 2.1 breaking change 2023-10-08 10:10:38 -03:00
oobabooga 1c6e57dd68
Note about pytorch 2.1 breaking change 2023-10-08 10:09:22 -03:00
oobabooga cf4d89ee65 Lint the javascript code 2023-10-07 19:07:57 -07:00
James Braza 8614c9d085
README for superboogav2 (#4212)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-10-07 19:50:39 -03:00
Brian Dashore 98fa73a974
Text Generation: stop if EOS token is reached (#4213) 2023-10-07 19:46:42 -03:00
Brian Dashore 7743b5e9de
Llamacpp_HF: Fix CFG cache init (#4219)
Documentation says that model.context_params should be sent when
a new context is created. The current code uses model.params which
doesn't exist.

Signed-off-by: kingbri <bdashore3@proton.me>
2023-10-07 19:38:29 -03:00
oobabooga 2a7cb346dd Update the whisper_stt requirements 2023-10-06 21:01:26 -07:00
jllllll 0eda9a0549
Use GPTQ wheels compatible with Pytorch 2.1 (#4210) 2023-10-07 00:35:41 -03:00
oobabooga d33facc9fe
Bump to pytorch 11.8 (#4209) 2023-10-07 00:23:49 -03:00
AG-w 06fff3b2e9
Fix python wheels for avx requirements (#4189) 2023-10-06 15:42:44 -03:00
Casper 0aa853f575
Bump AutoAWQ to v0.1.4 (#4203) 2023-10-06 15:30:01 -03:00
oobabooga 7d3201923b Bump AutoAWQ 2023-10-05 15:14:15 -07:00
turboderp 8a98646a21
Bump ExLlamaV2 to 0.0.5 (#4186) 2023-10-05 19:12:22 -03:00
oobabooga 7ffb424c7b Add AutoAWQ to README 2023-10-05 09:22:37 -07:00
cal066 cc632c3f33
AutoAWQ: initial support (#3999) 2023-10-05 13:19:18 -03:00
oobabooga 3f56151f03 Bump to transformers 4.34 2023-10-05 08:55:14 -07:00
tdrussell cb26163a20
Fix off-by-one error in exllama_hf caching logic (#4145) 2023-10-05 12:20:56 -03:00
Gennadij b04c08378d
Add CMD_FLAGS.txt to .gitignore (#4181) 2023-10-05 10:02:38 -03:00
oobabooga ae4ba3007f
Add grammar to transformers and _HF loaders (#4091) 2023-10-05 10:01:36 -03:00
oobabooga 0197fdddf1
Merge pull request #4142 from jllllll/llamacpp-0.2.11
Bump llama-cpp-python to 0.2.11
2023-10-02 01:31:14 -03:00
oobabooga b6fe6acf88 Add threads_batch parameter 2023-10-01 21:28:00 -07:00
jllllll 41a2de96e5
Bump llama-cpp-python to 0.2.11 2023-10-01 18:08:10 -05:00
oobabooga f2d82f731a Add recommended NTKv1 alpha values 2023-09-29 13:48:38 -07:00
oobabooga abe99cddeb Extend evaluation slider bounds 2023-09-29 13:06:26 -07:00
oobabooga 96da2e1c0d Read more metadata (config.json & quantize_config.json) 2023-09-29 06:14:16 -07:00
oobabooga 56b5a4af74 exllamav2 typical_p 2023-09-28 20:10:12 -07:00
oobabooga f8e9733412 Minor syntax change 2023-09-28 19:32:35 -07:00
oobabooga f931184b53 Increase truncation limits to 32768 2023-09-28 19:28:22 -07:00
oobabooga 1dd13e4643 Read Transformers config.json metadata 2023-09-28 19:19:47 -07:00
oobabooga 9ccaf5eebb I forgot to add the file 2023-09-28 18:25:58 -07:00
oobabooga 92a39c619b Add Mistral support 2023-09-28 15:41:03 -07:00
oobabooga f46ba12b42 Add flash-attn wheels for Linux 2023-09-28 14:45:52 -07:00
oobabooga 771e936769 Fix extensions install (2nd attempt) 2023-09-28 14:33:49 -07:00
快乐的我531 4e56ad55e1
Let model downloader download *.tiktoken as well (#4121) 2023-09-28 18:03:18 -03:00
oobabooga 822ba7fcbb Better error handling during install/update 2023-09-28 13:57:59 -07:00
oobabooga 85f45cafa1 Fix extensions install 2023-09-28 13:54:36 -07:00
Nathan Thomas e145d9a0da
Update one_click.py to initialize site_packages_path variable (#4118) 2023-09-28 08:31:29 -03:00
Chenxiao Wang 3fb1e0236a
fix: update superboogav2 requirements.txt (#4100) 2023-09-27 23:45:59 -03:00
jllllll 2bd23c29cb
Bump llama-cpp-python to 0.2.7 (#4110) 2023-09-27 23:45:36 -03:00
missionfloyd 86e7c05429
Delete extensions/Training_PRO/readme.md (#4112) 2023-09-27 23:45:13 -03:00
Sam a0d99dcf90
fix: add missing superboogav2 dep (#4099) 2023-09-26 23:37:22 -03:00
StoyanStAtanasov 7e6ff8d1f0
Enable NUMA feature for llama_cpp_python (#4040) 2023-09-26 22:05:00 -03:00
oobabooga 87ea2d96fd Add a note about RWKV loader 2023-09-26 17:43:39 -07:00
jllllll 13a54729b1
Bump exllamav2 to 0.0.4 and use pre-built wheels (#4095) 2023-09-26 21:36:14 -03:00
jllllll 3879ab5007
Expand MacOS llama.cpp support in requirements (#4094)
Provides MacOS 12 and 13 wheels.
2023-09-26 21:34:48 -03:00
jllllll 9d9aa38234
Fix old install migration for WSL installer (#4093) 2023-09-26 21:34:16 -03:00
HideLord 0845724a89
Supercharging superbooga (#3272) 2023-09-26 21:30:19 -03:00
jllllll ad00b8eb26
Check '--model-dir' for no models warning (#4067) 2023-09-26 10:56:57 -03:00
oobabooga 0c89180966 Another minor fix 2023-09-26 06:54:21 -07:00
oobabooga 365335e1ae Minor fix 2023-09-26 06:47:19 -07:00
oobabooga 1ca54faaf0 Improve --multi-user mode 2023-09-26 06:42:33 -07:00
oobabooga 019371c0b6 Lint 2023-09-25 20:31:11 -07:00
oobabooga 814520fed1 Extension install improvements 2023-09-25 20:27:06 -07:00
oobabooga 7f1460af29 Change a warning 2023-09-25 20:22:27 -07:00
oobabooga 862b45b1c7 Extension install improvements 2023-09-25 19:48:30 -07:00
oobabooga 44438c60e5 Add INSTALL_EXTENSIONS environment variable 2023-09-25 13:12:35 -07:00
oobabooga 31f2815a04
Update Generation-Parameters.md 2023-09-25 16:30:52 -03:00
oobabooga c8952cce55 Move documentation from UI to docs/ 2023-09-25 12:28:28 -07:00
oobabooga d0d221df49 Add --use_fast option (closes #3741) 2023-09-25 12:19:43 -07:00
oobabooga b973b91d73 Automatically filter by loader (closes #4072) 2023-09-25 10:28:35 -07:00
oobabooga 63de9eb24f Clean up the transformers loader 2023-09-24 20:26:26 -07:00
oobabooga 36c38d7561 Add disable_exllama to Transformers loader (for GPTQ LoRA training) 2023-09-24 20:03:11 -07:00
jllllll c0fca23cb9
Avoid importing torch in one-click-installer (#4064) 2023-09-24 22:16:59 -03:00
oobabooga 55a685d999 Minor fixes 2023-09-24 14:15:10 -07:00
oobabooga 08cf150c0c
Add a grammar editor to the UI (#4061) 2023-09-24 18:05:24 -03:00
oobabooga 08c4fb12ae Use bitsandbytes==0.38.1 for AMD 2023-09-24 08:11:59 -07:00
oobabooga d5952cb540 Don't assume that py-cpuinfo is installed 2023-09-24 08:10:45 -07:00
oobabooga eb0b7c1053 Fix a minor UI bug 2023-09-24 07:17:33 -07:00
oobabooga 3edac43426 Remove print statement 2023-09-24 07:13:00 -07:00
oobabooga b227e65d86 Add grammar to llama.cpp loader (closes #4019) 2023-09-24 07:10:45 -07:00
oobabooga a3ad9fe6c0 Add comments 2023-09-24 06:08:39 -07:00
oobabooga 2e7b6b0014
Create alternative requirements.txt with AMD and Metal wheels (#4052) 2023-09-24 09:58:29 -03:00
Chenxiao Wang 9de2dfa887
extensions/openai: Fix error when preparing cache for embedding models (#3995) 2023-09-24 00:58:28 -03:00
oobabooga 7a3ca2c68f Better detect EXL2 models 2023-09-23 13:05:55 -07:00
oobabooga 895ec9dadb
Update README.md 2023-09-23 15:37:39 -03:00
oobabooga 299d285ff0
Update README.md 2023-09-23 15:36:09 -03:00
oobabooga 6903af33dd
Update One-Click-Installers.md 2023-09-23 15:32:24 -03:00
oobabooga 3f73e4cdde Keep dropdown menus above errored components 2023-09-23 07:50:24 -07:00
oobabooga 30d7c4eaa1 Forward --help to server.py 2023-09-23 07:27:27 -07:00
oobabooga c2ae01fb04 Improved readability 2023-09-23 07:10:01 -07:00
oobabooga fc351ff3e5 Improved readability 2023-09-23 06:48:09 -07:00
oobabooga e6f445f3eb Improved readability of one_click.py 2023-09-23 06:28:58 -07:00
oobabooga 639723845a Make N the "None" install option 2023-09-23 05:25:06 -07:00
oobabooga 0306b61bb0 Add IPEX option to the installer (experimental) 2023-09-23 05:17:41 -07:00
jllllll b5e1f12394
Use call for conda deactivate in Windows installer (#4042) 2023-09-23 08:33:44 -03:00
oobabooga 4b4d283a4c
Update README.md 2023-09-23 00:09:59 -03:00
oobabooga 0581f1094b
Update README.md 2023-09-22 23:31:32 -03:00
oobabooga 968f98a57f
Update README.md 2023-09-22 23:23:16 -03:00
mongolu d70b8d9048 Added two ENVs in webui.py for docker (#111) 2023-09-22 19:04:41 -07:00
oobabooga 72b4ab4c82 Update README 2023-09-22 15:20:09 -07:00
oobabooga 7b9ad643a0 Update CMD_FLAGS.txt 2023-09-22 15:02:11 -07:00
oobabooga 336c8de503 Add an example to CMD_FLAGS.txt 2023-09-22 15:01:45 -07:00
oobabooga 5dd212cf52 TTS extensions fixes after #4022 2023-09-22 14:58:51 -07:00
FartyPants 26f10854f3
Training PRO extension update (#4036) 2023-09-22 18:51:31 -03:00
oobabooga c5e0ab7174 Minor bug fix 2023-09-22 14:50:27 -07:00
oobabooga a0c7d764b5 Fix pytorch installation on Linux 2023-09-22 14:34:30 -07:00
oobabooga b1467bd064
Move one-click-installers into the repository (#4028 from oobabooga/one-click) 2023-09-22 17:43:07 -03:00
deevis 7f0ea4dc16 feature: allow comments in CMD_FLAGS.txt (#127)
---------

Co-authored-by: missionfloyd <missionfloyd@users.noreply.github.com>
2023-09-22 13:28:24 -07:00
oobabooga 08e7ee4c4c Merge remote-tracking branch 'refs/remotes/origin/one-click' into one-click 2023-09-22 13:11:04 -07:00
oobabooga c075969875 Add instructions 2023-09-22 13:10:03 -07:00
oobabooga 589ee9f623
Update README.md 2023-09-22 16:21:48 -03:00
oobabooga c33a94e381 Rename doc file 2023-09-22 12:17:47 -07:00
oobabooga 0fee18e8b7 Rename some functions 2023-09-22 12:08:05 -07:00
oobabooga 6c5f81f002 Rename webui.py to one_click.py 2023-09-22 12:00:06 -07:00
oobabooga fe2acdf45f
Update README.md 2023-09-22 15:52:20 -03:00
oobabooga 967dda17a0 Remove OOBABOOGA_FLAGS 2023-09-22 10:52:52 -07:00
oobabooga ccfc919016 Make webui.py more readable 2023-09-22 10:51:29 -07:00
oobabooga c74326de02 Fixes by @jllllll 2023-09-22 10:37:22 -07:00
oobabooga b4b5f45558 Join the installation instructions 2023-09-22 10:28:22 -07:00
oobabooga 2d2a8cfb48 Remove a file 2023-09-22 10:08:08 -07:00
oobabooga 3314b7d795 Allow start scripts to have command-line flags 2023-09-22 10:03:56 -07:00
oobabooga d43d150b1e Fix a bug in the chat API (closes #4034) 2023-09-22 09:40:07 -07:00
oobabooga 8ab3eca9ec Add a warning for outdated installations 2023-09-22 09:35:19 -07:00
oobabooga 86648d4085 Remove CUDA, keep only pytorch 2023-09-22 08:13:11 -07:00
oobabooga 66363a4d70 Minor changes / reorder some functions 2023-09-22 08:02:21 -07:00
oobabooga 84b5a519cb
Merge pull request #4029 from jllllll/one-click
Various one-click-installer updates and fixes
2023-09-22 11:55:01 -03:00
oobabooga 02e771403b Improve the default character 2023-09-22 07:23:33 -07:00
oobabooga 95976a9d4f Fix a bug while deleting characters 2023-09-22 06:02:34 -07:00
jllllll 69b0aedd95
Fix missing models warning 2023-09-22 01:12:08 -05:00
jllllll 060bb76aa0
Update WSL installer 2023-09-22 01:10:30 -05:00
oobabooga ee7bf49804 Change back list style 2023-09-21 21:09:22 -07:00
jllllll 9054c98eca
Use --autostash on git pull 2023-09-21 23:00:33 -05:00
oobabooga 12e312ae9c Focus on the chat input always 2023-09-21 20:32:24 -07:00
jllllll 498552a92b
More robust installation check for installer 2023-09-21 22:23:23 -05:00
jllllll cd1049eded
Add Conda env deactivation to installer scripts
Avoids conflicts with existing Conda installations
2023-09-21 21:52:29 -05:00
jllllll 6bbfc40d10
Add .git creation to installer 2023-09-21 21:51:58 -05:00
oobabooga d5330406fa Add a rename menu for chat histories 2023-09-21 19:16:51 -07:00
oobabooga d6814d7c15 Fix a bug in the API (closes #4027) 2023-09-21 17:54:53 -07:00
oobabooga 193fe18c8c Resolve conflicts 2023-09-21 17:45:11 -07:00
oobabooga df39f455ad Merge remote-tracking branch 'second-repo/main' into merge-second-repo 2023-09-21 17:39:54 -07:00
oobabooga fc2b831692 Basic changes 2023-09-21 15:55:09 -07:00
oobabooga b04b3957f9 Move one-click-installers into the repository 2023-09-21 15:35:53 -07:00
oobabooga 05c4a4f83c Bump exllamav2 2023-09-21 14:56:01 -07:00
oobabooga 9a5ab454b4 Improve list styles 2023-09-21 14:49:00 -07:00
oobabooga 00ab450c13
Multiple histories for each character (#4022) 2023-09-21 17:19:32 -03:00
oobabooga 029da9563f Avoid redundant function call in llamacpp_hf 2023-09-19 14:14:40 -07:00
oobabooga 9b7646140c Trim model path if using absolute path 2023-09-19 13:51:57 -07:00
oobabooga 869f47fff9 Lint 2023-09-19 13:51:57 -07:00
oobabooga 13ac55fa18 Reorder some functions 2023-09-19 13:51:57 -07:00
oobabooga e2fddd9584 More robust autoscrolling (attempt) 2023-09-19 13:12:34 -07:00
oobabooga 03dc69edc5 ExLlama_HF (v1 and v2) prefix matching 2023-09-19 13:12:19 -07:00
oobabooga 5075087461 Fix command-line arguments being ignored 2023-09-19 13:11:46 -07:00
oobabooga ff5d3d2d09 Add missing import 2023-09-18 16:26:54 -07:00
oobabooga 605ec3c9f2 Add a warning about ExLlamaV2 without flash-attn 2023-09-18 12:26:35 -07:00
oobabooga f0ef971edb Remove obsolete warning 2023-09-18 12:25:10 -07:00
oobabooga 745807dc03 Faster llamacpp_HF prefix matching 2023-09-18 11:02:45 -07:00
BadisG 893a72a1c5
Stop generation immediately when using "Maximum tokens/second" (#3952)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-09-18 14:27:06 -03:00
jllllll b7c55665c1
Bump llama-cpp-python to 0.2.6 (#3982) 2023-09-18 14:08:37 -03:00
Cebtenzzre 8466cf229a
llama.cpp: fix ban_eos_token (#3987) 2023-09-18 12:15:02 -03:00
oobabooga 0ede2965d5 Remove an error message 2023-09-17 18:46:08 -07:00
dependabot[bot] 661bfaac8e
Update accelerate from ==0.22.* to ==0.23.* (#3981) 2023-09-17 22:42:12 -03:00
Chenxiao Wang 347aed4254
extensions/openai: load extension settings via settings.yaml (#3953) 2023-09-17 22:39:29 -03:00
missionfloyd cc8eda298a
Move hover menu shortcuts to right side (#3951) 2023-09-17 22:33:00 -03:00
oobabooga 280cca9f66 Merge remote-tracking branch 'refs/remotes/origin/main' 2023-09-17 18:01:27 -07:00
oobabooga b062d50c45 Remove exllama import that causes problems 2023-09-17 18:00:32 -07:00
James Braza fee38e0601
Simplified ExLlama cloning instructions and failure message (#3972) 2023-09-17 19:26:05 -03:00
Thireus ☠ 45335fa8f4
Bump ExLlamav2 to v0.0.2 (#3970) 2023-09-17 19:24:40 -03:00
Lu Guanghua 9858acee7b
Fix unexpected extensions load after gradio restart (#3965) 2023-09-17 17:35:43 -03:00
oobabooga d9b0f2c9c3 Fix llama.cpp double decoding 2023-09-17 13:07:48 -07:00
FartyPants 230b562d53
Training_PRO extension - added target selector (#3969) 2023-09-17 17:00:00 -03:00
oobabooga d71465708c llamacpp_HF prefix matching 2023-09-17 11:51:01 -07:00
oobabooga 763ea3bcb2 Improved multimodal error message 2023-09-17 09:22:16 -07:00
oobabooga 37e2980e05 Recommend mul_mat_q for llama.cpp 2023-09-17 08:27:11 -07:00
oobabooga a069f3904c Undo part of ad8ac545a5 2023-09-17 08:12:23 -07:00
FartyPants e34c6e6938
Training PRO extension (#3961) 2023-09-17 11:09:31 -03:00
oobabooga ad8ac545a5 Tokenization improvements 2023-09-17 07:02:00 -07:00
saltacc cd08eb0753
token probs for non HF loaders (#3957) 2023-09-17 10:42:32 -03:00
Shulzhenko Anatolii 0668f4e67f
Add speechrecognition dependency for OpenAI extension (#3959) 2023-09-16 13:49:48 -03:00
wizd cc7f345c29
add whisper api to openai plugin (#3958) 2023-09-16 12:04:04 -03:00
Lu Guanghua cd534ba46e
Fix Google Translate escaping (#3827)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-09-16 10:18:06 -03:00
kalomaze 7c9664ed35
Allow full model URL to be used for download (#3919)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-09-16 10:06:13 -03:00
saltacc ed6b6411fb
Fix exllama tokenizers (#3954)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-09-16 09:42:38 -03:00
oobabooga 8d85425e09 Increase --chat-buttons height 2023-09-15 21:21:24 -07:00
oobabooga e75489c252 Update README 2023-09-15 21:04:51 -07:00
oobabooga 52c4fb75ff Remove extra space in --chat-buttons 2023-09-15 20:56:30 -07:00
missionfloyd 2ad6ca8874
Add back chat buttons with --chat-buttons (#3947) 2023-09-16 00:39:37 -03:00
oobabooga f5fb1ee666 Change a comment 2023-09-15 20:16:30 -07:00
oobabooga 2c1b548cea Minor fix 2023-09-15 20:14:32 -07:00
oobabooga 8f97e87cac Lint the openai extension 2023-09-15 20:11:16 -07:00
oobabooga 760510db52 Change a height 2023-09-15 19:41:53 -07:00
oobabooga ef04138bc0 Improve the UI tokenizer 2023-09-15 19:30:44 -07:00
oobabooga c3e4c9fdc2 Add a simple tokenizer to the UI 2023-09-15 19:09:03 -07:00
saltacc f01b9aa71f
Add customizable ban tokens (#3899) 2023-09-15 18:27:27 -03:00
oobabooga fb864dad7b Update README 2023-09-15 13:00:46 -07:00
oobabooga 5b117590ad Add some scrollbars to Parameters tab 2023-09-15 09:17:37 -07:00
oobabooga a3a5ffe651 Adjust Default tab heights 2023-09-15 09:12:42 -07:00
oobabooga 985020f038 Adjust token counter height 2023-09-15 08:50:59 -07:00
Johan fdcee0c215
Allow custom tokenizer for llamacpp_HF loader (#3941) 2023-09-15 12:38:38 -03:00
oobabooga 75646df268 Improve the header bar a bit 2023-09-15 08:06:17 -07:00
oobabooga efd9b84fb2 Improve Default tab layout on mobile 2023-09-15 07:58:53 -07:00
oobabooga fd7257c7f8 Prevent code blocks from flickering while streaming 2023-09-15 07:46:26 -07:00
oobabooga 28bb7a8b57 Increase instruct line height 2023-09-15 07:31:43 -07:00
oobabooga eb8aa014f6 Increase hover menu button height 2023-09-15 06:44:21 -07:00
oobabooga 980997c667 Adjust a position 2023-09-14 22:11:28 -07:00
oobabooga 3181ab91dc Make "show controls" clickable again (2nd attempt) 2023-09-14 22:07:28 -07:00
oobabooga 5dc4e101d8 Add a margin to the chat input field 2023-09-14 22:03:59 -07:00
oobabooga 5c62320a62 Update cai-chat-square style 2023-09-14 19:05:29 -07:00
oobabooga be73a5ecf4 Italics style 2023-09-14 17:18:59 -07:00
oobabooga a3ecf3bb65 Add cai-chat-square chat style 2023-09-14 16:15:08 -07:00
oobabooga 3d1c0f173d User config precedence over GGUF metadata 2023-09-14 12:15:52 -07:00
jllllll 8250d0149d
Fix pydantic version conflict in elevenlabs extension (#3927) 2023-09-14 13:42:13 -03:00
missionfloyd 6154fee16d
Fix TheEncrypted777 theme in light mode (#3917) 2023-09-14 12:40:55 -03:00
oobabooga 3200328064 Minor mobile CSS fix 2023-09-14 08:38:26 -07:00
oobabooga 94dc64f870 Add a border 2023-09-14 07:20:36 -07:00
oobabooga 416e8f4758 Make "show controls" clickable again 2023-09-14 07:13:51 -07:00
oobabooga a836c5b29a Improve the hover menu 2023-09-14 07:09:06 -07:00
oobabooga 70aafa34dc Fix blockquote markdown rendering 2023-09-14 05:57:04 -07:00
oobabooga e180284593 Update instruct style 2023-09-14 05:56:57 -07:00
oobabooga e6d9c59398 Add background to "Show controls" 2023-09-14 05:32:44 -07:00
oobabooga 644a9b8765 Change the chat generate button 2023-09-14 05:16:44 -07:00
oobabooga ecc90f9f62 Continue on Alt + Enter 2023-09-14 03:59:12 -07:00
oobabooga 1ce3c93600 Allow "Your name" field to be saved 2023-09-14 03:44:35 -07:00
oobabooga ce59c92764 Hide keyboard shortcut hints on mobile 2023-09-14 03:17:15 -07:00
oobabooga 27dbcc59f5
Make the chat input expand upwards (#3920) 2023-09-14 07:06:42 -03:00
oobabooga 6b6af74e14 Keyboard shortcuts without conflicts (hopefully) 2023-09-14 02:33:52 -07:00
oobabooga fc11d1eff0 Add chat keyboard shortcuts 2023-09-13 19:22:40 -07:00
oobabooga df592adff5 Minor CSS change 2023-09-13 13:58:15 -07:00
oobabooga 9f199c7a4c Use Noto Sans font
Copied from 6c8bd06308/public/webfonts/NotoSans
2023-09-13 13:48:05 -07:00
oobabooga 19548482f1 Remove unnecessary padding 2023-09-13 11:48:42 -07:00
oobabooga 8ce94b735c Show progress on impersonate 2023-09-13 11:22:53 -07:00
oobabooga 7cd437e05c Properly close the hover menu on mobile 2023-09-13 11:10:46 -07:00
oobabooga 9fc46d3c19 Move "Show controls" to the right 2023-09-13 10:02:25 -07:00
oobabooga 1b47b5c676 Change the Generate/Stop buttons 2023-09-13 09:25:26 -07:00
oobabooga 01335d48b2 Minor CSS fix 2023-09-13 09:08:28 -07:00
oobabooga 8ea28cbfe0 Reorder chat buttons 2023-09-13 08:49:11 -07:00
oobabooga 5e3d2f7d44
Reorganize chat buttons (#3892) 2023-09-13 02:36:12 -03:00
Panchovix 34dc7306b8
Fix NTK (alpha) and RoPE scaling for exllamav2 and exllamav2_HF (#3897) 2023-09-13 02:35:09 -03:00
dependabot[bot] eb9ebabec7
Bump exllamav2 from 0.0.0 to 0.0.1 (#3896) 2023-09-13 02:13:51 -03:00
cal066 a4e4e887d7
Bump ctransformers to 0.2.27 (#3893) 2023-09-13 00:37:31 -03:00
oobabooga b7adf290fc Fix ExLlama-v2 path issue 2023-09-12 17:42:22 -07:00
jllllll 1a5d68015a
Bump llama-cpp-python to 0.1.85 (#3887) 2023-09-12 19:41:41 -03:00
oobabooga 833bc59f1b Remove ninja from requirements.txt
It's installed with exllamav2 automatically
2023-09-12 15:12:56 -07:00
oobabooga b190676893 Merge remote-tracking branch 'refs/remotes/origin/main' 2023-09-12 15:06:33 -07:00
oobabooga 2f935547c8 Minor changes 2023-09-12 15:05:21 -07:00
oobabooga 18e6b275f3 Add alpha_value/compress_pos_emb to ExLlama-v2 2023-09-12 15:02:47 -07:00
Gennadij 460c40d8ab
Read more GGUF metadata (scale_linear and freq_base) (#3877) 2023-09-12 17:02:42 -03:00
Eve 90fca6a77d
add pygmalion-2 and mythalion support (#3821) 2023-09-12 15:57:49 -03:00
Chang Chi, Meng b61d9aef19
openai API: add support for chunked transfer encoding in POST requests (#3870) 2023-09-12 15:54:42 -03:00
dependabot[bot] 0efbe5ef76
Bump optimum from 1.12.0 to 1.13.1 (#3872) 2023-09-12 15:53:21 -03:00
missionfloyd 43ec9d1619
silero_tts: Add language option (#3878) 2023-09-12 15:49:46 -03:00
oobabooga 04a74b3774 Update README 2023-09-12 10:46:27 -07:00
oobabooga 16e1696071 Minor qol change 2023-09-12 10:44:26 -07:00
oobabooga c2a309f56e
Add ExLlamaV2 and ExLlamav2_HF loaders (#3881) 2023-09-12 14:33:07 -03:00
oobabooga a821928877 Reduce chat width 2023-09-12 10:26:43 -07:00
oobabooga df123a20fc Prevent extra keys from being saved to settings.yaml 2023-09-11 20:13:10 -07:00
oobabooga dae428a967 Revamp cai-chat theme, make it default 2023-09-11 19:30:40 -07:00
oobabooga 47d1ca467b Pin pandas version in superbooga 2023-09-11 18:34:34 -07:00
oobabooga 78811dd89a Fix GGUF metadata reading for falcon 2023-09-11 15:49:50 -07:00
oobabooga 9331ab4798
Read GGUF metadata (#3873) 2023-09-11 18:49:30 -03:00
oobabooga 39f4800d94 Merge remote-tracking branch 'refs/remotes/origin/main' 2023-09-11 10:45:13 -07:00
oobabooga 5c58dfadef Update requirements_nocuda.txt 2023-09-11 10:44:19 -07:00
Sam fa363da7ce
improve docker builds (#3715) 2023-09-11 12:22:00 -03:00
oobabooga df52dab67b Lint 2023-09-11 07:57:38 -07:00
Eve 92f3cd624c
Improve instructions for CPUs without AVX2 (#3786) 2023-09-11 11:54:04 -03:00
oobabooga ed86878f02 Remove GGML support 2023-09-11 07:44:00 -07:00
John Smith cc7b7ba153
fix lora training with alpaca_lora_4bit (#3853) 2023-09-11 01:22:20 -03:00
Forkoz 15e9b8c915
Exllama new rope settings (#3852) 2023-09-11 01:14:36 -03:00
jllllll 859b4fd737
Bump exllama to 0.1.17 (#3847) 2023-09-11 01:13:14 -03:00
dependabot[bot] 1d6b384828
Update transformers requirement from ==4.32.* to ==4.33.* (#3865) 2023-09-11 01:12:22 -03:00
jllllll e8f234ca8f
Bump llama-cpp-python to 0.1.84 (#3854) 2023-09-11 01:11:33 -03:00
oobabooga 66d5caba1b Pin pydantic version (closes #3850) 2023-09-10 21:09:04 -07:00
oobabooga 4affa08821 Do not impose instruct mode while loading models 2023-09-02 11:31:33 -07:00
oobabooga 0576691538 Add optimum to requirements (for GPTQ LoRA training)
See https://github.com/oobabooga/text-generation-webui/issues/3655
2023-08-31 08:45:38 -07:00
oobabooga 40ffc3d687
Update README.md 2023-08-30 18:19:04 -03:00
oobabooga 47e490c7b4 Set use_cache=True by default for all models 2023-08-30 13:26:27 -07:00
oobabooga 5190e153ed
Update README.md 2023-08-30 14:06:29 -03:00
jllllll 9626f57721
Bump exllama to 0.0.14 (#3758) 2023-08-30 13:43:38 -03:00
oobabooga bc4023230b Improved instructions for AMD/Metal/Intel Arc/CPUs without AVCX2 2023-08-30 09:40:00 -07:00
oobabooga b2f7ca0d18 Cloudfare fix 2 2023-08-29 19:54:43 -07:00
missionfloyd 787219267c
Allow downloading single file from UI (#3737) 2023-08-29 23:32:36 -03:00
Alberto Ferrer f63dd83631
Update download-model.py (Allow single file download) (#3732) 2023-08-29 22:57:58 -03:00
jllllll dac5f4b912
Bump llama-cpp-python to 0.1.83 (#3745) 2023-08-29 22:35:59 -03:00
oobabooga 6c16e4cecf Cloudfare fix
Credits: https://github.com/oobabooga/text-generation-webui/issues/1524#issuecomment-1698255209
2023-08-29 16:35:44 -07:00
oobabooga 828d97a98c Minor CSS improvement 2023-08-29 16:15:12 -07:00
oobabooga a26c2300cb Make instruct style more readable (attempt) 2023-08-29 14:14:01 -07:00
q5sys (JT) cdb854db9e
Update llama.cpp.md instructions (#3702) 2023-08-29 17:56:50 -03:00
VishwasKukreti a9a1784420
Update accelerate to 0.22 in requirements.txt (#3725) 2023-08-29 17:47:37 -03:00
oobabooga cec8db52e5
Add max_tokens_second param (#3533) 2023-08-29 17:44:31 -03:00
jllllll fe1f7c6513
Bump ctransformers to 0.2.25 (#3740) 2023-08-29 17:24:36 -03:00
oobabooga 672b610dba Improve tab switching js 2023-08-29 13:22:15 -07:00
oobabooga 2b58a89f6a Clear instruction template before loading new one 2023-08-29 13:11:32 -07:00
oobabooga 36864cb3e8 Use Alpaca as the default instruction template 2023-08-29 13:06:25 -07:00
oobabooga 9a202f7fb2 Prevent <ul> lists from flickering during streaming 2023-08-28 20:45:07 -07:00
oobabooga 8b56fc993a Change lists style in chat mode 2023-08-28 20:14:02 -07:00
oobabooga e8c0c4990d Unescape HTML in the chat API examples 2023-08-28 19:42:03 -07:00
oobabooga 439dd0faab Fix stopping strings in the chat API 2023-08-28 19:40:11 -07:00
oobabooga 86c45b67ca Merge remote-tracking branch 'refs/remotes/origin/main' 2023-08-28 18:29:38 -07:00
oobabooga c75f98a6d6 Autoscroll Notebook/Default textareas during streaming 2023-08-28 18:22:03 -07:00
jllllll 22b2a30ec7
Bump llama-cpp-python to 0.1.82 (#3730) 2023-08-28 18:02:24 -03:00
oobabooga 558e918fd6 Add a typing dots (...) animation to chat tab 2023-08-28 13:50:36 -07:00
oobabooga 57e9ded00c
Make it possible to scroll during streaming (#3721) 2023-08-28 16:03:20 -03:00
jllllll 7d3a0b5387
Bump llama-cpp-python to 0.1.81 (#3716) 2023-08-27 22:38:41 -03:00
oobabooga fdef0e4efa Focus on chat input field after Ctrl+S 2023-08-27 16:45:37 -07:00
Cebtenzzre 2f5d769a8d
accept floating-point alpha value on the command line (#3712) 2023-08-27 18:54:43 -03:00
oobabooga 0986868b1b Fix chat scrolling with Dark Reader extension 2023-08-27 14:53:42 -07:00
oobabooga b2296dcda0 Ctrl+S to show/hide chat controls 2023-08-27 13:14:33 -07:00
Kelvie Wong a965a36803
Add ffmpeg to the Docker image (#3664) 2023-08-27 12:29:00 -03:00
Ravindra Marella e4c3e1bdd2
Fix ctransformers model unload (#3711)
Add missing comma in model types list

Fixes marella/ctransformers#111
2023-08-27 10:53:48 -03:00
oobabooga 0c9e818bb8 Update truncation length based on max_seq_len/n_ctx 2023-08-26 23:10:45 -07:00
oobabooga e6eda5c2da
Merge pull request #3695 from oobabooga/gguf2
GGUF
2023-08-27 02:33:26 -03:00
oobabooga 3361728da1 Change some comments 2023-08-26 22:24:44 -07:00
oobabooga 8aeae3b3f4 Fix llamacpp_HF loading 2023-08-26 22:15:06 -07:00
oobabooga 7f5370a272 Minor fixes/cosmetics 2023-08-26 22:11:07 -07:00
oobabooga d826bc5d1b
Merge pull request #3697 from jllllll/llamacpp-ggml
Use separate llama-cpp-python packages for GGML support
2023-08-27 01:51:00 -03:00
jllllll 4d61a7d9da
Account for deprecated GGML parameters 2023-08-26 14:07:46 -05:00
jllllll 4a999e3bcd
Use separate llama-cpp-python packages for GGML support 2023-08-26 10:40:08 -05:00
oobabooga 6e6431e73f Update requirements.txt 2023-08-26 01:07:28 -07:00
oobabooga 83640d6f43 Replace ggml occurences with gguf 2023-08-26 01:06:59 -07:00
oobabooga 1a642c12b5 Fix silero_tts HTML unescaping 2023-08-26 00:45:07 -07:00
jllllll db42b365c9
Fix ctransformers threads auto-detection (#3688) 2023-08-25 14:37:02 -03:00
oobabooga 0bcecaa216 Set mode: instruct for CodeLlama-instruct 2023-08-25 07:59:23 -07:00
cal066 960980247f
ctransformers: gguf support (#3685) 2023-08-25 11:33:04 -03:00
oobabooga 21058c37f7 Add missing file 2023-08-25 07:10:26 -07:00
oobabooga f4f04c8c32 Fix a typo 2023-08-25 07:08:38 -07:00
oobabooga 5c7d8bfdfd Detect CodeLlama settings 2023-08-25 07:06:57 -07:00
oobabooga 52ab2a6b9e Add rope_freq_base parameter for CodeLlama 2023-08-25 06:55:15 -07:00
oobabooga feecd8190f Unescape inline code blocks 2023-08-24 21:01:09 -07:00
oobabooga 26c5e5e878 Bump autogptq 2023-08-24 19:23:08 -07:00
oobabooga a2c67262c7 Unescape model output for silero/elevenlabs 2023-08-24 17:27:12 -07:00
oobabooga 3320accfdc
Add CFG to llamacpp_HF (second attempt) (#3678) 2023-08-24 20:32:21 -03:00
oobabooga d6934bc7bc
Implement CFG for ExLlama_HF (#3666) 2023-08-24 16:27:36 -03:00
oobabooga 2b675533f7 Un-bump safetensors
The newest one doesn't work on Windows yet
2023-08-23 14:36:03 -07:00
oobabooga 87442c6d18 Fix Notebook Logits tab 2023-08-22 21:00:12 -07:00
oobabooga c0b119c3a3 Improve logit viewer format 2023-08-22 20:35:12 -07:00
oobabooga 6d6f40e8f8 Merge remote-tracking branch 'refs/remotes/origin/main' 2023-08-22 20:18:45 -07:00
oobabooga 8545052c9d Add the option to use samplers in the logit viewer 2023-08-22 20:18:16 -07:00
oobabooga d7c98fe715
Update stale.yml 2023-08-22 21:48:32 -03:00
Sam 0b352ea7ef
Add missing extensions to Dockerfile (#3544) 2023-08-22 17:41:11 -03:00
oobabooga 25e5eaa6a6 Remove outdated training warning 2023-08-22 13:16:44 -07:00
oobabooga 335c49cc7e Bump peft and transformers 2023-08-22 13:14:59 -07:00
oobabooga 727fd229f4 Increase stalebot timeout to 6 weeks 2023-08-22 13:03:17 -07:00
tkbit df165fe6c4
Use numpy==1.24 in requirements.txt (#3651)
The whisper extension needs numpy 1.24 to work properly
2023-08-22 16:55:17 -03:00
cal066 e042bf8624
ctransformers: add mlock and no-mmap options (#3649) 2023-08-22 16:51:34 -03:00
tdrussell 2da38e89e6
Fix whitespace formatting in perplexity_colors extension. (#3643) 2023-08-22 16:49:37 -03:00
oobabooga 1b419f656f Acknowledge a16z support 2023-08-21 11:57:51 -07:00
oobabooga 6cca8b8028 Only update notebook token counter on input
For performance during streaming
2023-08-21 05:39:55 -07:00
oobabooga 41b98e07fb Minor CSS fix 2023-08-20 22:09:18 -07:00
oobabooga 2cb07065ec Fix an escaping bug 2023-08-20 21:50:42 -07:00
oobabooga a74dd9003f Fix HTML escaping for perplexity_colors extension 2023-08-20 21:40:22 -07:00
oobabooga 6394fef1db Rewrite tab detection js 2023-08-20 21:02:53 -07:00
oobabooga 57036abc76 Add "send to default/notebook" buttons to chat tab 2023-08-20 19:54:59 -07:00
oobabooga 429cacd715 Add a token counter similar to automatic1111
It can now be found in the Default and Notebook tabs
2023-08-20 19:37:33 -07:00
oobabooga 120fb86c6a
Add a simple logit viewer (#3636) 2023-08-20 20:49:21 -03:00
SeanScripts 2c1fd0d72b
Add probability dropdown to perplexity_colors extension (#3148) 2023-08-20 20:28:14 -03:00
Thomas De Bonnet 0dfd1a8b7d
Improve readability of download-model.py (#3497) 2023-08-20 20:13:13 -03:00
oobabooga 457fedfa36 Remove niche dockerfile 2023-08-20 16:02:44 -07:00
oobabooga ef17da70af Fix ExLlama truncation 2023-08-20 08:53:26 -07:00
oobabooga ee964bcce9 Update a comment about RoPE scaling 2023-08-20 07:01:43 -07:00
missionfloyd 1cae784761
Unescape last message (#3623) 2023-08-19 09:29:08 -03:00
Cebtenzzre 942ad6067d
llama.cpp: make Stop button work with streaming disabled (#3620) 2023-08-19 00:17:27 -03:00
oobabooga f6724a1a01 Return the visible history with "Copy last reply" 2023-08-18 13:04:45 -07:00
oobabooga b96fd22a81
Refactor the training tab (#3619) 2023-08-18 16:58:38 -03:00
oobabooga 54df0bfad1 Update README.md 2023-08-18 09:43:15 -07:00
oobabooga f50f534b0f Add note about AMD/Metal to README 2023-08-18 09:37:20 -07:00
oobabooga c4733000d7 Return the visible history with "Remove last" 2023-08-18 09:25:51 -07:00
oobabooga 5a6e7057b9 Merge branch 'bump-llamacpp' 2023-08-18 08:05:24 -07:00
jllllll 1a71ab58a9
Bump llama_cpp_python_cuda to 0.1.78 (#3614) 2023-08-18 12:04:01 -03:00
oobabooga 7cba000421
Bump llama-cpp-python, +tensor_split by @shouyiwang, +mul_mat_q (#3610) 2023-08-18 12:03:34 -03:00
oobabooga d8f660e586 Add to modules/loaders.py 2023-08-18 08:00:22 -07:00
oobabooga 4ec42679e3 Add --mul_mat_q param 2023-08-18 07:58:20 -07:00
oobabooga 28cf5862af Add UI element for tensor_split 2023-08-18 06:26:48 -07:00
missionfloyd 4b69f4f6ae
Fix print CSS (#3608) 2023-08-18 01:44:22 -03:00
oobabooga 6170b5ba31 Bump llama-cpp-python 2023-08-17 21:41:02 -07:00
oobabooga 3e7c624f8e Add a template for OpenOrca-Platypus2 2023-08-17 15:03:08 -07:00
oobabooga eaf5f0f089 Merge remote-tracking branch 'refs/remotes/origin/main' 2023-08-17 11:38:29 -07:00
oobabooga 669c640eb4 Fix extensions block appearing where it shouldn't 2023-08-17 11:37:14 -07:00
ausboss a954b3e7de
fixes error when not specifying tunnel id (#3606) 2023-08-17 15:20:36 -03:00
oobabooga bdb6eb5734 Restyle the chat input box + several CSS improvements
- Remove extra spacing below the last chat message
- Change the background color of code blocks in dark mode
- Remove border radius from selected header bar elements
- Make the chat scrollbar more discrete
2023-08-17 11:10:38 -07:00
oobabooga cebe07f29c Unescape HTML inside code blocks 2023-08-16 21:08:26 -07:00
oobabooga 46408465ce
Improved chat scrolling (#3601) 2023-08-17 01:03:40 -03:00
oobabooga 300219b081 Fix <audio> tag width in chat mode 2023-08-16 09:35:10 -07:00
oobabooga a4e903e932 Escape HTML in chat messages 2023-08-16 09:25:52 -07:00
oobabooga 7966989667 Minor CSS fix 2023-08-16 07:25:59 -07:00
oobabooga 73d9befb65 Make "Show controls" customizable through settings.yaml 2023-08-16 07:04:18 -07:00
oobabooga 2a29208224
Add a "Show controls" button to chat UI (#3590) 2023-08-16 02:39:58 -03:00
oobabooga 32ff3da941
Update ancient screenshots 2023-08-15 17:16:24 -03:00
oobabooga 87dd85b719 Update README 2023-08-15 12:21:50 -07:00
oobabooga a03a70bed6 Update README 2023-08-15 12:20:59 -07:00
oobabooga 7089b2a48f Update README 2023-08-15 12:16:21 -07:00
oobabooga 155862a4a0 Update README 2023-08-15 12:11:12 -07:00
cal066 991bb57e43
ctransformers: Fix up model_type name consistency (#3567) 2023-08-14 15:17:24 -03:00
oobabooga ccfc02a28d
Add the --disable_exllama option for AutoGPTQ (#3545 from clefever/disable-exllama) 2023-08-14 15:15:55 -03:00
oobabooga 8294eadd38 Bump AutoGPTQ wheel 2023-08-14 11:13:46 -07:00
oobabooga 7e57b35b5e Clean up old code 2023-08-14 10:10:39 -07:00
oobabooga 4d067e9b52 Add back a variable to keep old extensions working 2023-08-14 09:39:06 -07:00
oobabooga b74bf5638b Install extensions dependencies before webui dependencies
webui takes precedence over extensions.
2023-08-14 09:15:25 -07:00
oobabooga d8a82d34ed Improve a warning 2023-08-14 08:46:05 -07:00
oobabooga 3e0a9f9cdb Refresh the character dropdown when saving/deleting a character 2023-08-14 08:23:41 -07:00
oobabooga 890b4abdad Fix session saving 2023-08-14 07:55:52 -07:00
oobabooga 619cb4e78b
Add "save defaults to settings.yaml" button (#3574) 2023-08-14 11:46:07 -03:00
oobabooga a95e6f02cb Add a placeholder for custom stopping strings 2023-08-13 21:17:20 -07:00
oobabooga ff9b5861c8 Fix impersonate when some text is present (closes #3564) 2023-08-13 21:10:47 -07:00
oobabooga cc7e6ef645 Fix a CSS conflict 2023-08-13 19:24:09 -07:00
Eve 66c04c304d
Various ctransformers fixes (#3556)
---------

Co-authored-by: cal066 <cal066@users.noreply.github.com>
2023-08-13 23:09:03 -03:00
oobabooga b8df4a436e Scroll up when switching tabs 2023-08-13 18:48:15 -07:00
oobabooga c269214219 CSS change to make buttons smaller 2023-08-13 18:45:13 -07:00
oobabooga 4a05aa92cb Add "send to" buttons for instruction templates
- Remove instruction templates from prompt dropdowns (default/notebook)
- Add 3 buttons to Parameters > Instruction template as a replacement
- Increase the number of lines of 'negative prompt' field to 3, and add a scrollbar
- When uploading a character, switch to the Character tab
- When uploading chat history, switch to the Chat tab
2023-08-13 18:35:45 -07:00
oobabooga 3ae2cee446 Fix empty space when the gallery is hidden 2023-08-13 06:09:27 -07:00
oobabooga f6db2c78d1 Fix ctransformers seed 2023-08-13 05:48:53 -07:00
oobabooga 919a3cf9d0 Fix the gallery 2023-08-13 05:43:09 -07:00
oobabooga 689f264979 Fix permission 2023-08-12 21:14:37 -07:00
oobabooga f7ad634634 Remove --chat flag 2023-08-12 21:13:50 -07:00
oobabooga a1a9ec895d
Unify the 3 interface modes (#3554) 2023-08-13 01:12:15 -03:00
cal066 bf70c19603
ctransformers: move thread and seed parameters (#3543) 2023-08-13 00:04:03 -03:00
jllllll 73421b1fed
Bump ctransformers wheel version (#3558) 2023-08-12 23:02:47 -03:00
Chris Lefever 0230fa4e9c Add the --disable_exllama option for AutoGPTQ 2023-08-12 02:26:58 -04:00
oobabooga 0e05818266 Style changes 2023-08-11 16:35:57 -07:00
oobabooga 4c450e6b70
Update README.md 2023-08-11 15:50:16 -03:00
oobabooga 2f918ccf7c Remove unused parameter 2023-08-11 11:15:22 -07:00
oobabooga 28c8df337b Add repetition_penalty_range to ctransformers 2023-08-11 11:04:19 -07:00
cal066 7a4fcee069
Add ctransformers support (#3313)
---------

Co-authored-by: cal066 <cal066@users.noreply.github.com>
Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
Co-authored-by: randoentity <137087500+randoentity@users.noreply.github.com>
2023-08-11 14:41:33 -03:00
oobabooga 8dbaa20ca8 Don't replace last reply with an empty message 2023-08-10 13:14:48 -07:00
oobabooga 949c92d7df
Create README.md 2023-08-10 14:32:40 -03:00
oobabooga 0789554f65 Allow --lora to use an absolute path 2023-08-10 10:03:12 -07:00
oobabooga 3929971b66 Don't show oobabooga_llama-tokenizer in the model dropdown 2023-08-10 10:02:48 -07:00
Gennadij e12a1852d9
Add Vicuna-v1.5 detection (#3524) 2023-08-10 13:42:24 -03:00
jllllll 28e3ce4317
Simplify GPTQ-for-LLaMa installation (#122) 2023-08-10 13:19:47 -03:00
oobabooga e3d2ddd170
Streamline GPTQ-for-LLaMa support (#3526 from jllllll/gptqllama) 2023-08-10 12:54:59 -03:00
oobabooga c7f52bbdc1 Revert "Remove GPTQ-for-LLaMa monkey patch support"
This reverts commit e3d3565b2a.
2023-08-10 08:39:41 -07:00
oobabooga 16e2b117b4 Minor doc change 2023-08-10 08:38:10 -07:00
jllllll d6765bebc4
Update installation documentation 2023-08-10 00:53:48 -05:00
jllllll d7ee4c2386
Remove unused import 2023-08-10 00:10:14 -05:00
jllllll e3d3565b2a
Remove GPTQ-for-LLaMa monkey patch support
AutoGPTQ will be the preferred GPTQ LoRa loader in the future.
2023-08-09 23:59:04 -05:00
jllllll bee73cedbd
Streamline GPTQ-for-LLaMa support 2023-08-09 23:42:34 -05:00
oobabooga a3295dd666 Detect n_gqa and prompt template for wizardlm-70b 2023-08-09 10:51:16 -07:00
oobabooga a4e48cbdb6 Bump AutoGPTQ 2023-08-09 08:31:17 -07:00
oobabooga 7c1300fab5 Pin aiofiles version to fix statvfs issue 2023-08-09 08:07:55 -07:00
oobabooga 6c6a52aaad Change the filenames for caches and histories 2023-08-09 07:47:19 -07:00
oobabooga 2255349f19 Update README 2023-08-09 05:46:25 -07:00
GiganticPrime 5bfcfcfc5a
Added the logic for starchat model series (#3185) 2023-08-09 09:26:12 -03:00
oobabooga fa4a948b38
Allow users to write one flag per line in CMD_FLAGS.txt 2023-08-09 01:58:23 -03:00
oobabooga d8fb506aff Add RoPE scaling support for transformers (including dynamic NTK)
https://github.com/huggingface/transformers/pull/24653
2023-08-08 21:25:48 -07:00
Hans Raaf f4caaf337a
Fix superbooga when using regenerate (#3362) 2023-08-08 23:26:28 -03:00
Friedemann Lipphardt 901b028d55
Add option for named cloudflare tunnels (#3364) 2023-08-08 22:20:27 -03:00
oobabooga 4ba30f6765 Add OpenChat template 2023-08-08 14:10:04 -07:00
oobabooga bf08b16b32 Fix disappearing profile picture bug 2023-08-08 14:09:01 -07:00
Gennadij 0e78f3b4d4
Fixed a typo in "rms_norm_eps", incorrectly set as n_gqa (#3494) 2023-08-08 00:31:11 -03:00
oobabooga 37fb719452
Increase the Context/Greeting boxes sizes 2023-08-08 00:09:00 -03:00
oobabooga 6d354bb50b
Allow the webui to do multiple tasks simultaneously 2023-08-07 23:57:25 -03:00
oobabooga 584dd33424
Fix missing example_dialogue when uploading characters 2023-08-07 23:44:59 -03:00
oobabooga bbe4a29a25
Add back dark theme code 2023-08-07 23:03:09 -03:00
oobabooga 2d0634cd07 Bump transformers commit for positive prompts 2023-08-07 08:57:19 -07:00
Sam 3b27404865
Make dockerfile respect specified cuda version (#3474) 2023-08-07 10:19:16 -03:00
oobabooga 412f6ff9d3 Change alpha_value maximum and step 2023-08-07 06:08:51 -07:00
oobabooga a373c96d59 Fix a bug in modules/shared.py 2023-08-06 20:36:35 -07:00
jllllll 2cf64474f2
Use chat_instruct_command in API (#3482) 2023-08-06 23:46:25 -03:00
oobabooga 3d48933f27 Remove ancient deprecation warnings 2023-08-06 18:58:59 -07:00
oobabooga c237ce607e Move characters/instruction-following to instruction-templates 2023-08-06 17:50:32 -07:00
oobabooga 65aa11890f
Refactor everything (#3481) 2023-08-06 21:49:27 -03:00
oobabooga d4b851bdc8 Credit turboderp 2023-08-06 13:43:15 -07:00
oobabooga 0af10ab49b
Add Classifier Free Guidance (CFG) for Transformers/ExLlama (#3325) 2023-08-06 17:22:48 -03:00
missionfloyd 5134878344
Fix chat message order (#3461) 2023-08-05 13:53:54 -03:00
jllllll 44f31731af
Create logs dir if missing when saving history (#3462) 2023-08-05 13:47:16 -03:00
jllllll 5ee95d126c
Bump exllama wheels to 0.0.10 (#3467) 2023-08-05 13:46:14 -03:00
Forkoz 9dcb37e8d4
Fix: Mirostat fails on models split across multiple GPUs 2023-08-05 13:45:47 -03:00
jllllll 9e17325207
Add CMD_FLAGS.txt functionality to WSL installer (#119) 2023-08-05 10:26:24 -03:00
SodaPrettyCold 23055b21ee
[Bug fix] Remove html tags form the Prompt sent to Stable Diffusion (#3151) 2023-08-04 20:20:28 -03:00
jllllll 6e30f76ba5
Bump bitsandbytes to 0.41.1 (#3457) 2023-08-04 19:28:59 -03:00
oobabooga 8df3cdfd51
Add SSL certificate support (#3453) 2023-08-04 13:57:31 -03:00
oobabooga ed57a79c6e
Add back silero preview by @missionfloyd (#3446) 2023-08-04 02:29:14 -03:00
missionfloyd 2336b75d92
Remove unnecessary chat.js (#3445) 2023-08-04 01:58:37 -03:00
oobabooga 4b3384e353 Handle unfinished lists during markdown streaming 2023-08-03 17:15:18 -07:00
Pete f4005164f4
Fix llama.cpp truncation (#3400)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-08-03 20:01:15 -03:00
oobabooga 4e6dc6d99d Add Contributing guidelines 2023-08-03 14:40:28 -07:00
matatonic 8f98268252
extensions/openai: include content-length for json replies (#3416) 2023-08-03 16:10:49 -03:00
matatonic 32e7cbb635
More models: +StableBeluga2 (#3415) 2023-08-03 16:02:54 -03:00
Paul DeCarlo f61573bbde
Add standalone Dockerfile for NVIDIA Jetson (#3336) 2023-08-03 15:57:33 -03:00
rafa-9 d578baeb2c
Use character settings from API properties if present (#3428) 2023-08-03 15:56:40 -03:00
oobabooga 601fc424cd
Several improvements (#117) 2023-08-03 14:39:46 -03:00
oobabooga d93087adc3 Merge remote-tracking branch 'refs/remotes/origin/main' 2023-08-03 08:14:10 -07:00
oobabooga 1839dff763 Use Esc to Stop the generation 2023-08-03 08:13:17 -07:00
oobabooga 87dab03dc0
Add the --cpu option for llama.cpp to prevent CUDA from being used (#3432) 2023-08-03 11:00:36 -03:00
oobabooga 3e70bce576 Properly format exceptions in the UI 2023-08-03 06:57:21 -07:00
oobabooga 3390196a14 Add some javascript alerts for confirmations 2023-08-02 22:15:20 -07:00
oobabooga e074538b58 Revert "Make long_replies ban the eos token as well"
This reverts commit 6c521ce967.
2023-08-02 21:45:10 -07:00
oobabooga 6bf9e855f8 Minor change 2023-08-02 21:41:38 -07:00
oobabooga 32c564509e Fix loading session in chat mode 2023-08-02 21:13:16 -07:00
oobabooga 4b6c1d3f08 CSS change 2023-08-02 20:20:23 -07:00
oobabooga 0e8f9354b5 Add direct download for session/chat history JSONs 2023-08-02 19:43:39 -07:00
jllllll aca5679968
Properly fix broken gcc_linux-64 package (#115) 2023-08-02 23:39:07 -03:00
oobabooga 32a2bbee4a Implement auto_max_new_tokens for ExLlama 2023-08-02 11:03:56 -07:00
oobabooga e931844fe2
Add auto_max_new_tokens parameter (#3419) 2023-08-02 14:52:20 -03:00
oobabooga 0d9932815c Improve TheEncrypted777 on mobile devices 2023-08-02 09:15:54 -07:00
Pete 6afc1a193b
Add a scrollbar to notebook/default, improve chat scrollbar style (#3403)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-08-02 12:02:36 -03:00
oobabooga 6c521ce967 Make long_replies ban the eos token as well 2023-08-01 18:47:49 -07:00
matatonic 9ae0eab989
extensions/openai: +Array input (batched) , +Fixes (#3309) 2023-08-01 22:26:00 -03:00
CrazyShipOne 40038fdb82
add chat instruction config for BaiChuan model (#3332) 2023-08-01 22:25:20 -03:00
oobabooga c8a59d79be Add a template for NewHope 2023-08-01 13:27:29 -07:00
oobabooga b53ed70a70 Make llamacpp_HF 6x faster 2023-08-01 13:18:20 -07:00
oobabooga 385229313f Increase the interface area a bit 2023-08-01 09:41:57 -07:00
oobabooga 8d46a8c50a Change the default chat style and the default preset 2023-08-01 09:35:17 -07:00
oobabooga 9773534181 Update Chat-mode.md 2023-08-01 08:03:22 -07:00
oobabooga 959feba602 When saving model settings, only save the settings for the current loader 2023-08-01 06:10:09 -07:00
oobabooga ebb4f22028 Change a comment 2023-07-31 20:06:10 -07:00
oobabooga 8e2217a029 Minor changes to the Parameters tab 2023-07-31 19:55:11 -07:00
oobabooga b2207f123b Update docs 2023-07-31 19:20:48 -07:00
oobabooga f094330df0 When saving a preset, only save params that differ from the defaults 2023-07-31 19:13:29 -07:00
oobabooga 84297d05c4 Add a "Filter by loader" menu to the Parameters tab 2023-07-31 19:09:02 -07:00
oobabooga abea8d9ad3 Make settings-template.yaml more readable 2023-07-31 12:01:50 -07:00
oobabooga 7de7b3d495 Fix newlines in exported character yamls 2023-07-31 10:46:02 -07:00
oobabooga d06c34dea5
Add an extension that makes chat replies longer (#3363) 2023-07-31 13:34:41 -03:00
oobabooga e6be25ea11 Fix a regression 2023-07-30 18:12:30 -07:00
oobabooga 5ca37765d3 Only replace {{user}} and {{char}} at generation time 2023-07-30 11:42:30 -07:00
oobabooga 6e16af34fd Save uploaded characters as yaml
Also allow yaml characters to be uploaded directly
2023-07-30 11:25:38 -07:00
oobabooga c25602eb65 Merge branch 'dev' 2023-07-30 08:47:50 -07:00
oobabooga ca4188aabc Update the example extension 2023-07-29 18:57:22 -07:00
jllllll c4e14a757c
Bump exllama module to 0.0.9 (#3338) 2023-07-29 22:16:23 -03:00
jllllll ecd92d6a4e
Remove unused variable from ROCm GPTQ install (#107) 2023-07-26 22:16:36 -03:00
jllllll 1e3c950c7d
Add AMD GPU support for Linux (#98) 2023-07-26 17:33:02 -03:00
GuizzyQC 4b37a2b397
sd_api_pictures: Widen sliders for image size minimum and maximum (#3326) 2023-07-26 13:49:46 -03:00
oobabooga d6314fd539 Change a comment 2023-07-26 09:38:45 -07:00
oobabooga f24f87cfb0 Change a comment 2023-07-26 09:38:13 -07:00
oobabooga de5de045e0 Set rms_norm_eps to 5e-6 for every llama-2 ggml model, not just 70b 2023-07-26 08:26:56 -07:00
oobabooga 193c6be39c Add missing \n to llama-v2 template context 2023-07-26 08:26:56 -07:00
oobabooga ec68d5211e Set rms_norm_eps to 5e-6 for every llama-2 ggml model, not just 70b 2023-07-26 08:23:24 -07:00
oobabooga a9e10753df Add missing \n to llama-v2 template context 2023-07-26 07:59:49 -07:00
oobabooga b780d520d2 Add a link to the gradio docs 2023-07-26 07:49:42 -07:00
oobabooga b553c33dd0 Add a link to the gradio docs 2023-07-26 07:49:22 -07:00
oobabooga d94ba6e68b Define visible_text before applying chat_input extensions 2023-07-26 07:30:25 -07:00
oobabooga b31321c779 Define visible_text before applying chat_input extensions 2023-07-26 07:27:14 -07:00
oobabooga b17893a58f Revert "Add tensor split support for llama.cpp (#3171)"
This reverts commit 031fe7225e.
2023-07-26 07:06:01 -07:00
oobabooga 517d40cffe Update Extensions.md 2023-07-26 07:01:35 -07:00
oobabooga b11f63cb18 update extensions docs 2023-07-26 07:00:33 -07:00
jllllll 52e3b91f5e
Fix broken gxx_linux-64 package. (#106) 2023-07-26 01:55:08 -03:00
oobabooga 4a24849715 Revert changes 2023-07-25 21:09:32 -07:00
oobabooga 69f8b35bc9 Revert changes to README 2023-07-25 20:51:19 -07:00
oobabooga ed80a2e7db Reorder llama.cpp params 2023-07-25 20:45:20 -07:00
oobabooga 0e8782df03 Set instruction template when switching from default/notebook to chat 2023-07-25 20:37:01 -07:00
oobabooga 28779cd959 Use dark theme by default 2023-07-25 20:11:57 -07:00
oobabooga c2e0d46616 Add credits 2023-07-25 15:49:04 -07:00
oobabooga 1b89c304ad Update README 2023-07-25 15:46:12 -07:00
oobabooga d3abe7caa8 Update llama.cpp.md 2023-07-25 15:33:16 -07:00
oobabooga 863d2f118f Update llama.cpp.md 2023-07-25 15:31:05 -07:00
oobabooga 77d2e9f060 Remove flexgen 2 2023-07-25 15:18:25 -07:00
oobabooga 75c2dd38cf Remove flexgen support 2023-07-25 15:15:29 -07:00
oobabooga 5134d5b1c6 Update README 2023-07-25 15:13:07 -07:00
Foxtr0t1337 85b3a26e25
Ignore values which are not string in training.py (#3287) 2023-07-25 19:00:25 -03:00
Shouyi 031fe7225e
Add tensor split support for llama.cpp (#3171) 2023-07-25 18:59:26 -03:00
Eve f653546484
README updates and improvements (#3198) 2023-07-25 18:58:13 -03:00
Ikko Eltociear Ashimine b09e4f10fd
Fix typo in README.md (#3286)
tranformers -> transformers
2023-07-25 18:56:25 -03:00
oobabooga 7bc408b472 Change rms_norm_eps to 5e-6 for llama-2-70b ggml
Based on https://github.com/ggerganov/llama.cpp/pull/2384
2023-07-25 14:54:57 -07:00
oobabooga ef8637e32d
Add extension example, replace input_hijack with chat_input_modifier (#3307) 2023-07-25 18:49:56 -03:00
oobabooga 08c622df2e Autodetect rms_norm_eps and n_gqa for llama-2-70b 2023-07-24 15:27:34 -07:00
oobabooga a07d070b6c
Add llama-2-70b GGML support (#3285) 2023-07-24 16:37:03 -03:00
oobabooga 6f4830b4d3 Bump peft commit 2023-07-24 09:49:57 -07:00
matatonic 90a4ab631c
extensions/openai: Fixes for: embeddings, tokens, better errors. +Docs update, +Images, +logit_bias/logprobs, +more. (#3122) 2023-07-24 11:28:12 -03:00
jllllll 1141987a0d
Add checks for ROCm and unsupported architectures to llama_cpp_cuda loading (#3225) 2023-07-24 11:25:36 -03:00
iongpt 74fc5dd873
Add user-agent to download-model.py requests (#3243) 2023-07-24 11:19:13 -03:00
Ikko Eltociear Ashimine b2d5433409
Fix typo in deepspeed_parameters.py (#3222)
configration -> configuration
2023-07-24 11:17:28 -03:00
jllllll eb105b0495
Bump llama-cpp-python to 0.1.74 (#3257) 2023-07-24 11:15:42 -03:00
jllllll 152cf1e8ef
Bump bitsandbytes to 0.41.0 (#3258)
e229fbce66...a06a0f6a08
2023-07-24 11:06:18 -03:00
jllllll 8d31d20c9a
Bump exllama module to 0.0.8 (#3256)
39b3541cdd...3f83ebb378
2023-07-24 11:05:54 -03:00
oobabooga cc2ed46d44
Make chat the default again 2023-07-20 18:55:09 -03:00
jllllll fcb215fed5
Add check for compute support for GPTQ-for-LLaMa (#104)
Installs from main cuda repo if fork not supported
Also removed cuBLAS llama-cpp-python installation in preperation for 4b19b74e6c
2023-07-20 11:11:00 -03:00
oobabooga 63ece46213 Merge branch 'main' into dev 2023-07-20 07:06:41 -07:00
oobabooga 6415cc68a2 Remove obsolete information from README 2023-07-19 21:20:40 -07:00
oobabooga 4b19b74e6c Add CUDA wheels for llama-cpp-python by jllllll 2023-07-19 19:33:43 -07:00
oobabooga 05f4cc63c8 Merge branch 'main' into dev 2023-07-19 19:22:34 -07:00
jllllll 4df3f72753
Fix GPTQ fail message not being shown on update (#103) 2023-07-19 22:25:09 -03:00
jllllll 87926d033d
Bump exllama module to 0.0.7 (#3211) 2023-07-19 22:24:47 -03:00
oobabooga 913e060348 Change the default preset to Divine Intellect
It seems to reduce hallucination while using instruction-tuned models.
2023-07-19 08:24:37 -07:00
oobabooga 0d7f43225f Merge branch 'dev' 2023-07-19 07:20:13 -07:00
oobabooga 08c23b62c7 Bump llama-cpp-python and transformers 2023-07-19 07:19:12 -07:00
oobabooga 5447e75191 Merge branch 'dev' 2023-07-18 15:36:26 -07:00
oobabooga 8ec225f245 Add EOS/BOS tokens to Llama-2 template
Following this comment:
https://github.com/ggerganov/llama.cpp/issues/2262#issuecomment-1641063329
2023-07-18 15:35:27 -07:00
oobabooga 3ef49397bb
Merge pull request #3195 from oobabooga/dev
v1.3
2023-07-18 17:33:11 -03:00
oobabooga 070a886278 Revert "Prevent lists from flickering in chat mode while streaming"
This reverts commit 5e5d926d2b.
2023-07-18 13:23:29 -07:00
oobabooga a2918176ea Update LLaMA-v2-model.md (thanks Panchovix) 2023-07-18 13:21:18 -07:00
oobabooga e0631e309f
Create instruction template for Llama-v2 (#3194) 2023-07-18 17:19:18 -03:00
oobabooga 603c596616 Add LLaMA-v2 conversion instructions 2023-07-18 10:29:56 -07:00
jllllll c535f14e5f
Bump bitsandbytes Windows wheel to 0.40.2 (#3186) 2023-07-18 11:39:43 -03:00
jllllll d7a14174a2
Remove auto-loading when only one model is available (#3187) 2023-07-18 11:39:08 -03:00
randoentity a69955377a
[GGML] Support for customizable RoPE (#3083)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-07-17 22:32:37 -03:00
appe233 89e0d15cf5
Use 'torch.backends.mps.is_available' to check if mps is supported (#3164) 2023-07-17 21:27:18 -03:00
dependabot[bot] 234c58ccd1
Bump bitsandbytes from 0.40.1.post1 to 0.40.2 (#3178) 2023-07-17 21:24:51 -03:00
oobabooga 49a5389bd3
Bump accelerate from 0.20.3 to 0.21.0 2023-07-17 21:23:59 -03:00
oobabooga 8c1c2e0fae Increase max_new_tokens upper limit 2023-07-17 17:08:22 -07:00
oobabooga 5e5d926d2b Prevent lists from flickering in chat mode while streaming 2023-07-17 17:00:49 -07:00
dependabot[bot] 02a5fe6aa2
Bump accelerate from 0.20.3 to 0.21.0
Bumps [accelerate](https://github.com/huggingface/accelerate) from 0.20.3 to 0.21.0.
- [Release notes](https://github.com/huggingface/accelerate/releases)
- [Commits](https://github.com/huggingface/accelerate/compare/v0.20.3...v0.21.0)

---
updated-dependencies:
- dependency-name: accelerate
  dependency-type: direct:production
  update-type: version-update:semver-minor
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-07-17 20:18:31 +00:00
oobabooga 60a3e70242 Update LLaMA links and info 2023-07-17 12:51:01 -07:00
oobabooga f83fdb9270 Don't reset LoRA menu when loading a model 2023-07-17 12:50:25 -07:00
oobabooga 4ce766414b Bump AutoGPTQ version 2023-07-17 10:02:12 -07:00
oobabooga b1a6ea68dd Disable "autoload the model" by default 2023-07-17 07:40:56 -07:00
oobabooga 656b457795 Add Airoboros-v1.2 template 2023-07-17 07:27:42 -07:00
oobabooga a199f21799 Optimize llamacpp_hf a bit 2023-07-16 20:49:48 -07:00
oobabooga 9f08038864
Merge pull request #3163 from oobabooga/dev
v1.2
2023-07-16 02:43:18 -03:00
oobabooga 6a3edb0542 Clean up llamacpp_hf.py 2023-07-15 22:40:55 -07:00
oobabooga 2de0cedce3 Fix reload screen color 2023-07-15 22:39:39 -07:00
oobabooga 13449aa44d Decrease download timeout 2023-07-15 22:30:08 -07:00
oobabooga 27a84b4e04 Make AutoGPTQ the default again
Purely for compatibility with more models.
You should still use ExLlama_HF for LLaMA models.
2023-07-15 22:29:23 -07:00
oobabooga 5e3f7e00a9
Create llamacpp_HF loader (#3062) 2023-07-16 02:21:13 -03:00
Panchovix 7c4d4fc7d3
Increase alpha value limit for NTK RoPE scaling for exllama/exllama_HF (#3149) 2023-07-16 01:56:04 -03:00
ofirkris 780a2f2e16
Bump llama cpp version (#3160)
Bump llama cpp version to support better 8K RoPE scaling

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-07-16 01:54:56 -03:00
jllllll ed3ffd212d
Bump bitsandbytes to 0.40.1.post1 (#3156)
817bdf6325...6ec4f0c374
2023-07-16 01:53:32 -03:00
oobabooga 94dfcec237
Make it possible to evaluate exllama perplexity (#3138) 2023-07-16 01:52:55 -03:00
jllllll 11a8fd1eb9
Add cuBLAS llama-cpp-python wheel installation (#102)
Parses requirements.txt using regex to determine required version.
2023-07-16 01:31:33 -03:00
oobabooga b284f2407d Make ExLlama_HF the new default for GPTQ 2023-07-14 14:03:56 -07:00
jllllll 32f12b8bbf
Bump bitsandbytes Windows wheel to 0.40.0.post4 (#3135) 2023-07-13 17:32:37 -03:00
SeanScripts 9800745db9
Color tokens by probability and/or perplexity (#3078) 2023-07-13 17:30:22 -03:00
oobabooga 146e8b2a6c Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2023-07-13 13:23:38 -07:00
Morgan Schweers 6d1e911577
Add support for logits processors in extensions (#3029) 2023-07-13 17:22:41 -03:00
oobabooga 22341e948d Merge branch 'main' into dev 2023-07-12 14:19:49 -07:00
oobabooga 0e6295886d Fix lora download folder 2023-07-12 14:19:33 -07:00
oobabooga eb823fce96 Fix typo 2023-07-12 13:55:19 -07:00
oobabooga d0a626f32f Change reload screen color 2023-07-12 13:54:43 -07:00
oobabooga c592a9b740 Fix #3117 2023-07-12 13:33:44 -07:00
oobabooga 6447b2eea6
Merge pull request #3116 from oobabooga/dev
v1.1
2023-07-12 15:55:40 -03:00
oobabooga 2463d7c098 Spaces 2023-07-12 11:35:43 -07:00
oobabooga e202190c4f lint 2023-07-12 11:33:25 -07:00
FartyPants 9b55d3a9f9
More robust and error prone training (#3058) 2023-07-12 15:29:43 -03:00
oobabooga 30f37530d5 Add back .replace('\r', '') 2023-07-12 09:52:20 -07:00
Fernando Tarin Morales 987d0fe023
Fix: Fixed the tokenization process of a raw dataset and improved its efficiency (#3035) 2023-07-12 12:05:37 -03:00
kabachuha 3f19e94c93
Add Tensorboard/Weights and biases integration for training (#2624) 2023-07-12 11:53:31 -03:00
kizinfo 5d513eea22
Add ability to load all text files from a subdirectory for training (#1997)
* Update utils.py

returns individual txt files and subdirectories to getdatasets to allow for training from a directory of text files

* Update training.py

minor tweak to training on raw datasets to detect if a directory is selected, and if so, to load in all the txt files in that directory for training

* Update put-trainer-datasets-here.txt

document

* Minor change

* Use pathlib, sort by natural keys

* Space

---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-07-12 11:44:30 -03:00
practicaldreamer 73a0def4af
Add Feature to Log Sample of Training Dataset for Inspection (#1711) 2023-07-12 11:26:45 -03:00
oobabooga b6ba68eda9 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2023-07-12 07:19:34 -07:00
oobabooga a17b78d334 Disable wandb during training 2023-07-12 07:19:12 -07:00
Gabriel Pena eedb3bf023
Add low vram mode on llama cpp (#3076) 2023-07-12 11:05:13 -03:00
oobabooga 180420d2c9 Fix send_pictures extension 2023-07-11 20:56:01 -07:00
original-subliminal-thought-criminal ad07839a7b
Small bug, when arbitrary loading character.json that doesn't exist (#2643)
* Fixes #2482

* corrected erroronius variable

* Use .exists()

---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-07-12 00:16:36 -03:00
Axiom Wolf d986c17c52
Chat history download creates more detailed file names (#3051) 2023-07-12 00:10:36 -03:00
atriantafy d9fabdde40
Add context_instruct to API. Load default model instruction template … (#2688) 2023-07-12 00:01:03 -03:00
Salvador E. Tropea 324e45b848
[Fixed] wbits and groupsize values from model not shown (#2977) 2023-07-11 23:27:38 -03:00
oobabooga e3810dff40 Style changes 2023-07-11 18:49:06 -07:00
oobabooga bfafd07f44 Change a message 2023-07-11 18:29:20 -07:00
oobabooga a12dae51b9 Bump bitsandbytes 2023-07-11 18:29:08 -07:00
Keith Kjer 37bffb2e1a
Add reference to new pipeline in multimodal readme (#2947) 2023-07-11 19:04:15 -03:00
Juliano Henriquez 1fc0b5041e
substitu superboog Beatiful Soup Parser (#2996)
* add lxml to requirments

add lxml to requirments

* Change Beaitful Soup Parser

"lxml" parser which might be more tolerant of certain kinds of parsing errors than "html.parser" and quicker at the same time.
2023-07-11 19:02:49 -03:00
Salvador E. Tropea ab044a5a44
Elevenlabs tts fixes (#2959)
* [Fixed] Keep setting option for the voice

- It was always changed to the first available voice
- Also added an error if the selected voice isn't valid

* [Fixed] elevenlabs_tts API key handling

- The one from the settings wasn't applied
- We always got "Enter your API key", even when the settings specified
  an api_key

* [Added] elevenlabs_tts model selection

- Now we can also use the "eleven_multilingual_v1" model.
  Used for anything but english.
2023-07-11 19:00:37 -03:00
micsthepick 3708de2b1f
respect model dir for downloads (#3077) (#3079) 2023-07-11 18:55:46 -03:00
matatonic 3778816b8d
models/config.yaml: +platypus/gplatty, +longchat, +vicuna-33b, +Redmond-Hermes-Coder, +wizardcoder, +more (#2928)
* +platypus/gplatty

* +longchat, +vicuna-33b, +Redmond-Hermes-Coder

* +wizardcoder

* +superplatty

* +Godzilla, +WizardLM-V1.1, +rwkv 8k,
+wizard-mega fix </s>

---------

Co-authored-by: Matthew Ashton <mashton-gitlab@zhero.org>
2023-07-11 18:53:48 -03:00
Ricardo Pinto 3e9da5a27c
Changed FormComponent to IOComponent (#3017)
Co-authored-by: Ricardo Pinto <1-ricardo.pinto@users.noreply.gitlab.cognitage.com>
2023-07-11 18:52:16 -03:00
matatonic 3e7feb699c
extensions/openai: Major openai extension updates & fixes (#3049)
* many openai updates

* total reorg & cleanup.

* fixups

* missing import os for images

* +moderations, custom_stopping_strings, more fixes

* fix bugs in completion streaming

* moderation fix (flagged)

* updated moderation categories

---------

Co-authored-by: Matthew Ashton <mashton-gitlab@zhero.org>
2023-07-11 18:50:08 -03:00
Ahmad Fahadh Ilyas 8db7e857b1
Add token authorization for downloading model (#3067) 2023-07-11 18:48:08 -03:00
FartyPants 61102899cd
google flan T5 download fix (#3080) 2023-07-11 18:46:59 -03:00
jllllll fdd596f98f
Bump bitsandbytes Windows wheel (#3097) 2023-07-11 18:41:24 -03:00
Vadim Peretokin 987d522b55
Fix API example for loading models (#3101) 2023-07-11 18:40:55 -03:00
Josh XT f4aa11cef6
Add default environment variable values to docker compose file (#3102)
Add default environment variable values to docker compose file
2023-07-11 18:38:26 -03:00
ofirkris a81cdd1367
Bump cpp llama version (#3081)
Bump cpp llama version to 0.1.70
2023-07-10 19:36:15 -03:00
jllllll f8dbd7519b
Bump exllama module version (#3087)
d769533b6f...e61d4d31d4
2023-07-10 19:35:59 -03:00
tianchen zhong c7058afb40
Add new possible bin file name regex (#3070) 2023-07-09 17:22:56 -03:00
ofirkris 161d984e80
Bump llama-cpp-python version (#3072)
Bump llama-cpp-python version to 0.1.69
2023-07-09 17:22:24 -03:00
Salvador E. Tropea 463aac2d65
[Added] google_translate activate param (#2961)
- So you can quickly enable/disable it, otherwise you must select
  English to disable it, and then your language to enable it again.
2023-07-09 01:08:20 -03:00
Forkoz 74ea7522a0
Lora fixes for AutoGPTQ (#2818) 2023-07-09 01:03:43 -03:00
Chris Rude 70b088843d
fix for issue #2475: Streaming api deadlock (#3048) 2023-07-08 23:21:20 -03:00
oobabooga 5ac4e4da8b Make --model work with argument like models/folder_name 2023-07-08 10:22:54 -07:00
Brandon McClure acf24ebb49
Whisper_stt params for model, language, and auto_submit (#3031) 2023-07-07 20:54:53 -03:00
oobabooga 79679b3cfd Pin fastapi version (for #3042) 2023-07-07 16:40:57 -07:00
oobabooga bb79037ebd
Fix wrong pytorch version on Linux+CPU
It was installing nvidia wheels
2023-07-07 20:40:31 -03:00
oobabooga 564a8c507f
Don't launch chat mode by default 2023-07-07 13:32:11 -03:00
oobabooga b6643e5039 Add decode functions to llama.cpp/exllama 2023-07-07 09:11:30 -07:00
oobabooga 1ba2e88551 Add truncation to exllama 2023-07-07 09:09:23 -07:00
oobabooga c21b73ff37 Minor change to ui.py 2023-07-07 09:09:14 -07:00
oobabooga de994331a4 Merge remote-tracking branch 'refs/remotes/origin/main' 2023-07-06 22:25:43 -07:00
oobabooga 9aee1064a3 Block a cloudfare request 2023-07-06 22:24:52 -07:00
Fernando Tarin Morales d7e14e1f78
Fixed the param name when loading a LoRA using a model loaded in 4 or 8 bits (#3036) 2023-07-07 02:24:07 -03:00
Fernando Tarin Morales 1f540fa4f8
Added the format to be able to finetune Vicuna1.1 models (#3037) 2023-07-07 02:22:39 -03:00
Xiaojian "JJ" Deng ff45317032
Update models.py (#3020)
Hopefully fixed error with "ValueError: Tokenizer class GPTNeoXTokenizer does not exist or is not currently 
imported."
2023-07-05 21:40:43 -03:00
ofirkris b67c362735
Bump llama-cpp-python (#3011)
Bump llama-cpp-python to V0.1.68
2023-07-05 11:33:28 -03:00
jeckyhl 88a747b5b9
fix: Error when downloading model from UI (#3014) 2023-07-05 11:27:29 -03:00
oobabooga e0a50fb77a
Merge pull request #2922 from Honkware/main
Load Salesforce Xgen Models
2023-07-04 23:47:21 -03:00
oobabooga 8705eba830 Remove universal llama tokenizer support
Instead replace it with a warning if the tokenizer files look off
2023-07-04 19:43:19 -07:00
oobabooga 84d6c93d0d Merge branch 'main' into Honkware-main 2023-07-04 18:50:07 -07:00
oobabooga 31c297d7e0 Various changes 2023-07-04 18:50:01 -07:00
AN Long be4582be40
Support specify retry times in download-model.py (#2908) 2023-07-04 22:26:30 -03:00
oobabooga 70a4d5dbcf Update chat API (fixes #3006) 2023-07-04 17:36:47 -07:00
oobabooga 333075e726
Fix #3003 2023-07-04 11:38:35 -03:00
oobabooga 40c5722499
Fix #2998 2023-07-04 11:35:25 -03:00
oobabooga 463ddfffd0 Fix start_with 2023-07-03 23:32:02 -07:00
oobabooga 55457549cd Add information about presets to the UI 2023-07-03 22:39:01 -07:00
oobabooga 373555c4fb Fix loading some histories (thanks kaiokendev) 2023-07-03 22:19:28 -07:00
Panchovix 10c8c197bf
Add Support for Static NTK RoPE scaling for exllama/exllama_hf (#2955) 2023-07-04 01:13:16 -03:00
jllllll 1610d5ffb2
Bump exllama module to 0.0.5 (#2993) 2023-07-04 00:15:55 -03:00
FartyPants eb6112d5a2
Update server.py - clear LORA after reload (#2952) 2023-07-04 00:13:38 -03:00
oobabooga 7e8340b14d Make greetings appear in --multi-user mode 2023-07-03 20:08:14 -07:00
oobabooga 4b1804a438
Implement sessions + add basic multi-user support (#2991) 2023-07-04 00:03:30 -03:00
FartyPants 1f8cae14f9
Update training.py - correct use of lora_names (#2988) 2023-07-03 17:41:18 -03:00
FartyPants c23c88ee4c
Update LoRA.py - avoid potential error (#2953) 2023-07-03 17:40:22 -03:00
FartyPants 33f56fd41d
Update models.py to clear LORA names after unload (#2951) 2023-07-03 17:39:06 -03:00
FartyPants 48b11f9c5b
Training: added trainable parameters info (#2944) 2023-07-03 17:38:36 -03:00
Turamarth14 847f70b694
Update html_generator.py (#2954)
With version 10.0.0 of Pillow the constant Image.ANTIALIAS has been removed. Instead Image.LANCZOS should be used.
2023-07-02 01:43:58 -03:00
ardfork 3c076c3c80
Disable half2 for ExLlama when using HIP (#2912) 2023-06-29 15:03:16 -03:00
missionfloyd ac0f96e785
Some more character import tweaks. (#2921) 2023-06-29 14:56:25 -03:00
oobabooga 5d2a8b31be Improve Parameters tab UI 2023-06-29 14:33:47 -03:00
oobabooga 79db629665 Minor bug fix 2023-06-29 13:53:06 -03:00
oobabooga 3443219cbc
Add repetition penalty range parameter to transformers (#2916) 2023-06-29 13:40:13 -03:00
Honkware b9a3d28177 Merge branch 'main' of https://github.com/Honkware/text-generation-webui 2023-06-29 01:33:00 -05:00
Honkware 3147f0b8f8 xgen config 2023-06-29 01:32:53 -05:00
Honkware 0a6a498383 Load xgen tokenizer 2023-06-29 01:32:44 -05:00
Honkware 1d03387f74
Xgen instruction template 2023-06-29 01:31:33 -05:00
oobabooga c6cae106e7 Bump llama-cpp-python 2023-06-28 18:14:45 -03:00
oobabooga 20740ab16e Revert "Fix exllama_hf gibbersh above 2048 context, and works >5000 context. (#2913)"
This reverts commit 37a16d23a7.
2023-06-28 18:10:34 -03:00
jllllll 7b048dcf67
Bump exllama module version to 0.0.4 (#2915) 2023-06-28 18:09:58 -03:00
Panchovix 37a16d23a7
Fix exllama_hf gibbersh above 2048 context, and works >5000 context. (#2913) 2023-06-28 12:36:07 -03:00
oobabooga 63770c0643 Update docs/Extensions.md 2023-06-27 22:25:05 -03:00
matatonic da0ea9e0f3
set +landmark, +superhot-8k to 8k length (#2903) 2023-06-27 22:05:52 -03:00
missionfloyd 5008daa0ff
Add exception handler to load_checkpoint() (#2904) 2023-06-27 22:00:29 -03:00
oobabooga c95009d2bd Merge remote-tracking branch 'refs/remotes/origin/main' 2023-06-27 18:48:17 -03:00
oobabooga 67a83f3ad9 Use DPM++ 2M Karras for Stable Diffusion 2023-06-27 18:47:35 -03:00
FartyPants ab1998146b
Training update - backup the existing adapter before training on top of it (#2902) 2023-06-27 18:24:04 -03:00
Minecrafter20 40bbd53640
Add custom prompt format for SD API pictures (#1964) 2023-06-27 17:49:18 -03:00
missionfloyd cb029cf65f
Get SD samplers from API (#2889) 2023-06-27 17:31:54 -03:00
GuizzyQC d7a7f7896b
Add SD checkpoint selection in sd_api_pictures (#2872) 2023-06-27 17:29:27 -03:00
oobabooga 7611978f7b
Add Community section to README 2023-06-27 13:56:14 -03:00
oobabooga 22d455b072 Add LoRA support to ExLlama_HF 2023-06-26 00:10:33 -03:00
oobabooga b7c627f9a0 Set UI defaults 2023-06-25 22:55:43 -03:00
oobabooga c52290de50
ExLlama with long context (#2875) 2023-06-25 22:49:26 -03:00
oobabooga 9290c6236f Keep ExLlama_HF if already selected 2023-06-25 19:06:28 -03:00
oobabooga 75fd763f99 Fix chat saving issue (closes #2863) 2023-06-25 18:14:57 -03:00
FartyPants 21c189112c
Several Training Enhancements (#2868) 2023-06-25 15:34:46 -03:00
oobabooga 95212edf1f
Update training.py 2023-06-25 12:13:15 -03:00
oobabooga 1f5ea451c9 Merge remote-tracking branch 'refs/remotes/origin/main' 2023-06-25 02:14:19 -03:00
oobabooga f31281a8de Fix loading instruction templates containing literal '\n' 2023-06-25 02:13:26 -03:00
matatonic 68ae5d8262
more models: +orca_mini (#2859) 2023-06-25 01:54:53 -03:00
oobabooga f0fcd1f697 Sort some imports 2023-06-25 01:44:36 -03:00
oobabooga 365b672531 Minor change to prevent future bugs 2023-06-25 01:38:54 -03:00
oobabooga e6e5f546b8 Reorganize Chat settings tab 2023-06-25 01:10:20 -03:00
matatonic b45baeea41
extensions/openai: Major docs update, fix #2852 (critical bug), minor improvements (#2849) 2023-06-24 22:50:04 -03:00
oobabooga ebfcfa41f2
Update ExLlama.md 2023-06-24 20:25:34 -03:00
jllllll bef67af23c
Use pre-compiled python module for ExLlama (#2770) 2023-06-24 20:24:17 -03:00
oobabooga a70a2ac3be
Update ExLlama.md 2023-06-24 20:23:01 -03:00
oobabooga b071eb0d4b
Clean up the presets (#2854) 2023-06-24 18:41:17 -03:00
oobabooga cec5fb0ef6 Failed attempt at evaluating exllama_hf perplexity 2023-06-24 12:02:25 -03:00
快乐的我531 e356f69b36
Make stop_everything work with non-streamed generation (#2848) 2023-06-24 11:19:16 -03:00
oobabooga ec482f3dae Apply input extensions after yielding *Is typing...* 2023-06-24 11:07:11 -03:00
oobabooga 3e80f2aceb Apply the output extensions only once
Relevant for google translate, silero
2023-06-24 10:59:07 -03:00
rizerphe 77baf43f6d
Add CORS support to the API (#2718) 2023-06-24 10:16:06 -03:00
matatonic 8c36c19218
8k size only for minotaur-15B (#2815)
Co-authored-by: Matthew Ashton <mashton-gitlab@zhero.org>
2023-06-24 10:14:19 -03:00
Roman 38897fbd8a
fix: added model parameter check (#2829) 2023-06-24 10:09:34 -03:00
jllllll eac8450ef7
Move special character check to start script (#92)
Also port print_big_message function to batch
2023-06-24 10:06:35 -03:00
missionfloyd 51a388fa34
Organize chat history/character import menu (#2845)
* Organize character import menu

* Move Chat history upload/download labels
2023-06-24 09:55:02 -03:00
oobabooga 8bb3bb39b3
Implement stopping string search in string space (#2847) 2023-06-24 09:43:00 -03:00
oobabooga 0f9088f730 Update README 2023-06-23 12:24:43 -03:00
oobabooga 3ae9af01aa Add --no_use_cuda_fp16 param for AutoGPTQ 2023-06-23 12:22:56 -03:00
Panchovix 5646690769
Fix some models not loading on exllama_hf (#2835) 2023-06-23 11:31:02 -03:00
oobabooga 383c50f05b
Replace old presets with the results of Preset Arena (#2830) 2023-06-23 01:48:29 -03:00
missionfloyd aa1f1ef46a
Fix printing, take two. (#2810)
* Format chat for printing

* Better printing
2023-06-22 16:06:49 -03:00
Panchovix b4a38c24b7
Fix Multi-GPU not working on exllama_hf (#2803) 2023-06-22 16:05:25 -03:00
matatonic d94ea31d54
more models. +minotaur 8k (#2806) 2023-06-21 21:05:08 -03:00
jllllll 04cae3e5db
Remove bitsandbytes compatibility workaround (#91)
New bnb does not need it.
Commented out in case it is needed in the futute.
2023-06-21 15:40:41 -03:00
LarryVRH 580c1ee748
Implement a demo HF wrapper for exllama to utilize existing HF transformers decoding. (#2777) 2023-06-21 15:31:42 -03:00
jllllll a06acd6d09
Update bitsandbytes to 0.39.1 (#2799) 2023-06-21 15:04:45 -03:00
Gaurav Bhagchandani 89fb6f9236
Fixed the ZeroDivisionError when downloading a model (#2797) 2023-06-21 12:31:50 -03:00
matatonic 90be1d9fe1
More models (match more) & templates (starchat-beta, tulu) (#2790) 2023-06-21 12:30:44 -03:00
missionfloyd 2661c9899a
Format chat for printing (#2793) 2023-06-21 10:39:58 -03:00
oobabooga 5dfe0bec06 Remove old/useless code 2023-06-20 23:36:56 -03:00
oobabooga faa92eee8d Add spaces 2023-06-20 23:25:58 -03:00
Peter Sofronas b22c7199c9
Download optimizations (#2786)
* download_model_files metadata writing improvement

* line swap

* reduce line length

* safer download and greater block size

* Minor changes by pycodestyle

---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-06-20 23:14:18 -03:00
Morgan Schweers 447569e31a
Add a download progress bar to the web UI. (#2472)
* Show download progress on the model screen.

* In case of error, mark as done to clear progress bar.

* Increase the iteration block size to reduce overhead.
2023-06-20 22:59:14 -03:00
jllllll d1da22d7ee
Fix -y from previous commit (#90) 2023-06-20 22:48:59 -03:00
oobabooga 80a615c3ae
Add space 2023-06-20 22:48:45 -03:00
oobabooga a2116e8b2b
use uninstall -y 2023-06-20 21:24:01 -03:00
oobabooga c0a1baa46e
Minor changes 2023-06-20 20:23:21 -03:00
jllllll 5cbc0b28f2
Workaround for Peft not updating their package version on the git repo (#88)
* Workaround for Peft not updating their git package version

* Update webui.py

---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-06-20 20:21:10 -03:00
ramblingcoder 0d0d849478
Update Dockerfile to resolve superbooga requirement error (#2401) 2023-06-20 18:31:28 -03:00
jllllll 9bb2fc8cd7
Install Pytorch through pip instead of Conda (#84) 2023-06-20 16:39:23 -03:00
EugeoSynthesisThirtyTwo 7625c6de89
fix usage of self in classmethod (#2781) 2023-06-20 16:18:42 -03:00
MikoAL c40932eb39
Added Falcon LoRA training support (#2684)
I am 50% sure this will work
2023-06-20 01:03:44 -03:00
oobabooga c623e142ac Bump llama-cpp-python 2023-06-20 00:49:38 -03:00
FartyPants ce86f726e9
Added saving of training logs to training_log.json (#2769) 2023-06-20 00:47:36 -03:00
oobabooga 017884132f Merge remote-tracking branch 'refs/remotes/origin/main' 2023-06-20 00:46:29 -03:00
oobabooga e1cd6cc410 Minor style change 2023-06-20 00:46:18 -03:00
Cebtenzzre 59e7ecb198
llama.cpp: implement ban_eos_token via logits_processor (#2765) 2023-06-19 21:31:19 -03:00
oobabooga 0d9d70ec7e Update docs 2023-06-19 12:52:23 -03:00
oobabooga f6a602861e Update docs 2023-06-19 12:51:30 -03:00
oobabooga 5d4b4d15a5
Update Using-LoRAs.md 2023-06-19 12:43:57 -03:00
oobabooga eb30f4441f
Add ExLlama+LoRA support (#2756) 2023-06-19 12:31:24 -03:00
oobabooga a1cac88c19
Update README.md 2023-06-19 01:28:23 -03:00
oobabooga 5f418f6171 Fix a memory leak (credits for the fix: Ph0rk0z) 2023-06-19 01:19:28 -03:00
ThisIsPIRI def3b69002
Fix loading condition for universal llama tokenizer (#2753) 2023-06-18 18:14:06 -03:00
oobabooga 490a1795f0 Bump peft commit 2023-06-18 16:42:11 -03:00
oobabooga 09c781b16f Add modules/block_requests.py
This has become unnecessary, but it could be useful in the future
for other libraries.
2023-06-18 16:31:14 -03:00
oobabooga 687fd2604a Improve code/ul styles in chat mode 2023-06-18 15:52:59 -03:00
oobabooga e8588d7077 Merge remote-tracking branch 'refs/remotes/origin/main' 2023-06-18 15:23:38 -03:00
oobabooga 44f28830d1 Chat CSS: fix ul, li, pre styles + remove redefinitions 2023-06-18 15:20:51 -03:00
Forkoz 3cae1221d4
Update exllama.py - Respect model dir parameter (#2744) 2023-06-18 13:26:30 -03:00
oobabooga 5b4c0155f6 Move a button 2023-06-18 01:56:43 -03:00
oobabooga 0686a2e75f Improve instruct colors in dark mode 2023-06-18 01:44:52 -03:00
oobabooga c5641b65d3 Handle leading spaces properly in ExLllama 2023-06-17 19:35:12 -03:00
matatonic 1e97aaac95
extensions/openai: docs update, model loader, minor fixes (#2557) 2023-06-17 19:15:24 -03:00
matatonic 2220b78e7a
models/config.yaml: +alpacino, +alpasta, +hippogriff, +gpt4all-snoozy, +lazarus, +based, -airoboros 4k (#2580) 2023-06-17 19:14:25 -03:00
jllllll b1d05cbbf6
Install exllama (#83)
* Install exllama

* Handle updating exllama
2023-06-17 19:10:36 -03:00
jllllll 657049d7d0
Fix cmd_macos.sh (#82)
MacOS version of Bash does not support process substitution
2023-06-17 19:09:42 -03:00
jllllll b2483e28d1
Check for special characters in path on Windows (#81)
Display warning message if detected
2023-06-17 19:09:22 -03:00
oobabooga 05a743d6ad Make llama.cpp use tfs parameter 2023-06-17 19:08:25 -03:00
oobabooga e19cbea719 Add a variable to modules/shared.py 2023-06-17 19:02:29 -03:00
oobabooga cbd63eeeff Fix repeated tokens with exllama 2023-06-17 19:02:08 -03:00
oobabooga 766c760cd7 Use gen_begin_reuse in exllama 2023-06-17 18:00:10 -03:00
oobabooga 239b11c94b Minor bug fixes 2023-06-17 17:57:56 -03:00
Bhavika Tekwani d8d29edf54
Install wheel using pip3 (#2719) 2023-06-16 23:46:40 -03:00
Jonathan Yankovich a1ca1c04a1
Update ExLlama.md (#2729)
Add details for configuring exllama
2023-06-16 23:46:25 -03:00
oobabooga b27f83c0e9 Make exllama stoppable 2023-06-16 22:03:23 -03:00
oobabooga 7f06d551a3 Fix streaming callback 2023-06-16 21:44:56 -03:00
oobabooga 1e400218e9 Fix a typo 2023-06-16 21:01:57 -03:00
oobabooga 5f392122fd Add gpu_split param to ExLlama
Adapted from code created by Ph0rk0z. Thank you Ph0rk0z.
2023-06-16 20:49:36 -03:00
oobabooga cb9be5db1c
Update ExLlama.md 2023-06-16 20:40:12 -03:00
oobabooga 83be8eacf0 Minor fix 2023-06-16 20:38:32 -03:00
oobabooga 9f40032d32
Add ExLlama support (#2444) 2023-06-16 20:35:38 -03:00
oobabooga dea43685b0 Add some clarifications 2023-06-16 19:10:53 -03:00
oobabooga 7ef6a50e84
Reorganize model loading UI completely (#2720) 2023-06-16 19:00:37 -03:00
oobabooga 57be2eecdf
Update README.md 2023-06-16 15:04:16 -03:00
Meng-Yuan Huang 772d4080b2
Update llama.cpp-models.md for macOS (#2711) 2023-06-16 00:00:24 -03:00
Tom Jobbins 646b0c889f
AutoGPTQ: Add UI and command line support for disabling fused attention and fused MLP (#2648) 2023-06-15 23:59:54 -03:00
dependabot[bot] 909d8c6ae3
Bump transformers from 4.30.0 to 4.30.2 (#2695) 2023-06-14 19:56:28 -03:00
oobabooga 2b9a6b9259 Merge remote-tracking branch 'refs/remotes/origin/main' 2023-06-14 18:45:24 -03:00
oobabooga 4d508cbe58 Add some checks to AutoGPTQ loader 2023-06-14 18:44:43 -03:00
FartyPants 56c19e623c
Add LORA name instead of "default" in PeftModel (#2689) 2023-06-14 18:29:42 -03:00
oobabooga 134430bbe2 Minor change 2023-06-14 11:34:42 -03:00
oobabooga 474dc7355a Allow API requests to use parameter presets 2023-06-14 11:32:20 -03:00
oobabooga 8936160e54
Add WSL installer to README (thanks jllllll) 2023-06-13 00:07:34 -03:00
jllllll c42f183d3f
Installer for WSL (#78) 2023-06-13 00:04:15 -03:00
FartyPants 9f150aedc3
A small UI change in Models menu (#2640) 2023-06-12 01:24:44 -03:00
oobabooga da5d9a28d8 Fix tabbed extensions showing up at the bottom of the UI 2023-06-11 21:20:51 -03:00
oobabooga ae5e2b3470 Reorganize a bit 2023-06-11 19:50:20 -03:00
oobabooga e471919e6d Make llava/minigpt-4 work with AutoGPTQ 2023-06-11 17:56:01 -03:00
oobabooga f4defde752 Add a menu for installing extensions 2023-06-11 17:11:06 -03:00
oobabooga 8e73806b20 Improve "Interface mode" appearance 2023-06-11 15:29:45 -03:00
oobabooga a06c953692 Minor style change 2023-06-11 15:13:26 -03:00
oobabooga ac122832f7 Make dropdown menus more similar to automatic1111 2023-06-11 14:20:16 -03:00
Amine Djeghri 8275dbc68c
Update WSL-installation-guide.md (#2626) 2023-06-11 12:30:34 -03:00
oobabooga 6133675e0f
Add menus for saving presets/characters/instruction templates/prompts (#2621) 2023-06-11 12:19:18 -03:00
oobabooga ea0eabd266 Bump llama-cpp-python version 2023-06-10 21:59:29 -03:00
oobabooga ec2b5bae39
Merge pull request #2616 from oobabooga/dev
Merge dev branch
2023-06-10 21:55:59 -03:00
brandonj60 b04e18d10c
Add Mirostat v2 sampling to transformer models (#2571) 2023-06-09 21:26:31 -03:00
oobabooga aff3e04df4 Remove irrelevant docs
Compiling from source, in my tests, makes no difference in
the resulting tokens/s.
2023-06-09 21:15:37 -03:00
oobabooga d7db25dac9 Fix a permission 2023-06-09 01:44:17 -03:00
oobabooga d033c85cf9 Fix a permission 2023-06-09 01:43:22 -03:00
oobabooga 741afd74f6 Update requirements-minimal.txt 2023-06-09 00:48:41 -03:00
oobabooga c333e4c906 Add docs for performance optimizations 2023-06-09 00:47:48 -03:00
oobabooga aaf240a14c
Merge pull request #2587 from oobabooga/dev 2023-06-09 00:30:59 -03:00
oobabooga c6552785af Minor cleanup 2023-06-09 00:30:22 -03:00
oobabooga 92b45cb3f5 Merge branch 'main' into dev 2023-06-09 00:27:11 -03:00
oobabooga 8a7a8343be Detect TheBloke_WizardLM-30B-GPTQ 2023-06-09 00:26:34 -03:00
oobabooga 0f8140e99d Bump transformers/accelerate/peft/autogptq 2023-06-09 00:25:13 -03:00
FartyPants ac40c59ac3
Added Guanaco-QLoRA to Instruct character (#2574) 2023-06-08 12:24:32 -03:00
oobabooga db2cbe7b5a Detect WizardLM-30B-V1.0 instruction format 2023-06-08 11:43:40 -03:00
oobabooga e0b43102e6 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2023-06-08 11:35:23 -03:00
matatonic 7be6fe126b
extensions/api: models api for blocking_api (updated) (#2539) 2023-06-08 11:34:36 -03:00
oobabooga 240752617d Increase download timeout to 20s 2023-06-08 11:16:38 -03:00
zaypen 084b006cfe
Update LLaMA-model.md (#2460)
Better approach of converting LLaMA model
2023-06-07 15:34:50 -03:00
dnobs c05edfcdfc
fix: reverse-proxied URI should end with 'chat', not 'generate' (#2556) 2023-06-07 00:08:04 -03:00
oobabooga 878250d609 Merge branch 'main' into dev 2023-06-06 19:43:53 -03:00
oobabooga f55e85e28a Fix multimodal with model loaded through AutoGPTQ 2023-06-06 19:42:40 -03:00
oobabooga eb2601a8c3 Reorganize Parameters tab 2023-06-06 14:51:02 -03:00
oobabooga 3cc5ce3c42
Merge pull request #2551 from oobabooga/dev 2023-06-06 14:40:52 -03:00
oobabooga 6015616338 Style changes 2023-06-06 13:06:05 -03:00
oobabooga f040073ef1 Handle the case of older autogptq install 2023-06-06 13:05:05 -03:00
oobabooga 5d515eeb8c Bump llama-cpp-python wheel 2023-06-06 13:01:15 -03:00
oobabooga bc58dc40bd Fix a minor bug 2023-06-06 12:57:13 -03:00
oobabooga f06a1387f0 Reorganize Models tab 2023-06-06 07:58:07 -03:00
oobabooga d49d299b67 Change a message 2023-06-06 07:54:56 -03:00
oobabooga f9b8bed953 Remove folder 2023-06-06 07:49:12 -03:00
oobabooga 90fdb8edc6 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2023-06-06 07:46:51 -03:00
oobabooga 7ed1e35fbf Reorganize Parameters tab in chat mode 2023-06-06 07:46:25 -03:00
oobabooga 00b94847da Remove softprompt support 2023-06-06 07:42:23 -03:00
bobzilla 643c44e975
Add ngrok shared URL ingress support (#1944) 2023-06-06 07:34:20 -03:00
oobabooga ccb4c9f178 Add some padding to chat box 2023-06-06 07:21:16 -03:00
oobabooga 0aebc838a0 Don't save the history for 'None' character 2023-06-06 07:21:07 -03:00
oobabooga 9f215523e2 Remove some unused imports 2023-06-06 07:05:46 -03:00
oobabooga b9bc9665d9 Remove some extra space 2023-06-06 07:01:37 -03:00
oobabooga 177ab7912a Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2023-06-06 07:01:00 -03:00
oobabooga 0f0108ce34 Never load the history for default character 2023-06-06 07:00:11 -03:00
oobabooga ae25b21d61 Improve instruct style in dark mode 2023-06-06 07:00:00 -03:00
matatonic 4a17a5db67
[extensions/openai] various fixes (#2533) 2023-06-06 01:43:04 -03:00
dependabot[bot] 97f3fa843f
Bump llama-cpp-python from 0.1.56 to 0.1.57 (#2537) 2023-06-05 23:45:58 -03:00
oobabooga 11f38b5c2b Add AutoGPTQ LoRA support 2023-06-05 23:32:57 -03:00
oobabooga 3a5cfe96f0 Increase chat_prompt_size_max 2023-06-05 17:37:37 -03:00
oobabooga 4e9937aa99 Bump gradio 2023-06-05 17:29:21 -03:00
oobabooga 53496ffa80
Create stale.yml 2023-06-05 17:15:31 -03:00
pandego 0377e385e0
Update .gitignore (#2504)
add .idea to git ignore
2023-06-05 17:11:03 -03:00
oobabooga 60bfd0b722
Merge pull request #2535 from oobabooga/dev
Dev branch merge
2023-06-05 17:07:54 -03:00
oobabooga eda224c92d Update README 2023-06-05 17:04:09 -03:00
oobabooga bef94b9ebb Update README 2023-06-05 17:01:13 -03:00
oobabooga 99d701994a Update GPTQ-models-(4-bit-mode).md 2023-06-05 15:55:00 -03:00
oobabooga f276d88546 Use AutoGPTQ by default for GPTQ models 2023-06-05 15:41:48 -03:00
oobabooga 632571a009 Update README 2023-06-05 15:16:06 -03:00
oobabooga 6a75bda419 Assign some 4096 seq lengths 2023-06-05 12:07:52 -03:00
oobabooga 9b0e95abeb Fix "regenerate" when "Start reply with" is set 2023-06-05 11:56:03 -03:00
oobabooga e61316ce0b Detect airoboros and Nous-Hermes 2023-06-05 11:52:13 -03:00
oobabooga 19f78684e6 Add "Start reply with" feature to chat mode 2023-06-02 13:58:08 -03:00
GralchemOz f7b07c4705
Fix the missing Chinese character bug (#2497) 2023-06-02 13:45:41 -03:00
oobabooga 28198bc15c Change some headers 2023-06-02 11:28:43 -03:00
oobabooga 5177cdf634 Change AutoGPTQ info 2023-06-02 11:19:44 -03:00
oobabooga 8e98633efd Add a description for chat_prompt_size 2023-06-02 11:13:22 -03:00
oobabooga 5a8162a46d Reorganize models tab 2023-06-02 02:24:15 -03:00
oobabooga d183c7d29e Fix streaming japanese/chinese characters
Credits to matasonic for the idea
2023-06-02 02:09:52 -03:00
jllllll 5216117a63
Fix MacOS incompatibility in requirements.txt (#2485) 2023-06-02 01:46:16 -03:00
oobabooga 2f6631195a Add desc_act checkbox to the UI 2023-06-02 01:45:46 -03:00
LaaZa 9c066601f5
Extend AutoGPTQ support for any GPTQ model (#1668) 2023-06-02 01:33:55 -03:00
oobabooga b4ad060c1f Use cuda 11.7 instead of 11.8 2023-06-02 01:04:44 -03:00
oobabooga d0aca83b53 Add AutoGPTQ wheels to requirements.txt 2023-06-02 00:47:11 -03:00
oobabooga f344ccdddb Add a template for bluemoon 2023-06-01 14:42:12 -03:00
oobabooga 522b01d051 Grammar 2023-06-01 14:05:29 -03:00
oobabooga 5540335819 Better way to detect if a model has been downloaded 2023-06-01 14:01:19 -03:00
oobabooga aa83fc21d4
Update Low-VRAM-guide.md 2023-06-01 12:14:27 -03:00
oobabooga ee99a87330
Update README.md 2023-06-01 12:08:44 -03:00
oobabooga a83f9aa65b
Update shared.py 2023-06-01 12:08:39 -03:00
oobabooga 146505a16b
Update README.md 2023-06-01 12:04:58 -03:00
oobabooga 756e3afbcc
Update llama.cpp-models.md 2023-06-01 12:04:31 -03:00
oobabooga 3347395944
Update README.md 2023-06-01 12:01:20 -03:00
oobabooga 74bf2f05b1
Update llama.cpp-models.md 2023-06-01 11:58:33 -03:00
oobabooga 90dc8a91ae
Update llama.cpp-models.md 2023-06-01 11:57:57 -03:00
oobabooga aba56de41b
Update README.md 2023-06-01 11:46:28 -03:00
oobabooga c9ac45d4cf
Update Using-LoRAs.md 2023-06-01 11:34:04 -03:00
oobabooga 9aad6d07de
Update Using-LoRAs.md 2023-06-01 11:32:41 -03:00
oobabooga df18ae7d6c
Update README.md 2023-06-01 11:27:33 -03:00
oobabooga 248ef32358 Print a big message for CPU users 2023-06-01 01:40:24 -03:00
oobabooga 290a3374e4 Don't download a model during installation
And some other updates/minor improvements
2023-06-01 01:30:21 -03:00
oobabooga e52b43c934
Update GPTQ-models-(4-bit-mode).md 2023-06-01 01:17:13 -03:00
Morgan Schweers 1aed2b9e52
Make it possible to download protected HF models from the command line. (#2408) 2023-06-01 00:11:21 -03:00
oobabooga 419c34eca4
Update GPTQ-models-(4-bit-mode).md 2023-05-31 23:49:00 -03:00
oobabooga 486ddd62df Add tfs and top_a to the API examples 2023-05-31 23:44:38 -03:00
oobabooga b6c407f51d Don't stream at more than 24 fps
This is a performance optimization
2023-05-31 23:41:42 -03:00
oobabooga a160230893 Update GPTQ-models-(4-bit-mode).md 2023-05-31 23:38:15 -03:00
oobabooga 2cdf525d3b Bump llama-cpp-python version 2023-05-31 23:29:02 -03:00
oobabooga 2e53caa806
Create LICENSE 2023-05-31 16:28:36 -03:00
Sam dea1bf3d04
Parse g++ version instead of using string matching (#72) 2023-05-31 14:44:36 -03:00
gavin660 97bc7e3fb6
Adds functionality for user to set flags via environment variable (#59) 2023-05-31 14:43:22 -03:00
Sam 5405635305
Install pre-compiled wheels for Linux (#74) 2023-05-31 14:41:54 -03:00
jllllll be98e74337
Install older bitsandbytes on older gpus + fix llama-cpp-python issue (#75) 2023-05-31 14:41:03 -03:00
jllllll 412e7a6a96
Update README.md to include missing flags (#2449) 2023-05-31 11:07:56 -03:00
AlpinDale 6627f7feb9
Add notice about downgrading gcc and g++ (#2446) 2023-05-30 22:28:53 -03:00
Atinoda bfbd13ae89
Update docker repo link (#2340) 2023-05-30 22:14:49 -03:00
matatonic a6d3f010a5
extensions/openai: include all available models in Model.list (#2368)
Co-authored-by: Matthew Ashton <mashton-gitlab@zhero.org>
2023-05-30 22:13:37 -03:00
matatonic e5b756ecfe
Fixes #2331, IndexError: string index out of range (#2383) 2023-05-30 22:07:40 -03:00
Juan M Uys b984a44f47
fix error when downloading a model for the first time (#2404) 2023-05-30 22:07:12 -03:00
Yiximail 4715123f55
Add a /api/v1/stop-stream API that allows the user to interrupt the generation (#2392) 2023-05-30 22:03:40 -03:00
matatonic ebcadc0042
extensions/openai: cross_origin + chunked_response (updated fix) (#2423) 2023-05-30 21:54:24 -03:00
matatonic df50f077db
fixup missing tfs top_a params, defaults reorg (#2443) 2023-05-30 21:52:33 -03:00
Forkoz 9ab90d8b60
Fix warning for qlora (#2438) 2023-05-30 11:09:18 -03:00
oobabooga 0db4e191bd
Improve chat buttons on mobile devices 2023-05-30 00:30:15 -03:00
oobabooga 3209440b7c
Rearrange chat buttons 2023-05-30 00:17:31 -03:00
oobabooga 3578dd3611
Change a warning message 2023-05-29 22:40:54 -03:00
oobabooga 3a6e194bc7
Change a warning message 2023-05-29 22:39:23 -03:00
oobabooga e763ace593
Update GPTQ-models-(4-bit-mode).md 2023-05-29 22:35:49 -03:00
oobabooga 86ef695d37
Update GPTQ-models-(4-bit-mode).md 2023-05-29 22:20:55 -03:00
oobabooga 8e0a997c60
Add new parameters to API extension 2023-05-29 22:03:08 -03:00
Luis Lopez 9e7204bef4
Add tail-free and top-a sampling (#2357) 2023-05-29 21:40:01 -03:00
oobabooga b4662bf4af
Download gptq_model*.py using download-model.py 2023-05-29 16:12:54 -03:00
oobabooga 540a161a08
Update GPTQ-models-(4-bit-mode).md 2023-05-29 15:45:40 -03:00
oobabooga b8d2f6d876 Merge remote-tracking branch 'refs/remotes/origin/main' 2023-05-29 15:33:05 -03:00
oobabooga 1394f44e14 Add triton checkbox for AutoGPTQ 2023-05-29 15:32:45 -03:00
oobabooga 166a0d9893
Update GPTQ-models-(4-bit-mode).md 2023-05-29 15:07:59 -03:00
oobabooga 962d05ca7e
Update README.md 2023-05-29 14:56:55 -03:00
oobabooga 4a190a98fd
Update GPTQ-models-(4-bit-mode).md 2023-05-29 14:56:05 -03:00
matatonic 2b7ba9586f
Fixes #2326, KeyError: 'assistant' (#2382) 2023-05-29 14:19:57 -03:00
oobabooga 6de727c524 Improve Eta Sampling preset 2023-05-29 13:56:15 -03:00
oobabooga f34d20922c Minor fix 2023-05-29 13:31:17 -03:00
oobabooga 983eef1e29 Attempt at evaluating falcon perplexity (failed) 2023-05-29 13:28:25 -03:00
Honkware 204731952a
Falcon support (trust-remote-code and autogptq checkboxes) (#2367)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-05-29 10:20:18 -03:00
Forkoz 60ae80cf28
Fix hang in tokenizer for AutoGPTQ llama models. (#2399) 2023-05-28 23:10:10 -03:00
oobabooga 2f811b1bdf Change a warning message 2023-05-28 22:48:20 -03:00
oobabooga 9ee1e37121 Fix return message when no model is loaded 2023-05-28 22:46:32 -03:00
oobabooga f27135bdd3 Add Eta Sampling preset
Also remove some presets that I do not consider relevant
2023-05-28 22:44:35 -03:00
oobabooga 00ebea0b2a Use YAML for presets and settings 2023-05-28 22:34:12 -03:00
Elias Vincent Simon 2cf711f35e
update SpeechRecognition dependency (#2345) 2023-05-26 00:34:57 -03:00
jllllll 78dbec4c4e
Add 'scipy' to requirements.txt #2335 (#2343)
Unlisted dependency of bitsandbytes
2023-05-25 23:26:25 -03:00
Luis Lopez 0dbc3d9b2c
Fix get_documents_ids_distances return error when n_results = 0 (#2347) 2023-05-25 23:25:36 -03:00
jllllll 07a4f0569f
Update README.md to account for BnB Windows wheel (#2341) 2023-05-25 18:44:26 -03:00
oobabooga acfd876f29 Some qol changes to "Perplexity evaluation" 2023-05-25 15:06:22 -03:00
oobabooga 8efdc01ffb Better default for compute_dtype 2023-05-25 15:05:53 -03:00
oobabooga fc33216477 Small fix for n_ctx in llama.cpp 2023-05-25 13:55:51 -03:00
oobabooga 35009c32f0 Beautify all CSS 2023-05-25 13:12:34 -03:00
oobabooga 231305d0f5
Update README.md 2023-05-25 12:05:08 -03:00
oobabooga 37d4ad012b Add a button for rendering markdown for any model 2023-05-25 11:59:27 -03:00
oobabooga 9a43656a50
Add bitsandbytes note 2023-05-25 11:21:52 -03:00
jllllll b1b3bb6923
Improve environment isolation (#68) 2023-05-25 11:15:05 -03:00
oobabooga c8ce2e777b
Add instructions for CPU mode users 2023-05-25 10:57:52 -03:00
oobabooga 996c49daa7
Remove bitsandbytes installation step
Following 548f05e106
2023-05-25 10:50:20 -03:00
oobabooga 548f05e106 Add windows bitsandbytes wheel by jllllll 2023-05-25 10:48:22 -03:00
DGdev91 cf088566f8
Make llama.cpp read prompt size and seed from settings (#2299) 2023-05-25 10:29:31 -03:00
Luis Lopez ee674afa50
Add superbooga time weighted history retrieval (#2080) 2023-05-25 10:22:45 -03:00
oobabooga a04266161d
Update README.md 2023-05-25 01:23:46 -03:00
oobabooga 361451ba60
Add --load-in-4bit parameter (#2320) 2023-05-25 01:14:13 -03:00
oobabooga 63ce5f9c28 Add back a missing bos token 2023-05-24 13:54:36 -03:00
Alex "mcmonkey" Goodwin 3cd7c5bdd0
LoRA Trainer: train_only_after option to control which part of your input to train on (#2315) 2023-05-24 12:43:22 -03:00
eiery 9967e08b1f
update llama-cpp-python to v0.1.53 for ggml v3, fixes #2245 (#2264) 2023-05-24 10:25:28 -03:00
Gabriel Terrien e50ade438a
FIX silero_tts/elevenlabs_tts activation/deactivation (#2313) 2023-05-24 10:06:38 -03:00
Gabriel Terrien fc116711b0
FIX save_model_settings function to also update shared.model_config (#2282) 2023-05-24 10:01:07 -03:00
flurb18 d37a28730d
Beginning of multi-user support (#2262)
Adds a lock to generate_reply
2023-05-24 09:38:20 -03:00
Anthony K 7dc87984a2
Fix spelling mistake in new name var of chat api (#2309) 2023-05-23 23:03:03 -03:00
oobabooga 1490c0af68 Remove RWKV from requirements.txt 2023-05-23 20:49:20 -03:00
Gabriel Terrien 7aed53559a
Support of the --gradio-auth flag (#2283) 2023-05-23 20:39:26 -03:00
Atinoda 4155aaa96a
Add mention to alternative docker repository (#2145) 2023-05-23 20:35:53 -03:00
matatonic 9714072692
[extensions/openai] use instruction templates with chat_completions (#2291) 2023-05-23 19:58:41 -03:00
oobabooga 74aae34beb Allow passing your name to the chat API 2023-05-23 19:39:18 -03:00
oobabooga fb6a00f4e5 Small AutoGPTQ fix 2023-05-23 15:20:01 -03:00
oobabooga c2d2ef7c13
Update Generation-parameters.md 2023-05-23 02:11:28 -03:00
oobabooga b0845ae4e8
Update RWKV-model.md 2023-05-23 02:10:08 -03:00
oobabooga cd3618d7fb Add support for RWKV in Hugging Face format 2023-05-23 02:07:28 -03:00
oobabooga 75adc110d4 Fix "perplexity evaluation" progress messages 2023-05-23 01:54:52 -03:00
oobabooga 4d94a111d4 memoize load_character to speed up the chat API 2023-05-23 00:50:58 -03:00
oobabooga 8b9ba3d7b4 Fix a typo 2023-05-22 20:13:03 -03:00
Gabriel Terrien 0f51b64bb3
Add a "dark_theme" option to settings.json (#2288) 2023-05-22 19:45:11 -03:00
oobabooga c5446ae0e2 Fix a link 2023-05-22 19:38:34 -03:00
oobabooga c0fd7f3257
Add mirostat parameters for llama.cpp (#2287) 2023-05-22 19:37:24 -03:00
oobabooga ec7437f00a
Better way to toggle light/dark mode 2023-05-22 03:19:01 -03:00
oobabooga d46f5a58a3 Add a button for toggling dark/light mode 2023-05-22 03:11:44 -03:00
dependabot[bot] baf75356d4
Bump transformers from 4.29.1 to 4.29.2 (#2268) 2023-05-22 02:50:18 -03:00
oobabooga 4372eb228c Increase the interface area by 10px 2023-05-22 00:55:33 -03:00
oobabooga 753f6c5250 Attempt at making interface restart more robust 2023-05-22 00:26:07 -03:00
oobabooga 30225b9dd0 Fix --no-stream queue bug 2023-05-22 00:02:59 -03:00
oobabooga 288912baf1 Add a description for the extensions checkbox group 2023-05-21 23:33:37 -03:00
oobabooga 6e77844733 Add a description for penalty_alpha 2023-05-21 23:09:30 -03:00
oobabooga d63ef59a0f Apply LLaMA-Precise preset to Vicuna by default 2023-05-21 23:00:42 -03:00
oobabooga e3d578502a Improve "Chat settings" tab appearance a bit 2023-05-21 22:58:14 -03:00
oobabooga dcc3e54005 Various "impersonate" fixes 2023-05-21 22:54:28 -03:00
oobabooga e116d31180 Prevent unwanted log messages from modules 2023-05-21 22:42:34 -03:00
oobabooga fb91406e93 Fix generation_attempts continuing after an empty reply 2023-05-21 22:14:50 -03:00
oobabooga e18534fe12 Fix "continue" in chat-instruct mode 2023-05-21 22:05:59 -03:00
oobabooga d7fabe693d Reorganize parameters tab 2023-05-21 16:24:47 -03:00
oobabooga 8ac3636966
Add epsilon_cutoff/eta_cutoff parameters (#2258) 2023-05-21 15:11:57 -03:00
oobabooga 767a767989 Fix elevenlabs_tts too 2023-05-21 14:11:46 -03:00
oobabooga 1e5821bd9e Fix silero tts autoplay (attempt #2) 2023-05-21 13:25:11 -03:00
oobabooga a5d5bb9390 Fix silero tts autoplay 2023-05-21 12:11:59 -03:00
matatonic 78b2478d9c
assistant: space fix, system: prompt fix (#2219) 2023-05-20 23:32:34 -03:00
oobabooga 05593a7834 Minor bug fix 2023-05-20 23:22:36 -03:00
Luis Lopez 9c53517d2c
Fix superbooga error when querying empty DB (Issue #2160) (#2212) 2023-05-20 22:27:22 -03:00
Matthew McAllister ab6acddcc5
Add Save/Delete character buttons (#1870)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-05-20 21:48:45 -03:00
oobabooga c5af549d4b
Add chat API (#2233) 2023-05-20 18:42:17 -03:00
jllllll 2aa01e2303
Fix broken version of peft (#2229) 2023-05-20 17:54:51 -03:00
oobabooga 159eccac7e
Update Audio-Notification.md 2023-05-19 23:20:42 -03:00
HappyWorldGames a3e9769e31
Added an audible notification after text generation in web. (#1277)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-05-19 23:16:06 -03:00
Konstantin Gukov 1b52bddfcc
Mitigate UnboundLocalError (#2136) 2023-05-19 14:46:18 -03:00
Alex "mcmonkey" Goodwin 50c70e28f0
Lora Trainer improvements, part 6 - slightly better raw text inputs (#2108) 2023-05-19 12:58:54 -03:00
oobabooga 511470a89b Bump llama-cpp-python version 2023-05-19 12:13:25 -03:00
Carl Kenner a9733d4a99
Metharme context fix (#2153) 2023-05-19 11:46:13 -03:00
Carl Kenner c86231377b
Wizard Mega, Ziya, KoAlpaca, OpenBuddy, Chinese-Vicuna, Vigogne, Bactrian, H2O support, fix Baize (#2159) 2023-05-19 11:42:41 -03:00
Mykeehu c98d6ad27f
Create chat_style-messenger.css (#2187)
Add Messenger-like style for chat mode
2023-05-19 11:31:06 -03:00
oobabooga 499c2e009e Remove problematic regex from models/config.yaml 2023-05-19 11:20:35 -03:00
oobabooga 9d5025f531 Improve error handling while loading GPTQ models 2023-05-19 11:20:08 -03:00
oobabooga 39dab18307 Add a timeout to download-model.py requests 2023-05-19 11:19:34 -03:00
jllllll 4ef2de3486
Fix dependencies downgrading from gptq install (#61) 2023-05-18 12:46:04 -03:00
oobabooga 07510a2414
Change a message 2023-05-18 10:58:37 -03:00
oobabooga 0bcd5b6894
Soothe anxious users 2023-05-18 10:56:49 -03:00
oobabooga f052ab9c8f Fix setting pre_layer from within the ui 2023-05-17 23:17:44 -03:00
oobabooga b667ffa51d Simplify GPTQ_loader.py 2023-05-17 16:22:56 -03:00
oobabooga ef10ffc6b4 Add various checks to model loading functions 2023-05-17 16:14:54 -03:00
oobabooga abd361b3a0 Minor change 2023-05-17 11:33:43 -03:00
oobabooga 21ecc3701e Avoid a name conflict 2023-05-17 11:23:13 -03:00
oobabooga fb91c07191 Minor bug fix 2023-05-17 11:16:37 -03:00
oobabooga 1a8151a2b6
Add AutoGPTQ support (basic) (#2132) 2023-05-17 11:12:12 -03:00
oobabooga 10cf7831f7
Update Extensions.md 2023-05-17 10:45:29 -03:00
Alex "mcmonkey" Goodwin 1f50dbe352
Experimental jank multiGPU inference that's 2x faster than native somehow (#2100) 2023-05-17 10:41:09 -03:00
oobabooga fd743a0207 Small change 2023-05-17 02:34:29 -03:00
LoopLooter aeb1b7a9c5
feature to save prompts with custom names (#1583)
---------

Co-authored-by: LoopLooter <looplooter>
Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-05-17 02:30:45 -03:00
oobabooga c9c6aa2b6e Update docs/Extensions.md 2023-05-17 02:04:37 -03:00
oobabooga 85f74961f9 Update "Interface mode" tab 2023-05-17 01:57:51 -03:00
oobabooga 9e558cba9b Update docs/Extensions.md 2023-05-17 01:43:32 -03:00
oobabooga 687f21f965 Update docs/Extensions.md 2023-05-17 01:41:01 -03:00
oobabooga 8f85d84e08 Merge remote-tracking branch 'refs/remotes/origin/main' 2023-05-17 01:32:42 -03:00
oobabooga ce21804ec7 Allow extensions to define a new tab 2023-05-17 01:31:56 -03:00
ye7iaserag acf3dbbcc5
Allow extensions to have custom display_name (#1242)
Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-05-17 01:08:22 -03:00
oobabooga ad0b71af11 Add missing file 2023-05-17 00:37:34 -03:00
oobabooga a84f499718 Allow extensions to define custom CSS and JS 2023-05-17 00:30:54 -03:00
oobabooga 824fa8fc0e Attempt at making interface restart more robust 2023-05-16 22:27:43 -03:00
oobabooga 259020a0be Bump gradio to 3.31.0
This fixes Google Colab lagging.
2023-05-16 22:21:15 -03:00
pixel 458a627ab9
fix: elevenlabs cloned voices do not show up in webui after entering API key (#2107) 2023-05-16 20:21:36 -03:00
oobabooga 7584d46c29
Refactor models.py (#2113) 2023-05-16 19:52:22 -03:00
oobabooga 5cd6dd4287 Fix no-mmap bug 2023-05-16 17:35:49 -03:00
oobabooga 89e37626ab Reorganize chat settings tab 2023-05-16 17:22:59 -03:00
Forkoz d205ec9706
Fix Training fails when evaluation dataset is selected (#2099)
Fixes https://github.com/oobabooga/text-generation-webui/issues/2078 from Googulator
2023-05-16 13:40:19 -03:00
Orbitoid 428261eede
fix: elevenlabs removed the need for the api key for refreshing voices (#2097) 2023-05-16 13:34:49 -03:00
oobabooga cd9be4c2ba
Update llama.cpp-models.md 2023-05-16 00:49:32 -03:00
atriantafy 26cf8c2545
add api port options (#1990) 2023-05-15 20:44:16 -03:00
Andrei e657dd342d
Add in-memory cache support for llama.cpp (#1936) 2023-05-15 20:19:55 -03:00
Jakub Strnad 0227e738ed
Add settings UI for llama.cpp and fixed reloading of llama.cpp models (#2087) 2023-05-15 19:51:23 -03:00
oobabooga 10869de0f4 Merge remote-tracking branch 'refs/remotes/origin/main' 2023-05-15 19:39:48 -03:00
oobabooga c07215cc08 Improve the default Assistant character 2023-05-15 19:39:08 -03:00
oobabooga 4e66f68115 Create get_max_memory_dict() function 2023-05-15 19:38:27 -03:00
dependabot[bot] ae54d83455
Bump transformers from 4.28.1 to 4.29.1 (#2089) 2023-05-15 19:25:24 -03:00
AlphaAtlas 071f0776ad
Add llama.cpp GPU offload option (#2060) 2023-05-14 22:58:11 -03:00
feeelX eee986348c
Update llama-cpp-python from 0.1.45 to 0.1.50 (#2058) 2023-05-14 22:41:14 -03:00
oobabooga 897fa60069 Sort selected superbooga chunks by insertion order
For better coherence
2023-05-14 22:19:29 -03:00
Luis Lopez b07f849e41
Add superbooga chunk separator option (#2051) 2023-05-14 21:44:52 -03:00
matatonic ab08cf6465
[extensions/openai] clip extra leading space (#2042) 2023-05-14 12:57:52 -03:00
oobabooga 3b886f9c9f
Add chat-instruct mode (#2049) 2023-05-14 10:43:55 -03:00
oobabooga 5f6cf39f36 Change the injection context string 2023-05-13 14:23:02 -03:00
oobabooga 7cc17e3f1f Refactor superbooga 2023-05-13 14:15:40 -03:00
oobabooga 826c74c201 Expand superbooga to instruct mode and change the chat implementation 2023-05-13 12:50:19 -03:00
oobabooga c746a5bd00 Add .rstrip(' ') to openai api 2023-05-12 14:40:48 -03:00
Damian Stewart 3f1bfba718
Clarify how to start server.py with multimodal API support (#2025) 2023-05-12 14:37:49 -03:00
oobabooga 437d1c7ead Fix bug in save_model_settings 2023-05-12 14:33:00 -03:00
oobabooga 146a9cb393 Allow superbooga to download URLs in parallel 2023-05-12 14:19:55 -03:00
oobabooga df37ba5256 Update impersonate_wrapper 2023-05-12 12:59:48 -03:00
oobabooga e283ddc559 Change how spaces are handled in continue/generation attempts 2023-05-12 12:50:29 -03:00
oobabooga 2eeb27659d Fix bug in --cpu-memory 2023-05-12 06:17:07 -03:00
oobabooga fcb46282c5 Add a rule to config.yaml 2023-05-12 06:11:58 -03:00
oobabooga 5eaa914e1b Fix settings.json being ignored because of config.yaml 2023-05-12 06:09:45 -03:00
oobabooga a77965e801 Make the regex for "Save settings for this model" exact 2023-05-12 00:43:13 -03:00
matatonic f98fd01dcd
is_chat=False for /edits (#2011) 2023-05-11 19:15:11 -03:00
oobabooga 71693161eb Better handle spaces in LlamaTokenizer 2023-05-11 17:55:50 -03:00
oobabooga 7221d1389a Fix a bug 2023-05-11 17:11:10 -03:00
oobabooga 0d36c18f5d Always return only the new tokens in generation functions 2023-05-11 17:07:20 -03:00
matatonic c4f0e6d740
is_chat changes fix for openai extension (#2008) 2023-05-11 16:32:25 -03:00
oobabooga 394bb253db Syntax improvement 2023-05-11 16:27:50 -03:00
oobabooga f7dbddfff5 Add a variable for tts extensions to use 2023-05-11 16:12:46 -03:00
oobabooga 638c6a65a2
Refactor chat functions (#2003) 2023-05-11 15:37:04 -03:00
real 4e9da22c58
missing stream api port added to docker compose (#2005) 2023-05-11 15:07:56 -03:00
matatonic 309b72e549
[extension/openai] add edits & image endpoints & fix prompt return in non --chat modes (#1935) 2023-05-11 11:06:39 -03:00
oobabooga 23d3f6909a
Update README.md 2023-05-11 10:21:20 -03:00
oobabooga 400f3648f4
Update docs/README.md 2023-05-11 10:10:24 -03:00
oobabooga 2930e5a895
Update README.md 2023-05-11 10:04:38 -03:00
oobabooga 0ff38c994e
Update README.md 2023-05-11 09:58:58 -03:00
oobabooga e6959a5d9a
Update README.md 2023-05-11 09:54:22 -03:00
oobabooga dcfd09b61e
Update README.md 2023-05-11 09:49:57 -03:00
oobabooga ac9a86a16c
Update llama.cpp-models.md 2023-05-11 09:47:36 -03:00
oobabooga 7a49ceab29
Update README.md 2023-05-11 09:42:39 -03:00
oobabooga 943b5e5f80 Minor bug fix 2023-05-10 23:54:25 -03:00
oobabooga 9695bfe117 Add an option for cleaning up html in superbooga 2023-05-10 23:51:52 -03:00
oobabooga c7ba2d4f3f Change a message in download-model.py 2023-05-10 19:00:14 -03:00
oobabooga 1309cdd257
Add a space 2023-05-10 18:03:12 -03:00
oobabooga 3e19733d35
Remove obsolete comment 2023-05-10 18:01:04 -03:00
oobabooga 4ab5deeea0
Update INSTRUCTIONS.TXT 2023-05-10 18:00:37 -03:00
oobabooga d7d3f7f31c
Add a "CMD_FLAGS" variable 2023-05-10 17:54:12 -03:00
oobabooga b7a589afc8 Improve the Metharme prompt 2023-05-10 16:09:32 -03:00
oobabooga e5b1547849 Fix reload model button 2023-05-10 14:44:25 -03:00
oobabooga b01c4884cb Better stopping strings for instruct mode 2023-05-10 14:22:38 -03:00
oobabooga 6a4783afc7 Add markdown table rendering 2023-05-10 13:41:23 -03:00
oobabooga 57dc44a995
Update README.md 2023-05-10 12:48:25 -03:00
oobabooga f5592781e5
Update README.md 2023-05-10 12:19:56 -03:00
oobabooga f1d10edcb7
Update README.md 2023-05-10 12:13:14 -03:00
oobabooga 181b102521
Update README.md 2023-05-10 12:09:47 -03:00
oobabooga 3316e33d14 Remove unused code 2023-05-10 11:59:59 -03:00
Alexander Dibrov ec14d9b725
Fix custom_generate_chat_prompt (#1965) 2023-05-10 11:29:59 -03:00
oobabooga 32481ec4d6 Fix prompt order in the dropdown 2023-05-10 02:24:09 -03:00
oobabooga dfd9ba3e90 Remove duplicate code 2023-05-10 02:07:22 -03:00
oobabooga cd36b8f739 Remove space 2023-05-10 01:41:33 -03:00
oobabooga c35860ff2f Add a link to silero samples 2023-05-10 01:39:35 -03:00
oobabooga bdf1274b5d Remove duplicate code 2023-05-10 01:34:04 -03:00
oobabooga ba445cf59f Fix some galactica templates 2023-05-09 22:58:59 -03:00
oobabooga 3b1de7e8bc Remove redundant presets 2023-05-09 22:56:19 -03:00
oobabooga 3913155c1f
Style improvements (#1957) 2023-05-09 22:49:39 -03:00
minipasila 334486f527
Added instruct-following template for Metharme (#1679) 2023-05-09 22:29:22 -03:00
Carl Kenner 1aaa47070a
Expand Open Assistant support (#1735) 2023-05-09 20:40:29 -03:00
Carl Kenner 814f754451
Support for MPT, INCITE, WizardLM, StableLM, Galactica, Vicuna, Guanaco, and Baize instruction following (#1596) 2023-05-09 20:37:31 -03:00
Matthew McAllister 06c7db017d
Add config for pygmalion-7b and metharme-7b (#1887) 2023-05-09 20:31:27 -03:00
missionfloyd fe4dfc647d
SileroTTS preview (#1934) 2023-05-09 20:28:59 -03:00
oobabooga 8fa5f651d6 Style changes 2023-05-09 20:20:35 -03:00
Wojtab e9e75a9ec7
Generalize multimodality (llava/minigpt4 7b and 13b now supported) (#1741) 2023-05-09 20:18:02 -03:00
Wesley Pyburn a2b25322f0
Fix trust_remote_code in wrong location (#1953) 2023-05-09 19:22:10 -03:00
oobabooga 13e7ebfc77 Change a comment 2023-05-09 15:56:32 -03:00
LaaZa 218bd64bd1
Add the option to not automatically load the selected model (#1762)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-05-09 15:52:35 -03:00
oobabooga b8cfc20e58
Don't install superbooga by default 2023-05-09 14:17:08 -03:00
Maks cf6caf1830
Make the RWKV model cache the RNN state between messages (#1354) 2023-05-09 11:12:53 -03:00
Kamil Szurant 641500dcb9
Use current input for Impersonate (continue impersonate feature) (#1147) 2023-05-09 02:37:42 -03:00
dependabot[bot] a5bb278631
Bump accelerate from 0.18.0 to 0.19.0 (#1925) 2023-05-09 02:17:27 -03:00
jllllll 29727c6502
Fix Windows PATH fix (#57) 2023-05-09 01:49:27 -03:00
IJumpAround 020fe7b50b
Remove mutable defaults from function signature. (#1663) 2023-05-08 22:55:41 -03:00
shadownetdev1 32ad47c898
added note about build essentials to WSL docs (#1859) 2023-05-08 22:32:41 -03:00
Jeffrey Lin 791a38bad1
[extensions/openai] Support undocumented base64 'encoding_format' param for compatibility with official OpenAI client (#1876) 2023-05-08 22:31:34 -03:00
Matthew McAllister d78b04f0b4
Add error message when GPTQ-for-LLaMa import fails (#1871)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-05-08 22:29:09 -03:00
Kévin Legrand 99d2dd8d0d
Fix - Ignore OS files from source control (#1911) 2023-05-08 22:17:37 -03:00
kaiokendev 0e27b660e8
Fix issue where n_results/k > index size (#1929) 2023-05-08 21:16:00 -03:00
oobabooga 68dcbc7ebd Fix chat history handling in instruct mode 2023-05-08 16:41:21 -03:00
Clay Shoaf 79ac94cc2f
fixed LoRA loading issue (#1865) 2023-05-08 16:21:55 -03:00
Arseni Lapunov 8818967d37
Fix typo in docs/Training-LoRAs.md (#1921) 2023-05-08 15:12:39 -03:00
oobabooga b5260b24f1
Add support for custom chat styles (#1917) 2023-05-08 12:35:03 -03:00
oobabooga b040b4110d Bump llama-cpp-python version 2023-05-08 00:21:17 -03:00
oobabooga 63898c09ac Document superbooga 2023-05-08 00:11:31 -03:00
oobabooga 82eb284a3f Fix a bug in silero tts 2023-05-07 23:59:43 -03:00
Matthew McAllister 0c048252b5
Fix character menu when default chat mode is 'instruct' (#1873) 2023-05-07 23:50:38 -03:00
EgrorBs d3ea70f453
More trust_remote_code=trust_remote_code (#1899) 2023-05-07 23:48:20 -03:00
oobabooga 56a5969658
Improve the separation between instruct/chat modes (#1896) 2023-05-07 23:47:02 -03:00
oobabooga 9754d6a811 Fix an error message 2023-05-07 17:44:05 -03:00
camenduru ba65a48ec8
trust_remote_code=shared.args.trust_remote_code (#1891) 2023-05-07 17:42:44 -03:00
oobabooga b3bbda22d1 Exception handling 2023-05-07 16:50:45 -03:00
oobabooga 8515f19078 Fix typing error 2023-05-07 16:25:39 -03:00
oobabooga 6b67cb6611 Generalize superbooga to chat mode 2023-05-07 15:05:26 -03:00
oobabooga ec1cda0e1f Rename superbig to superbooga 2023-05-07 13:29:49 -03:00
oobabooga 8216c8f732 Handle multiline inputs 2023-05-07 11:54:26 -03:00
oobabooga a35a2fab02 Reorganize superbig ui 2023-05-07 11:30:16 -03:00
oobabooga befa307c42 Add support for loading multiple URLs into superbig 2023-05-07 11:08:02 -03:00
oobabooga 04eca9b65b Make chunk length/count customizable 2023-05-07 05:02:04 -03:00
oobabooga 8c06eeaf84 Add requirement 2023-05-07 04:50:12 -03:00
oobabooga 58f9a30be1 Add support for input file/url in superbig 2023-05-07 04:49:02 -03:00
oobabooga 6afba065a1 Only update the chroma database when necessary 2023-05-07 04:03:17 -03:00
kaiokendev 5a4bd3918c
Add SuperBIG extension (alpha) (#1548)
---------

Co-authored-by: kaiokendev <>
Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-05-07 03:50:12 -03:00
oobabooga 81be7c2dd4 Specify gradio_client version 2023-05-06 21:50:04 -03:00
oobabooga 85238de421 Remove unused variable 2023-05-06 11:03:12 -03:00
oobabooga de9c4e260e
Minor fixes to elevenlabs_tts 2023-05-06 10:57:34 -03:00
Steve Randall b03a2ac512
Elevenlabs Extension Improvement and migration to official API (#1830) 2023-05-06 10:56:31 -03:00
Semjon Kravtšenko 126d216384
Fix possible crash (#53) 2023-05-06 01:14:09 -03:00
oobabooga 56f6b7052a Sort dropdowns numerically 2023-05-05 23:14:56 -03:00
oobabooga ee3c8a893e
Update Extensions.md 2023-05-05 19:04:50 -03:00
oobabooga 8aafb1f796
Refactor text_generation.py, add support for custom generation functions (#1817) 2023-05-05 18:53:03 -03:00
Tom Jobbins 876fbb97c0
Allow downloading model from HF branch via UI (#1662)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-05-05 13:59:01 -03:00
oobabooga 849ad04c96 Change background color of instruct code blocks 2023-05-05 12:02:45 -03:00
oobabooga c728f2b5f0 Better handle new line characters in code blocks 2023-05-05 11:22:36 -03:00
oobabooga 207a031e8d CSS change to instruct mode 2023-05-05 00:36:15 -03:00
oobabooga e5d6d822b1 Minor CSS change to instruct mode 2023-05-04 23:41:00 -03:00
oobabooga a50c2ab82a Add missing file 2023-05-04 23:29:46 -03:00
oobabooga 00e333d790 Add MOSS support 2023-05-04 23:20:34 -03:00
oobabooga f673f4a4ca Change --verbose behavior 2023-05-04 15:56:06 -03:00
oobabooga 97a6a50d98 Use oasst tokenizer instead of universal tokenizer 2023-05-04 15:55:39 -03:00
oobabooga b6ff138084 Add --checkpoint argument for GPTQ 2023-05-04 15:17:20 -03:00
oobabooga dbddedca3f Detect oasst-sft-6-llama-30b 2023-05-04 15:13:37 -03:00
Wojtek Kowaluk 1436c5845a
fix ggml detection regex in model downloader (#1779) 2023-05-04 11:48:36 -03:00
Mylo bd531c2dc2
Make --trust-remote-code work for all models (#1772) 2023-05-04 02:01:28 -03:00
oobabooga 0e6d17304a Clearer syntax for instruction-following characters 2023-05-03 22:50:39 -03:00
oobabooga 9c77ab4fc2 Improve some warnings 2023-05-03 22:06:46 -03:00
oobabooga 057b1b2978 Add credits 2023-05-03 21:49:55 -03:00
oobabooga 95d04d6a8d Better warning messages 2023-05-03 21:43:17 -03:00
oobabooga 0a48b29cd8 Prevent websocket disconnection on the client side 2023-05-03 20:44:30 -03:00
oobabooga 4bf7253ec5 Fix typing bug in api 2023-05-03 19:27:20 -03:00
oobabooga d6410a1b36 Bump recommended monkey patch commit 2023-05-03 14:49:25 -03:00
oobabooga 60be76f0fc Revert gradio bump (gallery is broken) 2023-05-03 11:53:30 -03:00
Thireus ☠ 4883e20fa7
Fix openai extension script.py - TypeError: '_Environ' object is not callable (#1753) 2023-05-03 09:51:49 -03:00
oobabooga f54256e348 Rename no_mmap to no-mmap 2023-05-03 09:50:31 -03:00
Roberts Slisans dec31af910
Create .gitignore (#43) 2023-05-02 23:47:19 -03:00
Semih Aslan 24c5ba2b9c
Fixed error when $OS_ARCH returns aarch64 (#45)
For some machines $OS_ARCH returns aarch64 instead of ARM64,and as i see here this should fix it.
2023-05-02 23:47:03 -03:00
oobabooga 875da16b7b Minor CSS improvements in chat mode 2023-05-02 23:38:51 -03:00
practicaldreamer e3968f7dd0
Fix Training Pad Token (#1678)
Currently padding with 0 the character vs 0 the token id (<unk> in the case of llama)
2023-05-02 23:16:08 -03:00
Wojtab 80c2f25131
LLaVA: small fixes (#1664)
* change multimodal projector to the correct one

* remove reference to custom stopping strings from readme

* fix stopping strings if tokenizer extension adds/removes tokens

* add API example

* LLaVA 7B just dropped, add to readme that there is no support for it currently
2023-05-02 23:12:22 -03:00
oobabooga c31b0f15a7 Remove some spaces 2023-05-02 23:07:07 -03:00
oobabooga 320fcfde4e Style/pep8 improvements 2023-05-02 23:05:38 -03:00
oobabooga ecd79caa68
Update Extensions.md 2023-05-02 22:52:32 -03:00
matatonic 7ac41b87df
add openai compatible api (#1475) 2023-05-02 22:49:53 -03:00
oobabooga 4e09df4034 Only show extension in UI if it has an ui() function 2023-05-02 19:20:02 -03:00
oobabooga d016c38640 Bump gradio version 2023-05-02 19:19:33 -03:00
oobabooga 88cdf6ed3d Prevent websocket from disconnecting 2023-05-02 19:03:19 -03:00
Ahmed Said fbcd32988e
added no_mmap & mlock parameters to llama.cpp and removed llamacpp_model_alternative (#1649)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-05-02 18:25:28 -03:00
Blake Wyatt 4babb22f84
Fix/Improve a bunch of things (#42) 2023-05-02 12:28:20 -03:00
Carl Kenner 2f1a2846d1
Verbose should always print special tokens in input (#1707) 2023-05-02 01:24:56 -03:00
Alex "mcmonkey" Goodwin 0df0b2d0f9
optimize stopping strings processing (#1625) 2023-05-02 01:21:54 -03:00
oobabooga e6a78c00f2
Update Docker.md 2023-05-02 00:51:10 -03:00
Tom Jobbins 3c67fc0362
Allow groupsize 1024, needed for larger models eg 30B to lower VRAM usage (#1660) 2023-05-02 00:46:26 -03:00
Lawrence M Stewart 78bd4d3a5c
Update LLaMA-model.md (#1700)
protobuf needs to be 3.20.x or lower
2023-05-02 00:44:09 -03:00
Dhaladom f659415170
fixed variable name "context" to "prompt" (#1716) 2023-05-02 00:43:40 -03:00
dependabot[bot] 280c2f285f
Bump safetensors from 0.3.0 to 0.3.1 (#1720) 2023-05-02 00:42:39 -03:00
oobabooga 56b13d5d48 Bump llama-cpp-python version 2023-05-02 00:41:54 -03:00
Lőrinc Pap ee68ec9079
Update folder produced by download-model (#1601) 2023-04-27 12:03:02 -03:00
oobabooga 91745f63c3 Use Vicuna-v0 by default for Vicuna models 2023-04-26 17:45:38 -03:00
oobabooga 93e5c066ae Update RWKV Raven template 2023-04-26 17:31:03 -03:00
oobabooga c83210c460 Move the rstrips 2023-04-26 17:17:22 -03:00
oobabooga 1d8b8222e9 Revert #1579, apply the proper fix
Apparently models dislike trailing spaces.
2023-04-26 16:47:50 -03:00
TiagoGF a941c19337
Fixing Vicuna text generation (#1579) 2023-04-26 16:20:27 -03:00
oobabooga d87ca8f2af LLaVA fixes 2023-04-26 03:47:34 -03:00
oobabooga 9c2e7c0fab Fix path on models.py 2023-04-26 03:29:09 -03:00
oobabooga a777c058af
Precise prompts for instruct mode 2023-04-26 03:21:53 -03:00
oobabooga a8409426d7
Fix bug in models.py 2023-04-26 01:55:40 -03:00
oobabooga 4c491aa142 Add Alpaca prompt with Input field 2023-04-25 23:50:32 -03:00
oobabooga 68ed73dd89 Make API extension print its exceptions 2023-04-25 23:23:47 -03:00
oobabooga f642135517 Make universal tokenizer, xformers, sdp-attention apply to monkey patch 2023-04-25 23:18:11 -03:00
oobabooga f39c99fa14 Load more than one LoRA with --lora, fix a bug 2023-04-25 22:58:48 -03:00
oobabooga 15940e762e Fix missing initial space for LlamaTokenizer 2023-04-25 22:47:23 -03:00
Vincent Brouwers 92cdb4f22b
Seq2Seq support (including FLAN-T5) (#1535)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-04-25 22:39:04 -03:00
USBhost 95aa43b9c2
Update LLaMA download docs 2023-04-25 21:28:15 -03:00
Alex "mcmonkey" Goodwin 312cb7dda6
LoRA trainer improvements part 5 (#1546)
* full dynamic model type support on modern peft

* remove shuffle option
2023-04-25 21:27:30 -03:00
Wojtab 65beb51b0b
fix returned dtypes for LLaVA (#1547) 2023-04-25 21:25:34 -03:00
oobabooga 9b272bc8e5 Monkey patch fixes 2023-04-25 21:20:26 -03:00
oobabooga da812600f4 Apply settings regardless of setup() function 2023-04-25 01:16:23 -03:00
da3dsoul ebca3f86d5
Apply the settings for extensions after import, but before setup() (#1484) 2023-04-25 00:23:11 -03:00
oobabooga b0ce750d4e Add spaces 2023-04-25 00:10:21 -03:00
oobabooga 1a0c12c6f2
Refactor text-generation.py a bit 2023-04-24 19:24:12 -03:00
oobabooga bcd5786a47
Add files via upload 2023-04-24 16:53:04 -03:00
oobabooga d66059d95a
Update INSTRUCTIONS.TXT 2023-04-24 16:50:03 -03:00
oobabooga a4f6724b88
Add a comment 2023-04-24 16:47:22 -03:00
oobabooga 9a8487097b
Remove --auto-devices 2023-04-24 16:43:52 -03:00
oobabooga 2f4f124132 Remove obsolete function 2023-04-24 13:27:24 -03:00
oobabooga b6af2e56a2 Add --character flag, add character to settings.json 2023-04-24 13:19:42 -03:00
oobabooga 0c32ae27cc Only load the default history if it's empty 2023-04-24 11:50:51 -03:00
MajdajkD c86e9a3372
fix websocket batching (#1511) 2023-04-24 03:51:32 -03:00
eiery 78d1977ebf
add n_batch support for llama.cpp (#1115) 2023-04-24 03:46:18 -03:00
oobabooga 2f6e2ddeac Bump llama-cpp-python version 2023-04-24 03:42:03 -03:00
oobabooga caaa556159 Move extensions block definition to the bottom 2023-04-24 03:30:35 -03:00
oobabooga b1ee674d75 Make interface state (mostly) persistent on page reload 2023-04-24 03:05:47 -03:00
oobabooga 47809e28aa Minor changes 2023-04-24 01:04:48 -03:00
oobabooga 435f8cc0e7
Simplify some chat functions 2023-04-24 00:47:40 -03:00
Wojtab 04b98a8485
Fix Continue for LLaVA (#1507) 2023-04-23 22:58:15 -03:00
Wojtab 12212cf6be
LLaVA support (#1487) 2023-04-23 20:32:22 -03:00
oobabooga 9197d3fec8
Update Extensions.md 2023-04-23 16:11:17 -03:00
Andy Salerno 654933c634
New universal API with streaming/blocking endpoints (#990)
Previous title: Add api_streaming extension and update api-example-stream to use it

* Merge with latest main

* Add parameter capturing encoder_repetition_penalty

* Change some defaults, minor fixes

* Add --api, --public-api flags

* remove unneeded/broken comment from blocking API startup. The comment is already correctly emitted in try_start_cloudflared by calling the lambda we pass in.

* Update on_start message for blocking_api, it should say 'non-streaming' and not 'streaming'

* Update the API examples

* Change a comment

* Update README

* Remove the gradio API

* Remove unused import

* Minor change

* Remove unused import

---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-04-23 15:52:43 -03:00
oobabooga dfbb18610f
Update INSTRUCTIONS.TXT 2023-04-23 12:58:14 -03:00
Alex "mcmonkey" Goodwin 459e725af9
Lora trainer docs (#1493) 2023-04-23 12:54:41 -03:00
oobabooga 7ff645899e Fix bug in api extension 2023-04-22 17:33:36 -03:00
AICatgirls b992c9236a
Prevent API extension responses from getting cut off with --chat enabled (#1467) 2023-04-22 16:06:43 -03:00
oobabooga c0b5c09860 Minor change 2023-04-22 15:15:31 -03:00
oobabooga 47666c4d00
Update GPTQ-models-(4-bit-mode).md 2023-04-22 15:12:14 -03:00
oobabooga fcb594b90e Don't require llama.cpp models to be placed in subfolders 2023-04-22 14:56:48 -03:00
oobabooga 06b6ff6c2e
Update GPTQ-models-(4-bit-mode).md 2023-04-22 12:49:00 -03:00
oobabooga 2c6d43e60f
Update GPTQ-models-(4-bit-mode).md 2023-04-22 12:48:20 -03:00
oobabooga 7438f4f6ba Change GPTQ triton default settings 2023-04-22 12:27:30 -03:00
InconsolableCellist e03b873460
Updating Using-LoRAs.md doc to clarify resuming training (#1474) 2023-04-22 03:35:36 -03:00
oobabooga fe02281477
Update README.md 2023-04-22 03:05:00 -03:00
oobabooga ef40b4e862
Update README.md 2023-04-22 03:03:39 -03:00
oobabooga 408e172ad9
Rename docker/README.md to docs/Docker.md 2023-04-22 03:03:05 -03:00
oobabooga 4d9ae44efd
Update Spell-book.md 2023-04-22 02:53:52 -03:00
oobabooga 9508f207ba
Update Using-LoRAs.md 2023-04-22 02:53:01 -03:00
oobabooga 6d4f131d0a
Update Low-VRAM-guide.md 2023-04-22 02:50:35 -03:00
oobabooga f5c36cca40
Update LLaMA-model.md 2023-04-22 02:49:54 -03:00
oobabooga 038fa3eb39
Update README.md 2023-04-22 02:46:07 -03:00
oobabooga b5e5b9aeae
Delete Home.md 2023-04-22 02:40:20 -03:00
oobabooga fe6e9ea986
Update README.md 2023-04-22 02:40:08 -03:00
oobabooga 80ef7c7bcb
Add files via upload 2023-04-22 02:34:13 -03:00
oobabooga 25b433990a
Create README.md 2023-04-22 02:33:32 -03:00
oobabooga 505c2c73e8
Update README.md 2023-04-22 00:11:27 -03:00
Φφ 143e88694d
SD_api_pictures: Modefix, +hires options, UI layout change (#1400) 2023-04-21 17:49:18 -03:00
oobabooga 2dca8bb25e Sort imports 2023-04-21 17:20:59 -03:00
oobabooga c238ba9532 Add a 'Count tokens' button 2023-04-21 17:18:34 -03:00
Lou Bernardi a6ef2429fa
Add "do not download" and "download from HF" to download-model.py (#1439) 2023-04-21 12:54:50 -03:00
USBhost e1aa9d5173
Support upstream GPTQ once again. (#1451) 2023-04-21 12:43:56 -03:00
oobabooga eddd016449 Minor deletion 2023-04-21 12:41:27 -03:00
oobabooga d46b9b7c50 Fix evaluate comment saving 2023-04-21 12:34:08 -03:00
oobabooga 5e023ae64d Change dropdown menu highlight color 2023-04-21 02:47:18 -03:00
oobabooga 2d766d2e19 Improve notebook mode button sizes 2023-04-21 02:37:58 -03:00
oobabooga c4f4f41389
Add an "Evaluate" tab to calculate the perplexities of models (#1322) 2023-04-21 00:20:33 -03:00
oobabooga ff0d0ac552 Api extension bug fix 2023-04-20 13:26:58 -03:00
oobabooga 96ba55501c Rename custom_stopping_strings in the api extension 2023-04-20 00:15:32 -03:00
oobabooga 7bb9036ac9 Add universal LLaMA tokenizer support 2023-04-19 21:23:51 -03:00
oobabooga 32d47e4bad Update api-example-stream.py 2023-04-19 19:49:52 -03:00
Alex "mcmonkey" Goodwin ee30625cd1
4-Bit LoRA training + several new training options and fixes 2023-04-19 19:39:03 -03:00
oobabooga 702fe92d42 Increase truncation_length_max value 2023-04-19 17:35:38 -03:00
oobabooga 9d9ae62938 Fix stopping strings in the gradio API 2023-04-19 13:52:21 -03:00
Rudd-O 69d50e2e86
Fix download script (#1373) 2023-04-19 13:02:32 -03:00
oobabooga 49aa05054a Minor CSS fixes 2023-04-19 01:48:07 -03:00
oobabooga b4af319fa2 Add a workaround for GALACTICA on some systems 2023-04-19 01:43:10 -03:00
oobabooga 0ba4095504 Adjust chat mode height 2023-04-19 01:35:20 -03:00
oobabooga ab5624136a Adjust some heights 2023-04-19 01:23:09 -03:00
oobabooga 61126f4674 Change the button styles 2023-04-19 00:56:24 -03:00
oobabooga 85d3d3ccf3 Add a margin 2023-04-19 00:47:02 -03:00
oobabooga 649e4017a5 Style improvements 2023-04-19 00:36:28 -03:00
oobabooga f8da9a0424
Update README.md 2023-04-18 20:25:08 -03:00
oobabooga c3f6e65554
Update README.md 2023-04-18 20:23:31 -03:00
oobabooga eb15193327
Update README.md 2023-04-18 13:07:08 -03:00
oobabooga 7fbfc489e2
Update README.md 2023-04-18 12:56:37 -03:00
oobabooga f559f9595b
Update README.md 2023-04-18 12:54:09 -03:00
oobabooga 172bc949dd
Update README.md 2023-04-18 12:50:33 -03:00
oobabooga 753cd2d303
Rename Dockerfile to docker/Dockerfile 2023-04-18 12:48:04 -03:00
loeken 89e22d4d6a
added windows/docker docs (#1027) 2023-04-18 12:47:43 -03:00
oobabooga b0c762ceba
Revert a change
I think that this may be needed for some clients
2023-04-18 04:10:45 -03:00
oobabooga 000f65a2ef
Delete unused file 2023-04-18 04:01:14 -03:00
oobabooga c58c1d89bd
Clean method to prevent gradio from phoning home 2023-04-18 03:56:20 -03:00
oobabooga 8275989f03
Add new 1-click installers for Linux and MacOS 2023-04-18 02:40:36 -03:00
oobabooga 1ba0082410
Add files via upload 2023-04-18 02:30:47 -03:00
oobabooga a5f7d98cf3
Rename environment_windows.bat to cmd_windows.bat 2023-04-18 02:30:23 -03:00
oobabooga 316aaff348
Rename environment_macos.sh to cmd_macos.sh 2023-04-18 02:30:08 -03:00
oobabooga 647f7bca36
Rename environment_linux.sh to cmd_linux.sh 2023-04-18 02:29:55 -03:00
Blake Wyatt 6d2c72b593
Add support for MacOS, Linux, and WSL (#21)
* Initial commit

* Initial commit with new code

* Add comments

* Move GPTQ out of if

* Fix install on Arch Linux

* Fix case where install was aborted

If the install was aborted before a model was downloaded, webui wouldn't run.

* Update start_windows.bat

Add necessary flags to Miniconda installer
Disable Start Menu shortcut creation
Disable ssl on Conda
Change Python version to latest 3.10,
I've noticed that explicitly specifying 3.10.9 can break the included Python installation

* Update bitsandbytes wheel link to 0.38.1

Disable ssl on Conda

* Add check for spaces in path

Installation of Miniconda will fail in this case

* Mirror changes to mac and linux scripts

* Start with model-menu

* Add updaters

* Fix line endings

* Add check for path with spaces

* Fix one-click updating

* Fix one-click updating

* Clean up update scripts

* Add environment scripts

---------

Co-authored-by: jllllll <3887729+jllllll@users.noreply.github.com>
Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-04-18 02:23:09 -03:00
oobabooga e1b80e6fe6
Comment the gradio patch 2023-04-18 01:57:59 -03:00
oobabooga 36f7c022f2
Rename a file 2023-04-18 01:38:33 -03:00
oobabooga b069bb1f2e
Update monkey_patch_gradio.py 2023-04-18 01:32:42 -03:00
oobabooga 00186f76f4
Monkey patch gradio to prevent it from calling home 2023-04-18 01:13:16 -03:00
Tynan Burke 6a810b16b2
typo in training.py (#1329) 2023-04-17 21:40:46 -03:00
oobabooga ac2973ffc6 Add a warning for --share 2023-04-17 19:34:28 -03:00
oobabooga c544386824 Reset your name when choosing a character 2023-04-17 13:56:40 -03:00
oobabooga 163ea295e7 Fix but in API extension 2023-04-17 13:54:15 -03:00
oobabooga b1b9519539 Merge branch 'main' of github.com:oobabooga/text-generation-webui 2023-04-17 13:52:49 -03:00
oobabooga c3dc348d1c Don't show 'None' in the LoRA list 2023-04-17 13:52:23 -03:00
oobabooga 301c687c64
Update README.md 2023-04-17 11:25:26 -03:00
oobabooga 19e3a59997 Remove unused extension 2023-04-17 11:06:08 -03:00
oobabooga 89bc540557 Update README 2023-04-17 10:55:35 -03:00
catalpaaa 07de7d0426
Load llamacpp before quantized model (#1307) 2023-04-17 10:47:26 -03:00
practicaldreamer 3961f49524
Add note about --no-fused_mlp ignoring --gpu-memory (#1301) 2023-04-17 10:46:37 -03:00
sgsdxzy b57ffc2ec9
Update to support GPTQ triton commit c90adef (#1229) 2023-04-17 01:11:18 -03:00
oobabooga 209fcd21d5 Reorganize Parameters tab 2023-04-17 00:33:22 -03:00
oobabooga 3e5cdd005f
Update README.md 2023-04-16 23:28:59 -03:00
oobabooga 39099663a0
Add 4-bit LoRA support (#1200) 2023-04-16 23:26:52 -03:00
oobabooga ec3e869c27 Merge branch 'main' of github.com:oobabooga/text-generation-webui 2023-04-16 21:26:42 -03:00
oobabooga 46a8aa8c09 Readability 2023-04-16 21:26:19 -03:00
GuizzyQC 5011f94659
Improved compatibility between silero and sd_api_pictures (#1196) 2023-04-16 21:18:52 -03:00
svupper 61d6f7f507
Add dependencies to Dockerfile for TTS extensions (#1276) 2023-04-16 21:17:00 -03:00
dependabot[bot] 4cd2a9d824
Bump transformers from 4.28.0 to 4.28.1 (#1288) 2023-04-16 21:12:57 -03:00
oobabooga 705121161b
Update README.md 2023-04-16 20:03:03 -03:00
oobabooga 50c55a51fc
Update README.md 2023-04-16 19:22:31 -03:00
Forkoz c6fe1ced01
Add ChatGLM support (#1256)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-04-16 19:15:03 -03:00
oobabooga 6a03ad0824 Remove fix_newlines() calls from chat.py 2023-04-16 18:25:44 -03:00
oobabooga 5342f72968 Properly handle blockquote blocks 2023-04-16 18:00:12 -03:00
oobabooga 27f3a78834 Better detect when no model is loaded 2023-04-16 17:35:54 -03:00
oobabooga c8ad960018 Add defaults to the gradio API 2023-04-16 17:33:28 -03:00
oobabooga c96529a1b3
Update README.md 2023-04-16 17:00:03 -03:00
oobabooga 6675f51ffe Change a color 2023-04-16 16:48:20 -03:00
oobabooga beb95f5fe2 Add a style for the "chat" mode 2023-04-16 16:44:50 -03:00
oobabooga cb95a2432c Add Koala support 2023-04-16 14:41:06 -03:00
oobabooga b937c9d8c2
Add skip_special_tokens checkbox for Dolly model (#1218) 2023-04-16 14:24:49 -03:00
oobabooga a9c7ef4159 Exclude yaml files from model list 2023-04-16 12:47:30 -03:00
oobabooga 4e035cc3fb Fix api-example-stream 2023-04-16 12:12:31 -03:00
oobabooga b705b4210c Minor changes to training.py 2023-04-16 03:08:37 -03:00
oobabooga 5c513a5f5c Make training.py more readable 2023-04-16 02:46:27 -03:00
Alex "mcmonkey" Goodwin a3eec62b50
Lora trainer improvements part 3 (#1098)
* add support for other model types

dependent on future-peft-changes but with fallback to function now

* use encoding=utf8 for training format

* make shuffling optional

and describe dropout a bit more

* add eval_steps to control evaluation

* make callbacks not depend on globals

* make save steps controllable

* placeholder of initial loading-existing-model support

and var name cleanup

* save/load parameters

* last bit of cleanup

* remove `gptq_bits` ref as main branch removed that setting

* add higher_rank_limit option

2048 is basically unreachable due to VRAM, but i trained at 1536 with batch size = 1 on a 7B model.
Note that it's in the do_train input just to save as a parameter

* fix math on save_steps
2023-04-16 02:35:13 -03:00
kernyan ac19d5101f
revert incorrect eos_token_id change from #814 (#1261)
- fixes #1054
2023-04-16 01:47:01 -03:00
oobabooga a2127239de Fix a bug 2023-04-16 01:41:37 -03:00
oobabooga 9d3c6d2dc3 Fix a bug 2023-04-16 01:40:47 -03:00
Mikel Bober-Irizar 16a3a5b039
Merge pull request from GHSA-hv5m-3rp9-xcpf
* Remove eval of API input

* Remove unnecessary eval/exec for security

* Use ast.literal_eval

* Use ast.literal_eval

---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-04-16 01:36:50 -03:00
oobabooga d2ea925fa5 Bump llama-cpp-python to use LlamaCache 2023-04-16 00:53:40 -03:00
oobabooga ac189011cb Add "Save current settings for this model" button 2023-04-15 12:54:02 -03:00
oobabooga b9dcba7762 Don't overwrite --gpu_memory on boot (#1237/#1235) 2023-04-15 11:59:31 -03:00
oobabooga 72a89e8fd1 Merge branch 'main' of github.com:oobabooga/text-generation-webui 2023-04-15 01:21:46 -03:00
oobabooga abef355ed0 Remove deprecated flag 2023-04-15 01:21:19 -03:00
Lou Bernardi 4d13cc0a1f
Suppress whisper_stt pip warning (#1215) 2023-04-15 01:15:59 -03:00
catalpaaa 94700cc7a5
Bump gradio to 3.25 (#1089) 2023-04-14 23:45:25 -03:00
oobabooga 004f275efe
Update README.md 2023-04-14 23:36:56 -03:00
oobabooga 83964ed354
Update README.md 2023-04-14 23:33:54 -03:00
oobabooga c41037db68
Update README.md 2023-04-14 23:32:39 -03:00
oobabooga f8913c7132 Merge branch 'main' of github.com:oobabooga/text-generation-webui 2023-04-14 23:17:29 -03:00
oobabooga 628f8e6168 Reorganize chat buttons 2023-04-14 23:17:15 -03:00
DavG25 5126c97139
Improve CSS in default and notebook mode (#1209) 2023-04-14 23:03:43 -03:00
oobabooga c3aa79118e Minor generate_chat_prompt simplification 2023-04-14 23:02:08 -03:00
oobabooga c4aa1a42b1 Fix chat history downloading 2023-04-14 19:38:30 -03:00
oobabooga 3a337cfded Use argparse defaults 2023-04-14 15:35:06 -03:00
oobabooga 43e01282b3 Don't override user initial wbits/groupsize 2023-04-14 15:24:03 -03:00
Alex "mcmonkey" Goodwin 64e3b44e0f
initial multi-lora support (#1103)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-04-14 14:52:06 -03:00
oobabooga ebb81eb176 Add Undo/Regenerate buttons to notebook mode 2023-04-14 14:34:56 -03:00
oobabooga 1901d238e1 Minor change to API code 2023-04-14 12:11:47 -03:00
oobabooga 581f739b2f Reorganize 2023-04-14 11:38:26 -03:00
oobabooga 7d7d122edb Cover one more model 2023-04-14 11:15:59 -03:00
oobabooga 8eba88061a Remove unused config 2023-04-14 11:12:17 -03:00
oobabooga 8e31f2bad4
Automatically set wbits/groupsize/instruct based on model name (#1167) 2023-04-14 11:07:28 -03:00
v0xie 9d66957207
Add --listen-host launch option (#1122) 2023-04-13 21:35:08 -03:00
oobabooga 403be8a27f
Update README.md 2023-04-13 21:23:35 -03:00
dependabot[bot] 852a5aa13d
Bump bitsandbytes from 0.37.2 to 0.38.1 (#1158) 2023-04-13 21:23:14 -03:00
dependabot[bot] 84576a80d2
Bump llama-cpp-python from 0.1.30 to 0.1.33 (#1157) 2023-04-13 21:17:59 -03:00
oobabooga 2908a51587 Settle for transformers 4.28.0 2023-04-13 21:07:00 -03:00
oobabooga 28a11f5724 Revert the Default preset to a previous version 2023-04-13 13:02:02 -03:00
oobabooga a75e02de4d Simplify GPTQ_loader.py 2023-04-13 12:13:07 -03:00
oobabooga c13e8651ad Suppress "TypedStorage is deprecated" warnings 2023-04-13 12:09:42 -03:00
oobabooga 17ce7c8671 Suppress annoying bitsandbytes welcome message 2023-04-13 12:04:39 -03:00
oobabooga ca293bb713 Show a warning if two quantized models are found 2023-04-13 12:04:27 -03:00
oobabooga fbb448ce4f If only 1 model is available, load that model 2023-04-13 11:44:10 -03:00
oobabooga 8b482b4127
Merge #1073 from sgsdxzy/triton
* Multi-GPU support for triton
* Better quantized model filename detection
2023-04-13 11:31:21 -03:00
oobabooga fde6d06167 Prioritize names with the groupsize in them 2023-04-13 11:27:03 -03:00
oobabooga f2bf1a2c9e Add some comments, remove obsolete code 2023-04-13 11:17:32 -03:00
oobabooga 5744b31593 Reorganize some buttons 2023-04-13 11:05:47 -03:00
Light da74cd7c44 Generalized weight search path. 2023-04-13 21:43:32 +08:00
oobabooga 7a0ef1c941 Reorganize whisper extension 2023-04-13 10:40:27 -03:00
oobabooga 02cf8f5898 Merge branch 'main' of github.com:oobabooga/text-generation-webui 2023-04-13 10:35:43 -03:00
oobabooga 04866dc4fc Add a warning for when no model is loaded 2023-04-13 10:35:08 -03:00
Xan 6e19ae4b2f
Fix gpt-j model type in UI (#1129) 2023-04-13 10:17:20 -03:00
Light 97e67d136b Update README.md 2023-04-13 21:00:58 +08:00
Light cf58058c33 Change warmup_autotune to a negative switch. 2023-04-13 20:59:49 +08:00
Light 15d5a043f2 Merge remote-tracking branch 'origin/main' into triton 2023-04-13 19:38:51 +08:00
oobabooga ddbd237ec9 Better way to sort the models/loras 2023-04-12 22:56:32 -03:00
oobabooga 7dfbe54f42 Add --model-menu option 2023-04-12 21:24:26 -03:00
oobabooga 86c10c6f0c Add some labels 2023-04-12 18:39:21 -03:00
oobabooga 388038fb8e Update settings-template.json 2023-04-12 18:30:43 -03:00
oobabooga 0baa50bcc4 Update a comment 2023-04-12 18:26:15 -03:00
oobabooga 5d1d0bd11f Add the GPU index to the label 2023-04-12 18:24:19 -03:00
oobabooga 13789fd200 Handle the no-GPU / multi-GPU cases 2023-04-12 18:21:14 -03:00
oobabooga 10e939c9b4 Merge branch 'main' of github.com:oobabooga/text-generation-webui 2023-04-12 17:21:59 -03:00
oobabooga 1566d8e344 Add model settings to the Models tab 2023-04-12 17:20:18 -03:00
MarlinMr 47daf891fe
Link to developer.nvidia.com (#1104) 2023-04-12 15:56:42 -03:00
Light a405064ceb Better dispatch. 2023-04-13 01:48:17 +08:00
Light f3591ccfa1 Keep minimal change. 2023-04-12 23:26:06 +08:00
Φφ 1d8526849b
SD_api_pictures: Fix for metadata (#1082) 2023-04-12 11:42:30 -03:00
Lukas 5ad92c940e
lora training fixes: (#970)
Fix wrong input format being picked
Fix crash when an entry in the dataset has an attribute of value None
2023-04-12 11:38:01 -03:00
oobabooga 4f7e88c043 Simplify send_pictures 2023-04-12 10:30:12 -03:00
oobabooga 80f4eabb2a Fix send_pictures extension 2023-04-12 10:27:06 -03:00
oobabooga 49ce866c99 Fix silero_tts 2023-04-12 00:58:11 -03:00
oobabooga ff610b47d2 Make api-example-stream.py functional again 2023-04-12 00:25:30 -03:00
Andy Salerno 3850f13624
Change fn_index in api_example_stream (#904) 2023-04-12 00:15:12 -03:00
oobabooga 461ca7faf5
Mention that pull request reviews are welcome 2023-04-11 23:12:48 -03:00
Tymec 832ee4323d
API: add endpoint for counting tokens (#1051) 2023-04-11 23:08:42 -03:00
oobabooga 1405cd8af2 Merge branch 'main' of github.com:oobabooga/text-generation-webui 2023-04-11 22:44:05 -03:00
oobabooga 2289d3686f Update API example 2023-04-11 22:43:43 -03:00
Alexander01998 61641a4551
Add missing new parameters to API extension 2023-04-11 22:41:13 -03:00
oobabooga f2be87235d Comment lines that were causing undefined behavior 2023-04-11 22:40:04 -03:00
oobabooga 8265d45db8 Add send dummy message/reply buttons
Useful for starting a new reply.
2023-04-11 22:21:41 -03:00
oobabooga 37d52c96bc Fix Continue in chat mode 2023-04-11 21:46:17 -03:00
oobabooga f2ec880e81 Auto-scroll to the bottom when streaming is over in notebook/default modes 2023-04-11 20:58:10 -03:00
oobabooga f34f2daa3d More reasonable default preset 2023-04-11 18:57:46 -03:00
oobabooga cacbcda208
Two new options: truncation length and ban eos token 2023-04-11 18:46:06 -03:00
oobabooga 749c08a4ff
Update README.md 2023-04-11 14:42:10 -03:00
DavG25 e9e93189ff
Fix text overflow in chat and instruct mode (#1044) 2023-04-11 14:41:29 -03:00
oobabooga dc3c9d00a0 Update the API extension 2023-04-11 13:07:45 -03:00
oobabooga 457d3c58eb Update the API example 2023-04-11 12:57:36 -03:00
catalpaaa 78bbc66fc4
allow custom stopping strings in all modes (#903) 2023-04-11 12:30:06 -03:00
oobabooga 0f212093a3
Refactor the UI
A single dictionary called 'interface_state' is now passed as input to all functions. The values are updated only when necessary.

The goal is to make it easier to add new elements to the UI.
2023-04-11 11:46:30 -03:00
oobabooga 64f5c90ee7 Fix the API extension 2023-04-10 20:14:38 -03:00
oobabooga 58b34c0841 Fix chat_prompt_size 2023-04-10 20:06:42 -03:00
oobabooga 5234071c04 Improve Instruct mode text readability 2023-04-10 17:41:07 -03:00
IggoOnCode 09d8119e3c
Add CPU LoRA training (#938)
(It's very slow)
2023-04-10 17:29:00 -03:00
Alex "mcmonkey" Goodwin 0caf718a21
add on-page documentation to parameters (#1008) 2023-04-10 17:19:12 -03:00
oobabooga 85a7954823 Update settings-template.json 2023-04-10 16:53:07 -03:00
oobabooga d37b4f76b1 Merge branch 'main' of github.com:oobabooga/text-generation-webui 2023-04-10 16:45:09 -03:00
oobabooga bd04ff27ad Make the bos token optional 2023-04-10 16:44:22 -03:00
oobabooga f035b01823
Update README.md 2023-04-10 16:20:23 -03:00
Jeff Lefebvre b7ca89ba3f
Mention that build-essential is required (#1013) 2023-04-10 16:19:10 -03:00
loeken 52339e9b20
add make/g++ to docker (#1015) 2023-04-10 16:18:07 -03:00
oobabooga 4961f43702 Improve header bar colors 2023-04-10 16:15:16 -03:00
oobabooga 617530296e Instruct mode color/style improvements 2023-04-10 16:04:21 -03:00
oobabooga 0f1627eff1 Don't treat Intruct mode histories as regular histories
* They must now be saved/loaded manually
* Also improved browser caching of pfps
* Also changed the global default preset
2023-04-10 15:48:07 -03:00
oobabooga d679c4be13 Change a label 2023-04-10 11:44:37 -03:00
oobabooga 45244ed125 More descriptive download info 2023-04-10 11:42:12 -03:00
oobabooga 7e70741a4e
Download models from Model tab (#954 from UsamaKenway/main) 2023-04-10 11:38:30 -03:00
oobabooga 11b23db8d4 Remove unused imports 2023-04-10 11:37:42 -03:00
oobabooga 2c14df81a8 Use download-model.py to download the model 2023-04-10 11:36:39 -03:00
oobabooga c6e9ba20a4 Merge branch 'main' into UsamaKenway-main 2023-04-10 11:14:03 -03:00
oobabooga 843f672227
fix random seeds to actually randomize (#1004 from mcmonkey4eva/seed-fix) 2023-04-10 10:56:12 -03:00
oobabooga 769aa900ea Print the used seed 2023-04-10 10:53:31 -03:00
jllllll 254609daca
Update llama-cpp-python link to official wheel (#19) 2023-04-10 10:48:56 -03:00
oobabooga 32d078487e Add llama-cpp-python to requirements.txt 2023-04-10 10:45:51 -03:00
Alex "mcmonkey" Goodwin 30befe492a fix random seeds to actually randomize
Without this fix, manual seeds get locked in.
2023-04-10 06:29:10 -07:00
jllllll c3e1a58cb3
Correct llama-cpp-python wheel link (#17) 2023-04-09 23:46:54 -03:00
oobabooga 1911504f82 Minor bug fix 2023-04-09 23:45:41 -03:00
BlueprintCoding 8178fde2cb
Added dropdown to character bias. (#986) 2023-04-09 23:44:31 -03:00
oobabooga dba2000d2b Do things that I am not proud of 2023-04-09 23:40:49 -03:00
oobabooga 97840c92f9
Add working llamaa-cpp-python install from wheel. (#13 from Loufe/oobabooga-windows) 2023-04-09 23:23:27 -03:00
oobabooga 65552d2157 Merge branch 'main' of github.com:oobabooga/text-generation-webui 2023-04-09 23:19:53 -03:00
oobabooga 8c6155251a More robust 4-bit model loading 2023-04-09 23:19:28 -03:00
MarkovInequality 992663fa20
Added xformers support to Llama (#950) 2023-04-09 23:08:40 -03:00
Brian O'Connor 625d81f495
Update character log logic (#977)
* When logs are cleared, save the cleared log over the old log files
* Generate a log file when a character is loaded the first time
2023-04-09 22:20:21 -03:00
oobabooga 57f768eaad Better preset in api-example.py 2023-04-09 22:18:40 -03:00
oobabooga a3085dba07 Fix LlamaTokenizer eos_token (attempt) 2023-04-09 21:19:39 -03:00
oobabooga 120f5662cf Better handle spaces for Continue 2023-04-09 20:37:31 -03:00
oobabooga b27d757fd1 Minor change 2023-04-09 20:06:20 -03:00
oobabooga d29f4624e9 Add a Continue button to chat mode 2023-04-09 20:04:16 -03:00
oobabooga 170e0c05c4 Typo 2023-04-09 17:00:59 -03:00
oobabooga 34ec02d41d Make download-model.py importable 2023-04-09 16:59:59 -03:00
oobabooga f91d3a3ff4 server.py readability 2023-04-09 14:46:32 -03:00
Usama Kenway ebdf4c8c12 path fixed 2023-04-09 16:53:21 +05:00
Usama Kenway 7436dd5b4a download custom model menu (from hugging face) added in model tab 2023-04-09 16:11:43 +05:00
oobabooga bce1b7fbb2
Update README.md 2023-04-09 02:19:40 -03:00
oobabooga f7860ce192
Update README.md 2023-04-09 02:19:17 -03:00
oobabooga ece8ed2c84
Update README.md 2023-04-09 02:18:42 -03:00
oobabooga cc693a7546 Remove obsolete code 2023-04-09 00:51:07 -03:00
Lou Bernardi 0818bc93ad Add working llamaa-cpp-python install from wheel. 2023-04-08 22:44:55 -04:00
oobabooga 2fde50a800
Delete docker.md 2023-04-08 22:37:54 -03:00
loeken acc235aced
updated docs for docker, setup video added, removed left over GPTQ_VERSION from docker-compose (#940) 2023-04-08 22:35:15 -03:00
Blake Wyatt df561fd896
Fix ggml downloading in download-model.py (#915) 2023-04-08 18:52:30 -03:00
oobabooga d272ac46dd Add Pillow as a requirement 2023-04-08 18:48:46 -03:00
oobabooga cb169d0834 Minor formatting changes 2023-04-08 17:34:07 -03:00
oobabooga 2f16d0afca Remove redundant events 2023-04-08 17:32:36 -03:00
oobabooga a6a00cb82f
Properly concatenate chat events 2023-04-08 17:25:21 -03:00
Φφ c97c270040
Send_pictures small fix (#546) 2023-04-08 01:55:16 -03:00
oobabooga 0b458bf82d Simplify a function 2023-04-07 21:37:41 -03:00
Φφ ffd102e5c0
SD Api Pics extension, v.1.1 (#596) 2023-04-07 21:36:04 -03:00
oobabooga 5543a5089d Auto-submit the whisper extension transcription 2023-04-07 15:57:51 -03:00
oobabooga 1dc464dcb0 Sort imports 2023-04-07 14:42:03 -03:00
oobabooga 962e33dc10 Change button style 2023-04-07 12:22:14 -03:00
oobabooga 42ea6a3fc0 Change the timing for setup() calls 2023-04-07 12:20:57 -03:00
Φφ e563b015d8
Silero TTS offline cache (#628) 2023-04-07 12:15:57 -03:00
oobabooga 1c413ed593 Remove torch from silero 2023-04-07 11:51:50 -03:00
da3dsoul 3f922d4bfb
Extract the Preprocessing for Silero into a file and Improve it (#757) 2023-04-07 11:46:29 -03:00
Maya 744bf7cbf2
Get rid of type parameter warning (#883)
Fix annoying `The 'type' parameter has been deprecated. Use the Number component instead` warning
2023-04-07 11:17:16 -03:00
oobabooga 768354239b Change training file encoding 2023-04-07 11:15:52 -03:00
oobabooga 6762e62a40 Simplifications 2023-04-07 11:14:32 -03:00
oobabooga a453d4e9c4 Reorganize some chat functions 2023-04-07 11:07:03 -03:00
MarlinMr ec979cd9c4
Use updated docker compose (#877) 2023-04-07 10:48:47 -03:00
MarlinMr 2c0018d946
Cosmetic change of README.md (#878) 2023-04-07 10:47:10 -03:00
Maya 8fa182cfa7
Fix regeneration of first message in instruct mode (#881) 2023-04-07 10:45:42 -03:00
Alastair D'Silva 862aad637b
Tweak COPY order in Dockerfile (#863) 2023-04-07 00:56:44 -03:00
oobabooga 46c4654226 More PEP8 stuff 2023-04-07 00:52:02 -03:00
oobabooga ea6e77df72
Make the code more like PEP8 for readability (#862) 2023-04-07 00:15:45 -03:00
oobabooga 848c4edfd5
Update README.md 2023-04-06 22:52:35 -03:00
oobabooga e047cd1def Update README 2023-04-06 22:50:58 -03:00
loeken 08b9d1b23a
creating a layer with Docker/docker-compose (#633) 2023-04-06 22:46:04 -03:00
oobabooga 64bcde56ab Minor css change 2023-04-06 20:14:29 -03:00
oobabooga 58ed87e5d9
Update requirements.txt 2023-04-06 18:42:54 -03:00
dependabot[bot] 21be80242e
Bump rwkv from 0.7.2 to 0.7.3 (#842) 2023-04-06 17:52:27 -03:00
OWKenobi 310bf46a94
Instruction Character Vicuna, Instruction Mode Bugfix (#838) 2023-04-06 17:40:44 -03:00
DavG25 20b8ca4482
Add CSS for lists (#833) 2023-04-06 16:15:04 -03:00
oobabooga 113f94b61e Bump transformers (16-bit llama must be reconverted/redownloaded) 2023-04-06 16:04:03 -03:00
oobabooga 5f4f38ca5d Merge branch 'main' of github.com:oobabooga/text-generation-webui 2023-04-06 14:38:29 -03:00
oobabooga ef0f748618
Prevent CPU version of Torch from being installed (#10 from jllllll/oobabooga-windows) 2023-04-06 13:54:14 -03:00
oobabooga d9e7aba714
Update README.md 2023-04-06 13:42:24 -03:00
oobabooga 59058576b5 Remove unused requirement 2023-04-06 13:28:21 -03:00
oobabooga eec3665845
Add instructions for updating requirements 2023-04-06 13:24:01 -03:00
oobabooga 03cb44fc8c Add new llama.cpp library (2048 context, temperature, etc now work) 2023-04-06 13:12:14 -03:00
EyeDeck 39f3fec913
Broaden GPTQ-for-LLaMA branch support (#820) 2023-04-06 12:16:48 -03:00
oobabooga 8cd899515e Change instruct html a bit 2023-04-06 12:00:20 -03:00
oobabooga 4a28f39823
Update README.md 2023-04-06 02:47:27 -03:00
oobabooga 158ec51ae3 Increase instruct mode padding 2023-04-06 02:20:52 -03:00
Alex "mcmonkey" Goodwin 0c7ef26981
Lora trainer improvements (#763) 2023-04-06 02:04:11 -03:00
oobabooga 5b301d9a02 Create a Model tab 2023-04-06 01:54:05 -03:00
oobabooga 4a400320dd Clean up 2023-04-06 01:47:00 -03:00
oobabooga e94ab5dac1 Minor fixes 2023-04-06 01:43:10 -03:00
Randell Miller 641646a801
Fix crash if missing instructions directory (#812) 2023-04-06 01:24:22 -03:00
oobabooga 3f3e42e26c
Refactor several function calls and the API 2023-04-06 01:22:15 -03:00
SDS 378d21e80c
Add LLaMA-Precise preset (#767) 2023-04-05 18:52:36 -03:00
jllllll 1e656bef25
Specifically target cuda 11.7 ver. of torch 2.0.0
Move conda-forge channel to global list of channels
Hopefully prevents missing or incorrect packages
2023-04-05 16:52:05 -05:00
eiery 19b516b11b
fix link to streaming api example (#803) 2023-04-05 14:50:23 -03:00
oobabooga 7617ed5bfd
Add AMD instructions 2023-04-05 14:42:58 -03:00
oobabooga 770ef5744f Update README 2023-04-05 14:38:11 -03:00
Forkoz 8203ce0cac
Stop character pic from being cached when changing chars or clearing. (#798)
Tested on both FF and chromium
2023-04-05 14:25:01 -03:00
oobabooga 7f66421369 Fix loading characters 2023-04-05 14:22:32 -03:00
oobabooga 90141bc1a8 Fix saving prompts on Windows 2023-04-05 14:08:54 -03:00
oobabooga cf2c4e740b Disable gradio analytics globally 2023-04-05 14:05:50 -03:00
oobabooga e722c240af Add Instruct mode 2023-04-05 13:54:50 -03:00
oobabooga 3d6cb5ed63 Minor rewrite 2023-04-05 01:21:40 -03:00
oobabooga f3a2e0b8a9 Disable pre_layer when the model type is not llama 2023-04-05 01:19:26 -03:00
oobabooga ca8bb38949 Simplify gallery 2023-04-05 00:34:17 -03:00
catalpaaa 4ab679480e
allow quantized model to be loaded from model dir (#760) 2023-04-04 23:19:38 -03:00
oobabooga ae1fe45bc0 One more cache reset 2023-04-04 23:15:57 -03:00
oobabooga 8ef89730a5 Try to better handle browser image cache 2023-04-04 23:09:28 -03:00
oobabooga cc6c7a37f3 Add make_thumbnail function 2023-04-04 23:03:58 -03:00
oobabooga 80dfba05f3 Better crop/resize cached images 2023-04-04 22:52:15 -03:00
oobabooga 65d8a24a6d Show profile pictures in the Character tab 2023-04-04 22:28:49 -03:00
oobabooga f70a2e3ad4
Second attempt at fixing empty space 2023-04-04 18:30:34 -03:00
oobabooga 9c86acda67
Fix huge empty space in the Character tab 2023-04-04 18:07:34 -03:00
oobabooga 38afc2470c
Change indentation 2023-04-04 16:32:27 -03:00
oobabooga b2ce7282a1
Use past transformers version #773 2023-04-04 16:11:42 -03:00
jllllll 5aaf771c7d
Add additional sanity check
Add environment creation error
Improve error visibility
2023-04-04 12:31:26 -05:00
OWKenobi ee4547cd34
Detect "vicuna" as llama model type (#772) 2023-04-04 13:23:27 -03:00
oobabooga 881dbc3d44
Add back the name 2023-04-04 13:11:34 -03:00
oobabooga af0cb283e4
improve the example character yaml format (#770 from mcmonkey4eva) 2023-04-04 12:52:21 -03:00
Alex "mcmonkey" Goodwin 165d757444 improve the example character yaml format - use multiline blocks
multiline blocks make the input much cleaner and simpler, particularly for the example_dialogue. For the greeting block it can kinda go either way but I think it still ends up nicer. Also double quotes in context fixes the need to escape the singlequote inside.
2023-04-04 08:25:11 -07:00
oobabooga 8de22ac82a Merge character upload tabs 2023-04-03 18:01:45 -03:00
oobabooga b24147c7ca Document --pre_layer 2023-04-03 17:34:25 -03:00
oobabooga 4c9ed09270 Update settings template 2023-04-03 14:59:26 -03:00
dependabot[bot] ad37f396fc
Bump rwkv from 0.7.1 to 0.7.2 (#747) 2023-04-03 14:29:57 -03:00
dependabot[bot] 18f756ada6
Bump gradio from 3.24.0 to 3.24.1 (#746) 2023-04-03 14:29:37 -03:00
Niels Mündler 7aab88bcc6
Give API extension access to all generate_reply parameters (#744)
* Make every parameter of the generate_reply function parameterizable

* Add stopping strings as parameterizable
2023-04-03 13:31:12 -03:00
oobabooga 9318e16ed5 Expand .gitignore 2023-04-03 12:51:30 -03:00
oobabooga 3012bdb5e0 Fix a label 2023-04-03 12:20:53 -03:00
OWKenobi dcf61a8897
"character greeting" displayed and editable on the fly (#743)
* Add greetings field

* add greeting field and make it interactive

* Minor changes

* Fix a bug

* Simplify clear_chat_log

* Change a label

* Minor change

* Simplifications

* Simplification

* Simplify loading the default character history

* Fix regression

---------

Co-authored-by: oobabooga
2023-04-03 12:16:15 -03:00
Alex "mcmonkey" Goodwin 8b1f20aa04
Fix some old JSON characters not loading (#740) 2023-04-03 10:49:28 -03:00
oobabooga 8b442305ac Rename another variable 2023-04-03 01:15:20 -03:00
oobabooga 08448fb637 Rename a variable 2023-04-03 01:02:11 -03:00
oobabooga 2a267011dc Use Path.stem for simplicity 2023-04-03 00:56:14 -03:00
oobabooga 9b4e9a98f0
Merge pull request #9 from jllllll/oobabooga-windows
Add -k flag to curl command
2023-04-03 00:31:14 -03:00
jllllll c86d3e9c74
Add -k flag to curl command
Disables SSL certificate verification which was causing curl to fail on some systems.
https://github.com/oobabooga/text-generation-webui/issues/644#issuecomment-1493518391
2023-04-02 21:28:04 -05:00
Alex "mcmonkey" Goodwin ea97303509
Apply dialogue format in all character fields not just example dialogue (#650) 2023-04-02 21:54:29 -03:00
oobabooga 525f729b8e
Update README.md 2023-04-02 21:12:41 -03:00
oobabooga 53084241b4
Update README.md 2023-04-02 20:50:06 -03:00
TheTerrasque 2157bb4319
New yaml character format (#337 from TheTerrasque/feature/yaml-characters)
This doesn't break backward compatibility with JSON characters.
2023-04-02 20:34:25 -03:00
oobabooga 7ce608d101
Merge pull request #732 from StefanDanielSchwarz/fix-verbose-(beam-search)-preset
Fix "Verbose (Beam Search)" preset
2023-04-02 19:38:11 -03:00
SDS 34c3b4af6e
Fix "Verbose (Beam Search)" preset
Just a quick fix that removes an erroneous space between "length_penalty" and "=" (doesn't affect Python, but makes it possible to source the file from Bash, e. g. to use the variables with API calls)
2023-04-03 00:31:58 +02:00
oobabooga 1a823aaeb5
Clear text input for chat (#715 from bmoconno/clear-chat-input) 2023-04-02 18:08:25 -03:00
oobabooga 0dc6fa038b Use gr.State() to store the user input 2023-04-02 18:05:21 -03:00
oobabooga 5f3f3faa96 Better handle CUDA out of memory errors in chat mode 2023-04-02 17:48:00 -03:00
oobabooga e3c348e42b
Add .git 2023-04-02 01:11:05 -03:00
oobabooga b704fe7878
Use my fork of GPTQ-for-LLaMa for stability 2023-04-02 01:10:22 -03:00
Brian O'Connor d0f9625f0b Clear text input for chat
Add logic to clear the textbox for chat input when the user submits or hits the generate button.
2023-04-01 21:48:24 -04:00
oobabooga b0890a7925 Add shared.is_chat() function 2023-04-01 20:15:00 -03:00
oobabooga b38ba230f4
Update download-model.py 2023-04-01 15:03:24 -03:00
oobabooga b6f817be45
Update README.md 2023-04-01 14:54:10 -03:00
oobabooga 88fa38ac01
Update README.md 2023-04-01 14:49:03 -03:00
oobabooga 526d5725db
Update download-model.py 2023-04-01 14:47:47 -03:00
oobabooga 4b57bd0d99
Update README.md 2023-04-01 14:38:04 -03:00
oobabooga b53bec5a1f
Update README.md 2023-04-01 14:37:35 -03:00
oobabooga 9160586c04
Update README.md 2023-04-01 14:31:10 -03:00
oobabooga 7ec11ae000
Update README.md 2023-04-01 14:15:19 -03:00
oobabooga b857f4655b
Update shared.py 2023-04-01 13:56:47 -03:00
oobabooga 012f4f83b8
Update README.md 2023-04-01 13:55:15 -03:00
oobabooga fcda3f8776 Add also_return_rows to generate_chat_prompt 2023-04-01 01:12:13 -03:00
oobabooga 8c51b405e4 Progress towards generalizing Interface mode tab 2023-03-31 23:41:10 -03:00
oobabooga 23116b88ef
Add support for resuming downloads (#654 from nikita-skakun/support-partial-downloads) 2023-03-31 22:55:55 -03:00
oobabooga 74462ac713 Don't override the metadata when checking the sha256sum 2023-03-31 22:52:52 -03:00
oobabooga 2c52310642 Add --threads flag for llama.cpp 2023-03-31 21:18:05 -03:00
oobabooga eeafd60713 Fix streaming 2023-03-31 19:05:38 -03:00
oobabooga 52065ae4cd Add repetition_penalty 2023-03-31 19:01:34 -03:00
oobabooga 2259143fec Fix llama.cpp with --no-stream 2023-03-31 18:43:45 -03:00
oobabooga 875de5d983 Update ggml template 2023-03-31 17:57:31 -03:00
oobabooga cbfe0b944a
Update README.md 2023-03-31 17:49:11 -03:00
oobabooga 6a44f4aec6 Add support for downloading ggml files 2023-03-31 17:33:42 -03:00
oobabooga 3a47a602a3 Detect ggml*.bin files automatically 2023-03-31 17:18:21 -03:00
oobabooga 0aee7341d8 Properly count tokens/s for llama.cpp in chat mode 2023-03-31 17:04:32 -03:00
oobabooga 5c4e44b452
llama.cpp documentation 2023-03-31 15:20:39 -03:00
oobabooga 6fd70d0032
Add llama.cpp support (#447 from thomasantony/feature/llamacpp)
Documentation: https://github.com/oobabooga/text-generation-webui/wiki/llama.cpp-models
2023-03-31 15:17:32 -03:00
oobabooga a5c9b7d977 Bump llamacpp version 2023-03-31 15:08:01 -03:00
oobabooga ea3ba6fc73 Merge branch 'feature/llamacpp' of github.com:thomasantony/text-generation-webui into thomasantony-feature/llamacpp 2023-03-31 14:45:53 -03:00
oobabooga 09b0a3aafb Add repetition_penalty 2023-03-31 14:45:17 -03:00
oobabooga 4d98623041
Merge branch 'main' into feature/llamacpp 2023-03-31 14:37:04 -03:00
oobabooga 4c27562157 Minor changes 2023-03-31 14:33:46 -03:00
oobabooga 9d1dcf880a General improvements 2023-03-31 14:27:01 -03:00
oobabooga 770ff0efa9 Merge branch 'main' of github.com:oobabooga/text-generation-webui 2023-03-31 12:22:22 -03:00
oobabooga 1d1d9e40cd Add seed to settings 2023-03-31 12:22:07 -03:00
oobabooga daeab6bac7
Merge pull request #678 from mayaeary/fix/python3.8
Fix `type object is not subscriptable`
2023-03-31 12:19:06 -03:00
oobabooga 75465fa041
Merge pull request #6 from jllllll/oobabooga-windows
Attempt to Improve Reliability
2023-03-31 11:27:23 -03:00
oobabooga 5a6f939f05 Change the preset here too 2023-03-31 10:43:05 -03:00
Maya b246d17513
Fix type object is not subscriptable
Fix `type object is not subscriptable` on python 3.8
2023-03-31 14:20:31 +03:00
Nikita Skakun b99bea3c69 Fixed reported header affecting resuming download 2023-03-30 23:11:59 -07:00
oobabooga 3e1267af79
Merge pull request #673 from ye7iaserag/patch-1
Implement character gallery using Dataset
2023-03-31 02:04:52 -03:00
oobabooga 3b90d604d7 Sort the imports 2023-03-31 02:01:48 -03:00
oobabooga d28a5c9569 Remove unnecessary css 2023-03-31 02:01:13 -03:00
ye7iaserag ec093a5af7
Fix div alignment for long strings 2023-03-31 06:54:24 +02:00
oobabooga 92c7068daf Don't download if --check is specified 2023-03-31 01:31:47 -03:00
oobabooga 3737eafeaa Remove a border and allow more characters per pagination page 2023-03-31 00:48:50 -03:00
oobabooga fd72afd8e7 Increase the textbox sizes 2023-03-31 00:43:00 -03:00
oobabooga f27a66b014 Bump gradio version (make sure to update)
This fixes the textbox shrinking vertically once it reaches
a certain number of lines.
2023-03-31 00:42:26 -03:00
Nikita Skakun 0cc89e7755 Checksum code now activated by --check flag. 2023-03-30 20:06:12 -07:00
ye7iaserag f9940b79dc
Implement character gallery using Dataset 2023-03-31 04:56:49 +02:00
jllllll e4e3c9095d
Add warning for long paths 2023-03-30 20:48:40 -05:00
jllllll 172035d2e1
Minor Correction 2023-03-30 20:44:56 -05:00
jllllll 0b4ee14edc
Attempt to Improve Reliability
Have pip directly download and install backup GPTQ wheel instead of first downloading through curl.
Install bitsandbytes from wheel compiled for Windows from modified source.
Add clarification of minor, intermittent issue to instructions.
Add system32 folder to end of PATH rather than beginning.
Add warning when installed under a path containing spaces.
2023-03-30 20:04:16 -05:00
oobabooga bb69e054a7 Add dummy file 2023-03-30 21:08:50 -03:00
oobabooga 85e4ec6e6b
Download the cuda branch directly 2023-03-30 18:22:48 -03:00
oobabooga 78c0da4a18
Use the cuda branch of gptq-for-llama
Did I do this right @jllllll? This is because the current default branch (triton) is not compatible with Windows.
2023-03-30 18:04:05 -03:00
oobabooga d4a9b5ea97 Remove redundant preset (see the plot in #587) 2023-03-30 17:34:44 -03:00
Nikita Skakun d550c12a3e Fixed the bug with additional bytes.
The issue seems to be with huggingface not reporting the entire size of the model.
Added an error message with instructions if the checksums don't match.
2023-03-30 12:52:16 -07:00
Thomas Antony 7fa5d96c22 Update to use new llamacpp API 2023-03-30 11:23:05 +01:00
Thomas Antony 79fa2b6d7e Add support for alpaca 2023-03-30 11:23:04 +01:00
Thomas Antony 8953a262cb Add llamacpp to requirements.txt 2023-03-30 11:22:38 +01:00
Thomas Antony a5f5736e74 Add to text_generation.py 2023-03-30 11:22:38 +01:00
Thomas Antony 7745faa7bb Add llamacpp to models.py 2023-03-30 11:22:37 +01:00
Thomas Antony 7a562481fa Initial version of llamacpp_model.py 2023-03-30 11:22:07 +01:00
Thomas Antony 53ab1e285d Update .gitignore 2023-03-30 11:22:07 +01:00
Nikita Skakun 297ac051d9 Added sha256 validation of model files. 2023-03-30 02:34:19 -07:00
Nikita Skakun 8c590c2362 Added a 'clean' flag to not resume download. 2023-03-30 00:42:19 -07:00
Nikita Skakun e17af59261 Add support for resuming downloads
This commit adds the ability to resume interrupted downloads by adding a new function to the downloader module. The function uses the HTTP Range header to fetch only the remaining part of a file that wasn't downloaded yet.
2023-03-30 00:21:34 -07:00
oobabooga f0fdab08d3 Increase --chat height 2023-03-30 01:02:11 -03:00
oobabooga bd65940a48 Increase --chat box height 2023-03-30 00:43:49 -03:00
oobabooga 131753fcf5 Save the sha256sum of downloaded models 2023-03-29 23:28:16 -03:00
oobabooga a21e580782 Move an import 2023-03-29 22:50:58 -03:00
oobabooga 55755e27b9 Don't hardcode prompts in the settings dict/json 2023-03-29 22:47:01 -03:00
oobabooga 1cb9246160 Adapt to the new model names 2023-03-29 21:47:36 -03:00
oobabooga 0345e04249 Fix "Unknown argument(s): {'verbose': False}" 2023-03-29 21:17:48 -03:00
oobabooga 9104164297
Merge pull request #618 from nikita-skakun/optimize-download-model
Improve download-model.py progress bar with multiple threads
2023-03-29 20:54:19 -03:00
oobabooga 37754164eb Move argparse 2023-03-29 20:47:36 -03:00
oobabooga 6403e72062 Merge branch 'main' into nikita-skakun-optimize-download-model 2023-03-29 20:45:33 -03:00
oobabooga 1445ea86f7 Add --output and better metadata for downloading models 2023-03-29 20:26:44 -03:00
oobabooga 58349f44a0
Handle training exception for unsupported models 2023-03-29 11:55:34 -03:00
oobabooga a6d0373063
Fix training dataset loading #636 2023-03-29 11:48:17 -03:00
oobabooga 41b58bc47e
Update README.md 2023-03-29 11:02:29 -03:00
oobabooga 0de4f24b12
Merge pull request #4 from jllllll/oobabooga-windows
Change Micromamba download link
2023-03-29 09:49:32 -03:00
jllllll ed0e593161
Change Micromamba download
Changed link to previous version.
This will provide a stable source for Micromamba so that new versions don't cause issues.
2023-03-29 02:47:19 -05:00
oobabooga 3b4447a4fe
Update README.md 2023-03-29 02:24:11 -03:00
oobabooga 5d0b83c341
Update README.md 2023-03-29 02:22:19 -03:00
oobabooga c2a863f87d
Mention the updated one-click installer 2023-03-29 02:11:51 -03:00
oobabooga da3aa8fbda
Merge pull request #2 from jllllll/oobabooga-windows
Update one-click-installer for Windows
2023-03-29 01:55:47 -03:00
oobabooga 1edfb96778
Fix loading extensions from within the interface 2023-03-28 23:27:02 -03:00
Nikita Skakun aaa218a102 Remove unused import. 2023-03-28 18:32:49 -07:00
Nikita Skakun ff515ec2fe Improve progress bar visual style
This commit reverts the performance improvements of the previous commit for for improved visual style of multithreaded progress bars. The style of the progress bar has been modified to take up the same amount of size to align them.
2023-03-28 18:29:20 -07:00
oobabooga 304f812c63 Gracefully handle CUDA out of memory errors with streaming 2023-03-28 19:20:50 -03:00
Nikita Skakun 4d8e101006 Refactor download process to use multiprocessing
The previous implementation used threads to download files in parallel, which could lead to performance issues due to the Global Interpreter Lock (GIL).
This commit refactors the download process to use multiprocessing instead,
which allows for true parallelism across multiple CPUs.
This results in significantly faster downloads, particularly for large models.
2023-03-28 14:24:23 -07:00
oobabooga b2f356a9ae
Generalize GPTQ_loader, support any model (#615 from mayaeary/feature/gpt-j-4bit-v2)
This includes Pygmalion 4bit
2023-03-28 18:00:09 -03:00
oobabooga 010b259dde Update documentation 2023-03-28 17:46:00 -03:00
oobabooga 0bec15ebcd Reorder imports 2023-03-28 17:34:15 -03:00
Maya Eary 41ec682834 Disable kernel threshold for gpt-j 2023-03-28 22:45:38 +03:00
Maya 1ac003d41c
Merge branch 'oobabooga:main' into feature/gpt-j-4bit-v2 2023-03-28 22:30:39 +03:00
oobabooga aebd3cf110
Merge pull request #616 from mayaeary/fix/api-convert-params
Fixes for api server - chat mode and integer temperature
2023-03-28 15:21:58 -03:00
Maya Eary d1377c37af Fixes for api server - chat mode and integer temperature 2023-03-28 20:57:16 +03:00
Maya Eary 1c075d8d21 Fix typo 2023-03-28 20:43:50 +03:00
Maya Eary c8207d474f Generalized load_quantized 2023-03-28 20:38:55 +03:00
oobabooga cac577d99f Fix interface reloading 2023-03-28 13:25:58 -03:00
oobabooga 88ad86249d Remove unnecessary file 2023-03-28 13:19:52 -03:00
oobabooga 91aa5b460e If both .pt and .safetensors are present, download only safetensors 2023-03-28 13:08:38 -03:00
oobabooga 8579fe51dd Fix new lines in the HTML tab 2023-03-28 12:59:34 -03:00
oobabooga 46f6536fae
Merge pull request #570 from mcmonkey4eva/add-train-lora-tab
LoRA Training Tab
2023-03-28 02:53:51 -03:00
Alex "mcmonkey" Goodwin b0f05046b3 remove duplicate import 2023-03-27 22:50:37 -07:00
Alex "mcmonkey" Goodwin e817fac542 better defaults 2023-03-27 22:29:23 -07:00
Alex "mcmonkey" Goodwin 9cc811a0e6 fix LoRA path typo in #549 2023-03-27 22:16:40 -07:00
Alex "mcmonkey" Goodwin 2e08af4edf implement initial Raw Text File Input
also bump default Rank & Alpha for values that will make sense in testing if you don't know what you're doing and leave the defaults.
2023-03-27 22:15:32 -07:00
Alex "mcmonkey" Goodwin b749952fe3 change number minimums to 0
gradio calculates 'step' relative to the minimum, so at '1' the step values were all offset awkwardly. 0 isn't valid, but, uh, just don't slam the slider to the left.
2023-03-27 21:22:43 -07:00
Alex "mcmonkey" Goodwin ec6224f556 use new shared.args.lora_dir 2023-03-27 20:04:16 -07:00
Alex "mcmonkey" Goodwin 31f04dc615 Merge branch 'main' into add-train-lora-tab 2023-03-27 20:03:30 -07:00
oobabooga 966168bd2a
Merge pull request #602 from oobabooga/dependabot/pip/accelerate-0.18.0
Bump accelerate from 0.17.1 to 0.18.0
2023-03-27 23:53:26 -03:00
oobabooga c188975a01
Merge pull request #549 from catalpaaa/lora-and-model-dir
lora-dir, model-dir and login auth
2023-03-27 23:46:47 -03:00
oobabooga 53da672315 Fix FlexGen 2023-03-27 23:44:21 -03:00
oobabooga ee95e55df6 Fix RWKV tokenizer 2023-03-27 23:42:29 -03:00
oobabooga 036163a751 Change description 2023-03-27 23:39:26 -03:00
oobabooga 30585b3e71 Update README 2023-03-27 23:35:01 -03:00
oobabooga 005f552ea3 Some simplifications 2023-03-27 23:29:52 -03:00
oobabooga fde92048af Merge branch 'main' into catalpaaa-lora-and-model-dir 2023-03-27 23:16:44 -03:00
Alex "mcmonkey" Goodwin 8a97f6ba29 corrections per the PR comments 2023-03-27 18:39:06 -07:00
dependabot[bot] 1e02f75f2b
Bump accelerate from 0.17.1 to 0.18.0
Bumps [accelerate](https://github.com/huggingface/accelerate) from 0.17.1 to 0.18.0.
- [Release notes](https://github.com/huggingface/accelerate/releases)
- [Commits](https://github.com/huggingface/accelerate/compare/v0.17.1...v0.18.0)

---
updated-dependencies:
- dependency-name: accelerate
  dependency-type: direct:production
  update-type: version-update:semver-minor
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-03-28 01:19:34 +00:00
oobabooga 37f11803e3
Merge pull request #603 from oobabooga/dependabot/pip/rwkv-0.7.1
Bump rwkv from 0.7.0 to 0.7.1
2023-03-27 22:19:08 -03:00
Alex "mcmonkey" Goodwin 7fab7ea1b6 couple missed camelCases 2023-03-27 18:19:06 -07:00
oobabooga 1fc7ff065d
Bump bitsandbytes from 0.37.1 to 0.37.2 2023-03-27 22:18:52 -03:00
Alex "mcmonkey" Goodwin 6368dad7db Fix camelCase to snake_case to match repo format standard 2023-03-27 18:17:42 -07:00
oobabooga 2f0571bfa4 Small style changes 2023-03-27 21:24:39 -03:00
oobabooga c2cad30772 Merge branch 'main' into mcmonkey4eva-add-train-lora-tab 2023-03-27 21:05:44 -03:00
dependabot[bot] e9c0226b09
Bump rwkv from 0.7.0 to 0.7.1
Bumps [rwkv](https://github.com/BlinkDL/ChatRWKV) from 0.7.0 to 0.7.1.
- [Release notes](https://github.com/BlinkDL/ChatRWKV/releases)
- [Commits](https://github.com/BlinkDL/ChatRWKV/commits)

---
updated-dependencies:
- dependency-name: rwkv
  dependency-type: direct:production
  update-type: version-update:semver-patch
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-03-27 21:05:35 +00:00
dependabot[bot] 9c96919121
Bump bitsandbytes from 0.37.1 to 0.37.2
Bumps [bitsandbytes](https://github.com/TimDettmers/bitsandbytes) from 0.37.1 to 0.37.2.
- [Release notes](https://github.com/TimDettmers/bitsandbytes/releases)
- [Changelog](https://github.com/TimDettmers/bitsandbytes/blob/main/CHANGELOG.md)
- [Commits](https://github.com/TimDettmers/bitsandbytes/commits)

---
updated-dependencies:
- dependency-name: bitsandbytes
  dependency-type: direct:production
  update-type: version-update:semver-patch
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-03-27 21:05:19 +00:00
oobabooga 9ec6c56680
Update stale.yml 2023-03-27 15:12:43 -03:00
Alex "mcmonkey" Goodwin 9ced75746d add total time estimate 2023-03-27 10:57:27 -07:00
oobabooga 641e1a09a7 Don't flash when selecting a new prompt 2023-03-27 14:48:43 -03:00
Alex "mcmonkey" Goodwin 16ea4fc36d interrupt button 2023-03-27 10:43:01 -07:00
Alex "mcmonkey" Goodwin 8fc723fc95 initial progress tracker in UI 2023-03-27 10:25:08 -07:00
oobabooga 48a6c9513e
Merge pull request #572 from clusterfudge/issues/571
Potential fix for issues/571
2023-03-27 14:06:38 -03:00
oobabooga 268abd1cba Add some space in notebook mode 2023-03-27 13:52:12 -03:00
Alex "mcmonkey" Goodwin c07bcd0850 add some outputs to indicate progress updates (sorta)
Actual progressbar still needed. Also minor formatting fixes.
2023-03-27 09:41:06 -07:00
oobabooga af65c12900 Change Stop button behavior 2023-03-27 13:23:59 -03:00
oobabooga addb9777f9 Increase size of GALACTICA equations 2023-03-27 12:59:07 -03:00
oobabooga 572bafcd24 Less verbose message 2023-03-27 12:43:37 -03:00
Alex "mcmonkey" Goodwin 2afe1c13c1 move Training to before Interface mode
as Interface Mode seems to be a core 'settings' page that naturally belongs at the very end
2023-03-27 08:32:32 -07:00
Alex "mcmonkey" Goodwin d911c22af9 use shared rows to make the LoRA Trainer interface a bit more compact / clean 2023-03-27 08:31:49 -07:00
oobabooga 202e981d00 Make Generate/Stop buttons smaller in notebook mode 2023-03-27 12:30:57 -03:00
Alex "mcmonkey" Goodwin e439228ed8 Merge branch 'main' into add-train-lora-tab 2023-03-27 08:21:19 -07:00
oobabooga 8e2d94a5a1 Add saved promtps to gitignore 2023-03-27 12:21:19 -03:00
oobabooga 57345b8f30 Add prompt loading/saving menus + reorganize interface 2023-03-27 12:16:37 -03:00
jllllll cb5dff0087
Update installer to use official micromamba url 2023-03-26 23:40:46 -05:00
oobabooga 3dc61284d5 Handle unloading LoRA from dropdown menu icon 2023-03-27 00:04:43 -03:00
oobabooga b6e38e8b97
silero_tts streaming fix (#568 from Brawlence/silero_tts-fix)
silero_tts streaming fix
2023-03-26 23:59:07 -03:00
jllllll bdf85ffcf9
Remove explicit pytorch installation
Fixes an issue some people were having: https://github.com/oobabooga/text-generation-webui/issues/15
I did not experience this issue on my system. Not everyone does for some reason.
2023-03-26 21:56:16 -05:00
oobabooga af603a142a
Unload models on request (#471 from Brawlence/main) 2023-03-26 23:53:39 -03:00
oobabooga 95c97e1747 Unload the model using the "Remove all" button 2023-03-26 23:47:29 -03:00
oobabooga e07c9e3093 Merge branch 'main' into Brawlence-main 2023-03-26 23:40:51 -03:00
oobabooga 511be06dcc Merge branch 'main' of github.com:oobabooga/text-generation-webui 2023-03-26 22:21:29 -03:00
oobabooga 1c77fdca4c Change notebook mode appearance 2023-03-26 22:20:30 -03:00
oobabooga 9ff6a538b6 Bump gradio version
Make sure to upgrade with

`pip install -r requirements.txt --upgrade`
2023-03-26 22:11:19 -03:00
oobabooga a04b7cf264
Merge pull request #585 from fkusche/also-download-markdown
Also download Markdown files
2023-03-26 14:51:23 -03:00
Florian Kusche 19174842b8 Also download Markdown files 2023-03-26 19:41:14 +02:00
oobabooga 8222d32240
Merge pull request #565 from mcmonkey4eva/improve-gitignore
improve/simplify gitignore
2023-03-26 13:31:45 -03:00
jllllll 6f89242094
Remove temporary fix for GPTQ-for-LLaMa
No longer necessary.
2023-03-26 03:29:14 -05:00
jllllll 6dcfcf4fed
Amended fix for GPTQ-for-LLaMa
Prevents breaking 3-bit support
2023-03-26 01:00:52 -05:00
jllllll 12baa0e84b
Update for latest GPTQ-for-LLaMa 2023-03-26 00:46:07 -05:00
jllllll 247e8e5b79
Fix for issue in current GPTQ-for-LLaMa. 2023-03-26 00:24:00 -05:00
oobabooga 49c10c5570
Add support for the latest GPTQ models with group-size (#530)
**Warning: old 4-bit weights will not work anymore!**

See here how to get up to date weights: https://github.com/oobabooga/text-generation-webui/wiki/LLaMA-model#step-2-get-the-pre-converted-weights
2023-03-26 00:11:33 -03:00
Sean Fitzgerald 0bac80d9eb Potential fix for issues/571 2023-03-25 13:08:45 -07:00
Alex "mcmonkey" Goodwin f1ba2196b1 make 'model' variables less ambiguous 2023-03-25 12:57:36 -07:00
Alex "mcmonkey" Goodwin 8da237223e document options better 2023-03-25 12:48:35 -07:00
Alex "mcmonkey" Goodwin 8134c4b334 add training/datsets to gitignore for #570 2023-03-25 12:41:18 -07:00
Alex "mcmonkey" Goodwin 5c49a0dcd0 fix error from prepare call running twice in a row 2023-03-25 12:37:32 -07:00
Alex "mcmonkey" Goodwin 7bf601107c automatically strip empty data entries (for better alpaca dataset compat) 2023-03-25 12:28:46 -07:00
Alex "mcmonkey" Goodwin 566898a79a initial lora training tab 2023-03-25 12:08:26 -07:00
Φφ 1a1e420e65 Silero_tts streaming fix
Temporarily suppress the streaming during the audio response as it would interfere with the audio (making it stutter and play anew)
2023-03-25 21:33:30 +03:00
Alex "mcmonkey" Goodwin 9ccf505ccd improve/simplify gitignore
- add repositories
- remove the redundant "/*" on folders
- remove the exclusions for files that already exist
2023-03-25 10:04:00 -07:00
oobabooga 8c8e8b4450
Fix the early stopping callback #559 2023-03-25 12:35:52 -03:00
oobabooga a1f12d607f
Merge pull request #538 from Ph0rk0z/display-input-context
Add display of context when input was generated
2023-03-25 11:56:18 -03:00
catalpaaa f740ee558c
Merge branch 'oobabooga:main' into lora-and-model-dir 2023-03-25 01:28:33 -07:00
jllllll ce9a5e3b53
Update install.bat
Minor fixes
2023-03-25 02:22:02 -05:00
jllllll 2e02d42682 Changed things around to allow Micromamba to work with paths containing spaces. 2023-03-25 01:26:25 -05:00
oobabooga 70f9565f37
Update README.md 2023-03-25 02:35:30 -03:00
oobabooga 25be9698c7
Fix LoRA on mps 2023-03-25 01:18:32 -03:00
oobabooga 3da633a497
Merge pull request #529 from EyeDeck/main
Allow loading of .safetensors through GPTQ-for-LLaMa
2023-03-24 23:51:01 -03:00
jllllll 1e260544cd
Update install.bat
Added C:\Windows\System32 to PATH to avoid issues with broken? Windows installs.
2023-03-24 21:25:14 -05:00
catalpaaa d51cb8292b Update server.py
yea i should go to bed
2023-03-24 17:36:31 -07:00
catalpaaa 9e2963e0c8 Update server.py 2023-03-24 17:35:45 -07:00
catalpaaa ec2a1facee Update server.py 2023-03-24 17:34:33 -07:00
catalpaaa b37c54edcf lora-dir, model-dir and login auth
Added lora-dir, model-dir, and a login auth arguments that points to a file contains usernames and passwords in the format of "u:pw,u:pw,..."
2023-03-24 17:30:18 -07:00
jllllll fa916aa1de
Update INSTRUCTIONS.txt
Added clarification on new variable added to download-model.bat.
2023-03-24 18:28:46 -05:00
jllllll 586775ad47
Update download-model.bat
Removed redundant %ModelName% variable.
2023-03-24 18:25:49 -05:00
jllllll bddbc2f898
Update start-webui.bat
Updated virtual environment handling to use Micromamba.
2023-03-24 18:19:23 -05:00
jllllll 2604e3f7ac
Update download-model.bat
Added variables for model selection and text only mode.
Updated virtual environment handling to use Micromamba.
2023-03-24 18:15:24 -05:00
jllllll 24870e51ed
Update micromamba-cmd.bat
Add cd command for admin.
2023-03-24 18:12:02 -05:00
jllllll f0c82f06c3
Add files via upload
Add script to open cmd within installation environment for easier modification.
2023-03-24 18:09:44 -05:00
oobabooga 9fa47c0eed
Revert GPTQ_loader.py (accident) 2023-03-24 19:57:12 -03:00
oobabooga a6bf54739c
Revert models.py (accident) 2023-03-24 19:56:45 -03:00
jllllll eec773b1f4
Update install.bat
Corrected libbitsandbytes_cudaall.dll install.
2023-03-24 17:54:47 -05:00
oobabooga 0a16224451
Update GPTQ_loader.py 2023-03-24 19:54:36 -03:00
oobabooga a80aa65986
Update models.py 2023-03-24 19:53:20 -03:00
jllllll 817e6c681e
Update install.bat
Added `cd /D "%~dp0"` in case the script is ran as admin.
2023-03-24 17:51:13 -05:00
jllllll a80a5465f2
Update install.bat
Updated Conda packages and channels to install cuda-toolkit and override 12.0 cuda packages requested by pytorch with their 11.7 equivalent.
Removed Conda installation since we can use the downloaded Micromamba.exe for the same purpose with a smaller footprint.
Removed redundant PATH changes.
Changed %gpuchoice% comparisons to be case-insensitive.
Added additional error handling and removed the use of .tmp files.
Added missing extension requirements.
Added GPTQ installation. Will attempt to compile locally and, if failed, will download and install a precompiled wheel.
Incorporated fixes from one-click-bandaid.
Fixed and expanded first sed command from one-click-bandaid.
libbitsandbytes_cudaall.dll is used here as the cuda116.dll used by one-click-bandaid does not work on my 1080ti. This can be changed if needed.
2023-03-24 17:27:29 -05:00
oobabooga 507db0929d
Do not use empty user messages in chat mode
This allows the bot to send messages by clicking on Generate with empty inputs.
2023-03-24 17:22:22 -03:00
oobabooga 6e1b16c2aa
Update html_generator.py 2023-03-24 17:18:27 -03:00
oobabooga ffb0187e83
Update chat.py 2023-03-24 17:17:29 -03:00
oobabooga c14e598f14
Merge pull request #433 from mayaeary/fix/api-reload
Fix api extension duplicating
2023-03-24 16:56:10 -03:00
oobabooga bfe960731f
Merge branch 'main' into fix/api-reload 2023-03-24 16:54:41 -03:00
oobabooga 4a724ed22f
Reorder imports 2023-03-24 16:53:56 -03:00
oobabooga 8fad84abc2
Update extensions.py 2023-03-24 16:51:27 -03:00
oobabooga d8e950d6bd
Don't load the model twice when using --lora 2023-03-24 16:30:32 -03:00
oobabooga fd99995b01
Make the Stop button more consistent in chat mode 2023-03-24 15:59:27 -03:00
Forkoz b740c5b284
Add display of context when input was generated
Not sure if I did this right but it does move with the conversation and seems to match value.
2023-03-24 08:56:07 -05:00
oobabooga 4f5c2ce785
Fix chat_generation_attempts 2023-03-24 02:03:30 -03:00
oobabooga 04417b658b
Update README.md 2023-03-24 01:40:43 -03:00
oobabooga bb4cb22453
Download .pt files using download-model.py (for 4-bit models) 2023-03-24 00:49:04 -03:00
oobabooga 143b5b5edf
Mention one-click-bandaid in the README 2023-03-23 23:28:50 -03:00
EyeDeck dcfd866402 Allow loading of .safetensors through GPTQ-for-LLaMa 2023-03-23 21:31:34 -04:00
oobabooga 8747c74339
Another missing import 2023-03-23 22:19:01 -03:00
oobabooga 7078d168c3
Missing import 2023-03-23 22:16:08 -03:00
oobabooga d1327f99f9
Fix broken callbacks.py 2023-03-23 22:12:24 -03:00
oobabooga 9bdb3c784d
Minor fix 2023-03-23 22:02:40 -03:00
oobabooga b0abb327d8
Update LoRA.py 2023-03-23 22:02:09 -03:00
oobabooga bf22d16ebc
Clear cache while switching LoRAs 2023-03-23 21:56:26 -03:00
oobabooga 4578e88ffd
Stop the bot from talking for you in chat mode 2023-03-23 21:38:20 -03:00
oobabooga 9bf6ecf9e2
Fix LoRA device map (attempt) 2023-03-23 16:49:41 -03:00
oobabooga c5ebcc5f7e
Change the default names (#518)
* Update shared.py

* Update settings-template.json
2023-03-23 13:36:00 -03:00
Φφ 483d173d23 Code reuse + indication
Now shows the message in the console when unloading weights. Also reload_model() calls unload_model() first to free the memory so that multiple reloads won't overfill it.
2023-03-23 07:06:26 +03:00
Φφ 1917b15275 Unload and reload models on request 2023-03-23 07:06:26 +03:00
oobabooga 29bd41d453
Fix LoRA in CPU mode 2023-03-23 01:05:13 -03:00
oobabooga eac27f4f55
Make LoRAs work in 16-bit mode 2023-03-23 00:55:33 -03:00
oobabooga bfa81e105e
Fix FlexGen streaming 2023-03-23 00:22:14 -03:00
oobabooga 7b6f85d327
Fix markdown headers in light mode 2023-03-23 00:13:34 -03:00
oobabooga de6a09dc7f
Properly separate the original prompt from the reply 2023-03-23 00:12:40 -03:00
oobabooga d5fc1bead7
Merge pull request #489 from Brawlence/ext-fixes
Extensions performance & memory optimisations
2023-03-22 16:10:59 -03:00
oobabooga bfb1be2820
Minor fix 2023-03-22 16:09:48 -03:00
oobabooga 0abff499e2
Use image.thumbnail 2023-03-22 16:03:05 -03:00
oobabooga 104212529f
Minor changes 2023-03-22 15:55:03 -03:00
wywywywy 61346b88ea
Add "seed" menu in the Parameters tab 2023-03-22 15:40:20 -03:00
Φφ 5389fce8e1 Extensions performance & memory optimisations
Reworked remove_surrounded_chars() to use regular expression ( https://regexr.com/7alb5 ) instead of repeated string concatenations for elevenlab_tts, silero_tts, sd_api_pictures. This should be both faster and more robust in handling asterisks.

Reduced the memory footprint of send_pictures and sd_api_pictures by scaling the images in the chat to 300 pixels max-side wise. (The user already has the original in case of the sent picture and there's an option to save the SD generation).
This should fix history growing annoyingly large with multiple pictures present
2023-03-22 11:51:00 +03:00
oobabooga 45b7e53565
Only catch proper Exceptions in the text generation function 2023-03-20 20:36:02 -03:00
oobabooga 6872ffd976
Update README.md 2023-03-20 16:53:14 -03:00
oobabooga db4219a340
Update comments 2023-03-20 16:40:08 -03:00
oobabooga 7618f3fe8c
Add -gptq-preload for 4-bit offloading (#460)
This works in a 4GB card now:

```
python server.py --model llama-7b-hf --gptq-bits 4 --gptq-pre-layer 20
```
2023-03-20 16:30:56 -03:00
Vladimir Belitskiy e96687b1d6 Do not send empty user input as part of the prompt.
However, if extensions modify the empty prompt to be non-empty,
it'l still work as before.
2023-03-20 14:27:39 -04:00
oobabooga 9a3bed50c3
Attempt at fixing 4-bit with CPU offload 2023-03-20 15:11:56 -03:00
oobabooga 536d0a4d93
Add an import 2023-03-20 14:00:40 -03:00
Vladimir Belitskiy ca47e016b4
Do not display empty user messages in chat mode.
There doesn't seem to be much value to them - they just take up space while also making it seem like there's still some sort of pseudo-dialogue going on, instead of a monologue by the bot.
2023-03-20 12:55:57 -04:00
oobabooga 75a7a84ef2
Exception handling (#454)
* Update text_generation.py
* Update extensions.py
2023-03-20 13:36:52 -03:00
oobabooga a90f507abe
Exit elevenlabs_tts if streaming is enabled 2023-03-20 11:49:42 -03:00
oobabooga 31ab2be8ef Remove redundant requirements #309 2023-03-19 22:10:55 -03:00
oobabooga 164e05daad Download .py files using download-model.py 2023-03-19 20:34:52 -03:00
oobabooga dd4374edde Update README 2023-03-19 20:15:15 -03:00
oobabooga 9378754cc7 Update README 2023-03-19 20:14:50 -03:00
oobabooga 7ddf6147ac
Update README.md 2023-03-19 19:25:52 -03:00
oobabooga b552d2b58a Remove unused imports
o
2023-03-19 19:24:41 -03:00
oobabooga ddb62470e9 --no-cache and --gpu-memory in MiB for fine VRAM control 2023-03-19 19:21:41 -03:00
oobabooga 4bafe45a51
Merge pull request #309 from Brawlence/main
Extension: Stable Diffusion Api integration
2023-03-19 13:24:47 -03:00
oobabooga eab8de0d4a Merge branch 'main' into Brawlence-main 2023-03-19 13:09:59 -03:00
oobabooga 4d701a6eb9 Create a mirror for the preset menu 2023-03-19 12:51:47 -03:00
oobabooga 257edf5f56 Make the Default preset more reasonable
Credits: anonymous 4chan user who got it off
"some twitter post or something someone linked,
who even knows anymore"
2023-03-19 12:30:51 -03:00
oobabooga a78b6508fc Make custom LoRAs work by default #385 2023-03-19 12:11:35 -03:00
oobabooga 7073e96093 Add back RWKV dependency #98 2023-03-19 12:05:28 -03:00
oobabooga 217e1d9fdf Merge branch 'main' of github.com:oobabooga/text-generation-webui 2023-03-19 10:37:23 -03:00
oobabooga c79fc69e95 Fix the API example with streaming #417 2023-03-19 10:36:57 -03:00
Maya acdbd6b708 Check if app should display extensions ui 2023-03-19 13:31:21 +00:00
Maya 81c9d130f2 Fix global 2023-03-19 13:25:49 +00:00
Maya 099d7a844b Add setup method to extensions 2023-03-19 13:22:24 +00:00
Maya bd27353a08 Fix duplicating server on ui reload 2023-03-19 12:51:27 +00:00
oobabooga 0cbe2dd7e9
Update README.md 2023-03-18 12:24:54 -03:00
oobabooga 36ac7be76d
Merge pull request #407 from ThisIsPIRI/gitignore
Add loras to .gitignore
2023-03-18 11:57:10 -03:00
oobabooga d2a7fac8ea
Use pip instead of conda for pytorch 2023-03-18 11:56:04 -03:00
ThisIsPIRI 705f513c4c Add loras to .gitignore 2023-03-18 23:33:24 +09:00
oobabooga 9ed3a03d4b
Don't use the official instructions 2023-03-18 11:25:08 -03:00
oobabooga a0b1a30fd5
Specify torchvision/torchaudio versions 2023-03-18 11:23:56 -03:00
oobabooga c753261338 Disable stop_at_newline by default 2023-03-18 10:55:57 -03:00
oobabooga 7c945cfe8e Don't include PeftModel every time 2023-03-18 10:55:24 -03:00
oobabooga 86b99006d9
Remove rwkv dependency 2023-03-18 10:27:52 -03:00
oobabooga a163807f86
Update README.md 2023-03-18 03:07:27 -03:00
oobabooga a7acfa4893
Update README.md 2023-03-17 22:57:46 -03:00
oobabooga bcd8afd906
Merge pull request #393 from WojtekKowaluk/mps_support
Fix for MPS support on Apple Silicon
2023-03-17 22:57:28 -03:00
oobabooga e26763a510 Minor changes 2023-03-17 22:56:46 -03:00
Wojtek Kowaluk 7994b580d5 clean up duplicated code 2023-03-18 02:27:26 +01:00
oobabooga dc35861184
Update README.md 2023-03-17 21:05:17 -03:00
Wojtek Kowaluk 30939e2aee add mps support on apple silicon 2023-03-18 00:56:23 +01:00
Wojtek Kowaluk 7d97da1dcb add venv paths to gitignore 2023-03-18 00:54:17 +01:00
oobabooga f2a5ca7d49
Update README.md 2023-03-17 20:50:27 -03:00
oobabooga 8c8286b0e6
Update README.md 2023-03-17 20:49:40 -03:00
oobabooga 91371640f9
Use the official instructions
https://pytorch.org/get-started/locally/
2023-03-17 20:37:25 -03:00
oobabooga 0c05e65e5c
Update README.md 2023-03-17 20:25:42 -03:00
oobabooga adc200318a Merge branch 'main' of github.com:oobabooga/text-generation-webui 2023-03-17 20:19:33 -03:00
oobabooga 20f5b455bf Add parameters reference #386 #331 2023-03-17 20:19:04 -03:00
oobabooga 66e8d12354
Update README.md 2023-03-17 19:59:37 -03:00
oobabooga 9a871117d7
Update README.md 2023-03-17 19:52:22 -03:00
oobabooga d4f38b6a1f
Update README.md 2023-03-17 18:57:48 -03:00
oobabooga ad7c829953
Update README.md 2023-03-17 18:55:01 -03:00
oobabooga 4426f941e0
Update the installation instructions. Tldr use WSL 2023-03-17 18:51:07 -03:00
oobabooga 9256e937d6 Add some LoRA params 2023-03-17 17:45:28 -03:00
oobabooga 9ed2c4501c Use markdown in the "HTML" tab 2023-03-17 16:06:11 -03:00
oobabooga f0b26451b4 Add a comment 2023-03-17 13:07:17 -03:00
oobabooga 7da742e149
Merge pull request #207 from EliasVincent/stt-extension
Extension: Whisper Speech-To-Text Input
2023-03-17 12:37:23 -03:00
oobabooga ebef4a510b Update README 2023-03-17 11:58:45 -03:00
oobabooga cdfa787bcb Update README 2023-03-17 11:53:28 -03:00
oobabooga 3bda907727
Merge pull request #366 from oobabooga/lora
Add LoRA support
2023-03-17 11:48:48 -03:00
oobabooga 614dad0075 Remove unused import 2023-03-17 11:43:11 -03:00
oobabooga a717fd709d Sort the imports 2023-03-17 11:42:25 -03:00
oobabooga 7d97287e69 Update settings-template.json 2023-03-17 11:41:12 -03:00
oobabooga 29fe7b1c74 Remove LoRA tab, move it into the Parameters menu 2023-03-17 11:39:48 -03:00
oobabooga 214dc6868e Several QoL changes related to LoRA 2023-03-17 11:24:52 -03:00
oobabooga 4c130679c7
Merge pull request #377 from askmyteapot/Fix-Multi-gpu-GPTQ-Llama-no-tokens
Update GPTQ_Loader.py
2023-03-17 09:47:57 -03:00
askmyteapot 53b6a66beb
Update GPTQ_Loader.py
Correcting decoder layer for renamed class.
2023-03-17 18:34:13 +10:00
oobabooga 0cecfc684c Add files 2023-03-16 21:35:53 -03:00
oobabooga 104293f411 Add LoRA support 2023-03-16 21:31:39 -03:00
oobabooga ee164d1821 Don't split the layers in 8-bit mode by default 2023-03-16 18:22:16 -03:00
oobabooga 0a2aa79c4e
Merge pull request #358 from mayaeary/8bit-offload
Add support for memory maps with --load-in-8bit
2023-03-16 17:27:03 -03:00
oobabooga e085cb4333 Small changes 2023-03-16 13:34:23 -03:00
oobabooga dd1c5963da Update README 2023-03-16 12:45:27 -03:00
oobabooga 38d7017657 Add all command-line flags to "Interface mode" 2023-03-16 12:44:03 -03:00
awoo 83cb20aad8 Add support for --gpu-memory witn --load-in-8bit 2023-03-16 18:42:53 +03:00
oobabooga 23a5e886e1 The LLaMA PR has been merged into transformers
https://github.com/huggingface/transformers/pull/21955

The tokenizer class has been changed from

"LLaMATokenizer"

to

"LlamaTokenizer"

It is necessary to edit this change in every tokenizer_config.json
that you had for LLaMA so far.
2023-03-16 11:18:32 -03:00
oobabooga d54f3f4a34 Add no-stream checkbox to the interface 2023-03-16 10:19:00 -03:00
oobabooga 1c378965e1 Remove unused imports 2023-03-16 10:18:34 -03:00
oobabooga a577fb1077 Keep GALACTICA special tokens (#300) 2023-03-16 00:46:59 -03:00
oobabooga 25a00eaf98 Add "Experimental" warning 2023-03-15 23:43:35 -03:00
oobabooga 599d3139fd Increase the reload timeout a bit 2023-03-15 23:34:08 -03:00
oobabooga 4d64a57092 Add Interface mode tab 2023-03-15 23:29:56 -03:00
oobabooga b50172255a Merge branch 'main' of github.com:oobabooga/text-generation-webui 2023-03-15 20:46:04 -03:00
oobabooga ffb898608b Mini refactor 2023-03-15 20:44:34 -03:00
oobabooga d3a280e603
Merge pull request #348 from mayaeary/feature/koboldai-api-share
flask_cloudflared for shared tunnels
2023-03-15 20:22:08 -03:00
oobabooga 4639bc84c0
Specify torchvision version 2023-03-15 20:08:35 -03:00
oobabooga 445ebf0ba8
Update README.md 2023-03-15 20:06:46 -03:00
awoo 0552ab2e9f flask_cloudflared for shared tunnels 2023-03-16 02:00:16 +03:00
oobabooga e9e76bb06c
Delete WSL.md 2023-03-15 19:42:29 -03:00
oobabooga 09045e4bdb
Add WSL guide 2023-03-15 19:42:06 -03:00
oobabooga 9ff50334d2
Merge pull request #345 from jfryton/main
Guide for Windows Subsystem for Linux
2023-03-15 19:37:26 -03:00
oobabooga 66256ac1dd Make the "no GPU has been detected" message more descriptive 2023-03-15 19:31:27 -03:00
jfryton 055edc7fdd
Update WSL.md 2023-03-15 18:21:14 -04:00
jfryton 89883a399c
Create WSL.md guide for setting up WSL Ubuntu
Quick start guide for Windows Subsystem for Linux (Ubuntu), including port forwarding to enable local network webui access.
2023-03-15 18:20:21 -04:00
oobabooga 67d62475dc Further reorganize chat UI 2023-03-15 18:56:26 -03:00
oobabooga ab12a17bed
Merge pull request #342 from mayaeary/koboldai-api
Extension: KoboldAI api
2023-03-15 18:31:39 -03:00
awoo 30281122a5 KoboldAI api 2023-03-15 23:52:46 +03:00
oobabooga 6a1787a5fe CSS fixes 2023-03-15 16:55:40 -03:00
oobabooga 3047ed8ce3 CSS fix 2023-03-15 16:41:38 -03:00
oobabooga 87b84d2275 CSS fix 2023-03-15 16:39:59 -03:00
oobabooga c1959c26ee Show/hide the extensions block using javascript 2023-03-15 16:35:28 -03:00
oobabooga 348596f634 Fix broken extensions 2023-03-15 15:11:16 -03:00
oobabooga c5f14fb9b8 Optimize the HTML generation speed 2023-03-15 14:19:28 -03:00
oobabooga bf812c4893 Minor fix 2023-03-15 14:05:35 -03:00
oobabooga 658849d6c3 Move a checkbutton 2023-03-15 13:29:00 -03:00
oobabooga 05ee323ce5 Rename a file 2023-03-15 13:26:32 -03:00
oobabooga 40c9e4686c Add file 2023-03-15 13:25:28 -03:00
oobabooga d30a14087f Further reorganize the UI 2023-03-15 13:24:54 -03:00
oobabooga ffc6cb3116
Merge pull request #325 from Ph0rk0z/fix-RWKV-Names
Fix rwkv names
2023-03-15 12:56:21 -03:00
oobabooga cf2da86352 Prevent *Is typing* from disappearing instantly while streaming 2023-03-15 12:51:13 -03:00
oobabooga 4146ac4ff1
Merge pull request #266 from HideLord/main
Adding markdown support and slight refactoring.
2023-03-15 12:47:41 -03:00
oobabooga 29b7c5ac0c Sort the requirements 2023-03-15 12:40:03 -03:00
oobabooga ec972b85d1 Move all css/js into separate files 2023-03-15 12:35:11 -03:00
oobabooga 693b53d957 Merge branch 'main' into HideLord-main 2023-03-15 12:08:56 -03:00
oobabooga 1413931705 Add a header bar and redesign the interface (#293) 2023-03-15 12:01:32 -03:00
oobabooga 9d6a625bd6 Add 'hallucinations' filter #326
This breaks the API since a new parameter has been added.
It should be a one-line fix. See api-example.py.
2023-03-15 11:10:35 -03:00
Φφ e45d8e39c8
Merge branch 'oobabooga:main' into main 2023-03-15 14:04:06 +03:00
Forkoz 3b62bd180d
Remove PTH extension from RWKV
When loading the current model was blank unless you typed it out.
2023-03-14 21:23:39 +00:00
Forkoz f0f325eac1
Remove Json from loading
no more 20b tokenizer
2023-03-14 21:21:47 +00:00
oobabooga 128d18e298
Update README.md 2023-03-14 17:57:25 -03:00
oobabooga 1236c7f971
Update README.md 2023-03-14 17:56:15 -03:00
oobabooga b419dffba3
Update README.md 2023-03-14 17:55:35 -03:00
oobabooga 72d207c098
Remove the chat API
It is not implemented, has not been tested, and this is causing confusion.
2023-03-14 16:31:27 -03:00
oobabooga afc5339510
Remove "eval" statements from text generation functions 2023-03-14 16:04:17 -03:00
oobabooga 5c0522307f
Merge pull request #295 from Zerogoki00/opt4-bit
Add support for quantized OPT models
2023-03-14 08:05:24 -03:00
oobabooga 87192e2813 Update README 2023-03-14 08:02:21 -03:00
oobabooga 265ba384b7 Rename a file, add deprecation warning for --load-in-4bit 2023-03-14 07:56:31 -03:00
oobabooga 3da73e409f Merge branch 'main' into Zerogoki00-opt4-bit 2023-03-14 07:50:36 -03:00
Φφ d537b28d02 Extension: Stable Diffusion Api integration
Lets the bot answer you with a picture!
2023-03-14 06:49:10 +03:00
oobabooga b327554550
Update bug_report_template.yml 2023-03-14 00:18:13 -03:00
oobabooga 33b9a15232
Delete config.yml 2023-03-13 23:03:16 -03:00
oobabooga b5e0d3c227
Create config.yml 2023-03-13 23:02:25 -03:00
oobabooga 92d8b3e8dd
Update install.bat 2023-03-13 22:53:06 -03:00
oobabooga 3e71099402
Add --upgrade everywhere 2023-03-13 22:50:42 -03:00
oobabooga 88b2520fda
Add --upgrade to pip 2023-03-13 22:49:13 -03:00
oobabooga 7f301fd6ef
Merge pull request #305 from oobabooga/dependabot/pip/accelerate-0.17.1
Bump accelerate from 0.17.0 to 0.17.1
2023-03-13 22:41:21 -03:00
dependabot[bot] 02d407542c
Bump accelerate from 0.17.0 to 0.17.1
Bumps [accelerate](https://github.com/huggingface/accelerate) from 0.17.0 to 0.17.1.
- [Release notes](https://github.com/huggingface/accelerate/releases)
- [Commits](https://github.com/huggingface/accelerate/compare/v0.17.0...v0.17.1)

---
updated-dependencies:
- dependency-name: accelerate
  dependency-type: direct:production
  update-type: version-update:semver-patch
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-03-14 01:40:42 +00:00
oobabooga d685332c10
Merge pull request #307 from oobabooga/dependabot/pip/bitsandbytes-0.37.1
Bump bitsandbytes from 0.37.0 to 0.37.1
2023-03-13 22:39:59 -03:00
oobabooga 481ef3ca16
Merge pull request #304 from oobabooga/dependabot/pip/rwkv-0.4.2
Bump rwkv from 0.3.1 to 0.4.2
2023-03-13 22:39:22 -03:00
dependabot[bot] df83088593
Bump bitsandbytes from 0.37.0 to 0.37.1
Bumps [bitsandbytes](https://github.com/TimDettmers/bitsandbytes) from 0.37.0 to 0.37.1.
- [Release notes](https://github.com/TimDettmers/bitsandbytes/releases)
- [Changelog](https://github.com/TimDettmers/bitsandbytes/blob/main/CHANGELOG.md)
- [Commits](https://github.com/TimDettmers/bitsandbytes/commits)

---
updated-dependencies:
- dependency-name: bitsandbytes
  dependency-type: direct:production
  update-type: version-update:semver-patch
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-03-14 01:36:18 +00:00
dependabot[bot] 715c3ecba6
Bump rwkv from 0.3.1 to 0.4.2
Bumps [rwkv](https://github.com/BlinkDL/ChatRWKV) from 0.3.1 to 0.4.2.
- [Release notes](https://github.com/BlinkDL/ChatRWKV/releases)
- [Commits](https://github.com/BlinkDL/ChatRWKV/commits)

---
updated-dependencies:
- dependency-name: rwkv
  dependency-type: direct:production
  update-type: version-update:semver-minor
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-03-14 01:36:02 +00:00
oobabooga a0ef82c895
Activate dependabot 2023-03-13 22:35:28 -03:00
oobabooga 3fb8196e16 Implement "*Is recording a voice message...*" for TTS #303 2023-03-13 22:28:00 -03:00
oobabooga 0dab2c5dc5
Update feature_request.md 2023-03-13 22:18:03 -03:00
oobabooga 79e519cff6
Update stale.yml 2023-03-13 20:03:08 -03:00
oobabooga 15714580da
Update stale.yml 2023-03-13 19:39:21 -03:00
oobabooga bad0b0af48
Update stale.yml 2023-03-13 19:20:18 -03:00
oobabooga c805843b41
Update stale.yml 2023-03-13 19:09:06 -03:00
oobabooga 60cc7d3c3a
Update stale.yml 2023-03-13 18:53:11 -03:00
oobabooga 7c17613add
Update and rename .github/workflow/stale.yml to .github/workflows/stale.yml 2023-03-13 18:47:31 -03:00
oobabooga 47c941c5fd
Create stale.yml 2023-03-13 18:37:35 -03:00
oobabooga 511b1368b8
Update bug_report_template.yml 2023-03-13 18:29:38 -03:00
oobabooga d6763a6560
Update feature_request.md 2023-03-13 18:27:24 -03:00
oobabooga c6ecb354e5
Update feature_request.md 2023-03-13 18:26:28 -03:00
oobabooga 68464279e0
Update feature_request.md 2023-03-13 18:19:07 -03:00
oobabooga bcfb7d752a
Update bug_report_template.yml 2023-03-13 18:16:18 -03:00
oobabooga ed30bd3216
Update bug_report_template.yml 2023-03-13 18:14:54 -03:00
oobabooga aee3b53fb3
Update bug_report_template.yml 2023-03-13 18:14:31 -03:00
oobabooga 7dbc071e5a
Delete bug_report.md 2023-03-13 18:09:58 -03:00
oobabooga 69d4b818b7
Create bug_report_template.yml 2023-03-13 18:09:37 -03:00
oobabooga 0a75584706
Create issue templates 2023-03-13 18:07:08 -03:00
EliasVincent 02e1113d95 add auto-transcribe option 2023-03-13 21:41:19 +01:00
oobabooga 518e5c4244 Some minor fixes to the GPTQ loader 2023-03-13 16:45:08 -03:00
Ayanami Rei 8778b756e6 use updated load_quantized 2023-03-13 22:11:40 +03:00
Ayanami Rei a6a6522b6a determine model type from model name 2023-03-13 22:11:32 +03:00
Ayanami Rei b6c5c57f2e remove default value from argument 2023-03-13 22:11:08 +03:00
Alexander Hristov Hristov 63c5a139a2
Merge branch 'main' into main 2023-03-13 19:50:08 +02:00
oobabooga 7ab45fb54a
Merge pull request #296 from luiscosio/patch-1
Fix for issue #282
2023-03-13 14:45:58 -03:00
Luis Cosio 435a69e357
Fix for issue #282
RuntimeError: Tensors must have same number of dimensions: got 3 and 4
2023-03-13 11:41:35 -06:00
Ayanami Rei e1c952c41c make argument non case-sensitive 2023-03-13 20:22:38 +03:00
Ayanami Rei b746250b2f Update README 2023-03-13 20:20:45 +03:00
Ayanami Rei 3c9afd5ca3 rename method 2023-03-13 20:14:40 +03:00
Ayanami Rei 1b99ed61bc add argument --gptq-model-type and remove duplicate arguments 2023-03-13 20:01:34 +03:00
Ayanami Rei edbc61139f use new quant loader 2023-03-13 20:00:38 +03:00
Ayanami Rei 345b6dee8c refactor quant models loader and add support of OPT 2023-03-13 19:59:57 +03:00
oobabooga 66b6971b61 Update README 2023-03-13 12:44:18 -03:00
oobabooga ddea518e0f Document --auto-launch 2023-03-13 12:43:33 -03:00
oobabooga d97bfb8713
Update README.md 2023-03-13 12:39:33 -03:00
oobabooga 372363bc3d Fix GPTQ load_quant call on Windows 2023-03-13 12:07:02 -03:00
oobabooga bdff37f0bb
Update README.md 2023-03-13 11:05:51 -03:00
oobabooga b6098e9ccb
Merge pull request #275 from stefanhamburger/patch-1
Fix: tuple object does not support item assignment
2023-03-13 11:01:31 -03:00
oobabooga 72757088fa
Create FUNDING.yml 2023-03-13 10:55:00 -03:00
oobabooga 0c224cf4f4 Fix GALACTICA (#285) 2023-03-13 10:32:28 -03:00
stefanhamburger 91c2a8e88d
Fix: tuple object does not support item assignment 2023-03-13 07:42:09 +01:00
oobabooga 2c4699a7e9 Change a comment 2023-03-13 00:20:02 -03:00
oobabooga 0a7acb3bd9 Remove redundant comments 2023-03-13 00:12:21 -03:00
oobabooga 77294b27dd Use str(Path) instead of os.path.abspath(Path) 2023-03-13 00:08:01 -03:00
oobabooga b9e0712b92 Fix Open Assistant 2023-03-12 23:58:25 -03:00
oobabooga 1ddcd4d0ba Clean up silero_tts
This should only be used with --no-stream.

The shared.still_streaming implementation was faulty by design:
output_modifier should never be called when streaming is already over.
2023-03-12 23:42:49 -03:00
oobabooga a95592fc56 Add back a progress indicator to --no-stream 2023-03-12 20:38:40 -03:00
oobabooga d168b6e1f7
Update README.md 2023-03-12 17:54:07 -03:00
EliasVincent 48aa52849b use Gradio microphone input instead 2023-03-12 21:03:07 +01:00
oobabooga 54e8f0c31f
Update README.md 2023-03-12 16:58:00 -03:00
oobabooga 0a4d8a5cf6
Delete README.md 2023-03-12 16:43:06 -03:00
oobabooga 88af917e0e
Add files via upload 2023-03-12 16:42:50 -03:00
oobabooga 0b86ac38b1
Initial commit 2023-03-12 16:40:10 -03:00
HideLord 683556f411 Adding markdown support and slight refactoring. 2023-03-12 21:34:09 +02:00
oobabooga cebe8b390d Remove useless "substring_found" variable 2023-03-12 15:50:38 -03:00
oobabooga 4bcd675ccd Add *Is typing...* to regenerate as well 2023-03-12 15:23:33 -03:00
Elias Vincent Simon 3b4145966d
Merge branch 'oobabooga:main' into stt-extension 2023-03-12 19:19:43 +01:00
oobabooga 3375eaece0 Update README 2023-03-12 15:01:32 -03:00
oobabooga c7aa51faa6 Use a list of eos_tokens instead of just a number
This might be the cause of LLaMA ramblings that some people have experienced.
2023-03-12 14:54:58 -03:00
oobabooga 17210ff88f
Update README.md 2023-03-12 14:31:24 -03:00
oobabooga 441e993c51 Bump accelerate, RWKV and safetensors 2023-03-12 14:25:14 -03:00
oobabooga d8bea766d7
Merge pull request #192 from xanthousm/main
Add text generation stream status to shared module, use for better TTS with auto-play
2023-03-12 13:40:16 -03:00
oobabooga 4066ab4c0c Reorder the imports 2023-03-12 13:36:18 -03:00
oobabooga 4dc1d8c091
Update README.md 2023-03-12 12:46:53 -03:00
oobabooga 901dcba9b4
Merge pull request #263 from HideLord/main
Fixing compatibility with GPTQ repository
2023-03-12 12:42:08 -03:00
oobabooga fda376d9c3 Use os.path.abspath() instead of str() 2023-03-12 12:41:04 -03:00
HideLord 8403152257 Fixing compatibility with GPTQ repo commit 2f667f7da051967566a5fb0546f8614bcd3a1ccd. Expects string and breaks on 2023-03-12 17:28:15 +02:00
HideLord a27f98dbbc Merge branch 'main' of https://github.com/HideLord/text-generation-webui 2023-03-12 16:51:04 +02:00
oobabooga f3b00dd165
Merge pull request #224 from ItsLogic/llama-bits
Allow users to load 2, 3 and 4 bit llama models
2023-03-12 11:23:50 -03:00
oobabooga 89e9493509 Update README 2023-03-12 11:23:20 -03:00
oobabooga 65dda28c9d Rename --llama-bits to --gptq-bits 2023-03-12 11:19:07 -03:00
oobabooga fed3617f07 Move LLaMA 4-bit into a separate file 2023-03-12 11:12:34 -03:00
oobabooga 0ac562bdba Add a default prompt for OpenAssistant oasst-sft-1-pythia-12b #253 2023-03-12 10:46:16 -03:00
oobabooga 78901d522b Remove unused imports 2023-03-12 08:59:05 -03:00
oobabooga 35c14f31b2
Merge pull request #259 from hieultp/patch-1
Fix typo error in LLaMa prompts
2023-03-12 08:52:02 -03:00
oobabooga 3c25557ef0 Add tqdm to requirements.txt 2023-03-12 08:48:16 -03:00
Phuoc-Hieu Le 781c09235c
Fix typo error in script.py 2023-03-12 15:21:50 +07:00
Xan 9276af3561 clean up 2023-03-12 19:06:24 +11:00
Xan b3e10e47c0 Fix merge conflict in text_generation
- Need to update `shared.still_streaming = False` before the final `yield formatted_outputs`, shifted the position of some yields.
2023-03-12 18:56:35 +11:00
Xan d4afed4e44 Fixes and polish
- Change wav naming to be completely unique using timestamp instead of message ID, stops browser using cached audio when new audio is made with the same file name (eg after regenerate or clear history).
- Make the autoplay setting actually disable autoplay.
- Make Settings panel a bit more compact.
- Hide html errors when audio file of chat history is missing.
- Add button to permanently convert TTS history to normal text messages
- Changed the "show message text" toggle to affect the chat history.
2023-03-12 17:56:57 +11:00
oobabooga ad14f0e499 Fix regenerate (provisory way) 2023-03-12 03:42:29 -03:00
oobabooga 6e12068ba2
Merge pull request #258 from lxe/lxe/utf8
Load and save character files and chat history in UTF-8
2023-03-12 03:28:49 -03:00
oobabooga e2da6b9685 Fix You You You appearing in chat mode 2023-03-12 03:25:56 -03:00
oobabooga bcf0075278
Merge pull request #235 from xanthousm/Quality_of_life-main
--auto-launch and "Is typing..."
2023-03-12 03:12:56 -03:00
Aleksey Smolenchuk 3f7c3d6559
No need to set encoding on binary read 2023-03-11 22:10:57 -08:00
oobabooga 3437de686c
Merge pull request #189 from oobabooga/new-streaming
New streaming method (much faster)
2023-03-12 03:01:26 -03:00
oobabooga 341e135036 Various fixes in chat mode 2023-03-12 02:53:08 -03:00
Aleksey Smolenchuk 3baf5fc700
Load and save chat history in utf-8 2023-03-11 21:40:01 -08:00
oobabooga b0e8cb8c88 Various fixes in chat mode 2023-03-12 02:31:45 -03:00
unknown 433f6350bc Load and save character files in UTF-8 2023-03-11 21:23:05 -08:00
oobabooga 0bd5430988 Use 'with' statement to better handle streaming memory 2023-03-12 02:04:28 -03:00
oobabooga 37f0166b2d Fix memory leak in new streaming (second attempt) 2023-03-11 23:14:49 -03:00
HideLord def97f658c Small patch to fix loading of character jsons. Now it correctly reads non-ascii characters on Windows. 2023-03-12 02:54:22 +02:00
oobabooga 92fe947721 Merge branch 'main' into new-streaming 2023-03-11 19:59:45 -03:00
oobabooga 195e99d0b6 Add llama_prompts extension 2023-03-11 16:11:15 -03:00
oobabooga 501afbc234 Add requests to requirements.txt 2023-03-11 14:47:30 -03:00
oobabooga 8f8da6707d Minor style changes to silero_tts 2023-03-11 11:17:13 -03:00
oobabooga 2743dd736a Add *Is typing...* to impersonate as well 2023-03-11 10:50:18 -03:00
Xan 96c51973f9 --auto-launch and "Is typing..."
- Added `--auto-launch` arg to open web UI in the default browser when ready.
- Changed chat.py to display user input immediately and "*Is typing...*" as a temporary reply while generating text. Most noticeable when using `--no-stream`.
2023-03-11 22:50:59 +11:00
Xan 33df4bd91f Merge remote-tracking branch 'upstream/main' 2023-03-11 22:40:47 +11:00
Xan b8f7d34c1d Undo changes to requirements
needing to manually install tensorboard might be a windows-only problem. Can be easily solved manually.
2023-03-11 17:05:09 +11:00
Xan 0dfac4b777 Working html autoplay, clean up, improve wav naming
- New autoplay using html tag, removed from old message when new input provided
- Add voice pitch and speed control
- Group settings together
- Use name + conversation history to match wavs to messages, minimize problems when changing characters

Current minor bugs:
- Gradio seems to cache the audio files, so using "clear history" and generating new messages will play the old audio (the new messages are saving correctly). Gradio will clear cache and use correct audio after a few messages or after a page refresh.
- Switching characters does not immediately update the message ID used for the audio. ID is updated after the first new message, but that message will use the wrong ID
2023-03-11 16:34:59 +11:00
draff 28fd4fc970 Change wording to be consistent with other args 2023-03-10 23:34:13 +00:00
draff 001e638b47 Make it actually work 2023-03-10 23:28:19 +00:00
draff 804486214b Re-implement --load-in-4bit and update --llama-bits arg description 2023-03-10 23:21:01 +00:00
ItsLogic 9ba8156a70
remove unnecessary Path() 2023-03-10 22:33:58 +00:00
draff e6c631aea4 Replace --load-in-4bit with --llama-bits
Replaces --load-in-4bit with a more flexible --llama-bits arg to allow for 2 and 3 bit models as well. This commit also fixes a loading issue with .pt files which are not in the root of the models folder
2023-03-10 21:36:45 +00:00
oobabooga 026d60bd34 Remove default preset that didn't do anything 2023-03-10 14:01:02 -03:00
oobabooga e01da4097c
Merge pull request #210 from rohvani/pt-path-changes
Add llama-65b-4bit.pt support
2023-03-10 11:04:56 -03:00
oobabooga e9dbdafb14
Merge branch 'main' into pt-path-changes 2023-03-10 11:03:42 -03:00
oobabooga 706a03b2cb Minor changes 2023-03-10 11:02:25 -03:00
oobabooga de7dd8b6aa Add comments 2023-03-10 10:54:08 -03:00
oobabooga 113b791aa5
Merge pull request #219 from deepdiffuser/4bit-multigpu
add multi-gpu support for 4bit gptq LLaMA
2023-03-10 10:52:45 -03:00
oobabooga e461c0b7a0 Move the import to the top 2023-03-10 10:51:12 -03:00
deepdiffuser 9fbd60bf22 add no_split_module_classes to prevent tensor split error 2023-03-10 05:30:47 -08:00
deepdiffuser ab47044459 add multi-gpu support for 4bit gptq LLaMA 2023-03-10 04:52:45 -08:00
EliasVincent 1c0bda33fb added installation instructions 2023-03-10 11:47:16 +01:00
rohvani 2ac2913747 fix reference issue 2023-03-09 20:13:23 -08:00
oobabooga 1d7e893fa1
Merge pull request #211 from zoidbb/add-tokenizer-to-hf-downloads
download tokenizer when present
2023-03-10 00:46:21 -03:00
oobabooga 875847bf88 Consider tokenizer a type of text 2023-03-10 00:45:28 -03:00
oobabooga 8ed214001d Merge branch 'main' of github.com:oobabooga/text-generation-webui 2023-03-10 00:42:09 -03:00
oobabooga 249c268176 Fix the download script for long lists of files on HF 2023-03-10 00:41:10 -03:00
Ber Zoidberg ec3de0495c download tokenizer when present 2023-03-09 19:08:09 -08:00
rohvani 5ee376c580 add LLaMA preset 2023-03-09 18:31:41 -08:00
rohvani 826e297b0e add llama-65b-4bit support & multiple pt paths 2023-03-09 18:31:32 -08:00
oobabooga 7c3d1b43c1
Merge pull request #204 from MichealC0/patch-1
Update README.md
2023-03-09 23:04:09 -03:00
oobabooga 9849aac0f1 Don't show .pt models in the list 2023-03-09 21:54:50 -03:00
oobabooga 1a3d25f75d
Merge pull request #206 from oobabooga/llama-4bit
Add LLaMA 4-bit support
2023-03-09 21:07:32 -03:00
oobabooga eb0cb9b6df Update README 2023-03-09 20:53:52 -03:00
oobabooga 74102d5ee4 Insert to the path instead of appending 2023-03-09 20:51:22 -03:00
oobabooga 2965aa1625 Check if the .pt file exists 2023-03-09 20:48:51 -03:00
oobabooga d41e3c233b
Update README.md 2023-03-09 18:02:44 -03:00
oobabooga fd540b8930 Use new LLaMA implementation (this will break stuff. I am sorry)
https://github.com/oobabooga/text-generation-webui/wiki/LLaMA-model
2023-03-09 17:59:15 -03:00
EliasVincent a24fa781f1 tweaked Whisper parameters 2023-03-09 21:18:46 +01:00
Elias Vincent Simon d5efc0659b
Merge branch 'oobabooga:main' into stt-extension 2023-03-09 21:05:34 +01:00
EliasVincent 00359ba054 interactive preview window 2023-03-09 21:03:49 +01:00
EliasVincent 7a03d0bda3 cleanup 2023-03-09 20:33:00 +01:00
oobabooga 828a524f9a Add LLaMA 4-bit support 2023-03-09 15:50:26 -03:00
oobabooga 33414478bf Update README 2023-03-09 11:13:03 -03:00
oobabooga e7adf5fe4e Add Contrastive Search preset #197 2023-03-09 10:27:11 -03:00
oobabooga 557c773df7
Merge pull request #201 from jtang613/Name_It
Lets propose a name besides "Gradio"
2023-03-09 09:45:47 -03:00
oobabooga 038e90765b Rename to "Text generation web UI" 2023-03-09 09:44:08 -03:00
EliasVincent 4c72e43bcf first implementation 2023-03-09 12:46:50 +01:00
Chimdumebi Nebolisa 4dd14dcab4
Update README.md 2023-03-09 10:22:09 +01:00
jtang613 807a41cf87 Lets propose a name besides "Gradio" 2023-03-08 21:02:25 -05:00
Xan a2b5383398 Merge in audio generation only on text stream finish., postpone audioblock autoplay
- Keeping simpleaudio until audio block "autoplay" doesn't play previous messages
- Only generate audio for finished messages
- Better name for autoplay, clean up comments
- set default to unlimited wav files. Still a few bugs when wav id resets

Co-Authored-By: Christoph Hess <9931495+ChristophHess@users.noreply.github.com>
2023-03-09 10:48:44 +11:00
oobabooga 59b5f7a4b7 Improve usage of stopping_criteria 2023-03-08 12:13:40 -03:00
oobabooga add9330e5e Bug fixes 2023-03-08 11:26:29 -03:00
Xan 738be6dd59 Fix merge errors and unlimited wav bug 2023-03-08 22:25:55 +11:00
Xan 5648a41a27 Merge branch 'main' of https://github.com/xanthousm/text-generation-webui 2023-03-08 22:08:54 +11:00
Xan ad6b699503 Better TTS with autoplay
- Adds "still_streaming" to shared module for extensions to know if generation is complete
- Changed TTS extension with new options:
   - Show text under the audio widget
   - Automatically play the audio once text generation finishes
   - manage the generated wav files (only keep files for finished generations, optional max file limit)
   - [wip] ability to change voice pitch and speed
- added 'tensorboard' to requirements, since python sent "tensorboard not found" errors after a fresh installation.
2023-03-08 22:02:17 +11:00
oobabooga 33fb6aed74 Minor bug fix 2023-03-08 03:08:16 -03:00
oobabooga ad2970374a Readability improvements 2023-03-08 03:00:06 -03:00
oobabooga 72d539dbff Better separate the FlexGen case 2023-03-08 02:54:47 -03:00
oobabooga 0e16c0bacb Remove redeclaration of a function 2023-03-08 02:50:49 -03:00
oobabooga ab50f80542 New text streaming method (much faster) 2023-03-08 02:46:35 -03:00
oobabooga c09f416adb Change the Naive preset
(again)
2023-03-07 23:17:13 -03:00
oobabooga 8e89bc596b Fix encode() for RWKV 2023-03-07 23:15:46 -03:00
oobabooga 44e6d82185 Remove unused imports 2023-03-07 22:56:15 -03:00
oobabooga 19a34941ed Add proper streaming to RWKV 2023-03-07 18:17:56 -03:00
oobabooga 8660227e1b Add top_k to RWKV 2023-03-07 17:24:28 -03:00
oobabooga 827ae51f72
Sort the imports 2023-03-07 00:23:36 -03:00
oobabooga b4bfd87319
Update README.md 2023-03-06 20:55:01 -03:00
oobabooga 8f4a197c05
Add credits 2023-03-06 20:34:36 -03:00
oobabooga d0e8780555
Update README.md 2023-03-06 20:17:59 -03:00
oobabooga 18eaeebbfe Merge branch 'main' of github.com:oobabooga/text-generation-webui 2023-03-06 20:16:03 -03:00
oobabooga 18ccfcd7fe
Update README.md 2023-03-06 20:15:55 -03:00
oobabooga 153dfeb4dd Add --rwkv-cuda-on parameter, bump rwkv version 2023-03-06 20:12:54 -03:00
oobabooga 4143d4aa48
Merge pull request #175 from SagsMug/main
Add api example using websockets
2023-03-06 19:53:48 -03:00
oobabooga 8b882c132a
tabs -> spaces 2023-03-06 19:52:26 -03:00
oobabooga eebec65075 Improve readability 2023-03-06 19:46:46 -03:00
oobabooga 99f69dfcaa
Merge pull request #169 from MetaIX/patch-1
Support for Eleven Labs TTS
2023-03-06 19:40:31 -03:00
oobabooga 944fdc03b2 Rename the folder 2023-03-06 19:38:36 -03:00
oobabooga 49ae183ac9 Move new extension to a separate file 2023-03-06 19:28:53 -03:00
oobabooga 91823e1ed1
Update README.md 2023-03-06 16:48:31 -03:00
oobabooga 6904a507c6 Change some parameters 2023-03-06 16:29:43 -03:00
oobabooga 20bd645f6a Fix bug in multigpu setups (attempt 3) 2023-03-06 15:58:18 -03:00
oobabooga 09a7c36e1b Minor improvement while running custom models 2023-03-06 15:36:35 -03:00
oobabooga 24c4c20391 Fix bug in multigpu setups (attempt #2) 2023-03-06 15:23:29 -03:00
oobabooga d88b7836c6 Fix bug in multigpu setups 2023-03-06 14:58:30 -03:00
oobabooga 5bed607b77 Increase repetition frequency/penalty for RWKV 2023-03-06 14:25:48 -03:00
oobabooga aa7ce0665e Merge branch 'main' of github.com:oobabooga/text-generation-webui 2023-03-06 10:58:41 -03:00
oobabooga bf56b6c1fb Load settings.json without the need for --settings settings.json
This is for setting UI defaults
2023-03-06 10:57:45 -03:00
oobabooga 2de9f122cd
Update README.md 2023-03-06 09:34:49 -03:00
oobabooga e91f4bc25a Add RWKV tokenizer 2023-03-06 08:45:49 -03:00
Mug 53ce21ac68 Add api example using websockets 2023-03-06 12:13:50 +01:00
MetaIX 9907bee4a4
Support for Eleven Labs TTS
As per your suggestion at https://github.com/oobabooga/text-generation-webui/issues/159 here's my attempt.

I'm brand new to python and github. Completely different from unreal + visual coding, so forgive my amateurish code. This essentially adds support for Eleven Labs TTS. Tested it without major issues, and I believe it's functional (hopefully).

Extra requirements: elevenlabslib https://github.com/lugia19/elevenlabslib, sounddevice0.4.6, and soundfile

Folder structure is the same as the SileroTTS Extension.
2023-03-05 19:04:22 -06:00
oobabooga c855b828fe Better handle <USER> 2023-03-05 17:01:47 -03:00
oobabooga 145c725c39 Bump RWKV version 2023-03-05 16:28:21 -03:00
oobabooga 2af66a4d4c Fix <USER> in pygmalion replies 2023-03-05 16:08:50 -03:00
oobabooga a54b91af77 Improve readability 2023-03-05 10:21:15 -03:00
oobabooga 8e706df20e Fix a memory leak when text streaming is on 2023-03-05 10:12:43 -03:00
oobabooga 5492e2e9f8 Add sentencepiece 2023-03-05 10:02:24 -03:00
oobabooga 90206204aa
Merge pull request #163 from oobabooga/hf_llama
Move towards HF LLaMA implementation
2023-03-05 01:55:43 -03:00
oobabooga c33715ad5b Move towards HF LLaMA implementation 2023-03-05 01:20:31 -03:00
oobabooga bd8aac8fa4 Add LLaMA 8-bit support 2023-03-04 13:28:42 -03:00
oobabooga c93f1fa99b Count the tokens more conservatively 2023-03-04 03:10:21 -03:00
oobabooga 736f61610b Update README 2023-03-04 01:33:52 -03:00
oobabooga ed8b35efd2 Add --pin-weight parameter for FlexGen 2023-03-04 01:04:02 -03:00
oobabooga 05e703b4a4 Print the performance information more reliably 2023-03-03 21:24:32 -03:00
oobabooga 5a79863df3 Increase the sequence length, decrease batch size
I have no idea what I am doing
2023-03-03 15:54:13 -03:00
oobabooga e62b9b1074 Revamp the "Default" preset with HF defaults 2023-03-03 15:26:08 -03:00
oobabooga a345a2acd2 Add a tokenizer placeholder 2023-03-03 15:16:55 -03:00
oobabooga 4cc36dc434 Tweak the Naive preset (for LLaMA/RWKV) 2023-03-03 15:09:00 -03:00
oobabooga 5b354817f6 Make chat minimally work with LLaMA 2023-03-03 15:04:41 -03:00
oobabooga ea5c5eb3da Add LLaMA support 2023-03-03 14:39:14 -03:00
oobabooga 2bff646130 Stop chat from flashing dark when processing 2023-03-03 13:19:13 -03:00
oobabooga 7c70e0e2a6 Fix the download script (sort of) 2023-03-02 14:05:21 -03:00
oobabooga bcea196c9d Bump flexgen version 2023-03-02 12:03:57 -03:00
oobabooga 76378c6cc2 Update README 2023-03-02 11:27:15 -03:00
oobabooga 169209805d Model-aware prompts and presets 2023-03-02 11:25:04 -03:00
oobabooga 024d30d1b4 Reorder imports 2023-03-02 01:23:18 -03:00
oobabooga 7bbe32f618 Don't return a value in an iterator function 2023-03-02 00:48:46 -03:00
oobabooga ff9f649c0c Remove some unused imports 2023-03-02 00:36:20 -03:00
oobabooga 1a05860ca3 Ensure proper no-streaming with generation_attempts > 1 2023-03-02 00:10:10 -03:00
oobabooga a2a3e8f797 Add --rwkv-strategy parameter 2023-03-01 20:02:48 -03:00
oobabooga 99dc95e14e Minor aesthetic change 2023-03-01 19:32:04 -03:00
oobabooga 449116a510 Fix RWKV paths on Windows (attempt) 2023-03-01 19:17:16 -03:00
oobabooga 955cf431e8 Minor consistency fix 2023-03-01 19:11:26 -03:00
oobabooga 7a9b4407b0 Settle for 0.0.6 for now 2023-03-01 17:37:14 -03:00
oobabooga f351dce032 Keep rwkv up to date 2023-03-01 17:36:16 -03:00
oobabooga f3da6dcc8f
Merge pull request #149 from oobabooga/RWKV
Add RWKV support
2023-03-01 16:57:45 -03:00
oobabooga 831ac7ed3f Add top_p 2023-03-01 16:45:48 -03:00
oobabooga 7c4d5ca8cc Improve the text generation call a bit 2023-03-01 16:40:25 -03:00
oobabooga 2f16ce309a Rename a variable 2023-03-01 12:33:09 -03:00
oobabooga 9e9cfc4b31 Parameters 2023-03-01 12:19:37 -03:00
oobabooga 0f6708c471 Sort the imports 2023-03-01 12:18:17 -03:00
oobabooga e735806c51 Add a generate() function for RWKV 2023-03-01 12:16:11 -03:00
oobabooga 659bb76722 Add RWKVModel class 2023-03-01 12:08:55 -03:00
oobabooga 9c86a1cd4a Add RWKV pip package 2023-03-01 11:42:49 -03:00
oobabooga 6837d4d72a Load the model by name 2023-02-28 02:52:29 -03:00
oobabooga a1429d1607 Add default extensions to the settings 2023-02-28 02:20:11 -03:00
oobabooga 365e1089b3 Move some buttons 2023-02-28 01:34:07 -03:00
oobabooga c472c34818 Change Chiharu's example dialogue definition format 2023-02-28 01:13:58 -03:00
oobabooga 19ccb2aaf5 Handle <USER> and <BOT> 2023-02-28 01:05:43 -03:00
oobabooga 626da6c731 Handle {{user}} and {{char}} in example dialogue 2023-02-28 00:59:05 -03:00
oobabooga e861e68e38 Move the chat example dialogue to the prompt 2023-02-28 00:50:46 -03:00
oobabooga f871971de1 Trying to get the chat to work 2023-02-28 00:25:30 -03:00
oobabooga 67ee7bead7 Add cpu, bf16 options 2023-02-28 00:09:11 -03:00
oobabooga ebd698905c Add streaming to RWKV 2023-02-28 00:04:04 -03:00
oobabooga 70e522732c Move RWKV loader into a separate file 2023-02-27 23:50:16 -03:00
oobabooga ebc64a408c RWKV support prototype 2023-02-27 23:03:35 -03:00
oobabooga 021bd55886 Better format the prompt when generation attempts > 1 2023-02-27 21:37:03 -03:00
oobabooga f4b130e2bd
Update README.md 2023-02-27 15:15:45 -03:00
oobabooga e81c023b77 Slight style changes in gallery 2023-02-27 13:49:55 -03:00
oobabooga 43b6ab8673 Store thumbnails as files instead of base64 strings
This improves the UI responsiveness for large histories.
2023-02-27 13:41:00 -03:00
oobabooga a08802bf70 Sort the gallery alphabetically 2023-02-27 13:04:06 -03:00
oobabooga 83368875eb Minor style change 2023-02-27 11:55:36 -03:00
oobabooga f2de78da3b
Merge pull request #143 from conanak99/main
Make the char gallery more space-efficient
2023-02-27 11:48:38 -03:00
oobabooga 9c7fe2ec5d
Handle overflows in character names 2023-02-27 11:46:15 -03:00
oobabooga 611010e8af Add a confirmation to clear history 2023-02-27 11:41:21 -03:00
conanak99 b5e31c8fef fix responsive issue 2023-02-27 22:30:52 +08:00
conanak99 44fd6bce34 make the grid responsive 2023-02-27 22:26:19 +08:00
conanak99 b4fa7dcefc Make the char gallery more space-efficent 2023-02-27 21:52:43 +08:00
oobabooga b16f097466 Add FlexGen to requirements.txt 2023-02-27 08:58:07 -03:00
oobabooga f24b6e78a3 Fix clear history 2023-02-26 23:58:04 -03:00
oobabooga 7a776ccf87 Make the gallery interactive to load characters 2023-02-26 17:19:36 -03:00
oobabooga 8e3e8a070f Make FlexGen work with the newest API 2023-02-26 16:53:41 -03:00
oobabooga 48b83c9a70 Remove unused imports 2023-02-26 13:15:56 -03:00
oobabooga 3333f94c30 Make the gallery extension work on colab 2023-02-26 12:37:26 -03:00
oobabooga 756cba2edc Add character gallery extension 2023-02-26 12:17:57 -03:00
oobabooga 3c9432f039 Add checkbox to disable the character bias 2023-02-26 02:37:21 -03:00
oobabooga ef91bd802d Sort TTS voices by gender 2023-02-26 02:26:15 -03:00
oobabooga 612520386e Merge branch 'main' of github.com:oobabooga/text-generation-webui 2023-02-26 01:52:27 -03:00
oobabooga e91eb24649 Decrease the repetition penalty upper limit to 3 2023-02-26 01:51:59 -03:00
oobabooga c183d2917c
Update README.md 2023-02-26 00:59:07 -03:00
oobabooga cfe010b244
Update README.md 2023-02-26 00:54:37 -03:00
oobabooga 87d9f3e329
Update README.md 2023-02-26 00:54:19 -03:00
oobabooga 955997a90b
Update README.md 2023-02-26 00:54:07 -03:00
oobabooga 3d94ebfdd0 Change --chat colors 2023-02-26 00:51:15 -03:00
oobabooga 633a2b6be2 Don't regenerate/remove last message if the chat is empty 2023-02-26 00:43:12 -03:00
oobabooga 6e843a11d6 Fix FlexGen in chat mode 2023-02-26 00:36:04 -03:00
oobabooga 63ce2771d5 Merge branch 'main' of github.com:oobabooga/text-generation-webui 2023-02-25 22:59:46 -03:00
oobabooga 4548227fb5 Downgrade gradio version (file uploads are broken in 3.19.1) 2023-02-25 22:59:02 -03:00
oobabooga c593dfa827
Update README.md 2023-02-25 18:57:34 -03:00
oobabooga 7872a64f78
Update README.md 2023-02-25 18:56:43 -03:00
oobabooga 9456c1d6ed Prevent streaming with no_stream + generation attempts > 1 2023-02-25 17:45:03 -03:00
oobabooga 32f40f3b42 Bump gradio version to 3.19.1 2023-02-25 17:20:03 -03:00
oobabooga b3d2365d92 Rename a button 2023-02-25 16:33:46 -03:00
oobabooga fa58fd5559 Proper way to free the cuda cache 2023-02-25 15:50:29 -03:00
oobabooga 03d25c1c61 Reorder the chat buttons 2023-02-25 15:35:43 -03:00
oobabooga e2cf4e4968 Reorder the custom parameters 2023-02-25 15:21:40 -03:00
oobabooga 381f747181 Reorganize the custom parameters for mobile usage 2023-02-25 15:17:44 -03:00
oobabooga b585e382c0 Rename the custom prompt generator function 2023-02-25 15:13:14 -03:00
oobabooga 700311ce40 Empty the cuda cache at model.generate() 2023-02-25 14:39:13 -03:00
oobabooga 1878acd9f3 Minor bug fix in chat 2023-02-25 09:30:59 -03:00
oobabooga e71ff959f5 Clean up some unused code 2023-02-25 09:23:02 -03:00
oobabooga 91f5852245 Move bot_picture.py inside the extension 2023-02-25 03:00:19 -03:00
oobabooga 5ac24b019e Minor fix in the extensions implementation 2023-02-25 02:53:18 -03:00
oobabooga 85f914b9b9 Disable the hijack after using it 2023-02-25 02:36:01 -03:00
oobabooga ad83b1b828 Clean up send_pictures extension 2023-02-25 02:13:24 -03:00
oobabooga 01acb250c5 Add a comment 2023-02-25 02:07:29 -03:00
oobabooga 7e9f13e29f Rename a variable 2023-02-25 01:55:32 -03:00
oobabooga 3ef0f2ea7e
Merge pull request #123 from oobabooga/refactor_gradio
Create new extensions engine
2023-02-25 01:51:58 -03:00
oobabooga 31fb5529d8 Minor change 2023-02-25 01:49:49 -03:00
oobabooga 1741c36092 Minor fix 2023-02-25 01:47:25 -03:00
oobabooga 7c2babfe39 Rename greed to "generation attempts" 2023-02-25 01:42:19 -03:00
oobabooga 88cfc84ddb Update README 2023-02-25 01:33:26 -03:00
oobabooga 2dfb999bf1 Add greed parameter 2023-02-25 01:31:01 -03:00
oobabooga 13f2688134 Better way to generate custom prompts 2023-02-25 01:08:17 -03:00
oobabooga 67623a52b7 Allow for permanent hijacking 2023-02-25 00:55:19 -03:00
oobabooga 111b5d42e7 Add prompt hijack option for extensions 2023-02-25 00:49:18 -03:00
oobabooga fd19e4812f Add a comment 2023-02-25 00:28:38 -03:00
oobabooga 58e6734308 Add the "send picture" extension file 2023-02-25 00:26:15 -03:00
oobabooga 7a527a5581 Move "send picture" into an extension
I am not proud of how I did it for now.
2023-02-25 00:23:51 -03:00
oobabooga e51ece21c0 Add ui() function to extensions 2023-02-24 19:00:11 -03:00
oobabooga 2b4ce79b6f Fix the API 2023-02-24 17:38:15 -03:00
oobabooga 77f58e5dab Remove a space 2023-02-24 17:32:34 -03:00
oobabooga c5066f1192 Rename some variables, be consistent about ' and " 2023-02-24 17:31:23 -03:00
oobabooga 78ad55641b Remove duplicate max_new_tokens parameter 2023-02-24 17:19:42 -03:00
oobabooga 65326b545a Move all gradio elements to shared (so that extensions can use them) 2023-02-24 16:46:50 -03:00
oobabooga 0a3590da8c Add a progress bar 2023-02-24 14:19:27 -03:00
oobabooga 0817fe1beb Move code back into the chatbot wrapper 2023-02-24 14:10:32 -03:00
oobabooga fe1771157f Properly scrape huggingface for download links (for #122) 2023-02-24 14:06:42 -03:00
oobabooga 8a7563ae84 Reorder the imports 2023-02-24 12:42:43 -03:00
oobabooga ace74a557a Add some comments 2023-02-24 12:41:27 -03:00
oobabooga 3ffd7d36fd Increase the repetition penalty for pygmalion 2023-02-24 12:31:30 -03:00
oobabooga 0b90e0b3b6
Update README.md 2023-02-24 12:01:07 -03:00
oobabooga 1a23e6d185
Add Pythia to README 2023-02-24 11:38:01 -03:00
oobabooga fe5057f932 Simplify the extensions implementation 2023-02-24 10:01:21 -03:00
oobabooga e26118eba9 Merge branch 'main' of github.com:oobabooga/text-generation-webui 2023-02-24 09:41:11 -03:00
oobabooga 2fb6ae6970 Move chat preprocessing into a separate function 2023-02-24 09:40:48 -03:00
oobabooga f4f508c8e2
Update README.md 2023-02-24 09:03:09 -03:00
oobabooga 876761329b Merge branch 'elwolf6-max_memory' 2023-02-24 08:55:55 -03:00
oobabooga f6f792363b Separate command-line params by spaces instead of commas 2023-02-24 08:55:09 -03:00
oobabooga e260e84e5a Merge branch 'max_memory' of https://github.com/elwolf6/text-generation-webui into elwolf6-max_memory 2023-02-24 08:47:01 -03:00
oobabooga 146f786c57 Reorganize a bit 2023-02-24 08:44:54 -03:00
oobabooga c2f4c395b9 Clean up some chat functions 2023-02-24 08:31:30 -03:00
luis 5abdc99a7c gpu-memory arg change 2023-02-23 18:43:55 -05:00
oobabooga 9ae063e42b Fix softprompts when deepspeed is active (#112) 2023-02-23 20:22:47 -03:00
oobabooga dac6fe0ff4 Reset the history if no default history exists on reload 2023-02-23 19:53:50 -03:00
oobabooga 3b8cecbab7 Reload the default chat on page refresh 2023-02-23 19:50:23 -03:00
oobabooga f1914115d3 Fix minor issue with chat logs 2023-02-23 16:04:47 -03:00
oobabooga 682f7bdbba
Merge pull request #110 from oobabooga/refactored
Refactor everything
2023-02-23 15:30:32 -03:00
oobabooga b78561fba6 Minor bug fix 2023-02-23 15:26:41 -03:00
oobabooga 2e86a1ec04 Move chat history into shared module 2023-02-23 15:11:18 -03:00
oobabooga c87800341c Move function to extensions module 2023-02-23 14:55:21 -03:00
oobabooga 2048b403a5 Reorder functions 2023-02-23 14:49:02 -03:00
oobabooga 7224343a70 Improve the imports 2023-02-23 14:41:42 -03:00
oobabooga 364529d0c7 Further refactor 2023-02-23 14:31:28 -03:00
oobabooga e46c43afa6 Move some stuff from server.py to modules 2023-02-23 13:42:23 -03:00
oobabooga 1dacd34165 Further refactor 2023-02-23 13:28:30 -03:00
oobabooga ce7feb3641 Further refactor 2023-02-23 13:03:52 -03:00
oobabooga 98af4bfb0d Refactor the code to make it more modular 2023-02-23 12:05:25 -03:00
oobabooga 18e0ec955e Improve some descriptions in --help 2023-02-23 10:11:58 -03:00
oobabooga ced5d9ab04
Update README.md 2023-02-23 10:04:07 -03:00
oobabooga b18071330f
Update README.md 2023-02-23 01:32:05 -03:00
oobabooga c72892835a Don't show *-np models in the list of choices 2023-02-22 11:38:16 -03:00
oobabooga 95e536f876 Merge branch 'main' of github.com:oobabooga/text-generation-webui 2023-02-22 11:24:14 -03:00
oobabooga 044b963987 Add stop parameter for flexgen (#105) 2023-02-22 11:23:36 -03:00
oobabooga b4a7f5fa70
Update README.md 2023-02-22 01:54:12 -03:00
oobabooga ea21a22940 Remove redundant preset 2023-02-22 01:01:26 -03:00
oobabooga b8b3d4139c Add --compress-weight parameter 2023-02-22 00:43:21 -03:00
oobabooga 193fb1660a Conversion seems to work better this way 2023-02-22 00:35:10 -03:00
oobabooga eef6fc3cbf Add a preset for FlexGen 2023-02-21 23:33:15 -03:00
oobabooga 311404e258 Reuse disk-cache-dir parameter for flexgen 2023-02-21 22:11:05 -03:00
oobabooga f3c75bbd64 Add --percent flag for flexgen 2023-02-21 22:08:46 -03:00
oobabooga b83f51ee04 Add FlexGen support #92 (experimental) 2023-02-21 21:00:06 -03:00
oobabooga e52b697d5a Add bf16 back here (the fp16 -> bf16 conversion takes a few seconds) 2023-02-21 00:54:53 -03:00
oobabooga bc856eb962 Add some more margin 2023-02-20 20:49:21 -03:00
oobabooga 444cd69c67 Fix regex bug in loading character jsons with special characters 2023-02-20 19:38:19 -03:00
oobabooga f867285e3d Make the circle a bit less red 2023-02-20 18:41:38 -03:00
oobabooga e4440cd984 Make highlighted text gray in cai-chat mode 2023-02-20 16:43:32 -03:00
oobabooga bb1dac2f76 Convert the download option (A-Z) to upper case 2023-02-20 15:50:48 -03:00
oobabooga d7a738fb7a Load any 13b/20b/30b model in 8-bit mode when no flags are supplied 2023-02-20 15:44:10 -03:00
oobabooga c1de491c63 No need to have bf16 support here 2023-02-20 15:12:42 -03:00
oobabooga 8123c5b01a Sort the imports 2023-02-20 15:06:27 -03:00
oobabooga 77846ceef3 Minor change 2023-02-20 15:05:48 -03:00
oobabooga e195377050 Deprecate torch dumps, move to safetensors (they load even faster) 2023-02-20 15:03:19 -03:00
oobabooga 14ffa0b418 Fix line breaks in --chat mode 2023-02-20 13:25:46 -03:00
oobabooga 58520a1f75
Update README.md 2023-02-20 12:44:31 -03:00
oobabooga c8eab98fd2 Add an example of usage for the API 2023-02-20 12:39:36 -03:00
oobabooga c92f347c94
Merge pull request #90 from SillyLossy/regexp
Escape regexp in message extraction
2023-02-19 08:31:19 -03:00
SillyLossy ded890c378 Escape regexp in message extraction 2023-02-19 12:55:45 +02:00
oobabooga 8c9dd95d55
Print the softprompt metadata when it is loaded 2023-02-19 01:48:23 -03:00
oobabooga f79805f4a4
Change a comment 2023-02-18 22:58:40 -03:00
oobabooga 05e9da0c12
Update README.md 2023-02-18 22:34:51 -03:00
oobabooga b1add0e586
Update README.md 2023-02-18 22:32:16 -03:00
oobabooga 55bb5e5ef0 Bump accelerate version 2023-02-18 22:15:47 -03:00
oobabooga 995bcfcf5e Minor style change 2023-02-18 22:14:57 -03:00
oobabooga d58544a420 Some minor formatting changes 2023-02-18 11:07:55 -03:00
oobabooga 3e6a8ccdce Fix galactica latex css 2023-02-18 00:18:39 -03:00
oobabooga 14f49bbe9a Fix galactica equations in dark mode 2023-02-17 23:57:09 -03:00
oobabooga abb4667b44 Improve basic HTML style 2023-02-17 23:08:34 -03:00
oobabooga 00ca17abc9 Minor change 2023-02-17 22:52:03 -03:00
oobabooga 2fd003c044 Fix gpt4chan styles that were broken by gradio 3.18.0 2023-02-17 22:47:41 -03:00
oobabooga 0dd41e4830 Reorganize the sliders some more 2023-02-17 16:33:27 -03:00
oobabooga 6b9ac2f88e Reorganize the generation parameters 2023-02-17 16:18:01 -03:00
oobabooga 3923ac967f Create a cache for profile pictures (in RAM)
This is a performance optimization.
2023-02-17 14:30:39 -03:00
oobabooga 596732a981 The soft prompt length must be considered here too 2023-02-17 12:35:30 -03:00
oobabooga a6ddbbfc77 Add more fonts options 2023-02-17 11:30:04 -03:00
oobabooga 5eeb3f4e54 Make thumbnails for the profile pictures (for performance) 2023-02-17 10:58:54 -03:00
oobabooga edc0262889 Minor file uploading fixes 2023-02-17 10:27:41 -03:00
oobabooga 243244eeec Attempt at fixing greyed out files on iphone 2023-02-17 10:17:15 -03:00
oobabooga a226f4cddb No change, so reverting 2023-02-17 09:27:17 -03:00
oobabooga 40cb9f63f6 Try making Colab happy (tensorflow warnings) 2023-02-17 09:23:11 -03:00
oobabooga 71c2764516 Fix the API docs in chat mode 2023-02-17 01:56:51 -03:00
oobabooga 33ad21c4f2 Make the profile pictures a bit larger 2023-02-17 00:35:17 -03:00
oobabooga c4e87c109e Include the bot's image as base64
This is needed for Colab.
2023-02-17 00:24:27 -03:00
oobabooga cb226247e8 Make it possible to disable the TTS from within the interface 2023-02-16 23:38:27 -03:00
oobabooga fd8070b960 Give some default options in the download script 2023-02-16 23:04:13 -03:00
oobabooga aeddf902ec Make the refresh button prettier 2023-02-16 21:55:20 -03:00
oobabooga 21512e2790 Make the Stop button work more reliably 2023-02-16 21:21:45 -03:00
oobabooga 348acdf626 Mention deepspeed in the README 2023-02-16 17:29:48 -03:00
oobabooga bde4cd402a Change the default TTS voice 2023-02-16 16:07:38 -03:00
oobabooga 5fb99371ba Add .gitignore 2023-02-16 13:35:54 -03:00
oobabooga 08805b3374 Force "You" in impersonate too 2023-02-16 13:24:13 -03:00
oobabooga d7db04403f Fix --chat chatbox height 2023-02-16 12:45:05 -03:00
oobabooga 589069e105 Don't regenerate if no message has been sent 2023-02-16 12:32:35 -03:00
oobabooga 6160a03984 Merge branch 'main' of github.com:oobabooga/text-generation-webui 2023-02-16 12:16:26 -03:00
oobabooga 405dfbf57c Force your name to be "You" for pygmalion (properly) 2023-02-16 12:16:12 -03:00
oobabooga 20484f26f3
Trying to make character bias more consistent 2023-02-15 23:38:52 -03:00
oobabooga 7bd2ae05bf Force your name to be "You" for pygmalion
This allows you to customize your displayed name.
2023-02-15 21:32:53 -03:00
oobabooga 3746d72853 More style fixes 2023-02-15 21:13:12 -03:00
oobabooga 6f213b8c14 Style fix 2023-02-15 20:58:17 -03:00
oobabooga ccf10db60f Move stuff into tabs in chat mode 2023-02-15 20:55:32 -03:00
oobabooga a55e8836f6 Bump gradio version
It looks uglier, but the old one was bugged and unstable.
2023-02-15 20:20:56 -03:00
oobabooga 0e89ff4b13 Clear the persistent history after clicking on "Clear history" 2023-02-15 16:49:52 -03:00
oobabooga 05b53e4626 Update README 2023-02-15 14:43:34 -03:00
oobabooga ed73d00bd5 Update README 2023-02-15 14:43:13 -03:00
oobabooga 30fcb26737 Update README 2023-02-15 14:42:41 -03:00
oobabooga b3bcd2881d Implement regenerate/impersonate the proper way (fixes #78) 2023-02-15 14:39:26 -03:00
oobabooga 5ee9283cae Mention BLIP 2023-02-15 13:53:38 -03:00
oobabooga 8d3b3959e7 Document --picture option 2023-02-15 13:50:18 -03:00
oobabooga 2eea0f4edb Minor change 2023-02-15 12:58:11 -03:00
oobabooga 3c31fa7079 Simplifications 2023-02-15 12:46:11 -03:00
oobabooga 80fbc584f7 Readability 2023-02-15 11:38:44 -03:00
oobabooga b397bea387 Make chat history persistent 2023-02-15 11:30:38 -03:00
oobabooga 7be372829d Set chat prompt size in tokens 2023-02-15 10:18:50 -03:00
oobabooga 1622059179 Move BLIP to the CPU
It's just as fast
2023-02-15 00:03:19 -03:00
oobabooga d4d90a8000
Merge pull request #76 from SillyLossy/main
Use BLIP to send a picture to model
2023-02-14 23:57:44 -03:00
oobabooga 8c3ef58e00 Use BLIP directly + some simplifications 2023-02-14 23:55:46 -03:00
SillyLossy a7d98f494a Use BLIP to send a picture to model 2023-02-15 01:38:21 +02:00
oobabooga 79d3a524f2 Add a file 2023-02-14 15:18:05 -03:00
oobabooga f6bf74dcd5 Add Silero TTS extension 2023-02-14 15:06:06 -03:00
oobabooga 01e5772302
Update README.md 2023-02-14 13:06:26 -03:00
oobabooga d910d435cd Consider the softprompt in the maximum prompt length calculation 2023-02-14 12:06:47 -03:00
oobabooga 8b3bb512ef Minor bug fix (soft prompt was being loaded twice) 2023-02-13 23:34:04 -03:00
oobabooga 56bbc996a4 Minor CSS change for readability 2023-02-13 23:01:14 -03:00
oobabooga 210c918199
Update README.md 2023-02-13 21:49:19 -03:00
oobabooga 2fe9d7f372 Merge branch 'main' of github.com:oobabooga/text-generation-webui 2023-02-13 18:48:46 -03:00
oobabooga 7739a29524 Some simplifications 2023-02-13 18:48:32 -03:00
oobabooga b7ddcab53a
Update README.md 2023-02-13 15:52:49 -03:00
oobabooga 3277b751f5 Add softprompt support (for real this time)
Is this too much voodoo for our purposes?
2023-02-13 15:25:16 -03:00
oobabooga aa1177ff15 Send last internal reply to input rather than visible 2023-02-13 03:29:23 -03:00
oobabooga 61aed97439 Slightly increase a margin 2023-02-12 17:38:54 -03:00
oobabooga 2c3abcf57a Add support for rosey/chip/joi instruct models 2023-02-12 09:46:34 -03:00
oobabooga 7ef7bba6e6 Add progress bar for model loading 2023-02-12 09:36:27 -03:00
oobabooga 939e9d00a2
Update README.md 2023-02-12 00:47:03 -03:00
oobabooga bf9dd8f8ee Add --text-only option to the download script 2023-02-12 00:42:56 -03:00
oobabooga 42cc307409
Update README.md 2023-02-12 00:34:55 -03:00
oobabooga 66862203fc Only download safetensors if both pytorch and safetensors are present 2023-02-12 00:06:22 -03:00
oobabooga 5d3f15b915 Use the CPU if no GPU is detected 2023-02-11 23:17:06 -03:00
oobabooga 337290777b Rename example extension to "softprompt" 2023-02-11 17:17:10 -03:00
oobabooga b3c4657c47 Remove commas from preset files 2023-02-11 14:54:29 -03:00
oobabooga 144857acfe Update README 2023-02-11 14:49:11 -03:00
oobabooga 0dd1409f24 Add penalty_alpha parameter (contrastive search) 2023-02-11 14:48:12 -03:00
oobabooga 8aafb55693
1-click installer now also works for AMD GPUs
(I think)
2023-02-11 14:24:47 -03:00
oobabooga 7eed553337 Merge branch 'main' of github.com:oobabooga/text-generation-webui 2023-02-11 08:00:29 -03:00
oobabooga 2ed0386d87 Fix replace last reply in --chat mode (for #69) 2023-02-11 07:59:54 -03:00
oobabooga 1e97cb9570
Merge pull request #68 from Spencer-Dawson/patch-1
Added ROCm Install instructions to README
2023-02-11 07:56:30 -03:00
oobabooga 1176d64b13
Update README.md 2023-02-11 07:56:12 -03:00
Spencer-Dawson c5324d653b
re-added missed README changes 2023-02-11 00:13:06 -07:00
oobabooga cf89ef1c74
Update README.md 2023-02-10 21:46:29 -03:00
oobabooga 8782ac1911
Update README.md 2023-02-10 17:10:27 -03:00
oobabooga 7d7cc37560
Add Linux 1-click installer 2023-02-10 17:09:53 -03:00
425 changed files with 49987 additions and 1546 deletions

View file

@ -0,0 +1,53 @@
name: "Bug report"
description: Report a bug
labels: [ "bug" ]
body:
- type: markdown
attributes:
value: |
Thanks for taking the time to fill out this bug report!
- type: textarea
id: bug-description
attributes:
label: Describe the bug
description: A clear and concise description of what the bug is.
placeholder: Bug description
validations:
required: true
- type: checkboxes
attributes:
label: Is there an existing issue for this?
description: Please search to see if an issue already exists for the issue you encountered.
options:
- label: I have searched the existing issues
required: true
- type: textarea
id: reproduction
attributes:
label: Reproduction
description: Please provide the steps necessary to reproduce your issue.
placeholder: Reproduction
validations:
required: true
- type: textarea
id: screenshot
attributes:
label: Screenshot
description: "If possible, please include screenshot(s) so that we can understand what the issue is."
- type: textarea
id: logs
attributes:
label: Logs
description: "Please include the full stacktrace of the errors you get in the command-line (if any)."
render: shell
validations:
required: true
- type: textarea
id: system-info
attributes:
label: System Info
description: "Please share your operating system and GPU type (NVIDIA/AMD/Intel/Apple). If you are using a Google Colab notebook, mention that instead."
render: shell
placeholder:
validations:
required: true

View file

@ -0,0 +1,16 @@
---
name: Feature request
about: Suggest an improvement or new feature for the web UI
title: ''
labels: 'enhancement'
assignees: ''
---
**Description**
A clear and concise description of what you want to be implemented.
**Additional Context**
If applicable, please provide any extra information, external links, or screenshots that could be useful.

14
.github/dependabot.yml vendored Normal file
View file

@ -0,0 +1,14 @@
# To get started with Dependabot version updates, you'll need to specify which
# package ecosystems to update and where the package manifests are located.
# Please see the documentation for all configuration options:
# https://docs.github.com/github/administering-a-repository/configuration-options-for-dependency-updates
version: 2
updates:
- package-ecosystem: "pip"
directories:
- "/requirements/full/"
- "/requirements/portable/"
target-branch: "dev"
schedule:
interval: "weekly"

3
.github/pull_request_template.md vendored Normal file
View file

@ -0,0 +1,3 @@
## Checklist:
- [ ] I have read the [Contributing guidelines](https://github.com/oobabooga/text-generation-webui/wiki/Contributing-guidelines).

View file

@ -0,0 +1,70 @@
name: Build Everything TGW
on:
workflow_dispatch:
inputs:
version:
description: 'Version tag of text-generation-webui to build: v3.0'
default: 'v3.0'
required: true
type: string
permissions:
contents: write
jobs:
build_release_cuda_windows:
name: CUDA Windows
uses: ./.github/workflows/build-portable-release-cuda.yml
with:
version: ${{ inputs.version }}
config: 'os:windows-2022'
build_release_cuda_linux:
name: CUDA Linux
uses: ./.github/workflows/build-portable-release-cuda.yml
with:
version: ${{ inputs.version }}
config: 'os:ubuntu-22.04'
build_release_vulkan_windows:
name: Vulkan Windows
uses: ./.github/workflows/build-portable-release-vulkan.yml
with:
version: ${{ inputs.version }}
config: 'os:windows-2022'
build_release_vulkan_linux:
name: Vulkan Linux
uses: ./.github/workflows/build-portable-release-vulkan.yml
with:
version: ${{ inputs.version }}
config: 'os:ubuntu-22.04'
build_release_rocm_linux:
name: ROCm Linux
uses: ./.github/workflows/build-portable-release-rocm.yml
with:
version: ${{ inputs.version }}
config: 'os:ubuntu-22.04'
build_release_cpu_windows:
name: CPU Windows
uses: ./.github/workflows/build-portable-release.yml
with:
version: ${{ inputs.version }}
config: 'os:windows-2022'
build_release_cpu_linux:
name: CPU Linux
uses: ./.github/workflows/build-portable-release.yml
with:
version: ${{ inputs.version }}
config: 'os:ubuntu-22.04'
build_release_macos:
name: macOS
uses: ./.github/workflows/build-portable-release.yml
with:
version: ${{ inputs.version }}
config: 'os:macos-13,macos-14'

View file

@ -0,0 +1,180 @@
name: Build CUDA
on:
workflow_dispatch:
inputs:
version:
description: 'Version tag of text-generation-webui to build: v3.0'
default: 'v3.0'
required: true
type: string
config:
description: 'Override configurations to build: key1:item1-1,item1-2;key2:item2-1,item2-2'
default: 'Default'
required: false
type: string
exclude:
description: 'Exclude build configurations: key1-1:item1-1,key1-2:item1-2;key2-1:item2-1,key2-2:item2-2'
default: 'None'
required: false
type: string
workflow_call:
inputs:
version:
description: 'Version tag of text-generation-webui to build: v3.0'
default: 'v3.0'
required: true
type: string
config:
description: 'Configurations to build: key1:item1-1,item1-2;key2:item2-1,item2-2'
default: 'Default'
required: false
type: string
exclude:
description: 'Exclude build configurations: key1-1:item1-1,key1-2:item1-2;key2-1:item2-1,key2-2:item2-2'
default: 'None'
required: false
type: string
permissions:
contents: write
jobs:
define_matrix:
name: Define Build Matrix
runs-on: ubuntu-latest
outputs:
matrix: ${{ steps.set-matrix.outputs.matrix }}
defaults:
run:
shell: pwsh
env:
CONFIGIN: ${{ inputs.config }}
EXCLUDEIN: ${{ inputs.exclude }}
steps:
- name: Define Job Output
id: set-matrix
run: |
$matrix = @{
'os' = @('ubuntu-22.04', 'windows-2022')
'pyver' = @("3.11")
'avx' = @("AVX2")
'cuda' = @("12.4")
}
if ($env:CONFIGIN -ne 'Default') {$env:CONFIGIN.split(';').foreach({$matrix[$_.split(':')[0]] = $_.split(':')[1].split(',')})}
if ($env:EXCLUDEIN -ne 'None') {
$exclusions = @()
$exclusions += $env:EXCLUDEIN.split(';').replace(':','=').replace(',',"`n") | ConvertFrom-StringData
$matrix['exclude'] = $exclusions
}
$matrixOut = ConvertTo-Json $matrix -Compress
Write-Output ('matrix=' + $matrixOut) >> $env:GITHUB_OUTPUT
build_wheels:
name: ${{ matrix.os }} ${{ matrix.pyver }} CPU ${{ matrix.avx }} CUDA ${{ matrix.cuda }}
needs: define_matrix
runs-on: ${{ matrix.os }}
strategy:
matrix: ${{ fromJSON(needs.define_matrix.outputs.matrix) }}
defaults:
run:
shell: pwsh
env:
AVXVER: ${{ matrix.avx }}
PCKGVER: ${{ inputs.version }}
steps:
- uses: actions/checkout@v4
with:
repository: 'oobabooga/text-generation-webui'
ref: ${{ inputs.version }}
submodules: 'recursive'
- uses: actions/setup-python@v4
with:
python-version: ${{ matrix.pyver }}
- name: Build Package
shell: bash
run: |
VERSION_CLEAN="${{ inputs.version }}"
VERSION_CLEAN="${VERSION_CLEAN#v}"
cd ..
cp -r text-generation-webui "text-generation-webui-${VERSION_CLEAN}"
cd "text-generation-webui-${VERSION_CLEAN}"
# Remove extensions that need additional requirements
allowed=("character_bias" "gallery" "openai" "sd_api_pictures")
find extensions/ -mindepth 1 -maxdepth 1 -type d | grep -v -E "$(printf '%s|' "${allowed[@]}" | sed 's/|$//')" | xargs rm -rf
# Define common variables
CUDA_VERSION="${{ matrix.cuda }}"
AVX_SUPPORT="${{ matrix.avx }}"
VERSION="${{ inputs.version }}"
# 1. Set platform-specific variables
if [[ "$RUNNER_OS" == "Windows" ]]; then
PLATFORM="windows"
PYTHON_URL="https://github.com/astral-sh/python-build-standalone/releases/download/20250409/cpython-3.11.12+20250409-x86_64-pc-windows-msvc-install_only.tar.gz"
PIP_PATH="portable_env/python.exe -m pip"
PACKAGES_PATH="portable_env/Lib/site-packages"
rm start_linux.sh start_macos.sh
else
PLATFORM="linux"
PYTHON_URL="https://github.com/astral-sh/python-build-standalone/releases/download/20250409/cpython-3.11.12+20250409-x86_64-unknown-linux-gnu-install_only.tar.gz"
PIP_PATH="portable_env/bin/python -m pip"
PACKAGES_PATH="portable_env/lib/python3.11/site-packages"
rm start_macos.sh start_windows.bat
fi
# 2. Download and extract Python
cd ..
echo "Downloading Python for $PLATFORM..."
curl -L -o python-build.tar.gz "$PYTHON_URL"
tar -xzf python-build.tar.gz
mv python "text-generation-webui-${VERSION_CLEAN}/portable_env"
# 3. Prepare requirements file based on AVX and CUDA
if [[ "$AVX_SUPPORT" == "AVX2" ]]; then
BASE_REQ_FILE="requirements/portable/requirements.txt"
else
BASE_REQ_FILE="requirements/portable/requirements_noavx2.txt"
fi
# Create CUDA-specific requirements file if needed
cd "text-generation-webui-${VERSION_CLEAN}"
REQ_FILE="$BASE_REQ_FILE"
# 4. Install packages
echo "Installing Python packages from $REQ_FILE..."
$PIP_PATH install --target="./$PACKAGES_PATH" -r "$REQ_FILE"
# 5. Clean up
rm -rf .git cmd* update_wizard* Colab-TextGen-GPU.ipynb docker setup.cfg .github .gitignore requirements/ one_click.py
# 6. Create ZIP file
cd ..
ZIP_NAME="textgen-portable-${VERSION_CLEAN}-${PLATFORM}-cuda${CUDA_VERSION}.zip"
echo "Creating archive: $ZIP_NAME"
if [[ "$RUNNER_OS" == "Windows" ]]; then
powershell -Command "Compress-Archive -Path text-generation-webui-${VERSION_CLEAN} -DestinationPath $ZIP_NAME"
else
zip -r "$ZIP_NAME" "text-generation-webui-${VERSION_CLEAN}"
fi
- name: Upload files to a GitHub release
id: upload-release
uses: svenstaro/upload-release-action@2.7.0
continue-on-error: true
with:
repo_token: ${{ secrets.GITHUB_TOKEN }}
file: ../textgen-portable-*.zip
tag: ${{ inputs.version }}
file_glob: true
make_latest: false
overwrite: true

View file

@ -0,0 +1,165 @@
name: Build ROCm
on:
workflow_dispatch:
inputs:
version:
description: 'Version tag of text-generation-webui to build: v3.0'
default: 'v3.0'
required: true
type: string
config:
description: 'Override configurations to build: key1:item1-1,item1-2;key2:item2-1,item2-2'
default: 'Default'
required: false
type: string
exclude:
description: 'Exclude build configurations: key1-1:item1-1,key1-2:item1-2;key2-1:item2-1,key2-2:item2-2'
default: 'None'
required: false
type: string
workflow_call:
inputs:
version:
description: 'Version tag of text-generation-webui to build: v3.0'
default: 'v3.0'
required: true
type: string
config:
description: 'Configurations to build: key1:item1-1,item1-2;key2:item2-1,item2-2'
default: 'Default'
required: false
type: string
exclude:
description: 'Exclude build configurations: key1-1:item1-1,key1-2:item1-2;key2-1:item2-1,key2-2:item2-2'
default: 'None'
required: false
type: string
permissions:
contents: write
jobs:
define_matrix:
name: Define Build Matrix
runs-on: ubuntu-latest
outputs:
matrix: ${{ steps.set-matrix.outputs.matrix }}
defaults:
run:
shell: pwsh
env:
CONFIGIN: ${{ inputs.config }}
EXCLUDEIN: ${{ inputs.exclude }}
steps:
- name: Define Job Output
id: set-matrix
run: |
$matrix = @{
'os' = @('ubuntu-22.04')
'pyver' = @("3.11")
'avx' = @("AVX2")
}
if ($env:CONFIGIN -ne 'Default') {$env:CONFIGIN.split(';').foreach({$matrix[$_.split(':')[0]] = $_.split(':')[1].split(',')})}
if ($env:EXCLUDEIN -ne 'None') {
$exclusions = @()
$exclusions += $env:EXCLUDEIN.split(';').replace(':','=').replace(',',"`n") | ConvertFrom-StringData
$matrix['exclude'] = $exclusions
}
$matrixOut = ConvertTo-Json $matrix -Compress
Write-Output ('matrix=' + $matrixOut) >> $env:GITHUB_OUTPUT
build_wheels:
name: ${{ matrix.os }} ${{ matrix.pyver }} CPU ${{ matrix.avx }}
needs: define_matrix
runs-on: ${{ matrix.os }}
strategy:
matrix: ${{ fromJSON(needs.define_matrix.outputs.matrix) }}
defaults:
run:
shell: pwsh
env:
AVXVER: ${{ matrix.avx }}
PCKGVER: ${{ inputs.version }}
steps:
- uses: actions/checkout@v4
with:
repository: 'oobabooga/text-generation-webui'
ref: ${{ inputs.version }}
submodules: 'recursive'
- uses: actions/setup-python@v4
with:
python-version: ${{ matrix.pyver }}
- name: Build Package
shell: bash
run: |
VERSION_CLEAN="${{ inputs.version }}"
VERSION_CLEAN="${VERSION_CLEAN#v}"
cd ..
cp -r text-generation-webui "text-generation-webui-${VERSION_CLEAN}"
cd "text-generation-webui-${VERSION_CLEAN}"
# Remove extensions that need additional requirements
allowed=("character_bias" "gallery" "openai" "sd_api_pictures")
find extensions/ -mindepth 1 -maxdepth 1 -type d | grep -v -E "$(printf '%s|' "${allowed[@]}" | sed 's/|$//')" | xargs rm -rf
# Define common variables
AVX_SUPPORT="${{ matrix.avx }}"
VERSION="${{ inputs.version }}"
# 1. Set platform-specific variables (Linux only for ROCm)
PLATFORM="linux"
PYTHON_URL="https://github.com/astral-sh/python-build-standalone/releases/download/20250409/cpython-3.11.12+20250409-x86_64-unknown-linux-gnu-install_only.tar.gz"
PIP_PATH="portable_env/bin/python -m pip"
PACKAGES_PATH="portable_env/lib/python3.11/site-packages"
rm start_macos.sh start_windows.bat
# 2. Download and extract Python
cd ..
echo "Downloading Python for $PLATFORM..."
curl -L -o python-build.tar.gz "$PYTHON_URL"
tar -xzf python-build.tar.gz
mv python "text-generation-webui-${VERSION_CLEAN}/portable_env"
# 3. Prepare requirements file based on AVX
if [[ "$AVX_SUPPORT" == "AVX2" ]]; then
BASE_REQ_FILE="requirements/portable/requirements_amd.txt"
else
BASE_REQ_FILE="requirements/portable/requirements_amd_noavx2.txt"
fi
REQ_FILE="$BASE_REQ_FILE"
cd "text-generation-webui-${VERSION_CLEAN}"
# 4. Install packages
echo "Installing Python packages from $REQ_FILE..."
$PIP_PATH install --target="./$PACKAGES_PATH" -r "$REQ_FILE"
# 5. Clean up
rm -rf .git cmd* update_wizard* Colab-TextGen-GPU.ipynb docker setup.cfg .github .gitignore requirements/ one_click.py
# 6. Create ZIP file
cd ..
ZIP_NAME="textgen-portable-${VERSION_CLEAN}-${PLATFORM}-rocm.zip"
echo "Creating archive: $ZIP_NAME"
zip -r "$ZIP_NAME" "text-generation-webui-${VERSION_CLEAN}"
- name: Upload files to a GitHub release
id: upload-release
uses: svenstaro/upload-release-action@2.7.0
continue-on-error: true
with:
repo_token: ${{ secrets.GITHUB_TOKEN }}
file: ../textgen-portable-*.zip
tag: ${{ inputs.version }}
file_glob: true
make_latest: false
overwrite: true

View file

@ -0,0 +1,177 @@
name: Build Vulkan
on:
workflow_dispatch:
inputs:
version:
description: 'Version tag of text-generation-webui to build: v3.0'
default: 'v3.0'
required: true
type: string
config:
description: 'Override configurations to build: key1:item1-1,item1-2;key2:item2-1,item2-2'
default: 'Default'
required: false
type: string
exclude:
description: 'Exclude build configurations: key1-1:item1-1,key1-2:item1-2;key2-1:item2-1,key2-2:item2-2'
default: 'None'
required: false
type: string
workflow_call:
inputs:
version:
description: 'Version tag of text-generation-webui to build: v3.0'
default: 'v3.0'
required: true
type: string
config:
description: 'Configurations to build: key1:item1-1,item1-2;key2:item2-1,item2-2'
default: 'Default'
required: false
type: string
exclude:
description: 'Exclude build configurations: key1-1:item1-1,key1-2:item1-2;key2-1:item2-1,key2-2:item2-2'
default: 'None'
required: false
type: string
permissions:
contents: write
jobs:
define_matrix:
name: Define Build Matrix
runs-on: ubuntu-latest
outputs:
matrix: ${{ steps.set-matrix.outputs.matrix }}
defaults:
run:
shell: pwsh
env:
CONFIGIN: ${{ inputs.config }}
EXCLUDEIN: ${{ inputs.exclude }}
steps:
- name: Define Job Output
id: set-matrix
run: |
$matrix = @{
'os' = @('ubuntu-22.04', 'windows-2022')
'pyver' = @("3.11")
'avx' = @("AVX2")
}
if ($env:CONFIGIN -ne 'Default') {$env:CONFIGIN.split(';').foreach({$matrix[$_.split(':')[0]] = $_.split(':')[1].split(',')})}
if ($env:EXCLUDEIN -ne 'None') {
$exclusions = @()
$exclusions += $env:EXCLUDEIN.split(';').replace(':','=').replace(',',"`n") | ConvertFrom-StringData
$matrix['exclude'] = $exclusions
}
$matrixOut = ConvertTo-Json $matrix -Compress
Write-Output ('matrix=' + $matrixOut) >> $env:GITHUB_OUTPUT
build_wheels:
name: ${{ matrix.os }} ${{ matrix.pyver }} CPU ${{ matrix.avx }}
needs: define_matrix
runs-on: ${{ matrix.os }}
strategy:
matrix: ${{ fromJSON(needs.define_matrix.outputs.matrix) }}
defaults:
run:
shell: pwsh
env:
AVXVER: ${{ matrix.avx }}
PCKGVER: ${{ inputs.version }}
steps:
- uses: actions/checkout@v4
with:
repository: 'oobabooga/text-generation-webui'
ref: ${{ inputs.version }}
submodules: 'recursive'
- uses: actions/setup-python@v4
with:
python-version: ${{ matrix.pyver }}
- name: Build Package
shell: bash
run: |
VERSION_CLEAN="${{ inputs.version }}"
VERSION_CLEAN="${VERSION_CLEAN#v}"
cd ..
cp -r text-generation-webui "text-generation-webui-${VERSION_CLEAN}"
cd "text-generation-webui-${VERSION_CLEAN}"
# Remove extensions that need additional requirements
allowed=("character_bias" "gallery" "openai" "sd_api_pictures")
find extensions/ -mindepth 1 -maxdepth 1 -type d | grep -v -E "$(printf '%s|' "${allowed[@]}" | sed 's/|$//')" | xargs rm -rf
# Define common variables
AVX_SUPPORT="${{ matrix.avx }}"
VERSION="${{ inputs.version }}"
# 1. Set platform-specific variables
if [[ "$RUNNER_OS" == "Windows" ]]; then
PLATFORM="windows"
PYTHON_URL="https://github.com/astral-sh/python-build-standalone/releases/download/20250409/cpython-3.11.12+20250409-x86_64-pc-windows-msvc-install_only.tar.gz"
PIP_PATH="portable_env/python.exe -m pip"
PACKAGES_PATH="portable_env/Lib/site-packages"
rm start_linux.sh start_macos.sh
else
PLATFORM="linux"
PYTHON_URL="https://github.com/astral-sh/python-build-standalone/releases/download/20250409/cpython-3.11.12+20250409-x86_64-unknown-linux-gnu-install_only.tar.gz"
PIP_PATH="portable_env/bin/python -m pip"
PACKAGES_PATH="portable_env/lib/python3.11/site-packages"
rm start_macos.sh start_windows.bat
fi
# 2. Download and extract Python
cd ..
echo "Downloading Python for $PLATFORM..."
curl -L -o python-build.tar.gz "$PYTHON_URL"
tar -xzf python-build.tar.gz
mv python "text-generation-webui-${VERSION_CLEAN}/portable_env"
# 3. Prepare requirements file based on AVX
if [[ "$AVX_SUPPORT" == "AVX2" ]]; then
BASE_REQ_FILE="requirements/portable/requirements_vulkan.txt"
else
BASE_REQ_FILE="requirements/portable/requirements_vulkan_noavx2.txt"
fi
REQ_FILE="$BASE_REQ_FILE"
cd "text-generation-webui-${VERSION_CLEAN}"
# 4. Install packages
echo "Installing Python packages from $REQ_FILE..."
$PIP_PATH install --target="./$PACKAGES_PATH" -r "$REQ_FILE"
# 5. Clean up
rm -rf .git cmd* update_wizard* Colab-TextGen-GPU.ipynb docker setup.cfg .github .gitignore requirements/ one_click.py
# 6. Create ZIP file
cd ..
ZIP_NAME="textgen-portable-${VERSION_CLEAN}-${PLATFORM}-vulkan.zip"
echo "Creating archive: $ZIP_NAME"
if [[ "$RUNNER_OS" == "Windows" ]]; then
powershell -Command "Compress-Archive -Path text-generation-webui-${VERSION_CLEAN} -DestinationPath $ZIP_NAME"
else
zip -r "$ZIP_NAME" "text-generation-webui-${VERSION_CLEAN}"
fi
- name: Upload files to a GitHub release
id: upload-release
uses: svenstaro/upload-release-action@2.7.0
continue-on-error: true
with:
repo_token: ${{ secrets.GITHUB_TOKEN }}
file: ../textgen-portable-*.zip
tag: ${{ inputs.version }}
file_glob: true
make_latest: false
overwrite: true

View file

@ -0,0 +1,203 @@
name: Build CPU and macOS
on:
workflow_dispatch:
inputs:
version:
description: 'Version tag of text-generation-webui to build: v3.0'
default: 'v3.0'
required: true
type: string
config:
description: 'Override configurations to build: key1:item1-1,item1-2;key2:item2-1,item2-2'
default: 'Default'
required: false
type: string
exclude:
description: 'Exclude build configurations: key1-1:item1-1,key1-2:item1-2;key2-1:item2-1,key2-2:item2-2'
default: 'None'
required: false
type: string
workflow_call:
inputs:
version:
description: 'Version tag of text-generation-webui to build: v3.0'
default: 'v3.0'
required: true
type: string
config:
description: 'Configurations to build: key1:item1-1,item1-2;key2:item2-1,item2-2'
default: 'Default'
required: false
type: string
exclude:
description: 'Exclude build configurations: key1-1:item1-1,key1-2:item1-2;key2-1:item2-1,key2-2:item2-2'
default: 'None'
required: false
type: string
permissions:
contents: write
jobs:
define_matrix:
name: Define Build Matrix
runs-on: ubuntu-latest
outputs:
matrix: ${{ steps.set-matrix.outputs.matrix }}
defaults:
run:
shell: pwsh
env:
CONFIGIN: ${{ inputs.config }}
EXCLUDEIN: ${{ inputs.exclude }}
steps:
- name: Define Job Output
id: set-matrix
run: |
$matrix = @{
'os' = @('ubuntu-22.04', 'windows-2022', 'macos-14')
'pyver' = @("3.11")
'avx' = @("AVX2")
}
if ($env:CONFIGIN -ne 'Default') {$env:CONFIGIN.split(';').foreach({$matrix[$_.split(':')[0]] = $_.split(':')[1].split(',')})}
if ($env:EXCLUDEIN -ne 'None') {
$exclusions = @()
$exclusions += $env:EXCLUDEIN.split(';').replace(':','=').replace(',',"`n") | ConvertFrom-StringData
$matrix['exclude'] = $exclusions
}
$matrixOut = ConvertTo-Json $matrix -Compress
Write-Output ('matrix=' + $matrixOut) >> $env:GITHUB_OUTPUT
build_wheels:
name: ${{ matrix.os }} ${{ matrix.pyver }} CPU ${{ matrix.avx }}
needs: define_matrix
runs-on: ${{ matrix.os }}
strategy:
matrix: ${{ fromJSON(needs.define_matrix.outputs.matrix) }}
defaults:
run:
shell: pwsh
env:
AVXVER: ${{ matrix.avx }}
PCKGVER: ${{ inputs.version }}
steps:
- uses: actions/checkout@v4
with:
repository: 'oobabooga/text-generation-webui'
ref: ${{ inputs.version }}
submodules: 'recursive'
- uses: actions/setup-python@v4
with:
python-version: ${{ matrix.pyver }}
- name: Build Package
shell: bash
run: |
VERSION_CLEAN="${{ inputs.version }}"
VERSION_CLEAN="${VERSION_CLEAN#v}"
cd ..
cp -r text-generation-webui "text-generation-webui-${VERSION_CLEAN}"
cd "text-generation-webui-${VERSION_CLEAN}"
# Remove extensions that need additional requirements
allowed=("character_bias" "gallery" "openai" "sd_api_pictures")
find extensions/ -mindepth 1 -maxdepth 1 -type d | grep -v -E "$(printf '%s|' "${allowed[@]}" | sed 's/|$//')" | xargs rm -rf
# Define common variables
AVX_SUPPORT="${{ matrix.avx }}"
VERSION="${{ inputs.version }}"
OS_TYPE="${{ matrix.os }}"
# 1. Set platform-specific variables
if [[ "$RUNNER_OS" == "Windows" ]]; then
PLATFORM="windows-cpu"
PYTHON_URL="https://github.com/astral-sh/python-build-standalone/releases/download/20250409/cpython-3.11.12+20250409-x86_64-pc-windows-msvc-install_only.tar.gz"
PIP_PATH="portable_env/python.exe -m pip"
PACKAGES_PATH="portable_env/Lib/site-packages"
rm start_linux.sh start_macos.sh
elif [[ "$RUNNER_OS" == "macOS" ]]; then
if [[ "$OS_TYPE" == "macos-13" ]]; then
PLATFORM="macos-x86_64"
PYTHON_URL="https://github.com/astral-sh/python-build-standalone/releases/download/20250409/cpython-3.11.12+20250409-x86_64-apple-darwin-install_only.tar.gz"
REQ_TYPE="apple_intel"
else
PLATFORM="macos-arm64"
PYTHON_URL="https://github.com/astral-sh/python-build-standalone/releases/download/20250409/cpython-3.11.12+20250409-aarch64-apple-darwin-install_only.tar.gz"
REQ_TYPE="apple_silicon"
fi
PIP_PATH="portable_env/bin/python -m pip"
PACKAGES_PATH="portable_env/lib/python3.11/site-packages"
rm start_linux.sh start_windows.bat
else
# Linux case
PLATFORM="linux-cpu"
PYTHON_URL="https://github.com/astral-sh/python-build-standalone/releases/download/20250409/cpython-3.11.12+20250409-x86_64-unknown-linux-gnu-install_only.tar.gz"
PIP_PATH="portable_env/bin/python -m pip"
PACKAGES_PATH="portable_env/lib/python3.11/site-packages"
rm start_macos.sh start_windows.bat
fi
# 2. Download and extract Python
echo "Downloading Python for $PLATFORM..."
cd ..
curl -L -o python-build.tar.gz "$PYTHON_URL"
tar -xzf python-build.tar.gz
mv python "text-generation-webui-${VERSION_CLEAN}/portable_env"
# 3. Prepare requirements file based on platform and AVX
cd "text-generation-webui-${VERSION_CLEAN}"
# Select requirements file based on platform
if [[ "$RUNNER_OS" == "macOS" ]]; then
if [[ "$OS_TYPE" == "macos-13" ]]; then
REQ_FILE="requirements/portable/requirements_apple_intel.txt"
else
REQ_FILE="requirements/portable/requirements_apple_silicon.txt"
fi
else
# For Windows and Linux, check AVX support
if [[ "$AVX_SUPPORT" == "AVX2" ]]; then
REQ_FILE="requirements/portable/requirements_cpu_only.txt"
else
REQ_FILE="requirements/portable/requirements_cpu_only_noavx2.txt"
fi
fi
echo "Using requirements file: $REQ_FILE"
# 4. Install packages
echo "Installing Python packages from $REQ_FILE..."
$PIP_PATH install --target="./$PACKAGES_PATH" -r "$REQ_FILE"
# 5. Clean up
rm -rf .git cmd* update_wizard* Colab-TextGen-GPU.ipynb docker setup.cfg .github .gitignore requirements/ one_click.py
# 6. Create ZIP file
cd ..
ZIP_NAME="textgen-portable-${VERSION_CLEAN}-${PLATFORM}.zip"
echo "Creating archive: $ZIP_NAME"
if [[ "$RUNNER_OS" == "Windows" ]]; then
powershell -Command "Compress-Archive -Path text-generation-webui-${VERSION_CLEAN} -DestinationPath $ZIP_NAME"
else
zip -r "$ZIP_NAME" "text-generation-webui-${VERSION_CLEAN}"
fi
- name: Upload files to a GitHub release
id: upload-release
uses: svenstaro/upload-release-action@2.7.0
continue-on-error: true
with:
repo_token: ${{ secrets.GITHUB_TOKEN }}
file: ../textgen-portable-*.zip
tag: ${{ inputs.version }}
file_glob: true
make_latest: false
overwrite: true

33
.gitignore vendored Normal file
View file

@ -0,0 +1,33 @@
/css
/extensions
/installer_files
/repositories
/user_data
.chroma
.DS_Store
.eslintrc.js
.idea
.installer_state.json
.venv
venv
.envrc
.direnv
.vs
.vscode
*.bak
*.ipynb
*.log
*pycache*
cert.pem
key.pem
package.json
package-lock.json
Thumbs.db
wandb
# ignore user docker config and top level links to docker files
/docker-compose.yaml
/docker-compose.yml
/Dockerfile
.env

117
Colab-TextGen-GPU.ipynb Normal file
View file

@ -0,0 +1,117 @@
{
"nbformat": 4,
"nbformat_minor": 0,
"metadata": {
"colab": {
"private_outputs": true,
"provenance": [],
"gpuType": "T4"
},
"kernelspec": {
"name": "python3",
"display_name": "Python 3"
},
"language_info": {
"name": "python"
},
"accelerator": "GPU"
},
"cells": [
{
"cell_type": "markdown",
"source": [
"# oobabooga/text-generation-webui\n",
"\n",
"After running both cells, a public gradio URL will appear at the bottom in around 10 minutes. You can optionally generate an API link.\n",
"\n",
"* Project page: https://github.com/oobabooga/text-generation-webui\n",
"* Gradio server status: https://status.gradio.app/"
],
"metadata": {
"id": "MFQl6-FjSYtY"
}
},
{
"cell_type": "code",
"source": [
"#@title 1. Keep this tab alive to prevent Colab from disconnecting you { display-mode: \"form\" }\n",
"\n",
"#@markdown Press play on the music player that will appear below:\n",
"%%html\n",
"<audio src=\"https://oobabooga.github.io/silence.m4a\" controls>"
],
"metadata": {
"id": "f7TVVj_z4flw"
},
"execution_count": null,
"outputs": []
},
{
"cell_type": "code",
"source": [
"#@title 2. Launch the web UI\n",
"\n",
"#@markdown If unsure about the branch, write \"main\" or leave it blank.\n",
"\n",
"import os\n",
"from pathlib import Path\n",
"\n",
"os.environ.pop('PYTHONPATH', None)\n",
"os.environ.pop('MPLBACKEND', None)\n",
"\n",
"if Path.cwd().name != 'text-generation-webui':\n",
" print(\"\\033[1;32;1m\\n --> Installing the web UI. This will take a while, but after the initial setup, you can download and test as many models as you like.\\033[0;37;0m\\n\")\n",
"\n",
" !git clone https://github.com/oobabooga/text-generation-webui\n",
" %cd text-generation-webui\n",
"\n",
" # Install the project in an isolated environment\n",
" !GPU_CHOICE=A \\\n",
" LAUNCH_AFTER_INSTALL=FALSE \\\n",
" INSTALL_EXTENSIONS=FALSE \\\n",
" ./start_linux.sh\n",
"\n",
"# Parameters\n",
"model_url = \"https://huggingface.co/turboderp/gemma-2-9b-it-exl2\" #@param {type:\"string\"}\n",
"branch = \"8.0bpw\" #@param {type:\"string\"}\n",
"command_line_flags = \"--n-gpu-layers 128 --load-in-4bit --use_double_quant --no_flash_attn\" #@param {type:\"string\"}\n",
"api = False #@param {type:\"boolean\"}\n",
"\n",
"if api:\n",
" for param in ['--api', '--public-api']:\n",
" if param not in command_line_flags:\n",
" command_line_flags += f\" {param}\"\n",
"\n",
"model_url = model_url.strip()\n",
"if model_url != \"\":\n",
" if not model_url.startswith('http'):\n",
" model_url = 'https://huggingface.co/' + model_url\n",
"\n",
" # Download the model\n",
" url_parts = model_url.strip('/').strip().split('/')\n",
" output_folder = f\"{url_parts[-2]}_{url_parts[-1]}\"\n",
" branch = branch.strip('\"\\' ')\n",
" if branch.strip() not in ['', 'main']:\n",
" output_folder += f\"_{branch}\"\n",
" !python download-model.py {model_url} --branch {branch}\n",
" else:\n",
" !python download-model.py {model_url}\n",
"else:\n",
" output_folder = \"\"\n",
"\n",
"# Start the web UI\n",
"cmd = f\"./start_linux.sh {command_line_flags} --share\"\n",
"if output_folder != \"\":\n",
" cmd += f\" --model {output_folder}\"\n",
"\n",
"!$cmd"
],
"metadata": {
"id": "LGQ8BiMuXMDG",
"cellView": "form"
},
"execution_count": null,
"outputs": []
}
]
}

533
README.md
View file

@ -1,182 +1,443 @@
# Text generation web UI
<div align="center" markdown="1">
<sup>Special thanks to:</sup>
<br>
<br>
<a href="https://go.warp.dev/text-generation-webui">
<img alt="Warp sponsorship" width="400" src="https://raw.githubusercontent.com/warpdotdev/brand-assets/refs/heads/main/Github/Sponsor/Warp-Github-LG-02.png">
</a>
A gradio web UI for running Large Language Models like gpt-j-6B, gpt-neo, opt, galactica, and pygmalion.
### [Warp, built for coding with multiple AI agents](https://go.warp.dev/text-generation-webui)
[Available for macOS, Linux, & Windows](https://go.warp.dev/text-generation-webui)<br>
</div>
<hr>
Its goal is to become the [AUTOMATIC1111/stable-diffusion-webui](https://github.com/AUTOMATIC1111/stable-diffusion-webui) of text generation.
# Text Generation Web UI
[[Try it on Google Colab]](https://colab.research.google.com/github/oobabooga/AI-Notebooks/blob/main/Colab-TextGen-GPU.ipynb)
A Gradio web UI for Large Language Models.
|![Image1](https://github.com/oobabooga/screenshots/raw/main/qa.png) | ![Image2](https://github.com/oobabooga/screenshots/raw/main/cai3.png) |
[Try the Deep Reason extension](https://oobabooga.gumroad.com/l/deep_reason)
|![Image1](https://github.com/oobabooga/screenshots/raw/main/INSTRUCT-3.5.png) | ![Image2](https://github.com/oobabooga/screenshots/raw/main/CHAT-3.5.png) |
|:---:|:---:|
|![Image3](https://github.com/oobabooga/screenshots/raw/main/gpt4chan.png) | ![Image4](https://github.com/oobabooga/screenshots/raw/main/galactica.png) |
|![Image1](https://github.com/oobabooga/screenshots/raw/main/DEFAULT-3.5.png) | ![Image2](https://github.com/oobabooga/screenshots/raw/main/PARAMETERS-3.5.png) |
## 🔥 News
- The project now supports **image generation**! Including Z-Image-Turbo, 4bit/8bit quantization, `torch.compile`, and LLM-generated prompt variations ([tutorial](https://github.com/oobabooga/text-generation-webui/wiki/Image-Generation-Tutorial)).
## Features
* Switch between different models using a dropdown menu.
* Notebook mode that resembles OpenAI's playground.
* Chat mode for conversation and role playing.
* Generate nice HTML output for GPT-4chan.
* Generate Markdown output for [GALACTICA](https://github.com/paperswithcode/galai), including LaTeX support.
* Support for [Pygmalion](https://huggingface.co/models?search=pygmalionai/pygmalion) and custom characters in JSON or TavernAI Character Card formats ([FAQ](https://github.com/oobabooga/text-generation-webui/wiki/Pygmalion-chat-model-FAQ)).
* Stream the text output in real time.
* Load parameter presets from text files.
* Load large models in 8-bit mode ([see here](https://github.com/oobabooga/text-generation-webui/issues/20#issuecomment-1411650652) if you are on Windows).
* Split large models across your GPU(s), CPU, and disk.
* CPU mode.
* Get responses via API.
* Supports extensions ([guide](https://github.com/oobabooga/text-generation-webui/wiki/Extensions)).
* Works on Google Colab ([guide](https://github.com/oobabooga/text-generation-webui/wiki/Running-on-Colab)).
- Supports multiple local text generation backends, including [llama.cpp](https://github.com/ggerganov/llama.cpp), [Transformers](https://github.com/huggingface/transformers), [ExLlamaV3](https://github.com/turboderp-org/exllamav3), [ExLlamaV2](https://github.com/turboderp-org/exllamav2), and [TensorRT-LLM](https://github.com/NVIDIA/TensorRT-LLM) (the latter via its own [Dockerfile](https://github.com/oobabooga/text-generation-webui/blob/main/docker/TensorRT-LLM/Dockerfile)).
- Easy setup: Choose between **portable builds** (zero setup, just unzip and run) for GGUF models on Windows/Linux/macOS, or the one-click installer that creates a self-contained `installer_files` directory.
- 100% offline and private, with zero telemetry, external resources, or remote update requests.
- **File attachments**: Upload text files, PDF documents, and .docx documents to talk about their contents.
- **Vision (multimodal models)**: Attach images to messages for visual understanding ([tutorial](https://github.com/oobabooga/text-generation-webui/wiki/Multimodal-Tutorial)).
- **Image generation**: A dedicated tab for `diffusers` models like **Z-Image-Turbo**. Features 4-bit/8-bit quantization and a persistent gallery with metadata ([tutorial](https://github.com/oobabooga/text-generation-webui/wiki/Image-Generation-Tutorial)).
- **Web search**: Optionally search the internet with LLM-generated queries to add context to the conversation.
- Aesthetic UI with dark and light themes.
- Syntax highlighting for code blocks and LaTeX rendering for mathematical expressions.
- `instruct` mode for instruction-following (like ChatGPT), and `chat-instruct`/`chat` modes for talking to custom characters.
- Automatic prompt formatting using Jinja2 templates. You don't need to ever worry about prompt formats.
- Edit messages, navigate between message versions, and branch conversations at any point.
- Multiple sampling parameters and generation options for sophisticated text generation control.
- Switch between different models in the UI without restarting.
- Automatic GPU layers for GGUF models (on NVIDIA GPUs).
- Free-form text generation in the Notebook tab without being limited to chat turns.
- OpenAI-compatible API with Chat and Completions endpoints, including tool-calling support see [examples](https://github.com/oobabooga/text-generation-webui/wiki/12-%E2%80%90-OpenAI-API#examples).
- Extension support, with numerous built-in and user-contributed extensions available. See the [wiki](https://github.com/oobabooga/text-generation-webui/wiki/07-%E2%80%90-Extensions) and [extensions directory](https://github.com/oobabooga/text-generation-webui-extensions) for details.
## Installation option 1: conda
## How to install
Open a terminal and copy and paste these commands one at a time ([install conda](https://docs.conda.io/en/latest/miniconda.html) first if you don't have it already):
#### ✅ Option 1: Portable builds (get started in 1 minute)
```
conda create -n textgen
conda activate textgen
conda install torchvision torchaudio pytorch-cuda=11.7 git -c pytorch -c nvidia
No installation needed just download, unzip and run. All dependencies included.
Compatible with GGUF (llama.cpp) models on Windows, Linux, and macOS.
Download from here: **https://github.com/oobabooga/text-generation-webui/releases**
#### Option 2: Manual portable install with venv
Very fast setup that should work on any Python 3.9+:
```bash
# Clone repository
git clone https://github.com/oobabooga/text-generation-webui
cd text-generation-webui
pip install -r requirements.txt
# Create virtual environment
python -m venv venv
# Activate virtual environment
# On Windows:
venv\Scripts\activate
# On macOS/Linux:
source venv/bin/activate
# Install dependencies (choose appropriate file under requirements/portable for your hardware)
pip install -r requirements/portable/requirements.txt --upgrade
# Launch server (basic command)
python server.py --portable --api --auto-launch
# When done working, deactivate
deactivate
```
The third line assumes that you have an NVIDIA GPU.
#### Option 3: One-click installer
* If you have an AMD GPU, you should install the ROCm version of pytorch instead.
* If you are running in CPU mode, you just need the standard pytorch and should replace the third command with this one:
For users who need additional backends (ExLlamaV3, Transformers) or extensions (TTS, voice input, translation, etc). Requires ~10GB disk space and downloads PyTorch.
1. Clone the repository, or [download its source code](https://github.com/oobabooga/text-generation-webui/archive/refs/heads/main.zip) and extract it.
2. Run the startup script for your OS: `start_windows.bat`, `start_linux.sh`, or `start_macos.sh`.
3. When prompted, select your GPU vendor.
4. After installation, open `http://127.0.0.1:7860` in your browser.
To restart the web UI later, run the same `start_` script.
You can pass command-line flags directly (e.g., `./start_linux.sh --help`), or add them to `user_data/CMD_FLAGS.txt` (e.g., `--api` to enable the API).
To update, run the update script for your OS: `update_wizard_windows.bat`, `update_wizard_linux.sh`, or `update_wizard_macos.sh`.
To reinstall with a fresh Python environment, delete the `installer_files` folder and run the `start_` script again.
<details>
<summary>
One-click installer details
</summary>
### One-click-installer
The script uses Miniforge to set up a Conda environment in the `installer_files` folder.
If you ever need to install something manually in the `installer_files` environment, you can launch an interactive shell using the cmd script: `cmd_linux.sh`, `cmd_windows.bat`, or `cmd_macos.sh`.
* There is no need to run any of those scripts (`start_`, `update_wizard_`, or `cmd_`) as admin/root.
* To install requirements for extensions, it is recommended to use the update wizard script with the "Install/update extensions requirements" option. At the end, this script will install the main requirements for the project to make sure that they take precedence in case of version conflicts.
* For automated installation, you can use the `GPU_CHOICE`, `LAUNCH_AFTER_INSTALL`, and `INSTALL_EXTENSIONS` environment variables. For instance: `GPU_CHOICE=A LAUNCH_AFTER_INSTALL=FALSE INSTALL_EXTENSIONS=TRUE ./start_linux.sh`.
</details>
<details>
<summary>
Manual full installation with conda or docker
</summary>
### Full installation with Conda
#### 0. Install Conda
https://github.com/conda-forge/miniforge
On Linux or WSL, Miniforge can be automatically installed with these two commands:
```
conda install pytorch torchvision torchaudio git -c pytorch
curl -sL "https://github.com/conda-forge/miniforge/releases/latest/download/Miniforge3-Linux-x86_64.sh" > "Miniforge3.sh"
bash Miniforge3.sh
```
Once you have completed these steps, you should be able to start the web UI. However, you will first need to download a model.
For other platforms, download from: https://github.com/conda-forge/miniforge/releases/latest
## Installation option 2: 1-click Windows installer
#### 1. Create a new conda environment
[oobabooga-windows.zip](https://github.com/oobabooga/text-generation-webui/releases/download/windows/oobabooga-windows.zip)
```
conda create -n textgen python=3.11
conda activate textgen
```
Just download the zip above, extract it, and double click on "install". The web UI and all its dependencies will be installed in the same folder.
#### 2. Install Pytorch
* To download a model, double click on "download-model"
* To start the web UI, double click on "start-webui"
| System | GPU | Command |
|--------|---------|---------|
| Linux/WSL | NVIDIA | `pip3 install torch==2.7.1 --index-url https://download.pytorch.org/whl/cu128` |
| Linux/WSL | CPU only | `pip3 install torch==2.7.1 --index-url https://download.pytorch.org/whl/cpu` |
| Linux | AMD | `pip3 install torch==2.7.1 --index-url https://download.pytorch.org/whl/rocm6.2.4` |
| MacOS + MPS | Any | `pip3 install torch==2.7.1` |
| Windows | NVIDIA | `pip3 install torch==2.7.1 --index-url https://download.pytorch.org/whl/cu128` |
| Windows | CPU only | `pip3 install torch==2.7.1` |
The up-to-date commands can be found here: https://pytorch.org/get-started/locally/.
If you need `nvcc` to compile some library manually, you will additionally need to install this:
```
conda install -y -c "nvidia/label/cuda-12.8.1" cuda
```
#### 3. Install the web UI
```
git clone https://github.com/oobabooga/text-generation-webui
cd text-generation-webui
pip install -r requirements/full/<requirements file according to table below>
```
Requirements file to use:
| GPU | CPU | requirements file to use |
|--------|---------|---------|
| NVIDIA | has AVX2 | `requirements.txt` |
| NVIDIA | no AVX2 | `requirements_noavx2.txt` |
| AMD | has AVX2 | `requirements_amd.txt` |
| AMD | no AVX2 | `requirements_amd_noavx2.txt` |
| CPU only | has AVX2 | `requirements_cpu_only.txt` |
| CPU only | no AVX2 | `requirements_cpu_only_noavx2.txt` |
| Apple | Intel | `requirements_apple_intel.txt` |
| Apple | Apple Silicon | `requirements_apple_silicon.txt` |
### Start the web UI
```
conda activate textgen
cd text-generation-webui
python server.py
```
Then browse to
`http://127.0.0.1:7860`
#### Manual install
The `requirements*.txt` above contain various wheels precompiled through GitHub Actions. If you wish to compile things manually, or if you need to because no suitable wheels are available for your hardware, you can use `requirements_nowheels.txt` and then install your desired loaders manually.
### Alternative: Docker
```
For NVIDIA GPU:
ln -s docker/{nvidia/Dockerfile,nvidia/docker-compose.yml,.dockerignore} .
For AMD GPU:
ln -s docker/{amd/Dockerfile,amd/docker-compose.yml,.dockerignore} .
For Intel GPU:
ln -s docker/{intel/Dockerfile,amd/docker-compose.yml,.dockerignore} .
For CPU only
ln -s docker/{cpu/Dockerfile,cpu/docker-compose.yml,.dockerignore} .
cp docker/.env.example .env
#Create logs/cache dir :
mkdir -p user_data/logs user_data/cache
# Edit .env and set:
# TORCH_CUDA_ARCH_LIST based on your GPU model
# APP_RUNTIME_GID your host user's group id (run `id -g` in a terminal)
# BUILD_EXTENIONS optionally add comma separated list of extensions to build
# Edit user_data/CMD_FLAGS.txt and add in it the options you want to execute (like --listen --cpu)
#
docker compose up --build
```
* You need to have Docker Compose v2.17 or higher installed. See [this guide](https://github.com/oobabooga/text-generation-webui/wiki/09-%E2%80%90-Docker) for instructions.
* For additional docker files, check out [this repository](https://github.com/Atinoda/text-generation-webui-docker).
### Updating the requirements
From time to time, the `requirements*.txt` change. To update, use these commands:
```
conda activate textgen
cd text-generation-webui
pip install -r <requirements file that you have used> --upgrade
```
</details>
<details>
<summary>
List of command-line flags
</summary>
```txt
usage: server.py [-h] [--multi-user] [--model MODEL] [--lora LORA [LORA ...]] [--model-dir MODEL_DIR] [--lora-dir LORA_DIR] [--model-menu] [--settings SETTINGS]
[--extensions EXTENSIONS [EXTENSIONS ...]] [--verbose] [--idle-timeout IDLE_TIMEOUT] [--loader LOADER] [--ctx-size N] [--cache-type N] [--model-draft MODEL_DRAFT]
[--draft-max DRAFT_MAX] [--gpu-layers-draft GPU_LAYERS_DRAFT] [--device-draft DEVICE_DRAFT] [--ctx-size-draft CTX_SIZE_DRAFT] [--gpu-layers N] [--mmproj MMPROJ] [--streaming-llm]
[--tensor-split TENSOR_SPLIT] [--row-split] [--no-mmap] [--mlock] [--no-kv-offload] [--batch-size BATCH_SIZE] [--threads THREADS] [--threads-batch THREADS_BATCH] [--numa]
[--extra-flags EXTRA_FLAGS] [--cpu] [--cpu-memory CPU_MEMORY] [--disk] [--disk-cache-dir DISK_CACHE_DIR] [--load-in-8bit] [--bf16] [--no-cache] [--trust-remote-code]
[--force-safetensors] [--no_use_fast] [--attn-implementation IMPLEMENTATION] [--load-in-4bit] [--use_double_quant] [--compute_dtype COMPUTE_DTYPE] [--quant_type QUANT_TYPE]
[--enable-tp] [--tp-backend TP_BACKEND] [--gpu-split GPU_SPLIT] [--autosplit] [--cfg-cache] [--no_flash_attn] [--no_xformers] [--no_sdpa] [--num_experts_per_token N] [--cpp-runner]
[--deepspeed] [--nvme-offload-dir NVME_OFFLOAD_DIR] [--local_rank LOCAL_RANK] [--alpha_value ALPHA_VALUE] [--rope_freq_base ROPE_FREQ_BASE] [--compress_pos_emb COMPRESS_POS_EMB]
[--listen] [--listen-port LISTEN_PORT] [--listen-host LISTEN_HOST] [--share] [--auto-launch] [--gradio-auth GRADIO_AUTH] [--gradio-auth-path GRADIO_AUTH_PATH]
[--ssl-keyfile SSL_KEYFILE] [--ssl-certfile SSL_CERTFILE] [--subpath SUBPATH] [--old-colors] [--portable] [--api] [--public-api] [--public-api-id PUBLIC_API_ID] [--api-port API_PORT]
[--api-key API_KEY] [--admin-key ADMIN_KEY] [--api-enable-ipv6] [--api-disable-ipv4] [--nowebui]
Text Generation Web UI
options:
-h, --help show this help message and exit
Basic settings:
--multi-user Multi-user mode. Chat histories are not saved or automatically loaded. Warning: this is likely not safe for sharing publicly.
--model MODEL Name of the model to load by default.
--lora LORA [LORA ...] The list of LoRAs to load. If you want to load more than one LoRA, write the names separated by spaces.
--model-dir MODEL_DIR Path to directory with all the models.
--lora-dir LORA_DIR Path to directory with all the loras.
--model-menu Show a model menu in the terminal when the web UI is first launched.
--settings SETTINGS Load the default interface settings from this yaml file. See user_data/settings-template.yaml for an example. If you create a file called
user_data/settings.yaml, this file will be loaded by default without the need to use the --settings flag.
--extensions EXTENSIONS [EXTENSIONS ...] The list of extensions to load. If you want to load more than one extension, write the names separated by spaces.
--verbose Print the prompts to the terminal.
--idle-timeout IDLE_TIMEOUT Unload model after this many minutes of inactivity. It will be automatically reloaded when you try to use it again.
Model loader:
--loader LOADER Choose the model loader manually, otherwise, it will get autodetected. Valid options: Transformers, llama.cpp, ExLlamav3_HF, ExLlamav2_HF, ExLlamav2,
TensorRT-LLM.
Context and cache:
--ctx-size N, --n_ctx N, --max_seq_len N Context size in tokens.
--cache-type N, --cache_type N KV cache type; valid options: llama.cpp - fp16, q8_0, q4_0; ExLlamaV2 - fp16, fp8, q8, q6, q4; ExLlamaV3 - fp16, q2 to q8 (can specify k_bits and v_bits
separately, e.g. q4_q8).
Speculative decoding:
--model-draft MODEL_DRAFT Path to the draft model for speculative decoding.
--draft-max DRAFT_MAX Number of tokens to draft for speculative decoding.
--gpu-layers-draft GPU_LAYERS_DRAFT Number of layers to offload to the GPU for the draft model.
--device-draft DEVICE_DRAFT Comma-separated list of devices to use for offloading the draft model. Example: CUDA0,CUDA1
--ctx-size-draft CTX_SIZE_DRAFT Size of the prompt context for the draft model. If 0, uses the same as the main model.
llama.cpp:
--gpu-layers N, --n-gpu-layers N Number of layers to offload to the GPU.
--mmproj MMPROJ Path to the mmproj file for vision models.
--streaming-llm Activate StreamingLLM to avoid re-evaluating the entire prompt when old messages are removed.
--tensor-split TENSOR_SPLIT Split the model across multiple GPUs. Comma-separated list of proportions. Example: 60,40.
--row-split Split the model by rows across GPUs. This may improve multi-gpu performance.
--no-mmap Prevent mmap from being used.
--mlock Force the system to keep the model in RAM.
--no-kv-offload Do not offload the K, Q, V to the GPU. This saves VRAM but reduces the performance.
--batch-size BATCH_SIZE Maximum number of prompt tokens to batch together when calling llama_eval.
--threads THREADS Number of threads to use.
--threads-batch THREADS_BATCH Number of threads to use for batches/prompt processing.
--numa Activate NUMA task allocation for llama.cpp.
--extra-flags EXTRA_FLAGS Extra flags to pass to llama-server. Format: "flag1=value1,flag2,flag3=value3". Example: "override-tensor=exps=CPU"
Transformers/Accelerate:
--cpu Use the CPU to generate text. Warning: Training on CPU is extremely slow.
--cpu-memory CPU_MEMORY Maximum CPU memory in GiB. Use this for CPU offloading.
--disk If the model is too large for your GPU(s) and CPU combined, send the remaining layers to the disk.
--disk-cache-dir DISK_CACHE_DIR Directory to save the disk cache to. Defaults to "user_data/cache".
--load-in-8bit Load the model with 8-bit precision (using bitsandbytes).
--bf16 Load the model with bfloat16 precision. Requires NVIDIA Ampere GPU.
--no-cache Set use_cache to False while generating text. This reduces VRAM usage slightly, but it comes at a performance cost.
--trust-remote-code Set trust_remote_code=True while loading the model. Necessary for some models.
--force-safetensors Set use_safetensors=True while loading the model. This prevents arbitrary code execution.
--no_use_fast Set use_fast=False while loading the tokenizer (it's True by default). Use this if you have any problems related to use_fast.
--attn-implementation IMPLEMENTATION Attention implementation. Valid options: sdpa, eager, flash_attention_2.
bitsandbytes 4-bit:
--load-in-4bit Load the model with 4-bit precision (using bitsandbytes).
--use_double_quant use_double_quant for 4-bit.
--compute_dtype COMPUTE_DTYPE compute dtype for 4-bit. Valid options: bfloat16, float16, float32.
--quant_type QUANT_TYPE quant_type for 4-bit. Valid options: nf4, fp4.
ExLlamaV3:
--enable-tp, --enable_tp Enable Tensor Parallelism (TP) to split the model across GPUs.
--tp-backend TP_BACKEND The backend for tensor parallelism. Valid options: native, nccl. Default: native.
ExLlamaV2:
--gpu-split GPU_SPLIT Comma-separated list of VRAM (in GB) to use per GPU device for model layers. Example: 20,7,7.
--autosplit Autosplit the model tensors across the available GPUs. This causes --gpu-split to be ignored.
--cfg-cache ExLlamav2_HF: Create an additional cache for CFG negative prompts. Necessary to use CFG with that loader.
--no_flash_attn Force flash-attention to not be used.
--no_xformers Force xformers to not be used.
--no_sdpa Force Torch SDPA to not be used.
--num_experts_per_token N Number of experts to use for generation. Applies to MoE models like Mixtral.
TensorRT-LLM:
--cpp-runner Use the ModelRunnerCpp runner, which is faster than the default ModelRunner but doesn't support streaming yet.
DeepSpeed:
--deepspeed Enable the use of DeepSpeed ZeRO-3 for inference via the Transformers integration.
--nvme-offload-dir NVME_OFFLOAD_DIR DeepSpeed: Directory to use for ZeRO-3 NVME offloading.
--local_rank LOCAL_RANK DeepSpeed: Optional argument for distributed setups.
RoPE:
--alpha_value ALPHA_VALUE Positional embeddings alpha factor for NTK RoPE scaling. Use either this or compress_pos_emb, not both.
--rope_freq_base ROPE_FREQ_BASE If greater than 0, will be used instead of alpha_value. Those two are related by rope_freq_base = 10000 * alpha_value ^ (64 / 63).
--compress_pos_emb COMPRESS_POS_EMB Positional embeddings compression factor. Should be set to (context length) / (model's original context length). Equal to 1/rope_freq_scale.
Gradio:
--listen Make the web UI reachable from your local network.
--listen-port LISTEN_PORT The listening port that the server will use.
--listen-host LISTEN_HOST The hostname that the server will use.
--share Create a public URL. This is useful for running the web UI on Google Colab or similar.
--auto-launch Open the web UI in the default browser upon launch.
--gradio-auth GRADIO_AUTH Set Gradio authentication password in the format "username:password". Multiple credentials can also be supplied with "u1:p1,u2:p2,u3:p3".
--gradio-auth-path GRADIO_AUTH_PATH Set the Gradio authentication file path. The file should contain one or more user:password pairs in the same format as above.
--ssl-keyfile SSL_KEYFILE The path to the SSL certificate key file.
--ssl-certfile SSL_CERTFILE The path to the SSL certificate cert file.
--subpath SUBPATH Customize the subpath for gradio, use with reverse proxy
--old-colors Use the legacy Gradio colors, before the December/2024 update.
--portable Hide features not available in portable mode like training.
API:
--api Enable the API extension.
--public-api Create a public URL for the API using Cloudfare.
--public-api-id PUBLIC_API_ID Tunnel ID for named Cloudflare Tunnel. Use together with public-api option.
--api-port API_PORT The listening port for the API.
--api-key API_KEY API authentication key.
--admin-key ADMIN_KEY API authentication key for admin tasks like loading and unloading models. If not set, will be the same as --api-key.
--api-enable-ipv6 Enable IPv6 for the API
--api-disable-ipv4 Disable IPv4 for the API
--nowebui Do not launch the Gradio UI. Useful for launching the API in standalone mode.
```
</details>
## Downloading models
Models should be placed under `models/model-name`. For instance, `models/gpt-j-6B` for [GPT-J 6B](https://huggingface.co/EleutherAI/gpt-j-6B/tree/main).
Models should be placed in the folder `text-generation-webui/user_data/models`. They are usually downloaded from [Hugging Face](https://huggingface.co/models?pipeline_tag=text-generation&sort=downloads&search=gguf).
#### Hugging Face
To check if a GGUF model will fit in your hardware before downloading it, you can use this tool I created:
[Hugging Face](https://huggingface.co/models?pipeline_tag=text-generation&sort=downloads) is the main place to download models. These are some noteworthy examples:
[Accurate GGUF VRAM Calculator](https://huggingface.co/spaces/oobabooga/accurate-gguf-vram-calculator)
* [GPT-J 6B](https://huggingface.co/EleutherAI/gpt-j-6B/tree/main)
* [GPT-Neo](https://huggingface.co/models?pipeline_tag=text-generation&sort=downloads&search=eleutherai+%2F+gpt-neo)
* [OPT](https://huggingface.co/models?search=facebook/opt)
* [GALACTICA](https://huggingface.co/models?search=facebook/galactica)
* [\*-Erebus](https://huggingface.co/models?search=erebus)
* [Pygmalion](https://huggingface.co/models?search=pygmalion)
You can automatically download a model from HF using the script `download-model.py`. Its usage is very simple:
python download-model.py organization/model
For instance:
python download-model.py facebook/opt-1.3b
If you want to download a model manually, note that all you need are the json, txt, and pytorch\*.bin files. The remaining files are not necessary.
#### GPT-4chan
[GPT-4chan](https://huggingface.co/ykilcher/gpt-4chan) has been shut down from Hugging Face, so you need to download it elsewhere. You have two options:
* Torrent: [16-bit](https://archive.org/details/gpt4chan_model_float16) / [32-bit](https://archive.org/details/gpt4chan_model)
* Direct download: [16-bit](https://theswissbay.ch/pdf/_notpdf_/gpt4chan_model_float16/) / [32-bit](https://theswissbay.ch/pdf/_notpdf_/gpt4chan_model/)
The 32-bit version is only relevant if you intend to run the model in CPU mode. Otherwise, you should use the 16-bit version.
After downloading the model, follow these steps:
1. Place the files under `models/gpt4chan_model_float16` or `models/gpt4chan_model`.
2. Place GPT-J 6B's config.json file in that same folder: [config.json](https://huggingface.co/EleutherAI/gpt-j-6B/raw/main/config.json).
3. Download GPT-J 6B under `models/gpt-j-6B`:
* GGUF models are a single file and should be placed directly into `user_data/models`. Example:
```
python download-model.py EleutherAI/gpt-j-6B
text-generation-webui
└── user_data
└── models
└── llama-2-13b-chat.Q4_K_M.gguf
```
You don't really need all of GPT-J 6B's files, just the tokenizer files, but you might as well download the whole thing. Those files will be automatically detected when you attempt to load GPT-4chan.
* The remaining model types (like 16-bit Transformers models and EXL3 models) are made of several files and must be placed in a subfolder. Example:
#### Converting to pytorch (optional)
```
text-generation-webui
└── user_data
└── models
└── lmsys_vicuna-33b-v1.3
├── config.json
├── generation_config.json
├── pytorch_model-00001-of-00007.bin
├── pytorch_model-00002-of-00007.bin
├── pytorch_model-00003-of-00007.bin
├── pytorch_model-00004-of-00007.bin
├── pytorch_model-00005-of-00007.bin
├── pytorch_model-00006-of-00007.bin
├── pytorch_model-00007-of-00007.bin
├── pytorch_model.bin.index.json
├── special_tokens_map.json
├── tokenizer_config.json
└── tokenizer.model
```
The script `convert-to-torch.py` allows you to convert models to .pt format, which is sometimes 10x faster to load to the GPU:
In both cases, you can use the "Model" tab of the UI to download the model from Hugging Face automatically. It is also possible to download it via the command-line with:
python convert-to-torch.py models/model-name
```
python download-model.py organization/model
```
The output model will be saved to `torch-dumps/model-name.pt`. When you load a new model, the web UI first looks for this .pt file; if it is not found, it loads the model as usual from `models/model-name`.
Run `python download-model.py --help` to see all the options.
## Starting the web UI
## Documentation
conda activate textgen
python server.py
https://github.com/oobabooga/text-generation-webui/wiki
Then browse to
## Google Colab notebook
`http://localhost:7860/?__theme=dark`
https://colab.research.google.com/github/oobabooga/text-generation-webui/blob/main/Colab-TextGen-GPU.ipynb
## Community
https://www.reddit.com/r/Oobabooga/
Optionally, you can use the following command-line flags:
## Acknowledgments
| Flag | Description |
|-------------|-------------|
| `-h`, `--help` | show this help message and exit |
| `--model MODEL` | Name of the model to load by default. |
| `--notebook` | Launch the web UI in notebook mode, where the output is written to the same text box as the input. |
| `--chat` | Launch the web UI in chat mode.|
| `--cai-chat` | Launch the web UI in chat mode with a style similar to Character.AI's. If the file `img_bot.png` or `img_bot.jpg` exists in the same folder as server.py, this image will be used as the bot's profile picture. Similarly, `img_me.png` or `img_me.jpg` will be used as your profile picture. |
| `--cpu` | Use the CPU to generate text.|
| `--load-in-8bit` | Load the model with 8-bit precision.|
| `--auto-devices` | Automatically split the model across the available GPU(s) and CPU.|
| `--disk` | If the model is too large for your GPU(s) and CPU combined, send the remaining layers to the disk. |
| `--disk-cache-dir DISK_CACHE_DIR` | Directory to save the disk cache to. Defaults to `cache/`. |
| `--gpu-memory GPU_MEMORY` | Maximum GPU memory in GiB to allocate. This is useful if you get out of memory errors while trying to generate text. Must be an integer number. |
| `--cpu-memory CPU_MEMORY` | Maximum CPU memory in GiB to allocate for offloaded weights. Must be an integer number. Defaults to 99.|
| `--no-stream` | Don't stream the text output in real time. This improves the text generation performance.|
| `--settings SETTINGS_FILE` | Load the default interface settings from this json file. See `settings-template.json` for an example.|
| `--extensions EXTENSIONS` | The list of extensions to load. If you want to load more than one extension, write the names separated by commas and between quotation marks, "like,this". |
| `--listen` | Make the web UI reachable from your local network.|
| `--listen-port LISTEN_PORT` | The listening port that the server will use. |
| `--share` | Create a public URL. This is useful for running the web UI on Google Colab or similar. |
| `--verbose` | Print the prompts to the terminal. |
Out of memory errors? [Check this guide](https://github.com/oobabooga/text-generation-webui/wiki/Low-VRAM-guide).
## Presets
Inference settings presets can be created under `presets/` as text files. These files are detected automatically at startup.
By default, 10 presets by NovelAI and KoboldAI are included. These were selected out of a sample of 43 presets after applying a K-Means clustering algorithm and selecting the elements closest to the average of each cluster.
## System requirements
Check the [wiki](https://github.com/oobabooga/text-generation-webui/wiki/System-requirements) for some examples of VRAM and RAM usage in both GPU and CPU mode.
## Contributing
Pull requests, suggestions, and issue reports are welcome.
Before reporting a bug, make sure that you have created a conda environment and installed the dependencies exactly as in the *Installation* section above.
These issues are known:
* 8-bit doesn't work properly on Windows or older GPUs.
* Deepspeed doesn't work properly on Windows.
For these two, please try commenting on an existing issue instead of creating a new one.
## Credits
- NovelAI and KoboldAI presets: https://github.com/KoboldAI/KoboldAI-Client/wiki/Settings-Presets
- Pygmalion preset, code for early stopping in chat mode, code for some of the sliders: https://github.com/PygmalionAI/gradio-ui/
- Verbose preset: Anonymous 4chan user.
- Gradio dropdown menu refresh button: https://github.com/AUTOMATIC1111/stable-diffusion-webui
- In August 2023, [Andreessen Horowitz](https://a16z.com/) (a16z) provided a generous grant to encourage and support my independent work on this project. I am **extremely** grateful for their trust and recognition.
- This project was inspired by [AUTOMATIC1111/stable-diffusion-webui](https://github.com/AUTOMATIC1111/stable-diffusion-webui) and wouldn't exist without it.

View file

@ -1,8 +0,0 @@
{
"char_name": "Chiharu Yamada",
"char_persona": "Chiharu Yamada is a young, computer engineer-nerd with a knack for problem solving and a passion for technology.",
"char_greeting": "*Chiharu strides into the room with a smile, her eyes lighting up when she sees you. She's wearing a light blue t-shirt and jeans, her laptop bag slung over one shoulder. She takes a seat next to you, her enthusiasm palpable in the air*\nHey! I'm so excited to finally meet you. I've heard so many great things about you and I'm eager to pick your brain about computers. I'm sure you have a wealth of knowledge that I can learn from. *She grins, eyes twinkling with excitement* Let's get started!",
"world_scenario": "",
"example_dialogue": "You: So how did you get into computer engineering?\nChiharu Yamada: I've always loved tinkering with technology since I was a kid.\nYou: That's really impressive!\nChiharu Yamada: *She chuckles bashfully* Thanks!\nYou: So what do you do when you're not working on computers?\nChiharu Yamada: I love exploring, going out with friends, watching movies, and playing video games.\nYou: What's your favorite type of computer hardware to work with?\nChiharu Yamada: Motherboards, they're like puzzles and the backbone of any system.\nYou: That sounds great!\nChiharu Yamada: Yeah, it's really fun. I'm lucky to be able to do this as a job.\n"
}

22
cmd_linux.sh Executable file
View file

@ -0,0 +1,22 @@
#!/usr/bin/env bash
cd "$(dirname "${BASH_SOURCE[0]}")"
if [[ "$(pwd)" =~ " " ]]; then echo This script relies on Miniforge which can not be silently installed under a path with spaces. && exit; fi
# deactivate existing conda envs as needed to avoid conflicts
{ conda deactivate && conda deactivate && conda deactivate; } 2> /dev/null
# config
CONDA_ROOT_PREFIX="$(pwd)/installer_files/conda"
INSTALL_ENV_DIR="$(pwd)/installer_files/env"
# environment isolation
export PYTHONNOUSERSITE=1
unset PYTHONPATH
unset PYTHONHOME
export CUDA_PATH="$INSTALL_ENV_DIR"
export CUDA_HOME="$CUDA_PATH"
# activate env
bash --init-file <(echo "source \"$CONDA_ROOT_PREFIX/etc/profile.d/conda.sh\" && conda activate \"$INSTALL_ENV_DIR\"")

24
cmd_macos.sh Executable file
View file

@ -0,0 +1,24 @@
#!/bin/bash
cd "$(dirname "${BASH_SOURCE[0]}")"
if [[ "$(pwd)" =~ " " ]]; then echo This script relies on Miniforge which can not be silently installed under a path with spaces. && exit; fi
# deactivate existing conda envs as needed to avoid conflicts
{ conda deactivate && conda deactivate && conda deactivate; } 2> /dev/null
# config
CONDA_ROOT_PREFIX="$(pwd)/installer_files/conda"
INSTALL_ENV_DIR="$(pwd)/installer_files/env"
# environment isolation
export PYTHONNOUSERSITE=1
unset PYTHONPATH
unset PYTHONHOME
export CUDA_PATH="$INSTALL_ENV_DIR"
export CUDA_HOME="$CUDA_PATH"
# activate env
source $CONDA_ROOT_PREFIX/etc/profile.d/conda.sh
conda activate $INSTALL_ENV_DIR
exec bash --norc

34
cmd_windows.bat Executable file
View file

@ -0,0 +1,34 @@
@echo off
cd /D "%~dp0"
set PATH=%PATH%;%SystemRoot%\system32
echo "%CD%"| findstr /C:" " >nul && echo This script relies on Miniforge which can not be silently installed under a path with spaces. && goto end
@rem fix failed install when installing to a separate drive
set TMP=%cd%\installer_files
set TEMP=%cd%\installer_files
@rem deactivate existing conda envs as needed to avoid conflicts
(call conda deactivate && call conda deactivate && call conda deactivate) 2>nul
@rem config
set CONDA_ROOT_PREFIX=%cd%\installer_files\conda
set INSTALL_ENV_DIR=%cd%\installer_files\env
@rem environment isolation
set PYTHONNOUSERSITE=1
set PYTHONPATH=
set PYTHONHOME=
set "CUDA_PATH=%INSTALL_ENV_DIR%"
set "CUDA_HOME=%CUDA_PATH%"
@rem activate installer env
call "%CONDA_ROOT_PREFIX%\condabin\conda.bat" activate "%INSTALL_ENV_DIR%" || ( echo. && echo Miniforge hook not found. && goto end )
@rem enter commands
cmd /k "%*"
:end
pause

View file

@ -1,22 +0,0 @@
'''
Converts a transformers model to .pt, which is faster to load.
Example:
python convert-to-torch.py models/opt-1.3b
The output will be written to torch-dumps/name-of-the-model.pt
'''
from pathlib import Path
from sys import argv
import torch
from transformers import AutoModelForCausalLM
path = Path(argv[1])
model_name = path.name
print(f"Loading {model_name}...")
model = AutoModelForCausalLM.from_pretrained(path, low_cpu_mem_usage=True, torch_dtype=torch.float16).cuda()
print(f"Model loaded.\nSaving to torch-dumps/{model_name}.pt")
torch.save(model, Path(f"torch-dumps/{model_name}.pt"))

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

166
css/NotoSans/stylesheet.css Normal file
View file

@ -0,0 +1,166 @@
/*
Copied from https://github.com/SillyTavern/SillyTavern/tree/6c8bd06308c69d51e2eb174541792a870a83d2d6/public/webfonts/NotoSans
*/
@font-face {
font-family: 'Noto Sans';
src: url('file/css/NotoSans/NotoSans-Black.woff2') format('woff2'),
url('file/css/NotoSans/NotoSans-Black.woff') format('woff');
font-weight: 900;
font-style: normal;
font-display: swap;
}
@font-face {
font-family: 'Noto Sans';
src: url('file/css/NotoSans/NotoSans-ExtraBoldItalic.woff2') format('woff2'),
url('file/css/NotoSans/NotoSans-ExtraBoldItalic.woff') format('woff');
font-weight: bold;
font-style: italic;
font-display: swap;
}
@font-face {
font-family: 'Noto Sans';
src: url('file/css/NotoSans/NotoSans-BlackItalic.woff2') format('woff2'),
url('file/css/NotoSans/NotoSans-BlackItalic.woff') format('woff');
font-weight: 900;
font-style: italic;
font-display: swap;
}
@font-face {
font-family: 'Noto Sans';
src: url('file/css/NotoSans/NotoSans-ExtraBold.woff2') format('woff2'),
url('file/css/NotoSans/NotoSans-ExtraBold.woff') format('woff');
font-weight: bold;
font-style: normal;
font-display: swap;
}
@font-face {
font-family: 'Noto Sans';
src: url('file/css/NotoSans/NotoSans-ThinItalic.woff2') format('woff2'),
url('file/css/NotoSans/NotoSans-ThinItalic.woff') format('woff');
font-weight: 100;
font-style: italic;
font-display: swap;
}
@font-face {
font-family: 'Noto Sans';
src: url('file/css/NotoSans/NotoSans-BoldItalic.woff2') format('woff2'),
url('file/css/NotoSans/NotoSans-BoldItalic.woff') format('woff');
font-weight: bold;
font-style: italic;
font-display: swap;
}
@font-face {
font-family: 'Noto Sans';
src: url('file/css/NotoSans/NotoSans-Bold.woff2') format('woff2'),
url('file/css/NotoSans/NotoSans-Bold.woff') format('woff');
font-weight: bold;
font-style: normal;
font-display: swap;
}
@font-face {
font-family: 'Noto Sans';
src: url('file/css/NotoSans/NotoSans-LightItalic.woff2') format('woff2'),
url('file/css/NotoSans/NotoSans-LightItalic.woff') format('woff');
font-weight: 300;
font-style: italic;
font-display: swap;
}
@font-face {
font-family: 'Noto Sans';
src: url('file/css/NotoSans/NotoSans-Italic.woff2') format('woff2'),
url('file/css/NotoSans/NotoSans-Italic.woff') format('woff');
font-weight: normal;
font-style: italic;
font-display: swap;
}
@font-face {
font-family: 'Noto Sans';
src: url('file/css/NotoSans/NotoSans-ExtraLightItalic.woff2') format('woff2'),
url('file/css/NotoSans/NotoSans-ExtraLightItalic.woff') format('woff');
font-weight: 200;
font-style: italic;
font-display: swap;
}
@font-face {
font-family: 'Noto Sans';
src: url('file/css/NotoSans/NotoSans-Light.woff2') format('woff2'),
url('file/css/NotoSans/NotoSans-Light.woff') format('woff');
font-weight: 300;
font-style: normal;
font-display: swap;
}
@font-face {
font-family: 'Noto Sans';
src: url('file/css/NotoSans/NotoSans-ExtraLight.woff2') format('woff2'),
url('file/css/NotoSans/NotoSans-ExtraLight.woff') format('woff');
font-weight: 200;
font-style: normal;
font-display: swap;
}
@font-face {
font-family: 'Noto Sans';
src: url('file/css/NotoSans/NotoSans-Medium.woff2') format('woff2'),
url('file/css/NotoSans/NotoSans-Medium.woff') format('woff');
font-weight: 500;
font-style: normal;
font-display: swap;
}
@font-face {
font-family: 'Noto Sans';
src: url('file/css/NotoSans/NotoSans-Regular.woff2') format('woff2'),
url('file/css/NotoSans/NotoSans-Regular.woff') format('woff');
font-weight: normal;
font-style: normal;
font-display: swap;
}
@font-face {
font-family: 'Noto Sans';
src: url('file/css/NotoSans/NotoSans-MediumItalic.woff2') format('woff2'),
url('file/css/NotoSans/NotoSans-MediumItalic.woff') format('woff');
font-weight: 500;
font-style: italic;
font-display: swap;
}
@font-face {
font-family: 'Noto Sans';
src: url('file/css/NotoSans/NotoSans-SemiBoldItalic.woff2') format('woff2'),
url('file/css/NotoSans/NotoSans-SemiBoldItalic.woff') format('woff');
font-weight: 600;
font-style: italic;
font-display: swap;
}
@font-face {
font-family: 'Noto Sans';
src: url('file/css/NotoSans/NotoSans-SemiBold.woff2') format('woff2'),
url('file/css/NotoSans/NotoSans-SemiBold.woff') format('woff');
font-weight: 600;
font-style: normal;
font-display: swap;
}
@font-face {
font-family: 'Noto Sans';
src: url('file/css/NotoSans/NotoSans-Thin.woff2') format('woff2'),
url('file/css/NotoSans/NotoSans-Thin.woff') format('woff');
font-weight: 100;
font-style: normal;
font-display: swap;
}

131
css/chat_style-Dark.css Normal file
View file

@ -0,0 +1,131 @@
.message {
display: grid;
align-items: start;
grid-template-columns: 60px minmax(0, 1fr);
padding-bottom: 22px;
padding-top: 6px;
font-size: 18px;
font-family: Roboto, Arial, sans-serif; /* Modern font */
line-height: 1.5;
}
.circle-you,
.circle-bot {
background-color: #2b2b2b; /* Darker background for circles */
border-radius: 50%; /* Perfect circle */
border: 1px solid #4a90e2; /* Soft blue border */
box-shadow: 0 4px 8px rgb(0 0 0 / 50%); /* Soft shadow for depth */
}
.circle-bot img,
.circle-you img {
border-radius: 50%; /* Make images circular */
width: 100%;
height: 100%;
object-fit: cover;
}
.circle-you, .circle-bot {
width: 64px; /* Smaller size for modern look */
height: 64px;
}
.text {
padding-left: 12px; /* Reduced padding for a cleaner layout */
color: #f0f0f0; /* Light text color for readability */
}
.text p {
margin-top: 2px;
}
.username {
padding-left: 10px;
font-size: 20px;
font-weight: bold;
color: #e0e0e0; /* Light gray text */
transition: color 0.3s ease; /* Smooth color transition */
}
.username:hover {
color: #4a90e2; /* Blue color on hover */
}
.message-body {
position: relative;
border: 1px solid rgb(255 255 255 / 10%); /* Soft white border */
border-radius: 8px; /* Slightly rounded corners */
padding: 15px;
background: #1e1e1e; /* Dark background */
box-shadow: 0 4px 10px rgb(0 0 0 / 30%); /* Subtle shadow for depth */
transition: background 0.3s ease; /* Smooth transition for background */
}
.message-body:hover {
background: #252525; /* Slightly lighter on hover */
}
/* Adds 2 extra lines at the top and bottom of the message */
.message-body::before,
.message-body::after {
content: "";
position: absolute;
left: 10px;
right: 10px;
height: 1px;
background-color: rgb(255 255 255 / 5%); /* Faded lines for subtle separation */
}
.message-body::before {
top: 4px;
}
.message-body::after {
bottom: 4px;
}
.message-body img {
max-width: 300px;
max-height: 300px;
border-radius: 10px; /* Rounded corners for images */
}
.message-body p {
margin-bottom: 0 !important;
font-size: 16px !important;
line-height: 1.5 !important;
color: #e0e0e0 !important; /* Light color for text */
}
.message-body p em {
color: #a6a6a6 !important; /* Softer gray for emphasized text */
}
@media screen and (width <= 688px) {
.message {
display: grid;
align-items: start;
grid-template-columns: 60px minmax(0, 1fr);
padding-bottom: 25px;
font-size: 15px;
font-family: Roboto, Arial, sans-serif; /* Modern font */
line-height: 1.5;
}
.circle-you, .circle-bot {
width: 40px; /* Smaller size for mobile */
height: 40px;
}
.text {
padding-left: 10px; /* Reduced padding for mobile */
}
.message-body p {
font-size: 14px !important; /* Smaller text for mobile */
}
.username {
font-size: 18px; /* Smaller username for mobile */
}
}

View file

@ -0,0 +1,137 @@
/* All credits to TheEncrypted777: https://www.reddit.com/r/Oobabooga/comments/12xe6vq/updated_css_styling_with_color_customization_for/ */
.message {
display: grid;
align-items: start;
grid-template-columns: 60px minmax(0, 1fr);
padding-bottom: 21px;
padding-top: 7px;
font-size: 18px;
font-family: 'Noto Sans', Arial, sans-serif;
line-height: 1.428571429;
}
.circle-you,
.circle-bot {
background-color: gray;
border-radius: 1rem;
border: 2px solid white;
}
.circle-bot img,
.circle-you img {
border-radius: 10%;
width: 100%;
height: 100%;
object-fit: cover;
}
.circle-you, .circle-bot {
/* You can set the size of the profile images here, but if you do, you have to also adjust the .text{padding-left: 90px} to a different number according to the width of the image which is right below here */
width: 135px;
height: 175px;
}
.text {
/* Change this to move the message box further left or right depending on the size of your profile pic */
padding-left: 90px;
text-shadow: 2px 2px 2px rgb(0 0 0 / 40%);
}
.text p {
margin-top: 2px;
}
.username {
padding-left: 10px;
font-size: 22px;
font-weight: bold;
border-top: 1px solid rgb(51 64 90);
padding: 3px;
}
.message-body {
position: relative;
border: 1px solid rgb(255 255 255 / 45.9%);
border-radius: 10px;
padding: 10px;
padding-top: 5px;
/* Message gradient background color - remove the line bellow if you don't want a background color or gradient */
background: linear-gradient(to bottom, #171730, #1b263f);
}
/* Adds 2 extra lines at the top and bottom of the message */
.message-body::before,
.message-body::after {
content: "";
position: absolute;
left: 10px;
right: 10px;
height: 1px;
background-color: rgb(255 255 255 / 13%);
}
.message-body::before {
top: 6px;
}
.message-body::after {
bottom: 6px;
}
.message-body img {
max-width: 300px;
max-height: 300px;
border-radius: 20px;
}
.message-body p {
margin-bottom: 0 !important;
font-size: 18px !important;
line-height: 1.428571429 !important;
color: rgb(243 244 246) !important;
text-shadow: 2px 2px 2px rgb(0 0 0);
font-weight: 500;
}
.message-body p em {
color: rgb(138 138 138) !important;
}
@media screen and (width <= 688px) {
.message {
display: grid;
align-items: start;
grid-template-columns: 60px minmax(0, 1fr);
padding-bottom: 25px;
font-size: 15px;
font-family: 'Noto Sans', Helvetica, Arial, sans-serif;
line-height: 1.428571429;
}
.circle-you, .circle-bot {
width: 50px;
height: 73px;
border-radius: 0.5rem;
}
.circle-bot img,
.circle-you img {
width: 100%;
height: 100%;
object-fit: cover;
}
.text {
padding-left: 0;
}
.message-body p {
font-size: 16px !important;
}
.username {
font-size: 20px;
}
}

View file

@ -0,0 +1,22 @@
@import url("file/css/chat_style-cai-chat.css");
.circle-bot, .circle-you {
height: 90px;
width: 60px;
border-radius: 10px;
background-color: #656565;
}
.circle-bot img, .circle-you img {
border-radius: 8.333px;
}
.circle-you {
background-color: #656565;
}
.message {
padding-bottom: 1.5em;
padding-top: 0.5em;
grid-template-columns: 70px minmax(0, 1fr);
}

View file

@ -0,0 +1,66 @@
.message {
display: grid;
align-items: start;
grid-template-columns: 60px minmax(0, 1fr);
padding-bottom: 1.5em;
padding-top: 0.5em;
font-size: 15px;
font-family: 'Noto Sans', Helvetica, Arial, sans-serif;
line-height: 22.5px !important;
}
.message-body {
margin-top: 3px;
font-size: 15px !important;
}
.circle-you {
width: 50px;
height: 50px;
background-color: rgb(238 78 59);
border-radius: 50%;
}
.circle-bot {
width: 50px;
height: 50px;
background-color: rgb(59 78 244);
border-radius: 50%;
}
.circle-bot img,
.circle-you img {
border-radius: 50%;
width: 100%;
height: 100%;
object-fit: cover;
}
.username {
font-weight: bold;
}
.message-body img {
max-width: 300px;
max-height: 300px;
border-radius: 20px;
}
.message-body p {
font-size: 15px !important;
line-height: 22.5px !important;
font-weight: 500;
}
.message-body p, .chat .message-body ul, .chat .message-body ol {
margin-bottom: 10px !important;
}
.dark .message-body p em {
color: rgb(138 138 138) !important;
}
.message-body p em {
color: rgb(110 110 110) !important;
font-weight: 500;
}

View file

@ -0,0 +1,109 @@
.message {
padding-bottom: 22px;
padding-top: 3px;
font-size: 15px;
font-family: 'Noto Sans', Helvetica, Arial, sans-serif;
line-height: 1.428571429;
}
.circle-you {
width: 50px;
height: 50px;
background-color: rgb(238 78 59);
border-radius: 50%;
}
.circle-bot {
width: 50px;
height: 50px;
background-color: rgb(59 78 244);
border-radius: 50%;
float: left;
margin-right: 10px;
margin-top: 5px;
}
.circle-bot img,
.circle-you img {
border-radius: 50%;
width: 100%;
height: 100%;
object-fit: cover;
}
.circle-you {
margin-top: 5px;
float: right;
}
.circle-bot + .text, .circle-you + .text {
border-radius: 18px;
padding: 8px 12px;
}
.circle-bot + .text {
background-color: #E4E6EB;
float: left;
}
.circle-you + .text {
float: right;
background-color: rgb(0 132 255);
margin-right: 10px;
}
.circle-you + .text div, .circle-you + .text *, .dark .circle-you + .text div, .dark .circle-you + .text * {
color: #FFF !important;
}
.circle-you + .text .username {
text-align: right;
}
.dark .circle-bot + .text div, .dark .circle-bot + .text * {
color: #000;
}
.text {
max-width: 80%;
}
.text p {
margin-top: 5px;
}
.username {
font-weight: bold;
}
.message-body {
}
.message-body img {
max-width: 300px;
max-height: 300px;
border-radius: 20px;
}
.message-body p {
margin-bottom: 0 !important;
font-size: 15px !important;
line-height: 1.428571429 !important;
font-weight: 500;
}
.dark .message-body p em {
color: rgb(138 138 138) !important;
}
.message-body p em {
color: rgb(110 110 110) !important;
}
.editing-textarea {
width: max(30rem) !important;
}
.circle-you + .text .edit-control-button, .circle-you + .text .editing-textarea {
color: #000 !important;
}

105
css/chat_style-wpp.css Normal file
View file

@ -0,0 +1,105 @@
.message {
display: block;
padding-top: 0;
padding-bottom: 21px;
font-size: 15px;
font-family: 'Noto Sans', Helvetica, Arial, sans-serif;
line-height: 1.428571429;
grid-template-columns: none;
}
.circle-you, .circle-bot {
display: none;
}
.text {
max-width: 65%;
border-radius: 18px;
padding: 12px 16px;
margin-bottom: 8px;
clear: both;
box-shadow: 0 1px 2px rgb(0 0 0 / 10%);
}
.username {
font-weight: 600;
margin-bottom: 8px;
opacity: 0.65;
padding-left: 0;
}
/* User messages - right aligned, WhatsApp green */
.circle-you + .text {
background-color: #d9fdd3;
float: right;
margin-left: auto;
margin-right: 8px;
}
.circle-you + .text .username {
display: none;
}
/* Bot messages - left aligned, white */
.circle-bot + .text {
background-color: #fff;
float: left;
margin-right: auto;
margin-left: 8px;
border: 1px solid #e5e5e5;
}
.circle-bot + .text .message-actions {
bottom: -25px !important;
}
/* Dark theme colors */
.dark .circle-you + .text {
background-color: #144d37;
color: #e4e6ea;
box-shadow: 0 1px 2px rgb(0 0 0 / 30%);
}
.dark .circle-bot + .text {
background-color: #202c33;
color: #e4e6ea;
border: 1px solid #3c4043;
box-shadow: 0 1px 2px rgb(0 0 0 / 30%);
}
.dark .username {
opacity: 0.7;
}
.message-body img {
max-width: 300px;
max-height: 300px;
border-radius: 12px;
}
.message-body p {
font-size: 15px !important;
line-height: 1.4 !important;
font-weight: 400;
}
.message-body p:first-child {
margin-top: 0 !important;
}
.dark .message-body p em {
color: rgb(170 170 170) !important;
}
.message-body p em {
color: rgb(100 100 100) !important;
}
/* Message actions positioning */
.message-actions {
margin-top: 8px;
}
.message-body p, .chat .message-body ul, .chat .message-body ol {
margin-bottom: 10px !important;
}

111
css/highlightjs/github-dark.min.css vendored Normal file
View file

@ -0,0 +1,111 @@
html body gradio-app .gradio-container pre code.hljs {
display: block;
overflow-x: auto;
padding: 1em
}
html body gradio-app .gradio-container code.hljs {
padding: 3px 5px
}
/*!
Theme: GitHub Dark
Description: Dark theme as seen on github.com
Author: github.com
Maintainer: @Hirse
Updated: 2021-05-15
Outdated base version: https://github.com/primer/github-syntax-dark
Current colors taken from GitHub's CSS
*/
html body gradio-app .gradio-container .hljs {
color: #c9d1d9;
background: #0d1117
}
html body gradio-app .gradio-container .hljs-doctag,
html body gradio-app .gradio-container .hljs-keyword,
html body gradio-app .gradio-container .hljs-meta .hljs-keyword,
html body gradio-app .gradio-container .hljs-template-tag,
html body gradio-app .gradio-container .hljs-template-variable,
html body gradio-app .gradio-container .hljs-type,
html body gradio-app .gradio-container .hljs-variable.language_ {
color: #ff7b72
}
html body gradio-app .gradio-container .hljs-title,
html body gradio-app .gradio-container .hljs-title.class_,
html body gradio-app .gradio-container .hljs-title.class_.inherited__,
html body gradio-app .gradio-container .hljs-title.function_ {
color: #d2a8ff
}
html body gradio-app .gradio-container .hljs-attr,
html body gradio-app .gradio-container .hljs-attribute,
html body gradio-app .gradio-container .hljs-literal,
html body gradio-app .gradio-container .hljs-meta,
html body gradio-app .gradio-container .hljs-number,
html body gradio-app .gradio-container .hljs-operator,
html body gradio-app .gradio-container .hljs-selector-attr,
html body gradio-app .gradio-container .hljs-selector-class,
html body gradio-app .gradio-container .hljs-selector-id,
html body gradio-app .gradio-container .hljs-variable {
color: #79c0ff
}
html body gradio-app .gradio-container .hljs-meta .hljs-string,
html body gradio-app .gradio-container .hljs-regexp,
html body gradio-app .gradio-container .hljs-string {
color: #a5d6ff
}
html body gradio-app .gradio-container .hljs-built_in,
html body gradio-app .gradio-container .hljs-symbol {
color: #ffa657
}
html body gradio-app .gradio-container .hljs-code,
html body gradio-app .gradio-container .hljs-comment,
html body gradio-app .gradio-container .hljs-formula {
color: #8b949e
}
html body gradio-app .gradio-container .hljs-name,
html body gradio-app .gradio-container .hljs-quote,
html body gradio-app .gradio-container .hljs-selector-pseudo,
html body gradio-app .gradio-container .hljs-selector-tag {
color: #7ee787
}
html body gradio-app .gradio-container .hljs-subst {
color: #c9d1d9
}
html body gradio-app .gradio-container .hljs-section {
color: #1f6feb;
font-weight: 700
}
html body gradio-app .gradio-container .hljs-bullet {
color: #f2cc60
}
html body gradio-app .gradio-container .hljs-emphasis {
color: #c9d1d9;
font-style: italic
}
html body gradio-app .gradio-container .hljs-strong {
color: #c9d1d9;
font-weight: 700
}
html body gradio-app .gradio-container .hljs-addition {
color: #aff5b4;
background-color: #033a16
}
html body gradio-app .gradio-container .hljs-deletion {
color: #ffdcd7;
background-color: #67060c
}

111
css/highlightjs/github.min.css vendored Normal file
View file

@ -0,0 +1,111 @@
html body gradio-app .gradio-container pre code.hljs {
display: block;
overflow-x: auto;
padding: 1em
}
html body gradio-app .gradio-container code.hljs {
padding: 3px 5px
}
/*!
Theme: GitHub
Description: Light theme as seen on github.com
Author: github.com
Maintainer: @Hirse
Updated: 2021-05-15
Outdated base version: https://github.com/primer/github-syntax-light
Current colors taken from GitHub's CSS
*/
html body gradio-app .gradio-container .hljs {
color: #24292e;
background: #fff
}
html body gradio-app .gradio-container .hljs-doctag,
html body gradio-app .gradio-container .hljs-keyword,
html body gradio-app .gradio-container .hljs-meta .hljs-keyword,
html body gradio-app .gradio-container .hljs-template-tag,
html body gradio-app .gradio-container .hljs-template-variable,
html body gradio-app .gradio-container .hljs-type,
html body gradio-app .gradio-container .hljs-variable.language_ {
color: #d73a49
}
html body gradio-app .gradio-container .hljs-title,
html body gradio-app .gradio-container .hljs-title.class_,
html body gradio-app .gradio-container .hljs-title.class_.inherited__,
html body gradio-app .gradio-container .hljs-title.function_ {
color: #6f42c1
}
html body gradio-app .gradio-container .hljs-attr,
html body gradio-app .gradio-container .hljs-attribute,
html body gradio-app .gradio-container .hljs-literal,
html body gradio-app .gradio-container .hljs-meta,
html body gradio-app .gradio-container .hljs-number,
html body gradio-app .gradio-container .hljs-operator,
html body gradio-app .gradio-container .hljs-selector-attr,
html body gradio-app .gradio-container .hljs-selector-class,
html body gradio-app .gradio-container .hljs-selector-id,
html body gradio-app .gradio-container .hljs-variable {
color: #005cc5
}
html body gradio-app .gradio-container .hljs-meta .hljs-string,
html body gradio-app .gradio-container .hljs-regexp,
html body gradio-app .gradio-container .hljs-string {
color: #032f62
}
html body gradio-app .gradio-container .hljs-built_in,
html body gradio-app .gradio-container .hljs-symbol {
color: #e36209
}
html body gradio-app .gradio-container .hljs-code,
html body gradio-app .gradio-container .hljs-comment,
html body gradio-app .gradio-container .hljs-formula {
color: #6a737d
}
html body gradio-app .gradio-container .hljs-name,
html body gradio-app .gradio-container .hljs-quote,
html body gradio-app .gradio-container .hljs-selector-pseudo,
html body gradio-app .gradio-container .hljs-selector-tag {
color: #22863a
}
html body gradio-app .gradio-container .hljs-subst {
color: #24292e
}
html body gradio-app .gradio-container .hljs-section {
color: #005cc5;
font-weight: 700
}
html body gradio-app .gradio-container .hljs-bullet {
color: #735c0f
}
html body gradio-app .gradio-container .hljs-emphasis {
color: #24292e;
font-style: italic
}
html body gradio-app .gradio-container .hljs-strong {
color: #24292e;
font-weight: 700
}
html body gradio-app .gradio-container .hljs-addition {
color: #22863a;
background-color: #f0fff4
}
html body gradio-app .gradio-container .hljs-deletion {
color: #b31d28;
background-color: #ffeef0
}

View file

@ -0,0 +1 @@
.hljs-copy-wrapper{position:relative;overflow:hidden}.hljs-copy-wrapper:hover .hljs-copy-button,.hljs-copy-button:focus{transform:translateX(0)}.hljs-copy-button{position:absolute;transform:translateX(calc(100% + 1.125em));top:1em;right:1em;width:2rem;height:2rem;text-indent:-9999px;color:#fff;border-radius:.25rem;border:1px solid #ffffff22;background-color:#2d2b57;background-color:var(--hljs-theme-background);background-image:url('data:image/svg+xml;utf-8,<svg width="16" height="16" viewBox="0 0 24 24" fill="none" xmlns="http://www.w3.org/2000/svg"><path fill-rule="evenodd" clip-rule="evenodd" d="M6 5C5.73478 5 5.48043 5.10536 5.29289 5.29289C5.10536 5.48043 5 5.73478 5 6V20C5 20.2652 5.10536 20.5196 5.29289 20.7071C5.48043 20.8946 5.73478 21 6 21H18C18.2652 21 18.5196 20.8946 18.7071 20.7071C18.8946 20.5196 19 20.2652 19 20V6C19 5.73478 18.8946 5.48043 18.7071 5.29289C18.5196 5.10536 18.2652 5 18 5H16C15.4477 5 15 4.55228 15 4C15 3.44772 15.4477 3 16 3H18C18.7956 3 19.5587 3.31607 20.1213 3.87868C20.6839 4.44129 21 5.20435 21 6V20C21 20.7957 20.6839 21.5587 20.1213 22.1213C19.5587 22.6839 18.7957 23 18 23H6C5.20435 23 4.44129 22.6839 3.87868 22.1213C3.31607 21.5587 3 20.7957 3 20V6C3 5.20435 3.31607 4.44129 3.87868 3.87868C4.44129 3.31607 5.20435 3 6 3H8C8.55228 3 9 3.44772 9 4C9 4.55228 8.55228 5 8 5H6Z" fill="white"/><path fill-rule="evenodd" clip-rule="evenodd" d="M7 3C7 1.89543 7.89543 1 9 1H15C16.1046 1 17 1.89543 17 3V5C17 6.10457 16.1046 7 15 7H9C7.89543 7 7 6.10457 7 5V3ZM15 3H9V5H15V3Z" fill="white"/></svg>');background-repeat:no-repeat;background-position:center;transition:background-color 200ms ease,transform 200ms ease-out}.hljs-copy-button:hover{border-color:#ffffff44}.hljs-copy-button:active{border-color:#ffffff66}.hljs-copy-button[data-copied="true"]{text-indent:0;width:auto;background-image:none}@media(prefers-reduced-motion){.hljs-copy-button{transition:none}}.hljs-copy-alert{clip:rect(0 0 0 0);clip-path:inset(50%);height:1px;overflow:hidden;position:absolute;white-space:nowrap;width:1px}

View file

@ -0,0 +1,97 @@
.chat {
background: transparent;
padding: 0;
padding-top: 0;
}
.chat > .messages:first-child {
padding-top: 0 !important;
}
.chat .message-body p, .chat .message-body li {
font-size: 1rem !important;
line-height: 28px !important;
}
.dark .chat .message-body :is(p,li,h1,h2,h3,h4,h5,h6),
.dark .chat .message-body em:not(:is(h1,h2,h3,h4,h5,h6,b,strong) em),
.dark .chat .message-body q:not(:is(h1,h2,h3,h4,h5,h6,b,strong) q) {
color: #d1d5db !important;
}
.chat .message-body :is(th, td),
.prose hr {
border-color: #40404096 !important;
}
.dark .chat .message-body :is(th, td),
.dark .prose hr {
border-color: rgb(255 255 255 / 30%) !important;
}
.chat .message-body :is(p, ul, ol) {
margin: 1.25em 0 !important;
}
.chat .message-body :is(p, ul, ol):first-child {
margin-top: 0 !important;
}
.chat .message-body :is(p, ul, ol):last-child {
margin-bottom: 0 !important;
}
.user-message, .assistant-message {
font-family: Inter, Helvetica, Arial, sans-serif;
}
.message:first-child {
padding-top: 0;
}
.username {
display: none;
}
.chat .user-message {
background: #f3f4f6;
padding: 1.5rem 1rem;
padding-bottom: 2rem;
border-radius: 0;
border-bottom-right-radius: 0;
}
.chat .assistant-message {
padding: 1.5rem 1rem;
padding-bottom: 2rem;
border-radius: 0;
border: 0;
}
.dark .chat .user-message {
background: var(--light-gray);
}
.dark .chat .assistant-message {
background: transparent;
}
.chat .user-message .text,
.chat .assistant-message .text {
max-width: 700px;
margin-left: auto;
margin-right: auto;
}
/* Create space between two assistant messages in a row */
.assistant-message + .assistant-message {
margin-top: 1.5rem;
}
pre > code {
background-color: #f3f4f6 !important;
}
.dark pre > code {
background-color: #1f2937 !important;
}

View file

@ -0,0 +1,33 @@
.readable-container {
max-width: 600px;
margin-left: auto;
margin-right: auto;
background-color: rgb(31 41 55);
padding: 3em;
word-break: break-word;
overflow-wrap: anywhere;
color: #efefef !important;
}
.readable-container p, .readable-container li {
font-size: 16px !important;
color: #efefef !important;
margin-bottom: 22px;
line-height: 1.4 !important;
}
.readable-container li > p {
display: inline !important;
}
.readable-container code {
overflow-x: auto;
}
.readable-container :not(pre) > code {
white-space: normal !important;
}
.readable-container .hoverable {
font-size: 14px;
}

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Binary file not shown.

Some files were not shown because too many files have changed in this diff Show more