Commit graph

797 commits

Author SHA1 Message Date
Ayanami Rei edbc61139f use new quant loader 2023-03-13 20:00:38 +03:00
Ayanami Rei 345b6dee8c refactor quant models loader and add support of OPT 2023-03-13 19:59:57 +03:00
oobabooga 2c4699a7e9 Change a comment 2023-03-13 00:20:02 -03:00
oobabooga 0a7acb3bd9 Remove redundant comments 2023-03-13 00:12:21 -03:00
oobabooga 77294b27dd Use str(Path) instead of os.path.abspath(Path) 2023-03-13 00:08:01 -03:00
oobabooga b9e0712b92 Fix Open Assistant 2023-03-12 23:58:25 -03:00
oobabooga 1ddcd4d0ba Clean up silero_tts
This should only be used with --no-stream.

The shared.still_streaming implementation was faulty by design:
output_modifier should never be called when streaming is already over.
2023-03-12 23:42:49 -03:00
oobabooga a95592fc56 Add back a progress indicator to --no-stream 2023-03-12 20:38:40 -03:00
oobabooga d168b6e1f7
Update README.md 2023-03-12 17:54:07 -03:00
oobabooga 54e8f0c31f
Update README.md 2023-03-12 16:58:00 -03:00
oobabooga cebe8b390d Remove useless "substring_found" variable 2023-03-12 15:50:38 -03:00
oobabooga 4bcd675ccd Add *Is typing...* to regenerate as well 2023-03-12 15:23:33 -03:00
oobabooga 3375eaece0 Update README 2023-03-12 15:01:32 -03:00
oobabooga c7aa51faa6 Use a list of eos_tokens instead of just a number
This might be the cause of LLaMA ramblings that some people have experienced.
2023-03-12 14:54:58 -03:00
oobabooga 17210ff88f
Update README.md 2023-03-12 14:31:24 -03:00
oobabooga 441e993c51 Bump accelerate, RWKV and safetensors 2023-03-12 14:25:14 -03:00
oobabooga d8bea766d7
Merge pull request #192 from xanthousm/main
Add text generation stream status to shared module, use for better TTS with auto-play
2023-03-12 13:40:16 -03:00
oobabooga 4066ab4c0c Reorder the imports 2023-03-12 13:36:18 -03:00
oobabooga 4dc1d8c091
Update README.md 2023-03-12 12:46:53 -03:00
oobabooga 901dcba9b4
Merge pull request #263 from HideLord/main
Fixing compatibility with GPTQ repository
2023-03-12 12:42:08 -03:00
oobabooga fda376d9c3 Use os.path.abspath() instead of str() 2023-03-12 12:41:04 -03:00
HideLord 8403152257 Fixing compatibility with GPTQ repo commit 2f667f7da051967566a5fb0546f8614bcd3a1ccd. Expects string and breaks on 2023-03-12 17:28:15 +02:00
HideLord a27f98dbbc Merge branch 'main' of https://github.com/HideLord/text-generation-webui 2023-03-12 16:51:04 +02:00
oobabooga f3b00dd165
Merge pull request #224 from ItsLogic/llama-bits
Allow users to load 2, 3 and 4 bit llama models
2023-03-12 11:23:50 -03:00
oobabooga 89e9493509 Update README 2023-03-12 11:23:20 -03:00
oobabooga 65dda28c9d Rename --llama-bits to --gptq-bits 2023-03-12 11:19:07 -03:00
oobabooga fed3617f07 Move LLaMA 4-bit into a separate file 2023-03-12 11:12:34 -03:00
oobabooga 0ac562bdba Add a default prompt for OpenAssistant oasst-sft-1-pythia-12b #253 2023-03-12 10:46:16 -03:00
oobabooga 78901d522b Remove unused imports 2023-03-12 08:59:05 -03:00
oobabooga 35c14f31b2
Merge pull request #259 from hieultp/patch-1
Fix typo error in LLaMa prompts
2023-03-12 08:52:02 -03:00
oobabooga 3c25557ef0 Add tqdm to requirements.txt 2023-03-12 08:48:16 -03:00
Phuoc-Hieu Le 781c09235c
Fix typo error in script.py 2023-03-12 15:21:50 +07:00
Xan 9276af3561 clean up 2023-03-12 19:06:24 +11:00
Xan b3e10e47c0 Fix merge conflict in text_generation
- Need to update `shared.still_streaming = False` before the final `yield formatted_outputs`, shifted the position of some yields.
2023-03-12 18:56:35 +11:00
Xan d4afed4e44 Fixes and polish
- Change wav naming to be completely unique using timestamp instead of message ID, stops browser using cached audio when new audio is made with the same file name (eg after regenerate or clear history).
- Make the autoplay setting actually disable autoplay.
- Make Settings panel a bit more compact.
- Hide html errors when audio file of chat history is missing.
- Add button to permanently convert TTS history to normal text messages
- Changed the "show message text" toggle to affect the chat history.
2023-03-12 17:56:57 +11:00
oobabooga ad14f0e499 Fix regenerate (provisory way) 2023-03-12 03:42:29 -03:00
oobabooga 6e12068ba2
Merge pull request #258 from lxe/lxe/utf8
Load and save character files and chat history in UTF-8
2023-03-12 03:28:49 -03:00
oobabooga e2da6b9685 Fix You You You appearing in chat mode 2023-03-12 03:25:56 -03:00
oobabooga bcf0075278
Merge pull request #235 from xanthousm/Quality_of_life-main
--auto-launch and "Is typing..."
2023-03-12 03:12:56 -03:00
Aleksey Smolenchuk 3f7c3d6559
No need to set encoding on binary read 2023-03-11 22:10:57 -08:00
oobabooga 3437de686c
Merge pull request #189 from oobabooga/new-streaming
New streaming method (much faster)
2023-03-12 03:01:26 -03:00
oobabooga 341e135036 Various fixes in chat mode 2023-03-12 02:53:08 -03:00
Aleksey Smolenchuk 3baf5fc700
Load and save chat history in utf-8 2023-03-11 21:40:01 -08:00
oobabooga b0e8cb8c88 Various fixes in chat mode 2023-03-12 02:31:45 -03:00
unknown 433f6350bc Load and save character files in UTF-8 2023-03-11 21:23:05 -08:00
oobabooga 0bd5430988 Use 'with' statement to better handle streaming memory 2023-03-12 02:04:28 -03:00
oobabooga 37f0166b2d Fix memory leak in new streaming (second attempt) 2023-03-11 23:14:49 -03:00
HideLord def97f658c Small patch to fix loading of character jsons. Now it correctly reads non-ascii characters on Windows. 2023-03-12 02:54:22 +02:00
oobabooga 92fe947721 Merge branch 'main' into new-streaming 2023-03-11 19:59:45 -03:00
oobabooga 195e99d0b6 Add llama_prompts extension 2023-03-11 16:11:15 -03:00