Commit graph

96 commits

Author SHA1 Message Date
oobabooga 702fe92d42 Increase truncation_length_max value 2023-04-19 17:35:38 -03:00
oobabooga ac2973ffc6 Add a warning for --share 2023-04-17 19:34:28 -03:00
oobabooga 89bc540557 Update README 2023-04-17 10:55:35 -03:00
sgsdxzy b57ffc2ec9
Update to support GPTQ triton commit c90adef (#1229) 2023-04-17 01:11:18 -03:00
oobabooga 39099663a0
Add 4-bit LoRA support (#1200) 2023-04-16 23:26:52 -03:00
Forkoz c6fe1ced01
Add ChatGLM support (#1256)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-04-16 19:15:03 -03:00
oobabooga b937c9d8c2
Add skip_special_tokens checkbox for Dolly model (#1218) 2023-04-16 14:24:49 -03:00
Mikel Bober-Irizar 16a3a5b039
Merge pull request from GHSA-hv5m-3rp9-xcpf
* Remove eval of API input

* Remove unnecessary eval/exec for security

* Use ast.literal_eval

* Use ast.literal_eval

---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-04-16 01:36:50 -03:00
oobabooga 3a337cfded Use argparse defaults 2023-04-14 15:35:06 -03:00
Alex "mcmonkey" Goodwin 64e3b44e0f
initial multi-lora support (#1103)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-04-14 14:52:06 -03:00
oobabooga 8e31f2bad4
Automatically set wbits/groupsize/instruct based on model name (#1167) 2023-04-14 11:07:28 -03:00
v0xie 9d66957207
Add --listen-host launch option (#1122) 2023-04-13 21:35:08 -03:00
Light cf58058c33 Change warmup_autotune to a negative switch. 2023-04-13 20:59:49 +08:00
Light 15d5a043f2 Merge remote-tracking branch 'origin/main' into triton 2023-04-13 19:38:51 +08:00
oobabooga 7dfbe54f42 Add --model-menu option 2023-04-12 21:24:26 -03:00
oobabooga 388038fb8e Update settings-template.json 2023-04-12 18:30:43 -03:00
oobabooga 1566d8e344 Add model settings to the Models tab 2023-04-12 17:20:18 -03:00
Light f3591ccfa1 Keep minimal change. 2023-04-12 23:26:06 +08:00
oobabooga cacbcda208
Two new options: truncation length and ban eos token 2023-04-11 18:46:06 -03:00
catalpaaa 78bbc66fc4
allow custom stopping strings in all modes (#903) 2023-04-11 12:30:06 -03:00
IggoOnCode 09d8119e3c
Add CPU LoRA training (#938)
(It's very slow)
2023-04-10 17:29:00 -03:00
oobabooga bd04ff27ad Make the bos token optional 2023-04-10 16:44:22 -03:00
oobabooga 0f1627eff1 Don't treat Intruct mode histories as regular histories
* They must now be saved/loaded manually
* Also improved browser caching of pfps
* Also changed the global default preset
2023-04-10 15:48:07 -03:00
MarkovInequality 992663fa20
Added xformers support to Llama (#950) 2023-04-09 23:08:40 -03:00
oobabooga ea6e77df72
Make the code more like PEP8 for readability (#862) 2023-04-07 00:15:45 -03:00
SDS 378d21e80c
Add LLaMA-Precise preset (#767) 2023-04-05 18:52:36 -03:00
oobabooga e722c240af Add Instruct mode 2023-04-05 13:54:50 -03:00
oobabooga 65d8a24a6d Show profile pictures in the Character tab 2023-04-04 22:28:49 -03:00
oobabooga b24147c7ca Document --pre_layer 2023-04-03 17:34:25 -03:00
oobabooga 4c9ed09270 Update settings template 2023-04-03 14:59:26 -03:00
OWKenobi dcf61a8897
"character greeting" displayed and editable on the fly (#743)
* Add greetings field

* add greeting field and make it interactive

* Minor changes

* Fix a bug

* Simplify clear_chat_log

* Change a label

* Minor change

* Simplifications

* Simplification

* Simplify loading the default character history

* Fix regression

---------

Co-authored-by: oobabooga
2023-04-03 12:16:15 -03:00
oobabooga b0890a7925 Add shared.is_chat() function 2023-04-01 20:15:00 -03:00
oobabooga b857f4655b
Update shared.py 2023-04-01 13:56:47 -03:00
oobabooga 2c52310642 Add --threads flag for llama.cpp 2023-03-31 21:18:05 -03:00
oobabooga 1d1d9e40cd Add seed to settings 2023-03-31 12:22:07 -03:00
oobabooga d4a9b5ea97 Remove redundant preset (see the plot in #587) 2023-03-30 17:34:44 -03:00
oobabooga 55755e27b9 Don't hardcode prompts in the settings dict/json 2023-03-29 22:47:01 -03:00
oobabooga 1cb9246160 Adapt to the new model names 2023-03-29 21:47:36 -03:00
oobabooga 010b259dde Update documentation 2023-03-28 17:46:00 -03:00
oobabooga 036163a751 Change description 2023-03-27 23:39:26 -03:00
oobabooga 005f552ea3 Some simplifications 2023-03-27 23:29:52 -03:00
oobabooga fde92048af Merge branch 'main' into catalpaaa-lora-and-model-dir 2023-03-27 23:16:44 -03:00
oobabooga 49c10c5570
Add support for the latest GPTQ models with group-size (#530)
**Warning: old 4-bit weights will not work anymore!**

See here how to get up to date weights: https://github.com/oobabooga/text-generation-webui/wiki/LLaMA-model#step-2-get-the-pre-converted-weights
2023-03-26 00:11:33 -03:00
catalpaaa b37c54edcf lora-dir, model-dir and login auth
Added lora-dir, model-dir, and a login auth arguments that points to a file contains usernames and passwords in the format of "u:pw,u:pw,..."
2023-03-24 17:30:18 -07:00
oobabooga c5ebcc5f7e
Change the default names (#518)
* Update shared.py

* Update settings-template.json
2023-03-23 13:36:00 -03:00
oobabooga 7618f3fe8c
Add -gptq-preload for 4-bit offloading (#460)
This works in a 4GB card now:

```
python server.py --model llama-7b-hf --gptq-bits 4 --gptq-pre-layer 20
```
2023-03-20 16:30:56 -03:00
oobabooga ddb62470e9 --no-cache and --gpu-memory in MiB for fine VRAM control 2023-03-19 19:21:41 -03:00
oobabooga a78b6508fc Make custom LoRAs work by default #385 2023-03-19 12:11:35 -03:00
oobabooga c753261338 Disable stop_at_newline by default 2023-03-18 10:55:57 -03:00
oobabooga 29fe7b1c74 Remove LoRA tab, move it into the Parameters menu 2023-03-17 11:39:48 -03:00