Commit graph

488 commits

Author SHA1 Message Date
USBhost e1aa9d5173
Support upstream GPTQ once again. (#1451) 2023-04-21 12:43:56 -03:00
oobabooga eddd016449 Minor deletion 2023-04-21 12:41:27 -03:00
oobabooga d46b9b7c50 Fix evaluate comment saving 2023-04-21 12:34:08 -03:00
oobabooga 5e023ae64d Change dropdown menu highlight color 2023-04-21 02:47:18 -03:00
oobabooga c4f4f41389
Add an "Evaluate" tab to calculate the perplexities of models (#1322) 2023-04-21 00:20:33 -03:00
oobabooga 7bb9036ac9 Add universal LLaMA tokenizer support 2023-04-19 21:23:51 -03:00
Alex "mcmonkey" Goodwin ee30625cd1
4-Bit LoRA training + several new training options and fixes 2023-04-19 19:39:03 -03:00
oobabooga 702fe92d42 Increase truncation_length_max value 2023-04-19 17:35:38 -03:00
oobabooga 9d9ae62938 Fix stopping strings in the gradio API 2023-04-19 13:52:21 -03:00
oobabooga 649e4017a5 Style improvements 2023-04-19 00:36:28 -03:00
oobabooga 000f65a2ef
Delete unused file 2023-04-18 04:01:14 -03:00
oobabooga 36f7c022f2
Rename a file 2023-04-18 01:38:33 -03:00
oobabooga b069bb1f2e
Update monkey_patch_gradio.py 2023-04-18 01:32:42 -03:00
oobabooga 00186f76f4
Monkey patch gradio to prevent it from calling home 2023-04-18 01:13:16 -03:00
Tynan Burke 6a810b16b2
typo in training.py (#1329) 2023-04-17 21:40:46 -03:00
oobabooga ac2973ffc6 Add a warning for --share 2023-04-17 19:34:28 -03:00
oobabooga c544386824 Reset your name when choosing a character 2023-04-17 13:56:40 -03:00
oobabooga c3dc348d1c Don't show 'None' in the LoRA list 2023-04-17 13:52:23 -03:00
oobabooga 89bc540557 Update README 2023-04-17 10:55:35 -03:00
catalpaaa 07de7d0426
Load llamacpp before quantized model (#1307) 2023-04-17 10:47:26 -03:00
sgsdxzy b57ffc2ec9
Update to support GPTQ triton commit c90adef (#1229) 2023-04-17 01:11:18 -03:00
oobabooga 39099663a0
Add 4-bit LoRA support (#1200) 2023-04-16 23:26:52 -03:00
oobabooga 46a8aa8c09 Readability 2023-04-16 21:26:19 -03:00
Forkoz c6fe1ced01
Add ChatGLM support (#1256)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-04-16 19:15:03 -03:00
oobabooga 6a03ad0824 Remove fix_newlines() calls from chat.py 2023-04-16 18:25:44 -03:00
oobabooga 5342f72968 Properly handle blockquote blocks 2023-04-16 18:00:12 -03:00
oobabooga 27f3a78834 Better detect when no model is loaded 2023-04-16 17:35:54 -03:00
oobabooga c8ad960018 Add defaults to the gradio API 2023-04-16 17:33:28 -03:00
oobabooga beb95f5fe2 Add a style for the "chat" mode 2023-04-16 16:44:50 -03:00
oobabooga b937c9d8c2
Add skip_special_tokens checkbox for Dolly model (#1218) 2023-04-16 14:24:49 -03:00
oobabooga b705b4210c Minor changes to training.py 2023-04-16 03:08:37 -03:00
oobabooga 5c513a5f5c Make training.py more readable 2023-04-16 02:46:27 -03:00
Alex "mcmonkey" Goodwin a3eec62b50
Lora trainer improvements part 3 (#1098)
* add support for other model types

dependent on future-peft-changes but with fallback to function now

* use encoding=utf8 for training format

* make shuffling optional

and describe dropout a bit more

* add eval_steps to control evaluation

* make callbacks not depend on globals

* make save steps controllable

* placeholder of initial loading-existing-model support

and var name cleanup

* save/load parameters

* last bit of cleanup

* remove `gptq_bits` ref as main branch removed that setting

* add higher_rank_limit option

2048 is basically unreachable due to VRAM, but i trained at 1536 with batch size = 1 on a 7B model.
Note that it's in the do_train input just to save as a parameter

* fix math on save_steps
2023-04-16 02:35:13 -03:00
kernyan ac19d5101f
revert incorrect eos_token_id change from #814 (#1261)
- fixes #1054
2023-04-16 01:47:01 -03:00
oobabooga a2127239de Fix a bug 2023-04-16 01:41:37 -03:00
oobabooga 9d3c6d2dc3 Fix a bug 2023-04-16 01:40:47 -03:00
Mikel Bober-Irizar 16a3a5b039
Merge pull request from GHSA-hv5m-3rp9-xcpf
* Remove eval of API input

* Remove unnecessary eval/exec for security

* Use ast.literal_eval

* Use ast.literal_eval

---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-04-16 01:36:50 -03:00
oobabooga d2ea925fa5 Bump llama-cpp-python to use LlamaCache 2023-04-16 00:53:40 -03:00
oobabooga ac189011cb Add "Save current settings for this model" button 2023-04-15 12:54:02 -03:00
oobabooga abef355ed0 Remove deprecated flag 2023-04-15 01:21:19 -03:00
oobabooga c3aa79118e Minor generate_chat_prompt simplification 2023-04-14 23:02:08 -03:00
oobabooga 3a337cfded Use argparse defaults 2023-04-14 15:35:06 -03:00
Alex "mcmonkey" Goodwin 64e3b44e0f
initial multi-lora support (#1103)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-04-14 14:52:06 -03:00
oobabooga 1901d238e1 Minor change to API code 2023-04-14 12:11:47 -03:00
oobabooga 8e31f2bad4
Automatically set wbits/groupsize/instruct based on model name (#1167) 2023-04-14 11:07:28 -03:00
v0xie 9d66957207
Add --listen-host launch option (#1122) 2023-04-13 21:35:08 -03:00
oobabooga a75e02de4d Simplify GPTQ_loader.py 2023-04-13 12:13:07 -03:00
oobabooga ca293bb713 Show a warning if two quantized models are found 2023-04-13 12:04:27 -03:00
oobabooga 8b482b4127
Merge #1073 from sgsdxzy/triton
* Multi-GPU support for triton
* Better quantized model filename detection
2023-04-13 11:31:21 -03:00
oobabooga fde6d06167 Prioritize names with the groupsize in them 2023-04-13 11:27:03 -03:00