Commit graph

209 commits

Author SHA1 Message Date
oobabooga 990ee54ddd Move the example dialogue to the chat history, and keep it hidden.
This greatly improves the performance of text generation, as
histories can be quite long. It also makes more sense to implement
it this way.
2023-01-21 02:48:06 -03:00
oobabooga 3f2c1e7170
Merge pull request #16 from 81300/model-download
Allow specifying the Hugging Face Git branch when downloading models
2023-01-21 00:43:35 -03:00
oobabooga 1e541d4882
Update download-model.py 2023-01-21 00:43:00 -03:00
oobabooga 18ef72d7c0
Update download-model.py 2023-01-21 00:38:23 -03:00
oobabooga d7299df01f Rename parameters 2023-01-21 00:33:41 -03:00
oobabooga 86a2832f3b
Merge pull request #17 from Silver267/main
Add --disk-cache-dir and --max-cpu-mem parameters (Properly pushed to github)
2023-01-21 00:26:19 -03:00
oobabooga 5df03bf0fd
Merge branch 'main' into main 2023-01-21 00:25:34 -03:00
oobabooga faaafe7c0e Better parameter naming 2023-01-20 23:45:16 -03:00
Silver267 f4634e4c32 Update. 2023-01-20 17:05:43 -05:00
81300 fffd49e64e
Add --branch option to the model download script 2023-01-20 22:51:56 +02:00
oobabooga c0f2367b54 Minor fix 2023-01-20 17:09:25 -03:00
oobabooga 185587a33e Add a history size parameter to the chat
If too many messages are used in the prompt, the model
gets really slow. It is useful to have the ability to
limit this.
2023-01-20 17:03:09 -03:00
oobabooga 4067cecf67 Bump bitsandbytes version 2023-01-20 12:51:49 -03:00
oobabooga 8f3deec759 Prevent the history from being altered by the html script 2023-01-20 01:59:51 -03:00
oobabooga 78d5a999e6 Improve prompt formatation 2023-01-20 01:54:38 -03:00
oobabooga 70ff685736 Encode the input string correctly 2023-01-20 00:45:02 -03:00
oobabooga 83584ae2d7 Clearer installation instructions 2023-01-20 00:20:35 -03:00
oobabooga b66d18d5a0 Allow presets/characters with '.' in their names 2023-01-19 21:56:33 -03:00
oobabooga c4f7a874d5 Fix the regex... 2023-01-19 21:16:11 -03:00
oobabooga 8d4170826f Update README 2023-01-19 21:08:26 -03:00
oobabooga 11c3214981 Fix some regexes 2023-01-19 19:59:34 -03:00
oobabooga e61138bdad Minor fixes 2023-01-19 19:04:54 -03:00
oobabooga 2181fca709 Better defaults for chat 2023-01-19 18:58:45 -03:00
oobabooga cd7b07239f Add Colab guide 2023-01-19 17:58:04 -03:00
oobabooga 83808171d3 Add --share option for Colab 2023-01-19 17:31:29 -03:00
oobabooga b054367be2 Update README 2023-01-19 16:54:58 -03:00
oobabooga 8d788874d7 Add support for characters 2023-01-19 16:46:46 -03:00
oobabooga 3121f4788e Fix uploading chat log in --chat mode 2023-01-19 15:05:42 -03:00
oobabooga 849e4c7f90 Better way of finding the generated reply in the output string 2023-01-19 14:57:01 -03:00
oobabooga d03b0ad7a8 Implement saving/loading chat logs (#9) 2023-01-19 14:03:47 -03:00
oobabooga 39bfea5a22 Add a progress bar 2023-01-19 12:20:57 -03:00
oobabooga 5390fc87c8 add auto-devices when disk is used 2023-01-19 12:11:44 -03:00
oobabooga 759da435e3 Release 8-bit models memory 2023-01-19 12:03:16 -03:00
oobabooga f9faad4cfa Add low VRAM guide 2023-01-19 11:25:17 -03:00
oobabooga 7ace04864a Implement sending layers to disk with --disk (#10) 2023-01-19 11:09:24 -03:00
oobabooga 1ce95ee817 Mention text streaming 2023-01-19 10:46:41 -03:00
oobabooga 93fa9bbe01 Clean up the streaming implementation 2023-01-19 10:43:05 -03:00
oobabooga c90310e40e Small simplification 2023-01-19 00:41:57 -03:00
oobabooga 99536ef5bf Add no-stream option 2023-01-18 23:56:42 -03:00
oobabooga 116299b3ad Manual eos_token implementation 2023-01-18 22:57:39 -03:00
oobabooga 3cb30bed0a Add a "stop" button 2023-01-18 22:44:47 -03:00
oobabooga 8f27d33034 Fix another bug 2023-01-18 22:08:23 -03:00
oobabooga 6c7f187586 Minor change 2023-01-18 21:59:23 -03:00
oobabooga b3cba0b330 Bug 2023-01-18 21:54:44 -03:00
oobabooga df2e910421 Stop generating in chat mode when \nYou: is generated 2023-01-18 21:51:18 -03:00
oobabooga 022960a087 This is the correct way of sampling 1 token at a time 2023-01-18 21:37:21 -03:00
oobabooga 0f01a3b1fa Implement text streaming (#10)
Still experimental. There might be bugs.
2023-01-18 19:06:50 -03:00
oobabooga ca13acdfa0 Ensure that the chat prompt will always contain < 2048 tokens
This way, we can keep the context string at the top of the prompt
even if you keep talking to the bot for hours.

Before this commit, the prompt would be simply truncated and the
context string would eventually be lost.
2023-01-17 20:16:23 -03:00
oobabooga 6456777b09 Clean things up 2023-01-16 16:35:45 -03:00
oobabooga 3a99b2b030 Change a truncation parameter 2023-01-16 13:53:30 -03:00