erew123
f786aa3caa
Clean-up Ctrl+C Shutdown ( #4802 )
2023-12-05 02:16:16 -03:00
oobabooga
3f993280e4
Minor changes
2023-12-04 07:27:44 -08:00
Song Fuchang
0bfd5090be
Import accelerate very early to make Intel GPU happy ( #4704 )
2023-12-03 22:51:18 -03:00
oobabooga
ef6feedeb2
Add --nowebui flag for pure API mode ( #4651 )
2023-11-18 23:38:39 -03:00
oobabooga
4aabff3728
Remove old API, launch OpenAI API with --api
2023-11-10 06:39:08 -08:00
oobabooga
2358706453
Add /v1/internal/model/load endpoint (tentative)
2023-11-07 20:58:06 -08:00
oobabooga
fae8062d39
Bump to latest gradio (3.47) ( #4258 )
2023-10-10 22:20:49 -03:00
oobabooga
b973b91d73
Automatically filter by loader ( closes #4072 )
2023-09-25 10:28:35 -07:00
oobabooga
8ab3eca9ec
Add a warning for outdated installations
2023-09-22 09:35:19 -07:00
oobabooga
9b7646140c
Trim model path if using absolute path
2023-09-19 13:51:57 -07:00
oobabooga
df123a20fc
Prevent extra keys from being saved to settings.yaml
2023-09-11 20:13:10 -07:00
oobabooga
9331ab4798
Read GGUF metadata ( #3873 )
2023-09-11 18:49:30 -03:00
oobabooga
ed86878f02
Remove GGML support
2023-09-11 07:44:00 -07:00
oobabooga
4affa08821
Do not impose instruct mode while loading models
2023-09-02 11:31:33 -07:00
oobabooga
5c7d8bfdfd
Detect CodeLlama settings
2023-08-25 07:06:57 -07:00
oobabooga
73d9befb65
Make "Show controls" customizable through settings.yaml
2023-08-16 07:04:18 -07:00
oobabooga
619cb4e78b
Add "save defaults to settings.yaml" button ( #3574 )
2023-08-14 11:46:07 -03:00
oobabooga
a1a9ec895d
Unify the 3 interface modes ( #3554 )
2023-08-13 01:12:15 -03:00
oobabooga
6d354bb50b
Allow the webui to do multiple tasks simultaneously
2023-08-07 23:57:25 -03:00
oobabooga
bbe4a29a25
Add back dark theme code
2023-08-07 23:03:09 -03:00
oobabooga
65aa11890f
Refactor everything ( #3481 )
2023-08-06 21:49:27 -03:00
oobabooga
0af10ab49b
Add Classifier Free Guidance (CFG) for Transformers/ExLlama ( #3325 )
2023-08-06 17:22:48 -03:00
oobabooga
8df3cdfd51
Add SSL certificate support ( #3453 )
2023-08-04 13:57:31 -03:00
missionfloyd
2336b75d92
Remove unnecessary chat.js ( #3445 )
2023-08-04 01:58:37 -03:00
oobabooga
1839dff763
Use Esc to Stop the generation
2023-08-03 08:13:17 -07:00
oobabooga
3e70bce576
Properly format exceptions in the UI
2023-08-03 06:57:21 -07:00
oobabooga
3390196a14
Add some javascript alerts for confirmations
2023-08-02 22:15:20 -07:00
oobabooga
6bf9e855f8
Minor change
2023-08-02 21:41:38 -07:00
oobabooga
32c564509e
Fix loading session in chat mode
2023-08-02 21:13:16 -07:00
oobabooga
4b6c1d3f08
CSS change
2023-08-02 20:20:23 -07:00
oobabooga
0e8f9354b5
Add direct download for session/chat history JSONs
2023-08-02 19:43:39 -07:00
oobabooga
e931844fe2
Add auto_max_new_tokens parameter ( #3419 )
2023-08-02 14:52:20 -03:00
oobabooga
0d9932815c
Improve TheEncrypted777 on mobile devices
2023-08-02 09:15:54 -07:00
Pete
6afc1a193b
Add a scrollbar to notebook/default, improve chat scrollbar style ( #3403 )
...
---------
Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-08-02 12:02:36 -03:00
oobabooga
b53ed70a70
Make llamacpp_HF 6x faster
2023-08-01 13:18:20 -07:00
oobabooga
959feba602
When saving model settings, only save the settings for the current loader
2023-08-01 06:10:09 -07:00
oobabooga
ebb4f22028
Change a comment
2023-07-31 20:06:10 -07:00
oobabooga
8e2217a029
Minor changes to the Parameters tab
2023-07-31 19:55:11 -07:00
oobabooga
b2207f123b
Update docs
2023-07-31 19:20:48 -07:00
oobabooga
84297d05c4
Add a "Filter by loader" menu to the Parameters tab
2023-07-31 19:09:02 -07:00
oobabooga
e6be25ea11
Fix a regression
2023-07-30 18:12:30 -07:00
oobabooga
5ca37765d3
Only replace {{user}} and {{char}} at generation time
2023-07-30 11:42:30 -07:00
oobabooga
6e16af34fd
Save uploaded characters as yaml
...
Also allow yaml characters to be uploaded directly
2023-07-30 11:25:38 -07:00
oobabooga
ed80a2e7db
Reorder llama.cpp params
2023-07-25 20:45:20 -07:00
oobabooga
0e8782df03
Set instruction template when switching from default/notebook to chat
2023-07-25 20:37:01 -07:00
oobabooga
1b89c304ad
Update README
2023-07-25 15:46:12 -07:00
oobabooga
75c2dd38cf
Remove flexgen support
2023-07-25 15:15:29 -07:00
Shouyi
031fe7225e
Add tensor split support for llama.cpp ( #3171 )
2023-07-25 18:59:26 -03:00
oobabooga
7bc408b472
Change rms_norm_eps to 5e-6 for llama-2-70b ggml
...
Based on https://github.com/ggerganov/llama.cpp/pull/2384
2023-07-25 14:54:57 -07:00
oobabooga
08c622df2e
Autodetect rms_norm_eps and n_gqa for llama-2-70b
2023-07-24 15:27:34 -07:00
oobabooga
a07d070b6c
Add llama-2-70b GGML support ( #3285 )
2023-07-24 16:37:03 -03:00
jllllll
d7a14174a2
Remove auto-loading when only one model is available ( #3187 )
2023-07-18 11:39:08 -03:00
oobabooga
f83fdb9270
Don't reset LoRA menu when loading a model
2023-07-17 12:50:25 -07:00
oobabooga
2de0cedce3
Fix reload screen color
2023-07-15 22:39:39 -07:00
oobabooga
27a84b4e04
Make AutoGPTQ the default again
...
Purely for compatibility with more models.
You should still use ExLlama_HF for LLaMA models.
2023-07-15 22:29:23 -07:00
oobabooga
5e3f7e00a9
Create llamacpp_HF loader ( #3062 )
2023-07-16 02:21:13 -03:00
Panchovix
7c4d4fc7d3
Increase alpha value limit for NTK RoPE scaling for exllama/exllama_HF ( #3149 )
2023-07-16 01:56:04 -03:00
oobabooga
b284f2407d
Make ExLlama_HF the new default for GPTQ
2023-07-14 14:03:56 -07:00
oobabooga
22341e948d
Merge branch 'main' into dev
2023-07-12 14:19:49 -07:00
oobabooga
0e6295886d
Fix lora download folder
2023-07-12 14:19:33 -07:00
oobabooga
eb823fce96
Fix typo
2023-07-12 13:55:19 -07:00
oobabooga
d0a626f32f
Change reload screen color
2023-07-12 13:54:43 -07:00
oobabooga
c592a9b740
Fix #3117
2023-07-12 13:33:44 -07:00
Gabriel Pena
eedb3bf023
Add low vram mode on llama cpp ( #3076 )
2023-07-12 11:05:13 -03:00
Axiom Wolf
d986c17c52
Chat history download creates more detailed file names ( #3051 )
2023-07-12 00:10:36 -03:00
Salvador E. Tropea
324e45b848
[Fixed] wbits and groupsize values from model not shown ( #2977 )
2023-07-11 23:27:38 -03:00
oobabooga
bfafd07f44
Change a message
2023-07-11 18:29:20 -07:00
micsthepick
3708de2b1f
respect model dir for downloads ( #3077 ) ( #3079 )
2023-07-11 18:55:46 -03:00
oobabooga
9aee1064a3
Block a cloudfare request
2023-07-06 22:24:52 -07:00
oobabooga
40c5722499
Fix #2998
2023-07-04 11:35:25 -03:00
oobabooga
55457549cd
Add information about presets to the UI
2023-07-03 22:39:01 -07:00
Panchovix
10c8c197bf
Add Support for Static NTK RoPE scaling for exllama/exllama_hf ( #2955 )
2023-07-04 01:13:16 -03:00
FartyPants
eb6112d5a2
Update server.py - clear LORA after reload ( #2952 )
2023-07-04 00:13:38 -03:00
oobabooga
4b1804a438
Implement sessions + add basic multi-user support ( #2991 )
2023-07-04 00:03:30 -03:00
missionfloyd
ac0f96e785
Some more character import tweaks. ( #2921 )
2023-06-29 14:56:25 -03:00
oobabooga
5d2a8b31be
Improve Parameters tab UI
2023-06-29 14:33:47 -03:00
oobabooga
3443219cbc
Add repetition penalty range parameter to transformers ( #2916 )
2023-06-29 13:40:13 -03:00
oobabooga
22d455b072
Add LoRA support to ExLlama_HF
2023-06-26 00:10:33 -03:00
oobabooga
b7c627f9a0
Set UI defaults
2023-06-25 22:55:43 -03:00
oobabooga
c52290de50
ExLlama with long context ( #2875 )
2023-06-25 22:49:26 -03:00
oobabooga
f0fcd1f697
Sort some imports
2023-06-25 01:44:36 -03:00
oobabooga
e6e5f546b8
Reorganize Chat settings tab
2023-06-25 01:10:20 -03:00
jllllll
bef67af23c
Use pre-compiled python module for ExLlama ( #2770 )
2023-06-24 20:24:17 -03:00
missionfloyd
51a388fa34
Organize chat history/character import menu ( #2845 )
...
* Organize character import menu
* Move Chat history upload/download labels
2023-06-24 09:55:02 -03:00
oobabooga
3ae9af01aa
Add --no_use_cuda_fp16 param for AutoGPTQ
2023-06-23 12:22:56 -03:00
LarryVRH
580c1ee748
Implement a demo HF wrapper for exllama to utilize existing HF transformers decoding. ( #2777 )
2023-06-21 15:31:42 -03:00
Morgan Schweers
447569e31a
Add a download progress bar to the web UI. ( #2472 )
...
* Show download progress on the model screen.
* In case of error, mark as done to clear progress bar.
* Increase the iteration block size to reduce overhead.
2023-06-20 22:59:14 -03:00
oobabooga
09c781b16f
Add modules/block_requests.py
...
This has become unnecessary, but it could be useful in the future
for other libraries.
2023-06-18 16:31:14 -03:00
oobabooga
44f28830d1
Chat CSS: fix ul, li, pre styles + remove redefinitions
2023-06-18 15:20:51 -03:00
oobabooga
239b11c94b
Minor bug fixes
2023-06-17 17:57:56 -03:00
oobabooga
1e400218e9
Fix a typo
2023-06-16 21:01:57 -03:00
oobabooga
5f392122fd
Add gpu_split param to ExLlama
...
Adapted from code created by Ph0rk0z. Thank you Ph0rk0z.
2023-06-16 20:49:36 -03:00
oobabooga
83be8eacf0
Minor fix
2023-06-16 20:38:32 -03:00
oobabooga
9f40032d32
Add ExLlama support ( #2444 )
2023-06-16 20:35:38 -03:00
oobabooga
dea43685b0
Add some clarifications
2023-06-16 19:10:53 -03:00
oobabooga
7ef6a50e84
Reorganize model loading UI completely ( #2720 )
2023-06-16 19:00:37 -03:00
Tom Jobbins
646b0c889f
AutoGPTQ: Add UI and command line support for disabling fused attention and fused MLP ( #2648 )
2023-06-15 23:59:54 -03:00
oobabooga
474dc7355a
Allow API requests to use parameter presets
2023-06-14 11:32:20 -03:00
FartyPants
9f150aedc3
A small UI change in Models menu ( #2640 )
2023-06-12 01:24:44 -03:00
oobabooga
da5d9a28d8
Fix tabbed extensions showing up at the bottom of the UI
2023-06-11 21:20:51 -03:00