Commit graph

97 commits

Author SHA1 Message Date
oobabooga 44018c2f69
Add a "llamacpp_HF creator" menu (#5519) 2024-02-16 12:43:24 -03:00
oobabooga ee65f4f014 Downloader: don't assume that huggingface_hub is installed 2024-01-30 09:14:11 -08:00
Anthony Guijarro 828be63f2c
Downloader: use HF get_token function (#5381) 2024-01-27 17:13:09 -03:00
oobabooga 7bbe7e803a Minor fix 2023-12-08 05:01:25 -08:00
oobabooga d516815c9c Model downloader: download only fp16 if both fp16 and GGUF are present 2023-12-05 21:09:12 -08:00
oobabooga 510a01ef46 Lint 2023-11-16 18:03:06 -08:00
LightningDragon 0ced78fdfa
Replace hashlib.sha256 with hashlib.file_digest so we don't need to load entire files into ram before hashing them. (#4383) 2023-10-25 12:15:34 -03:00
oobabooga 613feca23b Make colab functional for llama.cpp
- Download only Q4_K_M for GGUF repositories by default
- Use maximum n-gpu-layers by default
2023-10-22 09:08:25 -07:00
oobabooga cd45635f53 tqdm improvement for colab 2023-10-21 22:00:29 -07:00
oobabooga 3a9d90c3a1 Download models with 4 threads by default 2023-10-10 13:52:10 -07:00
快乐的我531 4e56ad55e1
Let model downloader download *.tiktoken as well (#4121) 2023-09-28 18:03:18 -03:00
kalomaze 7c9664ed35
Allow full model URL to be used for download (#3919)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-09-16 10:06:13 -03:00
oobabooga df52dab67b Lint 2023-09-11 07:57:38 -07:00
oobabooga ed86878f02 Remove GGML support 2023-09-11 07:44:00 -07:00
missionfloyd 787219267c
Allow downloading single file from UI (#3737) 2023-08-29 23:32:36 -03:00
Alberto Ferrer f63dd83631
Update download-model.py (Allow single file download) (#3732) 2023-08-29 22:57:58 -03:00
oobabooga 7f5370a272 Minor fixes/cosmetics 2023-08-26 22:11:07 -07:00
jllllll 4a999e3bcd
Use separate llama-cpp-python packages for GGML support 2023-08-26 10:40:08 -05:00
oobabooga 83640d6f43 Replace ggml occurences with gguf 2023-08-26 01:06:59 -07:00
Thomas De Bonnet 0dfd1a8b7d
Improve readability of download-model.py (#3497) 2023-08-20 20:13:13 -03:00
oobabooga 4b3384e353 Handle unfinished lists during markdown streaming 2023-08-03 17:15:18 -07:00
oobabooga 13449aa44d Decrease download timeout 2023-07-15 22:30:08 -07:00
oobabooga e202190c4f lint 2023-07-12 11:33:25 -07:00
Ahmad Fahadh Ilyas 8db7e857b1
Add token authorization for downloading model (#3067) 2023-07-11 18:48:08 -03:00
FartyPants 61102899cd
google flan T5 download fix (#3080) 2023-07-11 18:46:59 -03:00
tianchen zhong c7058afb40
Add new possible bin file name regex (#3070) 2023-07-09 17:22:56 -03:00
jeckyhl 88a747b5b9
fix: Error when downloading model from UI (#3014) 2023-07-05 11:27:29 -03:00
AN Long be4582be40
Support specify retry times in download-model.py (#2908) 2023-07-04 22:26:30 -03:00
Roman 38897fbd8a
fix: added model parameter check (#2829) 2023-06-24 10:09:34 -03:00
Gaurav Bhagchandani 89fb6f9236
Fixed the ZeroDivisionError when downloading a model (#2797) 2023-06-21 12:31:50 -03:00
oobabooga 5dfe0bec06 Remove old/useless code 2023-06-20 23:36:56 -03:00
oobabooga faa92eee8d Add spaces 2023-06-20 23:25:58 -03:00
Peter Sofronas b22c7199c9
Download optimizations (#2786)
* download_model_files metadata writing improvement

* line swap

* reduce line length

* safer download and greater block size

* Minor changes by pycodestyle

---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-06-20 23:14:18 -03:00
Morgan Schweers 447569e31a
Add a download progress bar to the web UI. (#2472)
* Show download progress on the model screen.

* In case of error, mark as done to clear progress bar.

* Increase the iteration block size to reduce overhead.
2023-06-20 22:59:14 -03:00
oobabooga 240752617d Increase download timeout to 20s 2023-06-08 11:16:38 -03:00
oobabooga 9f215523e2 Remove some unused imports 2023-06-06 07:05:46 -03:00
Morgan Schweers 1aed2b9e52
Make it possible to download protected HF models from the command line. (#2408) 2023-06-01 00:11:21 -03:00
Juan M Uys b984a44f47
fix error when downloading a model for the first time (#2404) 2023-05-30 22:07:12 -03:00
oobabooga b4662bf4af
Download gptq_model*.py using download-model.py 2023-05-29 16:12:54 -03:00
oobabooga 39dab18307 Add a timeout to download-model.py requests 2023-05-19 11:19:34 -03:00
oobabooga c7ba2d4f3f Change a message in download-model.py 2023-05-10 19:00:14 -03:00
Wojtek Kowaluk 1436c5845a
fix ggml detection regex in model downloader (#1779) 2023-05-04 11:48:36 -03:00
Lou Bernardi a6ef2429fa
Add "do not download" and "download from HF" to download-model.py (#1439) 2023-04-21 12:54:50 -03:00
Rudd-O 69d50e2e86
Fix download script (#1373) 2023-04-19 13:02:32 -03:00
Forkoz c6fe1ced01
Add ChatGLM support (#1256)
---------

Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2023-04-16 19:15:03 -03:00
oobabooga 2c14df81a8 Use download-model.py to download the model 2023-04-10 11:36:39 -03:00
oobabooga 170e0c05c4 Typo 2023-04-09 17:00:59 -03:00
oobabooga 34ec02d41d Make download-model.py importable 2023-04-09 16:59:59 -03:00
Blake Wyatt df561fd896
Fix ggml downloading in download-model.py (#915) 2023-04-08 18:52:30 -03:00
oobabooga ea6e77df72
Make the code more like PEP8 for readability (#862) 2023-04-07 00:15:45 -03:00