Commit graph

4304 commits

Author SHA1 Message Date
oobabooga 5993ebeb1b Bump exllamav2 to 0.2.9 2025-04-25 05:27:59 -07:00
oobabooga 23399aff3c UI: minor style change 2025-04-24 20:39:00 -07:00
oobabooga 5861013e68 Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2025-04-24 20:36:20 -07:00
oobabooga a90df27ff5 UI: Add a greeting when the chat history is empty 2025-04-24 20:33:40 -07:00
oobabooga ae1fe87365
ExLlamaV2: Add speculative decoding (#6899) 2025-04-25 00:11:04 -03:00
Matthew Jenkins 8f2493cc60
Prevent llamacpp defaults from locking up consumer hardware (#6870) 2025-04-24 23:38:57 -03:00
oobabooga 370fe7b7cf Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2025-04-24 09:33:17 -07:00
oobabooga 8ebe868916 Fix typos in b313adf653 2025-04-24 09:32:17 -07:00
oobabooga 93fd4ad25d llama.cpp: Document the --device-draft syntax 2025-04-24 09:20:11 -07:00
oobabooga f1b64df8dd EXL2: add another torch.cuda.synchronize() call to prevent errors 2025-04-24 09:03:49 -07:00
Ziya 60ac495d59
extensions/superboogav2: existing embedding check bug fix (#6898) 2025-04-24 12:42:05 -03:00
oobabooga b313adf653 Bump llama.cpp, make the wheels work with any Python >= 3.7 2025-04-24 08:26:12 -07:00
oobabooga c71a2af5ab Handle CMD_FLAGS.txt in the main code (closes #6896) 2025-04-24 08:21:06 -07:00
oobabooga bfbde73409 Make 'instruct' the default chat mode 2025-04-24 07:08:49 -07:00
oobabooga e99c20bcb0
llama.cpp: Add speculative decoding (#6891) 2025-04-23 20:10:16 -03:00
oobabooga 9424ba17c8 UI: show only part 00001 of multipart GGUF models in the model menu 2025-04-22 19:56:42 -07:00
oobabooga bce1b68ca9 Minor fix after previous commit 2025-04-22 18:37:36 -07:00
oobabooga 812d878812 Make the dependabot less spammy 2025-04-22 18:35:22 -07:00
oobabooga 8228822a6c Revert "Temporary change"
This reverts commit 765de6f678.
2025-04-22 18:01:47 -07:00
oobabooga 765de6f678 Temporary change 2025-04-22 17:53:56 -07:00
oobabooga 89ec4c9ba6 Add vulkan workflow 2025-04-22 17:51:08 -07:00
oobabooga 06619e5f03 Add vulkan requirements.txt files 2025-04-22 17:46:54 -07:00
oobabooga 4335a24ff8 Fix the workflow 2025-04-22 08:14:13 -07:00
oobabooga 25cf3600aa Lint 2025-04-22 08:04:02 -07:00
oobabooga 39cbb5fee0 Lint 2025-04-22 08:03:25 -07:00
oobabooga da1919baae Update the README 2025-04-22 08:03:22 -07:00
oobabooga a3031795a3 Update the zip filename 2025-04-22 08:03:16 -07:00
oobabooga 008c6dd682 Lint 2025-04-22 08:02:37 -07:00
oobabooga ee09e44c85
Portable version (#6868) 2025-04-22 09:25:57 -03:00
oobabooga 78aeabca89 Fix the transformers loader 2025-04-21 18:33:14 -07:00
oobabooga 8320190184 Fix the exllamav2_HF and exllamav3_HF loaders 2025-04-21 18:32:23 -07:00
oobabooga 15989c2ed8 Make llama.cpp the default loader 2025-04-21 16:36:35 -07:00
oobabooga 86c3ed3218 Small change to the unload_model() function 2025-04-20 20:00:56 -07:00
oobabooga c178ea02fe Revert "Move the requirements*.txt to a requirements folder"
This reverts commit 6117ef7d64.
2025-04-20 19:27:38 -07:00
oobabooga 6117ef7d64 Move the requirements*.txt to a requirements folder 2025-04-20 19:12:04 -07:00
oobabooga fe8e80e04a Merge remote-tracking branch 'refs/remotes/origin/dev' into dev 2025-04-20 19:09:27 -07:00
oobabooga ff1c00bdd9 llama.cpp: set the random seed manually 2025-04-20 19:08:44 -07:00
Matthew Jenkins d3e7c655e5
Add support for llama-cpp builds from https://github.com/ggml-org/llama.cpp (#6862) 2025-04-20 23:06:24 -03:00
oobabooga 99588be576 Organize one_click.py 2025-04-20 18:57:26 -07:00
oobabooga e243424ba1 Fix an import 2025-04-20 17:51:28 -07:00
oobabooga 8cfd7f976b Revert "Remove the old --model-menu flag"
This reverts commit 109de34e3b.
2025-04-20 13:35:42 -07:00
oobabooga d5e1bccef9 Remove the SpeechRecognition requirement 2025-04-20 11:47:28 -07:00
oobabooga b3bf7a885d Fix ExLlamaV2_HF and ExLlamaV3_HF after ae02ffc605 2025-04-20 11:32:48 -07:00
oobabooga 9c59acf820 Remove the numba requirement (it's no longer used) 2025-04-20 10:02:40 -07:00
oobabooga ae02ffc605
Refactor the transformers loader (#6859) 2025-04-20 13:33:47 -03:00
oobabooga 6ba0164c70 Lint 2025-04-19 17:45:21 -07:00
oobabooga 5ab069786b llama.cpp: add back the two encode calls (they are harmless now) 2025-04-19 17:38:36 -07:00
oobabooga b9da5c7e3a Use 127.0.0.1 instead of localhost for faster llama.cpp on Windows 2025-04-19 17:36:04 -07:00
oobabooga 9c9df2063f llama.cpp: fix unicode decoding (closes #6856) 2025-04-19 16:38:15 -07:00
oobabooga ba976d1390 llama.cpp: avoid two 'encode' calls 2025-04-19 16:35:01 -07:00