oobabooga
8aeaa76365
Forward logit_bias, logprobs, and n to llama.cpp backend
...
- Forward logit_bias and logprobs natively to llama.cpp
- Support n>1 completions with seed increment for diversity
- Fix logprobs returning empty dict when not requested
2026-03-10 10:41:45 -03:00
oobabooga
328215b0c7
API: Stop generation on client disconnect for non-streaming requests
2026-03-07 06:06:13 -08:00
oobabooga
d0ac58ad31
API: Fix tool_calls placement and other response compatibility issues
2026-03-05 21:25:03 -08:00
oobabooga
f06583b2b9
API: Use \n instead of \r\n as the SSE separator to match OpenAI
2026-03-05 21:16:37 -08:00
oobabooga
9824c82cb6
API: Add parallel request support for llama.cpp and ExLlamaV3
2026-03-05 16:49:58 -08:00
Sense_wang
7bf15ad933
fix: replace bare except clauses with except Exception ( #7400 )
2026-03-04 18:06:17 -03:00
oobabooga
5763947c37
Image: Simplify the API code, add the llm_variations option
2025-12-04 10:23:00 -08:00
oobabooga
4468c49439
Add semaphore to image generation API endpoint
2025-12-03 12:02:47 -08:00
oobabooga
5433ef3333
Add an API endpoint for generating images
2025-12-03 11:50:56 -08:00
oobabooga
84617abdeb
Properly fix the /v1/models endpoint
2025-06-19 10:25:55 -07:00
oobabooga
dcdc42fa06
Fix the /v1/models output format ( closes #7089 )
2025-06-19 07:57:17 -07:00
NoxWorld2660
0b26650f47
Expose real model list via /v1/models endpoint ( #7088 )
2025-06-18 20:14:24 -03:00
djholtby
73bfc936a0
Close response generator when stopping API generation ( #7014 )
2025-05-26 22:39:03 -03:00
oobabooga
83bfd5c64b
Fix API issues
2025-05-18 12:45:01 -07:00
oobabooga
076aa67963
Fix API issues
2025-05-17 22:22:18 -07:00
oobabooga
470c822f44
API: Hide the uvicorn access logs from the terminal
2025-05-16 12:54:39 -07:00
oobabooga
c375b69413
API: Fix llama.cpp generating after disconnect, improve disconnect detection, fix deadlock on simultaneous requests
2025-05-13 11:23:33 -07:00
oobabooga
0c5fa3728e
Revert "Fix API failing to cancel streams (attempt), closes #6966 "
...
This reverts commit 006a866079 .
2025-05-10 19:12:40 -07:00
oobabooga
006a866079
Fix API failing to cancel streams (attempt), closes #6966
2025-05-10 17:55:48 -07:00
oobabooga
bbcaec75b4
API: Find a new port if the default one is taken ( closes #6918 )
2025-04-27 21:13:16 -07:00
oobabooga
35717a088c
API: Add an /v1/internal/health endpoint
2025-04-26 15:42:27 -07:00
oobabooga
bc55feaf3e
Improve host header validation in local mode
2025-04-26 15:42:17 -07:00
oobabooga
d5e1bccef9
Remove the SpeechRecognition requirement
2025-04-20 11:47:28 -07:00
oobabooga
ae02ffc605
Refactor the transformers loader ( #6859 )
2025-04-20 13:33:47 -03:00
BPplays
619265b32c
add ipv6 support to the API ( #6559 )
2025-01-09 10:23:44 -03:00
oobabooga
f27e1ba302
Add a /v1/internal/chat-prompt endpoint ( #5879 )
2024-04-19 00:24:46 -03:00
oobabooga
6247eafcc5
API: better handle temperature = 0
2024-01-22 04:12:23 -08:00
Stefan Daniel Schwarz
232c07bf1f
API: set do_sample=false when temperature=0 ( #5275 )
2024-01-17 23:58:11 -03:00
Philipp Claßen
3eca20c015
Typo fixed in variable names ( #5184 )
2024-01-06 03:05:03 -03:00
oobabooga
23818dc098
Better logger
...
Credits: vladmandic/automatic
2023-12-19 20:38:33 -08:00
oobabooga
b6d16a35b1
Minor API fix
2023-11-21 17:56:28 -08:00
oobabooga
771e62e476
Add /v1/internal/lora endpoints ( #4652 )
2023-11-19 00:35:22 -03:00
oobabooga
ef6feedeb2
Add --nowebui flag for pure API mode ( #4651 )
2023-11-18 23:38:39 -03:00
oobabooga
0fa1af296c
Add /v1/internal/logits endpoint ( #4650 )
2023-11-18 23:19:31 -03:00
oobabooga
8f4f4daf8b
Add --admin-key flag for API ( #4649 )
2023-11-18 22:33:27 -03:00
oobabooga
e0a7cc5e0f
Simplify CORS code
2023-11-16 20:11:55 -08:00
oobabooga
c0233bb9d3
Minor message change
2023-11-16 18:36:57 -08:00
oobabooga
510a01ef46
Lint
2023-11-16 18:03:06 -08:00
oobabooga
a475aa7816
Improve API documentation
2023-11-15 18:39:08 -08:00
oobabooga
a85ce5f055
Add more info messages for truncation / instruction template
2023-11-15 16:20:31 -08:00
oobabooga
be125e2708
Add /v1/internal/model/unload endpoint
2023-11-15 15:48:33 -08:00
oobabooga
c5be3f7acb
Make /v1/embeddings functional, add request/response types
2023-11-10 07:34:27 -08:00
oobabooga
4aabff3728
Remove old API, launch OpenAI API with --api
2023-11-10 06:39:08 -08:00
oobabooga
d86f1fd2c3
OpenAI API: stop streaming on client disconnect ( closes #4521 )
2023-11-09 06:37:32 -08:00
oobabooga
effb3aef42
Prevent deadlocks in OpenAI API with simultaneous requests
2023-11-08 20:55:39 -08:00
oobabooga
678fd73aef
Document /v1/internal/model/load and fix a bug
2023-11-08 17:41:12 -08:00
oobabooga
050ff36bd6
Revert "Add a comment to /v1/models"
...
This reverts commit 38b07493a0 .
2023-11-07 21:09:47 -08:00
oobabooga
38b07493a0
Add a comment to /v1/models
2023-11-07 21:07:12 -08:00
oobabooga
2358706453
Add /v1/internal/model/load endpoint (tentative)
2023-11-07 20:58:06 -08:00
oobabooga
43c53a7820
Refactor the /v1/models endpoint
2023-11-07 19:59:27 -08:00