text-generation-webui/extensions/openai
oobabooga 5a017aa338 API: Several OpenAI spec compliance fixes
- Return proper OpenAI error format ({"error": {...}}) instead of HTTP 500 for validation errors
- Send data: [DONE] at the end of SSE streams
- Fix finish_reason so "tool_calls" takes priority over "length"
- Stop including usage in streaming chunks when include_usage is not set
- Handle "developer" role in messages (treated same as "system")
- Add logprobs and top_logprobs parameters for chat completions
- Fix chat completions logprobs not working with llama.cpp and ExLlamav3 backends
- Add max_completion_tokens as an alias for max_tokens in chat completions
2026-03-12 13:30:38 -03:00
..
cache_embedding_model.py Make /v1/embeddings functional, add request/response types 2023-11-10 07:34:27 -08:00
completions.py API: Several OpenAI spec compliance fixes 2026-03-12 13:30:38 -03:00
embeddings.py Openai embedding fix to support jina-embeddings-v2 (#4642) 2023-11-18 20:24:29 -03:00
errors.py extensions/openai: Fixes for: embeddings, tokens, better errors. +Docs update, +Images, +logit_bias/logprobs, +more. (#3122) 2023-07-24 11:28:12 -03:00
images.py Image: Several fixes 2025-12-05 05:58:57 -08:00
logits.py New llama.cpp loader (#6846) 2025-04-18 09:59:37 -03:00
models.py Refactor to not import gradio in --nowebui mode 2026-03-09 19:29:24 -07:00
moderations.py Lint 2023-11-16 18:03:06 -08:00
script.py API: Several OpenAI spec compliance fixes 2026-03-12 13:30:38 -03:00
tokens.py Add types to the encode/decode/token-count endpoints 2023-11-07 19:32:14 -08:00
typing.py API: Several OpenAI spec compliance fixes 2026-03-12 13:30:38 -03:00
utils.py API: Add tool call parsing for DeepSeek, GLM, MiniMax, and Kimi models 2026-03-06 15:06:56 -03:00