mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2026-04-07 07:33:47 +00:00
Minor change after b1d06dcf96
This commit is contained in:
parent
05e4842033
commit
7b2f15e34a
17 changed files with 17 additions and 17 deletions
|
|
@ -9,6 +9,7 @@ flash-linear-attention==0.4.*
|
|||
huggingface-hub==1.5.*
|
||||
jinja2==3.1.6
|
||||
markdown
|
||||
mcp==1.27.0
|
||||
numpy==2.2.*
|
||||
pandas
|
||||
peft==0.18.*
|
||||
|
|
@ -48,4 +49,3 @@ https://github.com/turboderp-org/exllamav3/releases/download/v0.0.28/exllamav3-0
|
|||
https://github.com/turboderp-org/exllamav3/releases/download/v0.0.28/exllamav3-0.0.28+cu128.torch2.9.0-cp313-cp313-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64" and python_version == "3.13"
|
||||
https://github.com/kingbri1/flash-attention/releases/download/v2.8.3/flash_attn-2.8.3+cu128torch2.9.0cxx11abiFALSE-cp313-cp313-win_amd64.whl; platform_system == "Windows" and python_version == "3.13"
|
||||
https://github.com/kingbri1/flash-attention/releases/download/v2.8.3/flash_attn-2.8.3+cu128torch2.9.0cxx11abiFALSE-cp313-cp313-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64" and python_version == "3.13"
|
||||
mcp==1.27.0
|
||||
|
|
|
|||
|
|
@ -7,6 +7,7 @@ fastapi==0.112.4
|
|||
huggingface-hub==1.5.*
|
||||
jinja2==3.1.6
|
||||
markdown
|
||||
mcp==1.27.0
|
||||
numpy==2.2.*
|
||||
pandas
|
||||
peft==0.18.*
|
||||
|
|
@ -39,4 +40,3 @@ tiktoken
|
|||
# AMD wheels
|
||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+rocm7.2-py3-none-win_amd64.whl; platform_system == "Windows"
|
||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+rocm7.2-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
||||
mcp==1.27.0
|
||||
|
|
|
|||
|
|
@ -7,6 +7,7 @@ fastapi==0.112.4
|
|||
huggingface-hub==1.5.*
|
||||
jinja2==3.1.6
|
||||
markdown
|
||||
mcp==1.27.0
|
||||
numpy==2.2.*
|
||||
pandas
|
||||
peft==0.18.*
|
||||
|
|
@ -38,4 +39,3 @@ tiktoken
|
|||
|
||||
# Mac wheels
|
||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0-py3-none-macosx_13_0_x86_64.whl; platform_system == "Darwin"
|
||||
mcp==1.27.0
|
||||
|
|
|
|||
|
|
@ -7,6 +7,7 @@ fastapi==0.112.4
|
|||
huggingface-hub==1.5.*
|
||||
jinja2==3.1.6
|
||||
markdown
|
||||
mcp==1.27.0
|
||||
numpy==2.2.*
|
||||
pandas
|
||||
peft==0.18.*
|
||||
|
|
@ -38,4 +39,3 @@ tiktoken
|
|||
|
||||
# Mac wheels
|
||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0-py3-none-macosx_13_0_arm64.whl; platform_system == "Darwin"
|
||||
mcp==1.27.0
|
||||
|
|
|
|||
|
|
@ -7,6 +7,7 @@ fastapi==0.112.4
|
|||
huggingface-hub==1.5.*
|
||||
jinja2==3.1.6
|
||||
markdown
|
||||
mcp==1.27.0
|
||||
numpy==2.2.*
|
||||
pandas
|
||||
peft==0.18.*
|
||||
|
|
@ -41,4 +42,3 @@ https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama
|
|||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+cpu-py3-none-win_amd64.whl; platform_system == "Windows"
|
||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/ik_llama_cpp_binaries-0.106.0+cpu-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/ik_llama_cpp_binaries-0.106.0+cpu-py3-none-win_amd64.whl; platform_system == "Windows"
|
||||
mcp==1.27.0
|
||||
|
|
|
|||
|
|
@ -7,6 +7,7 @@ fastapi==0.112.4
|
|||
huggingface-hub==1.5.*
|
||||
jinja2==3.1.6
|
||||
markdown
|
||||
mcp==1.27.0
|
||||
numpy==2.2.*
|
||||
pandas
|
||||
peft==0.18.*
|
||||
|
|
@ -35,4 +36,3 @@ https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.19/gradio_cl
|
|||
flask_cloudflared==0.0.15
|
||||
sse-starlette==1.6.5
|
||||
tiktoken
|
||||
mcp==1.27.0
|
||||
|
|
|
|||
|
|
@ -3,6 +3,7 @@ fastapi==0.112.4
|
|||
huggingface-hub==1.5.*
|
||||
jinja2==3.1.6
|
||||
markdown
|
||||
mcp==1.27.0
|
||||
numpy==2.2.*
|
||||
pydantic==2.11.0
|
||||
pymupdf==1.27.*
|
||||
|
|
@ -25,4 +26,3 @@ tiktoken
|
|||
# CUDA wheels
|
||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+cu124-py3-none-win_amd64.whl; platform_system == "Windows"
|
||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+cu124-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
||||
mcp==1.27.0
|
||||
|
|
|
|||
|
|
@ -3,6 +3,7 @@ fastapi==0.112.4
|
|||
huggingface-hub==1.5.*
|
||||
jinja2==3.1.6
|
||||
markdown
|
||||
mcp==1.27.0
|
||||
numpy==2.2.*
|
||||
pydantic==2.11.0
|
||||
pymupdf==1.27.*
|
||||
|
|
@ -25,4 +26,3 @@ tiktoken
|
|||
# AMD wheels
|
||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+rocm7.2-py3-none-win_amd64.whl; platform_system == "Windows"
|
||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+rocm7.2-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
||||
mcp==1.27.0
|
||||
|
|
|
|||
|
|
@ -3,6 +3,7 @@ fastapi==0.112.4
|
|||
huggingface-hub==1.5.*
|
||||
jinja2==3.1.6
|
||||
markdown
|
||||
mcp==1.27.0
|
||||
numpy==2.2.*
|
||||
pydantic==2.11.0
|
||||
pymupdf==1.27.*
|
||||
|
|
@ -24,4 +25,3 @@ tiktoken
|
|||
|
||||
# Mac wheels
|
||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0-py3-none-macosx_13_0_x86_64.whl; platform_system == "Darwin"
|
||||
mcp==1.27.0
|
||||
|
|
|
|||
|
|
@ -3,6 +3,7 @@ fastapi==0.112.4
|
|||
huggingface-hub==1.5.*
|
||||
jinja2==3.1.6
|
||||
markdown
|
||||
mcp==1.27.0
|
||||
numpy==2.2.*
|
||||
pydantic==2.11.0
|
||||
pymupdf==1.27.*
|
||||
|
|
@ -24,4 +25,3 @@ tiktoken
|
|||
|
||||
# Mac wheels
|
||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0-py3-none-macosx_13_0_arm64.whl; platform_system == "Darwin"
|
||||
mcp==1.27.0
|
||||
|
|
|
|||
|
|
@ -3,6 +3,7 @@ fastapi==0.112.4
|
|||
huggingface-hub==1.5.*
|
||||
jinja2==3.1.6
|
||||
markdown
|
||||
mcp==1.27.0
|
||||
numpy==2.2.*
|
||||
pydantic==2.11.0
|
||||
pymupdf==1.27.*
|
||||
|
|
@ -25,4 +26,3 @@ tiktoken
|
|||
# llama.cpp (CPU only)
|
||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+cpu-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+cpu-py3-none-win_amd64.whl; platform_system == "Windows"
|
||||
mcp==1.27.0
|
||||
|
|
|
|||
|
|
@ -3,6 +3,7 @@ fastapi==0.112.4
|
|||
huggingface-hub==1.5.*
|
||||
jinja2==3.1.6
|
||||
markdown
|
||||
mcp==1.27.0
|
||||
numpy==2.2.*
|
||||
pydantic==2.11.0
|
||||
pymupdf==1.27.*
|
||||
|
|
@ -25,4 +26,3 @@ tiktoken
|
|||
# CUDA wheels
|
||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+cu131-py3-none-win_amd64.whl; platform_system == "Windows"
|
||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+cu131-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
||||
mcp==1.27.0
|
||||
|
|
|
|||
|
|
@ -3,6 +3,7 @@ fastapi==0.112.4
|
|||
huggingface-hub==1.5.*
|
||||
jinja2==3.1.6
|
||||
markdown
|
||||
mcp==1.27.0
|
||||
numpy==2.2.*
|
||||
pydantic==2.11.0
|
||||
pymupdf==1.27.*
|
||||
|
|
@ -25,4 +26,3 @@ tiktoken
|
|||
# CUDA wheels
|
||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/ik_llama_cpp_binaries-0.106.0+cu124-py3-none-win_amd64.whl; platform_system == "Windows"
|
||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/ik_llama_cpp_binaries-0.106.0+cu124-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
||||
mcp==1.27.0
|
||||
|
|
|
|||
|
|
@ -3,6 +3,7 @@ fastapi==0.112.4
|
|||
huggingface-hub==1.5.*
|
||||
jinja2==3.1.6
|
||||
markdown
|
||||
mcp==1.27.0
|
||||
numpy==2.2.*
|
||||
pydantic==2.11.0
|
||||
pymupdf==1.27.*
|
||||
|
|
@ -25,4 +26,3 @@ tiktoken
|
|||
# ik_llama.cpp (CPU only)
|
||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/ik_llama_cpp_binaries-0.106.0+cpu-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/ik_llama_cpp_binaries-0.106.0+cpu-py3-none-win_amd64.whl; platform_system == "Windows"
|
||||
mcp==1.27.0
|
||||
|
|
|
|||
|
|
@ -3,6 +3,7 @@ fastapi==0.112.4
|
|||
huggingface-hub==1.5.*
|
||||
jinja2==3.1.6
|
||||
markdown
|
||||
mcp==1.27.0
|
||||
numpy==2.2.*
|
||||
pydantic==2.11.0
|
||||
pymupdf==1.27.*
|
||||
|
|
@ -25,4 +26,3 @@ tiktoken
|
|||
# CUDA wheels
|
||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/ik_llama_cpp_binaries-0.106.0+cu131-py3-none-win_amd64.whl; platform_system == "Windows"
|
||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/ik_llama_cpp_binaries-0.106.0+cu131-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
||||
mcp==1.27.0
|
||||
|
|
|
|||
|
|
@ -3,6 +3,7 @@ fastapi==0.112.4
|
|||
huggingface-hub==1.5.*
|
||||
jinja2==3.1.6
|
||||
markdown
|
||||
mcp==1.27.0
|
||||
numpy==2.2.*
|
||||
pydantic==2.11.0
|
||||
pymupdf==1.27.*
|
||||
|
|
@ -21,4 +22,3 @@ https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.19/gradio_cl
|
|||
flask_cloudflared==0.0.15
|
||||
sse-starlette==1.6.5
|
||||
tiktoken
|
||||
mcp==1.27.0
|
||||
|
|
|
|||
|
|
@ -3,6 +3,7 @@ fastapi==0.112.4
|
|||
huggingface-hub==1.5.*
|
||||
jinja2==3.1.6
|
||||
markdown
|
||||
mcp==1.27.0
|
||||
numpy==2.2.*
|
||||
pydantic==2.11.0
|
||||
pymupdf==1.27.*
|
||||
|
|
@ -25,4 +26,3 @@ tiktoken
|
|||
# Vulkan wheels
|
||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+vulkan-py3-none-win_amd64.whl; platform_system == "Windows"
|
||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+vulkan-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
||||
mcp==1.27.0
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue