From 7b2f15e34ae57a6e86b0901482b4ed9b6b52ad8a Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Sun, 5 Apr 2026 21:16:32 -0700 Subject: [PATCH] Minor change after b1d06dcf96e2b5958ae004b8c9bbb0fc8518328b --- requirements/full/requirements.txt | 2 +- requirements/full/requirements_amd.txt | 2 +- requirements/full/requirements_apple_intel.txt | 2 +- requirements/full/requirements_apple_silicon.txt | 2 +- requirements/full/requirements_cpu_only.txt | 2 +- requirements/full/requirements_nowheels.txt | 2 +- requirements/portable/requirements.txt | 2 +- requirements/portable/requirements_amd.txt | 2 +- requirements/portable/requirements_apple_intel.txt | 2 +- requirements/portable/requirements_apple_silicon.txt | 2 +- requirements/portable/requirements_cpu_only.txt | 2 +- requirements/portable/requirements_cuda131.txt | 2 +- requirements/portable/requirements_ik.txt | 2 +- requirements/portable/requirements_ik_cpu_only.txt | 2 +- requirements/portable/requirements_ik_cuda131.txt | 2 +- requirements/portable/requirements_nowheels.txt | 2 +- requirements/portable/requirements_vulkan.txt | 2 +- 17 files changed, 17 insertions(+), 17 deletions(-) diff --git a/requirements/full/requirements.txt b/requirements/full/requirements.txt index 104cfdb2..d466e7e3 100644 --- a/requirements/full/requirements.txt +++ b/requirements/full/requirements.txt @@ -9,6 +9,7 @@ flash-linear-attention==0.4.* huggingface-hub==1.5.* jinja2==3.1.6 markdown +mcp==1.27.0 numpy==2.2.* pandas peft==0.18.* @@ -48,4 +49,3 @@ https://github.com/turboderp-org/exllamav3/releases/download/v0.0.28/exllamav3-0 https://github.com/turboderp-org/exllamav3/releases/download/v0.0.28/exllamav3-0.0.28+cu128.torch2.9.0-cp313-cp313-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64" and python_version == "3.13" https://github.com/kingbri1/flash-attention/releases/download/v2.8.3/flash_attn-2.8.3+cu128torch2.9.0cxx11abiFALSE-cp313-cp313-win_amd64.whl; platform_system == "Windows" and python_version == "3.13" https://github.com/kingbri1/flash-attention/releases/download/v2.8.3/flash_attn-2.8.3+cu128torch2.9.0cxx11abiFALSE-cp313-cp313-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64" and python_version == "3.13" -mcp==1.27.0 diff --git a/requirements/full/requirements_amd.txt b/requirements/full/requirements_amd.txt index 49db44db..e88ff7c5 100644 --- a/requirements/full/requirements_amd.txt +++ b/requirements/full/requirements_amd.txt @@ -7,6 +7,7 @@ fastapi==0.112.4 huggingface-hub==1.5.* jinja2==3.1.6 markdown +mcp==1.27.0 numpy==2.2.* pandas peft==0.18.* @@ -39,4 +40,3 @@ tiktoken # AMD wheels https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+rocm7.2-py3-none-win_amd64.whl; platform_system == "Windows" https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+rocm7.2-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64" -mcp==1.27.0 diff --git a/requirements/full/requirements_apple_intel.txt b/requirements/full/requirements_apple_intel.txt index 4584708f..eefd979e 100644 --- a/requirements/full/requirements_apple_intel.txt +++ b/requirements/full/requirements_apple_intel.txt @@ -7,6 +7,7 @@ fastapi==0.112.4 huggingface-hub==1.5.* jinja2==3.1.6 markdown +mcp==1.27.0 numpy==2.2.* pandas peft==0.18.* @@ -38,4 +39,3 @@ tiktoken # Mac wheels https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0-py3-none-macosx_13_0_x86_64.whl; platform_system == "Darwin" -mcp==1.27.0 diff --git a/requirements/full/requirements_apple_silicon.txt b/requirements/full/requirements_apple_silicon.txt index 4376a2b4..d1b4e09f 100644 --- a/requirements/full/requirements_apple_silicon.txt +++ b/requirements/full/requirements_apple_silicon.txt @@ -7,6 +7,7 @@ fastapi==0.112.4 huggingface-hub==1.5.* jinja2==3.1.6 markdown +mcp==1.27.0 numpy==2.2.* pandas peft==0.18.* @@ -38,4 +39,3 @@ tiktoken # Mac wheels https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0-py3-none-macosx_13_0_arm64.whl; platform_system == "Darwin" -mcp==1.27.0 diff --git a/requirements/full/requirements_cpu_only.txt b/requirements/full/requirements_cpu_only.txt index 2999d4a9..156ceb77 100644 --- a/requirements/full/requirements_cpu_only.txt +++ b/requirements/full/requirements_cpu_only.txt @@ -7,6 +7,7 @@ fastapi==0.112.4 huggingface-hub==1.5.* jinja2==3.1.6 markdown +mcp==1.27.0 numpy==2.2.* pandas peft==0.18.* @@ -41,4 +42,3 @@ https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+cpu-py3-none-win_amd64.whl; platform_system == "Windows" https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/ik_llama_cpp_binaries-0.106.0+cpu-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64" https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/ik_llama_cpp_binaries-0.106.0+cpu-py3-none-win_amd64.whl; platform_system == "Windows" -mcp==1.27.0 diff --git a/requirements/full/requirements_nowheels.txt b/requirements/full/requirements_nowheels.txt index 5a1e504e..19ac5183 100644 --- a/requirements/full/requirements_nowheels.txt +++ b/requirements/full/requirements_nowheels.txt @@ -7,6 +7,7 @@ fastapi==0.112.4 huggingface-hub==1.5.* jinja2==3.1.6 markdown +mcp==1.27.0 numpy==2.2.* pandas peft==0.18.* @@ -35,4 +36,3 @@ https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.19/gradio_cl flask_cloudflared==0.0.15 sse-starlette==1.6.5 tiktoken -mcp==1.27.0 diff --git a/requirements/portable/requirements.txt b/requirements/portable/requirements.txt index fb51c7cc..8a158f05 100644 --- a/requirements/portable/requirements.txt +++ b/requirements/portable/requirements.txt @@ -3,6 +3,7 @@ fastapi==0.112.4 huggingface-hub==1.5.* jinja2==3.1.6 markdown +mcp==1.27.0 numpy==2.2.* pydantic==2.11.0 pymupdf==1.27.* @@ -25,4 +26,3 @@ tiktoken # CUDA wheels https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+cu124-py3-none-win_amd64.whl; platform_system == "Windows" https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+cu124-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64" -mcp==1.27.0 diff --git a/requirements/portable/requirements_amd.txt b/requirements/portable/requirements_amd.txt index dbea7597..a4949a46 100644 --- a/requirements/portable/requirements_amd.txt +++ b/requirements/portable/requirements_amd.txt @@ -3,6 +3,7 @@ fastapi==0.112.4 huggingface-hub==1.5.* jinja2==3.1.6 markdown +mcp==1.27.0 numpy==2.2.* pydantic==2.11.0 pymupdf==1.27.* @@ -25,4 +26,3 @@ tiktoken # AMD wheels https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+rocm7.2-py3-none-win_amd64.whl; platform_system == "Windows" https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+rocm7.2-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64" -mcp==1.27.0 diff --git a/requirements/portable/requirements_apple_intel.txt b/requirements/portable/requirements_apple_intel.txt index d0f83a74..227823a6 100644 --- a/requirements/portable/requirements_apple_intel.txt +++ b/requirements/portable/requirements_apple_intel.txt @@ -3,6 +3,7 @@ fastapi==0.112.4 huggingface-hub==1.5.* jinja2==3.1.6 markdown +mcp==1.27.0 numpy==2.2.* pydantic==2.11.0 pymupdf==1.27.* @@ -24,4 +25,3 @@ tiktoken # Mac wheels https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0-py3-none-macosx_13_0_x86_64.whl; platform_system == "Darwin" -mcp==1.27.0 diff --git a/requirements/portable/requirements_apple_silicon.txt b/requirements/portable/requirements_apple_silicon.txt index 160c0646..9779dd4a 100644 --- a/requirements/portable/requirements_apple_silicon.txt +++ b/requirements/portable/requirements_apple_silicon.txt @@ -3,6 +3,7 @@ fastapi==0.112.4 huggingface-hub==1.5.* jinja2==3.1.6 markdown +mcp==1.27.0 numpy==2.2.* pydantic==2.11.0 pymupdf==1.27.* @@ -24,4 +25,3 @@ tiktoken # Mac wheels https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0-py3-none-macosx_13_0_arm64.whl; platform_system == "Darwin" -mcp==1.27.0 diff --git a/requirements/portable/requirements_cpu_only.txt b/requirements/portable/requirements_cpu_only.txt index 21695585..ff84907a 100644 --- a/requirements/portable/requirements_cpu_only.txt +++ b/requirements/portable/requirements_cpu_only.txt @@ -3,6 +3,7 @@ fastapi==0.112.4 huggingface-hub==1.5.* jinja2==3.1.6 markdown +mcp==1.27.0 numpy==2.2.* pydantic==2.11.0 pymupdf==1.27.* @@ -25,4 +26,3 @@ tiktoken # llama.cpp (CPU only) https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+cpu-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64" https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+cpu-py3-none-win_amd64.whl; platform_system == "Windows" -mcp==1.27.0 diff --git a/requirements/portable/requirements_cuda131.txt b/requirements/portable/requirements_cuda131.txt index 6b09a46b..89e43e1a 100644 --- a/requirements/portable/requirements_cuda131.txt +++ b/requirements/portable/requirements_cuda131.txt @@ -3,6 +3,7 @@ fastapi==0.112.4 huggingface-hub==1.5.* jinja2==3.1.6 markdown +mcp==1.27.0 numpy==2.2.* pydantic==2.11.0 pymupdf==1.27.* @@ -25,4 +26,3 @@ tiktoken # CUDA wheels https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+cu131-py3-none-win_amd64.whl; platform_system == "Windows" https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+cu131-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64" -mcp==1.27.0 diff --git a/requirements/portable/requirements_ik.txt b/requirements/portable/requirements_ik.txt index ca5ece2d..a23d8ff0 100644 --- a/requirements/portable/requirements_ik.txt +++ b/requirements/portable/requirements_ik.txt @@ -3,6 +3,7 @@ fastapi==0.112.4 huggingface-hub==1.5.* jinja2==3.1.6 markdown +mcp==1.27.0 numpy==2.2.* pydantic==2.11.0 pymupdf==1.27.* @@ -25,4 +26,3 @@ tiktoken # CUDA wheels https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/ik_llama_cpp_binaries-0.106.0+cu124-py3-none-win_amd64.whl; platform_system == "Windows" https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/ik_llama_cpp_binaries-0.106.0+cu124-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64" -mcp==1.27.0 diff --git a/requirements/portable/requirements_ik_cpu_only.txt b/requirements/portable/requirements_ik_cpu_only.txt index f8bafb27..a200e80f 100644 --- a/requirements/portable/requirements_ik_cpu_only.txt +++ b/requirements/portable/requirements_ik_cpu_only.txt @@ -3,6 +3,7 @@ fastapi==0.112.4 huggingface-hub==1.5.* jinja2==3.1.6 markdown +mcp==1.27.0 numpy==2.2.* pydantic==2.11.0 pymupdf==1.27.* @@ -25,4 +26,3 @@ tiktoken # ik_llama.cpp (CPU only) https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/ik_llama_cpp_binaries-0.106.0+cpu-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64" https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/ik_llama_cpp_binaries-0.106.0+cpu-py3-none-win_amd64.whl; platform_system == "Windows" -mcp==1.27.0 diff --git a/requirements/portable/requirements_ik_cuda131.txt b/requirements/portable/requirements_ik_cuda131.txt index 7825b959..8e9a097b 100644 --- a/requirements/portable/requirements_ik_cuda131.txt +++ b/requirements/portable/requirements_ik_cuda131.txt @@ -3,6 +3,7 @@ fastapi==0.112.4 huggingface-hub==1.5.* jinja2==3.1.6 markdown +mcp==1.27.0 numpy==2.2.* pydantic==2.11.0 pymupdf==1.27.* @@ -25,4 +26,3 @@ tiktoken # CUDA wheels https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/ik_llama_cpp_binaries-0.106.0+cu131-py3-none-win_amd64.whl; platform_system == "Windows" https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/ik_llama_cpp_binaries-0.106.0+cu131-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64" -mcp==1.27.0 diff --git a/requirements/portable/requirements_nowheels.txt b/requirements/portable/requirements_nowheels.txt index cde036d9..cafe3cee 100644 --- a/requirements/portable/requirements_nowheels.txt +++ b/requirements/portable/requirements_nowheels.txt @@ -3,6 +3,7 @@ fastapi==0.112.4 huggingface-hub==1.5.* jinja2==3.1.6 markdown +mcp==1.27.0 numpy==2.2.* pydantic==2.11.0 pymupdf==1.27.* @@ -21,4 +22,3 @@ https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.19/gradio_cl flask_cloudflared==0.0.15 sse-starlette==1.6.5 tiktoken -mcp==1.27.0 diff --git a/requirements/portable/requirements_vulkan.txt b/requirements/portable/requirements_vulkan.txt index 32f9e593..59524668 100644 --- a/requirements/portable/requirements_vulkan.txt +++ b/requirements/portable/requirements_vulkan.txt @@ -3,6 +3,7 @@ fastapi==0.112.4 huggingface-hub==1.5.* jinja2==3.1.6 markdown +mcp==1.27.0 numpy==2.2.* pydantic==2.11.0 pymupdf==1.27.* @@ -25,4 +26,3 @@ tiktoken # Vulkan wheels https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+vulkan-py3-none-win_amd64.whl; platform_system == "Windows" https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+vulkan-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64" -mcp==1.27.0