mirror of
https://github.com/oobabooga/text-generation-webui.git
synced 2026-04-06 07:03:37 +00:00
Compare commits
18 commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
b5afecc63b | ||
|
|
2fbaee58cd | ||
|
|
62e67adb55 | ||
|
|
fc35acab9b | ||
|
|
8ecdb41078 | ||
|
|
5fb8c4fbd6 | ||
|
|
0050a33f37 | ||
|
|
6b66da84d2 | ||
|
|
8e8e1ba898 | ||
|
|
131a9a0140 | ||
|
|
95d6c53e13 | ||
|
|
8bba9ecc3f | ||
|
|
66d1a22c73 | ||
|
|
000d776967 | ||
|
|
a1cb5b5dc0 | ||
|
|
b11379f328 | ||
|
|
42dfcdfc5b | ||
|
|
6e2b70bde6 |
25 changed files with 224 additions and 66 deletions
7
.github/workflows/build-everything-tgw.yml
vendored
7
.github/workflows/build-everything-tgw.yml
vendored
|
|
@ -41,6 +41,13 @@ jobs:
|
||||||
version: ${{ inputs.version }}
|
version: ${{ inputs.version }}
|
||||||
config: 'os:ubuntu-22.04'
|
config: 'os:ubuntu-22.04'
|
||||||
|
|
||||||
|
build_release_rocm_windows:
|
||||||
|
name: ROCm Windows
|
||||||
|
uses: ./.github/workflows/build-portable-release-rocm.yml
|
||||||
|
with:
|
||||||
|
version: ${{ inputs.version }}
|
||||||
|
config: 'os:windows-2022'
|
||||||
|
|
||||||
build_release_rocm_linux:
|
build_release_rocm_linux:
|
||||||
name: ROCm Linux
|
name: ROCm Linux
|
||||||
uses: ./.github/workflows/build-portable-release-rocm.yml
|
uses: ./.github/workflows/build-portable-release-rocm.yml
|
||||||
|
|
|
||||||
|
|
@ -210,6 +210,57 @@ def _expand_tool_sequence(tool_seq):
|
||||||
return messages
|
return messages
|
||||||
|
|
||||||
|
|
||||||
|
def _convert_to_tool_responses(messages):
|
||||||
|
"""Convert role:'tool' messages to tool_responses format.
|
||||||
|
|
||||||
|
Templates like Gemma 4 expect tool results as a ``tool_responses``
|
||||||
|
attribute on a message rather than separate ``role: 'tool'`` messages.
|
||||||
|
This function groups consecutive tool messages and rewrites them.
|
||||||
|
"""
|
||||||
|
result = []
|
||||||
|
tc_id_to_name = {}
|
||||||
|
|
||||||
|
i = 0
|
||||||
|
while i < len(messages):
|
||||||
|
msg = messages[i]
|
||||||
|
|
||||||
|
if msg.get('tool_calls'):
|
||||||
|
for tc in msg['tool_calls']:
|
||||||
|
tc_id = tc.get('id', '')
|
||||||
|
func_name = tc.get('function', {}).get('name', 'unknown')
|
||||||
|
if tc_id:
|
||||||
|
tc_id_to_name[tc_id] = func_name
|
||||||
|
|
||||||
|
if msg.get('role') == 'tool':
|
||||||
|
tool_responses = []
|
||||||
|
while i < len(messages) and messages[i].get('role') == 'tool':
|
||||||
|
tool_msg = messages[i]
|
||||||
|
tc_id = tool_msg.get('tool_call_id', '')
|
||||||
|
func_name = tc_id_to_name.get(tc_id, 'unknown')
|
||||||
|
|
||||||
|
content = tool_msg.get('content', '')
|
||||||
|
try:
|
||||||
|
response = json.loads(content)
|
||||||
|
except (json.JSONDecodeError, ValueError, TypeError):
|
||||||
|
response = content
|
||||||
|
|
||||||
|
tool_responses.append({
|
||||||
|
'name': func_name,
|
||||||
|
'response': response,
|
||||||
|
})
|
||||||
|
i += 1
|
||||||
|
|
||||||
|
result.append({
|
||||||
|
'role': 'tool',
|
||||||
|
'tool_responses': tool_responses,
|
||||||
|
})
|
||||||
|
else:
|
||||||
|
result.append(msg)
|
||||||
|
i += 1
|
||||||
|
|
||||||
|
return result
|
||||||
|
|
||||||
|
|
||||||
def _format_attachments(attachments, include_text=True):
|
def _format_attachments(attachments, include_text=True):
|
||||||
"""Build image ref and text attachment strings from a list of attachments."""
|
"""Build image ref and text attachment strings from a list of attachments."""
|
||||||
attachments_text = ""
|
attachments_text = ""
|
||||||
|
|
@ -267,6 +318,9 @@ def generate_chat_prompt(user_input, state, **kwargs):
|
||||||
tools=state['tools'] if 'tools' in state else None,
|
tools=state['tools'] if 'tools' in state else None,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
active_template_str = state['instruction_template_str'] if state['mode'] == 'instruct' else chat_template_str
|
||||||
|
uses_tool_responses = 'tool_responses' in active_template_str
|
||||||
|
|
||||||
messages = []
|
messages = []
|
||||||
|
|
||||||
if state['mode'] == 'instruct':
|
if state['mode'] == 'instruct':
|
||||||
|
|
@ -503,6 +557,9 @@ def generate_chat_prompt(user_input, state, **kwargs):
|
||||||
|
|
||||||
return prompt
|
return prompt
|
||||||
|
|
||||||
|
if uses_tool_responses:
|
||||||
|
messages = _convert_to_tool_responses(messages)
|
||||||
|
|
||||||
prompt = make_prompt(messages)
|
prompt = make_prompt(messages)
|
||||||
|
|
||||||
# Handle truncation
|
# Handle truncation
|
||||||
|
|
|
||||||
|
|
@ -1,6 +1,7 @@
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
|
|
||||||
from modules import shared, utils
|
from modules import shared, utils
|
||||||
|
from modules.utils import sanitize_filename
|
||||||
from modules.text_generation import get_encoded_length
|
from modules.text_generation import get_encoded_length
|
||||||
|
|
||||||
|
|
||||||
|
|
@ -18,6 +19,7 @@ def load_prompt(fname):
|
||||||
|
|
||||||
return initial_content
|
return initial_content
|
||||||
|
|
||||||
|
fname = sanitize_filename(fname)
|
||||||
file_path = shared.user_data_dir / 'logs' / 'notebook' / f'{fname}.txt'
|
file_path = shared.user_data_dir / 'logs' / 'notebook' / f'{fname}.txt'
|
||||||
if file_path.exists():
|
if file_path.exists():
|
||||||
with open(file_path, 'r', encoding='utf-8') as f:
|
with open(file_path, 'r', encoding='utf-8') as f:
|
||||||
|
|
|
||||||
|
|
@ -7,6 +7,7 @@ THINKING_FORMATS = [
|
||||||
('<|channel|>analysis<|message|>', '<|end|>', '<|channel|>final<|message|>'),
|
('<|channel|>analysis<|message|>', '<|end|>', '<|channel|>final<|message|>'),
|
||||||
('<|channel|>commentary<|message|>', '<|end|>', '<|channel|>final<|message|>'),
|
('<|channel|>commentary<|message|>', '<|end|>', '<|channel|>final<|message|>'),
|
||||||
('<seed:think>', '</seed:think>', None),
|
('<seed:think>', '</seed:think>', None),
|
||||||
|
('<|channel>thought', '<channel|>', None), # Gemma 4
|
||||||
('<|think|>', '<|end|>', '<|content|>'), # Solar Open
|
('<|think|>', '<|end|>', '<|content|>'), # Solar Open
|
||||||
# ('Thinking Process:', '</think>', None), # Qwen3.5 verbose thinking outside tags -- removed: too prone to false positives in streaming
|
# ('Thinking Process:', '</think>', None), # Qwen3.5 verbose thinking outside tags -- removed: too prone to false positives in streaming
|
||||||
(None, '</think>', None), # End-only variant (e.g., Qwen3-next)
|
(None, '</think>', None), # End-only variant (e.g., Qwen3-next)
|
||||||
|
|
|
||||||
|
|
@ -27,6 +27,7 @@ TOOL_CALL_OPENING_MARKERS = [
|
||||||
'[TOOL_CALLS]',
|
'[TOOL_CALLS]',
|
||||||
'to=functions.',
|
'to=functions.',
|
||||||
'<|channel|>commentary',
|
'<|channel|>commentary',
|
||||||
|
'<|tool_call>call:',
|
||||||
]
|
]
|
||||||
|
|
||||||
|
|
||||||
|
|
@ -400,6 +401,78 @@ def _parse_glm_tool_calls(answer: str, tool_names: list[str]):
|
||||||
return matches, start_pos
|
return matches, start_pos
|
||||||
|
|
||||||
|
|
||||||
|
def _extract_gemma4_balanced(text, start):
|
||||||
|
"""Extract balanced braces from Gemma 4 format, using <|"|> as string delimiters."""
|
||||||
|
if start >= len(text) or text[start] != '{':
|
||||||
|
return None
|
||||||
|
depth = 0
|
||||||
|
in_string = False
|
||||||
|
quote_token = '<|"|>'
|
||||||
|
quote_len = len(quote_token)
|
||||||
|
i = start
|
||||||
|
while i < len(text):
|
||||||
|
if text[i:i + quote_len] == quote_token:
|
||||||
|
in_string = not in_string
|
||||||
|
i += quote_len
|
||||||
|
continue
|
||||||
|
if in_string:
|
||||||
|
i += 1
|
||||||
|
continue
|
||||||
|
c = text[i]
|
||||||
|
if c == '{':
|
||||||
|
depth += 1
|
||||||
|
elif c == '}':
|
||||||
|
depth -= 1
|
||||||
|
if depth == 0:
|
||||||
|
return text[start:i + 1]
|
||||||
|
i += 1
|
||||||
|
return None
|
||||||
|
|
||||||
|
|
||||||
|
def _parse_gemma4_tool_calls(answer: str, tool_names: list[str]):
|
||||||
|
"""Parse Gemma 4-style tool calls.
|
||||||
|
|
||||||
|
Format:
|
||||||
|
<|tool_call>call:func_name{key:<|"|>value<|"|>,...}<tool_call|>
|
||||||
|
|
||||||
|
Values use <|"|> tokens instead of standard JSON quotes, and keys are
|
||||||
|
bare identifiers.
|
||||||
|
"""
|
||||||
|
matches = []
|
||||||
|
start_pos = None
|
||||||
|
|
||||||
|
for m in re.finditer(r'<\|tool_call>call:([^\s{]+)\s*', answer):
|
||||||
|
func_name = m.group(1).strip()
|
||||||
|
if func_name not in tool_names:
|
||||||
|
continue
|
||||||
|
|
||||||
|
brace_start = m.end()
|
||||||
|
if brace_start >= len(answer) or answer[brace_start] != '{':
|
||||||
|
continue
|
||||||
|
|
||||||
|
content = _extract_gemma4_balanced(answer, brace_start)
|
||||||
|
if content is None:
|
||||||
|
continue
|
||||||
|
|
||||||
|
# Convert to JSON: split on <|"|> tokens so that key quoting
|
||||||
|
# only applies outside string values (even-indexed parts),
|
||||||
|
# then rejoin with real quotes.
|
||||||
|
parts = content.split('<|"|>')
|
||||||
|
for idx in range(0, len(parts), 2):
|
||||||
|
parts[idx] = re.sub(r'(^|[{,\[])\s*(\w+)\s*:', r'\1"\2":', parts[idx])
|
||||||
|
json_str = '"'.join(parts)
|
||||||
|
|
||||||
|
try:
|
||||||
|
arguments = json.loads(json_str)
|
||||||
|
if start_pos is None:
|
||||||
|
start_pos = m.start()
|
||||||
|
matches.append(_make_tool_call(func_name, arguments))
|
||||||
|
except (json.JSONDecodeError, ValueError):
|
||||||
|
pass
|
||||||
|
|
||||||
|
return matches, start_pos
|
||||||
|
|
||||||
|
|
||||||
def _parse_pythonic_tool_calls(answer: str, tool_names: list[str]):
|
def _parse_pythonic_tool_calls(answer: str, tool_names: list[str]):
|
||||||
"""Parse pythonic-style tool calls used by Llama 4 and similar models.
|
"""Parse pythonic-style tool calls used by Llama 4 and similar models.
|
||||||
|
|
||||||
|
|
@ -472,6 +545,11 @@ TOOL_CALL_FORMATS = [
|
||||||
'parser': _parse_channel_tool_calls,
|
'parser': _parse_channel_tool_calls,
|
||||||
'markers': ['to=functions.', '<|channel|>commentary'],
|
'markers': ['to=functions.', '<|channel|>commentary'],
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
'template_hints': ['<|tool_call>call:'],
|
||||||
|
'parser': _parse_gemma4_tool_calls,
|
||||||
|
'markers': ['<|tool_call>call:'],
|
||||||
|
},
|
||||||
{
|
{
|
||||||
'template_hints': ['minimax:tool_call'],
|
'template_hints': ['minimax:tool_call'],
|
||||||
'parser': _parse_minimax_tool_calls,
|
'parser': _parse_minimax_tool_calls,
|
||||||
|
|
@ -504,6 +582,7 @@ ALL_PARSERS = [
|
||||||
_parse_deep_seek_tool_calls,
|
_parse_deep_seek_tool_calls,
|
||||||
_parse_kimi_tool_calls,
|
_parse_kimi_tool_calls,
|
||||||
_parse_channel_tool_calls,
|
_parse_channel_tool_calls,
|
||||||
|
_parse_gemma4_tool_calls,
|
||||||
_parse_minimax_tool_calls,
|
_parse_minimax_tool_calls,
|
||||||
_parse_glm_tool_calls,
|
_parse_glm_tool_calls,
|
||||||
_parse_xml_param_tool_calls,
|
_parse_xml_param_tool_calls,
|
||||||
|
|
@ -620,6 +699,8 @@ def parse_tool_call(answer: str, tool_names: list[str], return_prefix: bool = Fa
|
||||||
if not isinstance(candidates, list):
|
if not isinstance(candidates, list):
|
||||||
candidates = [candidates]
|
candidates = [candidates]
|
||||||
for candidate_dict in candidates:
|
for candidate_dict in candidates:
|
||||||
|
if not isinstance(candidate_dict, dict):
|
||||||
|
continue
|
||||||
checked_candidate = check_and_sanitize_tool_call_candidate(candidate_dict, tool_names)
|
checked_candidate = check_and_sanitize_tool_call_candidate(candidate_dict, tool_names)
|
||||||
if checked_candidate is not None:
|
if checked_candidate is not None:
|
||||||
matches.append(checked_candidate)
|
matches.append(checked_candidate)
|
||||||
|
|
|
||||||
|
|
@ -10,7 +10,7 @@ from modules.text_generation import (
|
||||||
stop_everything_event
|
stop_everything_event
|
||||||
)
|
)
|
||||||
from modules.ui_notebook import store_notebook_state_and_debounce
|
from modules.ui_notebook import store_notebook_state_and_debounce
|
||||||
from modules.utils import gradio
|
from modules.utils import gradio, sanitize_filename
|
||||||
|
|
||||||
inputs = ('textbox-default', 'interface_state')
|
inputs = ('textbox-default', 'interface_state')
|
||||||
outputs = ('output_textbox', 'html-default')
|
outputs = ('output_textbox', 'html-default')
|
||||||
|
|
@ -167,6 +167,7 @@ def handle_new_prompt():
|
||||||
|
|
||||||
|
|
||||||
def handle_delete_prompt_confirm_default(prompt_name):
|
def handle_delete_prompt_confirm_default(prompt_name):
|
||||||
|
prompt_name = sanitize_filename(prompt_name)
|
||||||
available_prompts = utils.get_available_prompts()
|
available_prompts = utils.get_available_prompts()
|
||||||
current_index = available_prompts.index(prompt_name) if prompt_name in available_prompts else 0
|
current_index = available_prompts.index(prompt_name) if prompt_name in available_prompts else 0
|
||||||
|
|
||||||
|
|
@ -199,6 +200,8 @@ def handle_rename_prompt_click_default(current_name):
|
||||||
|
|
||||||
|
|
||||||
def handle_rename_prompt_confirm_default(new_name, current_name):
|
def handle_rename_prompt_confirm_default(new_name, current_name):
|
||||||
|
new_name = sanitize_filename(new_name)
|
||||||
|
current_name = sanitize_filename(current_name)
|
||||||
old_path = shared.user_data_dir / "logs" / "notebook" / f"{current_name}.txt"
|
old_path = shared.user_data_dir / "logs" / "notebook" / f"{current_name}.txt"
|
||||||
new_path = shared.user_data_dir / "logs" / "notebook" / f"{new_name}.txt"
|
new_path = shared.user_data_dir / "logs" / "notebook" / f"{new_name}.txt"
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -11,7 +11,7 @@ from modules.text_generation import (
|
||||||
get_token_ids,
|
get_token_ids,
|
||||||
stop_everything_event
|
stop_everything_event
|
||||||
)
|
)
|
||||||
from modules.utils import gradio
|
from modules.utils import gradio, sanitize_filename
|
||||||
|
|
||||||
_notebook_file_lock = threading.Lock()
|
_notebook_file_lock = threading.Lock()
|
||||||
_notebook_auto_save_timer = None
|
_notebook_auto_save_timer = None
|
||||||
|
|
@ -202,6 +202,7 @@ def handle_new_prompt():
|
||||||
|
|
||||||
|
|
||||||
def handle_delete_prompt_confirm_notebook(prompt_name):
|
def handle_delete_prompt_confirm_notebook(prompt_name):
|
||||||
|
prompt_name = sanitize_filename(prompt_name)
|
||||||
available_prompts = utils.get_available_prompts()
|
available_prompts = utils.get_available_prompts()
|
||||||
current_index = available_prompts.index(prompt_name) if prompt_name in available_prompts else 0
|
current_index = available_prompts.index(prompt_name) if prompt_name in available_prompts else 0
|
||||||
|
|
||||||
|
|
@ -233,6 +234,8 @@ def handle_rename_prompt_click_notebook(current_name):
|
||||||
|
|
||||||
|
|
||||||
def handle_rename_prompt_confirm_notebook(new_name, current_name):
|
def handle_rename_prompt_confirm_notebook(new_name, current_name):
|
||||||
|
new_name = sanitize_filename(new_name)
|
||||||
|
current_name = sanitize_filename(current_name)
|
||||||
old_path = shared.user_data_dir / "logs" / "notebook" / f"{current_name}.txt"
|
old_path = shared.user_data_dir / "logs" / "notebook" / f"{current_name}.txt"
|
||||||
new_path = shared.user_data_dir / "logs" / "notebook" / f"{new_name}.txt"
|
new_path = shared.user_data_dir / "logs" / "notebook" / f"{new_name}.txt"
|
||||||
|
|
||||||
|
|
@ -249,6 +252,7 @@ def handle_rename_prompt_confirm_notebook(new_name, current_name):
|
||||||
|
|
||||||
def autosave_prompt(text, prompt_name):
|
def autosave_prompt(text, prompt_name):
|
||||||
"""Automatically save the text to the selected prompt file"""
|
"""Automatically save the text to the selected prompt file"""
|
||||||
|
prompt_name = sanitize_filename(prompt_name)
|
||||||
if prompt_name and text.strip():
|
if prompt_name and text.strip():
|
||||||
prompt_path = shared.user_data_dir / "logs" / "notebook" / f"{prompt_name}.txt"
|
prompt_path = shared.user_data_dir / "logs" / "notebook" / f"{prompt_name}.txt"
|
||||||
prompt_path.parent.mkdir(parents=True, exist_ok=True)
|
prompt_path.parent.mkdir(parents=True, exist_ok=True)
|
||||||
|
|
|
||||||
|
|
@ -105,6 +105,9 @@ def resolve_model_path(model_name_or_path, image_model=False):
|
||||||
before the default models directory.
|
before the default models directory.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
if model_name_or_path is None:
|
||||||
|
raise FileNotFoundError("No model specified.")
|
||||||
|
|
||||||
path_candidate = Path(model_name_or_path)
|
path_candidate = Path(model_name_or_path)
|
||||||
if path_candidate.exists():
|
if path_candidate.exists():
|
||||||
return path_candidate
|
return path_candidate
|
||||||
|
|
|
||||||
|
|
@ -31,8 +31,8 @@ tqdm
|
||||||
wandb
|
wandb
|
||||||
|
|
||||||
# Gradio
|
# Gradio
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio-4.37.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio-4.37.2+custom.18-py3-none-any.whl
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio_client-1.0.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio_client-1.0.2+custom.18-py3-none-any.whl
|
||||||
|
|
||||||
# API
|
# API
|
||||||
flask_cloudflared==0.0.15
|
flask_cloudflared==0.0.15
|
||||||
|
|
@ -40,10 +40,10 @@ sse-starlette==1.6.5
|
||||||
tiktoken
|
tiktoken
|
||||||
|
|
||||||
# CUDA wheels
|
# CUDA wheels
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/llama_cpp_binaries-0.102.0+cu124-py3-none-win_amd64.whl; platform_system == "Windows"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+cu124-py3-none-win_amd64.whl; platform_system == "Windows"
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/llama_cpp_binaries-0.102.0+cu124-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+cu124-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/ik_llama_cpp_binaries-0.102.0+cu124-py3-none-win_amd64.whl; platform_system == "Windows"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/ik_llama_cpp_binaries-0.106.0+cu124-py3-none-win_amd64.whl; platform_system == "Windows"
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/ik_llama_cpp_binaries-0.102.0+cu124-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/ik_llama_cpp_binaries-0.106.0+cu124-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
||||||
https://github.com/turboderp-org/exllamav3/releases/download/v0.0.28/exllamav3-0.0.28+cu128.torch2.9.0-cp313-cp313-win_amd64.whl; platform_system == "Windows" and python_version == "3.13"
|
https://github.com/turboderp-org/exllamav3/releases/download/v0.0.28/exllamav3-0.0.28+cu128.torch2.9.0-cp313-cp313-win_amd64.whl; platform_system == "Windows" and python_version == "3.13"
|
||||||
https://github.com/turboderp-org/exllamav3/releases/download/v0.0.28/exllamav3-0.0.28+cu128.torch2.9.0-cp313-cp313-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64" and python_version == "3.13"
|
https://github.com/turboderp-org/exllamav3/releases/download/v0.0.28/exllamav3-0.0.28+cu128.torch2.9.0-cp313-cp313-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64" and python_version == "3.13"
|
||||||
https://github.com/kingbri1/flash-attention/releases/download/v2.8.3/flash_attn-2.8.3+cu128torch2.9.0cxx11abiFALSE-cp313-cp313-win_amd64.whl; platform_system == "Windows" and python_version == "3.13"
|
https://github.com/kingbri1/flash-attention/releases/download/v2.8.3/flash_attn-2.8.3+cu128torch2.9.0cxx11abiFALSE-cp313-cp313-win_amd64.whl; platform_system == "Windows" and python_version == "3.13"
|
||||||
|
|
|
||||||
|
|
@ -28,8 +28,8 @@ trafilatura==2.0.0
|
||||||
wandb
|
wandb
|
||||||
|
|
||||||
# Gradio
|
# Gradio
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio-4.37.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio-4.37.2+custom.18-py3-none-any.whl
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio_client-1.0.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio_client-1.0.2+custom.18-py3-none-any.whl
|
||||||
|
|
||||||
# API
|
# API
|
||||||
flask_cloudflared==0.0.15
|
flask_cloudflared==0.0.15
|
||||||
|
|
@ -37,5 +37,5 @@ sse-starlette==1.6.5
|
||||||
tiktoken
|
tiktoken
|
||||||
|
|
||||||
# AMD wheels
|
# AMD wheels
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/llama_cpp_binaries-0.102.0+rocm7.2-py3-none-win_amd64.whl; platform_system == "Windows"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+rocm7.2-py3-none-win_amd64.whl; platform_system == "Windows"
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/llama_cpp_binaries-0.102.0+rocm7.2-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+rocm7.2-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
||||||
|
|
|
||||||
|
|
@ -28,8 +28,8 @@ trafilatura==2.0.0
|
||||||
wandb
|
wandb
|
||||||
|
|
||||||
# Gradio
|
# Gradio
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio-4.37.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio-4.37.2+custom.18-py3-none-any.whl
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio_client-1.0.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio_client-1.0.2+custom.18-py3-none-any.whl
|
||||||
|
|
||||||
# API
|
# API
|
||||||
flask_cloudflared==0.0.15
|
flask_cloudflared==0.0.15
|
||||||
|
|
@ -37,4 +37,4 @@ sse-starlette==1.6.5
|
||||||
tiktoken
|
tiktoken
|
||||||
|
|
||||||
# Mac wheels
|
# Mac wheels
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/llama_cpp_binaries-0.102.0-py3-none-macosx_13_0_x86_64.whl; platform_system == "Darwin"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0-py3-none-macosx_13_0_x86_64.whl; platform_system == "Darwin"
|
||||||
|
|
|
||||||
|
|
@ -28,8 +28,8 @@ trafilatura==2.0.0
|
||||||
wandb
|
wandb
|
||||||
|
|
||||||
# Gradio
|
# Gradio
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio-4.37.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio-4.37.2+custom.18-py3-none-any.whl
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio_client-1.0.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio_client-1.0.2+custom.18-py3-none-any.whl
|
||||||
|
|
||||||
# API
|
# API
|
||||||
flask_cloudflared==0.0.15
|
flask_cloudflared==0.0.15
|
||||||
|
|
@ -37,4 +37,4 @@ sse-starlette==1.6.5
|
||||||
tiktoken
|
tiktoken
|
||||||
|
|
||||||
# Mac wheels
|
# Mac wheels
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/llama_cpp_binaries-0.102.0-py3-none-macosx_13_0_arm64.whl; platform_system == "Darwin"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0-py3-none-macosx_13_0_arm64.whl; platform_system == "Darwin"
|
||||||
|
|
|
||||||
|
|
@ -28,8 +28,8 @@ trafilatura==2.0.0
|
||||||
wandb
|
wandb
|
||||||
|
|
||||||
# Gradio
|
# Gradio
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio-4.37.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio-4.37.2+custom.18-py3-none-any.whl
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio_client-1.0.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio_client-1.0.2+custom.18-py3-none-any.whl
|
||||||
|
|
||||||
# API
|
# API
|
||||||
flask_cloudflared==0.0.15
|
flask_cloudflared==0.0.15
|
||||||
|
|
@ -37,7 +37,7 @@ sse-starlette==1.6.5
|
||||||
tiktoken
|
tiktoken
|
||||||
|
|
||||||
# llama.cpp (CPU only)
|
# llama.cpp (CPU only)
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/llama_cpp_binaries-0.102.0+cpu-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+cpu-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/llama_cpp_binaries-0.102.0+cpu-py3-none-win_amd64.whl; platform_system == "Windows"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+cpu-py3-none-win_amd64.whl; platform_system == "Windows"
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/ik_llama_cpp_binaries-0.102.0+cpu-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/ik_llama_cpp_binaries-0.106.0+cpu-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/ik_llama_cpp_binaries-0.102.0+cpu-py3-none-win_amd64.whl; platform_system == "Windows"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/ik_llama_cpp_binaries-0.106.0+cpu-py3-none-win_amd64.whl; platform_system == "Windows"
|
||||||
|
|
|
||||||
|
|
@ -28,8 +28,8 @@ trafilatura==2.0.0
|
||||||
wandb
|
wandb
|
||||||
|
|
||||||
# Gradio
|
# Gradio
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio-4.37.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio-4.37.2+custom.18-py3-none-any.whl
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio_client-1.0.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio_client-1.0.2+custom.18-py3-none-any.whl
|
||||||
|
|
||||||
# API
|
# API
|
||||||
flask_cloudflared==0.0.15
|
flask_cloudflared==0.0.15
|
||||||
|
|
|
||||||
|
|
@ -14,8 +14,8 @@ trafilatura==2.0.0
|
||||||
tqdm
|
tqdm
|
||||||
|
|
||||||
# Gradio
|
# Gradio
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio-4.37.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio-4.37.2+custom.18-py3-none-any.whl
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio_client-1.0.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio_client-1.0.2+custom.18-py3-none-any.whl
|
||||||
|
|
||||||
# API
|
# API
|
||||||
flask_cloudflared==0.0.15
|
flask_cloudflared==0.0.15
|
||||||
|
|
@ -23,5 +23,5 @@ sse-starlette==1.6.5
|
||||||
tiktoken
|
tiktoken
|
||||||
|
|
||||||
# CUDA wheels
|
# CUDA wheels
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/llama_cpp_binaries-0.102.0+cu124-py3-none-win_amd64.whl; platform_system == "Windows"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+cu124-py3-none-win_amd64.whl; platform_system == "Windows"
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/llama_cpp_binaries-0.102.0+cu124-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+cu124-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
||||||
|
|
|
||||||
|
|
@ -14,8 +14,8 @@ trafilatura==2.0.0
|
||||||
tqdm
|
tqdm
|
||||||
|
|
||||||
# Gradio
|
# Gradio
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio-4.37.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio-4.37.2+custom.18-py3-none-any.whl
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio_client-1.0.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio_client-1.0.2+custom.18-py3-none-any.whl
|
||||||
|
|
||||||
# API
|
# API
|
||||||
flask_cloudflared==0.0.15
|
flask_cloudflared==0.0.15
|
||||||
|
|
@ -23,5 +23,5 @@ sse-starlette==1.6.5
|
||||||
tiktoken
|
tiktoken
|
||||||
|
|
||||||
# AMD wheels
|
# AMD wheels
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/llama_cpp_binaries-0.102.0+rocm7.2-py3-none-win_amd64.whl; platform_system == "Windows"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+rocm7.2-py3-none-win_amd64.whl; platform_system == "Windows"
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/llama_cpp_binaries-0.102.0+rocm7.2-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+rocm7.2-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
||||||
|
|
|
||||||
|
|
@ -14,8 +14,8 @@ trafilatura==2.0.0
|
||||||
tqdm
|
tqdm
|
||||||
|
|
||||||
# Gradio
|
# Gradio
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio-4.37.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio-4.37.2+custom.18-py3-none-any.whl
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio_client-1.0.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio_client-1.0.2+custom.18-py3-none-any.whl
|
||||||
|
|
||||||
# API
|
# API
|
||||||
flask_cloudflared==0.0.15
|
flask_cloudflared==0.0.15
|
||||||
|
|
@ -23,4 +23,4 @@ sse-starlette==1.6.5
|
||||||
tiktoken
|
tiktoken
|
||||||
|
|
||||||
# Mac wheels
|
# Mac wheels
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/llama_cpp_binaries-0.102.0-py3-none-macosx_13_0_x86_64.whl; platform_system == "Darwin"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0-py3-none-macosx_13_0_x86_64.whl; platform_system == "Darwin"
|
||||||
|
|
|
||||||
|
|
@ -14,8 +14,8 @@ trafilatura==2.0.0
|
||||||
tqdm
|
tqdm
|
||||||
|
|
||||||
# Gradio
|
# Gradio
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio-4.37.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio-4.37.2+custom.18-py3-none-any.whl
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio_client-1.0.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio_client-1.0.2+custom.18-py3-none-any.whl
|
||||||
|
|
||||||
# API
|
# API
|
||||||
flask_cloudflared==0.0.15
|
flask_cloudflared==0.0.15
|
||||||
|
|
@ -23,4 +23,4 @@ sse-starlette==1.6.5
|
||||||
tiktoken
|
tiktoken
|
||||||
|
|
||||||
# Mac wheels
|
# Mac wheels
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/llama_cpp_binaries-0.102.0-py3-none-macosx_13_0_arm64.whl; platform_system == "Darwin"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0-py3-none-macosx_13_0_arm64.whl; platform_system == "Darwin"
|
||||||
|
|
|
||||||
|
|
@ -14,8 +14,8 @@ trafilatura==2.0.0
|
||||||
tqdm
|
tqdm
|
||||||
|
|
||||||
# Gradio
|
# Gradio
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio-4.37.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio-4.37.2+custom.18-py3-none-any.whl
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio_client-1.0.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio_client-1.0.2+custom.18-py3-none-any.whl
|
||||||
|
|
||||||
# API
|
# API
|
||||||
flask_cloudflared==0.0.15
|
flask_cloudflared==0.0.15
|
||||||
|
|
@ -23,5 +23,5 @@ sse-starlette==1.6.5
|
||||||
tiktoken
|
tiktoken
|
||||||
|
|
||||||
# llama.cpp (CPU only)
|
# llama.cpp (CPU only)
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/llama_cpp_binaries-0.102.0+cpu-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+cpu-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/llama_cpp_binaries-0.102.0+cpu-py3-none-win_amd64.whl; platform_system == "Windows"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+cpu-py3-none-win_amd64.whl; platform_system == "Windows"
|
||||||
|
|
|
||||||
|
|
@ -14,8 +14,8 @@ trafilatura==2.0.0
|
||||||
tqdm
|
tqdm
|
||||||
|
|
||||||
# Gradio
|
# Gradio
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio-4.37.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio-4.37.2+custom.18-py3-none-any.whl
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio_client-1.0.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio_client-1.0.2+custom.18-py3-none-any.whl
|
||||||
|
|
||||||
# API
|
# API
|
||||||
flask_cloudflared==0.0.15
|
flask_cloudflared==0.0.15
|
||||||
|
|
@ -23,5 +23,5 @@ sse-starlette==1.6.5
|
||||||
tiktoken
|
tiktoken
|
||||||
|
|
||||||
# CUDA wheels
|
# CUDA wheels
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/llama_cpp_binaries-0.102.0+cu131-py3-none-win_amd64.whl; platform_system == "Windows"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+cu131-py3-none-win_amd64.whl; platform_system == "Windows"
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/llama_cpp_binaries-0.102.0+cu131-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+cu131-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
||||||
|
|
|
||||||
|
|
@ -14,8 +14,8 @@ trafilatura==2.0.0
|
||||||
tqdm
|
tqdm
|
||||||
|
|
||||||
# Gradio
|
# Gradio
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio-4.37.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio-4.37.2+custom.18-py3-none-any.whl
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio_client-1.0.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio_client-1.0.2+custom.18-py3-none-any.whl
|
||||||
|
|
||||||
# API
|
# API
|
||||||
flask_cloudflared==0.0.15
|
flask_cloudflared==0.0.15
|
||||||
|
|
@ -23,5 +23,5 @@ sse-starlette==1.6.5
|
||||||
tiktoken
|
tiktoken
|
||||||
|
|
||||||
# CUDA wheels
|
# CUDA wheels
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/ik_llama_cpp_binaries-0.102.0+cu124-py3-none-win_amd64.whl; platform_system == "Windows"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/ik_llama_cpp_binaries-0.106.0+cu124-py3-none-win_amd64.whl; platform_system == "Windows"
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/ik_llama_cpp_binaries-0.102.0+cu124-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/ik_llama_cpp_binaries-0.106.0+cu124-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
||||||
|
|
|
||||||
|
|
@ -14,8 +14,8 @@ trafilatura==2.0.0
|
||||||
tqdm
|
tqdm
|
||||||
|
|
||||||
# Gradio
|
# Gradio
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio-4.37.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio-4.37.2+custom.18-py3-none-any.whl
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio_client-1.0.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio_client-1.0.2+custom.18-py3-none-any.whl
|
||||||
|
|
||||||
# API
|
# API
|
||||||
flask_cloudflared==0.0.15
|
flask_cloudflared==0.0.15
|
||||||
|
|
@ -23,5 +23,5 @@ sse-starlette==1.6.5
|
||||||
tiktoken
|
tiktoken
|
||||||
|
|
||||||
# ik_llama.cpp (CPU only)
|
# ik_llama.cpp (CPU only)
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/ik_llama_cpp_binaries-0.102.0+cpu-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/ik_llama_cpp_binaries-0.106.0+cpu-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/ik_llama_cpp_binaries-0.102.0+cpu-py3-none-win_amd64.whl; platform_system == "Windows"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/ik_llama_cpp_binaries-0.106.0+cpu-py3-none-win_amd64.whl; platform_system == "Windows"
|
||||||
|
|
|
||||||
|
|
@ -14,8 +14,8 @@ trafilatura==2.0.0
|
||||||
tqdm
|
tqdm
|
||||||
|
|
||||||
# Gradio
|
# Gradio
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio-4.37.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio-4.37.2+custom.18-py3-none-any.whl
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio_client-1.0.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio_client-1.0.2+custom.18-py3-none-any.whl
|
||||||
|
|
||||||
# API
|
# API
|
||||||
flask_cloudflared==0.0.15
|
flask_cloudflared==0.0.15
|
||||||
|
|
@ -23,5 +23,5 @@ sse-starlette==1.6.5
|
||||||
tiktoken
|
tiktoken
|
||||||
|
|
||||||
# CUDA wheels
|
# CUDA wheels
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/ik_llama_cpp_binaries-0.102.0+cu131-py3-none-win_amd64.whl; platform_system == "Windows"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/ik_llama_cpp_binaries-0.106.0+cu131-py3-none-win_amd64.whl; platform_system == "Windows"
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/ik_llama_cpp_binaries-0.102.0+cu131-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/ik_llama_cpp_binaries-0.106.0+cu131-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
||||||
|
|
|
||||||
|
|
@ -14,8 +14,8 @@ trafilatura==2.0.0
|
||||||
tqdm
|
tqdm
|
||||||
|
|
||||||
# Gradio
|
# Gradio
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio-4.37.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio-4.37.2+custom.18-py3-none-any.whl
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio_client-1.0.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio_client-1.0.2+custom.18-py3-none-any.whl
|
||||||
|
|
||||||
# API
|
# API
|
||||||
flask_cloudflared==0.0.15
|
flask_cloudflared==0.0.15
|
||||||
|
|
|
||||||
|
|
@ -14,8 +14,8 @@ trafilatura==2.0.0
|
||||||
tqdm
|
tqdm
|
||||||
|
|
||||||
# Gradio
|
# Gradio
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio-4.37.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio-4.37.2+custom.18-py3-none-any.whl
|
||||||
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.14/gradio_client-1.0.2+custom.14-py3-none-any.whl
|
https://github.com/oobabooga/gradio/releases/download/4.37.2-custom.18/gradio_client-1.0.2+custom.18-py3-none-any.whl
|
||||||
|
|
||||||
# API
|
# API
|
||||||
flask_cloudflared==0.0.15
|
flask_cloudflared==0.0.15
|
||||||
|
|
@ -23,5 +23,5 @@ sse-starlette==1.6.5
|
||||||
tiktoken
|
tiktoken
|
||||||
|
|
||||||
# Vulkan wheels
|
# Vulkan wheels
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/llama_cpp_binaries-0.102.0+vulkan-py3-none-win_amd64.whl; platform_system == "Windows"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+vulkan-py3-none-win_amd64.whl; platform_system == "Windows"
|
||||||
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.102.0/llama_cpp_binaries-0.102.0+vulkan-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
https://github.com/oobabooga/llama-cpp-binaries/releases/download/v0.106.0/llama_cpp_binaries-0.106.0+vulkan-py3-none-linux_x86_64.whl; platform_system == "Linux" and platform_machine == "x86_64"
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue