From 552cb09f09de0d3a9442334580fd4a66f95a1ce3 Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Tue, 10 Jun 2025 18:45:42 -0700 Subject: [PATCH] Do not bump Transformers to 4.52 on CUDA 12.8 Performance is slow, and the older version works fine with torch 2.7. --- requirements/full/requirements_cuda128.txt | 3 +-- requirements/full/requirements_cuda128_noavx2.txt | 3 +-- 2 files changed, 2 insertions(+), 4 deletions(-) diff --git a/requirements/full/requirements_cuda128.txt b/requirements/full/requirements_cuda128.txt index 801009fb..7851041f 100644 --- a/requirements/full/requirements_cuda128.txt +++ b/requirements/full/requirements_cuda128.txt @@ -24,8 +24,7 @@ safetensors==0.5.* scipy sentencepiece tensorboard -transformers==4.52.* -triton-windows; platform_system == "Windows" +transformers==4.50.* tqdm wandb diff --git a/requirements/full/requirements_cuda128_noavx2.txt b/requirements/full/requirements_cuda128_noavx2.txt index 8b77e144..c8015166 100644 --- a/requirements/full/requirements_cuda128_noavx2.txt +++ b/requirements/full/requirements_cuda128_noavx2.txt @@ -24,8 +24,7 @@ safetensors==0.5.* scipy sentencepiece tensorboard -transformers==4.52.* -triton-windows; platform_system == "Windows" +transformers==4.50.* tqdm wandb