Skip to content

Commit 6091a19

Browse files
author
yefubao
committed
update install-requirements.sh
1 parent 99cc821 commit 6091a19

File tree

1 file changed

+8
-12
lines changed

1 file changed

+8
-12
lines changed

install/install-requirements.sh

Lines changed: 8 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -82,18 +82,14 @@ CMAKE_ARGS="-DCMAKE_POLICY_VERSION_MINIMUM=3.5" pip install nvidia-cudnn-fronten
8282
python -c "import torch; print('cuDNN version:', torch.backends.cudnn.version());"
8383
python -c "from transformer_engine.pytorch.utils import get_cudnn_version; get_cudnn_version()"
8484

85-
# # Megatron-LM requires flash-attn >= 2.1.1, <= 2.7.3
86-
# cu=$(nvcc --version | grep "Cuda compilation tools" | awk '{print $5}' | cut -d '.' -f 1)
87-
# torch=$(pip show torch | grep Version | awk '{print $2}' | cut -d '+' -f 1 | cut -d '.' -f 1,2)
88-
# cp=$(python3 --version | awk '{print $2}' | awk -F. '{print $1$2}')
89-
# cxx=$(g++ --version | grep 'g++' | awk '{print $3}' | cut -d '.' -f 1)
90-
# wget https://github.com/Dao-AILab/flash-attention/releases/download/v2.7.3/flash_attn-2.7.3+cu${cu}torch${torch}cxx${cxx}abiFALSE-cp${cp}-cp${cp}-linux_x86_64.whl
91-
# pip install flash_attn-2.7.3+cu${cu}torch${torch}cxx${cxx}abiFALSE-cp${cp}-cp${cp}-linux_x86_64.whl
92-
# rm flash_attn-2.7.3+cu${cu}torch${torch}cxx${cxx}abiFALSE-cp${cp}-cp${cp}-linux_x86_64.whl
93-
wget https://github.com/Dao-AILab/flash-attention/releases/download/v2.8.0.post2/flash_attn-2.8.0.post2+cu12torch2.7cxx11abiFALSE-cp312-cp312-linux_x86_64.whl
94-
pip install flash_attn-2.8.0.post2+cu12torch2.7cxx11abiFALSE-cp312-cp312-linux_x86_64.whl
95-
rm flash_attn-2.8.0.post2+cu12torch2.7cxx11abiFALSE-cp312-cp312-linux_x86_64.whl
96-
85+
# Megatron-LM requires flash-attn >= 2.1.1, <= 2.7.3
86+
cu=$(nvcc --version | grep "Cuda compilation tools" | awk '{print $5}' | cut -d '.' -f 1)
87+
torch=$(pip show torch | grep Version | awk '{print $2}' | cut -d '+' -f 1 | cut -d '.' -f 1,2)
88+
cp=$(python3 --version | awk '{print $2}' | awk -F. '{print $1$2}')
89+
cxx=$(g++ --version | grep 'g++' | awk '{print $3}' | cut -d '.' -f 1)
90+
wget https://github.com/Dao-AILab/flash-attention/releases/download/v2.8.0.post2/flash_attn-2.8.0.post2+cu${cu}torch${torch}cxx${cxx}abiFALSE-cp${cp}-cp${cp}-linux_x86_64.whl
91+
pip install --no-cache-dir flash_attn-2.8.0.post2+cu${cu}torch${torch}cxx${cxx}abiFALSE-cp${cp}-cp${cp}-linux_x86_64.whl
92+
rm flash_attn-2.8.0.post2+cu${cu}torch${torch}cxx${cxx}abiFALSE-cp${cp}-cp${cp}-linux_x86_64.whl
9793

9894
# From Megatron-LM log
9995
pip install "git+https://github.com/Dao-AILab/flash-attention.git@v2.7.2#egg=flashattn-hopper&subdirectory=hopper"

0 commit comments

Comments
 (0)