@@ -25,11 +25,11 @@ popd >/dev/null
2525if [[ " ${CUDA_VERSION_SHORT} " = " cu118" ]]; then
2626 apt-get install -y --no-install-recommends cuda-minimal-build-11-8
2727elif [[ " ${CUDA_VERSION_SHORT} " = " cu124" ]]; then
28- apt-get install -y --no-install-recommends cuda-minimal-build-12-4 dkms
28+ apt-get install -y --no-install-recommends cuda-minimal-build-12-4 numactl dkms
2929elif [[ " ${CUDA_VERSION_SHORT} " = " cu128" ]]; then
30- apt-get install -y --no-install-recommends cuda-minimal-build-12-8 dkms
30+ apt-get install -y --no-install-recommends cuda-minimal-build-12-8 numactl dkms
3131elif [[ " ${CUDA_VERSION_SHORT} " = " cu130" ]]; then
32- apt-get install -y --no-install-recommends cuda-minimal-build-13-0 dkms
32+ apt-get install -y --no-install-recommends cuda-minimal-build-13-0 numactl dkms
3333fi
3434
3535apt-get clean -y
6666pip install torch${TORCH_VERSION} --extra-index-url https://download.pytorch.org/whl/${CUDA_VERSION_SHORT}
6767pip install /wheels/* .whl
6868
69-
7069if [[ " ${CUDA_VERSION_SHORT} " != " cu118" ]] && [[ " ${PYTHON_VERSION} " != " 3.9" ]]; then
71- pip install cuda-python dlblas==0.0.6
70+ pip install cuda-python dlblas==0.0.6 dlslime==0.0.1.post10
71+ fi
72+
73+ # install pre-built flash attention 3 wheel
74+ if [[ " ${CUDA_VERSION_SHORT} " = " cu128" ]]; then
75+ FA3_WHEELS_URL=" https://windreamer.github.io/flash-attention3-wheels/cu128_torch280"
76+ pip install flash_attn_3 --find-links ${FA3_WHEELS_URL} --extra-index-url https://download.pytorch.org/whl/cu128
77+ elif [[ " ${CUDA_VERSION_SHORT} " = " cu130" ]]; then
78+ FA3_WHEELS_URL=" https://windreamer.github.io/flash-attention3-wheels/cu130_torch290"
79+ pip install flash_attn_3 --find-links ${FA3_WHEELS_URL} --extra-index-url https://download.pytorch.org/whl/cu130
7280fi
7381
74- # install pre-compiled flash attention wheel
82+ # install pre-built flash attention wheel
7583PLATFORM=" linux_x86_64"
7684PY_VERSION=$( python3 - << 'PY '
7785import torch, sys
0 commit comments