8000 Merge branch 'upstream-main' into liangan1/flex_attention · pytorch/pytorch@962ee1b · GitHub
[go: up one dir, main page]

Skip to content

Commit 962ee1b

Browse files
committed
Merge branch 'upstream-main' into liangan1/flex_attention
2 parents 800bed4 + 129a297 commit 962ee1b

File tree

196 files changed

+2710
-1867
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

196 files changed

+2710
-1867
lines changed

.ci/caffe2/README.md

-2
Original file line numberDiff line numberDiff line change
@@ -10,5 +10,3 @@ example: `py2-cuda9.0-cudnn7-ubuntu16.04`. The Docker images that are
1010
built on Jenkins and are used in triggered builds already have this
1111
environment variable set in their manifest. Also see
1212
`./docker/jenkins/*/Dockerfile` and search for `BUILD_ENVIRONMENT`.
13-
14-
Our Jenkins installation is located at https://ci.pytorch.org/jenkins/.

.ci/docker/build.sh

+4-4
Original file line numberDiff line numberDiff line change
@@ -251,19 +251,19 @@ case "$tag" in
251251
UCC_COMMIT=${_UCC_COMMIT}
252252
INDUCTOR_BENCHMARKS=yes
253253
;;
254-
pytorch-linux-jammy-xpu-2024.0-py3)
254+
pytorch-linux-jammy-xpu-2025.0-py3)
255255
ANACONDA_PYTHON_VERSION=3.9
256256
GCC_VERSION=11
257257
VISION=yes
258-
XPU_VERSION=0.5
258+
XPU_VERSION=2025.0
259259
NINJA_VERSION=1.9.0
260260
TRITON=yes
261261
;;
262-
pytorch-linux-jammy-xpu-2025.0-py3)
262+
pytorch-linux-jammy-xpu-2025.1-py3)
263263
ANACONDA_PYTHON_VERSION=3.9
264264
GCC_VERSION=11
265265
VISION=yes
266-
XPU_VERSION=2025.0
266+
XPU_VERSION=2025.1
267267
NINJA_VERSION=1.9.0
268268
TRITON=yes
269269
;;

.ci/docker/centos-rocm/Dockerfile

+2-3
Original file line numberDiff line numberDiff line change
@@ -17,9 +17,8 @@ RUN bash ./install_base.sh && rm install_base.sh
1717
# Update CentOS git version
1818
RUN yum -y remove git
1919
RUN yum -y remove git-*
20-
RUN yum -y install https://packages.endpoint.com/rhel/7/os/x86_64/endpoint-repo-1.9-1.x86_64.rpm || \
21-
(yum -y install https://packages.endpointdev.com/rhel/7/os/x86_64/endpoint-repo-1.9-1.x86_64.rpm && \
22-
sed -i "s/packages.endpoint/packages.endpointdev/" /etc/yum.repos.d/endpoint.repo)
20+
RUN yum -y install https://packages.endpointdev.com/rhel/7/os/x86_64/endpoint-repo-1.9-1.x86_64.rpm && \
21+
sed -i 's/packages.endpoint/packages.endpointdev/' /etc/yum.repos.d/endpoint.repo
2322
RUN yum install -y git
2423

2524
# Install devtoolset

.ci/docker/common/install_conda.sh

+1-1
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ if [ -n "$ANACONDA_PYTHON_VERSION" ]; then
77
BASE_URL="https://repo.anaconda.com/miniconda"
88
CONDA_FILE="Miniconda3-latest-Linux-x86_64.sh"
99
if [[ $(uname -m) == "aarch64" ]] || [[ "$BUILD_ENVIRONMENT" == *xpu* ]]; then
10-
BASE_URL="https://github.com/conda-forge/miniforge/releases/latest/download"
10+
BASE_URL="https://github.com/conda-forge/miniforge/releases/latest/download" # @lint-ignore
1111
CONDA_FILE="Miniforge3-Linux-$(uname -m).sh"
1212
fi
1313

.ci/docker/common/install_cpython.sh

+1-1
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33
set -uex -o pipefail
44

55
PYTHON_DOWNLOAD_URL=https://www.python.org/ftp/python
6-
PYTHON_DOWNLOAD_GITHUB_BRANCH=https://github.com/python/cpython/archive/refs/heads
6+
PYTHON_DOWNLOAD_GITHUB_BRANCH=https://github.com/python/cpython/archive/refs/heads # @lint-ignore
77
GET_PIP_URL=https://bootstrap.pypa.io/get-pip.py
88

99
# Python versions to be installed in /opt/$VERSION_NO

.ci/docker/common/install_xpu.sh

+7-7
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@ function install_ubuntu() {
2626
wget -O- https://apt.repos.intel.com/intel-gpg-keys/GPG-PUB-KEY-INTEL-SW-PRODUCTS.PUB \
2727
| gpg --dearmor > /usr/share/keyrings/oneapi-archive-keyring.gpg.gpg
2828
echo "deb [signed-by=/usr/share/keyrings/oneapi-archive-keyring.gpg.gpg] \
29-
https://apt.repos.intel.com/${XPU_REPO_NAME} all main" \
29+
https://apt.repos.intel.com/oneapi all main" \
3030
| tee /etc/apt/sources.list.d/oneAPI.list
3131

3232
# Update the packages list and repository index
@@ -74,7 +74,7 @@ function install_rhel() {
7474
tee > /etc/yum.repos.d/oneAPI.repo << EOF
7575
[oneAPI]
7676
name=Intel for Pytorch GPU dev repository
77-
baseurl=https://yum.repos.intel.com/${XPU_REPO_NAME}
77+
baseurl=https://yum.repos.intel.com/oneapi
7878
enabled=1
7979
gpgcheck=1
8080
repo_gpgcheck=1
118118
https://repositories.intel.com/gpu/sles/${VERSION_SP}${XPU_DRIVER_VERSION}/unified/intel-gpu-${VERSION_SP}.repo
119119
rpm --import https://repositories.intel.com/gpu/intel-graphics.key
120120
# To add the online network network package repository for the Intel Support Packages
121-
zypper addrepo https://yum.repos.intel.com/${XPU_REPO_NAME} oneAPI
121+
zypper addrepo https://yum.repos.intel.com/oneapi oneAPI
122122
rpm --import https://yum.repos.intel.com/intel-gpg-keys/GPG-PUB-KEY-INTEL-SW-PRODUCTS.PUB
123123

124124
# The xpu-smi packages
@@ -141,10 +141,10 @@ if [[ "${XPU_DRIVER_TYPE,,}" == "rolling" ]]; then
141141
XPU_DRIVER_VERSION=""
142142
fi
143143

144-
XPU_REPO_NAME="intel-for-pytorch-gpu-dev"
145-
XPU_PACKAGES="intel-for-pytorch-gpu-dev-0.5 intel-pti-dev-0.9"
146-
if [[ "$XPU_VERSION" == "2025.0" ]]; then
147-
XPU_REPO_NAME="oneapi"
144+
# Default use Intel® oneAPI Deep Learning Essentials 2025.0
145+
if [[ "$XPU_VERSION" == "2025.1" ]]; then
146+
XPU_PACKAGES="intel-deep-learning-essentials-2025.1"
147+
else
148148
XPU_PACKAGES="intel-deep-learning-essentials-2025.0"
149149
fi
150150

.ci/docker/manywheel/Dockerfile_2_28

+1-1
Original file line numberDiff line numberDiff line change
@@ -174,6 +174,6 @@ ENV XPU_DRIVER_TYPE ROLLING
174174
RUN python3 -m pip install --upgrade pip && \
175175
python3 -mpip install cmake==3.28.4
176176
ADD ./common/install_xpu.sh install_xpu.sh
177-
ENV XPU_VERSION 2025.0
177+
ENV XPU_VERSION 2025.1
178178
RUN bash ./install_xpu.sh && rm install_xpu.sh
179179
RUN pushd /opt/_internal && tar -xJf static-libs-for-embedding-only.tar.xz && popd

.ci/docker/manywheel/build_scripts/build.sh

+1-1
Original file line numberDiff line numberDiff line change
@@ -97,7 +97,7 @@ find /opt/_internal -type f -print0 \
9797
| xargs -0 -n1 strip --strip-unneeded 2>/dev/null || true
9898
# We do not need the Python test suites, or indeed the precompiled .pyc and
9999
# .pyo files. Partially cribbed from:
100-
# https://github.com/docker-library/python/blob/master/3.4/slim/Dockerfile
100+
# https://github.com/docker-library/python/blob/master/3.4/slim/Dockerfile # @lint-ignore
101101
find /opt/_internal \
102102
\( -type d -a -name test -o -name tests \) \
103103
-o \( -type f -a -name '*.pyc' -o -name '*.pyo' \) \

.ci/docker/manywheel/build_scripts/build_utils.sh

+1-1
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
# Helper utilities for build
33
# Script used only in CD pipeline
44

5-
OPENSSL_DOWNLOAD_URL=https://www.openssl.org/source/old/1.1.1/
5+
OPENSSL_DOWNLOAD_URL=https://www.openssl.org/source/old/1.1.1/ # @lint-ignore
66
CURL_DOWNLOAD_URL=https://curl.se/download
77

88
AUTOCONF_DOWNLOAD_URL=https://ftp.gnu.org/gnu/autoconf

.ci/manywheel/build_xpu.sh

+4
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,11 @@ fi
2020
source /opt/intel/oneapi/compiler/latest/env/vars.sh
2121
source /opt/intel/oneapi/pti/latest/env/vars.sh
2222
source /opt/intel/oneapi/umf/latest/env/vars.sh
23+
source /opt/intel/oneapi/ccl/latest/env/vars.sh
24+
source /opt/intel/oneapi/mpi/latest/env/vars.sh
2325
export USE_STATIC_MKL=1
26+
export USE_ONEMKL=1
27+
export USE_XCCL=1
2428

2529
WHEELHOUSE_DIR="wheelhousexpu"
2630
LIBTORCH_HOUSE_DIR="libtorch_housexpu"

.ci/onnx/README.md

-2
Original file line numberDiff line numberDiff line change
@@ -10,5 +10,3 @@ example: `py2-cuda9.0-cudnn7-ubuntu16.04`. The Docker images that are
1010
built on Jenkins and are used in triggered builds already have this
1111
environment variable set in their manifest. Also see
1212
`./docker/jenkins/*/Dockerfile` and search for `BUILD_ENVIRONMENT`.
13-
14-
Our Jenkins installation is located at https://ci.pytorch.org/jenkins/.

.ci/pytorch/test.sh

+2-2
Original file line numberDiff line numberDiff line change
@@ -412,10 +412,10 @@ test_inductor_aoti() {
412412
BUILD_AOT_INDUCTOR_TEST=1 TORCH_CUDA_ARCH_LIST=8.6 USE_FLASH_ATTENTION=OFF python setup.py develop
413413
# TODO: Replace me completely, as one should not use conda libstdc++, nor need special path to TORCH_LIB
414414
LD_LIBRARY_PATH=/opt/conda/envs/py_3.10/lib/:${TORCH_LIB_DIR}:$LD_LIBRARY_PATH
415-
CPP_TESTS_DIR="${BUILD_BIN_DIR}" python test/run_test.py --cpp --verbose -i cpp/test_aoti_abi_check cpp/test_aoti_inference
415+
CPP_TESTS_DIR="${BUILD_BIN_DIR}" python test/run_test.py --cpp --verbose -i cpp/test_aoti_abi_check cpp/test_aoti_inference -dist=loadfile
416416
else
417417
BUILD_AOT_INDUCTOR_TEST=1 python setup.py develop
418-
CPP_TESTS_DIR="${BUILD_BIN_DIR}" LD_LIBRARY_PATH="${TORCH_LIB_DIR}" python test/run_test.py --cpp --verbose -i cpp/test_aoti_abi_check cpp/test_aoti_inference
418+
CPP_TESTS_DIR="${BUILD_BIN_DIR}" LD_LIBRARY_PATH="${TORCH_LIB_DIR}" python test/run_test.py --cpp --verbose -i cpp/test_aoti_abi_check cpp/test_aoti_inference -dist=loadfile
419419
fi
420420
}
421421

.ci/pytorch/win-test-helpers/build_pytorch.bat

+6-1
Original file line numberDiff line numberDiff line change
@@ -37,6 +37,11 @@ call %INSTALLER_DIR%\activate_miniconda3.bat
3737
if errorlevel 1 goto fail
3838
if not errorlevel 0 goto fail
3939

40+
:: Update CMake
41+
call choco upgrade -y cmake --no-progress --installargs 'ADD_CMAKE_TO_PATH=System' --apply-install-arguments-to-dependencies --version=3.27.9
42+
if errorlevel 1 goto fail
43+
if not errorlevel 0 goto fail
44+
4045
call pip install mkl-include==2021.4.0 mkl-devel==2021.4.0
4146
if errorlevel 1 goto fail
4247
if not errorlevel 0 goto fail
@@ -88,7 +93,7 @@ set PATH=%CUDA_PATH%\bin;%CUDA_PATH%\libnvvp;%PATH%
8893
:cuda_build_end
8994

9095
set DISTUTILS_USE_SDK=1
91-
set PATH=%TMP_DIR_WIN%\bin;%PATH%
96+
set PATH=%TMP_DIR_WIN%\bin;C:\Program Files\CMake\bin;%PATH%
9297

9398
:: The latest Windows CUDA test is running on AWS G5 runner with A10G GPU
9499
if "%TORCH_CUDA_ARCH_LIST%" == "" set TORCH_CUDA_ARCH_LIST=8.6

.ci/pytorch/win-test-helpers/installation-helpers/install_magma.bat

+1-1
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,7 @@ if "%CUDA_SUFFIX%" == "" (
2424

2525
if "%REBUILD%"=="" (
2626
if "%BUILD_ENVIRONMENT%"=="" (
27-
curl --retry 3 --retry-all-errors -k https://s3.amazonaws.com/ossci-windows/magma_2.5.4_%CUDA_SUFFIX%_%BUILD_TYPE%.7z --output %TMP_DIR_WIN%\magma_2.5.4_%CUDA_SUFFIX%_%BUILD_TYPE%.7z
27+
curl --retry 3 --retry-all-errors -k https://s3.amazonaws.com/ossci-windows/magma_2.5.4_%CUDA_SUFFIX%_%BUILD_TYPE%.7z --output %TMP_DIR_WIN%\magma_2.5.4_%CUDA_SUFFIX%_%BUILD_TYPE%.7z & REM @lint-ignore
2828
) else (
2929
aws s3 cp s3://ossci-windows/magma_2.5.4_%CUDA_SUFFIX%_%BUILD_TYPE%.7z %TMP_DIR_WIN%\magma_2.5.4_%CUDA_SUFFIX%_%BUILD_TYPE%.7z --quiet
3030
)

.ci/pytorch/windows/internal/7z_install.bat

+1-.ci/pytorch/windows/internal/clone.bat
+1-1
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ goto submodule
88

99
:clone_pytorch
1010

11-
git clone https://github.com/%PYTORCH_REPO%/%MODULE_NAME%
11+
git clone https://github.com/%PYTORCH_REPO%/%MODULE_NAME% & REM @lint-ignore
1212

1313
cd %MODULE_NAME%
1414

.ci/pytorch/windows/internal/cuda_install.bat

+8-8
Original file line numberDiff line numberDiff line change
@@ -35,7 +35,7 @@ exit /b 1
3535

3636
set CUDA_INSTALL_EXE=cuda_11.8.0_522.06_windows.exe
3737
if not exist "%SRC_DIR%\temp_build\%CUDA_INSTALL_EXE%" (
38-
curl -k -L "https://ossci-windows.s3.amazonaws.com/%CUDA_INSTALL_EXE%" --output "%SRC_DIR%\temp_build\%CUDA_INSTALL_EXE%"
38+
curl -k -L "https://ossci-windows.s3.amazonaws.com/%CUDA_INSTALL_EXE%" --output "%SRC_DIR%\temp_build\%CUDA_INSTALL_EXE%" & REM @lint-ignore
3939
if errorlevel 1 exit /b 1
4040
set "CUDA_SETUP_FILE=%SRC_DIR%\temp_build\%CUDA_INSTALL_EXE%"
4141
set "ARGS=cuda_profiler_api_11.8 thrust_11.8 nvcc_11.8 cuobjdump_11.8 nvprune_11.8 nvprof_11.8 cupti_11.8 cublas_11.8 cublas_dev_11.8 cudart_11.8 cufft_11.8 cufft_dev_11.8 curand_11.8 curand_dev_11.8 cusolver_11.8 cusolver_dev_11.8 cusparse_11.8 cusparse_dev_11.8 npp_11.8 npp_dev_11.8 nvrtc_11.8 nvrtc_dev_11.8 nvml_dev_11.8 nvtx_11.8"
@@ -45,7 +45,7 @@ set CUDNN_FOLDER=cudnn-windows-x86_64-9.5.0.50_cuda11-archive
4545
set CUDNN_LIB_FOLDER="lib"
4646
set "CUDNN_INSTALL_ZIP=%CUDNN_FOLDER%.zip"
4747
if not exist "%SRC_DIR%\temp_build\%CUDNN_INSTALL_ZIP%" (
48-
curl -k -L "http://s3.amazonaws.com/ossci-windows/%CUDNN_INSTALL_ZIP%" --output "%SRC_DIR%\temp_build\%CUDNN_INSTALL_ZIP%"
48+
curl -k -L "http://s3.amazonaws.com/ossci-windows/%CUDNN_INSTALL_ZIP%" --output "%SRC_DIR%\temp_build\%CUDNN_INSTALL_ZIP%" & REM @lint-ignore
4949
if errorlevel 1 exit /b 1
5050
set "CUDNN_SETUP_FILE=%SRC_DIR%\temp_build\%CUDNN_INSTALL_ZIP%"
5151
)
@@ -62,7 +62,7 @@ goto cuda_common
6262

6363
set CUDA_INSTALL_EXE=cuda_12.4.0_551.61_windows.exe
6464
if not exist "%SRC_DIR%\temp_build\%CUDA_INSTALL_EXE%" (
65-
curl -k -L "https://ossci-windows.s3.amazonaws.com/%CUDA_INSTALL_EXE%" --output "%SRC_DIR%\temp_build\%CUDA_INSTALL_EXE%"
65+
curl -k -L "https://ossci-windows.s3.amazonaws.com/%CUDA_INSTALL_EXE%" --output "%SRC_DIR%\temp_build\%CUDA_INSTALL_EXE%" & REM @lint-ignore
6666
if errorlevel 1 exit /b 1
6767
set "CUDA_SETUP_FILE=%SRC_DIR%\temp_build\%CUDA_INSTALL_EXE%"
6868
set "ARGS=cuda_profiler_api_12.4 thrust_12.4 nvcc_12.4 cuobjdump_12.4 nvprune_12.4 nvprof_12.4 cupti_12.4 cublas_12.4 cublas_dev_12.4 cudart_12.4 cufft_12.4 cufft_dev_12.4 curand_12.4 curand_dev_12.4 cusolver_12.4 cusolver_dev_12.4 cusparse_12.4 cusparse_dev_12.4 npp_12.4 npp_dev_12.4 nvrtc_12.4 nvrtc_dev_12.4 nvml_dev_12.4 nvjitlink_12.4 nvtx_12.4"
@@ -72,7 +72,7 @@ set CUDNN_FOLDER=cudnn-windows-x86_64-9.5.0.50_cuda12-archive
7272
set CUDNN_LIB_FOLDER="lib"
7373
set "CUDNN_INSTALL_ZIP=%CUDNN_FOLDER%.zip"
7474
if not exist "%SRC_DIR%\temp_build\%CUDNN_INSTALL_ZIP%" (
75-
curl -k -L "http://s3.amazonaws.com/ossci-windows/%CUDNN_INSTALL_ZIP%" --output "%SRC_DIR%\temp_build\%CUDNN_INSTALL_ZIP%"
75+
curl -k -L "http://s3.amazonaws.com/ossci-windows/%CUDNN_INSTALL_ZIP%" --output "%SRC_DIR%\temp_build\%CUDNN_INSTALL_ZIP%" & REM @lint-ignore
7676
if errorlevel 1 exit /b 1
7777
set "CUDNN_SETUP_FILE=%SRC_DIR%\temp_build\%CUDNN_INSTALL_ZIP%"
7878
)
@@ -89,7 +89,7 @@ goto cuda_common
8989

9090
set CUDA_INSTALL_EXE=cuda_12.6.2_560.94_windows.exe
9191
if not exist "%SRC_DIR%\temp_build\%CUDA_INSTALL_EXE%" (
92-
curl -k -L "https://ossci-windows.s3.amazonaws.com/%CUDA_INSTALL_EXE%" --output "%SRC_DIR%\temp_build\%CUDA_INSTALL_EXE%"
92+
curl -k -L "https://ossci-windows.s3.amazonaws.com/%CUDA_INSTALL_EXE%" --output "%SRC_DIR%\temp_build\%CUDA_INSTALL_EXE%" & REM @lint-ignore
9393
if errorlevel 1 exit /b 1
9494
set "CUDA_SETUP_FILE=%SRC_DIR%\temp_build\%CUDA_INSTALL_EXE%"
9595
set "ARGS=cuda_profiler_api_12.6 thrust_12.6 nvcc_12.6 cuobjdump_12.6 nvprune_12.6 nvprof_12.6 cupti_12.6 cublas_12.6 cublas_dev_12.6 cudart_12.6 cufft_12.6 cufft_dev_12.6 curand_12.6 curand_dev_12.6 cusolver_12.6 cusolver_dev_12.6 cusparse_12.6 cusparse_dev_12.6 npp_12.6 npp_dev_12.6 nvrtc_12.6 nvrtc_dev_12.6 nvml_dev_12.6 nvjitlink_12.6 nvtx_12.6"
@@ -99,7 +99,7 @@ set CUDNN_FOLDER=cudnn-windows-x86_64-9.5.0.50_cuda12-archive
9999
set CUDNN_LIB_FOLDER="lib"
100100
set "CUDNN_INSTALL_ZIP=%CUDNN_FOLDER%.zip"
101101
if not exist "%SRC_DIR%\temp_build\%CUDNN_INSTALL_ZIP%" (
102-
curl -k -L "http://s3.amazonaws.com/ossci-windows/%CUDNN_INSTALL_ZIP%" --output "%SRC_DIR%\temp_build\%CUDNN_INSTALL_ZIP%"
102+
curl -k -L "http://s3.amazonaws.com/ossci-windows/%CUDNN_INSTALL_ZIP%" --output "%SRC_DIR%\temp_build\%CUDNN_INSTALL_ZIP%" & REM @lint-ignore
103103
if errorlevel 1 exit /b 1
104104
set "CUDNN_SETUP_FILE=%SRC_DIR%\temp_build\%CUDNN_INSTALL_ZIP%"
105105
)
@@ -116,7 +116,7 @@ goto cuda_common
116116

117117
set CUDA_INSTALL_EXE=cuda_12.8.0_571.96_windows.exe
118118
if not exist "%SRC_DIR%\temp_build\%CUDA_INSTALL_EXE%" (
119-
curl -k -L "https://ossci-windows.s3.amazonaws.com/%CUDA_INSTALL_EXE%" --output "%SRC_DIR%\temp_build\%CUDA_INSTALL_EXE%"
119+
curl -k -L "https://ossci-windows.s3.amazonaws.com/%CUDA_INSTALL_EXE%" --output "%SRC_DIR%\temp_build\%CUDA_INSTALL_EXE%" & REM @lint-ignore
120120
if errorlevel 1 exit /b 1
121121
set "CUDA_SETUP_FILE=%SRC_DIR%\temp_build\%CUDA_INSTALL_EXE%"
122122
set "ARGS=cuda_profiler_api_12.8 thrust_12.8 nvcc_12.8 cuobjdump_12.8 nvprune_12.8 nvprof_12.8 cupti_12.8 cublas_12.8 cublas_dev_12.8 cudart_12.8 cufft_12.8 cufft_dev_12.8 curand_12.8 curand_dev_12.8 cusolver_12.8 cusolver_dev_12.8 cusparse_12.8 cusparse_dev_12.8 npp_12.8 npp_dev_12.8 nvrtc_12.8 nvrtc_dev_12.8 nvml_dev_12.8 nvjitlink_12.8 nvtx_12.8"
@@ -126,7 +126,7 @@ set CUDNN_FOLDER=cudnn-windows-x86_64-9.7.0.66_cuda12-archive
126126
set CUDNN_LIB_FOLDER="lib"
127127
set "CUDNN_INSTALL_ZIP=%CUDNN_FOLDER%.zip"
128128
if not exist "%SRC_DIR%\temp_build\%CUDNN_INSTALL_ZIP%" (
129-
curl -k -L "http://s3.amazonaws.com/ossci-windows/%CUDNN_INSTALL_ZIP%" --output "%SRC_DIR%\temp_build\%CUDNN_INSTALL_ZIP%"
129+
curl -k -L "http://s3.amazonaws.com/ossci-windows/%CUDNN_INSTALL_ZIP%" --output "%SRC_DIR%\temp_build\%CUDNN_INSTALL_ZIP%" & REM @lint-ignore
130130
if errorlevel 1 exit /b 1
131131
set "CUDNN_SETUP_FILE=%SRC_DIR%\temp_build\%CUDNN_INSTALL_ZIP%"
132132
)

.ci/pytorch/windows/internal/driver_update.bat

+1-1
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
set WIN_DRIVER_VN=528.89
2-
set "DRIVER_DOWNLOAD_LINK=https://ossci-windows.s3.amazonaws.com/%WIN_DRIVER_VN%-data-center-tesla-desktop-winserver-2016-2019-2022-dch-international.exe"
2+
set "DRIVER_DOWNLOAD_LINK=https://ossci-windows.s3.amazonaws.com/%WIN_DRIVER_VN%-data-center-tesla-desktop-winserver-2016-2019-2022-dch-international.exe" & REM @lint-ignore
33
curl --retry 3 -kL %DRIVER_DOWNLOAD_LINK% --output %WIN_DRIVER_VN%-data-center-tesla-desktop-winserver-2016-2019-2022-dch-international.exe
44
if errorlevel 1 exit /b 1
55

.ci/pytorch/windows/internal/static_lib_test.bat

+1-1
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,7 @@ if "%DEBUG%" == "1" (
3737
if not "%CUDA_VERSION%" == "cpu" (
3838
rmdir /s /q magma_%CUDA_PREFIX%_%BUILD_TYPE%
3939
del magma_%CUDA_PREFIX%_%BUILD_TYPE%.7z
40-
curl -k https://s3.amazonaws.com/ossci-windows/magma_%MAGMA_VERSION%_%CUDA_PREFIX%_%BUILD_TYPE%.7z -o magma_%CUDA_PREFIX%_%BUILD_TYPE%.7z
40+
curl -k https://s3.amazonaws.com/ossci-windows/magma_%MAGMA_VERSION%_%CUDA_PREFIX%_%BUILD_TYPE%.7z -o magma_%CUDA_PREFIX%_%BUILD_TYPE%.7z & REM @lint-ignore
4141
7z x -aoa magma_%CUDA_PREFIX%_%BUILD_TYPE%.7z -omagma_%CUDA_PREFIX%_%BUILD_TYPE%
4242
set LIB=%CD%\magma_%CUDA_PREFIX%_%BUILD_TYPE%\lib;%LIB%
4343
)

.ci/pytorch/windows/internal/xpu_install.bat

+9-39
Original file line numberDiff line numberDiff line change
@@ -10,53 +10,23 @@ if not "%CUDA_VERSION%" == "xpu" (
1010
set SRC_DIR=%NIGHTLIES_PYTORCH_ROOT%
1111
if not exist "%SRC_DIR%\temp_build" mkdir "%SRC_DIR%\temp_build"
1212

13-
set XPU_INSTALL_MODE=%~1
14-
if "%XPU_INSTALL_MODE%"=="" goto xpu_bundle_install_start
15-
if "%XPU_INSTALL_MODE%"=="bundle" goto xpu_bundle_install_start
16-
if "%XPU_INSTALL_MODE%"=="driver" goto xpu_driver_install_start
17-
if "%XPU_INSTALL_MODE%"=="all" goto xpu_driver_install_start
18-
19-
:arg_error
20-
21-
echo Illegal XPU installation mode. The value can be "bundle"/"driver"/"all"
22-
echo If keep the value as space, will use default "bundle" mode
23-
exit /b 1
24-
25-
:xpu_driver_install_start
26-
:: TODO Need more testing for driver installation
27-
set XPU_DRIVER_LINK=https://downloadmirror.intel.com/830975/gfx_win_101.5972.exe
28-
curl -o xpu_driver.exe --retry 3 --retry-all-errors -k %XPU_DRIVER_LINK%
29-
echo "XPU Driver installing..."
30-
start /wait "Intel XPU Driver Installer" "xpu_driver.exe"
31-
if errorlevel 1 exit /b 1
32-
del xpu_driver.exe
33-
if "%XPU_INSTALL_MODE%"=="driver" goto xpu_install_end
34-
3513
:xpu_bundle_install_start
3614

3715
set XPU_BUNDLE_PARENT_DIR=C:\Program Files (x86)\Intel\oneAPI
38-
set XPU_BUNDLE_URL=https://registrationcenter-download.intel.com/akdlm/IRC_NAS/9d1a91e2-e8b8-40a5-8c7f-5db768a6a60c/w_intel-for-pytorch-gpu-dev_p_0.5.3.37_offline.exe
39-
set XPU_BUNDLE_PRODUCT_NAME=intel.oneapi.win.intel-for-pytorch-gpu-dev.product
40-
set XPU_BUNDLE_VERSION=0.5.3+31
16+
set XPU_BUNDLE_URL=https://registrationcenter-download.intel.com/akdlm/IRC_NAS/9d6d6c17-ca2d-4735-9331-99447e4a1280/intel-deep-learning-essentials-2025.0.1.28_offline.exe
17+
set XPU_BUNDLE_PRODUCT_NAME=intel.oneapi.win.deep-learning-essentials.product
18+
set XPU_BUNDLE_VERSION=2025.0.1+20
4119
set XPU_BUNDLE_INSTALLED=0
4220
set XPU_BUNDLE_UNINSTALL=0
43-
set XPU_EXTRA_URL=https://registrationcenter-download.intel.com/akdlm/IRC_NAS/9d1a91e2-e8b8-40a5-8c7f-5db768a6a60c/w_intel-pti-dev_p_0.9.0.37_offline.exe
44-
set XPU_EXTRA_PRODUCT_NAME=intel.oneapi.win.intel-pti-dev.product
45-
set XPU_EXTRA_VERSION=0.9.0+36
21+
set XPU_EXTRA_URL=NULL
22+
set XPU_EXTRA_PRODUCT_NAME=intel.oneapi.win.compiler.product
23+
set XPU_EXTRA_VERSION=2025.0.1+1226
4624
set XPU_EXTRA_INSTALLED=0
4725
set XPU_EXTRA_UNINSTALL=0
4826

49-
if not [%XPU_VERSION%]==[] if [%XPU_VERSION%]==[2025.0] (
50-
set XPU_BUNDLE_URL=https://registrationcenter-download.intel.com/akdlm/IRC_NAS/9d6d6c17-ca2d-4735-9331-99447e4a1280/intel-deep-learning-essentials-2025.0.1.28_offline.exe
51-
set XPU_BUNDLE_PRODUCT_NAME=intel.oneapi.win.deep-learning-essentials.product
52-
set XPU_BUNDLE_VERSION=2025.0.1+20
53-
set XPU_BUNDLE_INSTALLED=0
54-
set XPU_BUNDLE_UNINSTALL=0
55-
set XPU_EXTRA_URL=NULL
56-
set XPU_EXTRA_PRODUCT_NAME=intel.oneapi.win.compiler.product
57-
set XPU_EXTRA_VERSION=2025.0.1+1226
58-
set XPU_EXTRA_INSTALLED=0
59-
set XPU_EXTRA_UNINSTALL=0
27+
if not [%XPU_VERSION%]==[] if [%XPU_VERSION%]==[2025.1] (
28+
set XPU_BUNDLE_URL=https://registrationcenter-download.intel.com/akdlm/IRC_NAS/1a9fff3d-04c2-4d77-8861-3d86c774b66f/intel-deep-learning-essentials-2025.1.1.26_offline.exe
29+
set XPU_BUNDLE_VERSION=2025.1.1+23
6030
)
6131

6232
:: Check if XPU bundle is target version or already installed

.ci/pytorch/windows/xpu.bat

+1
Original file line numberDiff line numberDiff line change
@@ -26,6 +26,7 @@ set VS2022INSTALLDIR=%VS15INSTALLDIR%
2626
set XPU_BUNDLE_ROOT=%ProgramFiles(x86)%\Intel\oneAPI
2727
call "%XPU_BUNDLE_ROOT%\compiler\latest\env\vars.bat"
2828
call "%XPU_BUNDLE_ROOT%\ocloc\latest\env\vars.bat"
29+
set USE_ONEMKL=1
2930
IF ERRORLEVEL 1 goto :eof
3031

3132
if exist "%NIGHTLIES_PYTORCH_ROOT%" cd %NIGHTLIES_PYTORCH_ROOT%\..

.circleci/scripts/binary_windows_build.sh

+1-1
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@ fi
1515
if [[ "$DESIRED_CUDA" == 'xpu' ]]; then
1616
export VC_YEAR=2022
1717
export USE_SCCACHE=0
18-
export XPU_VERSION=2025.0
18+
export XPU_VERSION=2025.1
1919
export XPU_ENABLE_KINETO=1
2020
fi
2121

.circleci/scripts/binary_windows_test.sh

+1-1
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ export VC_YEAR=2019
88

99
if [[ "$DESIRED_CUDA" == 'xpu' ]]; then
1010
export VC_YEAR=2022
11-
export XPU_VERSION=2025.0
11+
export XPU_VERSION=2025.1
1212
fi
1313

1414
pushd "$PYTORCH_ROOT/.ci/pytorch/"

.github/ci_commit_pins/audio.txt

+1-1
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
2f78c953bb4a81d269c7d4b7b36e218a1f090fab
1+
ea5de17755d657508c84c4dce8970b614008adcf

0 commit comments

Comments
 (0)
0