8000 Enable XPU distributed test for PT2.8 by daisyden · Pull Request #149916 · pytorch/pytorch · GitHub
[go: up one dir, main page]

Skip to content

Enable XPU distributed test for PT2.8 #149916

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
wants to merge 53 commits into from
Closed
Show file tree
Hide file tree
Changes from 1 commit
Commits
Show all changes
53 commits
Select commit Hold shift + click to select a range
d0d8271
make skipXPU work
daisyden May 11, 2024
c791db9
enabled torch-xpu ops in op_db
daisyden May 13, 2024
f5cbd50
clean up code
daisyden May 13, 2024
4d94417
Revert "clean up code"
daisyden May 13, 2024
6844101
Revert "enabled torch-xpu ops in op_db"
daisyden May 13, 2024
5051e3c
Revert "make skipXPU work"
daisyden May 13, 2024
e2aa92a
merge common code update from https://github.com/Chao1Han/pytorch/pul…
chunhuanMeng Mar 19, 2025
9e83095
Merge branch 'main' of https://github.com/daisyden/pytorch into distr…
chunhuanMeng Mar 20, 2025
06dd2aa
merge common code update from https://github.com/Chao1Han/pytorch/pul…
daisyden Mar 20, 2025
a4a732b
Add XPU support for distributed
daisyden Mar 20, 2025
6e3f6b8
Merge branch 'distributed_2.8' of https://github.com/daisyden/pytorch…
daisyden Mar 20, 2025
5f47367
Merge remote-tracking branch 'upstream/main' into distributed_2.8
daisyden Mar 21, 2025
345d7e6
ported fsdp and _composable/fsdp cases
daisyden Mar 21, 2025
4a5a522
Support XPU device for DDP test cases
PenghuiCheng Mar 24, 2025
20a4456
Support XPU device for pipeline cases
PenghuiCheng Mar 24, 2025
a90a603
ported fsdp tests
daisyden Mar 24, 2025
5b1aff7
Merge branch 'distributed_2.8' of https://github.com/daisyden/pytorch…
daisyden Mar 24, 2025
44d55b9
fixed backend mapping error for register_backend function
PenghuiCheng Mar 26, 2025
7dade1f
Update distributed UT cases
PenghuiCheng Apr 1, 2025
580aaee
remove fsdp_kwargs in test_fsdp_memory.py to align with cuda, added r…
daisyden Apr 1, 2025
c0f5713
Merge branch 'upstream_main4' into distributed_2.8
daisyden Apr 1, 2025
6dedbe3
Add test_dynamo_distributed cases
PenghuiCheng Apr 1, 2025
20d074c
Merge remote-tracking branch 'upstream/distributed_2.8' into distribu…
PenghuiCheng Apr 1, 2025
124ff16
update test_tp_random_state.py
PenghuiCheng Apr 2, 2025
0bea112
Merge from main branch
PenghuiCheng Apr 7, 2025
7409ade
support xccl in with_comms
daisyden Apr 8, 2025
636cbff
Merge branch 'distributed_2.8' of https://github.com/daisyden/pytorch…
daisyden Apr 8, 2025
0d5a86b
Merge branch 'upstream_main3' into distributed_2.8
daisyden Apr 8, 2025
3826e30
Enabled UT in test/distributed/tensor
PenghuiCheng Apr 9, 2025
cb711b7
Merge branch 'distributed_2.8' of https://github.com/daisyden/pytorch…
PenghuiCheng Apr 9, 2025
d6cd1b3
refine fsdp2 test case for xpu
daisyden Apr 9, 2025
624be3a
Merge branch 'distributed_2.8' of https://github.com/daisyden/pytorch…
daisyden Apr 9, 2025
8d8c5fe
fix some issues in test case, cuda specific code, world_size 8, etc.
daisyden Apr 11, 2025
1cf7887
merge from main branch
PenghuiCheng Apr 16, 2025
41475ac
Merge remote-tracking branch 'upstream/distributed_2.8' into distribu…
PenghuiCheng Apr 16, 2025
0628c76
Change world size in test_device_mesh.py
PenghuiCheng Apr 18, 2025
b0d935d
Merge remote-tracking branch 'origin/distributed_2.8' into distribute…
PenghuiCheng Apr 18, 2025
58eb87e
Merge remote-tracking branch 'upstream/main' into distributed_2.8
PenghuiCheng Apr 22, 2025
e558eaa
Enabled some UT cases of distributed
PenghuiCheng Apr 24, 2025
83ac56e
enable UT case in _shard and _tool folder
PenghuiCheng Apr 29, 2025
0e7a7b6
Fixed hard code error for world_size 8
PenghuiCheng May 5, 2025
a2b2fc6
merge from main branch
PenghuiCheng May 7, 2025
8de00b9
fix regex
daisyden May 8, 2025
91f5d10
Merge branch 'distributed_2.8' of https://github.com/daisyden/pytorch…
daisyden May 8, 2025
39e6c02
Fixed UT errors for cuda hard code
PenghuiCheng May 15, 2025
06d6c3e
Merge remote-tracking branch 'origin/distributed_2.8' into distribute…
PenghuiCheng May 15, 2025
a059005
Merge from upstream main branch
PenghuiCheng May 16, 2025
31ddfc0
Fixed XPU UT error for CUDA hard code
PenghuiCheng May 21, 2025
9a6df8a
Merge remote-tracking branch 'upstream0523/main' into distributed_2.8
daisyden May 23, 2025
50cb9e9
fix fsdp2 issue after rebase, fix #1618 dynamo issue
daisyden May 26, 2025
08559b9
remove duplicated device_type
daisyden May 27, 2025
f6a8c6a
fix rebase issue of test_fully_shard_overlap.py
daisyden May 27, 2025
dc0befa 10000
merge from main branch
PenghuiCheng May 30, 2025
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
Prev Previous commit
merge from main branch
Signed-off-by: Cheng, Penghui <penghui.cheng@intel.com>
  • Loading branch information
PenghuiCheng committed May 30, 2025
commit dc0befa75568dba81ceaa6fe0f11fac923c69867
41 changes: 22 additions & 19 deletions .ci/aarch64_linux/aarch64_wheel_ci_build.py
Original file line number Diff line number Diff line change
Expand Up @@ -31,27 +31,28 @@ def build_ArmComputeLibrary() -> None:
"build=native",
]
acl_install_dir = "/acl"
acl_checkout_dir = "ComputeLibrary"
os.makedirs(acl_install_dir)
check_call(
[
"git",
"clone",
"https://github.com/ARM-software/ComputeLibrary.git",
"-b",
"v25.02",
"--depth",
"1",
"--shallow-submodules",
]
)
acl_checkout_dir = os.getenv("ACL_SOURCE_DIR", "ComputeLibrary")
if os.path.isdir(acl_install_dir):
shutil.rmtree(acl_install_dir)
if not os.path.isdir(acl_checkout_dir) or not len(os.listdir(acl_checkout_dir)):
check_call(
[
"git",
"clone",
"https://github.com/ARM-software/ComputeLibrary.git",
"-b",
"v25.02",
"--depth",
"1",
"--shallow-submodules",
]
)

check_call(
["scons", "Werror=1", "-j8", f"build_dir=/{acl_install_dir}/build"]
+ acl_build_flags,
["scons", "Werror=1", f"-j{os.cpu_count()}"] + acl_build_flags,
cwd=acl_checkout_dir,
)
for d in ["arm_compute", "include", "utils", "support", "src"]:
for d in ["arm_compute", "include", "utils", "support", "src", "build"]:
shutil.copytree(f"{acl_checkout_dir}/{d}", f"{acl_install_dir}/{d}")


Expand Down Expand Up @@ -203,8 +204,10 @@ def parse_arguments():
).decode()

print("Building PyTorch wheel")
build_vars = "MAX_JOBS=5 CMAKE_SHARED_LINKER_FLAGS=-Wl,-z,max-page-size=0x10000 "
os.system("cd /pytorch; python setup.py clean")
build_vars = "CMAKE_SHARED_LINKER_FLAGS=-Wl,-z,max-page-size=0x10000 "
# MAX_JOB=5 is not required for CPU backend (see commit 465d98b)
if enable_cuda:
build_vars = "MAX_JOBS=5 " + build_vars

override_package_version = os.getenv("OVERRIDE_PACKAGE_VERSION")
desired_cuda = os.getenv("DESIRED_CUDA")
Expand Down
4 changes: 3 additions & 1 deletion .ci/docker/almalinux/Dockerfile
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
ARG CUDA_VERSION=12.4
ARG BASE_TARGET=cuda${CUDA_VERSION}
ARG ROCM_IMAGE=rocm/dev-almalinux-8:6.3-complete
FROM amd64/almalinux:8 as base
FROM amd64/almalinux:8.10-20250519 as base

ENV LC_ALL en_US.UTF-8
ENV LANG en_US.UTF-8
Expand All @@ -11,6 +11,8 @@ ARG DEVTOOLSET_VERSION=11

RUN yum -y update
RUN yum -y install epel-release
# install glibc-langpack-en make sure en_US.UTF-8 locale is available
RUN yum -y install glibc-langpack-en
RUN yum install -y sudo wget curl perl util-linux xz bzip2 git patch which perl zlib-devel openssl-devel yum-utils autoconf automake make gcc-toolset-${DEVTOOLSET_VERSION}-toolchain
# Just add everything as a safe.directory for git since these will be used in multiple places with git
RUN git config --global --add safe.directory '*'
Expand Down
16 changes: 8 additions & 8 deletions .ci/docker/build.sh
Original file line number Diff line number Diff line change
Expand Up @@ -109,8 +109,8 @@ case "$tag" in
UCC_COMMIT=${_UCC_COMMIT}
TRITON=yes
;;
pytorch-linux-focal-cuda12.8-cudnn9-py3-gcc9-inductor-benchmarks)
CUDA_VERSION=12.8.0
pytorch-linux-jammy-cuda12.8-cudnn9-py3-gcc9-inductor-benchmarks)
CUDA_VERSION=12.8
CUDNN_VERSION=9
ANACONDA_PYTHON_VERSION=3.10
GCC_VERSION=9
Expand All @@ -132,8 +132,8 @@ case "$tag" in
UCC_COMMIT 8000 =${_UCC_COMMIT}
TRITON=yes
;;
pytorch-linux-focal-cuda12.6-cudnn9-py3-gcc9-inductor-benchmarks)
CUDA_VERSION=12.6.3
pytorch-linux-jammy-cuda12.6-cudnn9-py3-gcc9-inductor-benchmarks)
CUDA_VERSION=12.6
CUDNN_VERSION=9
ANACONDA_PYTHON_VERSION=3.10
GCC_VERSION=9
Expand All @@ -144,8 +144,8 @@ case "$tag" in
TRITON=yes
INDUCTOR_BENCHMARKS=yes
;;
pytorch-linux-focal-cuda12.6-cudnn9-py3.12-gcc9-inductor-benchmarks)
CUDA_VERSION=12.6.3
pytorch-linux-jammy-cuda12.6-cudnn9-py3.12-gcc9-inductor-benchmarks)
CUDA_VERSION=12.6
CUDNN_VERSION=9
ANACONDA_PYTHON_VERSION=3.12
GCC_VERSION=9
Expand All @@ -156,8 +156,8 @@ case "$tag" in
TRITON=yes
INDUCTOR_BENCHMARKS=yes
;;
pytorch-linux-focal-cuda12.6-cudnn9-py3.13-gcc9-inductor-benchmarks)
CUDA_VERSION=12.6.3
pytorch-linux-jammy-cuda12.6-cudnn9-py3.13-gcc9-inductor-benchmarks)
CUDA_VERSION=12.6
CUDNN_VERSION=9
ANACONDA_PYTHON_VERSION=3.13
GCC_VERSION=9
Expand Down
6 changes: 3 additions & 3 deletions .ci/docker/common/install_cuda.sh
Original file line number Diff line number Diff line change
Expand Up @@ -183,9 +183,9 @@ function prune_126 {

function install_128 {
CUDNN_VERSION=9.8.0.87
echo "Installing CUDA 12.8.0 and cuDNN ${CUDNN_VERSION} and NCCL and cuSparseLt-0.6.3"
# install CUDA 12.8.0 in the same container
install_cuda 12.8.0 cuda_12.8.0_570.86.10_linux
echo "Installing CUDA 12.8.1 and cuDNN ${CUDNN_VERSION} and NCCL and cuSparseLt-0.6.3"
# install CUDA 12.8.1 in the same container
install_cuda 12.8.1 cuda_12.8.1_570.124.06_linux

# cuDNN license: https://developer.nvidia.com/cudnn/license_agreement
install_cudnn 12 $CUDNN_VERSION
Expand Down
3 changes: 1 addition & 2 deletions .ci/docker/common/install_onnx.sh
Original file line number Diff line number Diff line change
Expand Up @@ -31,8 +31,7 @@ pip_install \
pip_install coloredlogs packaging

pip_install onnxruntime==1.18.1
pip_install onnx==1.17.0
pip_install onnxscript==0.2.2 --no-deps
pip_install onnxscript==0.2.6 --no-deps
# required by onnxscript
pip_install ml_dtypes

Expand Down
16 changes: 8 additions & 8 deletions .ci/docker/requirements-ci.txt
Original file line number Diff line number Diff line change
Expand Up @@ -93,7 +93,7 @@ librosa>=0.6.2 ; python_version < "3.11"
#Pinned versions:
#test that import:

mypy==1.14.0
mypy==1.15.0
# Pin MyPy version because new errors are likely to appear with each release
#Description: linter
#Pinned versions: 1.14.0
Expand Down Expand Up @@ -166,10 +166,10 @@ pillow==11.0.0
#Pinned versions: 10.3.0
#test that import:

protobuf==3.20.2
#Description: Googles data interchange format
#Pinned versions: 3.20.1
#test that import: test_tensorboard.py
protobuf==5.29.4
#Description: Google's data interchange format
#Pinned versions: 5.29.4
#test that import: test_tensorboard.py, test/onnx/*

psutil
#Description: information on running processes and system utilization
Expand Down Expand Up @@ -337,12 +337,12 @@ sympy==1.13.3
#Pinned versions:
#test that import:

onnx==1.17.0
#Description: Required by mypy and test_public_bindings.py when checking torch.onnx._internal
onnx==1.18.0
#Description: Required by onnx tests, and mypy and test_public_bindings.py when checking torch.onnx._internal
#Pinned versions:
#test that import:

onnxscript==0.2.2
onnxscript==0.2.6
#Description: Required by mypy and test_public_bindings.py when checking torch.onnx._internal
#Pinned versions:
#test that import:
Expand Down
4 changes: 4 additions & 0 deletions .ci/docker/requirements-docs.txt
Original file line number Diff line number Diff line change
Expand Up @@ -15,6 +15,10 @@ sphinxext-opengraph==0.9.1
#Description: This is used to generate PyTorch docs
#Pinned versions: 0.9.1

sphinx_sitemap==2.6.0
#Description: This is used to generate sitemap for PyTorch docs
#Pinned versions: 2.6.0

matplotlib==3.5.3
#Description: This is used to generate PyTorch docs
#Pinned versions: 3.5.3
Expand Down
11 changes: 5 additions & 6 deletions .ci/manywheel/build_common.sh
Original file line number Diff line number Diff line change
Expand Up @@ -18,12 +18,10 @@ retry () {
$* || (sleep 1 && $*) || (sleep 2 && $*) || (sleep 4 && $*) || (sleep 8 && $*)
}

PLATFORM="manylinux2014_x86_64"
PLATFORM=""
# TODO move this into the Docker images
OS_NAME=$(awk -F= '/^NAME/{print $2}' /etc/os-release)
if [[ "$OS_NAME" == *"CentOS Linux"* ]]; then
retry yum install -q -y zip openssl
elif [[ "$OS_NAME" == *"AlmaLinux"* ]]; then
if [[ "$OS_NAME" == *"AlmaLinux"* ]]; then
retry yum install -q -y zip openssl
PLATFORM="manylinux_2_28_x86_64"
elif [[ "$OS_NAME" == *"Red Hat Enterprise Linux"* ]]; then
Expand All @@ -36,6 +34,9 @@ elif [[ "$OS_NAME" == *"Ubuntu"* ]]; then

retry apt-get update
retry apt-get -y install zip openssl
else
echo "Unknown OS: '$OS_NAME'"
exit 1
fi

# We use the package name to test the package by passing this to 'pip install'
Expand Down Expand Up @@ -79,8 +80,6 @@ if [[ -e /opt/openssl ]]; then
export CMAKE_INCLUDE_PATH="/opt/openssl/include":$CMAKE_INCLUDE_PATH
fi



mkdir -p /tmp/$WHEELHOUSE_DIR

export PATCHELF_BIN=/usr/local/bin/patchelf
Expand Down
71 changes: 26 additions & 45 deletions .ci/manywheel/build_cuda.sh
Original file line number Diff line number Diff line change
Expand Up @@ -36,10 +36,8 @@ if [[ -n "$DESIRED_CUDA" ]]; then
if [[ ${DESIRED_CUDA} =~ ^[0-9]+\.[0-9]+$ ]]; then
CUDA_VERSION=${DESIRED_CUDA}
else
# cu90, cu92, cu100, cu101
if [[ ${#DESIRED_CUDA} -eq 4 ]]; then
CUDA_VERSION="${DESIRED_CUDA:2:1}.${DESIRED_CUDA:3:1}"
elif [[ ${#DESIRED_CUDA} -eq 5 ]]; then
# cu126, cu128 etc...
if [[ ${#DESIRED_CUDA} -eq 5 ]]; then
CUDA_VERSION="${DESIRED_CUDA:2:2}.${DESIRED_CUDA:4:1}"
fi
fi
Expand All @@ -61,10 +59,6 @@ case ${CUDA_VERSION} in
TORCH_CUDA_ARCH_LIST="${TORCH_CUDA_ARCH_LIST};9.0"
EXTRA_CAFFE2_CMAKE_FLAGS+=("-DATEN_NO_TEST=ON")
;;
12.4)
TORCH_CUDA_ARCH_LIST="${TORCH_CUDA_ARCH_LIST};9.0"
EXTRA_CAFFE2_CMAKE_FLAGS+=("-DATEN_NO_TEST=ON")
;;
11.8)
TORCH_CUDA_ARCH_LIST="${TORCH_CUDA_ARCH_LIST};3.7;9.0"
EXTRA_CAFFE2_CMAKE_FLAGS+=("-DATEN_NO_TEST=ON")
Expand All @@ -91,14 +85,15 @@ fi
mkdir -p "$PYTORCH_FINAL_PACKAGE_DIR" || true

OS_NAME=$(awk -F= '/^NAME/{print $2}' /etc/os-release)
if [[ "$OS_NAME" == *"CentOS Linux"* ]]; then
LIBGOMP_PATH="/usr/lib64/libgomp.so.1"
elif [[ "$OS_NAME" == *"AlmaLinux"* ]]; then
if [[ "$OS_NAME" == *"AlmaLinux"* ]]; then
LIBGOMP_PATH="/usr/lib64/libgomp.so.1"
elif [[ "$OS_NAME" == *"Red Hat Enterprise Linux"* ]]; then
LIBGOMP_PATH="/usr/lib64/libgomp.so.1"
elif [[ "$OS_NAME" == *"Ubuntu"* ]]; then
LIBGOMP_PATH="/usr/lib/x86_64-linux-gnu/libgomp.so.1"
else
echo "Unknown OS: '$OS_NAME'"
exit 1
fi

DEPS_LIST=(
Expand All @@ -108,26 +103,8 @@ DEPS_SONAME=(
"libgomp.so.1"
)

# CUDA 11.8 have to ship the libcusparseLt.so.0 with the binary
# since nvidia-cusparselt-cu11 is not available in PYPI
if [[ $USE_CUSPARSELT == "1" && $CUDA_VERSION == "11.8" ]]; then
DEPS_SONAME+=(
"libcusparseLt.so.0"
)
DEPS_LIST+=(
"/usr/local/cuda/lib64/libcusparseLt.so.0"
)
fi


# Turn USE_CUFILE off for CUDA 11.8, 12.4 since nvidia-cufile-cu11 and 1.9.0.20 are
# not available in PYPI
if [[ $CUDA_VERSION == "11.8" || $CUDA_VERSION == "12.4" ]]; then
export USE_CUFILE=0
fi


# CUDA_VERSION 12.4, 12.6, 12.8
# CUDA_VERSION 12.6, 12.8
if [[ $CUDA_VERSION == 12* ]]; then
export USE_STATIC_CUDNN=0
# Try parallelizing nvcc as well
Expand All @@ -151,6 +128,8 @@ if [[ $CUDA_VERSION == 12* ]]; then
"/usr/local/cuda/lib64/libnvToolsExt.so.1"
"/usr/local/cuda/lib64/libnvrtc.so.12"
"/usr/local/cuda/lib64/libnvrtc-builtins.so"
"/usr/local/cuda/lib64/libcufile.so.0"
"/usr/local/cuda/lib64/libcufile_rdma.so.1"
)
DEPS_SONAME+=(
"libcudnn_adv.so.9"
Expand All @@ -168,17 +147,9 @@ if [[ $CUDA_VERSION == 12* ]]; then
"libnvToolsExt.so.1"
"libnvrtc.so.12"
"libnvrtc-builtins.so"
"libcufile.so.0"
"libcufile_rdma.so.1"
)
if [[ $USE_CUFILE == 1 ]]; then
DEPS_LIST+=(
"/usr/local/cuda/lib64/libcufile.so.0"
"/usr/local/cuda/lib64/libcufile_rdma.so.1"
)
DEPS_SONAME+=(
"libcufile.so.0"
"libcufile_rdma.so.1"
)
fi
else
echo "Using nvidia libs from pypi."
CUDA_RPATHS=(
Expand All @@ -194,12 +165,8 @@ if [[ $CUDA_VERSION == 12* ]]; then
'$ORIGIN/../../cusparselt/lib'
'$ORIGIN/../../nvidia/nccl/lib'
'$ORIGIN/../../nvidia/nvtx/lib'
'$ORIGIN/../../nvidia/cufile/lib'
)
if [[ $USE_CUFILE == 1 ]]; then
CUDA_RPATHS+=(
'$ORIGIN/../../nvidia/cufile/lib'
)
fi
CUDA_RPATHS=$(IFS=: ; echo "${CUDA_RPATHS[*]}")
export C_SO_RPATH=$CUDA_RPATHS':$ORIGIN:$ORIGIN/lib'
export LIB_SO_RPATH=$CUDA_RPATHS':$ORIGIN'
Expand All @@ -214,11 +181,25 @@ if [[ $CUDA_VERSION == 12* ]]; then
fi
elif [[ $CUDA_VERSION == "11.8" ]]; then
export USE_STATIC_CUDNN=0
# Turn USE_CUFILE off for CUDA 11.8 since nvidia-cufile-cu11 and 1.9.0.20 are
# not available in PYPI
export USE_CUFILE=0
# Try parallelizing nvcc as well
export TORCH_NVCC_FLAGS="-Xfatbin -compress-all --threads 2"
# Bundle ptxas into the wheel, see https://github.com/pytorch/pytorch/pull/119750
export BUILD_BUNDLE_PTXAS=1

# CUDA 11.8 have to ship the libcusparseLt.so.0 with the binary
# since nvidia-cusparselt-cu11 is not available in PYPI
if [[ $USE_CUSPARSELT == "1" ]]; then
DEPS_SONAME+=(
"libcusparseLt.so.0"
)
DEPS_LIST+=(
"/usr/local/cuda/lib64/libcusparseLt.so.0"
)
fi

if [[ -z "$PYTORCH_EXTRA_INSTALL_REQUIREMENTS" ]]; then
echo "Bundling with cudnn and cublas."
DEPS_LIST+=(
Expand Down
7 changes: 4 additions & 3 deletions .ci/manywheel/build_libtorch.sh
Original file line number Diff line number Diff line change
Expand Up @@ -22,9 +22,7 @@ retry () {

# TODO move this into the Docker images
OS_NAME=`awk -F= '/^NAME/{print $2}' /etc/os-release`
if [[ "$OS_NAME" == *"CentOS Linux"* ]]; then
retry yum install -q -y zip openssl
elif [[ "$OS_NAME" == *"AlmaLinux"* ]]; then
if [[ "$OS_NAME" == *"AlmaLinux"* ]]; then
retry yum install -q -y zip openssl
elif [[ "$OS_NAME" == *"Red Hat Enterprise Linux"* ]]; then
retry dnf install -q -y zip openssl
Expand All @@ -35,6 +33,9 @@ elif [[ "$OS_NAME" == *"Ubuntu"* ]]; then
sed -i 's/.*nvidia.*/# &/' $(find /etc/apt/ -type f -name "*.list")
retry apt-get update
retry apt-get -y install zip openssl
else
echo "Unknown OS: '$OS_NAME'"
exit 1
fi

# Version: setup.py uses $PYTORCH_BUILD_VERSION.post$PYTORCH_BUILD_NUMBER if
Expand Down
2 changes: 1 addition & 1 deletion .ci/pytorch/macos-build.sh
Original file line number Diff line number Diff line change
Expand Up @@ -40,7 +40,7 @@ if [[ ${BUILD_ENVIRONMENT} == *"distributed"* ]]; then
else
# Explicitly set USE_DISTRIBUTED=0 to align with the default build config on mac. This also serves as the sole CI config that tests
# that building with USE_DISTRIBUTED=0 works at all. See https://github.com/pytorch/pytorch/issues/86448
USE_DISTRIBUTED=0 USE_OPENMP=1 MACOSX_DEPLOYMENT_TARGET=11.0 WERROR=1 BUILD_TEST=OFF USE_PYTORCH_METAL=1 python setup.py bdist_wheel
USE_DISTRIBUTED=0 USE_OPENMP=1 MACOSX_DEPLOYMENT_TARGET=11.0 WERROR=1 BUILD_TEST=OFF USE_PYTORCH_METAL=1 python setup.py bdist_wheel --plat-name macosx_11_0_arm64
fi
if which sccache > /dev/null; then
print_sccache_stats
Expand Down
Loading
You are viewing a condensed version of this merge commit. You can view the full changes here.
0