10000 Update · pytorch/pytorch@e870dbc · GitHub
[go: up one dir, main page]

Skip to content

Commit e870dbc

Browse files
committed
Update
[ghstack-poisoned]
2 parents 12ba31f + 1b5cea2 commit e870dbc

File tree

908 files changed

+22640
-15462
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

908 files changed

+22640
-15462
lines changed

.ci/aarch64_linux/aarch64_wheel_ci_build.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -204,7 +204,7 @@ def parse_arguments():
204204
else:
205205
build_vars += f"BUILD_TEST=0 PYTORCH_BUILD_VERSION={version}.dev{build_date} PYTORCH_BUILD_NUMBER=1 "
206206
elif branch.startswith(("v1.", "v2.")):
207-
build_vars += f"BUILD_TEST=0 PYTORCH_BUILD_VERSION={branch[1:branch.find('-')]} PYTORCH_BUILD_NUMBER=1 "
207+
build_vars += f"BUILD_TEST=0 PYTORCH_BUILD_VERSION={branch[1 : branch.find('-')]} PYTORCH_BUILD_NUMBER=1 "
208208

209209
if enable_mkldnn:
210210
build_ArmComputeLibrary()

.ci/aarch64_linux/build_aarch64_wheel.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -761,7 +761,7 @@ def start_build(
761761
version = host.check_output("cat pytorch/version.txt").strip()[:-2]
762762
build_vars += f"BUILD_TEST=0 PYTORCH_BUILD_VERSION={version}.dev{build_date} PYTORCH_BUILD_NUMBER=1"
763763
if branch.startswith(("v1.", "v2.")):
764-
build_vars += f"BUILD_TEST=0 PYTORCH_BUILD_VERSION={branch[1:branch.find('-')]} PYTORCH_BUILD_NUMBER=1"
764+
build_vars += f"BUILD_TEST=0 PYTORCH_BUILD_VERSION={branch[1 : branch.find('-')]} PYTORCH_BUILD_NUMBER=1"
765765
if host.using_docker():
766766
build_vars += " CMAKE_SHARED_LINKER_FLAGS=-Wl,-z,max-page-size=0x10000"
767767
if enable_mkldnn:

.ci/docker/common/install_conda.sh

+1-1
Original file line numberDiff line numberDiff line change
@@ -66,7 +66,7 @@ if [ -n "$ANACONDA_PYTHON_VERSION" ]; then
6666

6767
# Install PyTorch conda deps, as per https://github.com/pytorch/pytorch README
6868
if [[ $(uname -m) == "aarch64" ]]; then
69-
conda_install "openblas==0.3.28=*openmp*"
69+
conda_install "openblas==0.3.29=*openmp*"
7070
else
7171
conda_install "mkl=2021.4.0 mkl-include=2021.4.0"
7272
fi

.ci/docker/common/install_openblas.sh

+1-1
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@
44
set -ex
55

66
cd /
7-
git clone https://github.com/OpenMathLib/OpenBLAS.git -b v0.3.28 --depth 1 --shallow-submodules
7+
git clone https://github.com/OpenMathLib/OpenBLAS.git -b v0.3.29 --depth 1 --shallow-submodules
88

99

1010
OPENBLAS_BUILD_FLAGS="

.ci/docker/common/install_rocm_drm.sh

+1-1
Original file line numberDiff line numberDiff line change
@@ -115,7 +115,7 @@ index a5007ffc..13fa07fc 100644
115115
if (!fp) {
116116
- fprintf(stderr, "%s: %s\n", AMDGPU_ASIC_ID_TABLE,
117117
- strerror(errno));
118-
+ fprintf(stderr, "amdgpu.ids: No such file or directory\n");
118+
+ //fprintf(stderr, "amdgpu.ids: No such file or directory\n");
119119
return;
120120
}
121121

.ci/docker/common/install_triton.sh

+3-3
Original file line numberDiff line numberDiff line change
@@ -60,15 +60,15 @@ if [ -n "${UBUNTU_VERSION}" ] && [ -n "${GCC_VERSION}" ] && [[ "${GCC_VERSION}"
6060
# Triton needs at least gcc-9 to build
6161
apt-get install -y g++-9
6262

63-
CXX=g++-9 pip_install -e .
63+
CXX=g++-9 pip_install .
6464
elif [ -n "${UBUNTU_VERSION}" ] && [ -n "${CLANG_VERSION}" ]; then
6565
# Triton needs <filesystem> which surprisingly is not available with clang-9 toolchain
6666
add-apt-repository -y ppa:ubuntu-toolchain-r/test
6767
apt-get install -y g++-9
6868

69-
CXX=g++-9 pip_install -e .
69+
CXX=g++-9 pip_install .
7070
else
71-
pip_install -e .
71+
pip_install .
7272
fi
7373

7474
if [ -n "${CONDA_CMAKE}" ]; then

.ci/docker/libtorch/build.sh

+1-1
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,7 @@ case ${GPU_ARCH_TYPE} in
3939
BASE_TARGET=rocm
4040
DOCKER_TAG=rocm${GPU_ARCH_VERSION}
4141
GPU_IMAGE=rocm/dev-ubuntu-20.04:${GPU_ARCH_VERSION}-complete
42-
PYTORCH_ROCM_ARCH="gfx900;gfx906;gfx908;gfx90a;gfx1030;gfx1100;gfx1101;gfx942"
42+
PYTORCH_ROCM_ARCH="gfx900;gfx906;gfx908;gfx90a;gfx1030;gfx1100;gfx1101;gfx1102;gfx942"
4343
DOCKER_GPU_BUILD_ARG="--build-arg PYTORCH_ROCM_ARCH=${PYTORCH_ROCM_ARCH}"
4444
;;
4545
*)

.ci/docker/manywheel/Dockerfile_2014

-153
This file was deleted.

.ci/docker/manywheel/build.sh

+3-2
Original file line numberDiff line numberDiff line change
@@ -97,7 +97,7 @@ case ${GPU_ARCH_TYPE} in
9797
DEVTOOLSET_VERSION="11"
9898
GPU_IMAGE=rocm/dev-almalinux-8:${GPU_ARCH_VERSION}-complete
9999
fi
100-
PYTORCH_ROCM_ARCH="gfx900;gfx906;gfx908;gfx90a;gfx942;gfx1030;gfx1100;gfx1101"
100+
PYTORCH_ROCM_ARCH="gfx900;gfx906;gfx908;gfx90a;gfx942;gfx1030;gfx1100;gfx1101;gfx1102"
101101
DOCKER_GPU_BUILD_ARG="--build-arg ROCM_VERSION=${GPU_ARCH_VERSION} --build-arg PYTORCH_ROCM_ARCH=${PYTORCH_ROCM_ARCH} --build-arg DEVTOOLSET_VERSION=${DEVTOOLSET_VERSION}"
102102
;;
103103
xpu)
@@ -121,7 +121,8 @@ fi
121121
(
122122
set -x
123123

124-
if [ "$(uname -m)" != "s390x" ]; then
124+
# Only activate this if in CI
125+
if [ "$(uname -m)" != "s390x" ] && [ -v CI ]; then
125126
# TODO: Remove LimitNOFILE=1048576 patch once https://github.com/pytorch/test-infra/issues/5712
126127
# is resolved. This patch is required in order to fix timing out of Docker build on Amazon Linux 2023.
127128
sudo sed -i s/LimitNOFILE=infinity/LimitNOFILE=1048576/ /usr/lib/systemd/system/docker.service

.ci/magma/Makefile

+1-1
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ DOCKER_RUN = set -eou pipefail; ${DOCKER_CMD} run --rm -i \
1212
-e PACKAGE_NAME=${PACKAGE_NAME}${DESIRED_CUDA_SHORT} \
1313
-e DESIRED_CUDA=${DESIRED_CUDA} \
1414
-e CUDA_ARCH_LIST="${CUDA_ARCH_LIST}" \
15-
"pytorch/manylinux-builder:cuda${DESIRED_CUDA}-main" \
15+
"pytorch/manylinux2_28-builder:cuda${DESIRED_CUDA}-main" \
1616
magma/build_magma.sh
1717

1818
.PHONY: all

.ci/manywheel/build_cuda.sh

+2-2
Original file line numberDiff line numberDiff line change
@@ -54,11 +54,11 @@ cuda_version_nodot=$(echo $CUDA_VERSION | tr -d '.')
5454
TORCH_CUDA_ARCH_LIST="5.0;6.0;7.0;7.5;8.0;8.6"
5555
case ${CUDA_VERSION} in
5656
12.8)
57-
TORCH_CUDA_ARCH_LIST="${TORCH_CUDA_ARCH_LIST};9.0;10.0;12.0+PTX" #Ripping out 5.0 and 6.0 due to ld error
57+
TORCH_CUDA_ARCH_LIST="7.0;7.5;8.0;8.6;9.0;10.0;12.0+PTX" #Ripping out 5.0 and 6.0 due to ld error
5858
EXTRA_CAFFE2_CMAKE_FLAGS+=("-DATEN_NO_TEST=ON")
5959
;;
6060
12.6)
61-
TORCH_CUDA_ARCH_LIST="${TORCH_CUDA_ARCH_LIST};9.0+PTX"
61+
TORCH_CUDA_ARCH_LIST="${TORCH_CUDA_ARCH_LIST};9.0"
6262
EXTRA_CAFFE2_CMAKE_FLAGS+=("-DATEN_NO_TEST=ON")
6363
;;
6464
12.4)

.ci/pytorch/build.sh

+6-3
Original file line numberDiff line numberDiff line change
@@ -173,6 +173,7 @@ if [[ "$BUILD_ENVIRONMENT" == *xpu* ]]; then
173173
source /opt/intel/oneapi/compiler/latest/env/vars.sh
174174
# XPU kineto feature dependencies are not fully ready, disable kineto build as temp WA
175175
export USE_KINETO=0
176+
export TORCH_XPU_ARCH_LIST=pvc
176177
fi
177178

178179
# sccache will fail for CUDA builds if all cores are used for compiling
@@ -191,7 +192,7 @@ fi
191192

192193
# We only build FlashAttention files for CUDA 8.0+, and they require large amounts of
193194
# memory to build and will OOM
194-
if [[ "$BUILD_ENVIRONMENT" == *cuda* ]] && [[ 1 -eq $(echo "${TORCH_CUDA_ARCH_LIST} >= 8.0" | bc) ]]; then
195+
if [[ "$BUILD_ENVIRONMENT" == *cuda* ]] && [[ 1 -eq $(echo "${TORCH_CUDA_ARCH_LIST} >= 8.0" | bc) ]] && [ -z "$MAX_JOBS_OVERRIDE" ]; then
195196
echo "WARNING: FlashAttention files require large amounts of memory to build and will OOM"
196197
echo "Setting MAX_JOBS=(nproc-2)/3 to reduce memory usage"
197198
export MAX_JOBS="$(( $(nproc --ignore=2) / 3 ))"
@@ -377,8 +378,10 @@ else
377378
# This is an attempt to mitigate flaky libtorch build OOM error. By default, the build parallelization
378379
# is set to be the number of CPU minus 2. So, let's try a more conservative value here. A 4xlarge has
379380
# 16 CPUs
380-
MAX_JOBS=$(nproc --ignore=4)
381-
export MAX_JOBS
381+
if [ -z "$MAX_JOBS_OVERRIDE" ]; then
382+
MAX_JOBS=$(nproc --ignore=4)
383+
export MAX_JOBS
384+
fi
382385

383386
# NB: Install outside of source directory (at the same level as the root
384387
# pytorch folder) so that it doesn't get cleaned away prior to docker push.

.ci/pytorch/smoke_test/max_autotune.py

+6-2
Original file line numberDiff line numberDiff line change
@@ -46,7 +46,9 @@ def train(args, model, device, train_loader, optimizer, epoch):
4646
optimizer.step()
4747
if batch_idx % args.log_interval == 0:
4848
print(
49-
f"Train Epoch: {epoch} [{batch_idx * len(data)}/{len(train_loader.dataset)} ({100. * batch_idx / len(train_loader):.0f}%)]\tLoss: {loss.item():.6f}" # noqa: B950
49+
f"Train Epoch: {epoch} "
50+
f"[{batch_idx * len(data)}/{len(train_loader.dataset)} "
51+
f"({100.0 * batch_idx / len(train_loader):.0f}%)]\tLoss: {loss.item():.6f}"
5052
)
5153
if args.dry_run:
5254
break
@@ -71,7 +73,9 @@ def test(model, device, test_loader):
7173
test_loss /= len(test_loader.dataset)
7274

7375
print(
74-
f"\nTest set: Average loss: {test_loss:.4f}, Accuracy: {correct}/{len(test_loader.dataset)} ({100. * correct / len(test_loader.dataset):.0f}%)\n" # noqa: B950
76+
f"\nTest set: Average loss: {test_loss:.4f}, "
77+
f"Accuracy: {correct}/{len(test_loader.dataset)} "
78+
f"({100.0 * correct / len(test_loader.dataset):.0f}%)\n"
7579
)
7680

7781

.ci/pytorch/smoke_test/smoke_test.py

+4
Original file line numberDiff line numberDiff line change
@@ -166,6 +166,10 @@ def test_cuda_gds_errors_captured() -> None:
166166
major_version = int(torch.version.cuda.split(".")[0])
167167
minor_version = int(torch.version.cuda.split(".")[1])
168168

169+
if target_os == "windows":
170+
print(f"{target_os} is not supported for GDS smoke test")
171+
return
172+
169173
if major_version < 12 or (major_version == 12 and minor_version < 6):
170174
print("CUDA version is not supported for GDS smoke test")
171175
return

0 commit comments

Comments
 (0)
0