Compare commits

...

49 Commits

Author SHA1 Message Date
f939415b29 Add shell script to automate run of categorize.py 2024-10-01 13:17:42 -07:00
7cee9bc341 Category updates for non frontend 2024-10-01 09:43:48 -07:00
6a6e192180 Categorized commits 2024-09-30 22:14:44 -07:00
636c463469 Addd commit list.csv 2024-09-30 19:08:21 -07:00
058d3de7b9 [dynamo] Do not treat user defined nn module attributes static for dynamic shape infra (#137025)
[dynamo] Do not treat user defined nn module attributes static for dynamic shape infra (#136516)

Fixes https://github.com/pytorch/pytorch/issues/136254

Th regression was introduced in https://github.com/pytorch/pytorch/pull/132736 where originally we were trying to fix another regression. This PR and the offending PR together say - "treat user defined nn module attributes as automatic dynamic, but for cudagraphs they will be considered static". This avoid recompilations. This can lead to a cudagraph recording, which is ok. This also maintains the state before inline_inbuilt_nn_modules flag was introduced.

Pull Request resolved: https://github.com/pytorch/pytorch/pull/136516
Approved by: https://github.com/williamwen42
2024-09-30 18:02:31 -07:00
17d25897b2 [FlexAttention] Fix output layout (#135882) (#136905)
We previously only supported the same v_head dim and + qk_head dim. When allowed for different head-dims I accidently kept the same query strides for the output. This PR fixes this bug as well it ensures that we always produce output in the same stride order as the input query.

Pull Request resolved: https://github.com/pytorch/pytorch/pull/135882
Approved by: https://github.com/yanboliang, https://github.com/Chillee

(cherry picked from commit ae02d663cdf493362699d2672ed7dc9019a7033b)
2024-09-30 18:01:24 -07:00
70298e91f9 [Cherry-pick][DSD] Fix distributed state dict full_state_dict option hang during set_state_dict (#135725) and Fix loading uneven full tensor into sharded state dict (#136365) (#136903)
* [DSD] Fix distributed state dict full_state_dict option hang during set_state_dict (#135725)

Fix https://github.com/pytorch/pytorch/issues/134095
This fix distributed state dict full_state_dict option hang during set_state_dict. We switch `_distribute_tensors` in _state_dict_utils.py to use `DTensor.from_local` instead of `distribute_tensor` to support FSDP2+TP 2D strided sharding use case, as `distribute_tensor` cannot handle strided sharding yet. `distribute_tensor` incurs a scatter behind the scenes, while `DTensor.from_local` takes the local slice from the full tensor on each rank to create the DTensor (no collective).  This means it's the user's responsibility to make sure the full_tensor from the full_state_dict is the same across all ranks.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/135725
Approved by: https://github.com/fegin

(cherry picked from commit 0cdc6a8dcd7e294b01d8914385bbe45e79c1770d)

* [DSD] Fix loading uneven full tensor into sharded state dict (#136365)

Fix #136228.

This is a follow up on https://github.com/pytorch/pytorch/pull/135725. We need to pass shape and stride from the original dtensor, since for uneven case, `from_local` would calculate shape and stride assuming the tensor is evenly-sharded based on the local tensor.

Pull Request resolved: https://github.com/pytorch/pytorch/pull/136365
Approved by: https://github.com/fegin

(cherry picked from commit 637d5c4b7eb0fb82e019eed29efc0aa6ba92dc24)
2024-09-30 16:49:45 -07:00
69ed7c7093 [SymmetricMemory] improve multicast initialization/fallback logic (#136894)
[SymmetricMemory] improve multicast initialization/fallback logic (#136577)

Fixes https://github.com/pytorch/pytorch/issues/136494

Currently, CUDASymmetricMemory::rendezvous() initializes a multicast address if multicast support is present. However, if we believe multicast support is present but cuMulticastCreate still fails for some reason, we do not fallback gracefully.

- In addition to CUDART and driver version check, query CU_DEVICE_ATTRIBUTE_MULTICAST_SUPPORTED to determine multicast support for a rank/device.
- Before initializing multicast for a block, ensure all ranks/devices have multicast support.
- This is unlikely, but if cuMulticastCreate still fails on rank 0, print the corresponding driver error message as a warning, and gracefully skip multicast initialization for the block.
- Introduced an environment variable (TORCH_SYMM_MEM_DISABLE_MULTICAST) to allow users to explicitly disable multicast support as a workaround.

Pull Request resolved: https://github.com/pytorch/pytorch/pull/136577
Approved by: https://github.com/Chillee, https://github.com/eqy

(cherry picked from commit d55eef5c596b3955dd8ee43c721b1c311dbab5e0)
2024-09-30 15:54:05 -07:00
d80f521ee2 fix requirements.txt installation failure issue on Windows (#136893) 2024-09-30 15:51:36 -07:00
57717c8768 Fix lint (#137052)
Fix lint (#137023)

By migrating some of the workflows to Python-3.9 as 3.8 has been deprecated by https://github.com/pytorch/pytorch/pull/132138

Pull Request resolved: https://github.com/pytorch/pytorch/pull/137023
Approved by: https://github.com/ZainRizvi, https://github.com/janeyx99, https://github.com/seemethere, https://github.com/kit1980, https://github.com/Skylion007

(cherry picked from commit 40f80a70fab672e96a088796fa1b742a70b93070)

Co-authored-by: Nikita Shulga <2453524+malfet@users.noreply.github.com>
2024-09-30 14:56:13 -07:00
eqy
550ed97a89 [cuDNN][SDPA] cherrypick Support attn_bias in cuDNN (#130482) (#136885)
[cuDNN][SDPA] Support `attn_bias` in cuDNN (#130482)

CC @drisspg

Pull Request resolved: https://github.com/pytorch/pytorch/pull/130482
Approved by: https://github.com/drisspg, https://github.com/Skylion007, https://github.com/malfet

Co-authored-by: Nikita Shulga <2453524+malfet@users.noreply.github.com>
2024-09-30 13:17:37 -07:00
051df20ac2 [ROCm] Update to AOTriton 0.7b (Cherry-picked) (#135869)
* Always create seed and offset tensors on GPU memory.

* Adjust fudge_factors for test_flash_attention_vs_math_ref_grads

* Skip enable_gqa=True tests

* Fix cudagraph support for FA backend

* Update the AOTriton FA API to meet hipGraph demands.

* Enable test_fused_attention_vs_math_ref_grads_cudagraph and skip seq_len_q != seq_len_k when is_causal=True

* The main FA and ME tests passed after heavily hacking the fudge factors...

* [SDPA] Add experimental support to Navi31

* Changes aotriton_version.txt to 0.7b release

* Make the fudge factors more explicit.

* Code clean up.

* Claim GQA is not supported on ROCM in can_use_flash_attention

* Switch to .gz package

* Skip failures on test/test_native_mha.py

* Skip more GQA tests

* Skip nn_functional_scaled_dot_product_attention related tests

* Disable Efficient attention on fp32 + is_casual=True

* Revert "Disable Efficient attention on fp32 + is_casual=True"

This reverts commit 36324a49d2c322146adbd678902fa32d008b8b8b.

It's not very effective and forcing MATH backend does not help. Need
further investigations.

* Add missing imports

* Disable test_transformerencoderlayer and test_transformerdecoder

* Fix two more problems

* Fix lint

* Re-enable test_transformers

* Skip some tests in test_multiheadattention_fastpath_attn_mask on ROCM

* fix lint

* skip test_pointwise_associative_scan_tuple_reverse_True_combine_mode_pointwise_cuda on ROCm

* skip more test_pointwise_associative_scan

* Fix per suggestions from Nikita

* Update skip reason of test_transformerencoderlayer

* Add missing using
2024-09-30 13:14:55 -07:00
bc421d456e [RELEASE ONLY CHANGES] Revert XNNPACK Update (#136522)
* Revert "Update generate-xnnpack-wrappers.py parsing to handle build identifier (#134724)"

This reverts commit 679b8fe426babd40f8f9b013392df7a52901c926.

* Revert "Update PyTorch for XNNPACK 87ee0b4 (#134518)"

This reverts commit 3b40b07efbc13422d59fff83ffaddc5141a5fa5e.
2024-09-27 12:58:59 -07:00
aa574ab7e3 SDPA regression fix to work around high-precision by default (#136536)
Add option to configure reduced precision math backend for SDPA (#135964)

Summary: Address https://github.com/pytorch/pytorch/issues/135778 by adding a global flag to configure whether using high precision or low precision for math backend of SDPA.

Test Plan: buck2 run mode/opt //scripts/feikou/llm:run_attn_kernels

Differential Revision: D62625515

Pull Request resolved: https://github.com/pytorch/pytorch/pull/135964
Approved by: https://github.com/jbschlosser

(cherry picked from commit 0a35986cdb72383c4e173abbdfc140ae2f98d5b1)

Co-authored-by: Jianyu Huang <jianyuhuang@meta.com>
2024-09-27 09:44:04 -07:00
24bd87d5dd [Docs] fix inconsistent docs in conv1d, conv2d, and conv3d (#136813)
[Docs] fix inconsistent docs in conv1d, conv2d, and conv3d (#135894)

Addresses https://github.com/pytorch/pytorch/issues/135880
Pull Request resolved: https://github.com/pytorch/pytorch/pull/135894
Approved by: https://github.com/mikaylagawarecki, https://github.com/malfet

(cherry picked from commit c9de2efde6ba0e1f15fe3ea99646855fdd9debaa)

Co-authored-by: Sahan Paliskara <sahanp@meta.com>
2024-09-26 18:59:00 -07:00
6101aafa34 [Update] Update note for Getting Started with PyTorch on Intel GPUs (#136731)
[Update] Update note for Getting Started with PyTorch on Intel GPUs (#129946)

remove the hardware and software prerequisites and set up env part.
keep the prerequisites section and link to pytorch prerequistes for intel gpus for driver install, intel support package install and env set up
https://www.intel.com/content/www/us/en/developer/articles/tool/pytorch-prerequisites-for-intel-gpus.html
Update the support for Intel Client GPU MTL-H
Update inference & training examples

Pull Request resolved: https://github.com/pytorch/pytorch/pull/129946
Approved by: https://github.com/seemethere

(cherry picked from commit f3dd1721f48e503145669843a8fe90e0ad57d9f7)

Co-authored-by: Zheng, Zhaoqiong <zhaoqiong.zheng@intel.com>
2024-09-26 17:50:08 -07:00
396413f05c Fix ROCm skip decorator for test_ddp_tp and multiprocess UTs (#136161) (#136801)
skip_if_rocm is used only in multiprocess case (when UT test class is a child of MultiProcessTestCase). Each individual process can exit with a skip code. If used for single process UT, it will cause the UT to fail as the process returns a non-zero exit code. Use skipIfRocm in single process UTs.

To avoid the above confusion, this PR renamed skip_if_rocm to skip_if_rocm_multiprocess.

Fixes #ISSUE_NUMBER

Pull Request resolved: https://github.com/pytorch/pytorch/pull/136161
Approved by: https://github.com/jithunnair-amd, https://github.com/kwen2501, https://github.com/fegin

Co-authored-by: Prachi Gupta <prachi.gupta@amd.com>
2024-09-26 16:09:37 -07:00
c25781c5d2 Update current maintainers (#136769)
Update current maintainers (#136672)

This file didn't had an overall in a few years so long overdue. Most of the credit goes to @orionr for gathering all of this info.

The main rules we followed:
- No code contributor is removed, they're all placed as emeritus
- Breakdown too big categories to make this document useful to know who to ping
- No category where the code is still in the codebase is removed
- We did not rework the categories (for example to be closer to module: labels) and leave that for later
- All non-emeritus names are ordered by their number of comments on issues related to their topic
Pull Request resolved: https://github.com/pytorch/pytorch/pull/136672
Approved by: https://github.com/eqy, https://github.com/ezyang, https://github.com/seemethere, https://github.com/malfet

(cherry picked from commit 2421344d8f582084b69b7b00fe0304b1c9732f65)

Co-authored-by: albanD <desmaison.alban@gmail.com>
2024-09-26 11:45:15 -07:00
ecd330669e Constraint setuptools to 72.1.0 or older in requirements.txt (#136729)
Constraint setuptools to 72.1.0 or older in requirements.txt (#136489)

FIXES: https://github.com/pytorch/pytorch/issues/136541

Setuptools>=74.0.0 has deprecated support for some functions in distutils, and so the builds run into error such as ```AttributeError: module 'distutils' has no attribute '_msvccompiler'```. Also, the pytorch builds have setuptools pin to 72.1.0 according to these PRs: https://github.com/pytorch/builder/pull/1995 and 89d9a8cf6f. So, until there is a fix to change the function usage in accordance with latest setuptools, the 72.1.0 version works fine.

Also observed in CI jobs: https://github.com/pytorch/pytorch/actions/runs/10979326524
Pull Request resolved: https://github.com/pytorch/pytorch/pull/136489
Approved by: https://github.com/malfet

Co-authored-by: Nikita Shulga <2453524+malfet@users.noreply.github.com>
(cherry picked from commit 8d65d9f11bbbcb0ea4152510ffe9784135cfacc8)

Co-authored-by: ratnampa <ratnam.parikh@intel.com>
2024-09-26 00:05:27 -07:00
1715708183 Revert "Trace fwd graph under no_grad mode #134872" (#136734)
This reverts commit aca4d6b891ac1b7296195283edb8b64c828a6351.
2024-09-25 23:10:28 -07:00
2e2c00f74c Make test_skip_data_serialization regex more flexible (#136710)
Make test_skip_data_serialization regex more flexible (#136580)

Some CI machines seem to throw "Can't get local object" rather than
"Can't pickle local object".
Pull Request resolved: https://github.com/pytorch/pytorch/pull/136580
Approved by: https://github.com/mikaylagawarecki

(cherry picked from commit a0c76ea8533a4327fa36dfeeaf0863749d5e6dca)

Co-authored-by: Jez Ng <jezng@fb.com>
2024-09-25 18:33:27 -07:00
cbe476a5a7 Disable iOS workflow (#136706)
Disable iOS workflow (#136571)

See https://github.com/pytorch/pytorch/issues/136284
It's been broken for more than a week and it does not seem like anyone cares about fixing it.
Once it's landed I'll reassigned the issue on `oncall: mobile`
Pull Request resolved: https://github.com/pytorch/pytorch/pull/136571
Approved by: https://github.com/huydhn, https://github.com/kit1980

(cherry picked from commit 5340feb8aa97fe0a70251a52f2c9b570adf7dbe0)

Co-authored-by: Nikita Shulga <2453524+malfet@users.noreply.github.com>
2024-09-25 18:18:23 -07:00
4b030d47b1 [RELEASE-ONLY CHANGES] Don't push to https://ghcr.io/ (#136703)
Don't do push to https://ghcr.io/ on release branch: we don't need it and it fails with "unauthorized: unauthenticated: User cannot be authenticated with the token provided".
2024-09-25 18:03:12 -07:00
9b80ddecd6 Fix hardcoded ROCm paths in Caffe2Targets.cmake (#136700)
Fix hardcoded ROCm paths in `Caffe2Targets.cmake` (#136283)

Fixes #131701

Use CMake imported targets more consistently to eliminate hardcode paths.

Here is the new relevant sections of Caffe2Targets.cmake:
```
set_target_properties(c10_hip PROPERTIES
  INTERFACE_INCLUDE_DIRECTORIES "${_IMPORT_PREFIX}/include"
  INTERFACE_LINK_LIBRARIES "c10;hip::amdhip64"
)
```

```
set_target_properties(torch_hip PROPERTIES
  INTERFACE_COMPILE_DEFINITIONS "USE_C10D_NCCL"
  INTERFACE_COMPILE_OPTIONS "-fPIC;-D__HIP_PLATFORM_AMD__=1;-DCUDA_HAS_FP16=1;-DUSE_ROCM;-D__HIP_NO_HALF_OPERATORS__=1;-D__HIP_NO_HALF_CONVERSIONS__=1;-DTORCH_HIP_VERSION=602;-Wno-shift-count-negative;-Wno-shift-count-overflow;-Wno-duplicate-decl-specifier;-DCAFFE2_USE_MIOPEN;-DTHRUST_DEVICE_SYSTEM=THRUST_DEVICE_SYSTEM_HIP;-std=c++17;-DHIPBLAS_V2;-DHIP_NEW_TYPE_ENUMS"
  INTERFACE_INCLUDE_DIRECTORIES "${_IMPORT_PREFIX}/include"
  INTERFACE_LINK_LIBRARIES "c10_hip;torch_cpu_library;hip::amdhip64;MIOpen;hiprtc::hiprtc;roc::hipblaslt;roc::hipblas;hip::hipfft;hip::hiprand;roc::hipsparse;roc::hipsolver"
)
```

HIPCUB dependency was not actually used; which is why it is removed here as the imported target had undesirable side effects.

Pull Request resolved: https://github.com/pytorch/pytorch/pull/136283
Approved by: https://github.com/jeffdaily, https://github.com/Skylion007, https://github.com/jithunnair-amd, https://github.com/atalman

(cherry picked from commit e8f1dd6ba0675d3e11808e5198b0d927278a6f91)

Co-authored-by: Nichols A. Romero <nick.romero@amd.com>
2024-09-25 17:45:39 -07:00
6e86793f75 [ROCm] upgrade ROCm CI builds to py3.10 (#134108) (#136696)
Upgrade ROCm CI builds to py3.10

Pull Request resolved: https://github.com/pytorch/pytorch/pull/134108
Approved by: https://github.com/jeffdaily, https://github.com/jithunnair-amd, https://github.com/atalman

Co-authored-by: Jack Taylor <jack.taylor@amd.com>
2024-09-25 17:17:30 -07:00
7c550fea95 [ROCm][CI] upgrade CI to ROCm 6.2 (#132555) (#136467)
Pull Request resolved: https://github.com/pytorch/pytorch/pull/132555
Approved by: https://github.com/pruthvistony, https://github.com/malfet
2024-09-25 13:42:27 -07:00
7a00785c23 [ROCm] Cherry-pick unit test fixes to release/2.5 (#136557)
* [ROCm] skip test_fp8_cast_and_t on non-MI300 machines (#135917)

Fixes #ISSUE_NUMBER

Pull Request resolved: https://github.com/pytorch/pytorch/pull/135917
Approved by: https://github.com/malfet

(cherry picked from commit 6cdc70bccd369d0d7ec4b54d921c91d62bd931f7)

* Skip pointwise associative scan tests due to regression (changes based on PR https://github.com/pytorch/pytorch/pull/135995)

* Cherry-pick fix from https://github.com/pytorch/pytorch/pull/135702

---------

Co-authored-by: Prachi Gupta <prachi.gupta@amd.com>
Co-authored-by: Jithun Nair <jithun.nair@amd.com>
2024-09-25 16:12:47 -04:00
4e6a99e5f3 fix stride compare failed when size value equal to one in ForeachUtils.h (#136426)
fix stride compare failed when size value equal to one in ForeachUtils.h (#134546)

When size value equal to one, tensor strides value need be skipped to compare.
@ezyang
Pull Request resolved: https://github.com/pytorch/pytorch/pull/134546
Approved by: https://github.com/janeyx99
2024-09-25 12:26:29 -07:00
dd73223b90 [ROCm] [BUGFIX] Re-enable rocm-specific tuning parameters v2 (#133852) (#136139)
Small bug fix - https://github.com/pytorch/pytorch/pull/124592 replaced the torch.version.hip with device_props but made a mistake in porting the original logic.

The original code was:
`if torch.version.hip is not None:`

Which was incorrectly replaced by:
`if self.device_props.type != "hip":`

Another occurence of https://github.com/pytorch/pytorch/pull/130617

Pull Request resolved: https://github.com/pytorch/pytorch/pull/133852
Approved by: https://github.com/masnesral, https://github.com/malfet

(cherry picked from commit da587de9cb7cdbcd69f65783cd7a9589198bd6f6)
2024-09-25 11:56:24 -07:00
c5e5254a79 Fix xpu memory stats error (#135818) (#136420)
# Motivation
fix https://github.com/pytorch/pytorch/issues/135726
After merging two free blocks, I made a stupid mistake of ignoring the correct size to decrease the active memory size, which should be the original block size instead of the merged block size.

# Additional Context
Add a UT to guard this scenario.

Pull Request resolved: https://github.com/pytorch/pytorch/pull/135818
Approved by: https://github.com/EikanWang

(cherry picked from commit e6b68359d7c86aff25eefe77e0774c02b38f44b4)
2024-09-24 13:58:37 -04:00
ffed7b71e8 Fix dynamo benchmark skip logic for cpu device (#135193) (#135793)
Fixes #132380, adjust torchbench and huggingface skip models list, then we can remove `--no-skip` when running benchmarks on 3 suites.

Pull Request resolved: https://github.com/pytorch/pytorch/pull/135193
Approved by: https://github.com/chuanqi129, https://github.com/jansel

(cherry picked from commit 7ec17b49cf89cfeb97272a1baddcc30fa6fa66d8)
2024-09-20 15:47:42 -07:00
becdf8ae4f [Inductor] Increase multiplier to 3 for Inductor AMP FP16 benchmark correctness check (#135932) (#136262)
Fix https://github.com/pytorch/pytorch/issues/135657.
Aligned with AMP BF16, using multiplier 3 for Inductor AMP FP16 benchmark correctness check

Pull Request resolved: https://github.com/pytorch/pytorch/pull/135932
Approved by: https://github.com/CaoE, https://github.com/jgong5, https://github.com/jansel
2024-09-20 15:45:26 -07:00
fb276d2652 [ONNX] Fix numpy method to return the correct type (#136162) (#136203)
[ONNX] Fix numpy method to return the correct type (#136162)

Previous implementation of the `numpy()` method returns `fp64` when the tensor is `fp32`. This is unexpected but seems to be caused by calling `__array__(dtype=None)` on the numpy array. I updated the implementation to implement the `numpy()` method explicitly and added tests to guard the behavior.

This needs to be cherry-picked into torch 2.5
Pull Request resolved: https://github.com/pytorch/pytorch/pull/136162
Approved by: https://github.com/gramalingam, https://github.com/xadupre

(cherry picked from commit 67b14ce8bd9d4d0ad1920e57bc148644775646ac)
2024-09-20 15:44:05 -07:00
b7de7932fd Update torch-xpu-ops pin (ATen XPU implementation) (#135833)
Update torch-xpu-ops pin (ATen XPU implementation) (#135647)

Release cycle for PyTorch 2.5
1. Fixing runtime error on Windows: Fail to load torch_xpu_ops_unary_binary_kernels.dll as the bin size is large.

Pull Request resolved: https://github.com/pytorch/pytorch/pull/135647
Approved by: https://github.com/EikanWang
2024-09-20 15:39:53 -07:00
1954439802 Update document for autocast on CPU (#136082)
update document for autocast on CPU
2024-09-20 15:28:26 -07:00
3920988456 [Cherry-pick] [ONNX] Update fake mode usage in onnx docs (#135512) and Drop final None values as inputs for nodes in exporter graph (#135520) (#136005)
* [ONNX] Update fake mode usage in onnx docs (#135512)

Update fake mode usage in onnx docs
Pull Request resolved: https://github.com/pytorch/pytorch/pull/135512
Approved by: https://github.com/justinchuby

Co-authored-by: Justin Chu <justinchuby@users.noreply.github.com>
(cherry picked from commit 66db61f0d1b55323af958c3401d74842dc00b0c7)

* [ONNX] Drop final None values as inputs for nodes in exporter graph (#135520)

When value for an optional input is not provided, it is defaulted to `None`, which gets translates to "" in the onnx graph. To avoid this, if we have a list of inputs and the final few are all `None`, strip them in the graph.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/135520
Approved by: https://github.com/justinchuby

Co-authored-by: Justin Chu <justinchuby@users.noreply.github.com>
(cherry picked from commit e2f9a83b85af1bca54d97d2cc1259d72bf9579e5)
2024-09-20 15:25:25 -07:00
1db2a6562c [DCP] Fixes the stateless optimizer issue of distributed state_dict (… (#136000)
[DCP] Fixes the stateless optimizer issue of distributed state_dict (#135535)

Some optimizers don't have states that can cause get_state_dict/set_state_dict behave incorrectly. This PR fixes the issues.

fixes: https://github.com/pytorch/pytorch/issues/133415

Pull Request resolved: https://github.com/pytorch/pytorch/pull/135535
Approved by: https://github.com/wz337

Co-authored-by: Chien-Chin Huang <chienchin@fb.com>
2024-09-20 15:24:10 -07:00
4b5bf41476 [inductor] [cpp] fix the input contiguous check in max-autotune (#135561)
[inductor] [cpp] fix the input contiguous check in max-autotune (#134982)

## Description
Fixes the FP32 accuracy failure of `resmlp_12_224` and BF16 accuracy failure of `volo_d1_224` in timm.

In this PR, we check whether input is contiguous using the following way:
If it has `FixedLayout`, we know the accurate strides. For `FlexibleLayout`, if its data is a `ComputedBuffer`, we could get the fill order of the buffer to decide whether it's contiguous. For the other cases, we won't use GEMM template as we can't infer whether it's contiguous.

## Additional context
The current GEMM template only supports this case: `input.get_stride()[-1] == 1`. In `resmlp_12_224`, when we run into this check, the layout of `input` is a `FlexibleLayout`. The reason is that when realizing the input which is a `View` IR, the `convert_to_reinterpret_view` call fails:
d14fe3ffed/torch/_inductor/ir.py (L4712-L4715)

And it finally runs into this `copy_input` and returns a `FlexibleLayout`.
d14fe3ffed/torch/_inductor/ir.py (L4722)

When checking its stride, this `FlexibleLayout` indeed satisfies `input.get_stride()[-1] == 1` but it is later decided as a `FixedLayout` with `size = (3072, 196), stride = (1, 3072)`, which is not supported by the GEMM template, thus causing accuracy issue in this model.
The `FlexibleLayout` is converted to `FixedLayout` during [CppPackedGemmTemplate.add_choices](d14fe3ffed/torch/_inductor/mkldnn_lowerings.py (L1051)) which calls [slice_nd](d14fe3ffed/torch/_inductor/codegen/cpp_template_kernel.py (L150)) when rendering the kernel (`slice_nd(X)`). When creating the `SliceView` IR, [as_storage_and_layout](d14fe3ffed/torch/_inductor/ir.py (L2288)) invokes
[decide_layout](d14fe3ffed/torch/_inductor/ir.py (L2135)) and converts it to a `FixedLayout` with `size = (3072, 196), stride = (1, 3072)`.

Pull Request resolved: https://github.com/pytorch/pytorch/pull/134982
Approved by: https://github.com/jgong5, https://github.com/leslie-fang-intel, https://github.com/jansel
2024-09-20 15:20:16 -07:00
a889c85498 [Inductor] simplify indexing_exprs in LoopBody._init_with_copy (#135574) (#135935)
This PR uses `var_ranges` information to simplify `indexing_exprs` in `LoopBody._init_with_copy` to to reduce occurrences of `FloorDiv` and `ModularIndexing` in the `indexing_exprs`.

Pull Request resolved: https://github.com/pytorch/pytorch/pull/135574
Approved by: https://github.com/jgong5, https://github.com/leslie-fang-intel, https://github.com/jansel
2024-09-19 17:03:24 -07:00
813e06461c [ONNX] Fix symbolic values and numpy implementation (#135786) (#135868)
1. Remove `__eq__` to make `SymbolicTensor` hashable and test for that
2. Update the `__array__` method so that it works for tensor on GPU

Fixes https://github.com/pytorch/pytorch/issues/135700
Pull Request resolved: https://github.com/pytorch/pytorch/pull/135786
Approved by: https://github.com/titaiwangms
2024-09-19 16:58:05 -07:00
9887030485 Revert "[fx] Bypass custom __setattr__ in Node.__init__ (#135079)" (#… (#135625)
Revert "[fx] Bypass custom __setattr__ in Node.__init__ (#135079)" (#135562)

This reverts commit 66da3b3b2acacb116a9b23e91b24934830eaf6b8.

#135079 breaks internal tests and needs to be reverted. Revert with mergebot doesn't work as this PR is technically part of the stack, but, according to @jansel, it should be possible to revert it individually.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/135562
Approved by: https://github.com/jansel, https://github.com/seemethere

Co-authored-by: Ivan Zaitsev <ivanzaitsev@fb.com>
2024-09-19 16:50:49 -07:00
9e315fef22 Revert "[Release only] Temporary disable triton xpu build" (#136276)
Revert "[Release only] Temporary disable triton xpu build (#136206)"

This reverts commit 6b14e6cfddc62f71a65f250c58d2482fdf2126b4.
2024-09-18 11:17:52 -07:00
6b14e6cfdd [Release only] Temporary disable triton xpu build (#136206) 2024-09-17 12:12:09 -07:00
828d686e1c [ONNX] Fix scaled_dot_product_attention with float scale (#135710)
Fixes #125158

Pull Request resolved: https://github.com/pytorch/pytorch/pull/135594
Approved by: https://github.com/justinchuby

(cherry picked from commit e48ee2cf50d86d87ef7c7d0839267dbed4903ebf)
2024-09-12 09:50:58 -07:00
612fc7c447 [ONNX] Improves documentation of ONNX exporter (#135526)
The PR updates the documentation to reflect the changes introduced in pytorch 2.5 and related to onnx exporter.

Pull Request resolved: https://github.com/pytorch/pytorch/pull/135372
Approved by: https://github.com/justinchuby

Co-authored-by: Justin Chu <justinchuby@users.noreply.github.com>
(cherry picked from commit 5e145861f23370d342125118dd9b0dcba26147db)

Co-authored-by: Xavier Dupré <xadupre@users.noreply.github.com>
2024-09-12 09:49:45 -07:00
cea562006e [Cherry-Pick] Bump triton pin and release version, revert temporary changes to build from pin (#135613)
* Revert "[RELEASE-ONLY CHANGES] Temp changes to build triton from pin for first RC (#135517)"

This reverts commit 4a3dabd67f8ce63f2fc45f278421cca3cc532cfe.

* Build triton from release branch

* triton_pin

* fix

* Bump triton xpu pin and release version (#135638)

Similar with https://github.com/pytorch/pytorch/pull/135627

Pull Request resolved: https://github.com/pytorch/pytorch/pull/135638
Approved by: https://github.com/atalman

---------

Co-authored-by: chuanqiw <chuanqi.wang@intel.com>
2024-09-11 15:03:44 -04:00
ba27502501 Use upload-artifact@v4.4.0 for create_release.yml (#135534)
Use upload-artifact@v4.4.0 for create_release.yml (#135528)

Fixes failure: https://github.com/pytorch/pytorch/actions/runs/10780281005/job/29895846007

Due broken sync
```
actions/upload-artifact@v2
and
actions/download-artifact@v4.1.7
```

Pull Request resolved: https://github.com/pytorch/pytorch/pull/135528
Approved by: https://github.com/kit1980, https://github.com/malfet

(cherry picked from commit 9b764491e32755d64efd4e7ff1c90cf587f3e3ff)

Co-authored-by: atalman <atalman@fb.com>
2024-09-10 14:43:16 -04:00
4a3dabd67f [RELEASE-ONLY CHANGES] Temp changes to build triton from pin for first RC (#135517)
[Release only] Temp changes to build triton from pin
2024-09-09 15:08:05 -04:00
e130696270 [RELEASE-ONLY CHANGES] Branch Cut for Release 2.5 (#135506)
* [RELEASE-ONLY CHANGES] Branch Cut for Release 2.5

* fix_lint
2024-09-09 14:05:58 -04:00
201 changed files with 7368 additions and 1941 deletions

View File

@ -1,5 +1,5 @@
0.6b 0.7b
manylinux_2_17 manylinux_2_17
rocm6.2 rocm6.2
7f07e8a1cb1f99627eb6d77f5c0e9295c775f3c7 9be04068c3c0857a4cfd17d7e39e71d0423ebac2
e4ab195d2bd19e939c675a13280c29714c6ef9f2cf420690da150fa0cac043b1 3e9e1959d23b93d78a08fcc5f868125dc3854dece32fd9458be9ef4467982291

View File

@ -286,18 +286,7 @@ case "$image" in
TRITON=yes TRITON=yes
;; ;;
pytorch-linux-focal-rocm-n-1-py3) pytorch-linux-focal-rocm-n-1-py3)
ANACONDA_PYTHON_VERSION=3.8 ANACONDA_PYTHON_VERSION=3.10
GCC_VERSION=9
PROTOBUF=yes
DB=yes
VISION=yes
ROCM_VERSION=6.0
NINJA_VERSION=1.9.0
CONDA_CMAKE=yes
TRITON=yes
;;
pytorch-linux-focal-rocm-n-py3)
ANACONDA_PYTHON_VERSION=3.8
GCC_VERSION=9 GCC_VERSION=9
PROTOBUF=yes PROTOBUF=yes
DB=yes DB=yes
@ -307,6 +296,17 @@ case "$image" in
CONDA_CMAKE=yes CONDA_CMAKE=yes
TRITON=yes TRITON=yes
;; ;;
pytorch-linux-focal-rocm-n-py3)
ANACONDA_PYTHON_VERSION=3.10
GCC_VERSION=9
PROTOBUF=yes
DB=yes
VISION=yes
ROCM_VERSION=6.2
NINJA_VERSION=1.9.0
CONDA_CMAKE=yes
TRITON=yes
;;
pytorch-linux-jammy-xpu-2024.0-py3) pytorch-linux-jammy-xpu-2024.0-py3)
ANACONDA_PYTHON_VERSION=3.9 ANACONDA_PYTHON_VERSION=3.9
GCC_VERSION=11 GCC_VERSION=11

View File

@ -1 +1 @@
cc981feba10a3f4c2e46f3fe368e8fcf5f5643df 91b14bf5593cf58a8541f3e6b9125600a867d4ef

View File

@ -1 +1 @@
757b6a61e7df814ba806f498f8bb3160f84b120c 5fe38ffd73c2ac6ed6323b554205186696631c6f

View File

@ -4,12 +4,12 @@ set -ex
source "$(dirname "${BASH_SOURCE[0]}")/common_utils.sh" source "$(dirname "${BASH_SOURCE[0]}")/common_utils.sh"
TARBALL='aotriton.tar.bz2' TARBALL='aotriton.tar.gz'
# This read command alwasy returns with exit code 1 # This read command alwasy returns with exit code 1
read -d "\n" VER MANYLINUX ROCMBASE PINNED_COMMIT SHA256 < aotriton_version.txt || true read -d "\n" VER MANYLINUX ROCMBASE PINNED_COMMIT SHA256 < aotriton_version.txt || true
ARCH=$(uname -m) ARCH=$(uname -m)
AOTRITON_INSTALL_PREFIX="$1" AOTRITON_INSTALL_PREFIX="$1"
AOTRITON_URL="https://github.com/ROCm/aotriton/releases/download/${VER}/aotriton-${VER}-${MANYLINUX}_${ARCH}-${ROCMBASE}-shared.tar.bz2" AOTRITON_URL="https://github.com/ROCm/aotriton/releases/download/${VER}/aotriton-${VER}-${MANYLINUX}_${ARCH}-${ROCMBASE}-shared.tar.gz"
cd "${AOTRITON_INSTALL_PREFIX}" cd "${AOTRITON_INSTALL_PREFIX}"
# Must use -L to follow redirects # Must use -L to follow redirects

View File

@ -10,6 +10,21 @@ if [[ -z $ROCM_VERSION ]]; then
exit 1; exit 1;
fi fi
IS_UBUNTU=0
ID=$(grep -oP '(?<=^ID=).+' /etc/os-release | tr -d '"')
case "$ID" in
ubuntu)
IS_UBUNTU=1
;;
centos)
IS_UBUNTU=0
;;
*)
echo "Unable to determine OS..."
exit 1
;;
esac
# To make version comparison easier, create an integer representation. # To make version comparison easier, create an integer representation.
save_IFS="$IFS" save_IFS="$IFS"
IFS=. ROCM_VERSION_ARRAY=(${ROCM_VERSION}) IFS=. ROCM_VERSION_ARRAY=(${ROCM_VERSION})
@ -58,8 +73,7 @@ MIOPEN_CMAKE_COMMON_FLAGS="
" "
# Pull MIOpen repo and set DMIOPEN_EMBED_DB based on ROCm version # Pull MIOpen repo and set DMIOPEN_EMBED_DB based on ROCm version
if [[ $ROCM_INT -ge 60200 ]] && [[ $ROCM_INT -lt 60300 ]]; then if [[ $ROCM_INT -ge 60200 ]] && [[ $ROCM_INT -lt 60300 ]]; then
echo "ROCm 6.2 MIOpen does not need any patches, do not build from source" MIOPEN_BRANCH="release/rocm-rel-6.2-staging"
exit 0
elif [[ $ROCM_INT -ge 60100 ]] && [[ $ROCM_INT -lt 60200 ]]; then elif [[ $ROCM_INT -ge 60100 ]] && [[ $ROCM_INT -lt 60200 ]]; then
echo "ROCm 6.1 MIOpen does not need any patches, do not build from source" echo "ROCm 6.1 MIOpen does not need any patches, do not build from source"
exit 0 exit 0
@ -93,12 +107,21 @@ else
exit 1 exit 1
fi fi
yum remove -y miopen-hip
if [[ ${IS_UBUNTU} == 1 ]]; then
apt-get remove -y miopen-hip
else
yum remove -y miopen-hip
fi
git clone https://github.com/ROCm/MIOpen -b ${MIOPEN_BRANCH} git clone https://github.com/ROCm/MIOpen -b ${MIOPEN_BRANCH}
pushd MIOpen pushd MIOpen
# remove .git to save disk space since CI runner was running out # remove .git to save disk space since CI runner was running out
rm -rf .git rm -rf .git
# Don't build CK to save docker build time
if [[ $ROCM_INT -ge 60200 ]]; then
sed -i '/composable_kernel/d' requirements.txt
fi
# Don't build MLIR to save docker build time # Don't build MLIR to save docker build time
# since we are disabling MLIR backend for MIOpen anyway # since we are disabling MLIR backend for MIOpen anyway
if [[ $ROCM_INT -ge 50400 ]] && [[ $ROCM_INT -lt 50500 ]]; then if [[ $ROCM_INT -ge 50400 ]] && [[ $ROCM_INT -lt 50500 ]]; then
@ -111,10 +134,15 @@ cmake -P install_deps.cmake --minimum
# clean up since CI runner was running out of disk space # clean up since CI runner was running out of disk space
rm -rf /tmp/* rm -rf /tmp/*
yum clean all if [[ ${IS_UBUNTU} == 1 ]]; then
rm -rf /var/cache/yum apt-get autoclean && apt-get clean
rm -rf /var/lib/yum/yumdb rm -rf /var/lib/apt/lists/* /tmp/* /var/tmp/*
rm -rf /var/lib/yum/history else
yum clean all
rm -rf /var/cache/yum
rm -rf /var/lib/yum/yumdb
rm -rf /var/lib/yum/history
fi
## Build MIOpen ## Build MIOpen
mkdir -p build mkdir -p build
@ -131,7 +159,11 @@ make -j $(nproc) package
# clean up since CI runner was running out of disk space # clean up since CI runner was running out of disk space
rm -rf /usr/local/cget rm -rf /usr/local/cget
yum install -y miopen-*.rpm if [[ ${IS_UBUNTU} == 1 ]]; then
sudo dpkg -i miopen-hip*.deb
else
yum install -y miopen-*.rpm
fi
popd popd
rm -rf MIOpen rm -rf MIOpen

View File

@ -1 +1 @@
3.0.0 3.1.0

View File

@ -68,6 +68,8 @@ RUN rm install_rocm.sh
COPY ./common/install_rocm_magma.sh install_rocm_magma.sh COPY ./common/install_rocm_magma.sh install_rocm_magma.sh
RUN bash ./install_rocm_magma.sh RUN bash ./install_rocm_magma.sh
RUN rm install_rocm_magma.sh RUN rm install_rocm_magma.sh
ADD ./common/install_miopen.sh install_miopen.sh
RUN bash ./install_miopen.sh ${ROCM_VERSION} && rm install_miopen.sh
ENV ROCM_PATH /opt/rocm ENV ROCM_PATH /opt/rocm
ENV PATH /opt/rocm/bin:$PATH ENV PATH /opt/rocm/bin:$PATH
ENV PATH /opt/rocm/hcc/bin:$PATH ENV PATH /opt/rocm/hcc/bin:$PATH
@ -121,5 +123,8 @@ RUN bash ./install_cache.sh && rm install_cache.sh
ARG BUILD_ENVIRONMENT ARG BUILD_ENVIRONMENT
ENV BUILD_ENVIRONMENT ${BUILD_ENVIRONMENT} ENV BUILD_ENVIRONMENT ${BUILD_ENVIRONMENT}
# Install LLVM dev version (Defined in the pytorch/builder github repository)
COPY --from=pytorch/llvm:9.0.1 /opt/llvm /opt/llvm
USER jenkins USER jenkins
CMD ["bash"] CMD ["bash"]

View File

@ -49,13 +49,8 @@ if [[ ${BUILD_ENVIRONMENT} == *"parallelnative"* ]]; then
fi fi
# Enable LLVM dependency for TensorExpr testing # Enable LLVM dependency for TensorExpr testing
if [[ "$BUILD_ENVIRONMENT" == *rocm* ]]; then export USE_LLVM=/opt/llvm
export USE_LLVM=/opt/rocm/llvm export LLVM_DIR=/opt/llvm/lib/cmake/llvm
export LLVM_DIR=/opt/rocm/llvm/lib/cmake/llvm
else
export USE_LLVM=/opt/llvm
export LLVM_DIR=/opt/llvm/lib/cmake/llvm
fi
if [[ "$BUILD_ENVIRONMENT" == *executorch* ]]; then if [[ "$BUILD_ENVIRONMENT" == *executorch* ]]; then
# To build test_edge_op_registration # To build test_edge_op_registration

View File

@ -198,7 +198,7 @@ function install_torchrec_and_fbgemm() {
function clone_pytorch_xla() { function clone_pytorch_xla() {
if [[ ! -d ./xla ]]; then if [[ ! -d ./xla ]]; then
git clone --recursive --quiet https://github.com/pytorch/xla.git git clone --recursive -b r2.5 https://github.com/pytorch/xla.git
pushd xla pushd xla
# pin the xla hash so that we don't get broken by changes to xla # pin the xla hash so that we don't get broken by changes to xla
git checkout "$(cat ../.github/ci_commit_pins/xla.txt)" git checkout "$(cat ../.github/ci_commit_pins/xla.txt)"

View File

@ -78,7 +78,7 @@ TRITON_VERSION=$(cat $PYTORCH_ROOT/.ci/docker/triton_version.txt)
TRITON_CONSTRAINT="platform_system == 'Linux' and platform_machine == 'x86_64' and python_version < '3.13'" TRITON_CONSTRAINT="platform_system == 'Linux' and platform_machine == 'x86_64' and python_version < '3.13'"
if [[ "$PACKAGE_TYPE" =~ .*wheel.* && -n "${PYTORCH_EXTRA_INSTALL_REQUIREMENTS:-}" ]]; then if [[ "$PACKAGE_TYPE" =~ .*wheel.* && -n "${PYTORCH_EXTRA_INSTALL_REQUIREMENTS:-}" ]]; then
# Only linux Python < 3.13 are supported wheels for triton # Only linux Python < 3.13 are supported wheels for triton
TRITON_REQUIREMENT="triton==${TRITON_VERSION}; ${TRITON_CONSTRAINT}" TRITON_REQUIREMENT="pytorch-triton==${TRITON_VERSION}; ${TRITON_CONSTRAINT}"
if [[ -n "$PYTORCH_BUILD_VERSION" && "$PYTORCH_BUILD_VERSION" =~ .*dev.* ]]; then if [[ -n "$PYTORCH_BUILD_VERSION" && "$PYTORCH_BUILD_VERSION" =~ .*dev.* ]]; then
TRITON_SHORTHASH=$(cut -c1-10 $PYTORCH_ROOT/.ci/docker/ci_commit_pins/triton.txt) TRITON_SHORTHASH=$(cut -c1-10 $PYTORCH_ROOT/.ci/docker/ci_commit_pins/triton.txt)
TRITON_REQUIREMENT="pytorch-triton==${TRITON_VERSION}+${TRITON_SHORTHASH}; ${TRITON_CONSTRAINT}" TRITON_REQUIREMENT="pytorch-triton==${TRITON_VERSION}+${TRITON_SHORTHASH}; ${TRITON_CONSTRAINT}"

View File

@ -1 +1 @@
2eb4a60ed14a38260b85b0c765161f0ce45be6d1 r2.5

View File

@ -39,9 +39,9 @@ SUPPORTED_PERIODICAL_MODES: Dict[str, Callable[[Optional[str]], bool]] = {
} }
# The link to the published list of disabled jobs # The link to the published list of disabled jobs
DISABLED_JOBS_URL = "https://ossci-metrics.s3.amazonaws.com/disabled-jobs.json" DISABLED_JOBS_URL = "https://ossci-metrics.s3.amazonaws.com/disabled-jobs.json?versionId=sxzMTP57qj.Vwz8dN1glkTK560Txq9W3"
# and unstable jobs # and unstable jobs
UNSTABLE_JOBS_URL = "https://ossci-metrics.s3.amazonaws.com/unstable-jobs.json" UNSTABLE_JOBS_URL = "https://ossci-metrics.s3.amazonaws.com/unstable-jobs.json?versionId=8f1.4S3MupuHXH8t0waxyGnPsGHJYdv9"
# Some constants used to handle disabled and unstable jobs # Some constants used to handle disabled and unstable jobs
JOB_NAME_SEP = "/" JOB_NAME_SEP = "/"

View File

@ -8,7 +8,7 @@
# NOTE: If testing pytorch/builder changes you can change this variable to change what pytorch/builder reference # NOTE: If testing pytorch/builder changes you can change this variable to change what pytorch/builder reference
# the binary builds will check out # the binary builds will check out
{%- set builder_repo = "pytorch/builder" -%} {%- set builder_repo = "pytorch/builder" -%}
{%- set builder_branch = "main" -%} {%- set builder_branch = "release/2.5" -%}
{%- macro concurrency(build_environment) -%} {%- macro concurrency(build_environment) -%}
concurrency: concurrency:
@ -36,7 +36,7 @@ concurrency:
{%- macro setup_ec2_windows() -%} {%- macro setup_ec2_windows() -%}
!{{ display_ec2_information() }} !{{ display_ec2_information() }}
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}

View File

@ -142,10 +142,10 @@ jobs:
with: with:
name: !{{ config["build_name"] }} name: !{{ config["build_name"] }}
path: "${{ runner.temp }}/artifacts/" path: "${{ runner.temp }}/artifacts/"
!{{ common.checkout(deep_clone=False, directory="pytorch") }} !{{ common.checkout(deep_clone=False, directory="pytorch", checkout_pr_head=False) }}
!{{ common.checkout(deep_clone=False, directory="builder", repository=common.builder_repo, branch=common.builder_branch) }} !{{ common.checkout(deep_clone=False, directory="builder", repository=common.builder_repo, branch=common.builder_branch, checkout_pr_head=False) }}
- name: Pull Docker image - name: Pull Docker image
uses: pytorch/test-infra/.github/actions/pull-docker-image@main uses: pytorch/test-infra/.github/actions/pull-docker-image@release/2.5
with: with:
docker-image: !{{ config["container_image"] }} docker-image: !{{ config["container_image"] }}
- name: Test Pytorch binary - name: Test Pytorch binary
@ -164,13 +164,13 @@ jobs:
with: with:
name: !{{ config["build_name"] }} name: !{{ config["build_name"] }}
path: "${{ runner.temp }}/artifacts/" path: "${{ runner.temp }}/artifacts/"
!{{ common.checkout(deep_clone=False, directory="pytorch") }} !{{ common.checkout(deep_clone=False, directory="pytorch", checkout_pr_head=False) }}
!{{ common.checkout(deep_clone=False, directory="builder", repository=common.builder_repo, branch=common.builder_branch) }} !{{ common.checkout(deep_clone=False, directory="builder", repository=common.builder_repo, branch=common.builder_branch, checkout_pr_head=False) }}
- name: ROCm set GPU_FLAG - name: ROCm set GPU_FLAG
run: | run: |
echo "GPU_FLAG=--device=/dev/mem --device=/dev/kfd --device=/dev/dri --group-add video --group-add daemon" >> "${GITHUB_ENV}" echo "GPU_FLAG=--device=/dev/mem --device=/dev/kfd --device=/dev/dri --group-add video --group-add daemon" >> "${GITHUB_ENV}"
- name: Pull Docker image - name: Pull Docker image
uses: pytorch/test-infra/.github/actions/pull-docker-image@main uses: pytorch/test-infra/.github/actions/pull-docker-image@release/2.5
with: with:
docker-image: !{{ config["container_image"] }} docker-image: !{{ config["container_image"] }}
- name: Test Pytorch binary - name: Test Pytorch binary

View File

@ -78,8 +78,8 @@ jobs:
elif [ -d "/Applications/Xcode_13.3.1.app" ]; then elif [ -d "/Applications/Xcode_13.3.1.app" ]; then
echo "DEVELOPER_DIR=/Applications/Xcode_13.3.1.app/Contents/Developer" >> "${GITHUB_ENV}" echo "DEVELOPER_DIR=/Applications/Xcode_13.3.1.app/Contents/Developer" >> "${GITHUB_ENV}"
fi fi
!{{ common.checkout(deep_clone=False, directory="pytorch") }} !{{ common.checkout(deep_clone=False, directory="pytorch", checkout_pr_head=False) }}
!{{ common.checkout(deep_clone=False, directory="builder", repository=common.builder_repo, branch=common.builder_branch) }} !{{ common.checkout(deep_clone=False, directory="builder", repository=common.builder_repo, branch=common.builder_branch, checkout_pr_head=False) }}
- name: Install sccache (only for non-forked PRs, and pushes to trunk) - name: Install sccache (only for non-forked PRs, and pushes to trunk)
uses: nick-fields/retry@v3.0.0 uses: nick-fields/retry@v3.0.0
if: ${{ github.event_name == 'push' || github.event.pull_request.head.repo.full_name == github.repository }} if: ${{ github.event_name == 'push' || github.event.pull_request.head.repo.full_name == github.repository }}

View File

@ -79,8 +79,8 @@ jobs:
steps: steps:
!{{ common.setup_ec2_windows() }} !{{ common.setup_ec2_windows() }}
!{{ set_runner_specific_vars() }} !{{ set_runner_specific_vars() }}
!{{ common.checkout(deep_clone=False, directory="pytorch") }} !{{ common.checkout(deep_clone=False, directory="pytorch", checkout_pr_head=False) }}
!{{ common.checkout(deep_clone=False, directory="builder", repository=common.builder_repo, branch=common.builder_branch) }} !{{ common.checkout(deep_clone=False, directory="builder", repository=common.builder_repo, branch=common.builder_branch, checkout_pr_head=False) }}
- name: Populate binary env - name: Populate binary env
shell: bash shell: bash
run: | run: |
@ -121,8 +121,8 @@ jobs:
with: with:
name: !{{ config["build_name"] }} name: !{{ config["build_name"] }}
path: "${{ env.PYTORCH_FINAL_PACKAGE_DIR }}" path: "${{ env.PYTORCH_FINAL_PACKAGE_DIR }}"
!{{ common.checkout(deep_clone=False, directory="pytorch") }} !{{ common.checkout(deep_clone=False, directory="pytorch", checkout_pr_head=False) }}
!{{ common.checkout(deep_clone=False, directory="builder", repository=common.builder_repo, branch=common.builder_branch) }} !{{ common.checkout(deep_clone=False, directory="builder", repository=common.builder_repo, branch=common.builder_branch, checkout_pr_head=False) }}
- name: Populate binary env - name: Populate binary env
shell: bash shell: bash
run: | run: |

View File

@ -37,7 +37,7 @@ jobs:
keep-going: ${{ steps.filter.outputs.keep-going }} keep-going: ${{ steps.filter.outputs.keep-going }}
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
fetch-depth: 1 fetch-depth: 1
submodules: false submodules: false
@ -59,25 +59,25 @@ jobs:
runs-on: ${{ matrix.runner }} runs-on: ${{ matrix.runner }}
steps: steps:
- name: Setup SSH (Click me for login details) - name: Setup SSH (Click me for login details)
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
# [see note: pytorch repo ref] # [see note: pytorch repo ref]
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
- name: Setup Linux - name: Setup Linux
uses: ./.github/actions/setup-linux uses: ./.github/actions/setup-linux
- name: Calculate docker image - name: Calculate docker image
id: calculate-docker-image id: calculate-docker-image
uses: pytorch/test-infra/.github/actions/calculate-docker-image@main uses: pytorch/test-infra/.github/actions/calculate-docker-image@release/2.5
with: with:
docker-image-name: ${{ inputs.docker-image-name }} docker-image-name: ${{ inputs.docker-image-name }}
- name: Pull docker image - name: Pull docker image
uses: pytorch/test-infra/.github/actions/pull-docker-image@main uses: pytorch/test-infra/.github/actions/pull-docker-image@release/2.5
with: with:
docker-image: ${{ steps.calculate-docker-image.outputs.docker-image }} docker-image: ${{ steps.calculate-docker-image.outputs.docker-image }}
@ -141,5 +141,5 @@ jobs:
if: always() if: always()
- name: Teardown Linux - name: Teardown Linux
uses: pytorch/test-infra/.github/actions/teardown-linux@main uses: pytorch/test-infra/.github/actions/teardown-linux@release/2.5
if: always() if: always()

View File

@ -37,7 +37,7 @@ jobs:
keep-going: ${{ steps.filter.outputs.keep-going }} keep-going: ${{ steps.filter.outputs.keep-going }}
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
fetch-depth: 1 fetch-depth: 1
submodules: false submodules: false
@ -59,25 +59,25 @@ jobs:
runs-on: ${{ matrix.runner }} runs-on: ${{ matrix.runner }}
steps: steps:
- name: Setup SSH (Click me for login details) - name: Setup SSH (Click me for login details)
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
# [see note: pytorch repo ref] # [see note: pytorch repo ref]
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
- name: Setup Linux - name: Setup Linux
uses: ./.github/actions/setup-linux uses: ./.github/actions/setup-linux
- name: Calculate docker image - name: Calculate docker image
id: calculate-docker-image id: calculate-docker-image
uses: pytorch/test-infra/.github/actions/calculate-docker-image@main uses: pytorch/test-infra/.github/actions/calculate-docker-image@release/2.5
with: with:
docker-image-name: ${{ inputs.docker-image-name }} docker-image-name: ${{ inputs.docker-image-name }}
- name: Pull docker image - name: Pull docker image
uses: pytorch/test-infra/.github/actions/pull-docker-image@main uses: pytorch/test-infra/.github/actions/pull-docker-image@release/2.5
with: with:
docker-image: ${{ steps.calculate-docker-image.outputs.docker-image }} docker-image: ${{ steps.calculate-docker-image.outputs.docker-image }}
@ -186,5 +186,5 @@ jobs:
if: always() if: always()
- name: Teardown Linux - name: Teardown Linux
uses: pytorch/test-infra/.github/actions/teardown-linux@main uses: pytorch/test-infra/.github/actions/teardown-linux@release/2.5
if: always() if: always()

View File

@ -47,7 +47,7 @@ jobs:
reenabled-issues: ${{ steps.filter.outputs.reenabled-issues }} reenabled-issues: ${{ steps.filter.outputs.reenabled-issues }}
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
fetch-depth: 1 fetch-depth: 1
submodules: false submodules: false
@ -69,25 +69,25 @@ jobs:
runs-on: ${{ matrix.runner }} runs-on: ${{ matrix.runner }}
steps: steps:
- name: Setup SSH (Click me for login details) - name: Setup SSH (Click me for login details)
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
# [see note: pytorch repo ref] # [see note: pytorch repo ref]
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
- name: Setup Linux - name: Setup Linux
uses: ./.github/actions/setup-linux uses: ./.github/actions/setup-linux
- name: Calculate docker image - name: Calculate docker image
id: calculate-docker-image id: calculate-docker-image
uses: pytorch/test-infra/.github/actions/calculate-docker-image@main uses: pytorch/test-infra/.github/actions/calculate-docker-image@release/2.5
with: with:
docker-image-name: ${{ inputs.docker-image-name }} docker-image-name: ${{ inputs.docker-image-name }}
- name: Pull docker image - name: Pull docker image
uses: pytorch/test-infra/.github/actions/pull-docker-image@main uses: pytorch/test-infra/.github/actions/pull-docker-image@release/2.5
with: with:
docker-image: ${{ steps.calculate-docker-image.outputs.docker-image }} docker-image: ${{ steps.calculate-docker-image.outputs.docker-image }}
@ -97,7 +97,7 @@ jobs:
run: echo "IN_ARC_RUNNER=$([ -f /.inarc ] && echo true || echo false)" >> "$GITHUB_OUTPUT" run: echo "IN_ARC_RUNNER=$([ -f /.inarc ] && echo true || echo false)" >> "$GITHUB_OUTPUT"
- name: Install nvidia driver, nvidia-docker runtime, set GPU_FLAG - name: Install nvidia driver, nvidia-docker runtime, set GPU_FLAG
uses: pytorch/test-infra/.github/actions/setup-nvidia@main uses: pytorch/test-infra/.github/actions/setup-nvidia@release/2.5
if: ${{ inputs.cuda-version != 'cpu' && steps.check_arc_runner.outputs.IN_ARC_RUNNER == 'false' }} if: ${{ inputs.cuda-version != 'cpu' && steps.check_arc_runner.outputs.IN_ARC_RUNNER == 'false' }}
- name: Output disk space left - name: Output disk space left
@ -206,5 +206,5 @@ jobs:
file-suffix: bazel-${{ github.job }}_${{ steps.get-job-id.outputs.job-id }} file-suffix: bazel-${{ github.job }}_${{ steps.get-job-id.outputs.job-id }}
- name: Teardown Linux - name: Teardown Linux
uses: pytorch/test-infra/.github/actions/teardown-linux@main uses: pytorch/test-infra/.github/actions/teardown-linux@release/2.5
if: always() if: always()

View File

@ -159,13 +159,13 @@ jobs:
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
if: inputs.build_environment != 'linux-s390x-binary-manywheel' if: inputs.build_environment != 'linux-s390x-binary-manywheel'
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.github-token }} github-secret: ${{ secrets.github-token }}
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
no-sudo: ${{ inputs.build_environment == 'linux-aarch64-binary-manywheel' || inputs.build_environment == 'linux-s390x-binary-manywheel' }} no-sudo: ${{ inputs.build_environment == 'linux-aarch64-binary-manywheel' || inputs.build_environment == 'linux-s390x-binary-manywheel' }}
@ -195,7 +195,6 @@ jobs:
- name: Checkout PyTorch to pytorch dir - name: Checkout PyTorch to pytorch dir
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -209,7 +208,7 @@ jobs:
- name: Checkout pytorch/builder to builder dir - name: Checkout pytorch/builder to builder dir
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -235,7 +234,7 @@ jobs:
- name: Pull Docker image - name: Pull Docker image
if: ${{ steps.filter.outputs.is-test-matrix-empty == 'False' && inputs.build_environment != 'linux-s390x-binary-manywheel' }} if: ${{ steps.filter.outputs.is-test-matrix-empty == 'False' && inputs.build_environment != 'linux-s390x-binary-manywheel' }}
uses: pytorch/test-infra/.github/actions/pull-docker-image@main uses: pytorch/test-infra/.github/actions/pull-docker-image@release/2.5
with: with:
docker-image: ${{ inputs.DOCKER_IMAGE }} docker-image: ${{ inputs.DOCKER_IMAGE }}
@ -293,7 +292,7 @@ jobs:
- name: Teardown Linux - name: Teardown Linux
if: always() && inputs.build_environment != 'linux-s390x-binary-manywheel' if: always() && inputs.build_environment != 'linux-s390x-binary-manywheel'
uses: pytorch/test-infra/.github/actions/teardown-linux@main uses: pytorch/test-infra/.github/actions/teardown-linux@release/2.5
- name: Chown workspace - name: Chown workspace
if: always() && inputs.build_environment != 'linux-s390x-binary-manywheel' if: always() && inputs.build_environment != 'linux-s390x-binary-manywheel'

View File

@ -142,14 +142,14 @@ jobs:
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
if: inputs.build_environment != 'linux-s390x-binary-manywheel' if: inputs.build_environment != 'linux-s390x-binary-manywheel'
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.github-token }} github-secret: ${{ secrets.github-token }}
# Setup the environment # Setup the environment
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
no-sudo: ${{ inputs.build_environment == 'linux-aarch64-binary-manywheel' || inputs.build_environment == 'linux-s390x-binary-manywheel' }} no-sudo: ${{ inputs.build_environment == 'linux-aarch64-binary-manywheel' || inputs.build_environment == 'linux-s390x-binary-manywheel' }}
@ -172,7 +172,6 @@ jobs:
- name: Checkout PyTorch to pytorch dir - name: Checkout PyTorch to pytorch dir
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
@ -185,7 +184,7 @@ jobs:
- name: Checkout pytorch/builder to builder dir - name: Checkout pytorch/builder to builder dir
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -216,12 +215,12 @@ jobs:
path: "${{ runner.temp }}/artifacts/" path: "${{ runner.temp }}/artifacts/"
- name: Install nvidia driver, nvidia-docker runtime, set GPU_FLAG - name: Install nvidia driver, nvidia-docker runtime, set GPU_FLAG
uses: pytorch/test-infra/.github/actions/setup-nvidia@main uses: pytorch/test-infra/.github/actions/setup-nvidia@release/2.5
if: ${{ inputs.GPU_ARCH_TYPE == 'cuda' && steps.filter.outputs.is-test-matrix-empty == 'False' }} if: ${{ inputs.GPU_ARCH_TYPE == 'cuda' && steps.filter.outputs.is-test-matrix-empty == 'False' }}
- name: Pull Docker image - name: Pull Docker image
if: ${{ steps.filter.outputs.is-test-matrix-empty == 'False' && inputs.build_environment != 'linux-s390x-binary-manywheel' }} if: ${{ steps.filter.outputs.is-test-matrix-empty == 'False' && inputs.build_environment != 'linux-s390x-binary-manywheel' }}
uses: pytorch/test-infra/.github/actions/pull-docker-image@main uses: pytorch/test-infra/.github/actions/pull-docker-image@release/2.5
with: with:
docker-image: ${{ inputs.DOCKER_IMAGE }} docker-image: ${{ inputs.DOCKER_IMAGE }}
@ -231,7 +230,7 @@ jobs:
- name: Teardown Linux - name: Teardown Linux
if: always() && inputs.build_environment != 'linux-s390x-binary-manywheel' if: always() && inputs.build_environment != 'linux-s390x-binary-manywheel'
uses: pytorch/test-infra/.github/actions/teardown-linux@main uses: pytorch/test-infra/.github/actions/teardown-linux@release/2.5
- name: Chown workspace - name: Chown workspace
if: always() && inputs.build_environment != 'linux-s390x-binary-manywheel' if: always() && inputs.build_environment != 'linux-s390x-binary-manywheel'

View File

@ -103,7 +103,7 @@ jobs:
USE_SPLIT_BUILD: ${{ inputs.use_split_build }} USE_SPLIT_BUILD: ${{ inputs.use_split_build }}
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
no-sudo: true no-sudo: true

View File

@ -28,7 +28,7 @@ jobs:
keep-going: ${{ steps.filter.outputs.keep-going }} keep-going: ${{ steps.filter.outputs.keep-going }}
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
fetch-depth: 1 fetch-depth: 1
submodules: false submodules: false
@ -49,7 +49,7 @@ jobs:
runs-on: ${{ matrix.runner }} runs-on: ${{ matrix.runner }}
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
- name: Set up JDK 8 - name: Set up JDK 8
uses: actions/setup-java@v3 uses: actions/setup-java@v3
@ -58,7 +58,7 @@ jobs:
distribution: 'temurin' distribution: 'temurin'
- name: Setup miniconda - name: Setup miniconda
uses: pytorch/test-infra/.github/actions/setup-miniconda@main uses: pytorch/test-infra/.github/actions/setup-miniconda@release/2.5
with: with:
python-version: 3.8 python-version: 3.8
environment-file: .github/requirements/conda-env-${{ runner.os }}-${{ runner.arch }} environment-file: .github/requirements/conda-env-${{ runner.os }}-${{ runner.arch }}

View File

@ -84,7 +84,7 @@ jobs:
name: build-docs-${{ matrix.docs_type }}-${{ inputs.push }} name: build-docs-${{ matrix.docs_type }}-${{ inputs.push }}
steps: steps:
- name: Setup SSH (Click me for login details) - name: Setup SSH (Click me for login details)
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
instructions: | instructions: |
@ -95,7 +95,7 @@ jobs:
# [see note: pytorch repo ref] # [see note: pytorch repo ref]
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
- name: Setup Linux - name: Setup Linux
uses: ./.github/actions/setup-linux uses: ./.github/actions/setup-linux
@ -110,12 +110,12 @@ jobs:
- name: Calculate docker image - name: Calculate docker image
id: calculate-docker-image id: calculate-docker-image
uses: pytorch/test-infra/.github/actions/calculate-docker-image@main uses: pytorch/test-infra/.github/actions/calculate-docker-image@release/2.5
with: with:
docker-image-name: ${{ inputs.docker-image }} docker-image-name: ${{ inputs.docker-image }}
- name: Pull docker image - name: Pull docker image
uses: pytorch/test-infra/.github/actions/pull-docker-image@main uses: pytorch/test-infra/.github/actions/pull-docker-image@release/2.5
with: with:
docker-image: ${{ steps.calculate-docker-image.outputs.docker-image }} docker-image: ${{ steps.calculate-docker-image.outputs.docker-image }}
@ -222,5 +222,5 @@ jobs:
s3-prefix: pytorch/pytorch/${{ github.event.pull_request.number }}/functorchdocs s3-prefix: pytorch/pytorch/${{ github.event.pull_request.number }}/functorchdocs
- name: Teardown Linux - name: Teardown Linux
uses: pytorch/test-infra/.github/actions/teardown-linux@main uses: pytorch/test-infra/.github/actions/teardown-linux@release/2.5
if: always() if: always()

View File

@ -46,7 +46,7 @@ jobs:
keep-going: ${{ steps.filter.outputs.keep-going }} keep-going: ${{ steps.filter.outputs.keep-going }}
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
fetch-depth: 1 fetch-depth: 1
submodules: false submodules: false
@ -80,7 +80,7 @@ jobs:
steps: steps:
# [see note: pytorch repo ref] # [see note: pytorch repo ref]
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
- name: Populate CI build options - name: Populate CI build options
shell: bash shell: bash
@ -102,7 +102,7 @@ jobs:
brew install libtool brew install libtool
- name: Setup miniconda for iOS - name: Setup miniconda for iOS
uses: pytorch/test-infra/.github/actions/setup-miniconda@main uses: pytorch/test-infra/.github/actions/setup-miniconda@release/2.5
with: with:
python-version: "3.9" python-version: "3.9"
environment-file: .github/requirements/conda-env-iOS.txt environment-file: .github/requirements/conda-env-iOS.txt

View File

@ -108,7 +108,7 @@ jobs:
test-matrix: ${{ steps.filter.outputs.test-matrix }} test-matrix: ${{ steps.filter.outputs.test-matrix }}
steps: steps:
- name: Setup SSH (Click me for login details) - name: Setup SSH (Click me for login details)
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -117,7 +117,7 @@ jobs:
# checkout because when we run this action we don't *have* a local # checkout because when we run this action we don't *have* a local
# checkout. In other cases you should prefer a local checkout. # checkout. In other cases you should prefer a local checkout.
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
- name: Setup Linux - name: Setup Linux
uses: ./.github/actions/setup-linux uses: ./.github/actions/setup-linux
@ -132,7 +132,7 @@ jobs:
- name: Calculate docker image - name: Calculate docker image
id: calculate-docker-image id: calculate-docker-image
uses: pytorch/test-infra/.github/actions/calculate-docker-image@main uses: pytorch/test-infra/.github/actions/calculate-docker-image@release/2.5
with: with:
docker-image-name: ${{ inputs.docker-image-name }} docker-image-name: ${{ inputs.docker-image-name }}
@ -146,7 +146,7 @@ jobs:
echo "docker pull ghcr.io/pytorch/ci-image:${tag/:/-}" echo "docker pull ghcr.io/pytorch/ci-image:${tag/:/-}"
- name: Pull docker image - name: Pull docker image
uses: pytorch/test-infra/.github/actions/pull-docker-image@main uses: pytorch/test-infra/.github/actions/pull-docker-image@release/2.5
with: with:
docker-image: ${{ steps.calculate-docker-image.outputs.docker-image }} docker-image: ${{ steps.calculate-docker-image.outputs.docker-image }}
@ -269,5 +269,5 @@ jobs:
s3-bucket: ${{ inputs.s3-bucket }} s3-bucket: ${{ inputs.s3-bucket }}
- name: Teardown Linux - name: Teardown Linux
uses: pytorch/test-infra/.github/actions/teardown-linux@main uses: pytorch/test-infra/.github/actions/teardown-linux@release/2.5
if: always() if: always()

View File

@ -72,7 +72,7 @@ jobs:
timeout-minutes: ${{ matrix.mem_leak_check == 'mem_leak_check' && 600 || inputs.timeout-minutes }} timeout-minutes: ${{ matrix.mem_leak_check == 'mem_leak_check' && 600 || inputs.timeout-minutes }}
steps: steps:
- name: Setup SSH (Click me for login details) - name: Setup SSH (Click me for login details)
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
if: ${{ !contains(matrix.runner, 'gcp.a100') }} if: ${{ !contains(matrix.runner, 'gcp.a100') }}
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -81,7 +81,7 @@ jobs:
docker exec -it $(docker container ps --format '{{.ID}}') bash docker exec -it $(docker container ps --format '{{.ID}}') bash
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
- name: Setup Linux - name: Setup Linux
uses: ./.github/actions/setup-linux uses: ./.github/actions/setup-linux
@ -96,7 +96,7 @@ jobs:
- name: Calculate docker image - name: Calculate docker image
id: calculate-docker-image id: calculate-docker-image
uses: pytorch/test-infra/.github/actions/calculate-docker-image@main uses: pytorch/test-infra/.github/actions/calculate-docker-image@release/2.5
with: with:
docker-image-name: ${{ inputs.docker-image }} docker-image-name: ${{ inputs.docker-image }}
@ -110,7 +110,7 @@ jobs:
echo "docker pull ghcr.io/pytorch/ci-image:${tag/:/-}" echo "docker pull ghcr.io/pytorch/ci-image:${tag/:/-}"
- name: Pull docker image - name: Pull docker image
uses: pytorch/test-infra/.github/actions/pull-docker-image@main uses: pytorch/test-infra/.github/actions/pull-docker-image@release/2.5
with: with:
docker-image: ${{ steps.calculate-docker-image.outputs.docker-image }} docker-image: ${{ steps.calculate-docker-image.outputs.docker-image }}
@ -121,7 +121,7 @@ jobs:
- name: Install nvidia driver, nvidia-docker runtime, set GPU_FLAG - name: Install nvidia driver, nvidia-docker runtime, set GPU_FLAG
id: install-nvidia-driver id: install-nvidia-driver
uses: pytorch/test-infra/.github/actions/setup-nvidia@main uses: pytorch/test-infra/.github/actions/setup-nvidia@release/2.5
if: ${{ contains(inputs.build-environment, 'cuda') && !contains(matrix.config, 'nogpu') && steps.check_arc_runner.outputs.IN_ARC_RUNNER == 'false' }} if: ${{ contains(inputs.build-environment, 'cuda') && !contains(matrix.config, 'nogpu') && steps.check_arc_runner.outputs.IN_ARC_RUNNER == 'false' }}
- name: Lock NVIDIA A100 40GB Frequency - name: Lock NVIDIA A100 40GB Frequency
@ -342,7 +342,7 @@ jobs:
path: ./**/core.[1-9]* path: ./**/core.[1-9]*
- name: Teardown Linux - name: Teardown Linux
uses: pytorch/test-infra/.github/actions/teardown-linux@main uses: pytorch/test-infra/.github/actions/teardown-linux@release/2.5
if: always() if: always()
# NB: We are currently having an intermittent GPU-related issue on G5 runners with # NB: We are currently having an intermittent GPU-related issue on G5 runners with

View File

@ -71,11 +71,11 @@ jobs:
test-matrix: ${{ steps.filter.outputs.test-matrix }} test-matrix: ${{ steps.filter.outputs.test-matrix }}
steps: steps:
- name: Clean up disk space before running MacOS workflow - name: Clean up disk space before running MacOS workflow
uses: pytorch/test-infra/.github/actions/check-disk-space@main uses: pytorch/test-infra/.github/actions/check-disk-space@release/2.5
# [see note: pytorch repo ref] # [see note: pytorch repo ref]
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
- name: Set xcode version - name: Set xcode version
env: env:
@ -87,7 +87,7 @@ jobs:
- name: Setup miniconda - name: Setup miniconda
if: inputs.environment-file == '' if: inputs.environment-file == ''
uses: pytorch/test-infra/.github/actions/setup-miniconda@main uses: pytorch/test-infra/.github/actions/setup-miniconda@release/2.5
with: with:
python-version: ${{ inputs.python-version }} python-version: ${{ inputs.python-version }}
environment-file: .github/requirements/conda-env-${{ runner.os }}-${{ runner.arch }} environment-file: .github/requirements/conda-env-${{ runner.os }}-${{ runner.arch }}
@ -97,7 +97,7 @@ jobs:
# environment even though the arch is x86-64 # environment even though the arch is x86-64
- name: Setup miniconda using the provided environment file - name: Setup miniconda using the provided environment file
if: inputs.environment-file != '' if: inputs.environment-file != ''
uses: pytorch/test-infra/.github/actions/setup-miniconda@main uses: pytorch/test-infra/.github/actions/setup-miniconda@release/2.5
with: with:
python-version: ${{ inputs.python-version }} python-version: ${{ inputs.python-version }}
environment-file: ${{ inputs.environment-file }} environment-file: ${{ inputs.environment-file }}
@ -207,4 +207,4 @@ jobs:
- name: Clean up disk space - name: Clean up disk space
if: always() if: always()
continue-on-error: true continue-on-error: true
uses: pytorch/test-infra/.github/actions/check-disk-space@main uses: pytorch/test-infra/.github/actions/check-disk-space@release/2.5

View File

@ -41,7 +41,7 @@ jobs:
reenabled-issues: ${{ steps.filter.outputs.reenabled-issues }} reenabled-issues: ${{ steps.filter.outputs.reenabled-issues }}
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
submodules: false submodules: false
@ -82,7 +82,7 @@ jobs:
use-gha: true use-gha: true
- name: Setup miniconda - name: Setup miniconda
uses: pytorch/test-infra/.github/actions/setup-miniconda@main uses: pytorch/test-infra/.github/actions/setup-miniconda@release/2.5
with: with:
python-version: ${{ inputs.python-version }} python-version: ${{ inputs.python-version }}
environment-file: .github/requirements/conda-env-${{ runner.os }}-${{ runner.arch }} environment-file: .github/requirements/conda-env-${{ runner.os }}-${{ runner.arch }}
@ -161,4 +161,4 @@ jobs:
- name: Clean up disk space - name: Clean up disk space
if: always() if: always()
continue-on-error: true continue-on-error: true
uses: pytorch/test-infra/.github/actions/check-disk-space@main uses: pytorch/test-infra/.github/actions/check-disk-space@release/2.5

View File

@ -74,11 +74,11 @@ jobs:
done done
- name: Clean up disk space before running MacOS workflow - name: Clean up disk space before running MacOS workflow
uses: pytorch/test-infra/.github/actions/check-disk-space@main uses: pytorch/test-infra/.github/actions/check-disk-space@release/2.5
# [see note: pytorch repo ref] # [see note: pytorch repo ref]
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
- name: Download build artifacts - name: Download build artifacts
uses: ./.github/actions/download-build-artifacts uses: ./.github/actions/download-build-artifacts
@ -93,7 +93,7 @@ jobs:
use-gha: true use-gha: true
- name: Setup miniconda - name: Setup miniconda
uses: pytorch/test-infra/.github/actions/setup-miniconda@main uses: pytorch/test-infra/.github/actions/setup-miniconda@release/2.5
with: with:
python-version: ${{ inputs.python-version }} python-version: ${{ inputs.python-version }}
environment-file: .github/requirements/conda-env-${{ runner.os }}-${{ runner.arch }} environment-file: .github/requirements/conda-env-${{ runner.os }}-${{ runner.arch }}
@ -217,4 +217,4 @@ jobs:
- name: Clean up disk space - name: Clean up disk space
if: always() if: always()
continue-on-error: true continue-on-error: true
uses: pytorch/test-infra/.github/actions/check-disk-space@main uses: pytorch/test-infra/.github/actions/check-disk-space@release/2.5

View File

@ -58,7 +58,7 @@ jobs:
steps: steps:
# [see note: pytorch repo ref] # [see note: pytorch repo ref]
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
no-sudo: true no-sudo: true
@ -80,12 +80,12 @@ jobs:
- name: Calculate docker image - name: Calculate docker image
id: calculate-docker-image id: calculate-docker-image
uses: pytorch/test-infra/.github/actions/calculate-docker-image@main uses: pytorch/test-infra/.github/actions/calculate-docker-image@release/2.5
with: with:
docker-image-name: ${{ inputs.docker-image }} docker-image-name: ${{ inputs.docker-image }}
- name: Pull docker image - name: Pull docker image
uses: pytorch/test-infra/.github/actions/pull-docker-image@main uses: pytorch/test-infra/.github/actions/pull-docker-image@release/2.5
with: with:
docker-image: ${{ steps.calculate-docker-image.outputs.docker-image }} docker-image: ${{ steps.calculate-docker-image.outputs.docker-image }}

View File

@ -28,7 +28,7 @@ jobs:
keep-going: ${{ steps.filter.outputs.keep-going }} keep-going: ${{ steps.filter.outputs.keep-going }}
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
fetch-depth: 1 fetch-depth: 1
submodules: false submodules: false
@ -59,10 +59,10 @@ jobs:
SUPPORT_ABI: '${{ matrix.support_abi }}' SUPPORT_ABI: '${{ matrix.support_abi }}'
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
- name: Setup miniconda - name: Setup miniconda
uses: pytorch/test-infra/.github/actions/setup-miniconda@main uses: pytorch/test-infra/.github/actions/setup-miniconda@release/2.5
with: with:
python-version: 3.8 python-version: 3.8
environment-file: .github/requirements/conda-env-${{ runner.os }}-${{ runner.arch }}.txt environment-file: .github/requirements/conda-env-${{ runner.os }}-${{ runner.arch }}.txt

View File

@ -45,7 +45,7 @@ jobs:
ISSUE_OWNER: ${{ inputs.issue_owner }} ISSUE_OWNER: ${{ inputs.issue_owner }}
steps: steps:
# - name: Checkout PyTorch # - name: Checkout PyTorch
# uses: pytorch/pytorch/.github/actions/checkout-pytorch@main # uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
# with: # with:
# fetch-depth: 1 # fetch-depth: 1
# submodules: true # submodules: true

View File

@ -79,10 +79,10 @@ jobs:
git config --global core.fsmonitor false git config --global core.fsmonitor false
- name: Clean up leftover processes on non-ephemeral Windows runner - name: Clean up leftover processes on non-ephemeral Windows runner
uses: pytorch/test-infra/.github/actions/cleanup-runner@main uses: pytorch/test-infra/.github/actions/cleanup-runner@release/2.5
- name: Setup SSH (Click me for login details) - name: Setup SSH (Click me for login details)
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
instructions: | instructions: |
@ -97,7 +97,7 @@ jobs:
# [see note: pytorch repo ref] # [see note: pytorch repo ref]
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
no-sudo: true no-sudo: true

View File

@ -57,10 +57,10 @@ jobs:
git config --global core.fsmonitor false git config --global core.fsmonitor false
- name: Clean up leftover processes on non-ephemeral Windows runner - name: Clean up leftover processes on non-ephemeral Windows runner
uses: pytorch/test-infra/.github/actions/cleanup-runner@main uses: pytorch/test-infra/.github/actions/cleanup-runner@release/2.5
- name: Setup SSH (Click me for login details) - name: Setup SSH (Click me for login details)
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
instructions: | instructions: |
@ -76,7 +76,7 @@ jobs:
# [see note: pytorch repo ref] # [see note: pytorch repo ref]
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
no-sudo: true no-sudo: true

View File

@ -54,7 +54,7 @@ jobs:
steps: steps:
# [see note: pytorch repo ref] # [see note: pytorch repo ref]
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
- name: Setup XPU - name: Setup XPU
uses: ./.github/actions/setup-xpu uses: ./.github/actions/setup-xpu
@ -72,12 +72,12 @@ jobs:
- name: Calculate docker image - name: Calculate docker image
id: calculate-docker-image id: calculate-docker-image
uses: pytorch/test-infra/.github/actions/calculate-docker-image@main uses: pytorch/test-infra/.github/actions/calculate-docker-image@release/2.5
with: with:
docker-image-name: ${{ inputs.docker-image }} docker-image-name: ${{ inputs.docker-image }}
- name: Pull docker image - name: Pull docker image
uses: pytorch/test-infra/.github/actions/pull-docker-image@main uses: pytorch/test-infra/.github/actions/pull-docker-image@release/2.5
with: with:
docker-image: ${{ steps.calculate-docker-image.outputs.docker-image }} docker-image: ${{ steps.calculate-docker-image.outputs.docker-image }}

View File

@ -40,12 +40,12 @@ jobs:
CUDA_VERSION: ${{ matrix.cuda_version }} CUDA_VERSION: ${{ matrix.cuda_version }}
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
submodules: false submodules: false
- name: Calculate docker image - name: Calculate docker image
if: env.WITH_PUSH == 'false' if: env.WITH_PUSH == 'false'
uses: pytorch/test-infra/.github/actions/calculate-docker-image@main uses: pytorch/test-infra/.github/actions/calculate-docker-image@release/2.5
with: with:
docker-image-name: conda-builder${{ matrix.cuda_version == 'cpu' && '-' || '-cuda' }}${{matrix.cuda_version}} docker-image-name: conda-builder${{ matrix.cuda_version == 'cpu' && '-' || '-cuda' }}${{matrix.cuda_version}}
docker-build-dir: .ci/docker/conda docker-build-dir: .ci/docker/conda

View File

@ -40,12 +40,12 @@ jobs:
GPU_ARCH_VERSION: ${{ matrix.cuda_version }} GPU_ARCH_VERSION: ${{ matrix.cuda_version }}
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
submodules: false submodules: false
- name: Calculate docker image - name: Calculate docker image
if: env.WITH_PUSH == 'false' if: env.WITH_PUSH == 'false'
uses: pytorch/test-infra/.github/actions/calculate-docker-image@main uses: pytorch/test-infra/.github/actions/calculate-docker-image@release/2.5
with: with:
docker-image-name: libtorch-cxx11-builder-cuda${{matrix.cuda_version}} docker-image-name: libtorch-cxx11-builder-cuda${{matrix.cuda_version}}
docker-build-dir: .ci/docker/libtorch docker-build-dir: .ci/docker/libtorch
@ -75,12 +75,12 @@ jobs:
GPU_ARCH_VERSION: ${{ matrix.rocm_version }} GPU_ARCH_VERSION: ${{ matrix.rocm_version }}
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
submodules: false submodules: false
- name: Calculate docker image - name: Calculate docker image
if: env.WITH_PUSH == 'false' if: env.WITH_PUSH == 'false'
uses: pytorch/test-infra/.github/actions/calculate-docker-image@main uses: pytorch/test-infra/.github/actions/calculate-docker-image@release/2.5
with: with:
docker-image-name: libtorch-cxx11-builder-rocm${{matrix.rocm_version}} docker-image-name: libtorch-cxx11-builder-rocm${{matrix.rocm_version}}
docker-build-dir: .ci/docker/libtorch docker-build-dir: .ci/docker/libtorch
@ -104,12 +104,12 @@ jobs:
runs-on: linux.9xlarge.ephemeral runs-on: linux.9xlarge.ephemeral
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
submodules: false submodules: false
- name: Calculate docker image - name: Calculate docker image
if: env.WITH_PUSH == 'false' if: env.WITH_PUSH == 'false'
uses: pytorch/test-infra/.github/actions/calculate-docker-image@main uses: pytorch/test-infra/.github/actions/calculate-docker-image@release/2.5
with: with:
docker-image-name: libtorch-cxx11-builder-cpu docker-image-name: libtorch-cxx11-builder-cpu
docker-build-dir: .ci/docker/libtorch docker-build-dir: .ci/docker/libtorch

View File

@ -46,12 +46,12 @@ jobs:
- name: Purge tools folder (free space for build) - name: Purge tools folder (free space for build)
run: rm -rf /opt/hostedtoolcache run: rm -rf /opt/hostedtoolcache
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
submodules: false submodules: false
- name: Calculate docker image - name: Calculate docker image
if: env.WITH_PUSH == 'false' if: env.WITH_PUSH == 'false'
uses: pytorch/test-infra/.github/actions/calculate-docker-image@main uses: pytorch/test-infra/.github/actions/calculate-docker-image@release/2.5
with: with:
docker-image-name: manylinux-builder-cuda${{matrix.cuda_version}} docker-image-name: manylinux-builder-cuda${{matrix.cuda_version}}
docker-build-dir: .ci/docker/manywheel docker-build-dir: .ci/docker/manywheel
@ -84,12 +84,12 @@ jobs:
- name: Purge tools folder (free space for build) - name: Purge tools folder (free space for build)
run: rm -rf /opt/hostedtoolcache run: rm -rf /opt/hostedtoolcache
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
submodules: false submodules: false
- name: Calculate docker image - name: Calculate docker image
if: env.WITH_PUSH == 'false' if: env.WITH_PUSH == 'false'
uses: pytorch/test-infra/.github/actions/calculate-docker-image@main uses: pytorch/test-infra/.github/actions/calculate-docker-image@release/2.5
with: with:
docker-image-name: manylinux2_28-builder-cuda${{matrix.cuda_version}} docker-image-name: manylinux2_28-builder-cuda${{matrix.cuda_version}}
docker-build-dir: .ci/docker/manywheel docker-build-dir: .ci/docker/manywheel
@ -122,7 +122,7 @@ jobs:
uses: actions/checkout@v3 uses: actions/checkout@v3
- name: Calculate docker image - name: Calculate docker image
if: env.WITH_PUSH == 'false' if: env.WITH_PUSH == 'false'
uses: pytorch/test-infra/.github/actions/calculate-docker-image@main uses: pytorch/test-infra/.github/actions/calculate-docker-image@release/2.5
with: with:
docker-image-name: manylinuxaarch64-builder-cuda${{matrix.cuda_version}} docker-image-name: manylinuxaarch64-builder-cuda${{matrix.cuda_version}}
docker-build-dir: .ci/docker/manywheel docker-build-dir: .ci/docker/manywheel
@ -152,12 +152,12 @@ jobs:
GPU_ARCH_VERSION: ${{ matrix.rocm_version }} GPU_ARCH_VERSION: ${{ matrix.rocm_version }}
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
submodules: false submodules: false
- name: Calculate docker image - name: Calculate docker image
if: env.WITH_PUSH == 'false' if: env.WITH_PUSH == 'false'
uses: pytorch/test-infra/.github/actions/calculate-docker-image@main uses: pytorch/test-infra/.github/actions/calculate-docker-image@release/2.5
with: with:
docker-image-name: manylinux-builder-rocm${{matrix.rocm_version}} docker-image-name: manylinux-builder-rocm${{matrix.rocm_version}}
docker-build-dir: .ci/docker/manywheel docker-build-dir: .ci/docker/manywheel
@ -181,12 +181,12 @@ jobs:
runs-on: am2.linux.9xlarge.ephemeral runs-on: am2.linux.9xlarge.ephemeral
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
submodules: false submodules: false
- name: Calculate docker image - name: Calculate docker image
if: env.WITH_PUSH == 'false' if: env.WITH_PUSH == 'false'
uses: pytorch/test-infra/.github/actions/calculate-docker-image@main uses: pytorch/test-infra/.github/actions/calculate-docker-image@release/2.5
with: with:
docker-image-name: manylinux-builder-cpu docker-image-name: manylinux-builder-cpu
docker-build-dir: .ci/docker/manywheel docker-build-dir: .ci/docker/manywheel
@ -212,12 +212,12 @@ jobs:
GPU_ARCH_TYPE: cpu-manylinux_2_28 GPU_ARCH_TYPE: cpu-manylinux_2_28
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
submodules: false submodules: false
- name: Calculate docker image - name: Calculate docker image
if: env.WITH_PUSH == 'false' if: env.WITH_PUSH == 'false'
uses: pytorch/test-infra/.github/actions/calculate-docker-image@main uses: pytorch/test-infra/.github/actions/calculate-docker-image@release/2.5
with: with:
docker-image-name: manylinux2_28-builder-cpu docker-image-name: manylinux2_28-builder-cpu
docker-build-dir: .ci/docker/manywheel docker-build-dir: .ci/docker/manywheel
@ -243,12 +243,12 @@ jobs:
GPU_ARCH_TYPE: cpu-aarch64 GPU_ARCH_TYPE: cpu-aarch64
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
submodules: false submodules: false
- name: Calculate docker image - name: Calculate docker image
if: env.WITH_PUSH == 'false' if: env.WITH_PUSH == 'false'
uses: pytorch/test-infra/.github/actions/calculate-docker-image@main uses: pytorch/test-infra/.github/actions/calculate-docker-image@release/2.5
with: with:
docker-image-name: manylinuxaarch64-builder-cpu-aarch64 docker-image-name: manylinuxaarch64-builder-cpu-aarch64
docker-build-dir: .ci/docker/manywheel docker-build-dir: .ci/docker/manywheel
@ -274,12 +274,12 @@ jobs:
GPU_ARCH_TYPE: cpu-aarch64-2_28 GPU_ARCH_TYPE: cpu-aarch64-2_28
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
submodules: false submodules: false
- name: Calculate docker image - name: Calculate docker image
if: env.WITH_PUSH == 'false' if: env.WITH_PUSH == 'false'
uses: pytorch/test-infra/.github/actions/calculate-docker-image@main uses: pytorch/test-infra/.github/actions/calculate-docker-image@release/2.5
with: with:
docker-image-name: manylinux2_28_aarch64-builder-cpu-aarch64 docker-image-name: manylinux2_28_aarch64-builder-cpu-aarch64
docker-build-dir: .ci/docker/manywheel docker-build-dir: .ci/docker/manywheel
@ -308,12 +308,12 @@ jobs:
GPU_ARCH_TYPE: cpu-cxx11-abi GPU_ARCH_TYPE: cpu-cxx11-abi
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
submodules: false submodules: false
- name: Calculate docker image - name: Calculate docker image
if: env.WITH_PUSH == 'false' if: env.WITH_PUSH == 'false'
uses: pytorch/test-infra/.github/actions/calculate-docker-image@main uses: pytorch/test-infra/.github/actions/calculate-docker-image@release/2.5
with: with:
docker-image-name: manylinuxcxx11-abi-builder-cpu-cxx11-abi docker-image-name: manylinuxcxx11-abi-builder-cpu-cxx11-abi
docker-build-dir: .ci/docker/manywheel docker-build-dir: .ci/docker/manywheel
@ -339,12 +339,12 @@ jobs:
GPU_ARCH_TYPE: xpu GPU_ARCH_TYPE: xpu
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
submodules: false submodules: false
- name: Calculate docker image - name: Calculate docker image
if: env.WITH_PUSH == 'false' if: env.WITH_PUSH == 'false'
uses: pytorch/test-infra/.github/actions/calculate-docker-image@main uses: pytorch/test-infra/.github/actions/calculate-docker-image@release/2.5
with: with:
docker-image-name: manylinux2_28-builder-xpu docker-image-name: manylinux2_28-builder-xpu
docker-build-dir: .ci/docker/manywheel docker-build-dir: .ci/docker/manywheel

View File

@ -47,12 +47,12 @@ jobs:
BUILD_DEVICE: ${{ matrix.device }} BUILD_DEVICE: ${{ matrix.device }}
steps: steps:
- name: Setup SSH (Click me for login details) - name: Setup SSH (Click me for login details)
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
submodules: false submodules: false
@ -60,7 +60,7 @@ jobs:
uses: ./.github/actions/setup-linux uses: ./.github/actions/setup-linux
- name: Pull Docker image - name: Pull Docker image
uses: pytorch/test-infra/.github/actions/pull-docker-image@main uses: pytorch/test-infra/.github/actions/pull-docker-image@release/2.5
with: with:
docker-image: ${{ env.DOCKER_IMAGE }} docker-image: ${{ env.DOCKER_IMAGE }}
@ -125,7 +125,7 @@ jobs:
path: ${{ runner.temp }}/artifacts/* path: ${{ runner.temp }}/artifacts/*
- name: Teardown Linux - name: Teardown Linux
uses: pytorch/test-infra/.github/actions/teardown-linux@main uses: pytorch/test-infra/.github/actions/teardown-linux@release/2.5
if: always() if: always()
upload-wheel: upload-wheel:
@ -210,12 +210,12 @@ jobs:
PY_VERS: ${{ matrix.py_vers }} PY_VERS: ${{ matrix.py_vers }}
steps: steps:
- name: Setup SSH (Click me for login details) - name: Setup SSH (Click me for login details)
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
submodules: false submodules: false
@ -223,7 +223,7 @@ jobs:
uses: ./.github/actions/setup-linux uses: ./.github/actions/setup-linux
- name: Pull Docker image - name: Pull Docker image
uses: pytorch/test-infra/.github/actions/pull-docker-image@main uses: pytorch/test-infra/.github/actions/pull-docker-image@release/2.5
with: with:
docker-image: ${{ env.DOCKER_IMAGE }} docker-image: ${{ env.DOCKER_IMAGE }}
@ -258,7 +258,7 @@ jobs:
path: ${{ runner.temp }}/artifacts/* path: ${{ runner.temp }}/artifacts/*
- name: Teardown Linux - name: Teardown Linux
uses: pytorch/test-infra/.github/actions/teardown-linux@main uses: pytorch/test-infra/.github/actions/teardown-linux@release/2.5
if: always() if: always()
upload-conda: upload-conda:

View File

@ -35,7 +35,7 @@ jobs:
runs-on: linux.20_04.4x runs-on: linux.20_04.4x
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
submodules: false submodules: false
fetch-depth: 1 fetch-depth: 1

View File

@ -11,7 +11,7 @@ jobs:
runs-on: ubuntu-latest runs-on: ubuntu-latest
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
- name: Run close_nonexistent_disable_issues.py - name: Run close_nonexistent_disable_issues.py
env: env:

View File

@ -63,7 +63,7 @@ jobs:
files: ${{env.PT_RELEASE_FILE}} files: ${{env.PT_RELEASE_FILE}}
- name: Upload source distribution to GHA artifacts for release tags - name: Upload source distribution to GHA artifacts for release tags
if: ${{ github.event_name == 'push' && startsWith(github.ref, 'refs/tags/v') && contains(github.ref, 'rc') }} if: ${{ github.event_name == 'push' && startsWith(github.ref, 'refs/tags/v') && contains(github.ref, 'rc') }}
uses: actions/upload-artifact@v2 uses: actions/upload-artifact@v4.4.0
with: with:
name: ${{ env.PT_RELEASE_FILE }} name: ${{ env.PT_RELEASE_FILE }}
path: ${{ env.PT_RELEASE_FILE }} path: ${{ env.PT_RELEASE_FILE }}

View File

@ -82,28 +82,28 @@ jobs:
# [see note: pytorch repo ref] # [see note: pytorch repo ref]
# deep clone (fetch-depth 0) required for git merge-base # deep clone (fetch-depth 0) required for git merge-base
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
- name: Setup Linux - name: Setup Linux
uses: ./.github/actions/setup-linux uses: ./.github/actions/setup-linux
- name: Build docker image - name: Build docker image
id: build-docker-image id: build-docker-image
uses: pytorch/test-infra/.github/actions/calculate-docker-image@main uses: pytorch/test-infra/.github/actions/calculate-docker-image@release/2.5
with: with:
docker-image-name: ${{ matrix.docker-image-name }} docker-image-name: ${{ matrix.docker-image-name }}
always-rebuild: true always-rebuild: true
push: true push: true
- name: Pull docker image - name: Pull docker image
uses: pytorch/test-infra/.github/actions/pull-docker-image@main uses: pytorch/test-infra/.github/actions/pull-docker-image@release/2.5
with: with:
docker-image: ${{ steps.build-docker-image.outputs.docker-image }} docker-image: ${{ steps.build-docker-image.outputs.docker-image }}
- uses: nick-fields/retry@v3.0.0 - uses: nick-fields/retry@v3.0.0
name: Push to https://https://ghcr.io/ name: Push to https://https://ghcr.io/
id: push-to-ghcr-io id: push-to-ghcr-io
if: ${{ github.event_name == 'push' }} if: ${{ 0 && github.event_name == 'push' }}
env: env:
ECR_DOCKER_IMAGE: ${{ steps.build-docker-image.outputs.docker-image }} ECR_DOCKER_IMAGE: ${{ steps.build-docker-image.outputs.docker-image }}
GHCR_PAT: ${{ secrets.GHCR_PAT }} GHCR_PAT: ${{ secrets.GHCR_PAT }}
@ -128,5 +128,5 @@ jobs:
if: always() if: always()
- name: Teardown Linux - name: Teardown Linux
uses: pytorch/test-infra/.github/actions/teardown-linux@main uses: pytorch/test-infra/.github/actions/teardown-linux@release/2.5
if: always() if: always()

View File

@ -41,7 +41,7 @@ jobs:
matrix: ${{ steps.generate-matrix.outputs.matrix }} matrix: ${{ steps.generate-matrix.outputs.matrix }}
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
fetch-depth: 1 fetch-depth: 1
submodules: true submodules: true
@ -69,7 +69,7 @@ jobs:
CUDNN_VERSION: ${{ matrix.cudnn_version }} CUDNN_VERSION: ${{ matrix.cudnn_version }}
steps: steps:
- name: Setup SSH (Click me for login details) - name: Setup SSH (Click me for login details)
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
# [see note: pytorch repo ref] # [see note: pytorch repo ref]
@ -147,12 +147,12 @@ jobs:
fi fi
- name: Teardown Linux - name: Teardown Linux
uses: pytorch/test-infra/.github/actions/teardown-linux@main uses: pytorch/test-infra/.github/actions/teardown-linux@release/2.5
if: always() if: always()
validate: validate:
needs: build needs: build
uses: pytorch/builder/.github/workflows/validate-docker-images.yml@main uses: pytorch/builder/.github/workflows/validate-docker-images.yml@release/2.5
with: with:
channel: nightly channel: nightly
ref: main ref: main

View File

@ -57,7 +57,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu-aarch64 GPU_ARCH_TYPE: cpu-aarch64
DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cpu-aarch64-main DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cpu-aarch64-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
runs_on: linux.arm64.m7g.4xlarge.ephemeral runs_on: linux.arm64.m7g.4xlarge.ephemeral
@ -81,7 +81,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu-aarch64 GPU_ARCH_TYPE: cpu-aarch64
DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cpu-aarch64-main DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cpu-aarch64-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: manywheel-py3_9-cpu-aarch64 build_name: manywheel-py3_9-cpu-aarch64
@ -104,7 +104,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu-aarch64 GPU_ARCH_TYPE: cpu-aarch64
DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cpu-aarch64-main DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cpu-aarch64-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: manywheel-py3_9-cpu-aarch64 build_name: manywheel-py3_9-cpu-aarch64
@ -126,7 +126,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_TYPE: cuda-aarch64 GPU_ARCH_TYPE: cuda-aarch64
DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cuda12.4-2.5
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
@ -151,7 +151,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_TYPE: cuda-aarch64 GPU_ARCH_TYPE: cuda-aarch64
DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cuda12.4-2.5
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
@ -174,7 +174,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu-aarch64 GPU_ARCH_TYPE: cpu-aarch64
DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cpu-aarch64-main DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cpu-aarch64-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
runs_on: linux.arm64.m7g.4xlarge.ephemeral runs_on: linux.arm64.m7g.4xlarge.ephemeral
@ -198,7 +198,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu-aarch64 GPU_ARCH_TYPE: cpu-aarch64
DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cpu-aarch64-main DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cpu-aarch64-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
build_name: manywheel-py3_10-cpu-aarch64 build_name: manywheel-py3_10-cpu-aarch64
@ -221,7 +221,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu-aarch64 GPU_ARCH_TYPE: cpu-aarch64
DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cpu-aarch64-main DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cpu-aarch64-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
build_name: manywheel-py3_10-cpu-aarch64 build_name: manywheel-py3_10-cpu-aarch64
@ -243,7 +243,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_TYPE: cuda-aarch64 GPU_ARCH_TYPE: cuda-aarch64
DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cuda12.4-2.5
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
@ -268,7 +268,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_TYPE: cuda-aarch64 GPU_ARCH_TYPE: cuda-aarch64
DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cuda12.4-2.5
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
@ -291,7 +291,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu-aarch64 GPU_ARCH_TYPE: cpu-aarch64
DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cpu-aarch64-main DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cpu-aarch64-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
runs_on: linux.arm64.m7g.4xlarge.ephemeral runs_on: linux.arm64.m7g.4xlarge.ephemeral
@ -315,7 +315,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu-aarch64 GPU_ARCH_TYPE: cpu-aarch64
DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cpu-aarch64-main DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cpu-aarch64-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
build_name: manywheel-py3_11-cpu-aarch64 build_name: manywheel-py3_11-cpu-aarch64
@ -338,7 +338,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu-aarch64 GPU_ARCH_TYPE: cpu-aarch64
DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cpu-aarch64-main DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cpu-aarch64-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
build_name: manywheel-py3_11-cpu-aarch64 build_name: manywheel-py3_11-cpu-aarch64
@ -360,7 +360,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_TYPE: cuda-aarch64 GPU_ARCH_TYPE: cuda-aarch64
DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cuda12.4-2.5
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
@ -385,7 +385,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_TYPE: cuda-aarch64 GPU_ARCH_TYPE: cuda-aarch64
DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cuda12.4-2.5
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
@ -408,7 +408,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu-aarch64 GPU_ARCH_TYPE: cpu-aarch64
DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cpu-aarch64-main DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cpu-aarch64-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
runs_on: linux.arm64.m7g.4xlarge.ephemeral runs_on: linux.arm64.m7g.4xlarge.ephemeral
@ -432,7 +432,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu-aarch64 GPU_ARCH_TYPE: cpu-aarch64
DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cpu-aarch64-main DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cpu-aarch64-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
build_name: manywheel-py3_12-cpu-aarch64 build_name: manywheel-py3_12-cpu-aarch64
@ -455,7 +455,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu-aarch64 GPU_ARCH_TYPE: cpu-aarch64
DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cpu-aarch64-main DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cpu-aarch64-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
build_name: manywheel-py3_12-cpu-aarch64 build_name: manywheel-py3_12-cpu-aarch64
@ -477,7 +477,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_TYPE: cuda-aarch64 GPU_ARCH_TYPE: cuda-aarch64
DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cuda12.4-2.5
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
@ -502,7 +502,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_TYPE: cuda-aarch64 GPU_ARCH_TYPE: cuda-aarch64
DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinuxaarch64-builder:cuda12.4-2.5
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"

View File

@ -57,7 +57,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/conda-builder:cpu-main DOCKER_IMAGE: pytorch/conda-builder:cpu-2.5
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
build_name: conda-py3_9-cpu build_name: conda-py3_9-cpu
@ -78,7 +78,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/conda-builder:cpu-main DOCKER_IMAGE: pytorch/conda-builder:cpu-2.5
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: conda-py3_9-cpu build_name: conda-py3_9-cpu
build_environment: linux-binary-conda build_environment: linux-binary-conda
@ -100,7 +100,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/conda-builder:cpu-main DOCKER_IMAGE: pytorch/conda-builder:cpu-2.5
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: conda-py3_9-cpu build_name: conda-py3_9-cpu
secrets: secrets:
@ -122,7 +122,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda11.8-main DOCKER_IMAGE: pytorch/conda-builder:cuda11.8-2.5
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
runs_on: linux.24xlarge.ephemeral runs_on: linux.24xlarge.ephemeral
@ -145,7 +145,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda11.8-main DOCKER_IMAGE: pytorch/conda-builder:cuda11.8-2.5
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: conda-py3_9-cuda11_8 build_name: conda-py3_9-cuda11_8
build_environment: linux-binary-conda build_environment: linux-binary-conda
@ -168,7 +168,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda11.8-main DOCKER_IMAGE: pytorch/conda-builder:cuda11.8-2.5
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: conda-py3_9-cuda11_8 build_name: conda-py3_9-cuda11_8
secrets: secrets:
@ -190,7 +190,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda12.1-main DOCKER_IMAGE: pytorch/conda-builder:cuda12.1-2.5
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
runs_on: linux.24xlarge.ephemeral runs_on: linux.24xlarge.ephemeral
@ -213,7 +213,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda12.1-main DOCKER_IMAGE: pytorch/conda-builder:cuda12.1-2.5
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: conda-py3_9-cuda12_1 build_name: conda-py3_9-cuda12_1
build_environment: linux-binary-conda build_environment: linux-binary-conda
@ -236,7 +236,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda12.1-main DOCKER_IMAGE: pytorch/conda-builder:cuda12.1-2.5
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: conda-py3_9-cuda12_1 build_name: conda-py3_9-cuda12_1
secrets: secrets:
@ -258,7 +258,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda12.4-main DOCKER_IMAGE: pytorch/conda-builder:cuda12.4-2.5
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
runs_on: linux.24xlarge.ephemeral runs_on: linux.24xlarge.ephemeral
@ -281,7 +281,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda12.4-main DOCKER_IMAGE: pytorch/conda-builder:cuda12.4-2.5
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: conda-py3_9-cuda12_4 build_name: conda-py3_9-cuda12_4
build_environment: linux-binary-conda build_environment: linux-binary-conda
@ -304,7 +304,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda12.4-main DOCKER_IMAGE: pytorch/conda-builder:cuda12.4-2.5
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: conda-py3_9-cuda12_4 build_name: conda-py3_9-cuda12_4
secrets: secrets:
@ -325,7 +325,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/conda-builder:cpu-main DOCKER_IMAGE: pytorch/conda-builder:cpu-2.5
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
build_name: conda-py3_10-cpu build_name: conda-py3_10-cpu
@ -346,7 +346,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/conda-builder:cpu-main DOCKER_IMAGE: pytorch/conda-builder:cpu-2.5
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
build_name: conda-py3_10-cpu build_name: conda-py3_10-cpu
build_environment: linux-binary-conda build_environment: linux-binary-conda
@ -368,7 +368,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/conda-builder:cpu-main DOCKER_IMAGE: pytorch/conda-builder:cpu-2.5
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
build_name: conda-py3_10-cpu build_name: conda-py3_10-cpu
secrets: secrets:
@ -390,7 +390,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda11.8-main DOCKER_IMAGE: pytorch/conda-builder:cuda11.8-2.5
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
runs_on: linux.24xlarge.ephemeral runs_on: linux.24xlarge.ephemeral
@ -413,7 +413,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda11.8-main DOCKER_IMAGE: pytorch/conda-builder:cuda11.8-2.5
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
build_name: conda-py3_10-cuda11_8 build_name: conda-py3_10-cuda11_8
build_environment: linux-binary-conda build_environment: linux-binary-conda
@ -436,7 +436,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda11.8-main DOCKER_IMAGE: pytorch/conda-builder:cuda11.8-2.5
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
build_name: conda-py3_10-cuda11_8 build_name: conda-py3_10-cuda11_8
secrets: secrets:
@ -458,7 +458,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda12.1-main DOCKER_IMAGE: pytorch/conda-builder:cuda12.1-2.5
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
runs_on: linux.24xlarge.ephemeral runs_on: linux.24xlarge.ephemeral
@ -481,7 +481,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda12.1-main DOCKER_IMAGE: pytorch/conda-builder:cuda12.1-2.5
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
build_name: conda-py3_10-cuda12_1 build_name: conda-py3_10-cuda12_1
build_environment: linux-binary-conda build_environment: linux-binary-conda
@ -504,7 +504,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda12.1-main DOCKER_IMAGE: pytorch/conda-builder:cuda12.1-2.5
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
build_name: conda-py3_10-cuda12_1 build_name: conda-py3_10-cuda12_1
secrets: secrets:
@ -526,7 +526,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda12.4-main DOCKER_IMAGE: pytorch/conda-builder:cuda12.4-2.5
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
runs_on: linux.24xlarge.ephemeral runs_on: linux.24xlarge.ephemeral
@ -549,7 +549,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda12.4-main DOCKER_IMAGE: pytorch/conda-builder:cuda12.4-2.5
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
build_name: conda-py3_10-cuda12_4 build_name: conda-py3_10-cuda12_4
build_environment: linux-binary-conda build_environment: linux-binary-conda
@ -572,7 +572,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda12.4-main DOCKER_IMAGE: pytorch/conda-builder:cuda12.4-2.5
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
build_name: conda-py3_10-cuda12_4 build_name: conda-py3_10-cuda12_4
secrets: secrets:
@ -593,7 +593,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/conda-builder:cpu-main DOCKER_IMAGE: pytorch/conda-builder:cpu-2.5
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
build_name: conda-py3_11-cpu build_name: conda-py3_11-cpu
@ -614,7 +614,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/conda-builder:cpu-main DOCKER_IMAGE: pytorch/conda-builder:cpu-2.5
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
build_name: conda-py3_11-cpu build_name: conda-py3_11-cpu
build_environment: linux-binary-conda build_environment: linux-binary-conda
@ -636,7 +636,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/conda-builder:cpu-main DOCKER_IMAGE: pytorch/conda-builder:cpu-2.5
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
build_name: conda-py3_11-cpu build_name: conda-py3_11-cpu
secrets: secrets:
@ -658,7 +658,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda11.8-main DOCKER_IMAGE: pytorch/conda-builder:cuda11.8-2.5
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
runs_on: linux.24xlarge.ephemeral runs_on: linux.24xlarge.ephemeral
@ -681,7 +681,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda11.8-main DOCKER_IMAGE: pytorch/conda-builder:cuda11.8-2.5
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
build_name: conda-py3_11-cuda11_8 build_name: conda-py3_11-cuda11_8
build_environment: linux-binary-conda build_environment: linux-binary-conda
@ -704,7 +704,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda11.8-main DOCKER_IMAGE: pytorch/conda-builder:cuda11.8-2.5
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
build_name: conda-py3_11-cuda11_8 build_name: conda-py3_11-cuda11_8
secrets: secrets:
@ -726,7 +726,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda12.1-main DOCKER_IMAGE: pytorch/conda-builder:cuda12.1-2.5
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
runs_on: linux.24xlarge.ephemeral runs_on: linux.24xlarge.ephemeral
@ -749,7 +749,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda12.1-main DOCKER_IMAGE: pytorch/conda-builder:cuda12.1-2.5
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
build_name: conda-py3_11-cuda12_1 build_name: conda-py3_11-cuda12_1
build_environment: linux-binary-conda build_environment: linux-binary-conda
@ -772,7 +772,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda12.1-main DOCKER_IMAGE: pytorch/conda-builder:cuda12.1-2.5
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
build_name: conda-py3_11-cuda12_1 build_name: conda-py3_11-cuda12_1
secrets: secrets:
@ -794,7 +794,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda12.4-main DOCKER_IMAGE: pytorch/conda-builder:cuda12.4-2.5
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
runs_on: linux.24xlarge.ephemeral runs_on: linux.24xlarge.ephemeral
@ -817,7 +817,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda12.4-main DOCKER_IMAGE: pytorch/conda-builder:cuda12.4-2.5
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
build_name: conda-py3_11-cuda12_4 build_name: conda-py3_11-cuda12_4
build_environment: linux-binary-conda build_environment: linux-binary-conda
@ -840,7 +840,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda12.4-main DOCKER_IMAGE: pytorch/conda-builder:cuda12.4-2.5
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
build_name: conda-py3_11-cuda12_4 build_name: conda-py3_11-cuda12_4
secrets: secrets:
@ -861,7 +861,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/conda-builder:cpu-main DOCKER_IMAGE: pytorch/conda-builder:cpu-2.5
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
build_name: conda-py3_12-cpu build_name: conda-py3_12-cpu
@ -882,7 +882,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/conda-builder:cpu-main DOCKER_IMAGE: pytorch/conda-builder:cpu-2.5
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
build_name: conda-py3_12-cpu build_name: conda-py3_12-cpu
build_environment: linux-binary-conda build_environment: linux-binary-conda
@ -904,7 +904,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/conda-builder:cpu-main DOCKER_IMAGE: pytorch/conda-builder:cpu-2.5
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
build_name: conda-py3_12-cpu build_name: conda-py3_12-cpu
secrets: secrets:
@ -926,7 +926,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda11.8-main DOCKER_IMAGE: pytorch/conda-builder:cuda11.8-2.5
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
runs_on: linux.24xlarge.ephemeral runs_on: linux.24xlarge.ephemeral
@ -949,7 +949,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda11.8-main DOCKER_IMAGE: pytorch/conda-builder:cuda11.8-2.5
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
build_name: conda-py3_12-cuda11_8 build_name: conda-py3_12-cuda11_8
build_environment: linux-binary-conda build_environment: linux-binary-conda
@ -972,7 +972,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda11.8-main DOCKER_IMAGE: pytorch/conda-builder:cuda11.8-2.5
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
build_name: conda-py3_12-cuda11_8 build_name: conda-py3_12-cuda11_8
secrets: secrets:
@ -994,7 +994,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda12.1-main DOCKER_IMAGE: pytorch/conda-builder:cuda12.1-2.5
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
runs_on: linux.24xlarge.ephemeral runs_on: linux.24xlarge.ephemeral
@ -1017,7 +1017,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda12.1-main DOCKER_IMAGE: pytorch/conda-builder:cuda12.1-2.5
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
build_name: conda-py3_12-cuda12_1 build_name: conda-py3_12-cuda12_1
build_environment: linux-binary-conda build_environment: linux-binary-conda
@ -1040,7 +1040,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda12.1-main DOCKER_IMAGE: pytorch/conda-builder:cuda12.1-2.5
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
build_name: conda-py3_12-cuda12_1 build_name: conda-py3_12-cuda12_1
secrets: secrets:
@ -1062,7 +1062,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda12.4-main DOCKER_IMAGE: pytorch/conda-builder:cuda12.4-2.5
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
runs_on: linux.24xlarge.ephemeral runs_on: linux.24xlarge.ephemeral
@ -1085,7 +1085,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda12.4-main DOCKER_IMAGE: pytorch/conda-builder:cuda12.4-2.5
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
build_name: conda-py3_12-cuda12_4 build_name: conda-py3_12-cuda12_4
build_environment: linux-binary-conda build_environment: linux-binary-conda
@ -1108,7 +1108,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/conda-builder:cuda12.4-main DOCKER_IMAGE: pytorch/conda-builder:cuda12.4-2.5
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
build_name: conda-py3_12-cuda12_4 build_name: conda-py3_12-cuda12_4
secrets: secrets:

View File

@ -52,7 +52,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cpu-main DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cpu-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -74,7 +74,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cpu-main DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cpu-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
build_name: libtorch-cpu-shared-with-deps-cxx11-abi build_name: libtorch-cpu-shared-with-deps-cxx11-abi

View File

@ -57,7 +57,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cpu-main DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cpu-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -79,7 +79,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cpu-main DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cpu-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
build_name: libtorch-cpu-shared-with-deps-cxx11-abi build_name: libtorch-cpu-shared-with-deps-cxx11-abi
@ -102,7 +102,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cpu-main DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cpu-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
build_name: libtorch-cpu-shared-with-deps-cxx11-abi build_name: libtorch-cpu-shared-with-deps-cxx11-abi
@ -125,7 +125,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cuda11.8-main DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cuda11.8-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -148,7 +148,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cuda11.8-main DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cuda11.8-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
build_name: libtorch-cuda11_8-shared-with-deps-cxx11-abi build_name: libtorch-cuda11_8-shared-with-deps-cxx11-abi
@ -172,7 +172,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cuda11.8-main DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cuda11.8-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
build_name: libtorch-cuda11_8-shared-with-deps-cxx11-abi build_name: libtorch-cuda11_8-shared-with-deps-cxx11-abi
@ -195,7 +195,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cuda12.1-main DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cuda12.1-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -218,7 +218,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cuda12.1-main DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cuda12.1-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
build_name: libtorch-cuda12_1-shared-with-deps-cxx11-abi build_name: libtorch-cuda12_1-shared-with-deps-cxx11-abi
@ -242,7 +242,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cuda12.1-main DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cuda12.1-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
build_name: libtorch-cuda12_1-shared-with-deps-cxx11-abi build_name: libtorch-cuda12_1-shared-with-deps-cxx11-abi
@ -265,7 +265,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cuda12.4-main DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cuda12.4-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -288,7 +288,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cuda12.4-main DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cuda12.4-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
build_name: libtorch-cuda12_4-shared-with-deps-cxx11-abi build_name: libtorch-cuda12_4-shared-with-deps-cxx11-abi
@ -312,7 +312,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cuda12.4-main DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cuda12.4-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
build_name: libtorch-cuda12_4-shared-with-deps-cxx11-abi build_name: libtorch-cuda12_4-shared-with-deps-cxx11-abi
@ -335,7 +335,7 @@ jobs:
DESIRED_CUDA: rocm6.1 DESIRED_CUDA: rocm6.1
GPU_ARCH_VERSION: 6.1 GPU_ARCH_VERSION: 6.1
GPU_ARCH_TYPE: rocm GPU_ARCH_TYPE: rocm
DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:rocm6.1-main DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:rocm6.1-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -360,7 +360,7 @@ jobs:
GPU_ARCH_VERSION: 6.1 GPU_ARCH_VERSION: 6.1
GPU_ARCH_TYPE: rocm GPU_ARCH_TYPE: rocm
SKIP_ALL_TESTS: 1 SKIP_ALL_TESTS: 1
DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:rocm6.1-main DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:rocm6.1-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
steps: steps:
@ -374,7 +374,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -386,7 +385,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -400,9 +399,9 @@ jobs:
run: | run: |
echo "GPU_FLAG=--device=/dev/mem --device=/dev/kfd --device=/dev/dri --group-add video --group-add daemon" >> "${GITHUB_ENV}" echo "GPU_FLAG=--device=/dev/mem --device=/dev/kfd --device=/dev/dri --group-add video --group-add daemon" >> "${GITHUB_ENV}"
- name: Pull Docker image - name: Pull Docker image
uses: pytorch/test-infra/.github/actions/pull-docker-image@main uses: pytorch/test-infra/.github/actions/pull-docker-image@release/2.5
with: with:
docker-image: pytorch/libtorch-cxx11-builder:rocm6.1-main docker-image: pytorch/libtorch-cxx11-builder:rocm6.1-2.5
- name: Test Pytorch binary - name: Test Pytorch binary
uses: ./pytorch/.github/actions/test-pytorch-binary uses: ./pytorch/.github/actions/test-pytorch-binary
- name: Teardown ROCm - name: Teardown ROCm
@ -422,7 +421,7 @@ jobs:
DESIRED_CUDA: rocm6.1 DESIRED_CUDA: rocm6.1
GPU_ARCH_VERSION: 6.1 GPU_ARCH_VERSION: 6.1
GPU_ARCH_TYPE: rocm GPU_ARCH_TYPE: rocm
DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:rocm6.1-main DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:rocm6.1-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
build_name: libtorch-rocm6_1-shared-with-deps-cxx11-abi build_name: libtorch-rocm6_1-shared-with-deps-cxx11-abi
@ -445,7 +444,7 @@ jobs:
DESIRED_CUDA: rocm6.2 DESIRED_CUDA: rocm6.2
GPU_ARCH_VERSION: 6.2 GPU_ARCH_VERSION: 6.2
GPU_ARCH_TYPE: rocm GPU_ARCH_TYPE: rocm
DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:rocm6.2-main DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:rocm6.2-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -470,7 +469,7 @@ jobs:
GPU_ARCH_VERSION: 6.2 GPU_ARCH_VERSION: 6.2
GPU_ARCH_TYPE: rocm GPU_ARCH_TYPE: rocm
SKIP_ALL_TESTS: 1 SKIP_ALL_TESTS: 1
DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:rocm6.2-main DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:rocm6.2-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
steps: steps:
@ -484,7 +483,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -496,7 +494,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -510,9 +508,9 @@ jobs:
run: | run: |
echo "GPU_FLAG=--device=/dev/mem --device=/dev/kfd --device=/dev/dri --group-add video --group-add daemon" >> "${GITHUB_ENV}" echo "GPU_FLAG=--device=/dev/mem --device=/dev/kfd --device=/dev/dri --group-add video --group-add daemon" >> "${GITHUB_ENV}"
- name: Pull Docker image - name: Pull Docker image
uses: pytorch/test-infra/.github/actions/pull-docker-image@main uses: pytorch/test-infra/.github/actions/pull-docker-image@release/2.5
with: with:
docker-image: pytorch/libtorch-cxx11-builder:rocm6.2-main docker-image: pytorch/libtorch-cxx11-builder:rocm6.2-2.5
- name: Test Pytorch binary - name: Test Pytorch binary
uses: ./pytorch/.github/actions/test-pytorch-binary uses: ./pytorch/.github/actions/test-pytorch-binary
- name: Teardown ROCm - name: Teardown ROCm
@ -532,7 +530,7 @@ jobs:
DESIRED_CUDA: rocm6.2 DESIRED_CUDA: rocm6.2
GPU_ARCH_VERSION: 6.2 GPU_ARCH_VERSION: 6.2
GPU_ARCH_TYPE: rocm GPU_ARCH_TYPE: rocm
DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:rocm6.2-main DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:rocm6.2-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
build_name: libtorch-rocm6_2-shared-with-deps-cxx11-abi build_name: libtorch-rocm6_2-shared-with-deps-cxx11-abi

View File

@ -52,7 +52,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/manylinux-builder:cpu-main DOCKER_IMAGE: pytorch/manylinux-builder:cpu-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: pre-cxx11 DESIRED_DEVTOOLSET: pre-cxx11
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -74,7 +74,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/manylinux-builder:cpu-main DOCKER_IMAGE: pytorch/manylinux-builder:cpu-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: pre-cxx11 DESIRED_DEVTOOLSET: pre-cxx11
build_name: libtorch-cpu-shared-with-deps-pre-cxx11 build_name: libtorch-cpu-shared-with-deps-pre-cxx11

View File

@ -57,7 +57,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/manylinux-builder:cpu-main DOCKER_IMAGE: pytorch/manylinux-builder:cpu-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: pre-cxx11 DESIRED_DEVTOOLSET: pre-cxx11
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -79,7 +79,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/manylinux-builder:cpu-main DOCKER_IMAGE: pytorch/manylinux-builder:cpu-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: pre-cxx11 DESIRED_DEVTOOLSET: pre-cxx11
build_name: libtorch-cpu-shared-with-deps-pre-cxx11 build_name: libtorch-cpu-shared-with-deps-pre-cxx11
@ -102,7 +102,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/manylinux-builder:cpu-main DOCKER_IMAGE: pytorch/manylinux-builder:cpu-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: pre-cxx11 DESIRED_DEVTOOLSET: pre-cxx11
build_name: libtorch-cpu-shared-with-deps-pre-cxx11 build_name: libtorch-cpu-shared-with-deps-pre-cxx11
@ -125,7 +125,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: pre-cxx11 DESIRED_DEVTOOLSET: pre-cxx11
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -148,7 +148,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: pre-cxx11 DESIRED_DEVTOOLSET: pre-cxx11
build_name: libtorch-cuda11_8-shared-with-deps-pre-cxx11 build_name: libtorch-cuda11_8-shared-with-deps-pre-cxx11
@ -172,7 +172,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: pre-cxx11 DESIRED_DEVTOOLSET: pre-cxx11
build_name: libtorch-cuda11_8-shared-with-deps-pre-cxx11 build_name: libtorch-cuda11_8-shared-with-deps-pre-cxx11
@ -195,7 +195,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: pre-cxx11 DESIRED_DEVTOOLSET: pre-cxx11
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -218,7 +218,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: pre-cxx11 DESIRED_DEVTOOLSET: pre-cxx11
build_name: libtorch-cuda12_1-shared-with-deps-pre-cxx11 build_name: libtorch-cuda12_1-shared-with-deps-pre-cxx11
@ -242,7 +242,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: pre-cxx11 DESIRED_DEVTOOLSET: pre-cxx11
build_name: libtorch-cuda12_1-shared-with-deps-pre-cxx11 build_name: libtorch-cuda12_1-shared-with-deps-pre-cxx11
@ -265,7 +265,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: pre-cxx11 DESIRED_DEVTOOLSET: pre-cxx11
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -288,7 +288,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: pre-cxx11 DESIRED_DEVTOOLSET: pre-cxx11
build_name: libtorch-cuda12_4-shared-with-deps-pre-cxx11 build_name: libtorch-cuda12_4-shared-with-deps-pre-cxx11
@ -312,7 +312,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: pre-cxx11 DESIRED_DEVTOOLSET: pre-cxx11
build_name: libtorch-cuda12_4-shared-with-deps-pre-cxx11 build_name: libtorch-cuda12_4-shared-with-deps-pre-cxx11
@ -335,7 +335,7 @@ jobs:
DESIRED_CUDA: rocm6.1 DESIRED_CUDA: rocm6.1
GPU_ARCH_VERSION: 6.1 GPU_ARCH_VERSION: 6.1
GPU_ARCH_TYPE: rocm GPU_ARCH_TYPE: rocm
DOCKER_IMAGE: pytorch/manylinux-builder:rocm6.1-main DOCKER_IMAGE: pytorch/manylinux-builder:rocm6.1-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: pre-cxx11 DESIRED_DEVTOOLSET: pre-cxx11
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -360,7 +360,7 @@ jobs:
GPU_ARCH_VERSION: 6.1 GPU_ARCH_VERSION: 6.1
GPU_ARCH_TYPE: rocm GPU_ARCH_TYPE: rocm
SKIP_ALL_TESTS: 1 SKIP_ALL_TESTS: 1
DOCKER_IMAGE: pytorch/manylinux-builder:rocm6.1-main DOCKER_IMAGE: pytorch/manylinux-builder:rocm6.1-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: pre-cxx11 DESIRED_DEVTOOLSET: pre-cxx11
steps: steps:
@ -374,7 +374,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -386,7 +385,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -400,9 +399,9 @@ jobs:
run: | run: |
echo "GPU_FLAG=--device=/dev/mem --device=/dev/kfd --device=/dev/dri --group-add video --group-add daemon" >> "${GITHUB_ENV}" echo "GPU_FLAG=--device=/dev/mem --device=/dev/kfd --device=/dev/dri --group-add video --group-add daemon" >> "${GITHUB_ENV}"
- name: Pull Docker image - name: Pull Docker image
uses: pytorch/test-infra/.github/actions/pull-docker-image@main uses: pytorch/test-infra/.github/actions/pull-docker-image@release/2.5
with: with:
docker-image: pytorch/manylinux-builder:rocm6.1-main docker-image: pytorch/manylinux-builder:rocm6.1-2.5
- name: Test Pytorch binary - name: Test Pytorch binary
uses: ./pytorch/.github/actions/test-pytorch-binary uses: ./pytorch/.github/actions/test-pytorch-binary
- name: Teardown ROCm - name: Teardown ROCm
@ -422,7 +421,7 @@ jobs:
DESIRED_CUDA: rocm6.1 DESIRED_CUDA: rocm6.1
GPU_ARCH_VERSION: 6.1 GPU_ARCH_VERSION: 6.1
GPU_ARCH_TYPE: rocm GPU_ARCH_TYPE: rocm
DOCKER_IMAGE: pytorch/manylinux-builder:rocm6.1-main DOCKER_IMAGE: pytorch/manylinux-builder:rocm6.1-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: pre-cxx11 DESIRED_DEVTOOLSET: pre-cxx11
build_name: libtorch-rocm6_1-shared-with-deps-pre-cxx11 build_name: libtorch-rocm6_1-shared-with-deps-pre-cxx11
@ -445,7 +444,7 @@ jobs:
DESIRED_CUDA: rocm6.2 DESIRED_CUDA: rocm6.2
GPU_ARCH_VERSION: 6.2 GPU_ARCH_VERSION: 6.2
GPU_ARCH_TYPE: rocm GPU_ARCH_TYPE: rocm
DOCKER_IMAGE: pytorch/manylinux-builder:rocm6.2-main DOCKER_IMAGE: pytorch/manylinux-builder:rocm6.2-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: pre-cxx11 DESIRED_DEVTOOLSET: pre-cxx11
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -470,7 +469,7 @@ jobs:
GPU_ARCH_VERSION: 6.2 GPU_ARCH_VERSION: 6.2
GPU_ARCH_TYPE: rocm GPU_ARCH_TYPE: rocm
SKIP_ALL_TESTS: 1 SKIP_ALL_TESTS: 1
DOCKER_IMAGE: pytorch/manylinux-builder:rocm6.2-main DOCKER_IMAGE: pytorch/manylinux-builder:rocm6.2-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: pre-cxx11 DESIRED_DEVTOOLSET: pre-cxx11
steps: steps:
@ -484,7 +483,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -496,7 +494,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -510,9 +508,9 @@ jobs:
run: | run: |
echo "GPU_FLAG=--device=/dev/mem --device=/dev/kfd --device=/dev/dri --group-add video --group-add daemon" >> "${GITHUB_ENV}" echo "GPU_FLAG=--device=/dev/mem --device=/dev/kfd --device=/dev/dri --group-add video --group-add daemon" >> "${GITHUB_ENV}"
- name: Pull Docker image - name: Pull Docker image
uses: pytorch/test-infra/.github/actions/pull-docker-image@main uses: pytorch/test-infra/.github/actions/pull-docker-image@release/2.5
with: with:
docker-image: pytorch/manylinux-builder:rocm6.2-main docker-image: pytorch/manylinux-builder:rocm6.2-2.5
- name: Test Pytorch binary - name: Test Pytorch binary
uses: ./pytorch/.github/actions/test-pytorch-binary uses: ./pytorch/.github/actions/test-pytorch-binary
- name: Teardown ROCm - name: Teardown ROCm
@ -532,7 +530,7 @@ jobs:
DESIRED_CUDA: rocm6.2 DESIRED_CUDA: rocm6.2
GPU_ARCH_VERSION: 6.2 GPU_ARCH_VERSION: 6.2
GPU_ARCH_TYPE: rocm GPU_ARCH_TYPE: rocm
DOCKER_IMAGE: pytorch/manylinux-builder:rocm6.2-main DOCKER_IMAGE: pytorch/manylinux-builder:rocm6.2-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: pre-cxx11 DESIRED_DEVTOOLSET: pre-cxx11
build_name: libtorch-rocm6_2-shared-with-deps-pre-cxx11 build_name: libtorch-rocm6_2-shared-with-deps-pre-cxx11

View File

@ -53,7 +53,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -77,7 +77,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: manywheel-py3_9-cuda11_8 build_name: manywheel-py3_9-cuda11_8
@ -100,7 +100,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -124,7 +124,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: manywheel-py3_9-cuda12_1 build_name: manywheel-py3_9-cuda12_1
@ -147,7 +147,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -171,7 +171,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: manywheel-py3_9-cuda12_4 build_name: manywheel-py3_9-cuda12_4

View File

@ -53,7 +53,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -77,7 +77,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: manywheel-py3_9-cuda11_8 build_name: manywheel-py3_9-cuda11_8
@ -100,7 +100,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -124,7 +124,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: manywheel-py3_9-cuda12_1 build_name: manywheel-py3_9-cuda12_1
@ -147,7 +147,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -171,7 +171,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: manywheel-py3_9-cuda12_4 build_name: manywheel-py3_9-cuda12_4

View File

@ -58,7 +58,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -82,7 +82,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: manywheel-py3_9-cuda11_8 build_name: manywheel-py3_9-cuda11_8
@ -106,7 +106,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: manywheel-py3_9-cuda11_8 build_name: manywheel-py3_9-cuda11_8
@ -129,7 +129,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -153,7 +153,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: manywheel-py3_9-cuda12_1 build_name: manywheel-py3_9-cuda12_1
@ -177,7 +177,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: manywheel-py3_9-cuda12_1 build_name: manywheel-py3_9-cuda12_1
@ -200,7 +200,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -224,7 +224,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: manywheel-py3_9-cuda12_4 build_name: manywheel-py3_9-cuda12_4
@ -248,7 +248,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: manywheel-py3_9-cuda12_4 build_name: manywheel-py3_9-cuda12_4
@ -270,7 +270,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/manylinux-builder:cpu-main DOCKER_IMAGE: pytorch/manylinux-builder:cpu-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -292,7 +292,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/manylinux-builder:cpu-main DOCKER_IMAGE: pytorch/manylinux-builder:cpu-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: manywheel-py3_9-cpu build_name: manywheel-py3_9-cpu
@ -315,7 +315,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/manylinux-builder:cpu-main DOCKER_IMAGE: pytorch/manylinux-builder:cpu-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: manywheel-py3_9-cpu build_name: manywheel-py3_9-cpu
@ -338,7 +338,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -362,7 +362,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
build_name: manywheel-py3_10-cuda11_8 build_name: manywheel-py3_10-cuda11_8
@ -386,7 +386,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
build_name: manywheel-py3_10-cuda11_8 build_name: manywheel-py3_10-cuda11_8
@ -409,7 +409,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -433,7 +433,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
build_name: manywheel-py3_10-cuda12_1 build_name: manywheel-py3_10-cuda12_1
@ -457,7 +457,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
build_name: manywheel-py3_10-cuda12_1 build_name: manywheel-py3_10-cuda12_1
@ -480,7 +480,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -503,7 +503,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
build_name: manywheel-py3_10-cuda12_1-full build_name: manywheel-py3_10-cuda12_1-full
@ -527,7 +527,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
build_name: manywheel-py3_10-cuda12_1-full build_name: manywheel-py3_10-cuda12_1-full
@ -550,7 +550,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -574,7 +574,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
build_name: manywheel-py3_10-cuda12_4 build_name: manywheel-py3_10-cuda12_4
@ -598,7 +598,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
build_name: manywheel-py3_10-cuda12_4 build_name: manywheel-py3_10-cuda12_4
@ -620,7 +620,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/manylinux-builder:cpu-main DOCKER_IMAGE: pytorch/manylinux-builder:cpu-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -642,7 +642,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/manylinux-builder:cpu-main DOCKER_IMAGE: pytorch/manylinux-builder:cpu-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
build_name: manywheel-py3_10-cpu build_name: manywheel-py3_10-cpu
@ -665,7 +665,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/manylinux-builder:cpu-main DOCKER_IMAGE: pytorch/manylinux-builder:cpu-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
build_name: manywheel-py3_10-cpu build_name: manywheel-py3_10-cpu
@ -688,7 +688,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -712,7 +712,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
build_name: manywheel-py3_11-cuda11_8 build_name: manywheel-py3_11-cuda11_8
@ -736,7 +736,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
build_name: manywheel-py3_11-cuda11_8 build_name: manywheel-py3_11-cuda11_8
@ -759,7 +759,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -783,7 +783,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
build_name: manywheel-py3_11-cuda12_1 build_name: manywheel-py3_11-cuda12_1
@ -807,7 +807,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
build_name: manywheel-py3_11-cuda12_1 build_name: manywheel-py3_11-cuda12_1
@ -830,7 +830,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -854,7 +854,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
build_name: manywheel-py3_11-cuda12_4 build_name: manywheel-py3_11-cuda12_4
@ -878,7 +878,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
build_name: manywheel-py3_11-cuda12_4 build_name: manywheel-py3_11-cuda12_4
@ -900,7 +900,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/manylinux-builder:cpu-main DOCKER_IMAGE: pytorch/manylinux-builder:cpu-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -922,7 +922,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/manylinux-builder:cpu-main DOCKER_IMAGE: pytorch/manylinux-builder:cpu-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
build_name: manywheel-py3_11-cpu build_name: manywheel-py3_11-cpu
@ -945,7 +945,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/manylinux-builder:cpu-main DOCKER_IMAGE: pytorch/manylinux-builder:cpu-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
build_name: manywheel-py3_11-cpu build_name: manywheel-py3_11-cpu
@ -968,7 +968,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -992,7 +992,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
build_name: manywheel-py3_12-cuda11_8 build_name: manywheel-py3_12-cuda11_8
@ -1016,7 +1016,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
build_name: manywheel-py3_12-cuda11_8 build_name: manywheel-py3_12-cuda11_8
@ -1039,7 +1039,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -1063,7 +1063,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
build_name: manywheel-py3_12-cuda12_1 build_name: manywheel-py3_12-cuda12_1
@ -1087,7 +1087,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
build_name: manywheel-py3_12-cuda12_1 build_name: manywheel-py3_12-cuda12_1
@ -1110,7 +1110,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -1134,7 +1134,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
build_name: manywheel-py3_12-cuda12_4 build_name: manywheel-py3_12-cuda12_4
@ -1158,7 +1158,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
build_name: manywheel-py3_12-cuda12_4 build_name: manywheel-py3_12-cuda12_4
@ -1180,7 +1180,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/manylinux-builder:cpu-main DOCKER_IMAGE: pytorch/manylinux-builder:cpu-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -1202,7 +1202,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/manylinux-builder:cpu-main DOCKER_IMAGE: pytorch/manylinux-builder:cpu-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
build_name: manywheel-py3_12-cpu build_name: manywheel-py3_12-cpu
@ -1225,7 +1225,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/manylinux-builder:cpu-main DOCKER_IMAGE: pytorch/manylinux-builder:cpu-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
build_name: manywheel-py3_12-cpu build_name: manywheel-py3_12-cpu
@ -1248,7 +1248,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.13" DESIRED_PYTHON: "3.13"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -1272,7 +1272,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.13" DESIRED_PYTHON: "3.13"
build_name: manywheel-py3_13-cuda11_8 build_name: manywheel-py3_13-cuda11_8
@ -1296,7 +1296,7 @@ jobs:
DESIRED_CUDA: cu118 DESIRED_CUDA: cu118
GPU_ARCH_VERSION: 11.8 GPU_ARCH_VERSION: 11.8
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda11.8-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.13" DESIRED_PYTHON: "3.13"
build_name: manywheel-py3_13-cuda11_8 build_name: manywheel-py3_13-cuda11_8
@ -1319,7 +1319,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.13" DESIRED_PYTHON: "3.13"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -1343,7 +1343,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.13" DESIRED_PYTHON: "3.13"
build_name: manywheel-py3_13-cuda12_1 build_name: manywheel-py3_13-cuda12_1
@ -1367,7 +1367,7 @@ jobs:
DESIRED_CUDA: cu121 DESIRED_CUDA: cu121
GPU_ARCH_VERSION: 12.1 GPU_ARCH_VERSION: 12.1
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.1-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.13" DESIRED_PYTHON: "3.13"
build_name: manywheel-py3_13-cuda12_1 build_name: manywheel-py3_13-cuda12_1
@ -1390,7 +1390,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.13" DESIRED_PYTHON: "3.13"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -1414,7 +1414,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.13" DESIRED_PYTHON: "3.13"
build_name: manywheel-py3_13-cuda12_4 build_name: manywheel-py3_13-cuda12_4
@ -1438,7 +1438,7 @@ jobs:
DESIRED_CUDA: cu124 DESIRED_CUDA: cu124
GPU_ARCH_VERSION: 12.4 GPU_ARCH_VERSION: 12.4
GPU_ARCH_TYPE: cuda GPU_ARCH_TYPE: cuda
DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-main DOCKER_IMAGE: pytorch/manylinux-builder:cuda12.4-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.13" DESIRED_PYTHON: "3.13"
build_name: manywheel-py3_13-cuda12_4 build_name: manywheel-py3_13-cuda12_4
@ -1460,7 +1460,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/manylinux-builder:cpu-main DOCKER_IMAGE: pytorch/manylinux-builder:cpu-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.13" DESIRED_PYTHON: "3.13"
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
@ -1482,7 +1482,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/manylinux-builder:cpu-main DOCKER_IMAGE: pytorch/manylinux-builder:cpu-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.13" DESIRED_PYTHON: "3.13"
build_name: manywheel-py3_13-cpu build_name: manywheel-py3_13-cpu
@ -1505,7 +1505,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/manylinux-builder:cpu-main DOCKER_IMAGE: pytorch/manylinux-builder:cpu-2.5
use_split_build: True use_split_build: True
DESIRED_PYTHON: "3.13" DESIRED_PYTHON: "3.13"
build_name: manywheel-py3_13-cpu build_name: manywheel-py3_13-cpu

File diff suppressed because it is too large Load Diff

View File

@ -57,7 +57,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu-s390x GPU_ARCH_TYPE: cpu-s390x
DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-main DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
runs_on: linux.s390x runs_on: linux.s390x
@ -81,7 +81,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu-s390x GPU_ARCH_TYPE: cpu-s390x
DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-main DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: manywheel-py3_9-cpu-s390x build_name: manywheel-py3_9-cpu-s390x
@ -104,7 +104,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu-s390x GPU_ARCH_TYPE: cpu-s390x
DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-main DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: manywheel-py3_9-cpu-s390x build_name: manywheel-py3_9-cpu-s390x
@ -126,7 +126,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu-s390x GPU_ARCH_TYPE: cpu-s390x
DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-main DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
runs_on: linux.s390x runs_on: linux.s390x
@ -150,7 +150,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu-s390x GPU_ARCH_TYPE: cpu-s390x
DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-main DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
build_name: manywheel-py3_10-cpu-s390x build_name: manywheel-py3_10-cpu-s390x
@ -173,7 +173,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu-s390x GPU_ARCH_TYPE: cpu-s390x
DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-main DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
build_name: manywheel-py3_10-cpu-s390x build_name: manywheel-py3_10-cpu-s390x
@ -195,7 +195,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu-s390x GPU_ARCH_TYPE: cpu-s390x
DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-main DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
runs_on: linux.s390x runs_on: linux.s390x
@ -219,7 +219,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu-s390x GPU_ARCH_TYPE: cpu-s390x
DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-main DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
build_name: manywheel-py3_11-cpu-s390x build_name: manywheel-py3_11-cpu-s390x
@ -242,7 +242,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu-s390x GPU_ARCH_TYPE: cpu-s390x
DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-main DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
build_name: manywheel-py3_11-cpu-s390x build_name: manywheel-py3_11-cpu-s390x
@ -264,7 +264,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu-s390x GPU_ARCH_TYPE: cpu-s390x
DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-main DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
runs_on: linux.s390x runs_on: linux.s390x
@ -288,7 +288,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu-s390x GPU_ARCH_TYPE: cpu-s390x
DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-main DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
build_name: manywheel-py3_12-cpu-s390x build_name: manywheel-py3_12-cpu-s390x
@ -311,7 +311,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu-s390x GPU_ARCH_TYPE: cpu-s390x
DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-main DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
build_name: manywheel-py3_12-cpu-s390x build_name: manywheel-py3_12-cpu-s390x
@ -333,7 +333,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu-s390x GPU_ARCH_TYPE: cpu-s390x
DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-main DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.13" DESIRED_PYTHON: "3.13"
runs_on: linux.s390x runs_on: linux.s390x
@ -357,7 +357,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu-s390x GPU_ARCH_TYPE: cpu-s390x
DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-main DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.13" DESIRED_PYTHON: "3.13"
build_name: manywheel-py3_13-cpu-s390x build_name: manywheel-py3_13-cpu-s390x
@ -380,7 +380,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu-s390x GPU_ARCH_TYPE: cpu-s390x
DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-main DOCKER_IMAGE: pytorch/manylinuxs390x-builder:cpu-s390x-2.5
use_split_build: False use_split_build: False
DESIRED_PYTHON: "3.13" DESIRED_PYTHON: "3.13"
build_name: manywheel-py3_13-cpu-s390x build_name: manywheel-py3_13-cpu-s390x

View File

@ -74,7 +74,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -86,7 +85,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -138,7 +137,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/conda-builder:cpu-main DOCKER_IMAGE: pytorch/conda-builder:cpu-2.5
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: conda-py3_9-cpu build_name: conda-py3_9-cpu
use_s3: False use_s3: False
@ -189,7 +188,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -201,7 +199,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -253,7 +251,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/conda-builder:cpu-main DOCKER_IMAGE: pytorch/conda-builder:cpu-2.5
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
build_name: conda-py3_10-cpu build_name: conda-py3_10-cpu
use_s3: False use_s3: False
@ -304,7 +302,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -316,7 +313,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -368,7 +365,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/conda-builder:cpu-main DOCKER_IMAGE: pytorch/conda-builder:cpu-2.5
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
build_name: conda-py3_11-cpu build_name: conda-py3_11-cpu
use_s3: False use_s3: False
@ -419,7 +416,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -431,7 +427,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -483,7 +479,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/conda-builder:cpu-main DOCKER_IMAGE: pytorch/conda-builder:cpu-2.5
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
build_name: conda-py3_12-cpu build_name: conda-py3_12-cpu
use_s3: False use_s3: False

View File

@ -78,7 +78,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -90,7 +89,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -142,7 +141,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cpu-main DOCKER_IMAGE: pytorch/libtorch-cxx11-builder:cpu-2.5
LIBTORCH_VARIANT: shared-with-deps LIBTORCH_VARIANT: shared-with-deps
DESIRED_DEVTOOLSET: cxx11-abi DESIRED_DEVTOOLSET: cxx11-abi
build_name: libtorch-cpu-shared-with-deps-cxx11-abi build_name: libtorch-cpu-shared-with-deps-cxx11-abi

View File

@ -75,7 +75,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -87,7 +86,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -139,7 +138,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/manylinux-builder:cpu-main DOCKER_IMAGE: pytorch/manylinux-builder:cpu-2.5
DESIRED_PYTHON: "3.9" DESIRED_PYTHON: "3.9"
build_name: wheel-py3_9-cpu build_name: wheel-py3_9-cpu
use_s3: False use_s3: False
@ -191,7 +190,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -203,7 +201,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -255,7 +253,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/manylinux-builder:cpu-main DOCKER_IMAGE: pytorch/manylinux-builder:cpu-2.5
DESIRED_PYTHON: "3.10" DESIRED_PYTHON: "3.10"
build_name: wheel-py3_10-cpu build_name: wheel-py3_10-cpu
use_s3: False use_s3: False
@ -307,7 +305,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -319,7 +316,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -371,7 +368,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/manylinux-builder:cpu-main DOCKER_IMAGE: pytorch/manylinux-builder:cpu-2.5
DESIRED_PYTHON: "3.11" DESIRED_PYTHON: "3.11"
build_name: wheel-py3_11-cpu build_name: wheel-py3_11-cpu
use_s3: False use_s3: False
@ -423,7 +420,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -435,7 +431,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -487,7 +483,7 @@ jobs:
# favor of GPU_ARCH_VERSION # favor of GPU_ARCH_VERSION
DESIRED_CUDA: cpu DESIRED_CUDA: cpu
GPU_ARCH_TYPE: cpu GPU_ARCH_TYPE: cpu
DOCKER_IMAGE: pytorch/manylinux-builder:cpu-main DOCKER_IMAGE: pytorch/manylinux-builder:cpu-2.5
DESIRED_PYTHON: "3.12" DESIRED_PYTHON: "3.12"
build_name: wheel-py3_12-cpu build_name: wheel-py3_12-cpu
use_s3: False use_s3: False

View File

@ -71,7 +71,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -102,7 +102,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -114,7 +113,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -185,7 +184,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -221,7 +220,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -233,7 +231,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -317,7 +315,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -348,7 +346,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -360,7 +357,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -432,7 +429,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -468,7 +465,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -480,7 +476,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -565,7 +561,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -596,7 +592,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -608,7 +603,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -680,7 +675,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -716,7 +711,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -728,7 +722,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -813,7 +807,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -844,7 +838,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -856,7 +849,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -928,7 +921,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -964,7 +957,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -976,7 +968,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -1060,7 +1052,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -1091,7 +1083,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -1103,7 +1094,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -1174,7 +1165,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -1210,7 +1201,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -1222,7 +1212,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -1306,7 +1296,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -1337,7 +1327,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -1349,7 +1338,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -1421,7 +1410,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -1457,7 +1446,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -1469,7 +1457,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -1554,7 +1542,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -1585,7 +1573,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -1597,7 +1584,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -1669,7 +1656,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -1705,7 +1692,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -1717,7 +1703,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -1802,7 +1788,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -1833,7 +1819,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -1845,7 +1830,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -1917,7 +1902,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -1953,7 +1938,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -1965,7 +1949,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -2049,7 +2033,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -2080,7 +2064,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -2092,7 +2075,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -2163,7 +2146,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -2199,7 +2182,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -2211,7 +2193,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -2295,7 +2277,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -2326,7 +2308,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -2338,7 +2319,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -2410,7 +2391,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -2446,7 +2427,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -2458,7 +2438,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -2543,7 +2523,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -2574,7 +2554,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -2586,7 +2565,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -2658,7 +2637,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -2694,7 +2673,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -2706,7 +2684,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -2791,7 +2769,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -2822,7 +2800,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -2834,7 +2811,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -2906,7 +2883,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -2942,7 +2919,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -2954,7 +2930,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -3038,7 +3014,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -3069,7 +3045,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -3081,7 +3056,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -3152,7 +3127,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -3188,7 +3163,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -3200,7 +3174,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -3284,7 +3258,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -3315,7 +3289,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -3327,7 +3300,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -3399,7 +3372,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -3435,7 +3408,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -3447,7 +3419,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -3532,7 +3504,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -3563,7 +3535,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -3575,7 +3546,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -3647,7 +3618,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -3683,7 +3654,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -3695,7 +3665,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -3780,7 +3750,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -3811,7 +3781,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -3823,7 +3792,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -3895,7 +3864,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -3931,7 +3900,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -3943,7 +3911,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder

View File

@ -68,7 +68,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -99,7 +99,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -111,7 +110,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -186,7 +185,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -222,7 +221,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -234,7 +232,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder

View File

@ -75,7 +75,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -106,7 +106,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -118,7 +117,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -193,7 +192,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -229,7 +228,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -241,7 +239,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -333,7 +331,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -364,7 +362,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -376,7 +373,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -452,7 +449,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -488,7 +485,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -500,7 +496,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -593,7 +589,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -624,7 +620,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -636,7 +631,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -712,7 +707,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -748,7 +743,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -760,7 +754,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -853,7 +847,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -884,7 +878,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -896,7 +889,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -972,7 +965,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -1008,7 +1001,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -1020,7 +1012,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder

View File

@ -68,7 +68,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -99,7 +99,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -111,7 +110,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -186,7 +185,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -222,7 +221,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -234,7 +232,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder

View File

@ -75,7 +75,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -106,7 +106,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -118,7 +117,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -193,7 +192,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -229,7 +228,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -241,7 +239,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -333,7 +331,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -364,7 +362,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -376,7 +373,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -452,7 +449,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -488,7 +485,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -500,7 +496,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -593,7 +589,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -624,7 +620,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -636,7 +631,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -712,7 +707,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -748,7 +743,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -760,7 +754,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -853,7 +847,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -884,7 +878,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -896,7 +889,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -972,7 +965,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -1008,7 +1001,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -1020,7 +1012,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder

View File

@ -72,7 +72,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -103,7 +103,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -115,7 +114,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -186,7 +185,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -222,7 +221,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -234,7 +232,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -319,7 +317,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -350,7 +348,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -362,7 +359,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -434,7 +431,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -470,7 +467,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -482,7 +478,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -568,7 +564,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -599,7 +595,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -611,7 +606,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -683,7 +678,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -719,7 +714,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -731,7 +725,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -817,7 +811,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -848,7 +842,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -860,7 +853,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -932,7 +925,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -968,7 +961,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -980,7 +972,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -1064,7 +1056,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -1095,7 +1087,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -1107,7 +1098,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -1178,7 +1169,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -1214,7 +1205,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -1226,7 +1216,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -1310,7 +1300,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -1341,7 +1331,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -1353,7 +1342,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -1424,7 +1413,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -1460,7 +1449,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -1472,7 +1460,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -1557,7 +1545,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -1588,7 +1576,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -1600,7 +1587,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -1672,7 +1659,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -1708,7 +1695,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -1720,7 +1706,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -1806,7 +1792,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -1837,7 +1823,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -1849,7 +1834,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -1921,7 +1906,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -1957,7 +1942,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -1969,7 +1953,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -2055,7 +2039,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -2086,7 +2070,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -2098,7 +2081,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -2170,7 +2153,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -2206,7 +2189,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -2218,7 +2200,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -2302,7 +2284,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -2333,7 +2315,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -2345,7 +2326,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -2416,7 +2397,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -2452,7 +2433,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -2464,7 +2444,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -2548,7 +2528,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -2579,7 +2559,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -2591,7 +2570,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -2662,7 +2641,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -2698,7 +2677,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -2710,7 +2688,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -2795,7 +2773,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -2826,7 +2804,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -2838,7 +2815,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -2910,7 +2887,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -2946,7 +2923,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -2958,7 +2934,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -3044,7 +3020,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -3075,7 +3051,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -3087,7 +3062,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -3159,7 +3134,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -3195,7 +3170,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -3207,7 +3181,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -3293,7 +3267,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -3324,7 +3298,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -3336,7 +3309,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -3408,7 +3381,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -3444,7 +3417,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -3456,7 +3428,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -3540,7 +3512,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -3571,7 +3543,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -3583,7 +3554,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -3654,7 +3625,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -3690,7 +3661,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -3702,7 +3672,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -3786,7 +3756,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -3817,7 +3787,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -3829,7 +3798,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -3900,7 +3869,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -3936,7 +3905,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -3948,7 +3916,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -4033,7 +4001,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -4064,7 +4032,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -4076,7 +4043,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -4148,7 +4115,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -4184,7 +4151,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -4196,7 +4162,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -4282,7 +4248,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -4313,7 +4279,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -4325,7 +4290,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -4397,7 +4362,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -4433,7 +4398,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -4445,7 +4409,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -4531,7 +4495,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -4562,7 +4526,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -4574,7 +4537,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -4646,7 +4609,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -4682,7 +4645,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -4694,7 +4656,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -4778,7 +4740,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -4809,7 +4771,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -4821,7 +4782,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder
@ -4892,7 +4853,7 @@ jobs:
echo "instance-type: $(get_ec2_metadata instance-type)" echo "instance-type: $(get_ec2_metadata instance-type)"
echo "system info $(uname -a)" echo "system info $(uname -a)"
- name: "[FB EMPLOYEES] Enable SSH (Click me for login details)" - name: "[FB EMPLOYEES] Enable SSH (Click me for login details)"
uses: pytorch/test-infra/.github/actions/setup-ssh@main uses: pytorch/test-infra/.github/actions/setup-ssh@release/2.5
continue-on-error: true continue-on-error: true
with: with:
github-secret: ${{ secrets.GITHUB_TOKEN }} github-secret: ${{ secrets.GITHUB_TOKEN }}
@ -4928,7 +4889,6 @@ jobs:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: ${{ github.event_name == 'pull_request' && github.event.pull_request.head.sha || github.sha }}
submodules: recursive submodules: recursive
path: pytorch path: pytorch
quiet-checkout: true quiet-checkout: true
@ -4940,7 +4900,7 @@ jobs:
- name: Checkout pytorch/builder - name: Checkout pytorch/builder
uses: malfet/checkout@silent-checkout uses: malfet/checkout@silent-checkout
with: with:
ref: main ref: release/2.5
submodules: recursive submodules: recursive
repository: pytorch/builder repository: pytorch/builder
path: builder path: builder

View File

@ -22,11 +22,11 @@ concurrency:
permissions: read-all permissions: read-all
jobs: jobs:
linux-focal-rocm6_1-py3_8-inductor-build: linux-focal-rocm6_2-py3_10-inductor-build:
name: rocm6.1-py3.8-inductor name: rocm6.2-py3.10-inductor
uses: ./.github/workflows/_linux-build.yml uses: ./.github/workflows/_linux-build.yml
with: with:
build-environment: linux-focal-rocm6.1-py3.8 build-environment: linux-focal-rocm6.2-py3.10
docker-image-name: pytorch-linux-focal-rocm-n-py3 docker-image-name: pytorch-linux-focal-rocm-n-py3
test-matrix: | test-matrix: |
{ include: [ { include: [
@ -34,14 +34,14 @@ jobs:
{ config: "inductor", shard: 2, num_shards: 2, runner: "linux.rocm.gpu.2" }, { config: "inductor", shard: 2, num_shards: 2, runner: "linux.rocm.gpu.2" },
]} ]}
linux-focal-rocm6_1-py3_8-inductor-test: linux-focal-rocm6_2-py3_10-inductor-test:
permissions: permissions:
id-token: write id-token: write
contents: read contents: read
name: rocm6.1-py3.8-inductor name: rocm6.2-py3.10-inductor
uses: ./.github/workflows/_rocm-test.yml uses: ./.github/workflows/_rocm-test.yml
needs: linux-focal-rocm6_1-py3_8-inductor-build needs: linux-focal-rocm6_2-py3_10-inductor-build
with: with:
build-environment: linux-focal-rocm6.1-py3.8 build-environment: linux-focal-rocm6.2-py3.10
docker-image: ${{ needs.linux-focal-rocm6_1-py3_8-inductor-build.outputs.docker-image }} docker-image: ${{ needs.linux-focal-rocm6_2-py3_10-inductor-build.outputs.docker-image }}
test-matrix: ${{ needs.linux-focal-rocm6_1-py3_8-inductor-build.outputs.test-matrix }} test-matrix: ${{ needs.linux-focal-rocm6_2-py3_10-inductor-build.outputs.test-matrix }}

View File

@ -15,7 +15,7 @@ jobs:
runs-on: ubuntu-latest runs-on: ubuntu-latest
steps: steps:
- name: Run BC Lint Action - name: Run BC Lint Action
uses: pytorch/test-infra/.github/actions/bc-lint@main uses: pytorch/test-infra/.github/actions/bc-lint@release/2.5
with: with:
repo: ${{ github.event.pull_request.head.repo.full_name }} repo: ${{ github.event.pull_request.head.repo.full_name }}
base_sha: ${{ github.event.pull_request.base.sha }} base_sha: ${{ github.event.pull_request.base.sha }}

View File

@ -24,7 +24,7 @@ jobs:
curr_branch: ${{ github.head_ref || github.ref_name }} curr_branch: ${{ github.head_ref || github.ref_name }}
lintrunner-clang: lintrunner-clang:
uses: pytorch/test-infra/.github/workflows/linux_job.yml@main uses: pytorch/test-infra/.github/workflows/linux_job.yml@release/2.5
needs: get-label-type needs: get-label-type
with: with:
timeout: 120 timeout: 120
@ -41,7 +41,7 @@ jobs:
.github/scripts/lintrunner.sh .github/scripts/lintrunner.sh
lintrunner-noclang: lintrunner-noclang:
uses: pytorch/test-infra/.github/workflows/linux_job.yml@main uses: pytorch/test-infra/.github/workflows/linux_job.yml@release/2.5
needs: get-label-type needs: get-label-type
with: with:
timeout: 120 timeout: 120
@ -57,7 +57,7 @@ jobs:
.github/scripts/lintrunner.sh .github/scripts/lintrunner.sh
quick-checks: quick-checks:
uses: pytorch/test-infra/.github/workflows/linux_job.yml@main uses: pytorch/test-infra/.github/workflows/linux_job.yml@release/2.5
needs: get-label-type needs: get-label-type
with: with:
runner: "${{ needs.get-label-type.outputs.label-type }}linux.2xlarge" runner: "${{ needs.get-label-type.outputs.label-type }}linux.2xlarge"
@ -100,7 +100,7 @@ jobs:
if: github.event_name == 'pull_request' && !contains(github.event.pull_request.labels.*.name, 'skip-pr-sanity-checks') if: github.event_name == 'pull_request' && !contains(github.event.pull_request.labels.*.name, 'skip-pr-sanity-checks')
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
submodules: false submodules: false
fetch-depth: -1 fetch-depth: -1
@ -113,7 +113,7 @@ jobs:
bash .github/scripts/pr-sanity-check.sh bash .github/scripts/pr-sanity-check.sh
workflow-checks: workflow-checks:
uses: pytorch/test-infra/.github/workflows/linux_job.yml@main uses: pytorch/test-infra/.github/workflows/linux_job.yml@release/2.5
needs: get-label-type needs: get-label-type
with: with:
runner: "${{ needs.get-label-type.outputs.label-type }}linux.2xlarge" runner: "${{ needs.get-label-type.outputs.label-type }}linux.2xlarge"
@ -127,6 +127,7 @@ jobs:
conda activate "${CONDA_ENV}" conda activate "${CONDA_ENV}"
# Regenerate workflows # Regenerate workflows
export RELEASE_VERSION_TAG="2.5"
.github/scripts/generate_ci_workflows.py .github/scripts/generate_ci_workflows.py
RC=0 RC=0
@ -150,7 +151,7 @@ jobs:
exit $RC exit $RC
toc: toc:
uses: pytorch/test-infra/.github/workflows/linux_job.yml@main uses: pytorch/test-infra/.github/workflows/linux_job.yml@release/2.5
needs: get-label-type needs: get-label-type
with: with:
runner: "${{ needs.get-label-type.outputs.label-type }}linux.2xlarge" runner: "${{ needs.get-label-type.outputs.label-type }}linux.2xlarge"
@ -189,7 +190,7 @@ jobs:
test-tools: test-tools:
name: Test tools name: Test tools
if: ${{ github.repository == 'pytorch/pytorch' }} if: ${{ github.repository == 'pytorch/pytorch' }}
uses: pytorch/test-infra/.github/workflows/linux_job.yml@main uses: pytorch/test-infra/.github/workflows/linux_job.yml@release/2.5
needs: get-label-type needs: get-label-type
with: with:
runner: "${{ needs.get-label-type.outputs.label-type }}linux.2xlarge" runner: "${{ needs.get-label-type.outputs.label-type }}linux.2xlarge"
@ -211,14 +212,14 @@ jobs:
runs-on: linux.20_04.4x runs-on: linux.20_04.4x
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
submodules: false submodules: false
fetch-depth: 1 fetch-depth: 1
- name: Setup Python 3.8 - name: Setup Python 3.9
uses: actions/setup-python@v4 uses: actions/setup-python@v4
with: with:
python-version: '3.8' python-version: '3.9'
architecture: x64 architecture: x64
cache: pip cache: pip
- name: Install dependencies - name: Install dependencies
@ -241,7 +242,7 @@ jobs:
# [see note: pytorch repo ref] # [see note: pytorch repo ref]
# deep clone (fetch-depth 0) required, to allow us to use git log # deep clone (fetch-depth 0) required, to allow us to use git log
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
submodules: false submodules: false
fetch-depth: 1 fetch-depth: 1

View File

@ -48,7 +48,7 @@ jobs:
path: llm-target-determinator path: llm-target-determinator
- name: Setup miniconda - name: Setup miniconda
uses: pytorch/test-infra/.github/actions/setup-miniconda@main uses: pytorch/test-infra/.github/actions/setup-miniconda@release/2.5
with: with:
python-version: "3.9" python-version: "3.9"
@ -117,5 +117,5 @@ jobs:
AWS_REGION: "" AWS_REGION: ""
- name: Teardown Linux - name: Teardown Linux
uses: pytorch/test-infra/.github/actions/teardown-linux@main uses: pytorch/test-infra/.github/actions/teardown-linux@release/2.5
if: always() if: always()

View File

@ -21,7 +21,7 @@ jobs:
environment: upload-stats environment: upload-stats
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
fetch-depth: 1 fetch-depth: 1
submodules: false submodules: false

View File

@ -56,7 +56,7 @@ jobs:
if: ${{ github.event_name == 'schedule' }} if: ${{ github.event_name == 'schedule' }}
steps: steps:
- name: update-vision-commit-hash - name: update-vision-commit-hash
uses: pytorch/test-infra/.github/actions/update-commit-hash@main uses: pytorch/test-infra/.github/actions/update-commit-hash@release/2.5
with: with:
repo-name: vision repo-name: vision
branch: main branch: main
@ -71,7 +71,7 @@ jobs:
if: ${{ github.event_name == 'schedule' }} if: ${{ github.event_name == 'schedule' }}
steps: steps:
- name: update-audio-commit-hash - name: update-audio-commit-hash
uses: pytorch/test-infra/.github/actions/update-commit-hash@main uses: pytorch/test-infra/.github/actions/update-commit-hash@release/2.5
with: with:
repo-name: audio repo-name: audio
branch: main branch: main
@ -86,7 +86,7 @@ jobs:
if: ${{ github.event_name == 'schedule' }} if: ${{ github.event_name == 'schedule' }}
steps: steps:
- name: update-executorch-commit-hash - name: update-executorch-commit-hash
uses: pytorch/test-infra/.github/actions/update-commit-hash@main uses: pytorch/test-infra/.github/actions/update-commit-hash@release/2.5
with: with:
repo-name: executorch repo-name: executorch
branch: main branch: main

View File

@ -19,7 +19,7 @@ jobs:
if: ${{ github.event.pull_request.number != 26921 && github.repository_owner == 'pytorch' }} if: ${{ github.event.pull_request.number != 26921 && github.repository_owner == 'pytorch' }}
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
- uses: ethanis/nitpicker@v1 - uses: ethanis/nitpicker@v1
with: with:
nitpicks: '.github/nitpicks.yml' nitpicks: '.github/nitpicks.yml'

View File

@ -214,7 +214,9 @@ jobs:
# TODO: Figure out how to migrate this job to M1 runner # TODO: Figure out how to migrate this job to M1 runner
ios-build-test: ios-build-test:
name: ios-build-test name: ios-build-test
if: github.event_name != 'schedule' || github.event.schedule == '45 0,8,16 * * 1-5' || github.event.schedule == '45 4 * * 0,6' || github.event.schedule == '29 8 * * *' # Has been broken for a while, see https://github.com/pytorch/pytorch/issues/136284
# if: github.event_name != 'schedule' || github.event.schedule == '45 0,8,16 * * 1-5' || github.event.schedule == '45 4 * * 0,6' || github.event.schedule == '29 8 * * *'
if: false
uses: ./.github/workflows/_ios-build-test.yml uses: ./.github/workflows/_ios-build-test.yml
with: with:
trigger-event: ${{ github.event_name }} trigger-event: ${{ github.event_name }}
@ -293,13 +295,13 @@ jobs:
docker-image: ${{ needs.linux-vulkan-focal-py3_11-clang10-build.outputs.docker-image }} docker-image: ${{ needs.linux-vulkan-focal-py3_11-clang10-build.outputs.docker-image }}
test-matrix: ${{ needs.linux-vulkan-focal-py3_11-clang10-build.outputs.test-matrix }} test-matrix: ${{ needs.linux-vulkan-focal-py3_11-clang10-build.outputs.test-matrix }}
linux-focal-rocm6_1-py3_8-build: linux-focal-rocm6_2-py3_10-build:
name: linux-focal-rocm6.1-py3.8 name: linux-focal-rocm6.2-py3.10
uses: ./.github/workflows/_linux-build.yml uses: ./.github/workflows/_linux-build.yml
needs: get-label-type needs: get-label-type
with: with:
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
build-environment: linux-focal-rocm6.1-py3.8 build-environment: linux-focal-rocm6.2-py3.10
docker-image-name: pytorch-linux-focal-rocm-n-py3 docker-image-name: pytorch-linux-focal-rocm-n-py3
test-matrix: | test-matrix: |
{ include: [ { include: [
@ -308,19 +310,19 @@ jobs:
{ config: "distributed", shard: 3, num_shards: 3, runner: "linux.rocm.gpu" }, { config: "distributed", shard: 3, num_shards: 3, runner: "linux.rocm.gpu" },
]} ]}
linux-focal-rocm6_1-py3_8-test: linux-focal-rocm6_2-py3_10-test:
permissions: permissions:
id-token: write id-token: write
contents: read contents: read
name: linux-focal-rocm6.1-py3.8 name: linux-focal-rocm6.2-py3.10
uses: ./.github/workflows/_rocm-test.yml uses: ./.github/workflows/_rocm-test.yml
needs: needs:
- linux-focal-rocm6_1-py3_8-build - linux-focal-rocm6_2-py3_10-build
- target-determination - target-determination
with: with:
build-environment: linux-focal-rocm6.1-py3.8 build-environment: linux-focal-rocm6.2-py3.10
docker-image: ${{ needs.linux-focal-rocm6_1-py3_8-build.outputs.docker-image }} docker-image: ${{ needs.linux-focal-rocm6_2-py3_10-build.outputs.docker-image }}
test-matrix: ${{ needs.linux-focal-rocm6_1-py3_8-build.outputs.test-matrix }} test-matrix: ${{ needs.linux-focal-rocm6_2-py3_10-build.outputs.test-matrix }}
linux-focal-cuda12_1-py3_10-gcc9-experimental-split-build: linux-focal-cuda12_1-py3_10-gcc9-experimental-split-build:
name: linux-focal-cuda12.1-py3.10-gcc9-experimental-split-build name: linux-focal-cuda12.1-py3.10-gcc9-experimental-split-build

View File

@ -503,15 +503,15 @@ jobs:
]} ]}
secrets: inherit secrets: inherit
linux-focal-rocm6_1-py3_8-build: linux-focal-rocm6_2-py3_10-build:
# don't run build twice on main # don't run build twice on main
if: github.event_name == 'pull_request' if: github.event_name == 'pull_request'
name: linux-focal-rocm6.1-py3.8 name: linux-focal-rocm6.2-py3.10
uses: ./.github/workflows/_linux-build.yml uses: ./.github/workflows/_linux-build.yml
needs: get-label-type needs: get-label-type
with: with:
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
build-environment: linux-focal-rocm6.1-py3.8 build-environment: linux-focal-rocm6.2-py3.10
docker-image-name: pytorch-linux-focal-rocm-n-py3 docker-image-name: pytorch-linux-focal-rocm-n-py3
sync-tag: rocm-build sync-tag: rocm-build
test-matrix: | test-matrix: |

View File

@ -31,11 +31,11 @@ jobs:
id-token: write id-token: write
contents: read contents: read
linux-focal-rocm6_1-py3_8-build: linux-focal-rocm6_2-py3_10-build:
name: linux-focal-rocm6.1-py3.8 name: linux-focal-rocm6.2-py3.10
uses: ./.github/workflows/_linux-build.yml uses: ./.github/workflows/_linux-build.yml
with: with:
build-environment: linux-focal-rocm6.1-py3.8 build-environment: linux-focal-rocm6.2-py3.10
docker-image-name: pytorch-linux-focal-rocm-n-py3 docker-image-name: pytorch-linux-focal-rocm-n-py3
sync-tag: rocm-build sync-tag: rocm-build
test-matrix: | test-matrix: |
@ -48,16 +48,16 @@ jobs:
{ config: "default", shard: 6, num_shards: 6, runner: "linux.rocm.gpu.2" }, { config: "default", shard: 6, num_shards: 6, runner: "linux.rocm.gpu.2" },
]} ]}
linux-focal-rocm6_1-py3_8-test: linux-focal-rocm6_2-py3_10-test:
permissions: permissions:
id-token: write id-token: write
contents: read contents: read
name: linux-focal-rocm6.1-py3.8 name: linux-focal-rocm6.2-py3.10
uses: ./.github/workflows/_rocm-test.yml uses: ./.github/workflows/_rocm-test.yml
needs: needs:
- linux-focal-rocm6_1-py3_8-build - linux-focal-rocm6_2-py3_10-build
- target-determination - target-determination
with: with:
build-environment: linux-focal-rocm6.1-py3.8 build-environment: linux-focal-rocm6.2-py3.10
docker-image: ${{ needs.linux-focal-rocm6_1-py3_8-build.outputs.docker-image }} docker-image: ${{ needs.linux-focal-rocm6_2-py3_10-build.outputs.docker-image }}
test-matrix: ${{ needs.linux-focal-rocm6_1-py3_8-build.outputs.test-matrix }} test-matrix: ${{ needs.linux-focal-rocm6_2-py3_10-build.outputs.test-matrix }}

View File

@ -127,13 +127,13 @@ jobs:
docker-image: ${{ needs.linux-focal-py3_9-clang10-build.outputs.docker-image }} docker-image: ${{ needs.linux-focal-py3_9-clang10-build.outputs.docker-image }}
test-matrix: ${{ needs.linux-focal-py3_9-clang10-build.outputs.test-matrix }} test-matrix: ${{ needs.linux-focal-py3_9-clang10-build.outputs.test-matrix }}
linux-focal-rocm6_1-py3_8-build: linux-focal-rocm6_2-py3_10-build:
name: linux-focal-rocm6.1-py3.8 name: linux-focal-rocm6.2-py3.10
uses: ./.github/workflows/_linux-build.yml uses: ./.github/workflows/_linux-build.yml
needs: get-label-type needs: get-label-type
with: with:
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
build-environment: linux-focal-rocm6.1-py3.8 build-environment: linux-focal-rocm6.2-py3.10
docker-image-name: pytorch-linux-focal-rocm-n-py3 docker-image-name: pytorch-linux-focal-rocm-n-py3
test-matrix: | test-matrix: |
{ include: [ { include: [
@ -141,19 +141,19 @@ jobs:
{ config: "slow", shard: 2, num_shards: 2, runner: "linux.rocm.gpu" }, { config: "slow", shard: 2, num_shards: 2, runner: "linux.rocm.gpu" },
]} ]}
linux-focal-rocm6_1-py3_8-test: linux-focal-rocm6_2-py3_10-test:
permissions: permissions:
id-token: write id-token: write
contents: read contents: read
name: linux-focal-rocm6.1-py3.8 name: linux-focal-rocm6.2-py3.10
uses: ./.github/workflows/_rocm-test.yml uses: ./.github/workflows/_rocm-test.yml
needs: needs:
- linux-focal-rocm6_1-py3_8-build - linux-focal-rocm6_2-py3_10-build
- target-determination - target-determination
with: with:
build-environment: linux-focal-rocm6.1-py3.8 build-environment: linux-focal-rocm6.2-py3.10
docker-image: ${{ needs.linux-focal-rocm6_1-py3_8-build.outputs.docker-image }} docker-image: ${{ needs.linux-focal-rocm6_2-py3_10-build.outputs.docker-image }}
test-matrix: ${{ needs.linux-focal-rocm6_1-py3_8-build.outputs.test-matrix }} test-matrix: ${{ needs.linux-focal-rocm6_2-py3_10-build.outputs.test-matrix }}
linux-jammy-py3_10-clang15-asan-build: linux-jammy-py3_10-clang15-asan-build:
name: linux-jammy-py3.10-clang15-asan name: linux-jammy-py3.10-clang15-asan

View File

@ -24,7 +24,7 @@ jobs:
- name: Calculate docker image - name: Calculate docker image
id: calculate-docker-image id: calculate-docker-image
uses: pytorch/test-infra/.github/actions/calculate-docker-image@main uses: pytorch/test-infra/.github/actions/calculate-docker-image@release/2.5
with: with:
docker-image-name: pytorch-linux-focal-cuda12.1-cudnn9-py3-gcc9 docker-image-name: pytorch-linux-focal-cuda12.1-cudnn9-py3-gcc9
working-directory: pytorch working-directory: pytorch
@ -39,13 +39,13 @@ jobs:
echo "docker pull ghcr.io/pytorch/ci-image:${tag/:/-}" echo "docker pull ghcr.io/pytorch/ci-image:${tag/:/-}"
- name: Pull docker image - name: Pull docker image
uses: pytorch/test-infra/.github/actions/pull-docker-image@main uses: pytorch/test-infra/.github/actions/pull-docker-image@release/2.5
with: with:
docker-image: ${{ steps.calculate-docker-image.outputs.docker-image }} docker-image: ${{ steps.calculate-docker-image.outputs.docker-image }}
- name: Install nvidia driver, nvidia-docker runtime, set GPU_FLAG - name: Install nvidia driver, nvidia-docker runtime, set GPU_FLAG
id: install-nvidia-driver id: install-nvidia-driver
uses: pytorch/test-infra/.github/actions/setup-nvidia@main uses: pytorch/test-infra/.github/actions/setup-nvidia@release/2.5
- name: Clone CodeLlama - name: Clone CodeLlama
uses: actions/checkout@v3 uses: actions/checkout@v3
@ -136,7 +136,7 @@ jobs:
"s3://target-determinator-assets/indexes/latest/${ZIP_NAME}" "s3://target-determinator-assets/indexes/latest/${ZIP_NAME}"
- name: Teardown Linux - name: Teardown Linux
uses: pytorch/test-infra/.github/actions/teardown-linux@main uses: pytorch/test-infra/.github/actions/teardown-linux@release/2.5
if: always() if: always()
concurrency: concurrency:

View File

@ -25,7 +25,7 @@ jobs:
# checkout because when we run this action we don't *have* a local # checkout because when we run this action we don't *have* a local
# checkout. In other cases you should prefer a local checkout. # checkout. In other cases you should prefer a local checkout.
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
submodules: false submodules: false

View File

@ -223,13 +223,13 @@ jobs:
cuda-version: "12.1" cuda-version: "12.1"
runner: "${{ needs.get-label-type.outputs.label-type }}windows.4xlarge.nonephemeral" runner: "${{ needs.get-label-type.outputs.label-type }}windows.4xlarge.nonephemeral"
linux-focal-rocm6_1-py3_8-build: linux-focal-rocm6_2-py3_10-build:
name: linux-focal-rocm6.1-py3.8 name: linux-focal-rocm6.2-py3.10
uses: ./.github/workflows/_linux-build.yml uses: ./.github/workflows/_linux-build.yml
needs: get-label-type needs: get-label-type
with: with:
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}" runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
build-environment: linux-focal-rocm6.1-py3.8 build-environment: linux-focal-rocm6.2-py3.10
docker-image-name: pytorch-linux-focal-rocm-n-py3 docker-image-name: pytorch-linux-focal-rocm-n-py3
sync-tag: rocm-build sync-tag: rocm-build
test-matrix: | test-matrix: |
@ -240,19 +240,19 @@ jobs:
]} ]}
secrets: inherit secrets: inherit
linux-focal-rocm6_1-py3_8-test: linux-focal-rocm6_2-py3_10-test:
permissions: permissions:
id-token: write id-token: write
contents: read contents: read
name: linux-focal-rocm6.1-py3.8 name: linux-focal-rocm6.2-py3.10
uses: ./.github/workflows/_rocm-test.yml uses: ./.github/workflows/_rocm-test.yml
needs: needs:
- linux-focal-rocm6_1-py3_8-build - linux-focal-rocm6_2-py3_10-build
- target-determination - target-determination
with: with:
build-environment: linux-focal-rocm6.1-py3.8 build-environment: linux-focal-rocm6.2-py3.10
docker-image: ${{ needs.linux-focal-rocm6_1-py3_8-build.outputs.docker-image }} docker-image: ${{ needs.linux-focal-rocm6_2-py3_10-build.outputs.docker-image }}
test-matrix: ${{ needs.linux-focal-rocm6_1-py3_8-build.outputs.test-matrix }} test-matrix: ${{ needs.linux-focal-rocm6_2-py3_10-build.outputs.test-matrix }}
tests-to-include: "test_nn test_torch test_cuda test_ops test_unary_ufuncs test_binary_ufuncs test_autograd inductor/test_torchinductor distributed/test_c10d_common distributed/test_c10d_nccl" tests-to-include: "test_nn test_torch test_cuda test_ops test_unary_ufuncs test_binary_ufuncs test_autograd inductor/test_torchinductor distributed/test_c10d_common distributed/test_c10d_nccl"
linux-focal-cuda12_4-py3_10-gcc9-experimental-split-build: linux-focal-cuda12_4-py3_10-gcc9-experimental-split-build:

View File

@ -16,7 +16,7 @@ jobs:
environment: ${{ (github.event_name == 'schedule') && 'mergebot' || '' }} environment: ${{ (github.event_name == 'schedule') && 'mergebot' || '' }}
steps: steps:
- name: Update viable/strict - name: Update viable/strict
uses: pytorch/test-infra/.github/actions/update-viablestrict@main uses: pytorch/test-infra/.github/actions/update-viablestrict@release/2.5
with: with:
repository: pytorch/pytorch repository: pytorch/pytorch
stable-branch: viable/strict stable-branch: viable/strict

View File

@ -17,7 +17,7 @@ jobs:
contents: read contents: read
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
fetch-depth: 1 fetch-depth: 1
submodules: false submodules: false

View File

@ -44,7 +44,7 @@ jobs:
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }} GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
AWS_ACCESS_KEY_ID: ${{ secrets.AWS_ACCESS_KEY_ID }} AWS_ACCESS_KEY_ID: ${{ secrets.AWS_ACCESS_KEY_ID }}
AWS_SECRET_ACCESS_KEY: ${{ secrets.AWS_SECRET_ACCESS_KEY }} AWS_SECRET_ACCESS_KEY: ${{ secrets.AWS_SECRET_ACCESS_KEY }}
uses: pytorch/test-infra/.github/actions/upload-alerts@main uses: pytorch/test-infra/.github/actions/upload-alerts@release/2.5
with: with:
alerts: '${{ steps.alert_creation_step.outputs.script-output }}' alerts: '${{ steps.alert_creation_step.outputs.script-output }}'
organization: "pytorch" organization: "pytorch"

View File

@ -38,7 +38,7 @@ jobs:
run: echo "${TRIGGERING_WORKFLOW}" run: echo "${TRIGGERING_WORKFLOW}"
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
- name: Configure aws credentials - name: Configure aws credentials
uses: aws-actions/configure-aws-credentials@v3 uses: aws-actions/configure-aws-credentials@v3

View File

@ -31,7 +31,7 @@ jobs:
name: Upload dynamo performance stats for ${{ github.event.workflow_run.id }}, attempt ${{ github.event.workflow_run.run_attempt }} name: Upload dynamo performance stats for ${{ github.event.workflow_run.id }}, attempt ${{ github.event.workflow_run.run_attempt }}
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
submodules: false submodules: false
fetch-depth: 1 fetch-depth: 1

View File

@ -17,7 +17,7 @@ jobs:
environment: upload-stats environment: upload-stats
steps: steps:
- name: Checkout PyTorch - name: Checkout PyTorch
uses: pytorch/pytorch/.github/actions/checkout-pytorch@main uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.5
with: with:
fetch-depth: 1 fetch-depth: 1
submodules: false submodules: false

View File

@ -21,7 +21,7 @@ jobs:
fetch-depth: 0 fetch-depth: 0
- name: update-xla-commit-hash - name: update-xla-commit-hash
continue-on-error: true continue-on-error: true
uses: pytorch/test-infra/.github/actions/update-commit-hash@main uses: pytorch/test-infra/.github/actions/update-commit-hash@release/2.5
with: with:
repo-name: xla repo-name: xla
branch: master branch: master
@ -30,7 +30,7 @@ jobs:
updatebot-token: ${{ secrets.UPDATEBOT_TOKEN }} updatebot-token: ${{ secrets.UPDATEBOT_TOKEN }}
pytorchbot-token: ${{ secrets.GH_PYTORCHBOT_TOKEN }} pytorchbot-token: ${{ secrets.GH_PYTORCHBOT_TOKEN }}
- name: update-triton-commit-hash - name: update-triton-commit-hash
uses: pytorch/test-infra/.github/actions/update-commit-hash@main uses: pytorch/test-infra/.github/actions/update-commit-hash@release/2.5
with: with:
repo-owner: openai repo-owner: openai
repo-name: triton repo-name: triton

View File

@ -27,8 +27,8 @@ Our trunk health (Continuous Integration signals) can be found at [hud.pytorch.o
- [NVIDIA CUDA Support](#nvidia-cuda-support) - [NVIDIA CUDA Support](#nvidia-cuda-support)
- [AMD ROCm Support](#amd-rocm-support) - [AMD ROCm Support](#amd-rocm-support)
- [Intel GPU Support](#intel-gpu-support) - [Intel GPU Support](#intel-gpu-support)
- [Install Dependencies](#install-dependencies)
- [Get the PyTorch Source](#get-the-pytorch-source) - [Get the PyTorch Source](#get-the-pytorch-source)
- [Install Dependencies](#install-dependencies)
- [Install PyTorch](#install-pytorch) - [Install PyTorch](#install-pytorch)
- [Adjust Build Options (Optional)](#adjust-build-options-optional) - [Adjust Build Options (Optional)](#adjust-build-options-optional)
- [Docker Image](#docker-image) - [Docker Image](#docker-image)
@ -161,9 +161,34 @@ They require JetPack 4.2 and above, and [@dusty-nv](https://github.com/dusty-nv)
#### Prerequisites #### Prerequisites
If you are installing from source, you will need: If you are installing from source, you will need:
- Python 3.8 or later (for Linux, Python 3.8.1+ is needed) - Python 3.8 or later (for Linux, Python 3.8.1+ is needed)
- A compiler that fully supports C++17, such as clang or gcc (gcc 9.4.0 or newer is required) - A compiler that fully supports C++17, such as clang or gcc (gcc 9.4.0 or newer is required, on Linux)
- Visual Studio or Visual Studio Build Tool on Windows
We highly recommend installing an [Anaconda](https://www.anaconda.com/download) environment. You will get a high-quality BLAS library (MKL) and you get controlled dependency versions regardless of your Linux distro. \* PyTorch CI uses Visual C++ BuildTools, which come with Visual Studio Enterprise,
Professional, or Community Editions. You can also install the build tools from
https://visualstudio.microsoft.com/visual-cpp-build-tools/. The build tools *do not*
come with Visual Studio Code by default.
\* We highly recommend installing an [Anaconda](https://www.anaconda.com/download) environment. You will get a high-quality BLAS library (MKL) and you get controlled dependency versions regardless of your Linux distro.
An example of environment setup is shown below:
* Linux:
```bash
$ source <CONDA_INSTALL_DIR>/bin/activate
$ conda create -y -n <CONDA_NAME>
$ conda activate <CONDA_NAME>
```
* Windows:
```bash
$ source <CONDA_INSTALL_DIR>\Scripts\activate.bat
$ conda create -y -n <CONDA_NAME>
$ conda activate <CONDA_NAME>
$ call "C:\Program Files\Microsoft Visual Studio\<VERSION>\Community\VC\Auxiliary\Build\vcvarsall.bat" x64
```
##### NVIDIA CUDA Support ##### NVIDIA CUDA Support
If you want to compile with CUDA support, [select a supported version of CUDA from our support matrix](https://pytorch.org/get-started/locally/), then install the following: If you want to compile with CUDA support, [select a supported version of CUDA from our support matrix](https://pytorch.org/get-started/locally/), then install the following:
@ -194,12 +219,23 @@ If you want to compile with Intel GPU support, follow these
If you want to disable Intel GPU support, export the environment variable `USE_XPU=0`. If you want to disable Intel GPU support, export the environment variable `USE_XPU=0`.
Other potentially useful environment variables may be found in `setup.py`. Other potentially useful environment variables may be found in `setup.py`.
#### Get the PyTorch Source
```bash
git clone --recursive https://github.com/pytorch/pytorch
cd pytorch
# if you are updating an existing checkout
git submodule sync
git submodule update --init --recursive
```
#### Install Dependencies #### Install Dependencies
**Common** **Common**
```bash ```bash
conda install cmake ninja conda install cmake ninja
# Run this command on native Windows
conda install rust
# Run this command from the PyTorch directory after cloning the source code using the “Get the PyTorch Source“ section below # Run this command from the PyTorch directory after cloning the source code using the “Get the PyTorch Source“ section below
pip install -r requirements.txt pip install -r requirements.txt
``` ```
@ -235,15 +271,6 @@ pip install mkl-static mkl-include
conda install -c conda-forge libuv=1.39 conda install -c conda-forge libuv=1.39
``` ```
#### Get the PyTorch Source
```bash
git clone --recursive https://github.com/pytorch/pytorch
cd pytorch
# if you are updating an existing checkout
git submodule sync
git submodule update --init --recursive
```
#### Install PyTorch #### Install PyTorch
**On Linux** **On Linux**
@ -284,13 +311,6 @@ python3 setup.py develop
**On Windows** **On Windows**
Choose Correct Visual Studio Version.
PyTorch CI uses Visual C++ BuildTools, which come with Visual Studio Enterprise,
Professional, or Community Editions. You can also install the build tools from
https://visualstudio.microsoft.com/visual-cpp-build-tools/. The build tools *do not*
come with Visual Studio Code by default.
If you want to build legacy python code, please refer to [Building on legacy code and CUDA](https://github.com/pytorch/pytorch/blob/main/CONTRIBUTING.md#building-on-legacy-code-and-cuda) If you want to build legacy python code, please refer to [Building on legacy code and CUDA](https://github.com/pytorch/pytorch/blob/main/CONTRIBUTING.md#building-on-legacy-code-and-cuda)
**CPU-only builds** **CPU-only builds**
@ -298,7 +318,6 @@ If you want to build legacy python code, please refer to [Building on legacy cod
In this mode PyTorch computations will run on your CPU, not your GPU In this mode PyTorch computations will run on your CPU, not your GPU
```cmd ```cmd
conda activate
python setup.py develop python setup.py develop
``` ```

View File

@ -145,6 +145,14 @@ void Context::setSDPUseMath(bool e) {
enabled_mathSDP = e; enabled_mathSDP = e;
} }
bool Context::allowFP16BF16ReductionMathSDP() const {
return allow_fp16_bf16_reduction_mathSDP;
}
void Context::setAllowFP16BF16ReductionMathSDP(bool e) {
allow_fp16_bf16_reduction_mathSDP = e;
}
bool Context::userEnabledCuDNNSDP() const { bool Context::userEnabledCuDNNSDP() const {
return enabled_cudnnSDP; return enabled_cudnnSDP;
} }

View File

@ -234,6 +234,9 @@ class TORCH_API Context {
void setSDPUseCuDNN(bool); void setSDPUseCuDNN(bool);
bool userEnabledCuDNNSDP() const; bool userEnabledCuDNNSDP() const;
void setAllowFP16BF16ReductionMathSDP(bool);
bool allowFP16BF16ReductionMathSDP() const;
void setSDPUseOverrideable(bool); void setSDPUseOverrideable(bool);
bool userEnabledOverrideableSDP() const; bool userEnabledOverrideableSDP() const;
@ -390,6 +393,7 @@ class TORCH_API Context {
bool enabled_mathSDP = true; bool enabled_mathSDP = true;
bool enabled_cudnnSDP = true; bool enabled_cudnnSDP = true;
bool enabled_overrideable = true; bool enabled_overrideable = true;
bool allow_fp16_bf16_reduction_mathSDP = false;
#ifdef USE_ROCM #ifdef USE_ROCM
bool benchmark_cudnn = true; bool benchmark_cudnn = true;
#else #else

View File

@ -128,10 +128,26 @@ inline bool _check_tensors_share_device_and_dtype(
// corresponding tensors in tensor lists have the same sizes and strides. // corresponding tensors in tensor lists have the same sizes and strides.
inline bool _check_tensors_share_sizes_and_strides( inline bool _check_tensors_share_sizes_and_strides(
ArrayRef<TensorList> tensorLists) { ArrayRef<TensorList> tensorLists) {
auto is_diff_stride = [](const IntArrayRef& size,
const IntArrayRef& left_stride,
const IntArrayRef& right_stride) -> bool {
const size_t size_size = size.size();
for (const auto dim : c10::irange(size_size)) {
if (size[dim] == 1)
continue;
if (left_stride[dim] != right_stride[dim]) {
return true;
}
}
return false;
};
for (const auto i : c10::irange(1, tensorLists.size())) { for (const auto i : c10::irange(1, tensorLists.size())) {
for (const auto j : c10::irange(tensorLists[0].size())) { for (const auto j : c10::irange(tensorLists[0].size())) {
if (tensorLists[0][j].sizes() != tensorLists[i][j].sizes() || if (tensorLists[0][j].sizes() != tensorLists[i][j].sizes() ||
tensorLists[0][j].strides() != tensorLists[i][j].strides()) { is_diff_stride(
tensorLists[0][j].sizes(),
tensorLists[0][j].strides(),
tensorLists[i][j].strides())) {
return false; return false;
} }
} }

View File

@ -22,6 +22,7 @@ void run_cudnn_SDP_fprop(
const Tensor& q, const Tensor& q,
const Tensor& k, const Tensor& k,
const Tensor& v, const Tensor& v,
const std::optional<Tensor>& attn_bias,
Tensor& softmaxstats, Tensor& softmaxstats,
Tensor& o, Tensor& o,
Tensor& dropoutseed, Tensor& dropoutseed,
@ -43,6 +44,7 @@ void run_cudnn_SDP_bprop(
const Tensor& q, const Tensor& q,
const Tensor& k, const Tensor& k,
const Tensor& v, const Tensor& v,
const std::optional<Tensor>& attn_bias,
const Tensor& o, const Tensor& o,
const Tensor& dO, const Tensor& dO,
const Tensor& softmaxstats, const Tensor& softmaxstats,
@ -86,9 +88,9 @@ using graph_and_tensors = std::tuple<
std::shared_ptr<fe::graph::Tensor_attributes>, // Q, std::shared_ptr<fe::graph::Tensor_attributes>, // Q,
std::shared_ptr<fe::graph::Tensor_attributes>, // K, std::shared_ptr<fe::graph::Tensor_attributes>, // K,
std::shared_ptr<fe::graph::Tensor_attributes>, // V, std::shared_ptr<fe::graph::Tensor_attributes>, // V,
std::optional<std::shared_ptr<fe::graph::Tensor_attributes>>, // Bias
std::shared_ptr<fe::graph::Tensor_attributes>, // Attn_scale, std::shared_ptr<fe::graph::Tensor_attributes>, // Attn_scale,
// TODO(eqy): additional options // TODO(eqy): additional options
// std::shared_ptr<fe::graph::Tensor_attributes>, // Bias,
// std::shared_ptr<fe::graph::Tensor_attributes>, // SEQ_LEN_Q, // std::shared_ptr<fe::graph::Tensor_attributes>, // SEQ_LEN_Q,
// std::shared_ptr<fe::graph::Tensor_attributes>, // SEQ_LEN_KV, // std::shared_ptr<fe::graph::Tensor_attributes>, // SEQ_LEN_KV,
std::shared_ptr<fe::graph::Tensor_attributes>, // Seed, std::shared_ptr<fe::graph::Tensor_attributes>, // Seed,
@ -104,7 +106,8 @@ using graph_and_tensors_backward = std::tuple<
std::shared_ptr<fe::graph::Tensor_attributes>, // Q, std::shared_ptr<fe::graph::Tensor_attributes>, // Q,
std::shared_ptr<fe::graph::Tensor_attributes>, // K, std::shared_ptr<fe::graph::Tensor_attributes>, // K,
std::shared_ptr<fe::graph::Tensor_attributes>, // V, std::shared_ptr<fe::graph::Tensor_attributes>, // V,
std::shared_ptr<fe::graph::Tensor_attributes>, // Attn_scale std::optional<std::shared_ptr<fe::graph::Tensor_attributes>>, // Bias,
std::shared_ptr<fe::graph::Tensor_attributes>, // Attn_scale,
std::shared_ptr<fe::graph::Tensor_attributes>, // Seed, std::shared_ptr<fe::graph::Tensor_attributes>, // Seed,
std::shared_ptr<fe::graph::Tensor_attributes>, // Offset, std::shared_ptr<fe::graph::Tensor_attributes>, // Offset,
std::shared_ptr<fe::graph::Tensor_attributes>, // O, std::shared_ptr<fe::graph::Tensor_attributes>, // O,
@ -126,6 +129,8 @@ struct MHAParams {
std::array<int, MAX_MHA_DIM> q_stride; std::array<int, MAX_MHA_DIM> q_stride;
std::array<int, MAX_MHA_DIM> k_stride; std::array<int, MAX_MHA_DIM> k_stride;
std::array<int, MAX_MHA_DIM> v_stride; std::array<int, MAX_MHA_DIM> v_stride;
std::array<int, MAX_MHA_DIM> bias_dim;
std::array<int, MAX_MHA_DIM> bias_stride;
int64_t b; int64_t b;
int64_t h; int64_t h;
int64_t s_q; int64_t s_q;
@ -135,6 +140,9 @@ struct MHAParams {
double dropout_probability; double dropout_probability;
bool is_causal; bool is_causal;
bool return_softmaxstats; bool return_softmaxstats;
// might be redundant if we take 0 dim/stride
// as signaling no-bias
bool has_attn_bias;
}; };
void setMHAParams( void setMHAParams(
@ -148,6 +156,7 @@ void setMHAParams(
const Tensor& q, const Tensor& q,
const Tensor& k, const Tensor& k,
const Tensor& v, const Tensor& v,
const std::optional<Tensor>& attn_bias,
double dropout_probability, double dropout_probability,
bool is_causal, bool is_causal,
bool return_softmaxstats) { bool return_softmaxstats) {
@ -166,6 +175,7 @@ void setMHAParams(
params.dropout_probability = dropout_probability; params.dropout_probability = dropout_probability;
params.is_causal = is_causal; params.is_causal = is_causal;
params.return_softmaxstats = return_softmaxstats; params.return_softmaxstats = return_softmaxstats;
params.has_attn_bias = attn_bias.has_value();
TORCH_INTERNAL_ASSERT( TORCH_INTERNAL_ASSERT(
q.sizes().size() == MAX_MHA_DIM, q.sizes().size() == MAX_MHA_DIM,
"Q tensor has unexpected number of dims, please report a bug to PyTorch."); "Q tensor has unexpected number of dims, please report a bug to PyTorch.");
@ -190,6 +200,17 @@ void setMHAParams(
std::copy(k.strides().begin(), k.strides().end(), params.k_stride.begin()); std::copy(k.strides().begin(), k.strides().end(), params.k_stride.begin());
std::copy(v.sizes().begin(), v.sizes().end(), params.v_dim.begin()); std::copy(v.sizes().begin(), v.sizes().end(), params.v_dim.begin());
std::copy(v.strides().begin(), v.strides().end(), params.v_stride.begin()); std::copy(v.strides().begin(), v.strides().end(), params.v_stride.begin());
// uninit is OK as the struct is memset 0'd
if (params.has_attn_bias) {
std::copy(
attn_bias.value().sizes().begin(),
attn_bias.value().sizes().end(),
params.bias_dim.begin());
std::copy(
attn_bias.value().strides().begin(),
attn_bias.value().strides().end(),
params.bias_stride.begin());
}
} }
struct MHACacheKeyWrapper : ParamsWrapper<MHAParams> { struct MHACacheKeyWrapper : ParamsWrapper<MHAParams> {
@ -203,6 +224,7 @@ struct MHACacheKeyWrapper : ParamsWrapper<MHAParams> {
const Tensor& q, const Tensor& q,
const Tensor& k, const Tensor& k,
const Tensor& v, const Tensor& v,
const std::optional<Tensor>& attn_bias,
double dropout_probability, double dropout_probability,
bool is_causal, bool is_causal,
bool return_softmaxstats) { bool return_softmaxstats) {
@ -217,6 +239,7 @@ struct MHACacheKeyWrapper : ParamsWrapper<MHAParams> {
q, q,
k, k,
v, v,
attn_bias,
dropout_probability, dropout_probability,
is_causal, is_causal,
return_softmaxstats); return_softmaxstats);
@ -285,6 +308,7 @@ auto build_graph_and_tensors(
const Tensor& q, const Tensor& q,
const Tensor& k, const Tensor& k,
const Tensor& v, const Tensor& v,
const std::optional<Tensor>& attn_bias,
Tensor& softmaxstats, Tensor& softmaxstats,
Tensor& o, Tensor& o,
Tensor& dropoutseed, Tensor& dropoutseed,
@ -301,36 +325,6 @@ auto build_graph_and_tensors(
mha_graph->set_io_data_type(dtype) mha_graph->set_io_data_type(dtype)
.set_intermediate_data_type(fe::DataType_t::FLOAT) .set_intermediate_data_type(fe::DataType_t::FLOAT)
.set_compute_data_type(fe::DataType_t::FLOAT); .set_compute_data_type(fe::DataType_t::FLOAT);
auto Q = mha_graph->tensor(
fe::graph::Tensor_attributes()
.set_name("Q")
.set_dim(std::vector<int64_t>(
q.sizes().data(), q.sizes().data() + q.sizes().size()))
.set_stride(fixSizeOneDimStrideSDPA(
q.sizes(),
std::vector<int64_t>(
q.strides().data(),
q.strides().data() + q.strides().size()))));
auto K = mha_graph->tensor(
fe::graph::Tensor_attributes()
.set_name("K")
.set_dim(std::vector<int64_t>(
k.sizes().data(), k.sizes().data() + k.sizes().size()))
.set_stride(fixSizeOneDimStrideSDPA(
k.sizes(),
std::vector<int64_t>(
k.strides().data(),
k.strides().data() + k.strides().size()))));
auto V = mha_graph->tensor(
fe::graph::Tensor_attributes()
.set_name("V")
.set_dim(std::vector<int64_t>(
v.sizes().data(), v.sizes().data() + v.sizes().size()))
.set_stride(fixSizeOneDimStrideSDPA(
v.sizes(),
std::vector<int64_t>(
v.strides().data(),
v.strides().data() + v.strides().size()))));
auto attn_scale = auto attn_scale =
mha_graph->tensor(fe::graph::Tensor_attributes() mha_graph->tensor(fe::graph::Tensor_attributes()
.set_name("Attn_scale") .set_name("Attn_scale")
@ -338,11 +332,6 @@ auto build_graph_and_tensors(
.set_stride({1, 1, 1, 1}) .set_stride({1, 1, 1, 1})
.set_is_pass_by_value(true) .set_is_pass_by_value(true)
.set_data_type(fe::DataType_t::FLOAT)); .set_data_type(fe::DataType_t::FLOAT));
// TODO(eqy): support bias in the future in a follow-up PR
// auto bias = mha_graph->tensor(fe::graph::Tensor_attributes()
// .set_name("bias")
// .set_dim({b, 1, s_q, s_kv})
// .set_stride({s_q * s_kv, s_q * s_kv, s_kv, 1}));
auto seed = mha_graph->tensor(fe::graph::Tensor_attributes() auto seed = mha_graph->tensor(fe::graph::Tensor_attributes()
.set_name("Seed") .set_name("Seed")
.set_dim({1, 1, 1, 1}) .set_dim({1, 1, 1, 1})
@ -360,11 +349,30 @@ auto build_graph_and_tensors(
.set_causal_mask(is_causal) .set_causal_mask(is_causal)
.set_attn_scale(attn_scale) .set_attn_scale(attn_scale)
.set_dropout(dropout_probability, seed, offset); .set_dropout(dropout_probability, seed, offset);
// Optional bias in flash attention is only supported 8.9.3 onwards auto Q = mha_graph->tensor(
if (cudnnGetVersion() >= 8904) { fe::graph::Tensor_attributes()
// scaled_dot_product_flash_attention_options.set_alibi_mask(true); .set_name("Q")
.set_dim(q.sizes().vec())
.set_stride(fixSizeOneDimStrideSDPA(q.sizes(), q.strides().vec())));
auto K = mha_graph->tensor(
fe::graph::Tensor_attributes()
.set_name("K")
.set_dim(k.sizes().vec())
.set_stride(fixSizeOneDimStrideSDPA(k.sizes(), k.strides().vec())));
auto V = mha_graph->tensor(
fe::graph::Tensor_attributes()
.set_name("V")
.set_dim(v.sizes().vec())
.set_stride(fixSizeOneDimStrideSDPA(v.sizes(), v.strides().vec())));
std::optional<std::shared_ptr<fe::graph::Tensor_attributes>> bias;
if (attn_bias.has_value()) {
bias =
mha_graph->tensor(fe::graph::Tensor_attributes()
.set_name("bias")
.set_dim(attn_bias.value().sizes().vec())
.set_stride(attn_bias.value().strides().vec()));
scaled_dot_product_flash_attention_options.set_bias(bias.value());
} }
auto seq_q = mha_graph->tensor(fe::graph::Tensor_attributes() auto seq_q = mha_graph->tensor(fe::graph::Tensor_attributes()
.set_name("Seq_q") .set_name("Seq_q")
.set_dim({b, 1, 1, 1}) .set_dim({b, 1, 1, 1})
@ -376,20 +384,9 @@ auto build_graph_and_tensors(
.set_stride({1, 1, 1, 1}) .set_stride({1, 1, 1, 1})
.set_data_type(fe::DataType_t::INT32)); .set_data_type(fe::DataType_t::INT32));
// if (cudnnGetVersion() >= 8903) {
// scaled_dot_product_flash_attention_options.set_bias(bias)
// .set_padding_mask(true)
// .set_seq_len_q(seq_q)
// .set_seq_len_kv(seq_kv);
// }
auto [O, Stats] = auto [O, Stats] =
mha_graph->sdpa(Q, K, V, scaled_dot_product_flash_attention_options); mha_graph->sdpa(Q, K, V, scaled_dot_product_flash_attention_options);
O->set_output(true) O->set_output(true).set_dim(o.sizes().vec()).set_stride(o.strides().vec());
.set_dim(std::vector<int64_t>(
o.sizes().data(), o.sizes().data() + o.sizes().size()))
.set_stride(std::vector<int64_t>(
o.strides().data(), o.strides().data() + o.strides().size()));
if (Stats) { if (Stats) {
Stats->set_output(true).set_data_type(fe::DataType_t::FLOAT); Stats->set_output(true).set_data_type(fe::DataType_t::FLOAT);
@ -407,6 +404,7 @@ auto build_graph_and_tensors(
std::move(Q), std::move(Q),
std::move(K), std::move(K),
std::move(V), std::move(V),
std::move(bias),
std::move(attn_scale), std::move(attn_scale),
std::move(seed), std::move(seed),
std::move(offset), std::move(offset),
@ -427,6 +425,7 @@ auto build_graph_and_tensors_backward(
const Tensor& q, const Tensor& q,
const Tensor& k, const Tensor& k,
const Tensor& v, const Tensor& v,
const std::optional<Tensor>& attn_bias,
const Tensor& o, const Tensor& o,
const Tensor& dO, const Tensor& dO,
const Tensor& softmaxstats, const Tensor& softmaxstats,
@ -447,24 +446,6 @@ auto build_graph_and_tensors_backward(
mha_graph->set_io_data_type(dtype) mha_graph->set_io_data_type(dtype)
.set_intermediate_data_type(fe::DataType_t::FLOAT) .set_intermediate_data_type(fe::DataType_t::FLOAT)
.set_compute_data_type(fe::DataType_t::FLOAT); .set_compute_data_type(fe::DataType_t::FLOAT);
auto Q = mha_graph->tensor(
fe::graph::Tensor_attributes()
.set_name("Q")
.set_dim(std::vector<int64_t>(q.sizes().begin(), q.sizes().end()))
.set_stride(
std::vector<int64_t>(q.strides().begin(), q.strides().end())));
auto K = mha_graph->tensor(
fe::graph::Tensor_attributes()
.set_name("K")
.set_dim(std::vector<int64_t>(k.sizes().begin(), k.sizes().end()))
.set_stride(
std::vector<int64_t>(k.strides().begin(), k.strides().end())));
auto V = mha_graph->tensor(
fe::graph::Tensor_attributes()
.set_name("V")
.set_dim(std::vector<int64_t>(v.sizes().begin(), v.sizes().end()))
.set_stride(
std::vector<int64_t>(v.strides().begin(), v.strides().end())));
auto attn_scale = auto attn_scale =
mha_graph->tensor(fe::graph::Tensor_attributes() mha_graph->tensor(fe::graph::Tensor_attributes()
.set_name("Attn_scale") .set_name("Attn_scale")
@ -472,6 +453,31 @@ auto build_graph_and_tensors_backward(
.set_stride({1, 1, 1, 1}) .set_stride({1, 1, 1, 1})
.set_is_pass_by_value(true) .set_is_pass_by_value(true)
.set_data_type(fe::DataType_t::FLOAT)); .set_data_type(fe::DataType_t::FLOAT));
auto sdpa_backward_options = fe::graph::SDPA_backward_attributes()
.set_name("CUDNN_SDPA_BACKWARD")
.set_causal_mask(is_causal)
.set_attn_scale(attn_scale);
auto Q = mha_graph->tensor(fe::graph::Tensor_attributes()
.set_name("Q")
.set_dim(q.sizes().vec())
.set_stride(q.strides().vec()));
auto K = mha_graph->tensor(fe::graph::Tensor_attributes()
.set_name("K")
.set_dim(k.sizes().vec())
.set_stride(k.strides().vec()));
auto V = mha_graph->tensor(fe::graph::Tensor_attributes()
.set_name("V")
.set_dim(v.sizes().vec())
.set_stride(v.strides().vec()));
std::optional<std::shared_ptr<fe::graph::Tensor_attributes>> bias;
if (attn_bias.has_value()) {
bias =
mha_graph->tensor(fe::graph::Tensor_attributes()
.set_name("bias")
.set_dim(attn_bias.value().sizes().vec())
.set_stride(attn_bias.value().strides().vec()));
sdpa_backward_options.set_bias(bias.value());
}
auto Seed = mha_graph->tensor(fe::graph::Tensor_attributes() auto Seed = mha_graph->tensor(fe::graph::Tensor_attributes()
.set_name("Seed") .set_name("Seed")
.set_dim({1, 1, 1, 1}) .set_dim({1, 1, 1, 1})
@ -482,47 +488,27 @@ auto build_graph_and_tensors_backward(
.set_dim({1, 1, 1, 1}) .set_dim({1, 1, 1, 1})
.set_stride({1, 1, 1, 1}) .set_stride({1, 1, 1, 1})
.set_data_type(fe::DataType_t::INT32)); .set_data_type(fe::DataType_t::INT32));
auto O = mha_graph->tensor( auto O = mha_graph->tensor(fe::graph::Tensor_attributes()
fe::graph::Tensor_attributes() .set_name("O")
.set_name("O") .set_dim(o.sizes().vec())
.set_dim(std::vector<int64_t>(o.sizes().begin(), o.sizes().end())) .set_stride(o.strides().vec()));
.set_stride( auto STATS = mha_graph->tensor(fe::graph::Tensor_attributes()
std::vector<int64_t>(o.strides().begin(), o.strides().end()))); .set_name("Stats")
auto STATS = mha_graph->tensor( .set_dim(softmaxstats.sizes().vec())
fe::graph::Tensor_attributes() .set_stride(softmaxstats.strides().vec())
.set_name("Stats") .set_data_type(fe::DataType_t::FLOAT));
.set_dim(std::vector<int64_t>( auto DO = mha_graph->tensor(fe::graph::Tensor_attributes()
softmaxstats.sizes().begin(), softmaxstats.sizes().end())) .set_name("DO")
.set_stride(std::vector<int64_t>( .set_dim(dO.sizes().vec())
softmaxstats.strides().begin(), softmaxstats.strides().end())) .set_stride(dO.strides().vec()));
.set_data_type(fe::DataType_t::FLOAT));
auto DO = mha_graph->tensor(
fe::graph::Tensor_attributes()
.set_name("DO")
.set_dim(std::vector<int64_t>(dO.sizes().begin(), dO.sizes().end()))
.set_stride(
std::vector<int64_t>(dO.strides().begin(), dO.strides().end())));
auto sdpa_backward_options = fe::graph::SDPA_backward_attributes()
.set_name("CUDNN_SDPA_BACKWARD")
.set_causal_mask(is_causal)
.set_attn_scale(attn_scale);
if (dropout_probability != 0.0f) { if (dropout_probability != 0.0f) {
sdpa_backward_options.set_dropout(dropout_probability, Seed, Offset); sdpa_backward_options.set_dropout(dropout_probability, Seed, Offset);
} }
auto [DQ, DK, DV] = auto [DQ, DK, DV] =
mha_graph->sdpa_backward(Q, K, V, O, DO, STATS, sdpa_backward_options); mha_graph->sdpa_backward(Q, K, V, O, DO, STATS, sdpa_backward_options);
DQ->set_output(true) DQ->set_output(true).set_dim(dQ.sizes().vec()).set_stride(dQ.strides().vec());
.set_dim(std::vector<int64_t>(dQ.sizes().begin(), dQ.sizes().end())) DK->set_output(true).set_dim(dK.sizes().vec()).set_stride(dK.strides().vec());
.set_stride( DV->set_output(true).set_dim(dV.sizes().vec()).set_stride(dV.strides().vec());
std::vector<int64_t>(dQ.strides().begin(), dQ.strides().end()));
DK->set_output(true)
.set_dim(std::vector<int64_t>(dK.sizes().begin(), dK.sizes().end()))
.set_stride(
std::vector<int64_t>(dK.strides().begin(), dK.strides().end()));
DV->set_output(true)
.set_dim(std::vector<int64_t>(dV.sizes().begin(), dV.sizes().end()))
.set_stride(
std::vector<int64_t>(dV.strides().begin(), dV.strides().end()));
AT_CUDNN_FRONTEND_CHECK(mha_graph->validate()); AT_CUDNN_FRONTEND_CHECK(mha_graph->validate());
AT_CUDNN_FRONTEND_CHECK(mha_graph->build_operation_graph(handle)); AT_CUDNN_FRONTEND_CHECK(mha_graph->build_operation_graph(handle));
AT_CUDNN_FRONTEND_CHECK( AT_CUDNN_FRONTEND_CHECK(
@ -534,6 +520,7 @@ auto build_graph_and_tensors_backward(
std::move(Q), std::move(Q),
std::move(K), std::move(K),
std::move(V), std::move(V),
std::move(bias),
std::move(attn_scale), std::move(attn_scale),
std::move(Seed), std::move(Seed),
std::move(Offset), std::move(Offset),
@ -559,6 +546,7 @@ void run_cudnn_SDP_fprop(
const Tensor& q, const Tensor& q,
const Tensor& k, const Tensor& k,
const Tensor& v, const Tensor& v,
const std::optional<Tensor>& attn_bias,
Tensor& softmaxstats, Tensor& softmaxstats,
Tensor& o, Tensor& o,
Tensor& dropoutseed, Tensor& dropoutseed,
@ -573,6 +561,11 @@ void run_cudnn_SDP_fprop(
softmaxstats = at::empty({b, h, s_q}, q.options().dtype(kFloat)); softmaxstats = at::empty({b, h, s_q}, q.options().dtype(kFloat));
} }
// do nothing if we got 0-element tensors
if (!q.numel() || !k.numel() || !v.numel()) {
return;
}
auto key = MHACacheKeyWrapper( auto key = MHACacheKeyWrapper(
b, b,
h, h,
@ -583,6 +576,7 @@ void run_cudnn_SDP_fprop(
q, q,
k, k,
v, v,
attn_bias,
dropout_probability, dropout_probability,
is_causal, is_causal,
return_softmaxstats); return_softmaxstats);
@ -605,13 +599,14 @@ void run_cudnn_SDP_fprop(
q, q,
k, k,
v, v,
attn_bias,
softmaxstats, softmaxstats,
o, o,
dropoutseed, dropoutseed,
dropoutoffset, dropoutoffset,
handle); handle);
} }
auto [mha_graph, Q, K, V, attn_scale, seed, offset, O, Stats] = auto [mha_graph, Q, K, V, bias, attn_scale, seed, offset, O, Stats] =
graph_and_tensors_values; graph_and_tensors_values;
std::unordered_map<std::shared_ptr<fe::graph::Tensor_attributes>, void*> std::unordered_map<std::shared_ptr<fe::graph::Tensor_attributes>, void*>
variant_pack = { variant_pack = {
@ -619,13 +614,15 @@ void run_cudnn_SDP_fprop(
{K, k.data_ptr()}, {K, k.data_ptr()},
{V, v.data_ptr()}, {V, v.data_ptr()},
{attn_scale, &scaling_factor}, {attn_scale, &scaling_factor},
//{bias, bias.data_ptr()},
{seed, dropoutseed.data_ptr()}, {seed, dropoutseed.data_ptr()},
{offset, dropoutoffset.data_ptr()}, {offset, dropoutoffset.data_ptr()},
{O, o.data_ptr()}}; {O, o.data_ptr()}};
if (return_softmaxstats) { if (return_softmaxstats) {
variant_pack[Stats] = softmaxstats.data_ptr(); variant_pack[Stats] = softmaxstats.data_ptr();
} }
if (attn_bias.has_value()) {
variant_pack[bias.value()] = attn_bias.value().data_ptr();
}
auto workspace_size = mha_graph->get_workspace_size(); auto workspace_size = mha_graph->get_workspace_size();
auto workspace_ptr = auto workspace_ptr =
c10::cuda::CUDACachingAllocator::get()->allocate(workspace_size); c10::cuda::CUDACachingAllocator::get()->allocate(workspace_size);
@ -647,6 +644,7 @@ void run_cudnn_SDP_bprop(
const Tensor& q, const Tensor& q,
const Tensor& k, const Tensor& k,
const Tensor& v, const Tensor& v,
const std::optional<Tensor>& attn_bias,
const Tensor& o, const Tensor& o,
const Tensor& dO, const Tensor& dO,
const Tensor& softmaxstats, const Tensor& softmaxstats,
@ -655,6 +653,12 @@ void run_cudnn_SDP_bprop(
Tensor& dV, Tensor& dV,
const Tensor& dropoutseed, const Tensor& dropoutseed,
const Tensor& dropoutoffset) { const Tensor& dropoutoffset) {
// do nothing if we got 0-element tensors
if (!q.numel() || !k.numel() || !v.numel() || !o.numel() || !dO.numel() ||
!softmaxstats.numel()) {
return;
}
Tensor dO_ = dO; Tensor dO_ = dO;
if (!dO.strides()[dO.strides().size() - 1]) { if (!dO.strides()[dO.strides().size() - 1]) {
TORCH_WARN( TORCH_WARN(
@ -694,6 +698,7 @@ void run_cudnn_SDP_bprop(
q, q,
k, k,
v, v,
attn_bias,
dropout_probability, dropout_probability,
is_causal, is_causal,
true); true);
@ -715,6 +720,7 @@ void run_cudnn_SDP_bprop(
q, q,
k, k,
v, v,
attn_bias,
o, o,
dO_, dO_,
softmaxstats, softmaxstats,
@ -726,8 +732,20 @@ void run_cudnn_SDP_bprop(
handle); handle);
} }
auto auto
[mha_graph, Q, K, V, attn_scale, Seed, Offset, O, Do, Stats, Dq, Dk, Dv] = [mha_graph,
graph_and_tensors_backward_values; Q,
K,
V,
bias,
attn_scale,
Seed,
Offset,
O,
Do,
Stats,
Dq,
Dk,
Dv] = graph_and_tensors_backward_values;
std::unordered_map<std::shared_ptr<fe::graph::Tensor_attributes>, void*> std::unordered_map<std::shared_ptr<fe::graph::Tensor_attributes>, void*>
variant_pack = {// inputs variant_pack = {// inputs
{Q, q.data_ptr()}, {Q, q.data_ptr()},
@ -746,6 +764,9 @@ void run_cudnn_SDP_bprop(
variant_pack[Seed] = dropoutseed.data_ptr(); variant_pack[Seed] = dropoutseed.data_ptr();
variant_pack[Offset] = dropoutoffset.data_ptr(); variant_pack[Offset] = dropoutoffset.data_ptr();
} }
if (attn_bias.has_value()) {
variant_pack[bias.value()] = attn_bias.value().data_ptr();
}
auto workspace_size = mha_graph->get_workspace_size(); auto workspace_size = mha_graph->get_workspace_size();
auto workspace_ptr = auto workspace_ptr =
c10::cuda::CUDACachingAllocator::get()->allocate(workspace_size); c10::cuda::CUDACachingAllocator::get()->allocate(workspace_size);

View File

@ -18,6 +18,7 @@ void run_cudnn_SDP_fprop(
const Tensor& q, const Tensor& q,
const Tensor& k, const Tensor& k,
const Tensor& v, const Tensor& v,
const std::optional<Tensor>& attn_bias,
Tensor& softmaxstats, Tensor& softmaxstats,
Tensor& o, Tensor& o,
Tensor& dropoutseed, Tensor& dropoutseed,
@ -36,6 +37,7 @@ void run_cudnn_SDP_bprop(
const Tensor& q, const Tensor& q,
const Tensor& k, const Tensor& k,
const Tensor& v, const Tensor& v,
const std::optional<Tensor>& attn_bias,
const Tensor& o, const Tensor& o,
const Tensor& dO, const Tensor& dO,
const Tensor& softmaxstats, const Tensor& softmaxstats,

View File

@ -536,6 +536,24 @@ std::optional<Tensor> convert_boolean_attn_mask(const std::optional<Tensor>& att
// Otherwise, attn_mask represents an additive attention tensor // Otherwise, attn_mask represents an additive attention tensor
return attn_mask; return attn_mask;
} }
// alternate version to workaround -inf issue with cuDNN
// TODO(eqy): delete this when cuDNN -inf issue is resolved
std::optional<Tensor> convert_boolean_attn_mask_cudnn(const std::optional<Tensor>& attn_mask, caffe2::TypeMeta dtype) {
// Pass through
if(!attn_mask.has_value()){
return std::nullopt;
}
// Convert boolean mask to additive mask; need to invert mask to indicate what
// to mask *out*.
if (attn_mask->dtype() == at::kBool) {
// TODO Use the max type of the input and output
return at::where(attn_mask->logical_not(), -65504.0, at::scalar_tensor(0.0, at::TensorOptions().dtype(dtype)));
}
// Otherwise, attn_mask represents an additive attention tensor
return attn_mask;
}
// Memory Efficient Attention requires a padded attn mask bias // Memory Efficient Attention requires a padded attn mask bias
// This function pads the attn_mask bias to be a multiple of 16 // This function pads the attn_mask bias to be a multiple of 16
// Then slices the padded bias to the original size // Then slices the padded bias to the original size
@ -698,15 +716,16 @@ Tensor scaled_dot_product_attention(
query_, key, value, attn_mask_, dropout_p, is_causal, scale, enable_gqa); query_, key, value, attn_mask_, dropout_p, is_causal, scale, enable_gqa);
} }
sdp::SDPBackend backend = static_cast<sdp::SDPBackend>(choice_int); sdp::SDPBackend backend = static_cast<sdp::SDPBackend>(choice_int);
std::optional<Tensor> attn_mask = convert_boolean_attn_mask(attn_mask_, query_.dtype());
switch (backend) { switch (backend) {
case sdp::SDPBackend::cudnn_attention: { case sdp::SDPBackend::cudnn_attention: {
std::optional<Tensor> attn_mask = convert_boolean_attn_mask_cudnn(attn_mask_, query_.dtype());
bool compute_logsumexp = should_compute_logsumexp(query_, key, value); bool compute_logsumexp = should_compute_logsumexp(query_, key, value);
auto out_lse_softmax = at::_scaled_dot_product_cudnn_attention( auto out_lse_softmax = at::_scaled_dot_product_cudnn_attention(
query_, key, value, attn_mask_, compute_logsumexp, dropout_p, is_causal, false /*return_debug_mask*/, scale); query_, key, value, attn_mask, compute_logsumexp, dropout_p, is_causal, false /*return_debug_mask*/, scale);
return std::get<0>(out_lse_softmax); return std::get<0>(out_lse_softmax);
} }
case sdp::SDPBackend::flash_attention: { case sdp::SDPBackend::flash_attention: {
std::optional<Tensor> attn_mask = convert_boolean_attn_mask(attn_mask_, query_.dtype());
if(query_.device().type() == DeviceType::CUDA){ if(query_.device().type() == DeviceType::CUDA){
c10::SymInt og_size = query_.sym_size(-1); c10::SymInt og_size = query_.sym_size(-1);
Tensor query_padded = pad_last_dim<8, false>(query_); Tensor query_padded = pad_last_dim<8, false>(query_);
@ -723,6 +742,7 @@ Tensor scaled_dot_product_attention(
query_, key, value, dropout_p, is_causal, attn_mask, scale)); query_, key, value, dropout_p, is_causal, attn_mask, scale));
} }
case sdp::SDPBackend::efficient_attention: { case sdp::SDPBackend::efficient_attention: {
std::optional<Tensor> attn_mask = convert_boolean_attn_mask(attn_mask_, query_.dtype());
bool compute_logsumexp = should_compute_logsumexp(query_, key, value); bool compute_logsumexp = should_compute_logsumexp(query_, key, value);
if (attn_mask.has_value()) { if (attn_mask.has_value()) {
attn_mask.value() = preprocess_mask(attn_mask.value(), query_, key, value);; attn_mask.value() = preprocess_mask(attn_mask.value(), query_, key, value);;
@ -732,11 +752,13 @@ Tensor scaled_dot_product_attention(
return std::get<0>(out_and_lse); return std::get<0>(out_and_lse);
} }
case sdp::SDPBackend::overrideable: { case sdp::SDPBackend::overrideable: {
std::optional<Tensor> attn_mask = convert_boolean_attn_mask(attn_mask_, query_.dtype());
auto out_lse_softmax = at::_scaled_dot_product_fused_attention_overrideable( auto out_lse_softmax = at::_scaled_dot_product_fused_attention_overrideable(
query_, key, value, attn_mask, dropout_p, is_causal, false /*return_debug_mask*/, scale); query_, key, value, attn_mask, dropout_p, is_causal, false /*return_debug_mask*/, scale);
return std::get<0>(out_lse_softmax); return std::get<0>(out_lse_softmax);
} }
case sdp::SDPBackend::math: case sdp::SDPBackend::math: {
std::optional<Tensor> attn_mask = convert_boolean_attn_mask(attn_mask_, query_.dtype());
if ((!GradMode::is_enabled() || (!query_.requires_grad() && !key.requires_grad() && !value.requires_grad())) if ((!GradMode::is_enabled() || (!query_.requires_grad() && !key.requires_grad() && !value.requires_grad()))
&& query_.device().type() == DeviceType::MPS && dropout_p == 0.0 && query_.device().type() == DeviceType::MPS && dropout_p == 0.0
&& query_.is_contiguous() && key.is_contiguous() && value.is_contiguous() && query_.is_contiguous() && key.is_contiguous() && value.is_contiguous()
@ -761,6 +783,7 @@ Tensor scaled_dot_product_attention(
std::nullopt, /*dropout_mask*/ std::nullopt, /*dropout_mask*/
scale, scale,
enable_gqa)); enable_gqa));
}
default: default:
TORCH_CHECK( TORCH_CHECK(
false, false,
@ -780,22 +803,26 @@ std::tuple<Tensor, Tensor> _scaled_dot_product_attention_math(
value.is_contiguous(), value.is_contiguous(),
"scaled_dot_product_attention: If inputs are nested tensors they must be contiguous"); "scaled_dot_product_attention: If inputs are nested tensors they must be contiguous");
} }
auto& ctx = at::globalContext();
auto origin_dtype = query_.scalar_type(); auto origin_dtype = query_.scalar_type();
// Keep query, key, value in high precision for accuracy // Keep query, key, value in high precision for accuracy
// NestedTensor reports issues for backward with autograd so disabled: must be // NestedTensor reports issues for backward with autograd so disabled: must be
// contiguous to get buffer. // contiguous to get buffer.
auto query_acc = (query_.scalar_type() == at::kHalf || auto query_acc = !ctx.allowFP16BF16ReductionMathSDP() &&
query_.scalar_type() == at::kBFloat16) && (query_.scalar_type() == at::kHalf ||
query_.scalar_type() == at::kBFloat16) &&
!query_.is_nested() !query_.is_nested()
? query_.to(at::kFloat) ? query_.to(at::kFloat)
: query_; : query_;
auto key_acc = auto key_acc = !ctx.allowFP16BF16ReductionMathSDP() &&
(key.scalar_type() == at::kHalf || key.scalar_type() == at::kBFloat16) && (key.scalar_type() == at::kHalf ||
key.scalar_type() == at::kBFloat16) &&
!key.is_nested() !key.is_nested()
? key.to(at::kFloat) ? key.to(at::kFloat)
: key; : key;
auto value_acc = (value.scalar_type() == at::kHalf || auto value_acc = !ctx.allowFP16BF16ReductionMathSDP() &&
value.scalar_type() == at::kBFloat16) && (value.scalar_type() == at::kHalf ||
value.scalar_type() == at::kBFloat16) &&
!value.is_nested() !value.is_nested()
? value.to(at::kFloat) ? value.to(at::kFloat)
: value; : value;

View File

@ -774,6 +774,18 @@ std::tuple<Tensor, Tensor, Tensor, Tensor, c10::SymInt, c10::SymInt, Tensor, Ten
TORCH_CHECK( TORCH_CHECK(
max_seqlen_batch_k == max_seqlen_batch_v, max_seqlen_batch_k == max_seqlen_batch_v,
"Key and Value must have the same sequence length"); "Key and Value must have the same sequence length");
auto attn_bias_ = attn_bias;
if (attn_bias_.has_value()) {
const auto bias_dim = attn_bias_.value().dim();
if (bias_dim == 2) {
attn_bias_ = attn_bias_.value().expand({batch_size, 1, max_seqlen_batch_q, max_seqlen_batch_k});
} else if (bias_dim == 3) {
attn_bias_ = attn_bias_.value().expand({batch_size, 1, max_seqlen_batch_q, max_seqlen_batch_k});
} else {
attn_bias_ = attn_bias_.value().expand({batch_size, attn_bias_.value().size(1), max_seqlen_batch_q, max_seqlen_batch_k});
TORCH_CHECK(bias_dim == 4, "cuDNN SDPA expects either a 2D, 3D, or 4D attn_bias but got ", attn_bias_.value().dim(), "D");
}
}
Tensor attention, log_sumexp; Tensor attention, log_sumexp;
@ -818,13 +830,14 @@ std::tuple<Tensor, Tensor, Tensor, Tensor, c10::SymInt, c10::SymInt, Tensor, Ten
query/* Tensor q*/, query/* Tensor q*/,
key/* Tensor k*/, key/* Tensor k*/,
value/* Tensor v*/, value/* Tensor v*/,
attn_bias_ /* std::optional<Tensor> */,
log_sumexp/*Tensor softmaxstats*/, log_sumexp/*Tensor softmaxstats*/,
attention/*Tensor o*/, attention/*Tensor o*/,
cudnn_seed/*Tensor dropoutseed*/, cudnn_seed/*Tensor dropoutseed*/,
cudnn_offset/*Tensor dropoutoffset*/); cudnn_offset/*Tensor dropoutoffset*/);
// TODO(eqy): support debug_attn_mask // TODO(eqy): support debug_attn_mask
return std::make_tuple(attention, log_sumexp, Tensor(), Tensor(), max_seqlen_batch_q, max_seqlen_batch_k, cudnn_seed, cudnn_offset, Tensor()); return std::make_tuple(std::move(attention), std::move(log_sumexp), Tensor(), Tensor(), max_seqlen_batch_q, max_seqlen_batch_k, std::move(cudnn_seed), std::move(cudnn_offset), Tensor());
} }
std::tuple<Tensor, Tensor, Tensor, Tensor> _scaled_dot_product_efficient_attention_cuda( std::tuple<Tensor, Tensor, Tensor, Tensor> _scaled_dot_product_efficient_attention_cuda(
@ -1102,10 +1115,13 @@ std::tuple<Tensor, Tensor, Tensor, Tensor, c10::SymInt, c10::SymInt> _efficient_
offset_t = at::empty({}, at::dtype(at::kLong).device(device)); offset_t = at::empty({}, at::dtype(at::kLong).device(device));
} else { } else {
auto [seed, offset] = at::cuda::philox::unpack(philox_state); auto [seed, offset] = at::cuda::philox::unpack(philox_state);
seed_t = at::scalar_tensor( #ifdef USE_ROCM
at::Scalar(static_cast<int64_t>(seed)), at::dtype(at::kLong)); const auto options = at::dtype(at::kLong).device(at::kCUDA);
offset_t = at::scalar_tensor( #else
at::Scalar(static_cast<int64_t>(offset)), at::dtype(at::kLong)); const auto options = at::dtype(at::kLong);
#endif
seed_t = at::scalar_tensor(at::Scalar(static_cast<int64_t>(seed)), options);
offset_t = at::scalar_tensor(at::Scalar(static_cast<int64_t>(offset)), options);
} }
} else { } else {
// Not using dropout // Not using dropout
@ -1118,7 +1134,8 @@ std::tuple<Tensor, Tensor, Tensor, Tensor, c10::SymInt, c10::SymInt> _efficient_
auto ret = aotriton::v2::flash::check_gpu(stream); auto ret = aotriton::v2::flash::check_gpu(stream);
if (hipSuccess != ret) { if (hipSuccess != ret) {
TORCH_CHECK(false, TORCH_CHECK(false,
"[AOTriton] Accelerated SDPA only supports MI200/MI300X GPUs (gfx90a:sramecc+:xnack- or gfx94a:sramecc+:xnack-)") "[AOTriton] Accelerated SDPA only supports MI200/MI300X/Navi31 GPUs"
" (gfx90a:sramecc+:xnack-/gfx942:sramecc+:xnack-/gfx1100)")
} }
// AOTriton may accept aligned on logsumexp tensor in the future for better // AOTriton may accept aligned on logsumexp tensor in the future for better
@ -1147,8 +1164,16 @@ std::tuple<Tensor, Tensor, Tensor, Tensor, c10::SymInt, c10::SymInt> _efficient_
using aotriton::v2::flash::attn_fwd; using aotriton::v2::flash::attn_fwd;
using sdp::aotriton_adapter::mk_aotensor; using sdp::aotriton_adapter::mk_aotensor;
using sdp::aotriton_adapter::mk_aoscalartensor;
using sdp::aotriton_adapter::mk_philoxtensor;
aotriton::TensorView<4> empty_t4(0, {0, 0, 0, 0}, {0, 0, 0, 0}, aotriton::DType::kFloat16); aotriton::TensorView<4> empty_t4(0, {0, 0, 0, 0}, {0, 0, 0, 0}, aotriton::DType::kFloat16);
at::Tensor softmax_fa_t = at::empty({ 0, 0, 0, 0 }, query.options()); at::Tensor softmax_fa_t = at::empty({ 0, 0, 0, 0 }, query.options());
const bool use_philox_state = in_capture_stream;
auto seed = use_philox_state ? mk_philoxtensor(philox_state.seed_.ptr) : mk_aoscalartensor(seed_t);
auto offset1 = use_philox_state ? mk_philoxtensor(philox_state.offset_.ptr) : mk_aoscalartensor(offset_t);
auto offset2 = use_philox_state ? philox_state.offset_intragraph_ : 0;
auto seed_output = use_philox_state ? mk_philoxtensor(seed_t.data_ptr<int64_t>()) : mk_philoxtensor(nullptr);
auto offset_output = use_philox_state ? mk_philoxtensor(offset_t.data_ptr<int64_t>()) : mk_philoxtensor(nullptr);
hipError_t err; // TODO: Error handling hipError_t err; // TODO: Error handling
err = attn_fwd(mk_aotensor(q_t, "q"), err = attn_fwd(mk_aotensor(q_t, "q"),
mk_aotensor(k_t, "k"), mk_aotensor(k_t, "k"),
@ -1158,8 +1183,11 @@ std::tuple<Tensor, Tensor, Tensor, Tensor, c10::SymInt, c10::SymInt> _efficient_
mk_aotensor<2>(softmax_lse, "M"), mk_aotensor<2>(softmax_lse, "M"),
mk_aotensor(output_t, "Out"), mk_aotensor(output_t, "Out"),
dropout_p, dropout_p,
use_dropout ? *seed_t.data_ptr<int64_t>() : 0, seed,
use_dropout ? *offset_t.data_ptr<int64_t>() : 0, offset1,
offset2,
seed_output,
offset_output,
mk_aotensor(softmax_fa_t, "encoded_softmax"), mk_aotensor(softmax_fa_t, "encoded_softmax"),
is_causal, is_causal,
stream); stream);

View File

@ -195,6 +195,27 @@ std::tuple<Tensor, Tensor, Tensor> _scaled_dot_product_cudnn_attention_backward_
const int64_t num_heads = query.size(1); const int64_t num_heads = query.size(1);
const int64_t head_dim_qk = query.size(3); const int64_t head_dim_qk = query.size(3);
const int64_t head_dim_v = value.size(3); const int64_t head_dim_v = value.size(3);
const int64_t max_seqlen_batch_q = query.size(2);
const int64_t max_seqlen_batch_k = key.size(2);
// This is needed because SaveVariable automatically converts
// std::optional to undefined tensor
std::optional<Tensor> attn_bias_;
if (attn_bias.defined()) {
attn_bias_ = attn_bias;
}
if (attn_bias_.has_value()) {
const auto bias_dim = attn_bias_.value().dim();
if (bias_dim == 2) {
attn_bias_ = attn_bias_.value().expand({batch_size, 1, max_seqlen_batch_q, max_seqlen_batch_k});
} else if (bias_dim == 3) {
attn_bias_ = attn_bias_.value().expand({batch_size, 1, max_seqlen_batch_q, max_seqlen_batch_k});
} else {
attn_bias_ = attn_bias_.value().expand({batch_size, attn_bias_.value().size(1), max_seqlen_batch_q, max_seqlen_batch_k});
TORCH_CHECK(bias_dim == 4, "cuDNN SDPA expects either a 2D, 3D, or 4D attn_bias but got ", attn_bias_.value().dim(), "D");
}
}
const auto softmax_scale = sdp::calculate_scale(query, scale).as_float_unchecked(); const auto softmax_scale = sdp::calculate_scale(query, scale).as_float_unchecked();
auto dq = at::empty_like(query); auto dq = at::empty_like(query);
auto dk = at::empty_like(key); auto dk = at::empty_like(key);
@ -211,6 +232,7 @@ std::tuple<Tensor, Tensor, Tensor> _scaled_dot_product_cudnn_attention_backward_
query /*const Tensor& q*/, query /*const Tensor& q*/,
key /*const Tensor& k*/, key /*const Tensor& k*/,
value /*const Tensor& v*/, value /*const Tensor& v*/,
attn_bias_ /*const std::optional<Tensor>& attn_bias*/,
out /*const Tensor& o*/, out /*const Tensor& o*/,
grad_out/*const Tensor& dO*/, grad_out/*const Tensor& dO*/,
logsumexp.unsqueeze(-1)/*const Tensor& softmaxstats*/, logsumexp.unsqueeze(-1)/*const Tensor& softmaxstats*/,
@ -219,7 +241,7 @@ std::tuple<Tensor, Tensor, Tensor> _scaled_dot_product_cudnn_attention_backward_
dv/*Tensor& dV*/, dv/*Tensor& dV*/,
philox_seed/*Tensor& dropoutseed*/, philox_seed/*Tensor& dropoutseed*/,
philox_offset/*Tensor& dropoutoffset*/); philox_offset/*Tensor& dropoutoffset*/);
return std::make_tuple(dq, dk, dv); return std::make_tuple(std::move(dq), std::move(dk), std::move(dv));
} }
std::tuple<at::Tensor, at::Tensor, at::Tensor, at::Tensor> std::tuple<at::Tensor, at::Tensor, at::Tensor, at::Tensor>
@ -394,7 +416,8 @@ _efficient_attention_backward(
auto ret = aotriton::v2::flash::check_gpu(stream); auto ret = aotriton::v2::flash::check_gpu(stream);
if (hipSuccess != ret) { if (hipSuccess != ret) {
TORCH_CHECK(false, TORCH_CHECK(false,
"[AOTriton] Accelerated SDPA only supports MI200/MI300X GPUs (gfx90a:sramecc+:xnack- or gfx942:sramecc+:xnack-)") "[AOTriton] Accelerated SDPA only supports MI200/MI300X/Navi31 GPUs"
" (gfx90a:sramecc+:xnack-/gfx942:sramecc+:xnack-/gfx1100)")
} }
const auto softmax_scale = sdp::calculate_scale(query, scale).as_float_unchecked(); const auto softmax_scale = sdp::calculate_scale(query, scale).as_float_unchecked();
bool is_causal; bool is_causal;
@ -419,6 +442,7 @@ _efficient_attention_backward(
hipError_t err; hipError_t err;
using aotriton::v2::flash::attn_bwd; using aotriton::v2::flash::attn_bwd;
using sdp::aotriton_adapter::mk_aotensor; using sdp::aotriton_adapter::mk_aotensor;
using sdp::aotriton_adapter::mk_aoscalartensor;
using sdp::aotriton_adapter::cast_dtype; using sdp::aotriton_adapter::cast_dtype;
aotriton::TensorView<4> empty_t4(0, {0, 0, 0, 0}, {0, 0, 0, 0}, cast_dtype(query.dtype())); aotriton::TensorView<4> empty_t4(0, {0, 0, 0, 0}, {0, 0, 0, 0}, cast_dtype(query.dtype()));
err = attn_bwd(mk_aotensor(q_t, "q"), err = attn_bwd(mk_aotensor(q_t, "q"),
@ -435,8 +459,9 @@ _efficient_attention_backward(
mk_aotensor<2>(softmax_lse, "L"), mk_aotensor<2>(softmax_lse, "L"),
mk_aotensor<2>(delta, "delta"), mk_aotensor<2>(delta, "delta"),
float(dropout_p), float(dropout_p),
rng_engine_inputs.seed_.val, mk_aoscalartensor(philox_seed),
rng_engine_inputs.offset_.val, mk_aoscalartensor(philox_offset),
0,
is_causal, is_causal,
stream); stream);
#else #else

View File

@ -210,6 +210,7 @@ bool check_flash_attention_hardware_support(sdp_params const& params, bool debug
// Check that the gpu is capable of running flash attention // Check that the gpu is capable of running flash attention
using sm80 = SMVersion<8, 0>; using sm80 = SMVersion<8, 0>;
using sm90 = SMVersion<9, 0>; using sm90 = SMVersion<9, 0>;
auto dprops = at::cuda::getCurrentDeviceProperties();
#if USE_ROCM #if USE_ROCM
#if USE_AOTRITON #if USE_AOTRITON
auto stream = at::cuda::getCurrentCUDAStream().stream(); auto stream = at::cuda::getCurrentCUDAStream().stream();
@ -221,11 +222,19 @@ bool check_flash_attention_hardware_support(sdp_params const& params, bool debug
} }
return false; return false;
} }
c10::string_view arch(dprops->gcnArchName);
if (arch == "gfx1100") {
static const bool enable_navi3x = c10::utils::check_env("TORCH_ROCM_AOTRITON_ENABLE_EXPERIMENTAL") == true;
if (!enable_navi3x) {
TORCH_WARN_ONCE("Flash attention support on Navi31 GPU is still experimental."
" Enable it with TORCH_ROCM_AOTRITON_ENABLE_EXPERIMENTAL=1.");
return false;
}
}
#else #else
return false; return false;
#endif #endif
#else #else
auto dprops = at::cuda::getCurrentDeviceProperties();
if (!check_sm_version<sm80, sm90>(dprops)) { if (!check_sm_version<sm80, sm90>(dprops)) {
if (debug) { if (debug) {
TORCH_WARN( TORCH_WARN(
@ -245,6 +254,7 @@ bool check_mem_efficient_hardware_support(sdp_params const& params, bool debug)
// Mem Efficient attention supports hardware in the range [sm_50, sm_90] // Mem Efficient attention supports hardware in the range [sm_50, sm_90]
using sm50 = SMVersion<5, 0>; using sm50 = SMVersion<5, 0>;
using sm90 = SMVersion<9, 0>; using sm90 = SMVersion<9, 0>;
auto dprops = at::cuda::getCurrentDeviceProperties();
#if USE_ROCM #if USE_ROCM
#if USE_AOTRITON #if USE_AOTRITON
auto stream = at::cuda::getCurrentCUDAStream().stream(); auto stream = at::cuda::getCurrentCUDAStream().stream();
@ -256,11 +266,19 @@ bool check_mem_efficient_hardware_support(sdp_params const& params, bool debug)
} }
return false; return false;
} }
c10::string_view arch(dprops->gcnArchName);
if (arch == "gfx1100") {
static const bool enable_navi3x = c10::utils::check_env("TORCH_ROCM_AOTRITON_ENABLE_EXPERIMENTAL") == true;
if (!enable_navi3x) {
TORCH_WARN_ONCE("Memory Efficient attention on Navi31 GPU is still experimental."
" Enable it with TORCH_ROCM_AOTRITON_ENABLE_EXPERIMENTAL=1.");
return false;
}
}
#else #else
return false; return false;
#endif #endif
#else #else
auto dprops = at::cuda::getCurrentDeviceProperties();
if (!check_sm_version<sm50, sm90>(dprops)) { if (!check_sm_version<sm50, sm90>(dprops)) {
if (debug) { if (debug) {
TORCH_WARN( TORCH_WARN(
@ -561,7 +579,7 @@ bool can_use_cudnn_attention(const sdp_params& params, bool debug) {
check_cudnn_deterministic, check_cudnn_deterministic,
// check_is_causal, // check_is_causal,
check_dtypes_low_precision, check_dtypes_low_precision,
check_for_attn_mask_cudnn, check_attn_mask_shape,
check_cudnn_hardware_support check_cudnn_hardware_support
); );
for (auto& constraint : general_constraints) { for (auto& constraint : general_constraints) {
@ -616,9 +634,14 @@ bool can_use_flash_attention(sdp_params const& params, bool debug) {
} }
} }
} }
#if USE_ROCM
constexpr bool backend_supports_grouped_query_attention = false;
#else
constexpr bool backend_supports_grouped_query_attention = true;
#endif
if (has_only_dense_inputs(params)) { if (has_only_dense_inputs(params)) {
constexpr auto dense_constraints = array_of<bool (*)(sdp_params const&, bool)>( constexpr auto dense_constraints = array_of<bool (*)(sdp_params const&, bool)>(
check_batch_size_and_num_heads_dense<true /*supports_grouped_query_attention=*/>, check_batch_size_and_num_heads_dense<backend_supports_grouped_query_attention>,
check_nonzero_sequence_lengths_dense, check_nonzero_sequence_lengths_dense,
check_last_dim_stride_equals_1_dense<true /*ignore_singleton_dim=*/>); check_last_dim_stride_equals_1_dense<true /*ignore_singleton_dim=*/>);
for (auto& constraint : dense_constraints) { for (auto& constraint : dense_constraints) {

View File

@ -115,6 +115,18 @@ aotriton::TensorView<Rank> mk_aotensor(const at::Tensor& q, c10::string_view ten
cast_dtype(q.dtype())); cast_dtype(q.dtype()));
} }
inline aotriton::TensorView<0> mk_aoscalartensor(const at::Tensor& q)
{
return aotriton::TensorView<0>(reinterpret_cast<intptr_t>(q.data_ptr()),
cast_dtype(q.dtype()));
}
inline aotriton::TensorView<0> mk_philoxtensor(const int64_t* ptr)
{
return aotriton::TensorView<0>(reinterpret_cast<intptr_t>(ptr),
aotriton::DType::kUInt64); // AOTriton excepts unsigned int64
}
} // namespace aotriton_adapter } // namespace aotriton_adapter
} // namespace sdp } // namespace sdp

View File

@ -72,7 +72,8 @@ void check_gpu_arch(hipStream_t stream) {
auto ret = aotriton::v2::flash::check_gpu(stream); auto ret = aotriton::v2::flash::check_gpu(stream);
if (hipSuccess != ret) { if (hipSuccess != ret) {
TORCH_CHECK(false, TORCH_CHECK(false,
"FlashAttention only supports MI200/MI300X GPUs (gfx90a:sramecc+:xnack- or gfx942:sramecc+:xnack-)") "[AOTriton] Accelerated SDPA only supports MI200/MI300X/Navi31 GPUs"
" (gfx90a:sramecc+:xnack-/gfx942:sramecc+:xnack-/gfx1100)")
} }
} }
@ -164,6 +165,8 @@ mha_fwd(const at::Tensor &q, // batch_size x seqlen_q x num_heads x head
auto gen = at::get_generator_or_default<at::CUDAGeneratorImpl>(std::nullopt, at::cuda::detail::getDefaultCUDAGenerator()); auto gen = at::get_generator_or_default<at::CUDAGeneratorImpl>(std::nullopt, at::cuda::detail::getDefaultCUDAGenerator());
at::Tensor seed_t, offset_t; at::Tensor seed_t, offset_t;
at::PhiloxCudaState philox_state;
bool use_philox_state = false;
if (p_dropout > 0.0) { if (p_dropout > 0.0) {
// number of times random will be generated per thread, to offset philox counter in thc random // number of times random will be generated per thread, to offset philox counter in thc random
// state // state
@ -171,12 +174,14 @@ mha_fwd(const at::Tensor &q, // batch_size x seqlen_q x num_heads x head
int64_t counter_offset = batch_size * num_heads * 32; int64_t counter_offset = batch_size * num_heads * 32;
// See Note [Acquire lock when using random generators] // See Note [Acquire lock when using random generators]
std::lock_guard<std::mutex> lock(gen->mutex_); std::lock_guard<std::mutex> lock(gen->mutex_);
at::PhiloxCudaState philox_state = gen->philox_cuda_state(counter_offset); philox_state = gen->philox_cuda_state(counter_offset);
if (at::cuda::currentStreamCaptureStatus() == at::cuda::CaptureStatus::None) { if (at::cuda::currentStreamCaptureStatus() == at::cuda::CaptureStatus::None) {
auto [seed, offset] = at::cuda::philox::unpack(philox_state); auto [seed, offset] = at::cuda::philox::unpack(philox_state);
seed_t = at::scalar_tensor(at::Scalar(static_cast<int64_t>(seed)), at::dtype(at::kLong)); seed_t = at::scalar_tensor(at::Scalar(static_cast<int64_t>(seed)), at::dtype(at::kLong).device(at::kCUDA));
offset_t = at::scalar_tensor(at::Scalar(static_cast<int64_t>(offset)), at::dtype(at::kLong)); offset_t = at::scalar_tensor(at::Scalar(static_cast<int64_t>(offset)), at::dtype(at::kLong).device(at::kCUDA));
} else { } else {
// See Note [CUDA Graph-safe RNG states] about the design
use_philox_state = true;
seed_t = at::empty({}, at::dtype(at::kLong).device(at::kCUDA)); seed_t = at::empty({}, at::dtype(at::kLong).device(at::kCUDA));
offset_t = at::empty({}, at::dtype(at::kLong).device(at::kCUDA)); offset_t = at::empty({}, at::dtype(at::kLong).device(at::kCUDA));
} }
@ -185,8 +190,8 @@ mha_fwd(const at::Tensor &q, // batch_size x seqlen_q x num_heads x head
seed_t = at::empty({}, at::dtype(at::kLong).device(at::kCUDA)); seed_t = at::empty({}, at::dtype(at::kLong).device(at::kCUDA));
offset_t = at::empty({}, at::dtype(at::kLong).device(at::kCUDA)); offset_t = at::empty({}, at::dtype(at::kLong).device(at::kCUDA));
} else { } else {
seed_t = at::empty({}, at::dtype(at::kLong)); seed_t = at::empty({}, at::dtype(at::kLong).device(at::kCUDA));
offset_t = at::empty({}, at::dtype(at::kLong)); offset_t = at::empty({}, at::dtype(at::kLong).device(at::kCUDA));
} }
} }
@ -219,9 +224,17 @@ mha_fwd(const at::Tensor &q, // batch_size x seqlen_q x num_heads x head
hipError_t err; // TODO: Error handling hipError_t err; // TODO: Error handling
using aotriton::v2::flash::attn_fwd; using aotriton::v2::flash::attn_fwd;
using aotriton::TensorView;
using sdp::aotriton_adapter::mk_aotensor; using sdp::aotriton_adapter::mk_aotensor;
using sdp::aotriton_adapter::mk_aoscalartensor;
using sdp::aotriton_adapter::mk_philoxtensor;
using sdp::aotriton_adapter::cast_dtype; using sdp::aotriton_adapter::cast_dtype;
aotriton::TensorView<4> empty_bias(0, {0,0,0,0}, {0,0,0,0}, cast_dtype(q.dtype())); aotriton::TensorView<4> empty_bias(0, {0,0,0,0}, {0,0,0,0}, cast_dtype(q.dtype()));
auto seed = use_philox_state ? mk_philoxtensor(philox_state.seed_.ptr) : mk_aoscalartensor(seed_t);
auto offset1 = use_philox_state ? mk_philoxtensor(philox_state.offset_.ptr) : mk_aoscalartensor(offset_t);
auto offset2 = use_philox_state ? philox_state.offset_intragraph_ : 0;
auto seed_output = use_philox_state ? mk_philoxtensor(seed_t.data_ptr<int64_t>()) : mk_philoxtensor(nullptr);
auto offset_output = use_philox_state ? mk_philoxtensor(offset_t.data_ptr<int64_t>()) : mk_philoxtensor(nullptr);
err = attn_fwd(mk_aotensor(q_t, "q"), err = attn_fwd(mk_aotensor(q_t, "q"),
mk_aotensor(k_t, "k"), mk_aotensor(k_t, "k"),
mk_aotensor(v_t, "v"), mk_aotensor(v_t, "v"),
@ -230,8 +243,11 @@ mha_fwd(const at::Tensor &q, // batch_size x seqlen_q x num_heads x head
mk_aotensor<2>(M, "M"), mk_aotensor<2>(M, "M"),
mk_aotensor(output_t, "Out"), mk_aotensor(output_t, "Out"),
p_dropout, p_dropout,
philox_args.seed_.val, seed,
philox_args.offset_.val, offset1,
offset2,
seed_output,
offset_output,
mk_aotensor(softmax_fa_t, "encoded_softmax"), mk_aotensor(softmax_fa_t, "encoded_softmax"),
is_causal, is_causal,
stream); stream);
@ -420,6 +436,7 @@ mha_bwd(const at::Tensor &dout, // batch_size x seqlen_q x num_heads, x head_si
{ {
using aotriton::v2::flash::attn_bwd; using aotriton::v2::flash::attn_bwd;
using sdp::aotriton_adapter::mk_aotensor; using sdp::aotriton_adapter::mk_aotensor;
using sdp::aotriton_adapter::mk_aoscalartensor;
using sdp::aotriton_adapter::cast_dtype; using sdp::aotriton_adapter::cast_dtype;
aotriton::TensorView<4> empty_bias(0, {0,0,0,0}, {0,0,0,0}, cast_dtype(q.dtype())); aotriton::TensorView<4> empty_bias(0, {0,0,0,0}, {0,0,0,0}, cast_dtype(q.dtype()));
err = attn_bwd(mk_aotensor(q_t, "q"), err = attn_bwd(mk_aotensor(q_t, "q"),
@ -436,8 +453,9 @@ mha_bwd(const at::Tensor &dout, // batch_size x seqlen_q x num_heads, x head_si
mk_aotensor<2>(softmax_lse_cont, "L"), mk_aotensor<2>(softmax_lse_cont, "L"),
mk_aotensor<2>(delta, "delta"), mk_aotensor<2>(delta, "delta"),
p_dropout, p_dropout,
philox_args.seed_.val, mk_aoscalartensor(philox_seed),
philox_args.offset_.val, mk_aoscalartensor(philox_offset),
0,
is_causal, is_causal,
stream); stream);
} }

View File

@ -275,17 +275,6 @@ inline bool check_for_attn_mask(sdp_params const& params, bool debug) {
return true; return true;
} }
// TODO(eqy): remove this once support is added
inline bool check_for_attn_mask_cudnn(sdp_params const& params, bool debug) {
if (params.attn_mask.has_value()) {
if (debug) {
TORCH_WARN("cuDNN Attention does not support non-null attn_mask.");
}
return false;
}
return true;
}
inline bool check_attn_mask_shape(sdp_params const& params, bool debug) { inline bool check_attn_mask_shape(sdp_params const& params, bool debug) {
auto attn_mask = params.attn_mask; auto attn_mask = params.attn_mask;
if (!attn_mask.has_value()) { if (!attn_mask.has_value()) {

Some files were not shown because too many files have changed in this diff Show More