mirror of
https://github.com/pytorch/pytorch.git
synced 2025-11-16 23:44:53 +08:00
Add operator benchmarking run to CI nightly (#162530) This PR introduces a new "operator microbenchmark" CI workflow and GitHub Actions for operator microbenchmarks, updating test scripts and job matrices to support new parameters, and broadening the operator benchmark tests to include more data types, larger shapes, and gradient tests. The benchmark configurations now focus more on different cuda hardware and multiple dtypes (bf16, fp16, fp32), for both compile and eager mode. **Benchmark Configuration and Coverage:** * Expanded operator benchmark configurations in `addmm_test.py`, `bmm_test.py`, `matmul_test.py`, and `mm_test.py` to benchmark multiple dtypes on CUDA devices, in eager and compile mode, for forward and backward run. The configs with tag "long" for the above mentioned files are being run in CI. * The CI benchmarking is running on various hardwares: H100, A100. * The CI job also uploads the microbenchmarking outputs to a [HUD](https://hud.pytorch.org/benchmark/llms?repoName=pytorch%2Fpytorch&benchmarkName=PyTorch+operator+microbenchmark) dashboard. Pull Request resolved: https://github.com/pytorch/pytorch/pull/162530 Approved by: https://github.com/huydhn (cherry picked from commit 54b38f3b46c33a1cc4e8f7894619358afcbd7c89) Co-authored-by: jainapurva <apurvajain.kota@gmail.com> Co-authored-by: Huy Do <huydhn@gmail.com>
69 lines
2.0 KiB
Python
69 lines
2.0 KiB
Python
import operator_benchmark as op_bench
|
|
|
|
import torch
|
|
|
|
|
|
"""Microbenchmarks for MatMul operator"""
|
|
|
|
# Configs for PT Matmul operator
|
|
mm_short_configs = op_bench.config_list(
|
|
attr_names=["M", "N", "K", "trans_a", "trans_b"],
|
|
attrs=[
|
|
[1, 1, 1, True, False],
|
|
[128, 128, 128, True, False],
|
|
[256, 256, 256, False, True],
|
|
],
|
|
cross_product_configs={"device": ["cpu", "cuda"]},
|
|
tags=["short"],
|
|
)
|
|
|
|
|
|
mm_long_configs = op_bench.cross_product_configs(
|
|
M=[256, 1024, 3000],
|
|
N=[512, 4096],
|
|
K=[512, 4096],
|
|
trans_a=[False, True],
|
|
trans_b=[True, False],
|
|
device=["cuda"],
|
|
dtype=[torch.float16, torch.bfloat16, torch.float32],
|
|
tags=["long"],
|
|
)
|
|
|
|
|
|
class MatMulBenchmark(op_bench.TorchBenchmarkBase):
|
|
def init(self, M, N, K, trans_a, trans_b, device, dtype=torch.float):
|
|
# Create tensors without requires_grad first, then set it separately
|
|
# This avoids creating graph leaves that cannot be deep copied
|
|
if trans_a:
|
|
input_one = torch.rand(M, N, device=device, dtype=dtype)
|
|
else:
|
|
input_one = torch.rand(N, M, device=device, dtype=dtype).t()
|
|
|
|
if trans_b:
|
|
input_two = torch.rand(N, K, device=device, dtype=dtype)
|
|
else:
|
|
input_two = torch.rand(K, N, device=device, dtype=dtype).t()
|
|
|
|
# Set requires_grad after tensor creation to avoid graph leaf issues
|
|
if self.auto_set():
|
|
input_one.requires_grad_(True)
|
|
if self.auto_set():
|
|
input_two.requires_grad_(True)
|
|
|
|
self.inputs = {
|
|
"input_one": input_one,
|
|
"input_two": input_two,
|
|
}
|
|
self.set_module_name("matmul")
|
|
|
|
def forward(self, input_one, input_two):
|
|
return torch.matmul(input_one, input_two)
|
|
|
|
|
|
op_bench.generate_pt_test(mm_long_configs + mm_short_configs, MatMulBenchmark)
|
|
op_bench.generate_pt_gradient_test(mm_long_configs, MatMulBenchmark)
|
|
|
|
|
|
if __name__ == "__main__":
|
|
op_bench.benchmark_runner.main()
|