Skip to content

Commit 1cddbf8

Browse files
authored
Revert Add host-supports-nvptx requirement to lit tests (#66102 and #66129) (#66225)
1 parent 8b9bf3a commit 1cddbf8

25 files changed

+0
-48
lines changed

mlir/test/Integration/Dialect/SparseTensor/GPU/CUDA/dump-ptx.mlir

-2
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,3 @@
1-
// REQUIRES: host-supports-nvptx
2-
31
// RUN: mlir-opt %s \
42
// RUN: | mlir-opt -test-lower-to-nvvm -debug-only=serialize-to-isa \
53
// RUN: 2>&1 | FileCheck %s

mlir/test/Integration/Dialect/SparseTensor/GPU/CUDA/sparse-matvec-const.mlir

-1
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,3 @@
1-
// REQUIRES: host-supports-nvptx
21
//
32
// NOTE: this test requires gpu-sm80
43
//

mlir/test/Integration/Dialect/SparseTensor/GPU/CUDA/sparse-matvec.mlir

-1
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,3 @@
1-
// REQUIRES: host-supports-nvptx
21
//
32
// NOTE: this test requires gpu-sm80
43
//

mlir/test/Integration/Dialect/SparseTensor/GPU/CUDA/sparse-mma-2-4-f16.mlir

-2
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,3 @@
1-
// REQUIRES: host-supports-nvptx
2-
31
//
42
// NOTE: this test requires gpu-sm80
53
//

mlir/test/Integration/Dialect/Vector/GPU/CUDA/test-reduction-distribute.mlir

-2
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,3 @@
1-
// REQUIRES: host-supports-nvptx
2-
31
// RUN: mlir-opt %s -test-vector-warp-distribute="hoist-uniform distribute-transfer-write propagate-distribution" -canonicalize |\
42
// RUN: mlir-opt -test-vector-warp-distribute=rewrite-warp-ops-to-scf-if |\
53
// RUN: mlir-opt -lower-affine -convert-vector-to-scf -convert-scf-to-cf -convert-vector-to-llvm \

mlir/test/Integration/Dialect/Vector/GPU/CUDA/test-warp-distribute.mlir

-2
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,3 @@
1-
// REQUIRES: host-supports-nvptx
2-
31
// Run the test cases without distributing ops to test default lowering. Run
42
// everything on the same thread.
53
// RUN: mlir-opt %s -test-vector-warp-distribute=rewrite-warp-ops-to-scf-if -canonicalize | \

mlir/test/Integration/GPU/CUDA/TensorCore/sm80/transform-mma-sync-matmul-f16-f16-accum.mlir

-2
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,3 @@
1-
// REQUIRES: host-supports-nvptx
2-
31
// RUN: mlir-opt %s \
42
// RUN: -test-transform-dialect-interpreter \
53
// RUN: -test-transform-dialect-erase-schedule \

mlir/test/Integration/GPU/CUDA/TensorCore/sm80/transform-mma-sync-matmul-f32.mlir

-2
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,3 @@
1-
// REQUIRES: host-supports-nvptx
2-
//
31
// RUN: mlir-opt %s \
42
// RUN: -test-transform-dialect-interpreter \
53
// RUN: | FileCheck %s --check-prefix=CHECK-MMA-SYNC

mlir/test/Integration/GPU/CUDA/TensorCore/wmma-matmul-f16.mlir

-2
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,3 @@
1-
// REQUIRES: host-supports-nvptx
2-
31
// RUN: mlir-opt %s \
42
// RUN: | mlir-opt -test-lower-to-nvvm="cubin-chip=sm_70" \
53
// RUN: | mlir-cpu-runner \

mlir/test/Integration/GPU/CUDA/TensorCore/wmma-matmul-f32-bare-ptr.mlir

-2
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,3 @@
1-
// REQUIRES: host-supports-nvptx
2-
31
// Tests memref bare pointer lowering convention both host side and kernel-side;
42
// this works for only statically shaped memrefs.
53
// Similar to the wmma-matmul-f32 but but with the memref bare pointer lowering convention.

mlir/test/Integration/GPU/CUDA/TensorCore/wmma-matmul-f32.mlir

-2
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,3 @@
1-
// REQUIRES: host-supports-nvptx
2-
31
// RUN: mlir-opt %s \
42
// RUN: | mlir-opt -test-lower-to-nvvm="cubin-chip=sm_70" \
53
// RUN: | mlir-cpu-runner \

mlir/test/Integration/GPU/CUDA/all-reduce-and.mlir

-2
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,3 @@
1-
// REQUIRES: host-supports-nvptx
2-
31
// RUN: mlir-opt %s \
42
// RUN: | mlir-opt -test-lower-to-nvvm \
53
// RUN: | mlir-cpu-runner \

mlir/test/Integration/GPU/CUDA/all-reduce-max.mlir

-2
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,3 @@
1-
// REQUIRES: host-supports-nvptx
2-
31
// RUN: mlir-opt %s \
42
// RUN: | mlir-opt -test-lower-to-nvvm \
53
// RUN: | mlir-cpu-runner \

mlir/test/Integration/GPU/CUDA/all-reduce-min.mlir

-2
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,3 @@
1-
// REQUIRES: host-supports-nvptx
2-
31
// RUN: mlir-opt %s \
42
// RUN: | mlir-opt -test-lower-to-nvvm \
53
// RUN: | mlir-cpu-runner \

mlir/test/Integration/GPU/CUDA/all-reduce-op.mlir

-2
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,3 @@
1-
// REQUIRES: host-supports-nvptx
2-
31
// RUN: mlir-opt %s \
42
// RUN: | mlir-opt -test-lower-to-nvvm \
53
// RUN: | mlir-cpu-runner \

mlir/test/Integration/GPU/CUDA/all-reduce-or.mlir

-2
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,3 @@
1-
// REQUIRES: host-supports-nvptx
2-
31
// RUN: mlir-opt %s \
42
// RUN: | mlir-opt -test-lower-to-nvvm \
53
// RUN: | mlir-cpu-runner \

mlir/test/Integration/GPU/CUDA/all-reduce-region.mlir

-2
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,3 @@
1-
// REQUIRES: host-supports-nvptx
2-
31
// RUN: mlir-opt %s \
42
// RUN: | mlir-opt -test-lower-to-nvvm \
53
// RUN: | mlir-cpu-runner \

mlir/test/Integration/GPU/CUDA/all-reduce-xor.mlir

-2
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,3 @@
1-
// REQUIRES: host-supports-nvptx
2-
31
// RUN: mlir-opt %s \
42
// RUN: | mlir-opt -test-lower-to-nvvm \
53
// RUN: | mlir-cpu-runner \

mlir/test/Integration/GPU/CUDA/async.mlir

-2
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,3 @@
1-
// REQUIRES: host-supports-nvptx
2-
31
// RUN: mlir-opt %s \
42
// RUN: | mlir-opt -gpu-kernel-outlining \
53
// RUN: | mlir-opt -pass-pipeline='builtin.module(gpu.module(strip-debuginfo,convert-gpu-to-nvvm),nvvm-attach-target)' \

mlir/test/Integration/GPU/CUDA/gpu-to-cubin.mlir

-2
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,3 @@
1-
// REQUIRES: host-supports-nvptx
2-
31
// RUN: mlir-opt %s \
42
// RUN: | mlir-opt -test-lower-to-nvvm \
53
// RUN: | mlir-cpu-runner \

mlir/test/Integration/GPU/CUDA/multiple-all-reduce.mlir

-2
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,3 @@
1-
// REQUIRES: host-supports-nvptx
2-
31
// RUN: mlir-opt %s \
42
// RUN: | mlir-opt -test-lower-to-nvvm \
53
// RUN: | mlir-cpu-runner \

mlir/test/Integration/GPU/CUDA/printf.mlir

-2
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,3 @@
1-
// REQUIRES: host-supports-nvptx
2-
31
// RUN: mlir-opt %s \
42
// RUN: | mlir-opt -test-lower-to-nvvm \
53
// RUN: | mlir-cpu-runner \

mlir/test/Integration/GPU/CUDA/shuffle.mlir

-2
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,3 @@
1-
// REQUIRES: host-supports-nvptx
2-
31
// RUN: mlir-opt %s \
42
// RUN: | mlir-opt -test-lower-to-nvvm \
53
// RUN: | mlir-cpu-runner \

mlir/test/Integration/GPU/CUDA/sm90/transform-dialect/tma_load_64x8_8x128_noswizzle-transform.mlir

-2
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,3 @@
1-
// REQUIRES: host-supports-nvptx
2-
31
// RUN: mlir-opt %s \
42
// RUN: -test-transform-dialect-interpreter \
53
// RUN: -test-transform-dialect-erase-schedule \

mlir/test/Integration/GPU/CUDA/two-modules.mlir

-2
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,3 @@
1-
// REQUIRES: host-supports-nvptx
2-
31
// RUN: mlir-opt %s \
42
// RUN: | mlir-opt -test-lower-to-nvvm \
53
// RUN: | mlir-cpu-runner \

0 commit comments

Comments
 (0)