Skip to content

Commit 2a07f0f

Browse files
[mlir][sparse] Migrate more tests to use new syntax (#66443)
**Dense** `lvlTypes = [ "dense", "dense" ]` to `map = (d0, d1) -> (d0 : dense, d1 : dense)` `lvlTypes = [ "dense", "dense" ], dimToLvl = affine_map<(i,j) -> (j,i)>` to `map = (d0, d1) -> (d1 : dense, d0 : dense)` **DCSR** `lvlTypes = [ "compressed", "compressed" ]` to `map = (d0, d1) -> (d0 : compressed, d1 : compressed)` **DCSC** `lvlTypes = [ "compressed", "compressed" ], dimToLvl = affine_map<(i,j) -> (j,i)>` to `map = (d0, d1) -> (d1 : compressed, d0 : compressed)` **Block Row** `lvlTypes = [ "compressed", "dense" ]` to `map = (d0, d1) -> (d0 : compressed, d1 : dense)` **Block Column** `lvlTypes = [ "compressed", "dense" ], dimToLvl = affine_map<(i,j) -> (j,i)>` to `map = (d0, d1) -> (d1 : compressed, d0 : dense)` This is an ongoing effort: #66146, #66309
1 parent 4db6803 commit 2a07f0f

File tree

69 files changed

+128
-165
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

69 files changed

+128
-165
lines changed

mlir/include/mlir/Dialect/SparseTensor/IR/SparseTensorAttrDefs.td

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -206,8 +206,7 @@ def SparseTensorEncodingAttr : SparseTensor_Attr<"SparseTensorEncoding",
206206

207207
// Doubly compressed sparse column storage with specific bitwidths.
208208
#DCSC = #sparse_tensor.encoding<{
209-
lvlTypes = [ "compressed", "compressed" ],
210-
dimToLvl = affine_map<(i, j) -> (j, i)>,
209+
map = (d0, d1) -> (d1 : compressed, d0 : compressed),
211210
posWidth = 32,
212211
crdWidth = 8
213212
}>

mlir/lib/Dialect/SparseTensor/Transforms/Sparsification.cpp

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -351,8 +351,7 @@ static bool findDepIdxSet(Merger &merger, TensorId tensor, Level lvl,
351351
/// Get the total number of compound affine expressions in the
352352
/// `getMatchingIndexingMap` for the given tensor. For the following inputs:
353353
///
354-
/// map = (d0, d1, d2) => (d0 + d1, d2)
355-
/// lvlTypes = ["compressed", "compressed"]
354+
/// map = (d0, d1, d2) => (d0 + d1 : compressed, d2 : compressed)
356355
///
357356
/// Returns 1 (because the first level is compressed and its corresponding
358357
/// indexing-expression is `d0 + d1`)

mlir/test/Dialect/Bufferization/invalid.mlir

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -58,7 +58,7 @@ func.func @escape_attr_non_bufferizable(%m0: memref<?xf32>) {
5858

5959
// -----
6060

61-
#DCSR = #sparse_tensor.encoding<{ lvlTypes = [ "compressed", "compressed" ] }>
61+
#DCSR = #sparse_tensor.encoding<{ map = (d0, d1) -> (d0 : compressed, d1 : compressed) }>
6262

6363
func.func @sparse_alloc_direct_return() -> tensor<20x40xf32, #DCSR> {
6464
// expected-error @+1{{sparse tensor allocation should not escape function}}
@@ -68,7 +68,7 @@ func.func @sparse_alloc_direct_return() -> tensor<20x40xf32, #DCSR> {
6868

6969
// -----
7070

71-
#DCSR = #sparse_tensor.encoding<{ lvlTypes = [ "compressed", "compressed" ] }>
71+
#DCSR = #sparse_tensor.encoding<{ map = (d0, d1) -> (d0 : compressed, d1 : compressed) }>
7272

7373
func.func private @foo(tensor<20x40xf32, #DCSR>) -> ()
7474

mlir/test/Dialect/SparseTensor/codegen.mlir

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -9,13 +9,13 @@
99
}>
1010

1111
#Dense2D = #sparse_tensor.encoding<{
12-
lvlTypes = [ "dense", "dense" ],
12+
map = (d0, d1) -> (d0 : dense, d1 : dense),
1313
crdWidth = 64,
1414
posWidth = 32
1515
}>
1616

1717
#Row = #sparse_tensor.encoding<{
18-
lvlTypes = [ "compressed", "dense" ],
18+
map = (d0, d1) -> (d0 : compressed, d1 : dense),
1919
crdWidth = 64,
2020
posWidth = 32
2121
}>
@@ -35,7 +35,7 @@
3535
}>
3636

3737
#DCSR = #sparse_tensor.encoding<{
38-
lvlTypes = [ "compressed", "compressed" ],
38+
map = (d0, d1) -> (d0 : compressed, d1 : compressed),
3939
crdWidth = 64,
4040
posWidth = 32
4141
}>

mlir/test/Dialect/SparseTensor/dense.mlir

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@
77
// latter class is linearized into one-dimensional buffers that are backed
88
// by the runtime support library.
99

10-
#DenseMatrix = #sparse_tensor.encoding<{ lvlTypes = [ "dense", "dense" ] }>
10+
#DenseMatrix = #sparse_tensor.encoding<{ map = (d0, d1) -> (d0 : dense, d1 : dense) }>
1111

1212
#trait_2d = {
1313
indexing_maps = [

mlir/test/Dialect/SparseTensor/invalid.mlir

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -371,7 +371,7 @@ func.func @sparse_convert_unranked(%arg0: tensor<*xf32>) -> tensor<10xf32> {
371371

372372
// -----
373373

374-
#DCSR = #sparse_tensor.encoding<{lvlTypes = ["compressed", "compressed"]}>
374+
#DCSR = #sparse_tensor.encoding<{map = (d0, d1) -> (d0 : compressed, d1 : compressed)}>
375375

376376
func.func @sparse_convert_rank_mismatch(%arg0: tensor<10x10xf64, #DCSR>) -> tensor<?xf64> {
377377
// expected-error@+1 {{unexpected conversion mismatch in rank}}
@@ -714,7 +714,7 @@ func.func @invalid_concat_size_mismatch(%arg0: tensor<2x4xf64, #DC>,
714714

715715
// -----
716716

717-
#DCSR = #sparse_tensor.encoding<{lvlTypes = ["compressed", "compressed"]}>
717+
#DCSR = #sparse_tensor.encoding<{map = (d0, d1) -> (d0 : compressed, d1 : compressed)}>
718718
func.func @sparse_tensor_foreach(%arg0: tensor<2x4xf64, #DCSR>) -> () {
719719
// expected-error@+1 {{Unmatched number of arguments in the block}}
720720
sparse_tensor.foreach in %arg0 : tensor<2x4xf64, #DCSR> do {
@@ -725,7 +725,7 @@ func.func @sparse_tensor_foreach(%arg0: tensor<2x4xf64, #DCSR>) -> () {
725725

726726
// -----
727727

728-
#DCSR = #sparse_tensor.encoding<{lvlTypes = ["compressed", "compressed"]}>
728+
#DCSR = #sparse_tensor.encoding<{map = (d0, d1) -> (d0 : compressed, d1 : compressed)}>
729729
func.func @sparse_tensor_foreach(%arg0: tensor<2x4xf64, #DCSR>) -> () {
730730
// expected-error@+1 {{Expecting Index type for argument at index 1}}
731731
sparse_tensor.foreach in %arg0 : tensor<2x4xf64, #DCSR> do {
@@ -736,7 +736,7 @@ func.func @sparse_tensor_foreach(%arg0: tensor<2x4xf64, #DCSR>) -> () {
736736

737737
// -----
738738

739-
#DCSR = #sparse_tensor.encoding<{lvlTypes = ["compressed", "compressed"]}>
739+
#DCSR = #sparse_tensor.encoding<{map = (d0, d1) -> (d0 : compressed, d1 : compressed)}>
740740
func.func @sparse_tensor_foreach(%arg0: tensor<2x4xf64, #DCSR>) -> () {
741741
// expected-error@+1 {{Unmatched element type between input tensor and block argument}}
742742
sparse_tensor.foreach in %arg0 : tensor<2x4xf64, #DCSR> do {
@@ -747,7 +747,7 @@ func.func @sparse_tensor_foreach(%arg0: tensor<2x4xf64, #DCSR>) -> () {
747747

748748
// -----
749749

750-
#DCSR = #sparse_tensor.encoding<{lvlTypes = ["compressed", "compressed"]}>
750+
#DCSR = #sparse_tensor.encoding<{map = (d0, d1) -> (d0 : compressed, d1 : compressed)}>
751751
func.func @sparse_tensor_foreach(%arg0: tensor<2x4xf64, #DCSR>) -> () {
752752
// expected-error@+1 {{Unmatched element type between input tensor and block argument}}
753753
sparse_tensor.foreach in %arg0 : tensor<2x4xf64, #DCSR> do {
@@ -758,7 +758,7 @@ func.func @sparse_tensor_foreach(%arg0: tensor<2x4xf64, #DCSR>) -> () {
758758

759759
// -----
760760

761-
#DCSR = #sparse_tensor.encoding<{lvlTypes = ["compressed", "compressed"]}>
761+
#DCSR = #sparse_tensor.encoding<{map = (d0, d1) -> (d0 : compressed, d1 : compressed)}>
762762
func.func @sparse_tensor_foreach(%arg0: tensor<2x4xf64, #DCSR>, %arg1: f32) -> () {
763763
// expected-error@+1 {{Mismatch in number of init arguments and results}}
764764
sparse_tensor.foreach in %arg0 init(%arg1) : tensor<2x4xf64, #DCSR>, f32 do {
@@ -769,7 +769,7 @@ func.func @sparse_tensor_foreach(%arg0: tensor<2x4xf64, #DCSR>, %arg1: f32) -> (
769769

770770
// -----
771771

772-
#DCSR = #sparse_tensor.encoding<{lvlTypes = ["compressed", "compressed"]}>
772+
#DCSR = #sparse_tensor.encoding<{map = (d0, d1) -> (d0 : compressed, d1 : compressed)}>
773773
func.func @sparse_tensor_foreach(%arg0: tensor<2x4xf64, #DCSR>, %arg1: f32) -> () {
774774
// expected-error@+1 {{Mismatch in types of init arguments and results}}
775775
%1 = sparse_tensor.foreach in %arg0 init(%arg1) : tensor<2x4xf64, #DCSR>, f32 -> i32 do {
@@ -780,7 +780,7 @@ func.func @sparse_tensor_foreach(%arg0: tensor<2x4xf64, #DCSR>, %arg1: f32) -> (
780780

781781
// -----
782782

783-
#DCSR = #sparse_tensor.encoding<{lvlTypes = ["compressed", "compressed"]}>
783+
#DCSR = #sparse_tensor.encoding<{map = (d0, d1) -> (d0 : compressed, d1 : compressed)}>
784784
func.func @sparse_tensor_foreach(%arg0: tensor<2x4xf64, #DCSR>, %arg1: f32) -> () {
785785
// expected-error@+1 {{Mismatch in types of yield values and results}}
786786
%1 = sparse_tensor.foreach in %arg0 init(%arg1) : tensor<2x4xf64, #DCSR>, f32 -> f32 do {

mlir/test/Dialect/SparseTensor/one_shot_bufferize_tensor_copy_insertion.mlir

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -2,8 +2,7 @@
22
// RUN: mlir-opt %s -test-tensor-copy-insertion="bufferize-function-boundaries allow-return-allocs" | FileCheck %s --check-prefix=CHECK-FUNC
33

44
#DCSR = #sparse_tensor.encoding<{
5-
lvlTypes = [ "compressed", "compressed" ],
6-
dimToLvl = affine_map<(i,j) -> (i,j)>
5+
map = (d0, d1) -> (d0 : compressed, d1 : compressed)
76
}>
87

98
// CHECK-LABEL: func @bufferization_alloc_tensor

mlir/test/Dialect/SparseTensor/one_trip.mlir

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
// RUN: mlir-opt %s -sparsification -cse | FileCheck %s
22

33
#Dense = #sparse_tensor.encoding<{
4-
lvlTypes = [ "dense" , "dense" ]
4+
map = (d0, d1) -> (d0 : dense, d1 : dense)
55
}>
66

77
#trait_scale = {

mlir/test/Dialect/SparseTensor/post_rewriting.mlir

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@
55
}>
66

77
#SparseMatrix = #sparse_tensor.encoding<{
8-
lvlTypes = ["compressed", "compressed"]
8+
map = (d0, d1) -> (d0 : compressed, d1 : compressed)
99
}>
1010

1111
// CHECK-LABEL: func.func @expand_dense(

mlir/test/Dialect/SparseTensor/pre_rewriting.mlir

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@
99
}>
1010

1111
#DCSR = #sparse_tensor.encoding<{
12-
lvlTypes = ["compressed", "compressed"]
12+
map = (d0, d1) -> (d0 : compressed, d1 : compressed)
1313
}>
1414

1515
#Slice = #sparse_tensor.encoding<{

mlir/test/Dialect/SparseTensor/roundtrip.mlir

Lines changed: 13 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -283,7 +283,7 @@ func.func @sparse_noe(%arg0: tensor<128xf64, #SparseVector>) -> index {
283283

284284
// -----
285285

286-
#DenseMatrix = #sparse_tensor.encoding<{lvlTypes = ["dense","dense"]}>
286+
#DenseMatrix = #sparse_tensor.encoding<{map = (d0, d1) -> (d0 : dense, d1 : dense)}>
287287

288288
// CHECK-LABEL: func @sparse_load(
289289
// CHECK-SAME: %[[A:.*]]: tensor<16x32xf64, #{{.*}}>)
@@ -296,7 +296,7 @@ func.func @sparse_load(%arg0: tensor<16x32xf64, #DenseMatrix>) -> tensor<16x32xf
296296

297297
// -----
298298

299-
#DenseMatrix = #sparse_tensor.encoding<{lvlTypes = ["dense","dense"]}>
299+
#DenseMatrix = #sparse_tensor.encoding<{map = (d0, d1) -> (d0 : dense, d1 : dense)}>
300300

301301
// CHECK-LABEL: func @sparse_load_ins(
302302
// CHECK-SAME: %[[A:.*]]: tensor<16x32xf64, #{{.*}}>)
@@ -364,7 +364,7 @@ func.func @sparse_push_back_n(%arg0: index, %arg1: memref<?xf64>, %arg2: f64, %a
364364

365365
// -----
366366

367-
#SparseMatrix = #sparse_tensor.encoding<{lvlTypes = ["compressed", "compressed"]}>
367+
#SparseMatrix = #sparse_tensor.encoding<{map = (d0, d1) -> (d0 : compressed, d1 : compressed)}>
368368

369369
// CHECK-LABEL: func @sparse_expansion(
370370
// CHECK-SAME: %[[A:.*]]: tensor<8x8xf64, #sparse_tensor.encoding<{{.*}}>>)
@@ -378,7 +378,7 @@ func.func @sparse_expansion(%tensor: tensor<8x8xf64, #SparseMatrix>) -> index {
378378

379379
// -----
380380

381-
#SparseMatrix = #sparse_tensor.encoding<{lvlTypes = ["compressed", "compressed"]}>
381+
#SparseMatrix = #sparse_tensor.encoding<{map = (d0, d1) -> (d0 : compressed, d1 : compressed)}>
382382

383383
// CHECK-LABEL: func @sparse_compression(
384384
// CHECK-SAME: %[[A0:.*0]]: memref<?xf64>,
@@ -402,7 +402,7 @@ func.func @sparse_compression(%values: memref<?xf64>,
402402

403403
// -----
404404

405-
#SparseMatrix = #sparse_tensor.encoding<{lvlTypes = ["compressed", "compressed"]}>
405+
#SparseMatrix = #sparse_tensor.encoding<{map = (d0, d1) -> (d0 : compressed, d1 : compressed)}>
406406

407407
// CHECK-LABEL: func @sparse_out(
408408
// CHECK-SAME: %[[A:.*]]: tensor<?x?xf64, #sparse_tensor.encoding<{{.*}}>>,
@@ -416,7 +416,7 @@ func.func @sparse_out(%arg0: tensor<?x?xf64, #SparseMatrix>, %arg1: !llvm.ptr<i8
416416

417417
// -----
418418

419-
#SparseMatrix = #sparse_tensor.encoding<{lvlTypes = ["compressed", "compressed"]}>
419+
#SparseMatrix = #sparse_tensor.encoding<{map = (d0, d1) -> (d0 : compressed, d1 : compressed)}>
420420

421421
// CHECK-LABEL: func @sparse_binary(
422422
// CHECK-SAME: %[[A:.*]]: f64, %[[B:.*]]: i64) -> f64 {
@@ -450,7 +450,7 @@ func.func @sparse_binary(%arg0: f64, %arg1: i64) -> f64 {
450450

451451
// -----
452452

453-
#SparseMatrix = #sparse_tensor.encoding<{lvlTypes = ["compressed", "compressed"]}>
453+
#SparseMatrix = #sparse_tensor.encoding<{map = (d0, d1) -> (d0 : compressed, d1 : compressed)}>
454454

455455
// CHECK-LABEL: func @sparse_unary(
456456
// CHECK-SAME: %[[A:.*]]: f64) -> f64 {
@@ -480,7 +480,7 @@ func.func @sparse_unary(%arg0: f64) -> f64 {
480480

481481
// -----
482482

483-
#SparseMatrix = #sparse_tensor.encoding<{lvlTypes = ["compressed", "compressed"]}>
483+
#SparseMatrix = #sparse_tensor.encoding<{map = (d0, d1) -> (d0 : compressed, d1 : compressed)}>
484484

485485
// CHECK-LABEL: func @sparse_unary(
486486
// CHECK-SAME: %[[A:.*]]: f64) -> i64 {
@@ -507,7 +507,7 @@ func.func @sparse_unary(%arg0: f64) -> i64 {
507507

508508
// -----
509509

510-
#SparseMatrix = #sparse_tensor.encoding<{lvlTypes = ["compressed", "compressed"]}>
510+
#SparseMatrix = #sparse_tensor.encoding<{map = (d0, d1) -> (d0 : compressed, d1 : compressed)}>
511511

512512
// CHECK-LABEL: func @sparse_reduce_2d_to_1d(
513513
// CHECK-SAME: %[[A:.*]]: f64, %[[B:.*]]: f64) -> f64 {
@@ -529,7 +529,7 @@ func.func @sparse_reduce_2d_to_1d(%arg0: f64, %arg1: f64) -> f64 {
529529

530530
// -----
531531

532-
#SparseMatrix = #sparse_tensor.encoding<{lvlTypes = ["compressed", "compressed"]}>
532+
#SparseMatrix = #sparse_tensor.encoding<{map = (d0, d1) -> (d0 : compressed, d1 : compressed)}>
533533

534534
// CHECK-LABEL: func @sparse_select(
535535
// CHECK-SAME: %[[A:.*]]: f64) -> f64 {
@@ -553,7 +553,7 @@ func.func @sparse_select(%arg0: f64) -> f64 {
553553

554554
// -----
555555

556-
#SparseMatrix = #sparse_tensor.encoding<{lvlTypes = ["compressed", "compressed"]}>
556+
#SparseMatrix = #sparse_tensor.encoding<{map = (d0, d1) -> (d0 : compressed, d1 : compressed)}>
557557

558558
// CHECK-LABEL: func @concat_sparse_sparse(
559559
// CHECK-SAME: %[[A0:.*]]: tensor<2x4xf64
@@ -577,7 +577,7 @@ func.func @concat_sparse_sparse(%arg0: tensor<2x4xf64, #SparseMatrix>,
577577

578578
// -----
579579

580-
#DCSR = #sparse_tensor.encoding<{lvlTypes = ["compressed", "compressed"]}>
580+
#DCSR = #sparse_tensor.encoding<{map = (d0, d1) -> (d0 : compressed, d1 : compressed)}>
581581

582582
// CHECK-LABEL: func @sparse_tensor_foreach(
583583
// CHECK-SAME: %[[A0:.*]]: tensor<2x4xf64
@@ -592,7 +592,7 @@ func.func @sparse_tensor_foreach(%arg0: tensor<2x4xf64, #DCSR>) -> () {
592592

593593
// -----
594594

595-
#DCSR = #sparse_tensor.encoding<{lvlTypes = ["compressed", "compressed"]}>
595+
#DCSR = #sparse_tensor.encoding<{map = (d0, d1) -> (d0 : compressed, d1 : compressed)}>
596596

597597
// CHECK-LABEL: func @sparse_tensor_foreach(
598598
// CHECK-SAME: %[[A0:.*]]: tensor<2x4xf64, #sparse_tensor.encoding<{{{.*}}}>>,

mlir/test/Dialect/SparseTensor/sparse_2d.mlir

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,10 +1,10 @@
11
// NOTE: Assertions have been autogenerated by utils/generate-test-checks.py
22
// RUN: mlir-opt %s -sparsification | FileCheck %s
33

4-
#Tdd = #sparse_tensor.encoding<{ lvlTypes = [ "dense", "dense" ] }>
4+
#Tdd = #sparse_tensor.encoding<{ map = (d0, d1) -> (d0 : dense, d1 : dense) }>
55
#Tds = #sparse_tensor.encoding<{ map = (d0, d1) -> (d0 : dense, d1 : compressed) }>
6-
#Tsd = #sparse_tensor.encoding<{ lvlTypes = [ "compressed", "dense" ] }>
7-
#Tss = #sparse_tensor.encoding<{ lvlTypes = [ "compressed", "compressed" ] }>
6+
#Tsd = #sparse_tensor.encoding<{ map = (d0, d1) -> (d0 : compressed, d1 : dense) }>
7+
#Tss = #sparse_tensor.encoding<{ map = (d0, d1) -> (d0 : compressed, d1 : compressed) }>
88

99
#trait2 = {
1010
indexing_maps = [

mlir/test/Dialect/SparseTensor/sparse_affine.mlir

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33

44
#SpVec = #sparse_tensor.encoding<{ map = (d0) -> (d0 : compressed) }>
55
#CSR = #sparse_tensor.encoding<{ map = (d0, d1) -> (d0 : dense, d1 : compressed) }>
6-
#Row = #sparse_tensor.encoding<{ lvlTypes = [ "compressed", "dense" ] }>
6+
#Row = #sparse_tensor.encoding<{ map = (d0, d1) -> (d0 : compressed, d1 : dense) }>
77
#EncDenseVec = #sparse_tensor.encoding<{ map = (d0) -> (d0 : dense) }>
88

99
#trait1 = {

mlir/test/Dialect/SparseTensor/sparse_broadcast.mlir

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
// RUN: mlir-opt %s --sparsification --canonicalize --cse | FileCheck %s
22

3-
#DCSR = #sparse_tensor.encoding<{ lvlTypes = [ "compressed", "compressed" ] }>
3+
#DCSR = #sparse_tensor.encoding<{ map = (d0, d1) -> (d0 : compressed, d1 : compressed) }>
44
#SparseTensor = #sparse_tensor.encoding<{ lvlTypes = [ "compressed", "compressed", "compressed" ] }>
55

66
#trait = {

mlir/test/Dialect/SparseTensor/sparse_concat.mlir

Lines changed: 3 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -1,16 +1,10 @@
11
// RUN: mlir-opt %s --sparse-tensor-conversion --canonicalize --cse | FileCheck %s
22

3-
#SparseMatrix = #sparse_tensor.encoding<{lvlTypes = ["compressed", "compressed"]}>
3+
#SparseMatrix = #sparse_tensor.encoding<{map = (d0, d1) -> (d0 : compressed, d1 : compressed)}>
44

5-
#SparseMatrix_P = #sparse_tensor.encoding<{
6-
lvlTypes = [ "compressed", "compressed" ],
7-
dimToLvl = affine_map<(i,j) -> (j,i)>
8-
}>
5+
#SparseMatrix_P = #sparse_tensor.encoding<{map = (d0, d1) -> (d1 : compressed, d0 : compressed)}>
96

10-
#SparseMatrix_D_P = #sparse_tensor.encoding<{
11-
lvlTypes = [ "dense", "dense" ],
12-
dimToLvl = affine_map<(i,j) -> (j,i)>
13-
}>
7+
#SparseMatrix_D_P = #sparse_tensor.encoding<{map = (d0, d1) -> (d1 : dense, d0 : dense)}>
148

159
// CHECK-LABEL: func.func @concat_mix_dense(
1610
// CHECK-SAME: %[[TMP_arg0:.*]]: tensor<2x4xf64>,

mlir/test/Dialect/SparseTensor/sparse_concat_codegen.mlir

Lines changed: 3 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -1,12 +1,9 @@
11
// RUN: mlir-opt %s --post-sparsification-rewrite="enable-runtime-library=false enable-convert=false" \
22
// RUN: | FileCheck %s
33

4-
#DCSR = #sparse_tensor.encoding<{lvlTypes = ["compressed", "compressed"]}>
5-
#DENSE = #sparse_tensor.encoding<{lvlTypes = ["dense", "dense"]}>
6-
#DENSE_P = #sparse_tensor.encoding<{
7-
lvlTypes = ["dense", "dense"],
8-
dimToLvl = affine_map<(i,j) -> (j,i)>
9-
}>
4+
#DCSR = #sparse_tensor.encoding<{map = (d0, d1) -> (d0 : compressed, d1 : compressed)}>
5+
#DENSE = #sparse_tensor.encoding<{map = (d0, d1) -> (d0 : dense, d1 : dense)}>
6+
#DENSE_P = #sparse_tensor.encoding<{map = (d0, d1) -> (d1 : dense, d0 : dense)}>
107
// CHECK-LABEL: @concat_sparse_sparse(
118
// CHECK-SAME: %[[TMP_arg0:.*]]: tensor<2x4xf64, #sparse_tensor
129
// CHECK-SAME: %[[TMP_arg1:.*]]: tensor<3x4xf64, #sparse_tensor

mlir/test/Dialect/SparseTensor/sparse_conv_2d_slice_based.mlir

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@
44
#map1 = affine_map<(d0, d1, d2, d3) -> (d2, d3)>
55
#map2 = affine_map<(d0, d1, d2, d3) -> (d0, d1)>
66

7-
#DCSR = #sparse_tensor.encoding<{ lvlTypes = [ "compressed", "compressed" ] }>
7+
#DCSR = #sparse_tensor.encoding<{ map = (d0, d1) -> (d0 : compressed, d1 : compressed) }>
88

99
// CHECK-LABEL: func.func @conv2d_all_sparse_CSR(
1010
// CHECK-SAME: %[[VAL_0:.*]]: tensor<8x8xi32, #sparse_tensor.encoding<{{.*}}>>,

mlir/test/Dialect/SparseTensor/sparse_expand.mlir

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -16,8 +16,7 @@
1616
}>
1717

1818
#DCSC = #sparse_tensor.encoding<{
19-
lvlTypes = [ "compressed", "compressed" ],
20-
dimToLvl = affine_map<(i,j) -> (j,i)>
19+
map = (d0, d1) -> (d1 : compressed, d0 : compressed),
2120
}>
2221

2322
#SV = #sparse_tensor.encoding<{

mlir/test/Dialect/SparseTensor/sparse_fill_zero.mlir

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
// RUN: mlir-opt %s --linalg-generalize-named-ops --pre-sparsification-rewrite --sparsification --sparse-tensor-conversion --canonicalize --cse | FileCheck %s
22

3-
#DCSR = #sparse_tensor.encoding<{ lvlTypes = [ "compressed", "compressed" ] }>
3+
#DCSR = #sparse_tensor.encoding<{ map = (d0, d1) -> (d0 : compressed, d1 : compressed) }>
44

55
// CHECK-LABEL: func.func @fill_zero_after_alloc(
66
// CHECK-SAME: %[[Arg0:.*]]: !llvm.ptr<i8>,

0 commit comments

Comments
 (0)