From 0d1fa02ae88539c368a7f1f44dcc3f1384a8e242 Mon Sep 17 00:00:00 2001 From: Yinying Li Date: Tue, 12 Sep 2023 21:05:10 +0000 Subject: [PATCH 1/3] [mlir][sparse] Migrate tests to use new syntax lvlTypes = [ "compressed" ] to map = (d0) -> (d0 : compressed) lvlTypes = [ "dense" ] to map = (d0) -> (d0 : dense) --- .../SparseTensor/IR/SparseTensorAttrDefs.td | 2 +- mlir/test/Dialect/SparseTensor/codegen.mlir | 4 +-- .../codegen_buffer_initialization.mlir | 2 +- .../Dialect/SparseTensor/codegen_to_llvm.mlir | 2 +- .../SparseTensor/constant_index_map.mlir | 2 +- .../test/Dialect/SparseTensor/conversion.mlir | 6 ++-- .../SparseTensor/convert_dense2sparse.mlir | 2 +- .../SparseTensor/convert_sparse2dense.mlir | 2 +- .../SparseTensor/convert_sparse2sparse.mlir | 6 ++-- .../convert_sparse2sparse_element.mlir | 4 +-- mlir/test/Dialect/SparseTensor/fold.mlir | 2 +- mlir/test/Dialect/SparseTensor/invalid.mlir | 28 +++++++++---------- .../one_shot_bufferize_invalid.mlir | 2 +- ..._shot_bufferize_tensor_copy_insertion.mlir | 2 +- .../Dialect/SparseTensor/post_rewriting.mlir | 2 +- .../Dialect/SparseTensor/pre_rewriting.mlir | 2 +- mlir/test/Dialect/SparseTensor/rejected.mlir | 2 +- mlir/test/Dialect/SparseTensor/roundtrip.mlir | 28 +++++++++---------- .../SparseTensor/scf_1_N_conversion.mlir | 2 +- mlir/test/Dialect/SparseTensor/sparse_1d.mlir | 4 +-- mlir/test/Dialect/SparseTensor/sparse_3d.mlir | 2 +- .../Dialect/SparseTensor/sparse_affine.mlir | 4 +-- .../Dialect/SparseTensor/sparse_expand.mlir | 2 +- .../Dialect/SparseTensor/sparse_fp_ops.mlir | 2 +- .../Dialect/SparseTensor/sparse_fusion.mlir | 2 +- .../Dialect/SparseTensor/sparse_int_ops.mlir | 2 +- .../Dialect/SparseTensor/sparse_kernels.mlir | 2 +- .../Dialect/SparseTensor/sparse_outbuf.mlir | 2 +- .../Dialect/SparseTensor/sparse_reshape.mlir | 2 +- .../Dialect/SparseTensor/sparse_storage.mlir | 4 +-- .../Dialect/SparseTensor/sparse_vector.mlir | 6 ++-- .../SparseTensor/sparse_vector_index.mlir | 2 +- .../SparseTensor/sparse_vector_ops.mlir | 2 +- .../SparseTensor/sparse_vector_peeled.mlir | 2 +- .../SparseTensor/unsparsifiable_dense_op.mlir | 2 +- .../SparseTensor/vectorize_reduction.mlir | 14 +++++----- .../SparseTensor/CPU/dense_output_bf16.mlir | 4 +-- .../SparseTensor/CPU/dense_output_f16.mlir | 4 +-- .../Dialect/SparseTensor/CPU/sparse_abs.mlir | 2 +- .../SparseTensor/CPU/sparse_binary.mlir | 2 +- .../Dialect/SparseTensor/CPU/sparse_cast.mlir | 2 +- .../CPU/sparse_collapse_shape.mlir | 2 +- .../SparseTensor/CPU/sparse_complex32.mlir | 2 +- .../SparseTensor/CPU/sparse_complex64.mlir | 2 +- .../SparseTensor/CPU/sparse_complex_ops.mlir | 2 +- .../Dialect/SparseTensor/CPU/sparse_dot.mlir | 2 +- .../SparseTensor/CPU/sparse_expand_shape.mlir | 2 +- .../SparseTensor/CPU/sparse_index.mlir | 2 +- .../SparseTensor/CPU/sparse_index_dense.mlir | 2 +- .../SparseTensor/CPU/sparse_insert_1d.mlir | 2 +- .../SparseTensor/CPU/sparse_re_im.mlir | 2 +- .../CPU/sparse_reduce_custom.mlir | 2 +- .../CPU/sparse_reduce_custom_prod.mlir | 2 +- .../CPU/sparse_reduce_custom_sum.mlir | 2 +- .../SparseTensor/CPU/sparse_reductions.mlir | 2 +- .../CPU/sparse_reductions_min.mlir | 2 +- .../CPU/sparse_reductions_prod.mlir | 4 +-- .../SparseTensor/CPU/sparse_reshape.mlir | 2 +- .../SparseTensor/CPU/sparse_select.mlir | 2 +- .../Dialect/SparseTensor/CPU/sparse_sign.mlir | 2 +- .../Dialect/SparseTensor/CPU/sparse_tanh.mlir | 2 +- .../SparseTensor/CPU/sparse_unary.mlir | 2 +- .../SparseTensor/CPU/sparse_vector_ops.mlir | 4 +-- .../python/dialects/sparse_tensor/dialect.py | 4 +-- 64 files changed, 112 insertions(+), 112 deletions(-) diff --git a/mlir/include/mlir/Dialect/SparseTensor/IR/SparseTensorAttrDefs.td b/mlir/include/mlir/Dialect/SparseTensor/IR/SparseTensorAttrDefs.td index e2f3df005b70d..8e2de77e6850a 100644 --- a/mlir/include/mlir/Dialect/SparseTensor/IR/SparseTensorAttrDefs.td +++ b/mlir/include/mlir/Dialect/SparseTensor/IR/SparseTensorAttrDefs.td @@ -194,7 +194,7 @@ def SparseTensorEncodingAttr : SparseTensor_Attr<"SparseTensorEncoding", ```mlir // Sparse vector. #SparseVector = #sparse_tensor.encoding<{ - lvlTypes = [ "compressed" ] + map = (d0) -> (d0 : compressed) }> ... tensor ... diff --git a/mlir/test/Dialect/SparseTensor/codegen.mlir b/mlir/test/Dialect/SparseTensor/codegen.mlir index cf8b1ba87d303..5155e5ce6c454 100644 --- a/mlir/test/Dialect/SparseTensor/codegen.mlir +++ b/mlir/test/Dialect/SparseTensor/codegen.mlir @@ -1,9 +1,9 @@ // RUN: mlir-opt %s --sparse-tensor-codegen --canonicalize -cse | FileCheck %s -#SV = #sparse_tensor.encoding<{ lvlTypes = [ "compressed" ] }> +#SV = #sparse_tensor.encoding<{ map = (d0) -> (d0 : compressed) }> #SparseVector = #sparse_tensor.encoding<{ - lvlTypes = [ "compressed" ], + map = (d0) -> (d0 : compressed), crdWidth = 64, posWidth = 32 }> diff --git a/mlir/test/Dialect/SparseTensor/codegen_buffer_initialization.mlir b/mlir/test/Dialect/SparseTensor/codegen_buffer_initialization.mlir index 0a338064eff32..640d0f56a0f94 100644 --- a/mlir/test/Dialect/SparseTensor/codegen_buffer_initialization.mlir +++ b/mlir/test/Dialect/SparseTensor/codegen_buffer_initialization.mlir @@ -1,6 +1,6 @@ // RUN: mlir-opt %s --sparse-tensor-codegen=enable-buffer-initialization=true --canonicalize --cse | FileCheck %s -#SV = #sparse_tensor.encoding<{ lvlTypes = [ "compressed" ] }> +#SV = #sparse_tensor.encoding<{ map = (d0) -> (d0 : compressed) }> // CHECK-LABEL: func.func @sparse_alloc_sparse_vector( // CHECK-SAME: %[[VAL_0:.*]]: index) -> (memref, memref, memref, !sparse_tensor.storage_specifier diff --git a/mlir/test/Dialect/SparseTensor/codegen_to_llvm.mlir b/mlir/test/Dialect/SparseTensor/codegen_to_llvm.mlir index 99b8759806544..b0f7c62ef283a 100644 --- a/mlir/test/Dialect/SparseTensor/codegen_to_llvm.mlir +++ b/mlir/test/Dialect/SparseTensor/codegen_to_llvm.mlir @@ -1,6 +1,6 @@ // RUN: mlir-opt %s --sparse-tensor-codegen --sparse-storage-specifier-to-llvm | FileCheck %s -#SparseVector = #sparse_tensor.encoding<{ lvlTypes = ["compressed"] }> +#SparseVector = #sparse_tensor.encoding<{ map = (d0) -> (d0 : compressed) }> // CHECK-LABEL: func @sparse_nop( // CHECK-SAME: %[[A0:.*0]]: memref, diff --git a/mlir/test/Dialect/SparseTensor/constant_index_map.mlir b/mlir/test/Dialect/SparseTensor/constant_index_map.mlir index 532b95507d548..bfb4503edbc4e 100644 --- a/mlir/test/Dialect/SparseTensor/constant_index_map.mlir +++ b/mlir/test/Dialect/SparseTensor/constant_index_map.mlir @@ -5,7 +5,7 @@ #map1 = affine_map<(d0) -> (0, d0)> #map2 = affine_map<(d0) -> (d0)> -#SpVec = #sparse_tensor.encoding<{ lvlTypes = [ "compressed" ] }> +#SpVec = #sparse_tensor.encoding<{ map = (d0) -> (d0 : compressed) }> // CHECK-LABEL: func.func @main( // CHECK-SAME: %[[VAL_0:.*0]]: tensor<1x77xi1>, diff --git a/mlir/test/Dialect/SparseTensor/conversion.mlir b/mlir/test/Dialect/SparseTensor/conversion.mlir index aa432460173cf..ae9e312de7f27 100644 --- a/mlir/test/Dialect/SparseTensor/conversion.mlir +++ b/mlir/test/Dialect/SparseTensor/conversion.mlir @@ -1,17 +1,17 @@ // RUN: mlir-opt %s --sparse-tensor-conversion --canonicalize --cse | FileCheck %s #SparseVector = #sparse_tensor.encoding<{ - lvlTypes = ["compressed"] + map = (d0) -> (d0 : compressed) }> #SparseVector64 = #sparse_tensor.encoding<{ - lvlTypes = ["compressed"], + map = (d0) -> (d0 : compressed), posWidth = 64, crdWidth = 64 }> #SparseVector32 = #sparse_tensor.encoding<{ - lvlTypes = ["compressed"], + map = (d0) -> (d0 : compressed), posWidth = 32, crdWidth = 32 }> diff --git a/mlir/test/Dialect/SparseTensor/convert_dense2sparse.mlir b/mlir/test/Dialect/SparseTensor/convert_dense2sparse.mlir index ac9a613134ed5..f2ac0c22e035e 100644 --- a/mlir/test/Dialect/SparseTensor/convert_dense2sparse.mlir +++ b/mlir/test/Dialect/SparseTensor/convert_dense2sparse.mlir @@ -3,7 +3,7 @@ // RUN: --canonicalize --cse | FileCheck %s --check-prefix=CHECK-RWT #SparseVector = #sparse_tensor.encoding<{ - lvlTypes = ["compressed"] + map = (d0) -> (d0 : compressed) }> #CSR = #sparse_tensor.encoding<{ diff --git a/mlir/test/Dialect/SparseTensor/convert_sparse2dense.mlir b/mlir/test/Dialect/SparseTensor/convert_sparse2dense.mlir index 1adc9f9566da3..7328ede697d96 100644 --- a/mlir/test/Dialect/SparseTensor/convert_sparse2dense.mlir +++ b/mlir/test/Dialect/SparseTensor/convert_sparse2dense.mlir @@ -4,7 +4,7 @@ // RUN: --canonicalize --cse | FileCheck %s --check-prefix=CHECK-RWT #SparseVector = #sparse_tensor.encoding<{ - lvlTypes = ["compressed"] + map = (d0) -> (d0 : compressed) }> #SparseMatrix = #sparse_tensor.encoding<{ diff --git a/mlir/test/Dialect/SparseTensor/convert_sparse2sparse.mlir b/mlir/test/Dialect/SparseTensor/convert_sparse2sparse.mlir index a0435bd9b0edc..296e1bf9030c6 100644 --- a/mlir/test/Dialect/SparseTensor/convert_sparse2sparse.mlir +++ b/mlir/test/Dialect/SparseTensor/convert_sparse2sparse.mlir @@ -10,19 +10,19 @@ // RUN: --canonicalize --cse | FileCheck %s --check-prefix=CHECK-RWT #SparseVector64 = #sparse_tensor.encoding<{ - lvlTypes = ["compressed"], + map = (d0) -> (d0 : compressed), posWidth = 64, crdWidth = 64 }> #SparseVector32 = #sparse_tensor.encoding<{ - lvlTypes = ["compressed"], + map = (d0) -> (d0 : compressed), posWidth = 32, crdWidth = 32 }> #SparseVector = #sparse_tensor.encoding<{ - lvlTypes = ["compressed"] + map = (d0) -> (d0 : compressed) }> #SortedCOO2D = #sparse_tensor.encoding<{ diff --git a/mlir/test/Dialect/SparseTensor/convert_sparse2sparse_element.mlir b/mlir/test/Dialect/SparseTensor/convert_sparse2sparse_element.mlir index 6d20cc01a4eb7..19a789ea2449a 100644 --- a/mlir/test/Dialect/SparseTensor/convert_sparse2sparse_element.mlir +++ b/mlir/test/Dialect/SparseTensor/convert_sparse2sparse_element.mlir @@ -1,13 +1,13 @@ // RUN: mlir-opt %s --sparse-tensor-codegen --canonicalize --cse | FileCheck %s #SparseVector64 = #sparse_tensor.encoding<{ - lvlTypes = ["compressed"], + map = (d0) -> (d0 : compressed), posWidth = 64, crdWidth = 64 }> #SparseVector32 = #sparse_tensor.encoding<{ - lvlTypes = ["compressed"], + map = (d0) -> (d0 : compressed), posWidth = 32, crdWidth = 32 }> diff --git a/mlir/test/Dialect/SparseTensor/fold.mlir b/mlir/test/Dialect/SparseTensor/fold.mlir index 44eb8ac1fb64b..089431f9e18e9 100644 --- a/mlir/test/Dialect/SparseTensor/fold.mlir +++ b/mlir/test/Dialect/SparseTensor/fold.mlir @@ -1,6 +1,6 @@ // RUN: mlir-opt %s --canonicalize --cse | FileCheck %s -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> // CHECK-LABEL: func @sparse_nop_dense2dense_convert( // CHECK-SAME: %[[A:.*]]: tensor<64xf32>) diff --git a/mlir/test/Dialect/SparseTensor/invalid.mlir b/mlir/test/Dialect/SparseTensor/invalid.mlir index 4849a1f25da6c..360dfcce2ef2b 100644 --- a/mlir/test/Dialect/SparseTensor/invalid.mlir +++ b/mlir/test/Dialect/SparseTensor/invalid.mlir @@ -8,7 +8,7 @@ func.func @invalid_new_dense(%arg0: !llvm.ptr) -> tensor<32xf32> { // ----- -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"], posWidth=32, crdWidth=32}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed), posWidth=32, crdWidth=32}> func.func @non_static_pack_ret(%values: tensor<6xf64>, %pos: tensor<2xi32>, %coordinates: tensor<6x1xi32>) -> tensor { @@ -20,7 +20,7 @@ func.func @non_static_pack_ret(%values: tensor<6xf64>, %pos: tensor<2xi32>, %coo // ----- -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"], posWidth=32, crdWidth=32}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed), posWidth=32, crdWidth=32}> func.func @invalid_pack_type(%values: tensor<6xf64>, %pos: tensor<2xi32>, %coordinates: tensor<6x1xi32>) -> tensor<100xf32, #SparseVector> { @@ -56,7 +56,7 @@ func.func @invalid_pack_mis_position(%values: tensor<6xf64>, %coordinates: tenso // ----- -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"], posWidth=32, crdWidth=32}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed), posWidth=32, crdWidth=32}> func.func @invalid_unpack_type(%sp: tensor<100xf32, #SparseVector>, %values: tensor<6xf64>, %pos: tensor<2xi32>, %coordinates: tensor<6x1xi32>) { // expected-error@+1 {{input/output element-types don't match}} @@ -108,7 +108,7 @@ func.func @invalid_positions_unranked(%arg0: tensor<*xf64>) -> memref { // ----- -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"], posWidth=32}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed), posWidth=32}> func.func @mismatch_positions_types(%arg0: tensor<128xf64, #SparseVector>) -> memref { // expected-error@+1 {{unexpected type for positions}} @@ -118,7 +118,7 @@ func.func @mismatch_positions_types(%arg0: tensor<128xf64, #SparseVector>) -> me // ----- -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> func.func @positions_oob(%arg0: tensor<128xf64, #SparseVector>) -> memref { // expected-error@+1 {{requested level is out of bounds}} @@ -144,7 +144,7 @@ func.func @invalid_indices_unranked(%arg0: tensor<*xf64>) -> memref { // ----- -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> func.func @mismatch_indices_types(%arg0: tensor) -> memref { // expected-error@+1 {{unexpected type for coordinates}} @@ -154,7 +154,7 @@ func.func @mismatch_indices_types(%arg0: tensor) -> memref // ----- -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> func.func @indices_oob(%arg0: tensor<128xf64, #SparseVector>) -> memref { // expected-error@+1 {{requested level is out of bounds}} @@ -172,7 +172,7 @@ func.func @invalid_values_dense(%arg0: tensor<1024xf32>) -> memref { // ----- -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> func.func @indices_buffer_noncoo(%arg0: tensor<128xf64, #SparseVector>) -> memref { // expected-error@+1 {{expected sparse tensor with a COO region}} @@ -190,7 +190,7 @@ func.func @indices_buffer_dense(%arg0: tensor<1024xf32>) -> memref { // ----- -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> func.func @mismatch_values_types(%arg0: tensor) -> memref { // expected-error@+1 {{unexpected mismatch in element types}} @@ -226,7 +226,7 @@ func.func @sparse_slice_stride(%arg0: tensor<2x8xf64, #CSR_SLICE>) -> index { // ----- -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> func.func @sparse_get_md(%arg0: !sparse_tensor.storage_specifier<#SparseVector>) -> index { // expected-error@+1 {{redundant level argument for querying value memory size}} @@ -237,7 +237,7 @@ func.func @sparse_get_md(%arg0: !sparse_tensor.storage_specifier<#SparseVector>) // ----- -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> func.func @sparse_get_md(%arg0: !sparse_tensor.storage_specifier<#SparseVector>) -> i64 { // expected-error@+1 {{requested slice data on non-slice tensor}} @@ -248,7 +248,7 @@ func.func @sparse_get_md(%arg0: !sparse_tensor.storage_specifier<#SparseVector>) // ----- -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> func.func @sparse_get_md(%arg0: !sparse_tensor.storage_specifier<#SparseVector>) -> index { // expected-error@+1 {{missing level argument}} @@ -259,7 +259,7 @@ func.func @sparse_get_md(%arg0: !sparse_tensor.storage_specifier<#SparseVector>) // ----- -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> func.func @sparse_get_md(%arg0: !sparse_tensor.storage_specifier<#SparseVector>) -> index { // expected-error@+1 {{requested level is out of bounds}} @@ -656,7 +656,7 @@ func.func @invalid_concat_dim(%arg0: tensor<2x4xf64, #DC>, // ----- -#C = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#C = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> #DC = #sparse_tensor.encoding<{lvlTypes = ["dense", "compressed"]}> #DCC = #sparse_tensor.encoding<{lvlTypes = ["dense", "compressed", "compressed"]}> func.func @invalid_concat_rank_mismatch(%arg0: tensor<2xf64, #C>, diff --git a/mlir/test/Dialect/SparseTensor/one_shot_bufferize_invalid.mlir b/mlir/test/Dialect/SparseTensor/one_shot_bufferize_invalid.mlir index 25ecd20c38003..1540d1876d7f0 100644 --- a/mlir/test/Dialect/SparseTensor/one_shot_bufferize_invalid.mlir +++ b/mlir/test/Dialect/SparseTensor/one_shot_bufferize_invalid.mlir @@ -1,7 +1,7 @@ // RUN: mlir-opt %s -one-shot-bufferize -verify-diagnostics #SparseVector = #sparse_tensor.encoding<{ - lvlTypes = ["compressed"] + map = (d0) -> (d0 : compressed) }> func.func @sparse_tensor_op(%arg0: tensor<64xf32, #SparseVector>) -> tensor<64xf32, #SparseVector> { diff --git a/mlir/test/Dialect/SparseTensor/one_shot_bufferize_tensor_copy_insertion.mlir b/mlir/test/Dialect/SparseTensor/one_shot_bufferize_tensor_copy_insertion.mlir index 93bcfe0d0cefb..0ccce5121ce1a 100644 --- a/mlir/test/Dialect/SparseTensor/one_shot_bufferize_tensor_copy_insertion.mlir +++ b/mlir/test/Dialect/SparseTensor/one_shot_bufferize_tensor_copy_insertion.mlir @@ -41,7 +41,7 @@ func.func @sparse_tensor_convert() -> tensor<20x40xf32> { return %2 : tensor<20x40xf32> } -#SV = #sparse_tensor.encoding<{ lvlTypes = [ "compressed" ] }> +#SV = #sparse_tensor.encoding<{ map = (d0) -> (d0 : compressed) }> #trait = { indexing_maps = [ diff --git a/mlir/test/Dialect/SparseTensor/post_rewriting.mlir b/mlir/test/Dialect/SparseTensor/post_rewriting.mlir index 3a0c57c71fe0c..ab334496aaad5 100644 --- a/mlir/test/Dialect/SparseTensor/post_rewriting.mlir +++ b/mlir/test/Dialect/SparseTensor/post_rewriting.mlir @@ -1,7 +1,7 @@ // RUN: mlir-opt %s -post-sparsification-rewrite | FileCheck %s #SparseVector = #sparse_tensor.encoding<{ - lvlTypes = ["compressed"] + map = (d0) -> (d0 : compressed) }> #SparseMatrix = #sparse_tensor.encoding<{ diff --git a/mlir/test/Dialect/SparseTensor/pre_rewriting.mlir b/mlir/test/Dialect/SparseTensor/pre_rewriting.mlir index de63043ea5cf4..0c5f32b0b5510 100644 --- a/mlir/test/Dialect/SparseTensor/pre_rewriting.mlir +++ b/mlir/test/Dialect/SparseTensor/pre_rewriting.mlir @@ -1,7 +1,7 @@ // RUN: mlir-opt %s -pre-sparsification-rewrite | FileCheck %s #SparseVector = #sparse_tensor.encoding<{ - lvlTypes = ["compressed"] + map = (d0) -> (d0 : compressed) }> #SortedCOO = #sparse_tensor.encoding<{ diff --git a/mlir/test/Dialect/SparseTensor/rejected.mlir b/mlir/test/Dialect/SparseTensor/rejected.mlir index 5cc676e54a0a1..3e3af15fc1d44 100644 --- a/mlir/test/Dialect/SparseTensor/rejected.mlir +++ b/mlir/test/Dialect/SparseTensor/rejected.mlir @@ -3,7 +3,7 @@ // The file contains examples that will be rejected by sparse compiler // (we expect the linalg.generic unchanged). -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> #trait = { indexing_maps = [ diff --git a/mlir/test/Dialect/SparseTensor/roundtrip.mlir b/mlir/test/Dialect/SparseTensor/roundtrip.mlir index bd4baf57d8ba7..cb178e4257b1c 100644 --- a/mlir/test/Dialect/SparseTensor/roundtrip.mlir +++ b/mlir/test/Dialect/SparseTensor/roundtrip.mlir @@ -1,6 +1,6 @@ // RUN: mlir-opt %s -split-input-file | mlir-opt -split-input-file | FileCheck %s -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> // CHECK-LABEL: func @sparse_new( // CHECK-SAME: %[[A:.*]]: !llvm.ptr) @@ -13,7 +13,7 @@ func.func @sparse_new(%arg0: !llvm.ptr) -> tensor<128xf64, #SparseVector> { // ----- -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"], posWidth=32, crdWidth=32}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed), posWidth=32, crdWidth=32}> // CHECK-LABEL: func @sparse_pack( // CHECK-SAME: %[[D:.*]]: tensor<6xf64>, @@ -30,7 +30,7 @@ func.func @sparse_pack(%data: tensor<6xf64>, %pos: tensor<2xi32>, %index: tensor // ----- -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"], crdWidth=32}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed), crdWidth=32}> // CHECK-LABEL: func @sparse_unpack( // CHECK-SAME: %[[T:.*]]: tensor<100xf64, # // CHECK-SAME: %[[OD:.*]]: tensor<6xf64> @@ -51,7 +51,7 @@ func.func @sparse_unpack(%sp : tensor<100xf64, #SparseVector>, // ----- -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> // CHECK-LABEL: func @sparse_dealloc( // CHECK-SAME: %[[A:.*]]: tensor<128xf64, #{{.*}}> @@ -64,7 +64,7 @@ func.func @sparse_dealloc(%arg0: tensor<128xf64, #SparseVector>) { // ----- -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> // CHECK-LABEL: func @sparse_convert_1d_to_sparse( // CHECK-SAME: %[[A:.*]]: tensor<64xf32>) @@ -90,7 +90,7 @@ func.func @sparse_convert_3d_from_sparse(%arg0: tensor<8x8x8xf64, #SparseTensor> // ----- -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> // CHECK-LABEL: func @sparse_positions( // CHECK-SAME: %[[A:.*]]: tensor<128xf64, #{{.*}}>) @@ -116,7 +116,7 @@ func.func @sparse_indices_buffer(%arg0: tensor) -> memref +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> // CHECK-LABEL: func @sparse_indices( // CHECK-SAME: %[[A:.*]]: tensor<128xf64, #{{.*}}>) @@ -129,7 +129,7 @@ func.func @sparse_indices(%arg0: tensor<128xf64, #SparseVector>) -> memref +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> // CHECK-LABEL: func @sparse_values( // CHECK-SAME: %[[A:.*]]: tensor<128xf64, #{{.*}}>) @@ -174,7 +174,7 @@ func.func @sparse_slice_stride(%arg0: tensor<2x8xf64, #CSR_SLICE>) -> index { // ----- -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> // CHECK-LABEL: func @sparse_metadata_init( // CHECK: %[[T:.*]] = sparse_tensor.storage_specifier.init : !sparse_tensor.storage_specifier<#{{.*}}> @@ -186,7 +186,7 @@ func.func @sparse_metadata_init() -> !sparse_tensor.storage_specifier<#SparseVec // ----- -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> #SparseVector_Slice = #sparse_tensor.encoding<{ lvlTypes = ["compressed"], dimSlices = [ (?, ?, ?) ] @@ -205,7 +205,7 @@ func.func @sparse_metadata_init(%src : !sparse_tensor.storage_specifier<#SparseV // ----- -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> // CHECK-LABEL: func @sparse_get_md( // CHECK-SAME: %[[A:.*]]: !sparse_tensor.storage_specifier<#{{.*}}> @@ -254,7 +254,7 @@ func.func @sparse_get_md(%arg0: !sparse_tensor.storage_specifier<#SparseVector>) // ----- -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> // CHECK-LABEL: func @sparse_set_md( // CHECK-SAME: %[[A:.*]]: !sparse_tensor.storage_specifier<#{{.*}}>, @@ -270,7 +270,7 @@ func.func @sparse_set_md(%arg0: !sparse_tensor.storage_specifier<#SparseVector>, // ----- -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> // CHECK-LABEL: func @sparse_noe( // CHECK-SAME: %[[A:.*]]: tensor<128xf64, #{{.*}}>) @@ -309,7 +309,7 @@ func.func @sparse_load_ins(%arg0: tensor<16x32xf64, #DenseMatrix>) -> tensor<16x // ----- -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> // CHECK-LABEL: func @sparse_insert( // CHECK-SAME: %[[A:.*]]: tensor<128xf64, #sparse_tensor.encoding<{{.*}}>>, diff --git a/mlir/test/Dialect/SparseTensor/scf_1_N_conversion.mlir b/mlir/test/Dialect/SparseTensor/scf_1_N_conversion.mlir index fb29c315bea25..f5d6a08b7de31 100644 --- a/mlir/test/Dialect/SparseTensor/scf_1_N_conversion.mlir +++ b/mlir/test/Dialect/SparseTensor/scf_1_N_conversion.mlir @@ -1,6 +1,6 @@ // RUN: mlir-opt %s -sparse-tensor-codegen -cse | FileCheck %s -#SparseVector = #sparse_tensor.encoding<{ lvlTypes = [ "compressed" ] }> +#SparseVector = #sparse_tensor.encoding<{ map = (d0) -> (d0 : compressed) }> // CHECK-LABEL: func.func @for( // CHECK-SAME: %[[VAL_1:.*0]]: memref, diff --git a/mlir/test/Dialect/SparseTensor/sparse_1d.mlir b/mlir/test/Dialect/SparseTensor/sparse_1d.mlir index 8bbb35add2618..94c4513f8b5ec 100644 --- a/mlir/test/Dialect/SparseTensor/sparse_1d.mlir +++ b/mlir/test/Dialect/SparseTensor/sparse_1d.mlir @@ -1,8 +1,8 @@ // NOTE: Assertions have been autogenerated by utils/generate-test-checks.py // RUN: mlir-opt %s -sparsification | FileCheck %s -#DV = #sparse_tensor.encoding<{ lvlTypes = [ "dense" ] }> -#SV = #sparse_tensor.encoding<{ lvlTypes = [ "compressed" ] }> +#DV = #sparse_tensor.encoding<{ map = (d0) -> (d0 : dense) }> +#SV = #sparse_tensor.encoding<{ map = (d0) -> (d0 : compressed) }> #trait1 = { indexing_maps = [ diff --git a/mlir/test/Dialect/SparseTensor/sparse_3d.mlir b/mlir/test/Dialect/SparseTensor/sparse_3d.mlir index b6b4093ac463a..9019b9984d8f6 100644 --- a/mlir/test/Dialect/SparseTensor/sparse_3d.mlir +++ b/mlir/test/Dialect/SparseTensor/sparse_3d.mlir @@ -1,7 +1,7 @@ // NOTE: Assertions have been autogenerated by utils/generate-test-checks.py // RUN: mlir-opt %s -sparsification | FileCheck %s -#Td = #sparse_tensor.encoding<{ lvlTypes = [ "dense" ] }> +#Td = #sparse_tensor.encoding<{ map = (d0) -> (d0 : dense) }> #Tddd = #sparse_tensor.encoding<{ lvlTypes = [ "dense", "dense", "dense" ] }> #Tdds = #sparse_tensor.encoding<{ lvlTypes = [ "dense", "dense", "compressed" ] }> diff --git a/mlir/test/Dialect/SparseTensor/sparse_affine.mlir b/mlir/test/Dialect/SparseTensor/sparse_affine.mlir index 0d224bea9229d..3a58c5ea9c17b 100644 --- a/mlir/test/Dialect/SparseTensor/sparse_affine.mlir +++ b/mlir/test/Dialect/SparseTensor/sparse_affine.mlir @@ -1,10 +1,10 @@ // NOTE: Assertions have been autogenerated by utils/generate-test-checks.py // RUN: mlir-opt %s -sparsification | FileCheck %s -#SpVec = #sparse_tensor.encoding<{ lvlTypes = [ "compressed" ] }> +#SpVec = #sparse_tensor.encoding<{ map = (d0) -> (d0 : compressed) }> #CSR = #sparse_tensor.encoding<{ lvlTypes = [ "dense", "compressed" ] }> #Row = #sparse_tensor.encoding<{ lvlTypes = [ "compressed", "dense" ] }> -#EncDenseVec = #sparse_tensor.encoding<{ lvlTypes = [ "dense" ] }> +#EncDenseVec = #sparse_tensor.encoding<{ map = (d0) -> (d0 : dense) }> #trait1 = { indexing_maps = [ diff --git a/mlir/test/Dialect/SparseTensor/sparse_expand.mlir b/mlir/test/Dialect/SparseTensor/sparse_expand.mlir index d5ba7ecca6c9a..98dfb4b88d6d3 100644 --- a/mlir/test/Dialect/SparseTensor/sparse_expand.mlir +++ b/mlir/test/Dialect/SparseTensor/sparse_expand.mlir @@ -22,7 +22,7 @@ }> #SV = #sparse_tensor.encoding<{ - lvlTypes = [ "compressed" ] + map = (d0) -> (d0 : compressed) }> #rowsum = { diff --git a/mlir/test/Dialect/SparseTensor/sparse_fp_ops.mlir b/mlir/test/Dialect/SparseTensor/sparse_fp_ops.mlir index 5cb82cb30cd16..3e18b0c1b6c1b 100644 --- a/mlir/test/Dialect/SparseTensor/sparse_fp_ops.mlir +++ b/mlir/test/Dialect/SparseTensor/sparse_fp_ops.mlir @@ -1,6 +1,6 @@ // RUN: mlir-opt %s -sparsification | FileCheck %s -#SV = #sparse_tensor.encoding<{ lvlTypes = [ "compressed" ] }> +#SV = #sparse_tensor.encoding<{ map = (d0) -> (d0 : compressed) }> #trait1 = { indexing_maps = [ diff --git a/mlir/test/Dialect/SparseTensor/sparse_fusion.mlir b/mlir/test/Dialect/SparseTensor/sparse_fusion.mlir index 49af2b85f2fa6..8780baac199e1 100644 --- a/mlir/test/Dialect/SparseTensor/sparse_fusion.mlir +++ b/mlir/test/Dialect/SparseTensor/sparse_fusion.mlir @@ -1,6 +1,6 @@ // RUN: mlir-opt %s --linalg-fuse-elementwise-ops | FileCheck %s -#SV = #sparse_tensor.encoding<{ lvlTypes = ["compressed"] }> +#SV = #sparse_tensor.encoding<{ map = (d0) -> (d0 : compressed) }> #trait = { indexing_maps = [ diff --git a/mlir/test/Dialect/SparseTensor/sparse_int_ops.mlir b/mlir/test/Dialect/SparseTensor/sparse_int_ops.mlir index 1be3a1441cd40..0bc3962783576 100644 --- a/mlir/test/Dialect/SparseTensor/sparse_int_ops.mlir +++ b/mlir/test/Dialect/SparseTensor/sparse_int_ops.mlir @@ -1,7 +1,7 @@ // NOTE: Assertions have been autogenerated by utils/generate-test-checks.py // RUN: mlir-opt %s -sparsification | FileCheck %s -#SV = #sparse_tensor.encoding<{ lvlTypes = [ "compressed" ] }> +#SV = #sparse_tensor.encoding<{ map = (d0) -> (d0 : compressed) }> #trait2 = { indexing_maps = [ diff --git a/mlir/test/Dialect/SparseTensor/sparse_kernels.mlir b/mlir/test/Dialect/SparseTensor/sparse_kernels.mlir index 1ecdc6ff58137..8fd19c53249ec 100644 --- a/mlir/test/Dialect/SparseTensor/sparse_kernels.mlir +++ b/mlir/test/Dialect/SparseTensor/sparse_kernels.mlir @@ -2,7 +2,7 @@ // RUN: --linalg-generalize-named-ops --linalg-fuse-elementwise-ops \ // RUN: --sparsification | FileCheck %s -#SparseVector = #sparse_tensor.encoding<{ lvlTypes = [ "compressed" ] }> +#SparseVector = #sparse_tensor.encoding<{ map = (d0) -> (d0 : compressed) }> #DCSR = #sparse_tensor.encoding<{ lvlTypes = [ "compressed", "compressed" ] }> diff --git a/mlir/test/Dialect/SparseTensor/sparse_outbuf.mlir b/mlir/test/Dialect/SparseTensor/sparse_outbuf.mlir index afff4ea586de1..d6677e0a8ff15 100644 --- a/mlir/test/Dialect/SparseTensor/sparse_outbuf.mlir +++ b/mlir/test/Dialect/SparseTensor/sparse_outbuf.mlir @@ -1,6 +1,6 @@ // RUN: mlir-opt %s -sparsification | FileCheck %s -#SV = #sparse_tensor.encoding<{ lvlTypes = [ "compressed" ] }> +#SV = #sparse_tensor.encoding<{ map = (d0) -> (d0 : compressed) }> #trait = { indexing_maps = [ diff --git a/mlir/test/Dialect/SparseTensor/sparse_reshape.mlir b/mlir/test/Dialect/SparseTensor/sparse_reshape.mlir index 704a2b2bc64c5..f0e16189a94d0 100644 --- a/mlir/test/Dialect/SparseTensor/sparse_reshape.mlir +++ b/mlir/test/Dialect/SparseTensor/sparse_reshape.mlir @@ -3,7 +3,7 @@ // RUN: mlir-opt %s --post-sparsification-rewrite="enable-runtime-library=false enable-convert=false" \ // RUN: --cse --canonicalize | FileCheck %s --check-prefix=CHECK-RWT -#SparseVector = #sparse_tensor.encoding<{ lvlTypes = [ "compressed" ] }> +#SparseVector = #sparse_tensor.encoding<{ map = (d0) -> (d0 : compressed) }> #SparseMatrix = #sparse_tensor.encoding<{ lvlTypes = [ "compressed", "compressed" ] }> // diff --git a/mlir/test/Dialect/SparseTensor/sparse_storage.mlir b/mlir/test/Dialect/SparseTensor/sparse_storage.mlir index 0b55ed5160206..65ea2d1312779 100644 --- a/mlir/test/Dialect/SparseTensor/sparse_storage.mlir +++ b/mlir/test/Dialect/SparseTensor/sparse_storage.mlir @@ -1,13 +1,13 @@ // RUN: mlir-opt %s -sparsification= | FileCheck %s #SparseVector64 = #sparse_tensor.encoding<{ - lvlTypes = [ "compressed" ], + map = (d0) -> (d0 : compressed), posWidth = 64, crdWidth = 64 }> #SparseVector32 = #sparse_tensor.encoding<{ - lvlTypes = [ "compressed" ], + map = (d0) -> (d0 : compressed), posWidth = 32, crdWidth = 32 }> diff --git a/mlir/test/Dialect/SparseTensor/sparse_vector.mlir b/mlir/test/Dialect/SparseTensor/sparse_vector.mlir index 4399123ec842a..a438f9442fe20 100755 --- a/mlir/test/Dialect/SparseTensor/sparse_vector.mlir +++ b/mlir/test/Dialect/SparseTensor/sparse_vector.mlir @@ -7,7 +7,7 @@ // RUN: mlir-opt %s -sparsification -cse -sparse-vectorization="vl=4 enable-vla-vectorization=true" -cse -split-input-file | \ // RUN: FileCheck %s --check-prefix=CHECK-VEC4-SVE -#DenseVector = #sparse_tensor.encoding<{ lvlTypes = [ "dense" ] }> +#DenseVector = #sparse_tensor.encoding<{ map = (d0) -> (d0 : dense) }> #trait_scale_d = { indexing_maps = [ @@ -86,7 +86,7 @@ func.func @scale_d(%arga: tensor<1024xf32, #DenseVector>, %b: f32, %argx: tensor // ----- #SparseVector = #sparse_tensor.encoding<{ - lvlTypes = [ "compressed" ], + map = (d0) -> (d0 : compressed), posWidth = 32, crdWidth = 32 }> @@ -209,7 +209,7 @@ func.func @mul_s(%arga: tensor<1024xf32, #SparseVector>, // ----- -#DenseVector = #sparse_tensor.encoding<{ lvlTypes = [ "dense" ] }> +#DenseVector = #sparse_tensor.encoding<{ map = (d0) -> (d0 : dense) }> #trait_reduction_d = { indexing_maps = [ diff --git a/mlir/test/Dialect/SparseTensor/sparse_vector_index.mlir b/mlir/test/Dialect/SparseTensor/sparse_vector_index.mlir index 97227fd9e40d7..7101f294705aa 100644 --- a/mlir/test/Dialect/SparseTensor/sparse_vector_index.mlir +++ b/mlir/test/Dialect/SparseTensor/sparse_vector_index.mlir @@ -4,7 +4,7 @@ // NOTE: Assertions have been autogenerated by utils/generate-test-checks.py #SparseVector = #sparse_tensor.encoding<{ - lvlTypes = ["compressed"] + map = (d0) -> (d0 : compressed) }> #trait_1d = { diff --git a/mlir/test/Dialect/SparseTensor/sparse_vector_ops.mlir b/mlir/test/Dialect/SparseTensor/sparse_vector_ops.mlir index 1c474735d0af1..8e23f901da765 100644 --- a/mlir/test/Dialect/SparseTensor/sparse_vector_ops.mlir +++ b/mlir/test/Dialect/SparseTensor/sparse_vector_ops.mlir @@ -1,7 +1,7 @@ // RUN: mlir-opt %s -sparsification -cse -sparse-vectorization="vl=8" -cse | \ // RUN: FileCheck %s -#DenseVector = #sparse_tensor.encoding<{ lvlTypes = [ "dense" ] }> +#DenseVector = #sparse_tensor.encoding<{ map = (d0) -> (d0 : dense) }> #trait = { indexing_maps = [ diff --git a/mlir/test/Dialect/SparseTensor/sparse_vector_peeled.mlir b/mlir/test/Dialect/SparseTensor/sparse_vector_peeled.mlir index ac0d035424574..db830f29d7ca6 100644 --- a/mlir/test/Dialect/SparseTensor/sparse_vector_peeled.mlir +++ b/mlir/test/Dialect/SparseTensor/sparse_vector_peeled.mlir @@ -2,7 +2,7 @@ // RUN: FileCheck %s #SparseVector = #sparse_tensor.encoding<{ - lvlTypes = [ "compressed" ], + map = (d0) -> (d0 : compressed), posWidth = 32, crdWidth = 32 }> diff --git a/mlir/test/Dialect/SparseTensor/unsparsifiable_dense_op.mlir b/mlir/test/Dialect/SparseTensor/unsparsifiable_dense_op.mlir index 3304d00d96a1d..2cef1afad2b89 100644 --- a/mlir/test/Dialect/SparseTensor/unsparsifiable_dense_op.mlir +++ b/mlir/test/Dialect/SparseTensor/unsparsifiable_dense_op.mlir @@ -14,7 +14,7 @@ iterator_types = ["parallel", "parallel", "parallel", "reduction"] } -#VEC = #sparse_tensor.encoding<{ lvlTypes = [ "compressed" ], posWidth = 32, crdWidth = 32 }> +#VEC = #sparse_tensor.encoding<{ map = (d0) -> (d0 : compressed), posWidth = 32, crdWidth = 32 }> #COO = #sparse_tensor.encoding<{ lvlTypes = [ "compressed_nu", "singleton" ], posWidth = 32, crdWidth = 32 }> #CCC = #sparse_tensor.encoding<{ lvlTypes = [ "compressed", "compressed", "compressed" ], posWidth = 32, crdWidth = 32 }> diff --git a/mlir/test/Dialect/SparseTensor/vectorize_reduction.mlir b/mlir/test/Dialect/SparseTensor/vectorize_reduction.mlir index 09bbe410bfc16..cfa3e2cdea252 100644 --- a/mlir/test/Dialect/SparseTensor/vectorize_reduction.mlir +++ b/mlir/test/Dialect/SparseTensor/vectorize_reduction.mlir @@ -55,7 +55,7 @@ // CHECK-OFF: %[[VAL_16:.*]] = bufferization.to_tensor %[[VAL_6]] : memref // CHECK-OFF: return %[[VAL_16]] : tensor // CHECK-OFF: } -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> #trait = { indexing_maps = [ @@ -132,7 +132,7 @@ func.func @sparse_reduction_ori(%argx: tensor, // CHECK-OFF: %[[VAL_16:.*]] = bufferization.to_tensor %[[VAL_6]] : memref // CHECK-OFF: return %[[VAL_16]] : tensor // CHECK-OFF: } -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> #trait = { indexing_maps = [ @@ -207,7 +207,7 @@ func.func @sparse_reduction_ori_accumulator_on_rhs(%argx: tensor, // CHECK-OFF: %[[VAL_16:.*]] = bufferization.to_tensor %[[VAL_6]] : memref // CHECK-OFF: return %[[VAL_16]] : tensor // CHECK-OFF: } -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> #trait = { indexing_maps = [ @@ -283,7 +283,7 @@ func.func @sparse_reduction_subi(%argx: tensor, // CHECK-OFF: return %[[VAL_16]] : tensor // CHECK-OFF: } -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> #trait = { indexing_maps = [ @@ -359,7 +359,7 @@ func.func @sparse_reduction_xor(%argx: tensor, // CHECK-OFF: return %[[VAL_16]] : tensor // CHECK-OFF: } -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> #trait = { indexing_maps = [ @@ -435,7 +435,7 @@ func.func @sparse_reduction_addi(%argx: tensor, // CHECK-OFF: return %[[VAL_16]] : tensor // CHECK-OFF: } -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> #trait = { indexing_maps = [ @@ -511,7 +511,7 @@ func.func @sparse_reduction_subf(%argx: tensor, // CHECK-OFF: return %[[VAL_16]] : tensor // CHECK-OFF: } -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> #trait = { indexing_maps = [ diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/dense_output_bf16.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/dense_output_bf16.mlir index fe6fcea7cef7c..420d3d8c62327 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/dense_output_bf16.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/dense_output_bf16.mlir @@ -29,8 +29,8 @@ // UNSUPPORTED: target=aarch64{{.*}} -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> -#DenseVector = #sparse_tensor.encoding<{lvlTypes = ["dense"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> +#DenseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : dense)}> #trait_vec_op = { indexing_maps = [ diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/dense_output_f16.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/dense_output_f16.mlir index f88b38b467e89..96ea972bd6b5f 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/dense_output_f16.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/dense_output_f16.mlir @@ -30,8 +30,8 @@ // Do the same run, but now with direct IR generation and VLA vectorization. // RUN: %if mlir_arm_sve_tests %{ %{compile_sve} | %{run_sve} | FileCheck %s %} -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> -#DenseVector = #sparse_tensor.encoding<{lvlTypes = ["dense"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> +#DenseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : dense)}> #trait_vec_op = { indexing_maps = [ diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_abs.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_abs.mlir index b9e8eec032060..584906034d2d2 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_abs.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_abs.mlir @@ -30,7 +30,7 @@ // Do the same run, but now with direct IR generation and VLA vectorization. // RUN: %if mlir_arm_sve_tests %{ %{compile_sve} | %{run_sve} | FileCheck %s %} -#SparseVector = #sparse_tensor.encoding<{ lvlTypes = [ "compressed" ] }> +#SparseVector = #sparse_tensor.encoding<{ map = (d0) -> (d0 : compressed) }> #trait_op = { indexing_maps = [ diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_binary.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_binary.mlir index bd2b238a64369..ed34d3f71832e 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_binary.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_binary.mlir @@ -30,7 +30,7 @@ // Do the same run, but now with direct IR generation and VLA vectorization. // RUN: %if mlir_arm_sve_tests %{ %{compile_sve} | %{run_sve} | FileCheck %s %} -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> #DCSR = #sparse_tensor.encoding<{lvlTypes = ["compressed", "compressed"]}> // diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_cast.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_cast.mlir index 5cc465ab4d0db..5290b97b9b4f6 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_cast.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_cast.mlir @@ -30,7 +30,7 @@ // Do the same run, but now with direct IR generation and VLA vectorization. // RUN: %if mlir_arm_sve_tests %{ %{compile_sve} | %{run_sve} | FileCheck %s %} -#SV = #sparse_tensor.encoding<{ lvlTypes = [ "compressed" ] }> +#SV = #sparse_tensor.encoding<{ map = (d0) -> (d0 : compressed) }> #trait_cast = { indexing_maps = [ diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_collapse_shape.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_collapse_shape.mlir index 100025f218940..ada14f4696a92 100755 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_collapse_shape.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_collapse_shape.mlir @@ -31,7 +31,7 @@ // RUN: %if mlir_arm_sve_tests %{ %{compile_sve} | %{run_sve} | FileCheck %s %} #SparseVector = #sparse_tensor.encoding<{ - lvlTypes = ["compressed"] + map = (d0) -> (d0 : compressed) }> #SparseMatrix = #sparse_tensor.encoding<{ diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_complex32.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_complex32.mlir index 1490c87b64c8b..6e67e043043ec 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_complex32.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_complex32.mlir @@ -30,7 +30,7 @@ // Do the same run, but now with direct IR generation and VLA vectorization. // RUN: %if mlir_arm_sve_tests %{ %{compile_sve} | %{run_sve} | FileCheck %s %} -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> #trait_op = { indexing_maps = [ diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_complex64.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_complex64.mlir index 17e95146a8271..0f9ce030269d7 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_complex64.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_complex64.mlir @@ -27,7 +27,7 @@ // Do the same run, but now with direct IR generation and VLA vectorization. // RUN: %if mlir_arm_sve_tests %{ %{compile_sve} | %{run_sve} | FileCheck %s %} -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> #trait_op = { indexing_maps = [ diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_complex_ops.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_complex_ops.mlir index 96428d44061a7..26c0f8c3ec7d0 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_complex_ops.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_complex_ops.mlir @@ -30,7 +30,7 @@ // Do the same run, but now with direct IR generation and VLA vectorization. // RUN: %if mlir_arm_sve_tests %{ %{compile_sve} | %{run_sve} | FileCheck %s %} -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> #trait_op1 = { indexing_maps = [ diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_dot.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_dot.mlir index 2a0480bfc44cd..8c5ad40cd83c6 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_dot.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_dot.mlir @@ -30,7 +30,7 @@ // Do the same run, but now with direct IR generation and VLA vectorization. // RUN: %if mlir_arm_sve_tests %{ %{compile_sve} | %{run_sve} | FileCheck %s %} -#SparseVector = #sparse_tensor.encoding<{ lvlTypes = [ "compressed" ] }> +#SparseVector = #sparse_tensor.encoding<{ map = (d0) -> (d0 : compressed) }> module { diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_expand_shape.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_expand_shape.mlir index 75d85949a7658..9d47b3ab1a7ec 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_expand_shape.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_expand_shape.mlir @@ -31,7 +31,7 @@ // RUN: %if mlir_arm_sve_tests %{ %{compile_sve} | %{run_sve} | FileCheck %s %} #SparseVector = #sparse_tensor.encoding<{ - lvlTypes = ["compressed"] + map = (d0) -> (d0 : compressed) }> #SparseMatrix = #sparse_tensor.encoding<{ diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_index.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_index.mlir index 5d396e01357a9..1633194408ec0 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_index.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_index.mlir @@ -32,7 +32,7 @@ #SparseVector = #sparse_tensor.encoding<{ - lvlTypes = ["compressed"] + map = (d0) -> (d0 : compressed) }> #SparseMatrix = #sparse_tensor.encoding<{ diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_index_dense.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_index_dense.mlir index dc08a2142f49c..5a030ac93110e 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_index_dense.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_index_dense.mlir @@ -31,7 +31,7 @@ // RUN: %if mlir_arm_sve_tests %{ %{compile_sve} | %{run_sve} | FileCheck %s %} #SparseVector = #sparse_tensor.encoding<{ - lvlTypes = ["compressed"] + map = (d0) -> (d0 : compressed) }> #SparseMatrix = #sparse_tensor.encoding<{ diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_insert_1d.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_insert_1d.mlir index d8bd5c4a158b9..2f42e5f573177 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_insert_1d.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_insert_1d.mlir @@ -32,7 +32,7 @@ // Insertion example using pure codegen (no sparse runtime support lib). -#SparseVector = #sparse_tensor.encoding<{ lvlTypes = [ "compressed" ] }> +#SparseVector = #sparse_tensor.encoding<{ map = (d0) -> (d0 : compressed) }> #trait_mul_s = { indexing_maps = [ diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_re_im.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_re_im.mlir index fa6f4c99a9d0a..6dd2c4c497a48 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_re_im.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_re_im.mlir @@ -30,7 +30,7 @@ // Do the same run, but now with direct IR generation and VLA vectorization. // RUN: %if mlir_arm_sve_tests %{ %{compile_sve} | %{run_sve} | FileCheck %s %} -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> #trait_op = { indexing_maps = [ diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reduce_custom.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reduce_custom.mlir index 3489dcadfd543..1a1351ac6c6a6 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reduce_custom.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reduce_custom.mlir @@ -32,7 +32,7 @@ // Reduction in this file _are_ supported by the AArch64 SVE backend -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> #CSR = #sparse_tensor.encoding<{lvlTypes = ["dense", "compressed"]}> #CSC = #sparse_tensor.encoding<{ lvlTypes = [ "dense", "compressed" ], diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reduce_custom_prod.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reduce_custom_prod.mlir index 44a8c83830f4c..af30fb1bf8a1d 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reduce_custom_prod.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reduce_custom_prod.mlir @@ -31,7 +31,7 @@ // the AArch64 SVE backend (so the set-up is a bit different to // sparse_reducitons.mlir) -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> #CSR = #sparse_tensor.encoding<{lvlTypes = ["dense", "compressed"]}> #CSC = #sparse_tensor.encoding<{ lvlTypes = [ "dense", "compressed" ], diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reduce_custom_sum.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reduce_custom_sum.mlir index 3030eafe215f7..43a6e698ce5d6 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reduce_custom_sum.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reduce_custom_sum.mlir @@ -27,7 +27,7 @@ // REDEFINE: %{sparse_compiler_opts} = enable-runtime-library=false vl=2 reassociate-fp-reductions=true enable-index-optimizations=true // RUN: %{compile} | %{run} | FileCheck %s -#SV = #sparse_tensor.encoding<{ lvlTypes = [ "compressed" ] }> +#SV = #sparse_tensor.encoding<{ map = (d0) -> (d0 : compressed) }> #trait_reduction = { indexing_maps = [ diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reductions.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reductions.mlir index c32058093713c..b96aa4db285b4 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reductions.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reductions.mlir @@ -32,7 +32,7 @@ // Reduction in this file _are_ supported by the AArch64 SVE backend -#SV = #sparse_tensor.encoding<{ lvlTypes = [ "compressed" ] }> +#SV = #sparse_tensor.encoding<{ map = (d0) -> (d0 : compressed) }> #trait_reduction = { indexing_maps = [ diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reductions_min.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reductions_min.mlir index ab8f88c00a66c..3fe597100f37a 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reductions_min.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reductions_min.mlir @@ -30,7 +30,7 @@ // Do the same run, but now with direct IR generation and VLA vectorization. // RUN: %if mlir_arm_sve_tests %{ %{compile_sve} | %{run_sve} | FileCheck %s %} -#SV = #sparse_tensor.encoding<{ lvlTypes = [ "compressed" ] }> +#SV = #sparse_tensor.encoding<{ map = (d0) -> (d0 : compressed) }> #trait_reduction = { indexing_maps = [ diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reductions_prod.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reductions_prod.mlir index 628bc25aad08e..80a7e90a3f8f1 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reductions_prod.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reductions_prod.mlir @@ -27,8 +27,8 @@ // REDEFINE: %{sparse_compiler_opts} = enable-runtime-library=false enable-buffer-initialization=true vl=2 reassociate-fp-reductions=true enable-index-optimizations=true // RUN: %{compile} | %{run} | FileCheck %s -#SV = #sparse_tensor.encoding<{ lvlTypes = [ "compressed" ] }> -#DV = #sparse_tensor.encoding<{ lvlTypes = [ "dense" ] }> +#SV = #sparse_tensor.encoding<{ map = (d0) -> (d0 : compressed) }> +#DV = #sparse_tensor.encoding<{ map = (d0) -> (d0 : dense) }> #trait_reduction = { indexing_maps = [ diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reshape.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reshape.mlir index f141d65e6c26b..64d27dfb91bf8 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reshape.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_reshape.mlir @@ -31,7 +31,7 @@ // RUN: %if mlir_arm_sve_tests %{ %{compile_sve} | %{run_sve} | FileCheck %s %} #SparseVector = #sparse_tensor.encoding<{ - lvlTypes = ["compressed"] + map = (d0) -> (d0 : compressed) }> #SparseMatrix = #sparse_tensor.encoding<{ diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_select.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_select.mlir index 65eb624966e94..11301c4316352 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_select.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_select.mlir @@ -30,7 +30,7 @@ // Do the same run, but now with VLA vectorization. // RUN: %if mlir_arm_sve_tests %{ %{compile_sve} | %{run_sve} | FileCheck %s %} -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> #CSR = #sparse_tensor.encoding<{lvlTypes = ["dense", "compressed"]}> #CSC = #sparse_tensor.encoding<{ lvlTypes = [ "dense", "compressed" ], diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_sign.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_sign.mlir index 1009e50b20c9d..327ac8656f405 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_sign.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_sign.mlir @@ -30,7 +30,7 @@ // Do the same run, but now with VLA vectorization. // RUN: %if mlir_arm_sve_tests %{ %{compile_sve} | %{run_sve} | FileCheck %s %} -#SparseVector = #sparse_tensor.encoding<{ lvlTypes = [ "compressed" ] }> +#SparseVector = #sparse_tensor.encoding<{ map = (d0) -> (d0 : compressed) }> #trait_op = { indexing_maps = [ diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_tanh.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_tanh.mlir index e0de6dd7200a5..4b8ef2274c27a 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_tanh.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_tanh.mlir @@ -33,7 +33,7 @@ // Current fails for SVE, see https://github.com/llvm/llvm-project/issues/60626 // UNSUPPORTED: target=aarch64{{.*}} -#SparseVector = #sparse_tensor.encoding<{ lvlTypes = [ "compressed" ] }> +#SparseVector = #sparse_tensor.encoding<{ map = (d0) -> (d0 : compressed) }> #trait_op = { indexing_maps = [ diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_unary.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_unary.mlir index ddd52ef40c752..cf9113f81ef53 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_unary.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_unary.mlir @@ -30,7 +30,7 @@ // Do the same run, but now with VLA vectorization. // RUN: %if mlir_arm_sve_tests %{ %{compile_sve} | %{run_sve} | FileCheck %s %} -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> #DCSR = #sparse_tensor.encoding<{lvlTypes = ["compressed", "compressed"]}> // diff --git a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_vector_ops.mlir b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_vector_ops.mlir index ac7e9ff8f8756..d9e1125a79d85 100644 --- a/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_vector_ops.mlir +++ b/mlir/test/Integration/Dialect/SparseTensor/CPU/sparse_vector_ops.mlir @@ -30,8 +30,8 @@ // Do the same run, but now with VLA vectorization. // RUN: %if mlir_arm_sve_tests %{ %{compile_sve} | %{run_sve} | FileCheck %s %} -#SparseVector = #sparse_tensor.encoding<{lvlTypes = ["compressed"]}> -#DenseVector = #sparse_tensor.encoding<{lvlTypes = ["dense"]}> +#SparseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : compressed)}> +#DenseVector = #sparse_tensor.encoding<{map = (d0) -> (d0 : dense)}> // // Traits for 1-d tensor (aka vector) operations. diff --git a/mlir/test/python/dialects/sparse_tensor/dialect.py b/mlir/test/python/dialects/sparse_tensor/dialect.py index 8d98d670ee4d0..a4343545ffa12 100644 --- a/mlir/test/python/dialects/sparse_tensor/dialect.py +++ b/mlir/test/python/dialects/sparse_tensor/dialect.py @@ -16,7 +16,7 @@ def testEncodingAttr1D(): with Context() as ctx: parsed = Attribute.parse( "#sparse_tensor.encoding<{" - ' lvlTypes = [ "compressed" ],' + ' map = (d0) -> (d0 : compressed),' " posWidth = 16," " crdWidth = 32" "}>" @@ -94,7 +94,7 @@ def testEncodingAttrOnTensorType(): encoding = st.EncodingAttr( Attribute.parse( "#sparse_tensor.encoding<{" - ' lvlTypes = [ "compressed" ], ' + ' map = (d0) -> (d0 : compressed), ' " posWidth = 64," " crdWidth = 32" "}>" From c6fcc14d2f65928b14bb0ecc36f8e858774809a0 Mon Sep 17 00:00:00 2001 From: Yinying Li Date: Tue, 12 Sep 2023 22:07:19 +0000 Subject: [PATCH 2/3] fixup! [mlir][sparse] Migrate tests to use new syntax --- mlir/test/python/dialects/sparse_tensor/dialect.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/mlir/test/python/dialects/sparse_tensor/dialect.py b/mlir/test/python/dialects/sparse_tensor/dialect.py index a4343545ffa12..3600a7d95b138 100644 --- a/mlir/test/python/dialects/sparse_tensor/dialect.py +++ b/mlir/test/python/dialects/sparse_tensor/dialect.py @@ -16,7 +16,7 @@ def testEncodingAttr1D(): with Context() as ctx: parsed = Attribute.parse( "#sparse_tensor.encoding<{" - ' map = (d0) -> (d0 : compressed),' + " map = (d0) -> (d0 : compressed)," " posWidth = 16," " crdWidth = 32" "}>" @@ -56,7 +56,7 @@ def testEncodingAttr2D(): with Context() as ctx: parsed = Attribute.parse( "#sparse_tensor.encoding<{" - ' lvlTypes = [ "dense", "compressed" ],' + " lvlTypes = [ "dense", "compressed" ]," " dimToLvl = affine_map<(d0, d1) -> (d1, d0)>," " posWidth = 8," " crdWidth = 32" @@ -94,7 +94,7 @@ def testEncodingAttrOnTensorType(): encoding = st.EncodingAttr( Attribute.parse( "#sparse_tensor.encoding<{" - ' map = (d0) -> (d0 : compressed), ' + " map = (d0) -> (d0 : compressed), " " posWidth = 64," " crdWidth = 32" "}>" From c9a6c699231b1d874e0b875ea5c79f44c438c0a4 Mon Sep 17 00:00:00 2001 From: Yinying Li Date: Tue, 12 Sep 2023 22:16:53 +0000 Subject: [PATCH 3/3] fixup! fixup! [mlir][sparse] Migrate tests to use new syntax --- mlir/test/python/dialects/sparse_tensor/dialect.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/mlir/test/python/dialects/sparse_tensor/dialect.py b/mlir/test/python/dialects/sparse_tensor/dialect.py index 3600a7d95b138..bd5d9750eb574 100644 --- a/mlir/test/python/dialects/sparse_tensor/dialect.py +++ b/mlir/test/python/dialects/sparse_tensor/dialect.py @@ -56,7 +56,7 @@ def testEncodingAttr2D(): with Context() as ctx: parsed = Attribute.parse( "#sparse_tensor.encoding<{" - " lvlTypes = [ "dense", "compressed" ]," + ' lvlTypes = [ "dense", "compressed" ],' " dimToLvl = affine_map<(d0, d1) -> (d1, d0)>," " posWidth = 8," " crdWidth = 32"