From b01e04e0f7b8247e21a8ecfddd5101ccd74079fc Mon Sep 17 00:00:00 2001 From: klensy Date: Mon, 27 May 2024 16:42:47 +0300 Subject: [PATCH 1/2] fix filecheck test typos --- mlir/test/Analysis/test-liveness.mlir | 8 ++--- .../Conversion/GPUToNVVM/gpu-to-nvvm.mlir | 2 +- .../convert-dynamic-memref-ops.mlir | 4 +-- .../Conversion/NVGPUToNVVM/nvgpu-to-nvvm.mlir | 10 +++--- mlir/test/Dialect/AMX/roundtrip.mlir | 2 +- mlir/test/Dialect/Affine/loop-fusion-3.mlir | 2 +- mlir/test/Dialect/Affine/unroll.mlir | 2 +- .../ArmSME/tile-allocation-liveness.mlir | 2 +- .../Dialect/Linalg/drop-unit-extent-dims.mlir | 2 +- .../Linalg/fusion-elementwise-ops.mlir | 2 +- mlir/test/Dialect/Linalg/transform-ops.mlir | 6 ++-- mlir/test/Dialect/OpenMP/ops.mlir | 2 +- .../Dialect/SparseTensor/sparse_reshape.mlir | 12 +++---- mlir/test/Dialect/Tensor/canonicalize.mlir | 4 +-- mlir/test/Dialect/Tosa/canonicalize.mlir | 2 +- mlir/test/Dialect/Tosa/tosa-infer-shapes.mlir | 6 ++-- .../vector-dropleadunitdim-transforms.mlir | 2 +- mlir/test/IR/parser.mlir | 30 ++++++++--------- mlir/test/Target/LLVMIR/llvmir.mlir | 28 ++++++++++------ mlir/test/Target/LLVMIR/openmp-llvm.mlir | 32 +++++++++---------- mlir/test/Transforms/canonicalize.mlir | 2 +- mlir/test/mlir-cpu-runner/copy.mlir | 2 +- mlir/test/mlir-tblgen/bytecode-reserved.td | 2 +- mlir/test/python/ir/attributes.py | 4 +-- 24 files changed, 89 insertions(+), 81 deletions(-) diff --git a/mlir/test/Analysis/test-liveness.mlir b/mlir/test/Analysis/test-liveness.mlir index 61a1e5fffa888..47805e5d95dae 100644 --- a/mlir/test/Analysis/test-liveness.mlir +++ b/mlir/test/Analysis/test-liveness.mlir @@ -40,7 +40,7 @@ func.func @func_simpleBranch(%arg0: i32, %arg1 : i32) -> i32 { // CHECK-SAME: arg0@0 arg1@0 val_2 // CHECK: return // CHECK-SAME: val_2 - // CHECK-NEXT EndCurrentlyLive + // CHECK-NEXT:EndCurrentlyLive %result = arith.addi %arg0, %arg1 : i32 return %result : i32 } @@ -197,9 +197,9 @@ func.func @func_ranges(%cond : i1, %arg1 : i32, %arg2 : i32, %arg3 : i32) -> i32 // CHECK-NEXT: %2 = arith.addi // CHECK-NEXT: %3 = arith.muli // CHECK-NEXT: val_7 - // CHECK-NEXT %2 = arith.addi - // CHECK-NEXT %3 = arith.muli - // CHECK-NEXT %4 = arith.muli + // CHECK-NEXT: %2 = arith.addi + // CHECK-NEXT: %3 = arith.muli + // CHECK-NEXT: %4 = arith.muli // CHECK: val_8 // CHECK-NEXT: %3 = arith.muli // CHECK-NEXT: %4 = arith.muli diff --git a/mlir/test/Conversion/GPUToNVVM/gpu-to-nvvm.mlir b/mlir/test/Conversion/GPUToNVVM/gpu-to-nvvm.mlir index 8877ee083286b..cd701095d8e64 100644 --- a/mlir/test/Conversion/GPUToNVVM/gpu-to-nvvm.mlir +++ b/mlir/test/Conversion/GPUToNVVM/gpu-to-nvvm.mlir @@ -638,7 +638,7 @@ gpu.module @test_module_30 { } // CHECK-LABEL: @subgroup_reduce_xor gpu.func @subgroup_reduce_xor(%arg0 : i32) { - // CHECK nvvm.redux.sync xor {{.*}} + // CHECK: nvvm.redux.sync xor {{.*}} %result = gpu.subgroup_reduce xor %arg0 uniform {} : (i32) -> (i32) gpu.return } diff --git a/mlir/test/Conversion/MemRefToLLVM/convert-dynamic-memref-ops.mlir b/mlir/test/Conversion/MemRefToLLVM/convert-dynamic-memref-ops.mlir index 609fcb10b992c..9d8f4266adf27 100644 --- a/mlir/test/Conversion/MemRefToLLVM/convert-dynamic-memref-ops.mlir +++ b/mlir/test/Conversion/MemRefToLLVM/convert-dynamic-memref-ops.mlir @@ -378,14 +378,14 @@ func.func @memref_cast_ranked_to_unranked(%arg : memref<42x2x?xf32>) { // CHECK-DAG: %[[p:.*]] = llvm.alloca %[[c]] x !llvm.struct<(ptr, ptr, i64, array<3 x i64>, array<3 x i64>)> : (i64) -> !llvm.ptr // CHECK-DAG: llvm.store %{{.*}}, %[[p]] : !llvm.struct<(ptr, ptr, i64, array<3 x i64>, array<3 x i64>)>, !llvm.ptr // CHECK-DAG: %[[r:.*]] = llvm.mlir.constant(3 : index) : i64 -// CHECK : llvm.mlir.undef : !llvm.struct<(i64, ptr)> +// CHECK: llvm.mlir.undef : !llvm.struct<(i64, ptr)> // CHECK-DAG: llvm.insertvalue %[[r]], %{{.*}}[0] : !llvm.struct<(i64, ptr)> // CHECK-DAG: llvm.insertvalue %[[p]], %{{.*}}[1] : !llvm.struct<(i64, ptr)> // CHECK32-DAG: %[[c:.*]] = llvm.mlir.constant(1 : index) : i64 // CHECK32-DAG: %[[p:.*]] = llvm.alloca %[[c]] x !llvm.struct<(ptr, ptr, i32, array<3 x i32>, array<3 x i32>)> : (i64) -> !llvm.ptr // CHECK32-DAG: llvm.store %{{.*}}, %[[p]] : !llvm.struct<(ptr, ptr, i32, array<3 x i32>, array<3 x i32>)>, !llvm.ptr // CHECK32-DAG: %[[r:.*]] = llvm.mlir.constant(3 : index) : i32 -// CHECK32 : llvm.mlir.undef : !llvm.struct<(i32, ptr)> +// CHECK32: llvm.mlir.undef : !llvm.struct<(i32, ptr)> // CHECK32-DAG: llvm.insertvalue %[[r]], %{{.*}}[0] : !llvm.struct<(i32, ptr)> // CHECK32-DAG: llvm.insertvalue %[[p]], %{{.*}}[1] : !llvm.struct<(i32, ptr)> %0 = memref.cast %arg : memref<42x2x?xf32> to memref<*xf32> diff --git a/mlir/test/Conversion/NVGPUToNVVM/nvgpu-to-nvvm.mlir b/mlir/test/Conversion/NVGPUToNVVM/nvgpu-to-nvvm.mlir index 1b046d32f163a..73d2367915284 100644 --- a/mlir/test/Conversion/NVGPUToNVVM/nvgpu-to-nvvm.mlir +++ b/mlir/test/Conversion/NVGPUToNVVM/nvgpu-to-nvvm.mlir @@ -11,7 +11,7 @@ func.func @m16n8k16_fp16(%arg0: vector<4x2xf16>, %arg1: vector<2x2xf16>, %arg2: // CHECK: llvm.extractvalue %{{.*}}[1] : !llvm.array<2 x vector<2xf16>> // CHECK: llvm.extractvalue %{{.*}}[0] : !llvm.array<2 x vector<2xf16>> // CHECK: llvm.extractvalue %{{.*}}[1] : !llvm.array<2 x vector<2xf16>> - // CHECK-NOT llvm.extractvalue + // CHECK-NOT: llvm.extractvalue // CHECK: [[d:%.+]] = nvvm.mma.sync // CHECK-SAME: shape = #nvvm.shape %d = nvgpu.mma.sync (%arg0, %arg1, %arg2) {mmaShape = [16, 8, 16]} : (vector<4x2xf16>, vector<2x2xf16>, vector<2x2xf16>) -> vector<2x2xf16> @@ -56,7 +56,7 @@ func.func @m16n8k8_fp16(%arg0: vector<2x2xf16>, %arg1: vector<1x2xf16>, %arg2: v // CHECK: llvm.extractvalue %{{.*}}[0] : !llvm.array<1 x vector<2xf16>> // CHECK: llvm.extractvalue %{{.*}}[0] : !llvm.array<2 x vector<2xf16>> // CHECK: llvm.extractvalue %{{.*}}[1] : !llvm.array<2 x vector<2xf16>> - // CHECK-NOT llvm.extractvalue + // CHECK-NOT: llvm.extractvalue // CHECK: [[d:%.+]] = nvvm.mma.sync // CHECK-SAME: shape = #nvvm.shape %d = nvgpu.mma.sync (%arg0, %arg1, %arg2) {mmaShape = [16, 8, 8]} : (vector<2x2xf16>, vector<1x2xf16>, vector<2x2xf16>) -> vector<2x2xf16> @@ -360,7 +360,7 @@ func.func @mma_sp_sync_f16_16832(%arg0: vector<4x2xf16>, // CHECK: llvm.extractvalue %{{.*}}[0] : !llvm.array<2 x vector<2xf16>> // CHECK: llvm.extractvalue %{{.*}}[1] : !llvm.array<2 x vector<2xf16>> - // CHECK-NOT llvm.extractvalue + // CHECK-NOT: llvm.extractvalue // CHECK: %[[sparseMetadata:.+]] = llvm.bitcast %{{.+}} : vector<2xi16> to i32 @@ -396,7 +396,7 @@ func.func @mma_sp_sync_f16_16816(%arg0: vector<2x2xf16>, // CHECK: llvm.extractvalue %{{.*}}[0] : !llvm.array<2 x vector<2xf16>> // CHECK: llvm.extractvalue %{{.*}}[1] : !llvm.array<2 x vector<2xf16>> - // CHECK-NOT llvm.extractvalue + // CHECK-NOT: llvm.extractvalue // CHECK: %[[sparseMetadata:.+]] = llvm.bitcast %{{.+}} : vector<2xi16> to i32 @@ -455,7 +455,7 @@ func.func @mma_sp_sync_i8_16864(%arg0: vector<4x4xi8>, // CHECK: llvm.extractvalue %{{.*}}[{{.*}}] : !llvm.array<2 x vector<2xi32>> // CHECK: llvm.extractvalue %{{.*}}[{{.*}}] : !llvm.array<2 x vector<2xi32>> - // CHECK-NOT llvm.extractvalue + // CHECK-NOT: llvm.extractvalue // CHECK: %[[sparseMetadata:.+]] = llvm.bitcast %{{.+}} : vector<2xi16> to i32 diff --git a/mlir/test/Dialect/AMX/roundtrip.mlir b/mlir/test/Dialect/AMX/roundtrip.mlir index 8018962f1092d..f2ac5e47f6c35 100644 --- a/mlir/test/Dialect/AMX/roundtrip.mlir +++ b/mlir/test/Dialect/AMX/roundtrip.mlir @@ -2,7 +2,7 @@ // CHECK-LABEL: tzero // CHECK: amx.tile_zero : vector<16x16xbf16> -// CHECK amx.tile_store %{{.*}}[%{{.*}}, %{{.*}}], %{{.*}} : memref, vector<16x16xbf16> +// CHECK: amx.tile_store %{{.*}}[%{{.*}}, %{{.*}}], %{{.*}} : memref, vector<16x16xbf16> func.func @tzero(%arg0: memref) { %0 = arith.constant 0 : index %1 = amx.tile_zero : vector<16x16xbf16> diff --git a/mlir/test/Dialect/Affine/loop-fusion-3.mlir b/mlir/test/Dialect/Affine/loop-fusion-3.mlir index 37ad178235dc9..6bc4feadb8c98 100644 --- a/mlir/test/Dialect/Affine/loop-fusion-3.mlir +++ b/mlir/test/Dialect/Affine/loop-fusion-3.mlir @@ -889,7 +889,7 @@ func.func @reduce_add_non_innermost(%arg0: memref<64x64xf32, 1>, %arg1: memref<1 // CHECK: affine.for // CHECK-NEXT: affine.for // CHECK-NEXT: affine.for -// CHECK affine.for +// CHECK: affine.for diff --git a/mlir/test/Dialect/Affine/unroll.mlir b/mlir/test/Dialect/Affine/unroll.mlir index eb90d10362f9b..e398c3fe2011d 100644 --- a/mlir/test/Dialect/Affine/unroll.mlir +++ b/mlir/test/Dialect/Affine/unroll.mlir @@ -616,7 +616,7 @@ func.func @loop_nest_non_trivial_multiple_upper_bound_alt(%M : index, %N : index // UNROLL-BY-4-NEXT: "foo" // UNROLL-BY-4-NEXT: "foo" // UNROLL-BY-4-NEXT: "foo" - // UNROLL-BY-4-NOT for + // UNROLL-BY-4-NOT: for // UNROLL-BY-4: return return } diff --git a/mlir/test/Dialect/ArmSME/tile-allocation-liveness.mlir b/mlir/test/Dialect/ArmSME/tile-allocation-liveness.mlir index fe4c005c7c42f..9c22b29ac22e7 100644 --- a/mlir/test/Dialect/ArmSME/tile-allocation-liveness.mlir +++ b/mlir/test/Dialect/ArmSME/tile-allocation-liveness.mlir @@ -406,7 +406,7 @@ func.func @avoidable_spill(%a: vector<[4]xf32>, %b: vector<[4]xf32>, %c: vector< // CHECK: arm_sme.get_tile {tile_id = 2 : i32} : vector<[4]x[4]xf32> // CHECK: arm_sme.get_tile {tile_id = 3 : i32} : vector<[4]x[4]xf32> // CHECK: arm_sme.move_vector_to_tile_slice {{.*}} {tile_id = 0 : i32} : vector<[4]xf32> into vector<[4]x[4]xf32> -// CHECK-NOT tile_id = 16 +// CHECK-NOT: tile_id = 16 func.func @cond_branch_with_backedge(%slice: vector<[4]xf32>) { %tileA = arm_sme.get_tile : vector<[4]x[4]xf32> %tileB = arm_sme.get_tile : vector<[4]x[4]xf32> diff --git a/mlir/test/Dialect/Linalg/drop-unit-extent-dims.mlir b/mlir/test/Dialect/Linalg/drop-unit-extent-dims.mlir index a9cbaaf7fdc48..c72a676b3f37e 100644 --- a/mlir/test/Dialect/Linalg/drop-unit-extent-dims.mlir +++ b/mlir/test/Dialect/Linalg/drop-unit-extent-dims.mlir @@ -881,7 +881,7 @@ func.func @input_stays_same(%arg0 : memref>, %arg1 // CHECK: func @input_stays_same( // CHECK-SAME: %[[ARG0:.*]]: memref>, // CHECK-SAME: %[[ARG1:.*]]: f32, %[[ARG2:.*]]: memref) -// CHECK-SAME -> memref { +// CHECK-SAME: -> memref { // CHECK: %[[OUT:.*]] = memref.collapse_shape %[[ARG2]] {{\[}}[0, 1], [2, 3], [4]] // CHECK-SAME: : memref into memref // CHECK: linalg.generic diff --git a/mlir/test/Dialect/Linalg/fusion-elementwise-ops.mlir b/mlir/test/Dialect/Linalg/fusion-elementwise-ops.mlir index 15a4f6cdd3bbe..898070b2e61fc 100644 --- a/mlir/test/Dialect/Linalg/fusion-elementwise-ops.mlir +++ b/mlir/test/Dialect/Linalg/fusion-elementwise-ops.mlir @@ -532,7 +532,7 @@ func.func @scalar_generic_fusion // CHECK-SAME: ins(%[[ARG1]] : tensor) // CHECK: tensor.extract %[[ARG0]] // CHECK: linalg.yield -// CHECK return %[[T0]] +// CHECK: return %[[T0]] // ----- diff --git a/mlir/test/Dialect/Linalg/transform-ops.mlir b/mlir/test/Dialect/Linalg/transform-ops.mlir index 733f305f850c6..c152fc887a3a3 100644 --- a/mlir/test/Dialect/Linalg/transform-ops.mlir +++ b/mlir/test/Dialect/Linalg/transform-ops.mlir @@ -2,7 +2,7 @@ transform.sequence failures(propagate) { ^bb1(%arg0: !transform.any_op): - // CHECK %{{.*}}, %{{.*}}:2 = transform.structured.tile + // CHECK: %{{.*}}, %{{.*}}:2 = transform.structured.tile %0, %1:2 = transform.structured.tile_using_for %arg0 tile_sizes [2, 0, 3] : (!transform.any_op) -> (!transform.any_op, !transform.any_op, !transform.any_op) } @@ -10,9 +10,9 @@ transform.sequence failures(propagate) { // and parsing with and without use of the optional `interchange` Attribute. transform.sequence failures(propagate) { ^bb1(%arg0: !transform.any_op): - // CHECK %{{.*}}, %{{.*}}:2 = transform.structured.tile %arg0 [2, 0, 3] interchange = [2, 1] {test_attr1 = 1 : i64, test_attr2} + // CHECK: %{{.*}}, %{{.*}}:2 = transform.structured.tile_using_for %arg0 tile_sizes [2, 0, 3] interchange = [2, 1] {test_attr1 = 1 : i64, test_attr2} %0, %1:2 = transform.structured.tile_using_for %arg0 tile_sizes [2, 0, 3] interchange = [2, 1] {test_attr1 = 1 : i64, test_attr2}: (!transform.any_op) -> (!transform.any_op, !transform.any_op, !transform.any_op) - // CHECK %{{.*}}, %{{.*}}:2 = transform.structured.tile %arg0 [4, 5, 3] {test_attr3 = 1 : i64, test_attr4} + // CHECK: %{{.*}}, %{{.*}}:2 = transform.structured.tile_using_for %tiled_linalg_op tile_sizes [0, 5, 3] {test_attr3 = 1 : i64, test_attr4} %2, %3:2 = transform.structured.tile_using_for %0 tile_sizes [0, 5, 3] {test_attr3 = 1 : i64, test_attr4}: (!transform.any_op) -> (!transform.any_op, !transform.any_op, !transform.any_op) } diff --git a/mlir/test/Dialect/OpenMP/ops.mlir b/mlir/test/Dialect/OpenMP/ops.mlir index caf25a3cb59f0..af66d0c65dab8 100644 --- a/mlir/test/Dialect/OpenMP/ops.mlir +++ b/mlir/test/Dialect/OpenMP/ops.mlir @@ -142,7 +142,7 @@ func.func @omp_parallel_pretty(%data_var : memref, %if_cond : i1, %num_thre omp.terminator } - // CHECK omp.parallel if(%{{.*}}) num_threads(%{{.*}} : i32) private(%{{.*}} : memref) proc_bind(close) + // CHECK: omp.parallel if(%{{.*}}) num_threads(%{{.*}} : i32) proc_bind(close) omp.parallel num_threads(%num_threads : i32) if(%if_cond: i1) proc_bind(close) { omp.terminator } diff --git a/mlir/test/Dialect/SparseTensor/sparse_reshape.mlir b/mlir/test/Dialect/SparseTensor/sparse_reshape.mlir index c96f9c31443db..492dcd05dc909 100644 --- a/mlir/test/Dialect/SparseTensor/sparse_reshape.mlir +++ b/mlir/test/Dialect/SparseTensor/sparse_reshape.mlir @@ -76,9 +76,9 @@ func.func @sparse_expand(%arg0: tensor<100xf64, #SparseVector>) -> tensor<10x10x // CHECK: %[[T:.*]] = arith.muli %[[SI0]], %[[C10]] : index // CHECK: %[[DI:.*]] = arith.addi %[[T]], %[[SI1]] : index // CHECK: %[[R1:.*]] = tensor.insert %[[SV]] into %[[A1]]{{\[}}%[[DI]]] -// CHECK scf.yield %[[R1]] -// CHECK } -// CHECK scf.yield %[[RET_1]] +// CHECK: scf.yield %[[R1]] +// CHECK: } +// CHECK: scf.yield %[[RET_1]] // CHECK: } // CHECK: %[[NT1:.*]] = sparse_tensor.load %[[RET]] hasInserts // CHECK-NOT: sparse_tensor.convert @@ -170,9 +170,9 @@ func.func @dynamic_sparse_expand(%arg0: tensor, %sz0: inde // CHECK: %[[T4:.*]] = arith.muli %[[SI1]], %[[T3]] : index // CHECK: %[[DI:.*]] = arith.addi %[[T2]], %[[T4]] : index // CHECK: %[[NT:.*]] = tensor.insert %[[SV]] into %[[R1]]{{\[}}%[[DI]]] -// CHECK scf.yield %[[NT]] -// CHECK } -// CHECK scf.yield %[[RET_1]] +// CHECK: scf.yield %[[NT]] +// CHECK: } +// CHECK: scf.yield %[[RET_1]] // CHECK: } // CHECK: %[[NT1:.*]] = sparse_tensor.load %[[RET]] hasInserts // CHECK-NOT: sparse_tensor.convert diff --git a/mlir/test/Dialect/Tensor/canonicalize.mlir b/mlir/test/Dialect/Tensor/canonicalize.mlir index f7fbd3834288b..b378a4b410e09 100644 --- a/mlir/test/Dialect/Tensor/canonicalize.mlir +++ b/mlir/test/Dialect/Tensor/canonicalize.mlir @@ -1120,8 +1120,8 @@ func.func @compose_expand_of_expand_of_zero_dim(%arg0 : tensor) // CHECK-LABEL: func.func @collapse_of_cast( // CHECK-SAME: %[[IN:.*]]: tensor<8x12x32xf32>) -> tensor { // CHECK-NEXT: %[[COLLAPSE:.*]] = tensor.collapse_shape %[[IN]] {{\[}}[0, 1], [2]] : tensor<8x12x32xf32> into tensor<96x32xf32> -// CHECK-NEXT %[[CAST:.*]] = tensor.cast %[[COLLAPSE]] : tensor<96x32xf32> to tensor -// CHECK-NEXT return %[[CAST]] : tensor +// CHECK-NEXT: %[[CAST:.*]] = tensor.cast %[[COLLAPSE]] : tensor<96x32xf32> to tensor +// CHECK-NEXT: return %[[CAST]] : tensor func.func @collapse_of_cast(%t: tensor<8x12x32xf32>) -> tensor { %0 = tensor.cast %t : tensor<8x12x32xf32> to tensor %1 = tensor.collapse_shape %0 [[0, 1], [2]] : tensor into tensor diff --git a/mlir/test/Dialect/Tosa/canonicalize.mlir b/mlir/test/Dialect/Tosa/canonicalize.mlir index 6eac759a08364..862b1c3245791 100644 --- a/mlir/test/Dialect/Tosa/canonicalize.mlir +++ b/mlir/test/Dialect/Tosa/canonicalize.mlir @@ -561,7 +561,7 @@ func.func @canonicalize_concat_slice_on_non_concat_axis(%arg0 : tensor<1x12x12xf // ----- -// CHECK-LABEL +// CHECK-LABEL: @fold_log_exp func.func @fold_log_exp(%arg0: tensor) -> tensor { // CHECK: return %arg{{.*}} : tensor %0 = tosa.exp %arg0 : (tensor) -> tensor diff --git a/mlir/test/Dialect/Tosa/tosa-infer-shapes.mlir b/mlir/test/Dialect/Tosa/tosa-infer-shapes.mlir index 2be120439ed68..5f8afa57bc747 100644 --- a/mlir/test/Dialect/Tosa/tosa-infer-shapes.mlir +++ b/mlir/test/Dialect/Tosa/tosa-infer-shapes.mlir @@ -222,13 +222,13 @@ func.func @test_binary_i32(%arg0 : tensor<4xi32>, %arg1 : tensor) -> () { // CHECK-LABEL: @test_binary_i1 func.func @test_binary_i1(%arg0 : tensor<4xi1>, %arg1 : tensor) -> () { - // CHECK tosa.logical_and %arg0, %arg1 : (tensor<4xi1>, tensor) -> tensor<4xi1> + // CHECK: tosa.logical_and %arg0, %arg1 : (tensor<4xi1>, tensor) -> tensor<4xi1> %0 = tosa.logical_and %arg0, %arg1 : (tensor<4xi1>, tensor) -> tensor<*xi1> - // CHECK tosa.logical_or %arg0, %arg1 : (tensor<4xi1>, tensor) -> tensor<4xi1> + // CHECK: tosa.logical_or %arg0, %arg1 : (tensor<4xi1>, tensor) -> tensor<4xi1> %1 = tosa.logical_or %arg0, %arg1 : (tensor<4xi1>, tensor) -> tensor<*xi1> - // CHECK tosa.logical_xor %arg0, %arg1 : (tensor<4xi1>, tensor) -> tensor<*4i1> + // CHECK: tosa.logical_xor %arg0, %arg1 : (tensor<4xi1>, tensor) -> tensor<4xi1> %2 = tosa.logical_xor %arg0, %arg1 : (tensor<4xi1>, tensor) -> tensor<*xi1> return diff --git a/mlir/test/Dialect/Vector/vector-dropleadunitdim-transforms.mlir b/mlir/test/Dialect/Vector/vector-dropleadunitdim-transforms.mlir index 252aeb0c15cbe..9526d610e490e 100644 --- a/mlir/test/Dialect/Vector/vector-dropleadunitdim-transforms.mlir +++ b/mlir/test/Dialect/Vector/vector-dropleadunitdim-transforms.mlir @@ -241,7 +241,7 @@ func.func @cast_away_contraction_leading_one_dims_nonleadingunitdim_rank4_acctra // ----- // CHECK-LABEL: func.func @cast_away_contraction_does_not_transpose_leading_unit_dims -// CHECK-NOT vector.transpose +// CHECK-NOT: vector.transpose // CHECK: vector.contract func.func @cast_away_contraction_does_not_transpose_leading_unit_dims(%lhs: vector<1x1x8xi32>, %rhs: vector<1x8x8xi32>, diff --git a/mlir/test/IR/parser.mlir b/mlir/test/IR/parser.mlir index bcc146ea0101b..c8c4025945f7a 100644 --- a/mlir/test/IR/parser.mlir +++ b/mlir/test/IR/parser.mlir @@ -342,15 +342,15 @@ func.func @loop_bounds(%N : index) { // CHECK-LABEL: func @ifinst(%{{.*}}: index) { func.func @ifinst(%N: index) { - %c = arith.constant 200 : index // CHECK %{{.*}} = arith.constant 200 - affine.for %i = 1 to 10 { // CHECK affine.for %{{.*}} = 1 to 10 { - affine.if #set0(%i)[%N, %c] { // CHECK affine.if #set0(%{{.*}})[%{{.*}}, %{{.*}}] { + %c = arith.constant 200 : index // CHECK: %{{.*}} = arith.constant 200 + affine.for %i = 1 to 10 { // CHECK: affine.for %{{.*}} = 1 to 10 { + affine.if #set0(%i)[%N, %c] { // CHECK: affine.if #set(%{{.*}})[%{{.*}}, %{{.*}}] { %x = arith.constant 1 : i32 // CHECK: %{{.*}} = arith.constant 1 : i32 %y = "add"(%x, %i) : (i32, index) -> i32 // CHECK: %{{.*}} = "add"(%{{.*}}, %{{.*}}) : (i32, index) -> i32 %z = "mul"(%y, %y) : (i32, i32) -> i32 // CHECK: %{{.*}} = "mul"(%{{.*}}, %{{.*}}) : (i32, i32) -> i32 } else { // CHECK } else { - affine.if affine_set<(i)[N] : (i - 2 >= 0, 4 - i >= 0)>(%i)[%N] { // CHECK affine.if (#set1(%{{.*}})[%{{.*}}]) { + affine.if affine_set<(i)[N] : (i - 2 >= 0, 4 - i >= 0)>(%i)[%N] { // CHECK: affine.if #set1(%{{.*}})[%{{.*}}] { // CHECK: %{{.*}} = arith.constant 1 : index %u = arith.constant 1 : index // CHECK: %{{.*}} = affine.apply #map{{.*}}(%{{.*}}, %{{.*}})[%{{.*}}] @@ -358,24 +358,24 @@ func.func @ifinst(%N: index) { } else { // CHECK } else { %v = arith.constant 3 : i32 // %c3_i32 = arith.constant 3 : i32 } - } // CHECK } - } // CHECK } - return // CHECK return -} // CHECK } + } // CHECK: } + } // CHECK: } + return // CHECK: return +} // CHECK:} // CHECK-LABEL: func @simple_ifinst(%{{.*}}: index) { func.func @simple_ifinst(%N: index) { - %c = arith.constant 200 : index // CHECK %{{.*}} = arith.constant 200 - affine.for %i = 1 to 10 { // CHECK affine.for %{{.*}} = 1 to 10 { - affine.if #set0(%i)[%N, %c] { // CHECK affine.if #set0(%{{.*}})[%{{.*}}, %{{.*}}] { + %c = arith.constant 200 : index // CHECK: %{{.*}} = arith.constant 200 + affine.for %i = 1 to 10 { // CHECK: affine.for %{{.*}} = 1 to 10 { + affine.if #set0(%i)[%N, %c] { // CHECK: affine.if #set(%{{.*}})[%{{.*}}, %{{.*}}] { %x = arith.constant 1 : i32 // CHECK: %{{.*}} = arith.constant 1 : i32 %y = "add"(%x, %i) : (i32, index) -> i32 // CHECK: %{{.*}} = "add"(%{{.*}}, %{{.*}}) : (i32, index) -> i32 %z = "mul"(%y, %y) : (i32, i32) -> i32 // CHECK: %{{.*}} = "mul"(%{{.*}}, %{{.*}}) : (i32, i32) -> i32 - } // CHECK } - } // CHECK } - return // CHECK return -} // CHECK } + } // CHECK: } + } // CHECK: } + return // CHECK: return +} // CHECK:} // CHECK-LABEL: func @attributes() { func.func @attributes() { diff --git a/mlir/test/Target/LLVMIR/llvmir.mlir b/mlir/test/Target/LLVMIR/llvmir.mlir index 97f37939551d8..c83641a57537c 100644 --- a/mlir/test/Target/LLVMIR/llvmir.mlir +++ b/mlir/test/Target/LLVMIR/llvmir.mlir @@ -2331,39 +2331,47 @@ llvm.func @streaming_compatible_func() attributes {arm_streaming_compatible} { // ----- // CHECK-LABEL: @new_za_func -// CHECK: #[[ATTR:[0-9]*]] +// CHECK-SAME: #[[ATTR:[0-9]*]] llvm.func @new_za_func() attributes {arm_new_za} { llvm.return } -// CHECK #[[ATTR]] = { "aarch64_new_za" } +// CHECK: #[[ATTR]] = { "aarch64_new_za" } + +// ----- // CHECK-LABEL: @in_za_func -// CHECK: #[[ATTR:[0-9]*]] +// CHECK-SAME: #[[ATTR:[0-9]*]] llvm.func @in_za_func() attributes {arm_in_za } { llvm.return } -// CHECK #[[ATTR]] = { "aarch64_in_za" } +// CHECK: #[[ATTR]] = { "aarch64_in_za" } + +// ----- // CHECK-LABEL: @out_za_func -// CHECK: #[[ATTR:[0-9]*]] +// CHECK-SAME: #[[ATTR:[0-9]*]] llvm.func @out_za_func() attributes {arm_out_za } { llvm.return } -// CHECK #[[ATTR]] = { "aarch64_out_za" } +// CHECK: #[[ATTR]] = { "aarch64_out_za" } + +// ----- // CHECK-LABEL: @inout_za_func -// CHECK: #[[ATTR:[0-9]*]] +// CHECK-SAME: #[[ATTR:[0-9]*]] llvm.func @inout_za_func() attributes {arm_inout_za } { llvm.return } -// CHECK #[[ATTR]] = { "aarch64_inout_za" } +// CHECK: #[[ATTR]] = { "aarch64_inout_za" } + +// ----- // CHECK-LABEL: @preserves_za_func -// CHECK: #[[ATTR:[0-9]*]] +// CHECK-SAME: #[[ATTR:[0-9]*]] llvm.func @preserves_za_func() attributes {arm_preserves_za} { llvm.return } -// CHECK #[[ATTR]] = { "aarch64_preserves_za" } +// CHECK: #[[ATTR]] = { "aarch64_preserves_za" } // ----- diff --git a/mlir/test/Target/LLVMIR/openmp-llvm.mlir b/mlir/test/Target/LLVMIR/openmp-llvm.mlir index 8654899efefd2..a1cc76f9ab770 100644 --- a/mlir/test/Target/LLVMIR/openmp-llvm.mlir +++ b/mlir/test/Target/LLVMIR/openmp-llvm.mlir @@ -549,7 +549,7 @@ llvm.func @test_omp_wsloop_auto(%lb : i64, %ub : i64, %step : i64) -> () { // CHECK: call void @__kmpc_dispatch_init_8u // CHECK: %[[continue:.*]] = call i32 @__kmpc_dispatch_next_8u // CHECK: %[[cond:.*]] = icmp ne i32 %[[continue]], 0 - // CHECK br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} + // CHECK: br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} llvm.call @body(%iv) : (i64) -> () omp.yield } @@ -568,7 +568,7 @@ llvm.func @test_omp_wsloop_runtime(%lb : i64, %ub : i64, %step : i64) -> () { // CHECK: call void @__kmpc_dispatch_init_8u // CHECK: %[[continue:.*]] = call i32 @__kmpc_dispatch_next_8u // CHECK: %[[cond:.*]] = icmp ne i32 %[[continue]], 0 - // CHECK br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} + // CHECK: br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} llvm.call @body(%iv) : (i64) -> () omp.yield } @@ -587,7 +587,7 @@ llvm.func @test_omp_wsloop_guided(%lb : i64, %ub : i64, %step : i64) -> () { // CHECK: call void @__kmpc_dispatch_init_8u // CHECK: %[[continue:.*]] = call i32 @__kmpc_dispatch_next_8u // CHECK: %[[cond:.*]] = icmp ne i32 %[[continue]], 0 - // CHECK br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} + // CHECK: br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} llvm.call @body(%iv) : (i64) -> () omp.yield } @@ -606,7 +606,7 @@ llvm.func @test_omp_wsloop_dynamic_nonmonotonic(%lb : i64, %ub : i64, %step : i6 // CHECK: call void @__kmpc_dispatch_init_8u(ptr @{{.*}}, i32 %{{.*}}, i32 1073741859 // CHECK: %[[continue:.*]] = call i32 @__kmpc_dispatch_next_8u // CHECK: %[[cond:.*]] = icmp ne i32 %[[continue]], 0 - // CHECK br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} + // CHECK: br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} llvm.call @body(%iv) : (i64) -> () omp.yield } @@ -625,7 +625,7 @@ llvm.func @test_omp_wsloop_dynamic_monotonic(%lb : i64, %ub : i64, %step : i64) // CHECK: call void @__kmpc_dispatch_init_8u(ptr @{{.*}}, i32 %{{.*}}, i32 536870947 // CHECK: %[[continue:.*]] = call i32 @__kmpc_dispatch_next_8u // CHECK: %[[cond:.*]] = icmp ne i32 %[[continue]], 0 - // CHECK br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} + // CHECK: br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} llvm.call @body(%iv) : (i64) -> () omp.yield } @@ -644,7 +644,7 @@ llvm.func @test_omp_wsloop_runtime_simd(%lb : i64, %ub : i64, %step : i64) -> () // CHECK: call void @__kmpc_dispatch_init_8u(ptr @{{.*}}, i32 %{{.*}}, i32 1073741871 // CHECK: %[[continue:.*]] = call i32 @__kmpc_dispatch_next_8u // CHECK: %[[cond:.*]] = icmp ne i32 %[[continue]], 0 - // CHECK br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} + // CHECK: br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} llvm.call @body(%iv) : (i64) -> () omp.yield } @@ -663,7 +663,7 @@ llvm.func @test_omp_wsloop_guided_simd(%lb : i64, %ub : i64, %step : i64) -> () // CHECK: call void @__kmpc_dispatch_init_8u(ptr @{{.*}}, i32 %{{.*}}, i32 1073741870 // CHECK: %[[continue:.*]] = call i32 @__kmpc_dispatch_next_8u // CHECK: %[[cond:.*]] = icmp ne i32 %[[continue]], 0 - // CHECK br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} + // CHECK: br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} llvm.call @body(%iv) : (i64) -> () omp.yield } @@ -835,7 +835,7 @@ llvm.func @test_omp_wsloop_ordered(%lb : i64, %ub : i64, %step : i64) -> () { // CHECK: call void @__kmpc_dispatch_fini_8u // CHECK: %[[continue:.*]] = call i32 @__kmpc_dispatch_next_8u // CHECK: %[[cond:.*]] = icmp ne i32 %[[continue]], 0 - // CHECK br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} + // CHECK: br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} llvm.call @body(%iv) : (i64) -> () omp.yield } @@ -855,7 +855,7 @@ llvm.func @test_omp_wsloop_static_ordered(%lb : i64, %ub : i64, %step : i64) -> // CHECK: call void @__kmpc_dispatch_fini_8u // CHECK: %[[continue:.*]] = call i32 @__kmpc_dispatch_next_8u // CHECK: %[[cond:.*]] = icmp ne i32 %[[continue]], 0 - // CHECK br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} + // CHECK: br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} llvm.call @body(%iv) : (i64) -> () omp.yield } @@ -876,7 +876,7 @@ llvm.func @test_omp_wsloop_static_chunk_ordered(%lb : i32, %ub : i32, %step : i3 // CHECK: call void @__kmpc_dispatch_fini_4u // CHECK: %[[continue:.*]] = call i32 @__kmpc_dispatch_next_4u // CHECK: %[[cond:.*]] = icmp ne i32 %[[continue]], 0 - // CHECK br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} + // CHECK: br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} llvm.call @body(%iv) : (i32) -> () omp.yield } @@ -896,7 +896,7 @@ llvm.func @test_omp_wsloop_dynamic_ordered(%lb : i64, %ub : i64, %step : i64) -> // CHECK: call void @__kmpc_dispatch_fini_8u // CHECK: %[[continue:.*]] = call i32 @__kmpc_dispatch_next_8u // CHECK: %[[cond:.*]] = icmp ne i32 %[[continue]], 0 - // CHECK br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} + // CHECK: br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} llvm.call @body(%iv) : (i64) -> () omp.yield } @@ -916,7 +916,7 @@ llvm.func @test_omp_wsloop_auto_ordered(%lb : i64, %ub : i64, %step : i64) -> () // CHECK: call void @__kmpc_dispatch_fini_8u // CHECK: %[[continue:.*]] = call i32 @__kmpc_dispatch_next_8u // CHECK: %[[cond:.*]] = icmp ne i32 %[[continue]], 0 - // CHECK br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} + // CHECK: br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} llvm.call @body(%iv) : (i64) -> () omp.yield } @@ -936,7 +936,7 @@ llvm.func @test_omp_wsloop_runtime_ordered(%lb : i64, %ub : i64, %step : i64) -> // CHECK: call void @__kmpc_dispatch_fini_8u // CHECK: %[[continue:.*]] = call i32 @__kmpc_dispatch_next_8u // CHECK: %[[cond:.*]] = icmp ne i32 %[[continue]], 0 - // CHECK br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} + // CHECK: br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} llvm.call @body(%iv) : (i64) -> () omp.yield } @@ -956,7 +956,7 @@ llvm.func @test_omp_wsloop_guided_ordered(%lb : i64, %ub : i64, %step : i64) -> // CHECK: call void @__kmpc_dispatch_fini_8u // CHECK: %[[continue:.*]] = call i32 @__kmpc_dispatch_next_8u // CHECK: %[[cond:.*]] = icmp ne i32 %[[continue]], 0 - // CHECK br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} + // CHECK: br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} llvm.call @body(%iv) : (i64) -> () omp.yield } @@ -976,7 +976,7 @@ llvm.func @test_omp_wsloop_dynamic_nonmonotonic_ordered(%lb : i64, %ub : i64, %s // CHECK: call void @__kmpc_dispatch_fini_8u // CHECK: %[[continue:.*]] = call i32 @__kmpc_dispatch_next_8u // CHECK: %[[cond:.*]] = icmp ne i32 %[[continue]], 0 - // CHECK br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} + // CHECK: br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} llvm.call @body(%iv) : (i64) -> () omp.yield } @@ -996,7 +996,7 @@ llvm.func @test_omp_wsloop_dynamic_monotonic_ordered(%lb : i64, %ub : i64, %step // CHECK: call void @__kmpc_dispatch_fini_8u // CHECK: %[[continue:.*]] = call i32 @__kmpc_dispatch_next_8u // CHECK: %[[cond:.*]] = icmp ne i32 %[[continue]], 0 - // CHECK br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} + // CHECK: br i1 %[[cond]], label %omp_loop.header{{.*}}, label %omp_loop.exit{{.*}} llvm.call @body(%iv) : (i64) -> () omp.yield } diff --git a/mlir/test/Transforms/canonicalize.mlir b/mlir/test/Transforms/canonicalize.mlir index d2c2c12d32389..b34b87610a0e4 100644 --- a/mlir/test/Transforms/canonicalize.mlir +++ b/mlir/test/Transforms/canonicalize.mlir @@ -719,7 +719,7 @@ func.func @view(%arg0 : index) -> (f32, f32, f32, f32) { %r2 = memref.load %3[%c0, %c0] : memref // Test: folding static alloc and memref.cast into a view. - // CHECK memref.view %[[ALLOC_MEM]][%[[C15]]][] : memref<2048xi8> to memref<15x7xf32> + // CHECK: memref.view %[[ALLOC_MEM]][%[[C15]]][] : memref<2048xi8> to memref<15x7xf32> %4 = memref.cast %0 : memref<2048xi8> to memref %5 = memref.view %4[%c15][%c15, %c7] : memref to memref %r3 = memref.load %5[%c0, %c0] : memref diff --git a/mlir/test/mlir-cpu-runner/copy.mlir b/mlir/test/mlir-cpu-runner/copy.mlir index 8766f51050e09..7fa35fac70e4d 100644 --- a/mlir/test/mlir-cpu-runner/copy.mlir +++ b/mlir/test/mlir-cpu-runner/copy.mlir @@ -64,7 +64,7 @@ func.func @main() -> () { %unranked_scalar_copy = memref.cast %scalar_copy : memref to memref<*xf32> call @printMemrefF32(%unranked_scalar_copy) : (memref<*xf32>) -> () // CHECK: rank = 0 offset = 0 sizes = [] strides = [] - // CHECK-NEXT [42] + // CHECK-NEXT: [42] memref.dealloc %copy_empty : memref<3x0x1xf32> memref.dealloc %copy_empty_casted : memref<0x3x1xf32> diff --git a/mlir/test/mlir-tblgen/bytecode-reserved.td b/mlir/test/mlir-tblgen/bytecode-reserved.td index 1a7bcd306d075..5162a6b36317c 100644 --- a/mlir/test/mlir-tblgen/bytecode-reserved.td +++ b/mlir/test/mlir-tblgen/bytecode-reserved.td @@ -2,7 +2,7 @@ include "mlir/IR/BuiltinDialectBytecode.td" -// CHECK static ::mlir::Type readIntegerType +// CHECK: static ::mlir::Type readIntegerType def TestDialectTypes : DialectTypes<"Test"> { // CHECK: static Type readType diff --git a/mlir/test/python/ir/attributes.py b/mlir/test/python/ir/attributes.py index dbd6bad05e01d..3ccfa8842adf7 100644 --- a/mlir/test/python/ir/attributes.py +++ b/mlir/test/python/ir/attributes.py @@ -514,7 +514,7 @@ def testDictAttr(): a = DictAttr.get(dict_attr) - # CHECK attr: {integerattr = 42 : i32, stringattr = "string"} + # CHECK: attr: {integerattr = 42 : i32, stringattr = "string"} print("attr:", a) assert len(a) == 2 @@ -546,7 +546,7 @@ def testDictAttr(): else: assert False, "expected IndexError on accessing an out-of-bounds attribute" - # CHECK "empty: {}" + # CHECK: "empty: {}" print("empty: ", DictAttr.get()) From 23048b9b531f9d1d96fc48d14195f965abd89d6e Mon Sep 17 00:00:00 2001 From: klensy Date: Wed, 29 May 2024 10:33:15 +0300 Subject: [PATCH 2/2] review --- mlir/test/python/ir/attributes.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/mlir/test/python/ir/attributes.py b/mlir/test/python/ir/attributes.py index 3ccfa8842adf7..6c1ac88f32265 100644 --- a/mlir/test/python/ir/attributes.py +++ b/mlir/test/python/ir/attributes.py @@ -546,7 +546,7 @@ def testDictAttr(): else: assert False, "expected IndexError on accessing an out-of-bounds attribute" - # CHECK: "empty: {}" + # CHECK: empty: {} print("empty: ", DictAttr.get())