From 474ffd3e8bcced9a8204e07e7aa32c45c9fe6b53 Mon Sep 17 00:00:00 2001 From: Momchil Velikov Date: Tue, 19 Nov 2024 12:17:13 +0000 Subject: [PATCH 1/2] [AArch64] Generate zeroing forms of certain SVE2.2 instructions (6/11) --- .../lib/Target/AArch64/AArch64SVEInstrInfo.td | 10 +- llvm/lib/Target/AArch64/SVEInstrFormats.td | 7 +- .../AArch64/zeroing-forms-counts-not.ll | 1152 +++++++++++++++++ 3 files changed, 1163 insertions(+), 6 deletions(-) create mode 100644 llvm/test/CodeGen/AArch64/zeroing-forms-counts-not.ll diff --git a/llvm/lib/Target/AArch64/AArch64SVEInstrInfo.td b/llvm/lib/Target/AArch64/AArch64SVEInstrInfo.td index 7dd6d49bf2022..09689fe7e521b 100644 --- a/llvm/lib/Target/AArch64/AArch64SVEInstrInfo.td +++ b/llvm/lib/Target/AArch64/AArch64SVEInstrInfo.td @@ -4314,11 +4314,11 @@ let Predicates = [HasSVE2p2orSME2p2] in { defm FSQRT_ZPZz : sve_fp_z2op_p_zd_hsd<0b01101, "fsqrt">; // SVE2p2 integer unary arithmetic (bitwise), zeroing predicate - defm CLS_ZPzZ : sve_int_un_pred_arit_bitwise_z<0b000, "cls">; - defm CLZ_ZPzZ : sve_int_un_pred_arit_bitwise_z<0b001, "clz">; - defm CNT_ZPzZ : sve_int_un_pred_arit_bitwise_z<0b010, "cnt">; - defm CNOT_ZPzZ : sve_int_un_pred_arit_bitwise_z<0b011, "cnot">; - defm NOT_ZPzZ : sve_int_un_pred_arit_bitwise_z<0b110, "not">; + defm CLS_ZPzZ : sve_int_un_pred_arit_bitwise_z<0b000, "cls", AArch64cls_mt>; + defm CLZ_ZPzZ : sve_int_un_pred_arit_bitwise_z<0b001, "clz", AArch64clz_mt>; + defm CNT_ZPzZ : sve_int_un_pred_arit_bitwise_z<0b010, "cnt", AArch64cnt_mt>; + defm CNOT_ZPzZ : sve_int_un_pred_arit_bitwise_z<0b011, "cnot", AArch64cnot_mt>; + defm NOT_ZPzZ : sve_int_un_pred_arit_bitwise_z<0b110, " not", AArch64not_mt>; // floating point defm FABS_ZPzZ : sve_int_un_pred_arit_bitwise_fp_z<0b100, "fabs", AArch64fabs_mt>; diff --git a/llvm/lib/Target/AArch64/SVEInstrFormats.td b/llvm/lib/Target/AArch64/SVEInstrFormats.td index 0ef862fc1a27c..369ed1ab55875 100644 --- a/llvm/lib/Target/AArch64/SVEInstrFormats.td +++ b/llvm/lib/Target/AArch64/SVEInstrFormats.td @@ -4966,11 +4966,16 @@ multiclass sve_int_un_pred_arit_bitwise opc, string asm, defm : SVE_1_Op_PassthruUndef_Pat(NAME # _D_UNDEF)>; } -multiclass sve_int_un_pred_arit_bitwise_z opc, string asm> { +multiclass sve_int_un_pred_arit_bitwise_z opc, string asm, SDPatternOperator op> { def _B : sve_int_un_pred_arit_z<0b00, { opc, 0b1 }, asm, ZPR8>; def _H : sve_int_un_pred_arit_z<0b01, { opc, 0b1 }, asm, ZPR16>; def _S : sve_int_un_pred_arit_z<0b10, { opc, 0b1 }, asm, ZPR32>; def _D : sve_int_un_pred_arit_z<0b11, { opc, 0b1 }, asm, ZPR64>; + + def : SVE_1_Op_PassthruUndefZero_Pat(NAME # _B)>; + def : SVE_1_Op_PassthruUndefZero_Pat(NAME # _H)>; + def : SVE_1_Op_PassthruUndefZero_Pat(NAME # _S)>; + def : SVE_1_Op_PassthruUndefZero_Pat(NAME # _D)>; } multiclass sve_int_un_pred_arit_bitwise_fp opc, string asm, diff --git a/llvm/test/CodeGen/AArch64/zeroing-forms-counts-not.ll b/llvm/test/CodeGen/AArch64/zeroing-forms-counts-not.ll new file mode 100644 index 0000000000000..efdf4bd6a5c64 --- /dev/null +++ b/llvm/test/CodeGen/AArch64/zeroing-forms-counts-not.ll @@ -0,0 +1,1152 @@ +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 5 +; RUN: llc -mattr=+bf16,+sve < %s | FileCheck %s +; RUN: llc -mattr=+bf16,+sve2p2 < %s | FileCheck %s -check-prefix CHECK-2p2 + +; RUN: llc -mattr=+bf16,+sme -force-streaming < %s | FileCheck %s +; RUN: llc -mattr=+bf16,+sme2p2 -force-streaming < %s | FileCheck %s -check-prefix CHECK-2p2 + +target triple = "aarch64-linux" + +define @test_svcls_s8_x_1( %pg, %x) { +; CHECK-LABEL: test_svcls_s8_x_1: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: cls z0.b, p0/m, z0.b +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcls_s8_x_1: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cls z0.b, p0/z, z0.b +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cls.nxv16i8( undef, %pg, %x) + ret %0 +} + +define @test_svcls_s8_x_2( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcls_s8_x_2: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cls z0.b, p0/m, z1.b +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcls_s8_x_2: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cls z0.b, p0/z, z1.b +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cls.nxv16i8( undef, %pg, %x) + ret %0 +} + +define @test_svcls_s8_z( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcls_s8_z: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: mov z0.b, #0 // =0x0 +; CHECK-NEXT: cls z0.b, p0/m, z1.b +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcls_s8_z: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cls z0.b, p0/z, z1.b +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cls.nxv16i8( zeroinitializer, %pg, %x) + ret %0 +} + +define @test_svcls_s16_x_1( %pg, %x) { +; CHECK-LABEL: test_svcls_s16_x_1: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: cls z0.h, p0/m, z0.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcls_s16_x_1: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cls z0.h, p0/z, z0.h +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cls.nxv8i16( undef, %pg, %x) + ret %0 +} + +define @test_svcls_s16_x_2( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcls_s16_x_2: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cls z0.h, p0/m, z1.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcls_s16_x_2: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cls z0.h, p0/z, z1.h +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cls.nxv8i16( undef, %pg, %x) + ret %0 +} + +define @test_svcls_s16_z( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcls_s16_z: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: mov z0.h, #0 // =0x0 +; CHECK-NEXT: cls z0.h, p0/m, z1.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcls_s16_z: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cls z0.h, p0/z, z1.h +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cls.nxv8i16( zeroinitializer, %pg, %x) + ret %0 +} + +define @test_svcls_s32_x_1( %pg, %x) { +; CHECK-LABEL: test_svcls_s32_x_1: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: cls z0.s, p0/m, z0.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcls_s32_x_1: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cls z0.s, p0/z, z0.s +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cls.nxv4i32( undef, %pg, %x) + ret %0 +} + +define @test_svcls_s32_x_2( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcls_s32_x_2: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cls z0.s, p0/m, z1.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcls_s32_x_2: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cls z0.s, p0/z, z1.s +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cls.nxv4i32( undef, %pg, %x) + ret %0 +} + +define @test_svcls_s32_z( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcls_s32_z: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: mov z0.s, #0 // =0x0 +; CHECK-NEXT: cls z0.s, p0/m, z1.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcls_s32_z: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cls z0.s, p0/z, z1.s +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cls.nxv4i32( zeroinitializer, %pg, %x) + ret %0 +} + +define @test_svcls_s64_x_1( %pg, %x) { +; CHECK-LABEL: test_svcls_s64_x_1: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: cls z0.d, p0/m, z0.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcls_s64_x_1: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cls z0.d, p0/z, z0.d +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cls.nxv2i64( undef, %pg, %x) + ret %0 +} + +define @test_svcls_s64_x_2( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcls_s64_x_2: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cls z0.d, p0/m, z1.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcls_s64_x_2: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cls z0.d, p0/z, z1.d +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cls.nxv2i64( undef, %pg, %x) + ret %0 +} + +define @test_svcls_s64_z( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcls_s64_z: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: mov z0.d, #0 // =0x0 +; CHECK-NEXT: cls z0.d, p0/m, z1.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcls_s64_z: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cls z0.d, p0/z, z1.d +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cls.nxv2i64( zeroinitializer, %pg, %x) + ret %0 +} + +define @test_svclz_s8_x_1( %pg, %x) { +; CHECK-LABEL: test_svclz_s8_x_1: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: clz z0.b, p0/m, z0.b +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svclz_s8_x_1: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: clz z0.b, p0/z, z0.b +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.clz.nxv16i8( undef, %pg, %x) + ret %0 +} + +define @test_svclz_s8_x_2( %pg, double %z0, %x) { +; CHECK-LABEL: test_svclz_s8_x_2: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: clz z0.b, p0/m, z1.b +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svclz_s8_x_2: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: clz z0.b, p0/z, z1.b +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.clz.nxv16i8( undef, %pg, %x) + ret %0 +} + +define @test_svclz_s8_z( %pg, double %z0, %x) { +; CHECK-LABEL: test_svclz_s8_z: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: mov z0.b, #0 // =0x0 +; CHECK-NEXT: clz z0.b, p0/m, z1.b +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svclz_s8_z: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: clz z0.b, p0/z, z1.b +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.clz.nxv16i8( zeroinitializer, %pg, %x) + ret %0 +} + +define @test_svclz_s16_x_1( %pg, %x) { +; CHECK-LABEL: test_svclz_s16_x_1: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: clz z0.h, p0/m, z0.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svclz_s16_x_1: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: clz z0.h, p0/z, z0.h +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.clz.nxv8i16( undef, %pg, %x) + ret %0 +} + +define @test_svclz_s16_x_2( %pg, double %z0, %x) { +; CHECK-LABEL: test_svclz_s16_x_2: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: clz z0.h, p0/m, z1.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svclz_s16_x_2: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: clz z0.h, p0/z, z1.h +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.clz.nxv8i16( undef, %pg, %x) + ret %0 +} + +define @test_svclz_s16_z( %pg, double %z0, %x) { +; CHECK-LABEL: test_svclz_s16_z: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: mov z0.h, #0 // =0x0 +; CHECK-NEXT: clz z0.h, p0/m, z1.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svclz_s16_z: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: clz z0.h, p0/z, z1.h +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.clz.nxv8i16( zeroinitializer, %pg, %x) + ret %0 +} + +define @test_svclz_s32_x_1( %pg, %x) { +; CHECK-LABEL: test_svclz_s32_x_1: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: clz z0.s, p0/m, z0.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svclz_s32_x_1: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: clz z0.s, p0/z, z0.s +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.clz.nxv4i32( undef, %pg, %x) + ret %0 +} + +define @test_svclz_s32_x_2( %pg, double %z0, %x) { +; CHECK-LABEL: test_svclz_s32_x_2: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: clz z0.s, p0/m, z1.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svclz_s32_x_2: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: clz z0.s, p0/z, z1.s +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.clz.nxv4i32( undef, %pg, %x) + ret %0 +} + +define @test_svclz_s32_z( %pg, double %z0, %x) { +; CHECK-LABEL: test_svclz_s32_z: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: mov z0.s, #0 // =0x0 +; CHECK-NEXT: clz z0.s, p0/m, z1.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svclz_s32_z: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: clz z0.s, p0/z, z1.s +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.clz.nxv4i32( zeroinitializer, %pg, %x) + ret %0 +} + +define @test_svclz_s64_x_1( %pg, %x) { +; CHECK-LABEL: test_svclz_s64_x_1: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: clz z0.d, p0/m, z0.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svclz_s64_x_1: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: clz z0.d, p0/z, z0.d +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.clz.nxv2i64( undef, %pg, %x) + ret %0 +} + +define @test_svclz_s64_x_2( %pg, double %z0, %x) { +; CHECK-LABEL: test_svclz_s64_x_2: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: clz z0.d, p0/m, z1.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svclz_s64_x_2: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: clz z0.d, p0/z, z1.d +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.clz.nxv2i64( undef, %pg, %x) + ret %0 +} + +define @test_svclz_s64_z( %pg, double %z0, %x) { +; CHECK-LABEL: test_svclz_s64_z: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: mov z0.d, #0 // =0x0 +; CHECK-NEXT: clz z0.d, p0/m, z1.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svclz_s64_z: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: clz z0.d, p0/z, z1.d +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.clz.nxv2i64( zeroinitializer, %pg, %x) + ret %0 +} + +define @test_svcnt_s8_x_1( %pg, %x) { +; CHECK-LABEL: test_svcnt_s8_x_1: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: cnt z0.b, p0/m, z0.b +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_s8_x_1: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnt z0.b, p0/z, z0.b +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnt.nxv16i8( undef, %pg, %x) + ret %0 +} + +define @test_svcnt_s8_x_2( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcnt_s8_x_2: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cnt z0.b, p0/m, z1.b +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_s8_x_2: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnt z0.b, p0/z, z1.b +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnt.nxv16i8( undef, %pg, %x) + ret %0 +} + +define @test_svcnt_s8_z( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcnt_s8_z: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: mov z0.b, #0 // =0x0 +; CHECK-NEXT: cnt z0.b, p0/m, z1.b +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_s8_z: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnt z0.b, p0/z, z1.b +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnt.nxv16i8( zeroinitializer, %pg, %x) + ret %0 +} + +define @test_svcnt_s16_x_1( %pg, %x) { +; CHECK-LABEL: test_svcnt_s16_x_1: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: cnt z0.h, p0/m, z0.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_s16_x_1: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnt z0.h, p0/z, z0.h +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnt.nxv8i16( undef, %pg, %x) + ret %0 +} + +define @test_svcnt_s16_x_2( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcnt_s16_x_2: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cnt z0.h, p0/m, z1.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_s16_x_2: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnt z0.h, p0/z, z1.h +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnt.nxv8i16( undef, %pg, %x) + ret %0 +} + +define @test_svcnt_s16_z( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcnt_s16_z: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: mov z0.h, #0 // =0x0 +; CHECK-NEXT: cnt z0.h, p0/m, z1.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_s16_z: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnt z0.h, p0/z, z1.h +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnt.nxv8i16( zeroinitializer, %pg, %x) + ret %0 +} + +define @test_svcnt_s32_x_1( %pg, %x) { +; CHECK-LABEL: test_svcnt_s32_x_1: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: cnt z0.s, p0/m, z0.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_s32_x_1: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnt z0.s, p0/z, z0.s +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnt.nxv4i32( undef, %pg, %x) + ret %0 +} + +define @test_svcnt_s32_x_2( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcnt_s32_x_2: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cnt z0.s, p0/m, z1.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_s32_x_2: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnt z0.s, p0/z, z1.s +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnt.nxv4i32( undef, %pg, %x) + ret %0 +} + +define @test_svcnt_s32_z( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcnt_s32_z: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: mov z0.s, #0 // =0x0 +; CHECK-NEXT: cnt z0.s, p0/m, z1.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_s32_z: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnt z0.s, p0/z, z1.s +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnt.nxv4i32( zeroinitializer, %pg, %x) + ret %0 +} + +define @test_svcnt_s64_x_1( %pg, %x) { +; CHECK-LABEL: test_svcnt_s64_x_1: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: cnt z0.d, p0/m, z0.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_s64_x_1: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnt z0.d, p0/z, z0.d +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnt.nxv2i64( undef, %pg, %x) + ret %0 +} + +define @test_svcnt_s64_x_2( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcnt_s64_x_2: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cnt z0.d, p0/m, z1.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_s64_x_2: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnt z0.d, p0/z, z1.d +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnt.nxv2i64( undef, %pg, %x) + ret %0 +} + +define @test_svcnt_s64_z( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcnt_s64_z: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: mov z0.d, #0 // =0x0 +; CHECK-NEXT: cnt z0.d, p0/m, z1.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_s64_z: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnt z0.d, p0/z, z1.d +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnt.nxv2i64( zeroinitializer, %pg, %x) + ret %0 +} + +define @test_svcnt_f16_x_1( %pg, %x) { +; CHECK-LABEL: test_svcnt_f16_x_1: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: cnt z0.h, p0/m, z0.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_f16_x_1: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnt z0.h, p0/z, z0.h +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnt.nxv8f16( undef, %pg, %x) + ret %0 +} + +define @test_svcnt_f16_x_2( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcnt_f16_x_2: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cnt z0.h, p0/m, z1.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_f16_x_2: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnt z0.h, p0/z, z1.h +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnt.nxv8f16( undef, %pg, %x) + ret %0 +} + +define @test_svcnt_f16_z( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcnt_f16_z: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: mov z0.h, #0 // =0x0 +; CHECK-NEXT: cnt z0.h, p0/m, z1.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_f16_z: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnt z0.h, p0/z, z1.h +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnt.nxv8f16( zeroinitializer, %pg, %x) + ret %0 +} + +define @test_svcnt_bf16_x_1( %pg, %x) { +; CHECK-LABEL: test_svcnt_bf16_x_1: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: cnt z0.h, p0/m, z0.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_bf16_x_1: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnt z0.h, p0/z, z0.h +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnt.nxv8bf16( undef, %pg, %x) + ret %0 +} + +define @test_svcnt_bf16_x_2( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcnt_bf16_x_2: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cnt z0.h, p0/m, z1.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_bf16_x_2: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnt z0.h, p0/z, z1.h +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnt.nxv8bf16( undef, %pg, %x) + ret %0 +} + +define @test_svcnt_bf16_z( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcnt_bf16_z: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: mov z0.h, #0 // =0x0 +; CHECK-NEXT: cnt z0.h, p0/m, z1.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_bf16_z: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnt z0.h, p0/z, z1.h +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnt.nxv8bf16( zeroinitializer, %pg, %x) + ret %0 +} + +define @test_svcnt_f32_x_1( %pg, %x) { +; CHECK-LABEL: test_svcnt_f32_x_1: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: cnt z0.s, p0/m, z0.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_f32_x_1: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnt z0.s, p0/z, z0.s +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnt.nxv4f32( undef, %pg, %x) + ret %0 +} + +define @test_svcnt_f32_x_2( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcnt_f32_x_2: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cnt z0.s, p0/m, z1.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_f32_x_2: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnt z0.s, p0/z, z1.s +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnt.nxv4f32( undef, %pg, %x) + ret %0 +} + +define @test_svcnt_f32_z( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcnt_f32_z: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: mov z0.s, #0 // =0x0 +; CHECK-NEXT: cnt z0.s, p0/m, z1.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_f32_z: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnt z0.s, p0/z, z1.s +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnt.nxv4f32( zeroinitializer, %pg, %x) + ret %0 +} + +define @test_svcnt_f64_x_1( %pg, %x) { +; CHECK-LABEL: test_svcnt_f64_x_1: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: cnt z0.d, p0/m, z0.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_f64_x_1: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnt z0.d, p0/z, z0.d +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnt.nxv2f64( undef, %pg, %x) + ret %0 +} + +define @test_svcnt_f64_x_2( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcnt_f64_x_2: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cnt z0.d, p0/m, z1.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_f64_x_2: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnt z0.d, p0/z, z1.d +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnt.nxv2f64( undef, %pg, %x) + ret %0 +} + +define @test_svcnt_f64_z( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcnt_f64_z: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: mov z0.d, #0 // =0x0 +; CHECK-NEXT: cnt z0.d, p0/m, z1.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_f64_z: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnt z0.d, p0/z, z1.d +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnt.nxv2f64( zeroinitializer, %pg, %x) + ret %0 +} + +define @test_svcnot_s8_x_1( %pg, %x) { +; CHECK-LABEL: test_svcnot_s8_x_1: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: cnot z0.b, p0/m, z0.b +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnot_s8_x_1: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnot z0.b, p0/z, z0.b +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnot.nxv16i8( undef, %pg, %x) + ret %0 +} + +define @test_svcnot_s8_x_2( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcnot_s8_x_2: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cnot z0.b, p0/m, z1.b +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnot_s8_x_2: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnot z0.b, p0/z, z1.b +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnot.nxv16i8( undef, %pg, %x) + ret %0 +} + +define @test_svcnot_s8_z( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcnot_s8_z: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: mov z0.b, #0 // =0x0 +; CHECK-NEXT: cnot z0.b, p0/m, z1.b +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnot_s8_z: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnot z0.b, p0/z, z1.b +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnot.nxv16i8( zeroinitializer, %pg, %x) + ret %0 +} + +define @test_svcnot_s16_x_1( %pg, %x) { +; CHECK-LABEL: test_svcnot_s16_x_1: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: cnot z0.h, p0/m, z0.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnot_s16_x_1: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnot z0.h, p0/z, z0.h +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnot.nxv8i16( undef, %pg, %x) + ret %0 +} + +define @test_svcnot_s16_x_2( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcnot_s16_x_2: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cnot z0.h, p0/m, z1.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnot_s16_x_2: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnot z0.h, p0/z, z1.h +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnot.nxv8i16( undef, %pg, %x) + ret %0 +} + +define @test_svcnot_s16_z( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcnot_s16_z: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: mov z0.h, #0 // =0x0 +; CHECK-NEXT: cnot z0.h, p0/m, z1.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnot_s16_z: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnot z0.h, p0/z, z1.h +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnot.nxv8i16( zeroinitializer, %pg, %x) + ret %0 +} + +define @test_svcnot_s32_x_1( %pg, %x) { +; CHECK-LABEL: test_svcnot_s32_x_1: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: cnot z0.s, p0/m, z0.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnot_s32_x_1: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnot z0.s, p0/z, z0.s +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnot.nxv4i32( undef, %pg, %x) + ret %0 +} + +define @test_svcnot_s32_x_2( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcnot_s32_x_2: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cnot z0.s, p0/m, z1.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnot_s32_x_2: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnot z0.s, p0/z, z1.s +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnot.nxv4i32( undef, %pg, %x) + ret %0 +} + +define @test_svcnot_s32_z( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcnot_s32_z: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: mov z0.s, #0 // =0x0 +; CHECK-NEXT: cnot z0.s, p0/m, z1.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnot_s32_z: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnot z0.s, p0/z, z1.s +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnot.nxv4i32( zeroinitializer, %pg, %x) + ret %0 +} + +define @test_svcnot_s64_x_1( %pg, %x) { +; CHECK-LABEL: test_svcnot_s64_x_1: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: cnot z0.d, p0/m, z0.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnot_s64_x_1: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnot z0.d, p0/z, z0.d +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnot.nxv2i64( undef, %pg, %x) + ret %0 +} + +define @test_svcnot_s64_x_2( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcnot_s64_x_2: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cnot z0.d, p0/m, z1.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnot_s64_x_2: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnot z0.d, p0/z, z1.d +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnot.nxv2i64( undef, %pg, %x) + ret %0 +} + +define @test_svcnot_s64_z( %pg, double %z0, %x) { +; CHECK-LABEL: test_svcnot_s64_z: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: mov z0.d, #0 // =0x0 +; CHECK-NEXT: cnot z0.d, p0/m, z1.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnot_s64_z: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: cnot z0.d, p0/z, z1.d +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.cnot.nxv2i64( zeroinitializer, %pg, %x) + ret %0 +} + +define @test_svnot_s8_x_1( %pg, %x) { +; CHECK-LABEL: test_svnot_s8_x_1: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: not z0.b, p0/m, z0.b +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svnot_s8_x_1: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: not z0.b, p0/z, z0.b +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.not.nxv16i8( undef, %pg, %x) + ret %0 +} + +define @test_svnot_s8_x_2( %pg, double %z0, %x) { +; CHECK-LABEL: test_svnot_s8_x_2: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: not z0.b, p0/m, z1.b +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svnot_s8_x_2: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: not z0.b, p0/z, z1.b +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.not.nxv16i8( undef, %pg, %x) + ret %0 +} + +define @test_svnot_s8_z( %pg, double %z0, %x) { +; CHECK-LABEL: test_svnot_s8_z: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: mov z0.b, #0 // =0x0 +; CHECK-NEXT: not z0.b, p0/m, z1.b +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svnot_s8_z: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: not z0.b, p0/z, z1.b +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.not.nxv16i8( zeroinitializer, %pg, %x) + ret %0 +} + +define @test_svnot_s16_x_1( %pg, %x) { +; CHECK-LABEL: test_svnot_s16_x_1: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: not z0.h, p0/m, z0.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svnot_s16_x_1: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: not z0.h, p0/z, z0.h +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.not.nxv8i16( undef, %pg, %x) + ret %0 +} + +define @test_svnot_s16_x_2( %pg, double %z0, %x) { +; CHECK-LABEL: test_svnot_s16_x_2: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: not z0.h, p0/m, z1.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svnot_s16_x_2: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: not z0.h, p0/z, z1.h +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.not.nxv8i16( undef, %pg, %x) + ret %0 +} + +define @test_svnot_s16_z( %pg, double %z0, %x) { +; CHECK-LABEL: test_svnot_s16_z: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: mov z0.h, #0 // =0x0 +; CHECK-NEXT: not z0.h, p0/m, z1.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svnot_s16_z: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: not z0.h, p0/z, z1.h +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.not.nxv8i16( zeroinitializer, %pg, %x) + ret %0 +} + +define @test_svnot_s32_x_1( %pg, %x) { +; CHECK-LABEL: test_svnot_s32_x_1: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: not z0.s, p0/m, z0.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svnot_s32_x_1: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: not z0.s, p0/z, z0.s +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.not.nxv4i32( undef, %pg, %x) + ret %0 +} + +define @test_svnot_s32_x_2( %pg, double %z0, %x) { +; CHECK-LABEL: test_svnot_s32_x_2: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: not z0.s, p0/m, z1.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svnot_s32_x_2: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: not z0.s, p0/z, z1.s +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.not.nxv4i32( undef, %pg, %x) + ret %0 +} + +define @test_svnot_s32_z( %pg, double %z0, %x) { +; CHECK-LABEL: test_svnot_s32_z: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: mov z0.s, #0 // =0x0 +; CHECK-NEXT: not z0.s, p0/m, z1.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svnot_s32_z: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: not z0.s, p0/z, z1.s +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.not.nxv4i32( zeroinitializer, %pg, %x) + ret %0 +} + +define @test_svnot_s64_x_1( %pg, %x) { +; CHECK-LABEL: test_svnot_s64_x_1: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: not z0.d, p0/m, z0.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svnot_s64_x_1: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: not z0.d, p0/z, z0.d +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.not.nxv2i64( undef, %pg, %x) + ret %0 +} + +define @test_svnot_s64_x_2( %pg, double %z0, %x) { +; CHECK-LABEL: test_svnot_s64_x_2: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: not z0.d, p0/m, z1.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svnot_s64_x_2: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: not z0.d, p0/z, z1.d +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.not.nxv2i64( undef, %pg, %x) + ret %0 +} + +define @test_svnot_s64_z( %pg, double %z0, %x) { +; CHECK-LABEL: test_svnot_s64_z: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: mov z0.d, #0 // =0x0 +; CHECK-NEXT: not z0.d, p0/m, z1.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svnot_s64_z: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: not z0.d, p0/z, z1.d +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.not.nxv2i64( zeroinitializer, %pg, %x) + ret %0 +} + +define @test_svnot_u64_z( %pg, double %z0, %x) { +; CHECK-LABEL: test_svnot_u64_z: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: mov z0.d, #0 // =0x0 +; CHECK-NEXT: not z0.d, p0/m, z1.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svnot_u64_z: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: not z0.d, p0/z, z1.d +; CHECK-2p2-NEXT: ret +entry: + %0 = tail call @llvm.aarch64.sve.not.nxv2i64( zeroinitializer, %pg, %x) + ret %0 +} From 674c7de7087cbe204eea2cc139e81f030bc18731 Mon Sep 17 00:00:00 2001 From: Momchil Velikov Date: Fri, 20 Dec 2024 16:34:33 +0000 Subject: [PATCH 2/2] [fixup] Rebase, update tests, add tests for the new all-true patterns, change undef to poison --- .../lib/Target/AArch64/AArch64SVEInstrInfo.td | 2 +- llvm/lib/Target/AArch64/SVEInstrFormats.td | 8 +- .../AArch64/zeroing-forms-counts-not.ll | 1002 ++++++++++++++++- 3 files changed, 954 insertions(+), 58 deletions(-) diff --git a/llvm/lib/Target/AArch64/AArch64SVEInstrInfo.td b/llvm/lib/Target/AArch64/AArch64SVEInstrInfo.td index 09689fe7e521b..8d91fcf612f94 100644 --- a/llvm/lib/Target/AArch64/AArch64SVEInstrInfo.td +++ b/llvm/lib/Target/AArch64/AArch64SVEInstrInfo.td @@ -4318,7 +4318,7 @@ let Predicates = [HasSVE2p2orSME2p2] in { defm CLZ_ZPzZ : sve_int_un_pred_arit_bitwise_z<0b001, "clz", AArch64clz_mt>; defm CNT_ZPzZ : sve_int_un_pred_arit_bitwise_z<0b010, "cnt", AArch64cnt_mt>; defm CNOT_ZPzZ : sve_int_un_pred_arit_bitwise_z<0b011, "cnot", AArch64cnot_mt>; - defm NOT_ZPzZ : sve_int_un_pred_arit_bitwise_z<0b110, " not", AArch64not_mt>; + defm NOT_ZPzZ : sve_int_un_pred_arit_bitwise_z<0b110, "not", AArch64not_mt>; // floating point defm FABS_ZPzZ : sve_int_un_pred_arit_bitwise_fp_z<0b100, "fabs", AArch64fabs_mt>; diff --git a/llvm/lib/Target/AArch64/SVEInstrFormats.td b/llvm/lib/Target/AArch64/SVEInstrFormats.td index 369ed1ab55875..c6c8dfa5d547b 100644 --- a/llvm/lib/Target/AArch64/SVEInstrFormats.td +++ b/llvm/lib/Target/AArch64/SVEInstrFormats.td @@ -4972,10 +4972,10 @@ multiclass sve_int_un_pred_arit_bitwise_z opc, string asm, SDPatternOper def _S : sve_int_un_pred_arit_z<0b10, { opc, 0b1 }, asm, ZPR32>; def _D : sve_int_un_pred_arit_z<0b11, { opc, 0b1 }, asm, ZPR64>; - def : SVE_1_Op_PassthruUndefZero_Pat(NAME # _B)>; - def : SVE_1_Op_PassthruUndefZero_Pat(NAME # _H)>; - def : SVE_1_Op_PassthruUndefZero_Pat(NAME # _S)>; - def : SVE_1_Op_PassthruUndefZero_Pat(NAME # _D)>; + defm : SVE_1_Op_PassthruUndefZero_Pat(NAME # _B)>; + defm : SVE_1_Op_PassthruUndefZero_Pat(NAME # _H)>; + defm : SVE_1_Op_PassthruUndefZero_Pat(NAME # _S)>; + defm : SVE_1_Op_PassthruUndefZero_Pat(NAME # _D)>; } multiclass sve_int_un_pred_arit_bitwise_fp opc, string asm, diff --git a/llvm/test/CodeGen/AArch64/zeroing-forms-counts-not.ll b/llvm/test/CodeGen/AArch64/zeroing-forms-counts-not.ll index efdf4bd6a5c64..f7970ca81f608 100644 --- a/llvm/test/CodeGen/AArch64/zeroing-forms-counts-not.ll +++ b/llvm/test/CodeGen/AArch64/zeroing-forms-counts-not.ll @@ -18,7 +18,7 @@ define @test_svcls_s8_x_1( %pg, @llvm.aarch64.sve.cls.nxv16i8( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cls.nxv16i8( poison, %pg, %x) ret %0 } @@ -34,7 +34,7 @@ define @test_svcls_s8_x_2( %pg, double %z0, ; CHECK-2p2-NEXT: cls z0.b, p0/z, z1.b ; CHECK-2p2-NEXT: ret entry: - %0 = tail call @llvm.aarch64.sve.cls.nxv16i8( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cls.nxv16i8( poison, %pg, %x) ret %0 } @@ -65,7 +65,7 @@ define @test_svcls_s16_x_1( %pg, @llvm.aarch64.sve.cls.nxv8i16( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cls.nxv8i16( poison, %pg, %x) ret %0 } @@ -81,7 +81,7 @@ define @test_svcls_s16_x_2( %pg, double %z0, ; CHECK-2p2-NEXT: cls z0.h, p0/z, z1.h ; CHECK-2p2-NEXT: ret entry: - %0 = tail call @llvm.aarch64.sve.cls.nxv8i16( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cls.nxv8i16( poison, %pg, %x) ret %0 } @@ -112,7 +112,7 @@ define @test_svcls_s32_x_1( %pg, @llvm.aarch64.sve.cls.nxv4i32( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cls.nxv4i32( poison, %pg, %x) ret %0 } @@ -128,7 +128,7 @@ define @test_svcls_s32_x_2( %pg, double %z0, ; CHECK-2p2-NEXT: cls z0.s, p0/z, z1.s ; CHECK-2p2-NEXT: ret entry: - %0 = tail call @llvm.aarch64.sve.cls.nxv4i32( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cls.nxv4i32( poison, %pg, %x) ret %0 } @@ -159,7 +159,7 @@ define @test_svcls_s64_x_1( %pg, @llvm.aarch64.sve.cls.nxv2i64( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cls.nxv2i64( poison, %pg, %x) ret %0 } @@ -175,7 +175,7 @@ define @test_svcls_s64_x_2( %pg, double %z0, ; CHECK-2p2-NEXT: cls z0.d, p0/z, z1.d ; CHECK-2p2-NEXT: ret entry: - %0 = tail call @llvm.aarch64.sve.cls.nxv2i64( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cls.nxv2i64( poison, %pg, %x) ret %0 } @@ -206,7 +206,7 @@ define @test_svclz_s8_x_1( %pg, @llvm.aarch64.sve.clz.nxv16i8( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.clz.nxv16i8( poison, %pg, %x) ret %0 } @@ -222,7 +222,7 @@ define @test_svclz_s8_x_2( %pg, double %z0, ; CHECK-2p2-NEXT: clz z0.b, p0/z, z1.b ; CHECK-2p2-NEXT: ret entry: - %0 = tail call @llvm.aarch64.sve.clz.nxv16i8( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.clz.nxv16i8( poison, %pg, %x) ret %0 } @@ -253,7 +253,7 @@ define @test_svclz_s16_x_1( %pg, @llvm.aarch64.sve.clz.nxv8i16( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.clz.nxv8i16( poison, %pg, %x) ret %0 } @@ -269,7 +269,7 @@ define @test_svclz_s16_x_2( %pg, double %z0, ; CHECK-2p2-NEXT: clz z0.h, p0/z, z1.h ; CHECK-2p2-NEXT: ret entry: - %0 = tail call @llvm.aarch64.sve.clz.nxv8i16( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.clz.nxv8i16( poison, %pg, %x) ret %0 } @@ -300,7 +300,7 @@ define @test_svclz_s32_x_1( %pg, @llvm.aarch64.sve.clz.nxv4i32( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.clz.nxv4i32( poison, %pg, %x) ret %0 } @@ -316,7 +316,7 @@ define @test_svclz_s32_x_2( %pg, double %z0, ; CHECK-2p2-NEXT: clz z0.s, p0/z, z1.s ; CHECK-2p2-NEXT: ret entry: - %0 = tail call @llvm.aarch64.sve.clz.nxv4i32( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.clz.nxv4i32( poison, %pg, %x) ret %0 } @@ -347,7 +347,7 @@ define @test_svclz_s64_x_1( %pg, @llvm.aarch64.sve.clz.nxv2i64( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.clz.nxv2i64( poison, %pg, %x) ret %0 } @@ -363,7 +363,7 @@ define @test_svclz_s64_x_2( %pg, double %z0, ; CHECK-2p2-NEXT: clz z0.d, p0/z, z1.d ; CHECK-2p2-NEXT: ret entry: - %0 = tail call @llvm.aarch64.sve.clz.nxv2i64( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.clz.nxv2i64( poison, %pg, %x) ret %0 } @@ -394,7 +394,7 @@ define @test_svcnt_s8_x_1( %pg, @llvm.aarch64.sve.cnt.nxv16i8( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cnt.nxv16i8( poison, %pg, %x) ret %0 } @@ -410,7 +410,7 @@ define @test_svcnt_s8_x_2( %pg, double %z0, ; CHECK-2p2-NEXT: cnt z0.b, p0/z, z1.b ; CHECK-2p2-NEXT: ret entry: - %0 = tail call @llvm.aarch64.sve.cnt.nxv16i8( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cnt.nxv16i8( poison, %pg, %x) ret %0 } @@ -441,7 +441,7 @@ define @test_svcnt_s16_x_1( %pg, @llvm.aarch64.sve.cnt.nxv8i16( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cnt.nxv8i16( poison, %pg, %x) ret %0 } @@ -457,7 +457,7 @@ define @test_svcnt_s16_x_2( %pg, double %z0, ; CHECK-2p2-NEXT: cnt z0.h, p0/z, z1.h ; CHECK-2p2-NEXT: ret entry: - %0 = tail call @llvm.aarch64.sve.cnt.nxv8i16( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cnt.nxv8i16( poison, %pg, %x) ret %0 } @@ -488,7 +488,7 @@ define @test_svcnt_s32_x_1( %pg, @llvm.aarch64.sve.cnt.nxv4i32( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cnt.nxv4i32( poison, %pg, %x) ret %0 } @@ -504,7 +504,7 @@ define @test_svcnt_s32_x_2( %pg, double %z0, ; CHECK-2p2-NEXT: cnt z0.s, p0/z, z1.s ; CHECK-2p2-NEXT: ret entry: - %0 = tail call @llvm.aarch64.sve.cnt.nxv4i32( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cnt.nxv4i32( poison, %pg, %x) ret %0 } @@ -535,7 +535,7 @@ define @test_svcnt_s64_x_1( %pg, @llvm.aarch64.sve.cnt.nxv2i64( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cnt.nxv2i64( poison, %pg, %x) ret %0 } @@ -551,7 +551,7 @@ define @test_svcnt_s64_x_2( %pg, double %z0, ; CHECK-2p2-NEXT: cnt z0.d, p0/z, z1.d ; CHECK-2p2-NEXT: ret entry: - %0 = tail call @llvm.aarch64.sve.cnt.nxv2i64( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cnt.nxv2i64( poison, %pg, %x) ret %0 } @@ -582,7 +582,7 @@ define @test_svcnt_f16_x_1( %pg, @llvm.aarch64.sve.cnt.nxv8f16( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cnt.nxv8f16( poison, %pg, %x) ret %0 } @@ -598,7 +598,7 @@ define @test_svcnt_f16_x_2( %pg, double %z0, ; CHECK-2p2-NEXT: cnt z0.h, p0/z, z1.h ; CHECK-2p2-NEXT: ret entry: - %0 = tail call @llvm.aarch64.sve.cnt.nxv8f16( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cnt.nxv8f16( poison, %pg, %x) ret %0 } @@ -629,7 +629,7 @@ define @test_svcnt_bf16_x_1( %pg, @llvm.aarch64.sve.cnt.nxv8bf16( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cnt.nxv8bf16( poison, %pg, %x) ret %0 } @@ -645,7 +645,7 @@ define @test_svcnt_bf16_x_2( %pg, double %z0 ; CHECK-2p2-NEXT: cnt z0.h, p0/z, z1.h ; CHECK-2p2-NEXT: ret entry: - %0 = tail call @llvm.aarch64.sve.cnt.nxv8bf16( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cnt.nxv8bf16( poison, %pg, %x) ret %0 } @@ -676,7 +676,7 @@ define @test_svcnt_f32_x_1( %pg, @llvm.aarch64.sve.cnt.nxv4f32( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cnt.nxv4f32( poison, %pg, %x) ret %0 } @@ -692,7 +692,7 @@ define @test_svcnt_f32_x_2( %pg, double %z0, ; CHECK-2p2-NEXT: cnt z0.s, p0/z, z1.s ; CHECK-2p2-NEXT: ret entry: - %0 = tail call @llvm.aarch64.sve.cnt.nxv4f32( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cnt.nxv4f32( poison, %pg, %x) ret %0 } @@ -723,7 +723,7 @@ define @test_svcnt_f64_x_1( %pg, @llvm.aarch64.sve.cnt.nxv2f64( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cnt.nxv2f64( poison, %pg, %x) ret %0 } @@ -739,7 +739,7 @@ define @test_svcnt_f64_x_2( %pg, double %z0, ; CHECK-2p2-NEXT: cnt z0.d, p0/z, z1.d ; CHECK-2p2-NEXT: ret entry: - %0 = tail call @llvm.aarch64.sve.cnt.nxv2f64( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cnt.nxv2f64( poison, %pg, %x) ret %0 } @@ -770,7 +770,7 @@ define @test_svcnot_s8_x_1( %pg, @llvm.aarch64.sve.cnot.nxv16i8( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cnot.nxv16i8( poison, %pg, %x) ret %0 } @@ -786,7 +786,7 @@ define @test_svcnot_s8_x_2( %pg, double %z0 ; CHECK-2p2-NEXT: cnot z0.b, p0/z, z1.b ; CHECK-2p2-NEXT: ret entry: - %0 = tail call @llvm.aarch64.sve.cnot.nxv16i8( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cnot.nxv16i8( poison, %pg, %x) ret %0 } @@ -817,7 +817,7 @@ define @test_svcnot_s16_x_1( %pg, @llvm.aarch64.sve.cnot.nxv8i16( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cnot.nxv8i16( poison, %pg, %x) ret %0 } @@ -833,7 +833,7 @@ define @test_svcnot_s16_x_2( %pg, double %z0 ; CHECK-2p2-NEXT: cnot z0.h, p0/z, z1.h ; CHECK-2p2-NEXT: ret entry: - %0 = tail call @llvm.aarch64.sve.cnot.nxv8i16( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cnot.nxv8i16( poison, %pg, %x) ret %0 } @@ -864,7 +864,7 @@ define @test_svcnot_s32_x_1( %pg, @llvm.aarch64.sve.cnot.nxv4i32( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cnot.nxv4i32( poison, %pg, %x) ret %0 } @@ -880,7 +880,7 @@ define @test_svcnot_s32_x_2( %pg, double %z0 ; CHECK-2p2-NEXT: cnot z0.s, p0/z, z1.s ; CHECK-2p2-NEXT: ret entry: - %0 = tail call @llvm.aarch64.sve.cnot.nxv4i32( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cnot.nxv4i32( poison, %pg, %x) ret %0 } @@ -911,7 +911,7 @@ define @test_svcnot_s64_x_1( %pg, @llvm.aarch64.sve.cnot.nxv2i64( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cnot.nxv2i64( poison, %pg, %x) ret %0 } @@ -927,7 +927,7 @@ define @test_svcnot_s64_x_2( %pg, double %z0 ; CHECK-2p2-NEXT: cnot z0.d, p0/z, z1.d ; CHECK-2p2-NEXT: ret entry: - %0 = tail call @llvm.aarch64.sve.cnot.nxv2i64( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.cnot.nxv2i64( poison, %pg, %x) ret %0 } @@ -958,7 +958,7 @@ define @test_svnot_s8_x_1( %pg, @llvm.aarch64.sve.not.nxv16i8( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.not.nxv16i8( poison, %pg, %x) ret %0 } @@ -974,7 +974,7 @@ define @test_svnot_s8_x_2( %pg, double %z0, ; CHECK-2p2-NEXT: not z0.b, p0/z, z1.b ; CHECK-2p2-NEXT: ret entry: - %0 = tail call @llvm.aarch64.sve.not.nxv16i8( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.not.nxv16i8( poison, %pg, %x) ret %0 } @@ -1005,7 +1005,7 @@ define @test_svnot_s16_x_1( %pg, @llvm.aarch64.sve.not.nxv8i16( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.not.nxv8i16( poison, %pg, %x) ret %0 } @@ -1021,7 +1021,7 @@ define @test_svnot_s16_x_2( %pg, double %z0, ; CHECK-2p2-NEXT: not z0.h, p0/z, z1.h ; CHECK-2p2-NEXT: ret entry: - %0 = tail call @llvm.aarch64.sve.not.nxv8i16( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.not.nxv8i16( poison, %pg, %x) ret %0 } @@ -1052,7 +1052,7 @@ define @test_svnot_s32_x_1( %pg, @llvm.aarch64.sve.not.nxv4i32( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.not.nxv4i32( poison, %pg, %x) ret %0 } @@ -1068,7 +1068,7 @@ define @test_svnot_s32_x_2( %pg, double %z0, ; CHECK-2p2-NEXT: not z0.s, p0/z, z1.s ; CHECK-2p2-NEXT: ret entry: - %0 = tail call @llvm.aarch64.sve.not.nxv4i32( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.not.nxv4i32( poison, %pg, %x) ret %0 } @@ -1099,7 +1099,7 @@ define @test_svnot_s64_x_1( %pg, @llvm.aarch64.sve.not.nxv2i64( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.not.nxv2i64( poison, %pg, %x) ret %0 } @@ -1115,7 +1115,7 @@ define @test_svnot_s64_x_2( %pg, double %z0, ; CHECK-2p2-NEXT: not z0.d, p0/z, z1.d ; CHECK-2p2-NEXT: ret entry: - %0 = tail call @llvm.aarch64.sve.not.nxv2i64( undef, %pg, %x) + %0 = tail call @llvm.aarch64.sve.not.nxv2i64( poison, %pg, %x) ret %0 } @@ -1135,18 +1135,914 @@ entry: ret %0 } -define @test_svnot_u64_z( %pg, double %z0, %x) { -; CHECK-LABEL: test_svnot_u64_z: +define @test_svcls_nxv16i8_ptrue_u(double %z0, %x) { +; CHECK-LABEL: test_svcls_nxv16i8_ptrue_u: ; CHECK: // %bb.0: // %entry -; CHECK-NEXT: mov z0.d, #0 // =0x0 +; CHECK-NEXT: ptrue p0.b +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cls z0.b, p0/m, z1.b +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcls_nxv16i8_ptrue_u: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.b +; CHECK-2p2-NEXT: cls z0.b, p0/z, z1.b +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv16i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cls.nxv16i8( poison, %pg, %x) + ret %0 +} + +define @test_svcls_nxv16i8_ptrue(double %z0, %x, %y) { +; CHECK-LABEL: test_svcls_nxv16i8_ptrue: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.b +; CHECK-NEXT: movprfx z0, z2 +; CHECK-NEXT: cls z0.b, p0/m, z2.b +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcls_nxv16i8_ptrue: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.b +; CHECK-2p2-NEXT: cls z0.b, p0/z, z2.b +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv16i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cls.nxv16i8( %x, %pg, %y) + ret %0 +} + +define @test_svcls_nxv8i16_ptrue_u(double %z0, %x) { +; CHECK-LABEL: test_svcls_nxv8i16_ptrue_u: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.h +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cls z0.h, p0/m, z1.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcls_nxv8i16_ptrue_u: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.h +; CHECK-2p2-NEXT: cls z0.h, p0/z, z1.h +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv8i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cls.nxv8i16( poison, %pg, %x) + ret %0 +} + +define @test_svcls_nxv8i16_ptrue(double %z0, %x, %y) { +; CHECK-LABEL: test_svcls_nxv8i16_ptrue: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.h +; CHECK-NEXT: movprfx z0, z2 +; CHECK-NEXT: cls z0.h, p0/m, z2.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcls_nxv8i16_ptrue: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.h +; CHECK-2p2-NEXT: cls z0.h, p0/z, z2.h +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv8i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cls.nxv8i16( %x, %pg, %y) + ret %0 +} + +define @test_svcls_nxv4i32_ptrue_u(double %z0, %x) { +; CHECK-LABEL: test_svcls_nxv4i32_ptrue_u: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.s +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cls z0.s, p0/m, z1.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcls_nxv4i32_ptrue_u: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.s +; CHECK-2p2-NEXT: cls z0.s, p0/z, z1.s +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv4i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cls.nxv4i32( poison, %pg, %x) + ret %0 +} + +define @test_svcls_nxv4i32_ptrue(double %z0, %x, %y) { +; CHECK-LABEL: test_svcls_nxv4i32_ptrue: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.s +; CHECK-NEXT: movprfx z0, z2 +; CHECK-NEXT: cls z0.s, p0/m, z2.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcls_nxv4i32_ptrue: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.s +; CHECK-2p2-NEXT: cls z0.s, p0/z, z2.s +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv4i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cls.nxv4i32( %x, %pg, %y) + ret %0 +} + +define @test_svcls_nxv2i64_ptrue_u(double %z0, %x) { +; CHECK-LABEL: test_svcls_nxv2i64_ptrue_u: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.d +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cls z0.d, p0/m, z1.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcls_nxv2i64_ptrue_u: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.d +; CHECK-2p2-NEXT: cls z0.d, p0/z, z1.d +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv2i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cls.nxv2i64( poison, %pg, %x) + ret %0 +} + +define @test_svcls_nxv2i64_ptrue(double %z0, %x, %y) { +; CHECK-LABEL: test_svcls_nxv2i64_ptrue: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.d +; CHECK-NEXT: movprfx z0, z2 +; CHECK-NEXT: cls z0.d, p0/m, z2.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcls_nxv2i64_ptrue: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.d +; CHECK-2p2-NEXT: cls z0.d, p0/z, z2.d +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv2i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cls.nxv2i64( %x, %pg, %y) + ret %0 +} + +define @test_svclz_nxv16i8_ptrue_u(double %z0, %x) { +; CHECK-LABEL: test_svclz_nxv16i8_ptrue_u: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.b +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: clz z0.b, p0/m, z1.b +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svclz_nxv16i8_ptrue_u: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.b +; CHECK-2p2-NEXT: clz z0.b, p0/z, z1.b +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv16i1(i32 31) + %0 = tail call @llvm.aarch64.sve.clz.nxv16i8( poison, %pg, %x) + ret %0 +} + +define @test_svclz_nxv16i8_ptrue(double %z0, %x, %y) { +; CHECK-LABEL: test_svclz_nxv16i8_ptrue: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.b +; CHECK-NEXT: movprfx z0, z2 +; CHECK-NEXT: clz z0.b, p0/m, z2.b +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svclz_nxv16i8_ptrue: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.b +; CHECK-2p2-NEXT: clz z0.b, p0/z, z2.b +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv16i1(i32 31) + %0 = tail call @llvm.aarch64.sve.clz.nxv16i8( %x, %pg, %y) + ret %0 +} + +define @test_svclz_nxv8i16_ptrue_u(double %z0, %x) { +; CHECK-LABEL: test_svclz_nxv8i16_ptrue_u: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.h +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: clz z0.h, p0/m, z1.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svclz_nxv8i16_ptrue_u: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.h +; CHECK-2p2-NEXT: clz z0.h, p0/z, z1.h +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv8i1(i32 31) + %0 = tail call @llvm.aarch64.sve.clz.nxv8i16( poison, %pg, %x) + ret %0 +} + +define @test_svclz_nxv8i16_ptrue(double %z0, %x, %y) { +; CHECK-LABEL: test_svclz_nxv8i16_ptrue: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.h +; CHECK-NEXT: movprfx z0, z2 +; CHECK-NEXT: clz z0.h, p0/m, z2.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svclz_nxv8i16_ptrue: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.h +; CHECK-2p2-NEXT: clz z0.h, p0/z, z2.h +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv8i1(i32 31) + %0 = tail call @llvm.aarch64.sve.clz.nxv8i16( %x, %pg, %y) + ret %0 +} + +define @test_svclz_nxv4i32_ptrue_u(double %z0, %x) { +; CHECK-LABEL: test_svclz_nxv4i32_ptrue_u: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.s +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: clz z0.s, p0/m, z1.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svclz_nxv4i32_ptrue_u: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.s +; CHECK-2p2-NEXT: clz z0.s, p0/z, z1.s +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv4i1(i32 31) + %0 = tail call @llvm.aarch64.sve.clz.nxv4i32( poison, %pg, %x) + ret %0 +} + +define @test_svclz_nxv4i32_ptrue(double %z0, %x, %y) { +; CHECK-LABEL: test_svclz_nxv4i32_ptrue: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.s +; CHECK-NEXT: movprfx z0, z2 +; CHECK-NEXT: clz z0.s, p0/m, z2.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svclz_nxv4i32_ptrue: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.s +; CHECK-2p2-NEXT: clz z0.s, p0/z, z2.s +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv4i1(i32 31) + %0 = tail call @llvm.aarch64.sve.clz.nxv4i32( %x, %pg, %y) + ret %0 +} + +define @test_svclz_nxv2i64_ptrue_u(double %z0, %x) { +; CHECK-LABEL: test_svclz_nxv2i64_ptrue_u: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.d +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: clz z0.d, p0/m, z1.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svclz_nxv2i64_ptrue_u: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.d +; CHECK-2p2-NEXT: clz z0.d, p0/z, z1.d +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv2i1(i32 31) + %0 = tail call @llvm.aarch64.sve.clz.nxv2i64( poison, %pg, %x) + ret %0 +} + +define @test_svclz_nxv2i64_ptrue(double %z0, %x, %y) { +; CHECK-LABEL: test_svclz_nxv2i64_ptrue: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.d +; CHECK-NEXT: movprfx z0, z2 +; CHECK-NEXT: clz z0.d, p0/m, z2.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svclz_nxv2i64_ptrue: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.d +; CHECK-2p2-NEXT: clz z0.d, p0/z, z2.d +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv2i1(i32 31) + %0 = tail call @llvm.aarch64.sve.clz.nxv2i64( %x, %pg, %y) + ret %0 +} + +define @test_svcnt_nxv16i8_ptrue_u(double %z0, %x) { +; CHECK-LABEL: test_svcnt_nxv16i8_ptrue_u: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.b +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cnt z0.b, p0/m, z1.b +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_nxv16i8_ptrue_u: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.b +; CHECK-2p2-NEXT: cnt z0.b, p0/z, z1.b +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv16i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cnt.nxv16i8( poison, %pg, %x) + ret %0 +} + +define @test_svcnt_nxv16i8_ptrue(double %z0, %x, %y) { +; CHECK-LABEL: test_svcnt_nxv16i8_ptrue: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.b +; CHECK-NEXT: movprfx z0, z2 +; CHECK-NEXT: cnt z0.b, p0/m, z2.b +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_nxv16i8_ptrue: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.b +; CHECK-2p2-NEXT: cnt z0.b, p0/z, z2.b +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv16i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cnt.nxv16i8( %x, %pg, %y) + ret %0 +} + +define @test_svcnt_nxv8i16_ptrue_u(double %z0, %x) { +; CHECK-LABEL: test_svcnt_nxv8i16_ptrue_u: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.h +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cnt z0.h, p0/m, z1.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_nxv8i16_ptrue_u: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.h +; CHECK-2p2-NEXT: cnt z0.h, p0/z, z1.h +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv8i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cnt.nxv8i16( poison, %pg, %x) + ret %0 +} + +define @test_svcnt_nxv8i16_ptrue(double %z0, %x, %y) { +; CHECK-LABEL: test_svcnt_nxv8i16_ptrue: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.h +; CHECK-NEXT: movprfx z0, z2 +; CHECK-NEXT: cnt z0.h, p0/m, z2.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_nxv8i16_ptrue: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.h +; CHECK-2p2-NEXT: cnt z0.h, p0/z, z2.h +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv8i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cnt.nxv8i16( %x, %pg, %y) + ret %0 +} + +define @test_svcnt_nxv4i32_ptrue_u(double %z0, %x) { +; CHECK-LABEL: test_svcnt_nxv4i32_ptrue_u: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.s +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cnt z0.s, p0/m, z1.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_nxv4i32_ptrue_u: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.s +; CHECK-2p2-NEXT: cnt z0.s, p0/z, z1.s +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv4i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cnt.nxv4i32( poison, %pg, %x) + ret %0 +} + +define @test_svcnt_nxv4i32_ptrue(double %z0, %x, %y) { +; CHECK-LABEL: test_svcnt_nxv4i32_ptrue: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.s +; CHECK-NEXT: movprfx z0, z2 +; CHECK-NEXT: cnt z0.s, p0/m, z2.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_nxv4i32_ptrue: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.s +; CHECK-2p2-NEXT: cnt z0.s, p0/z, z2.s +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv4i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cnt.nxv4i32( %x, %pg, %y) + ret %0 +} + +define @test_svcnt_nxv2i64_ptrue_u(double %z0, %x) { +; CHECK-LABEL: test_svcnt_nxv2i64_ptrue_u: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.d +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cnt z0.d, p0/m, z1.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_nxv2i64_ptrue_u: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.d +; CHECK-2p2-NEXT: cnt z0.d, p0/z, z1.d +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv2i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cnt.nxv2i64( poison, %pg, %x) + ret %0 +} + +define @test_svcnt_nxv2i64_ptrue(double %z0, %x, %y) { +; CHECK-LABEL: test_svcnt_nxv2i64_ptrue: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.d +; CHECK-NEXT: movprfx z0, z2 +; CHECK-NEXT: cnt z0.d, p0/m, z2.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_nxv2i64_ptrue: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.d +; CHECK-2p2-NEXT: cnt z0.d, p0/z, z2.d +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv2i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cnt.nxv2i64( %x, %pg, %y) + ret %0 +} + +define @test_svcnt_nxv8f16_ptrue_u(double %z0, %x) { +; CHECK-LABEL: test_svcnt_nxv8f16_ptrue_u: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.h +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cnt z0.h, p0/m, z1.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_nxv8f16_ptrue_u: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.h +; CHECK-2p2-NEXT: cnt z0.h, p0/z, z1.h +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv8i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cnt.nxv8f16( poison, %pg, %x) + ret %0 +} + +define @test_svcnt_nxv8f16_ptrue(double %z0, %x, %y) { +; CHECK-LABEL: test_svcnt_nxv8f16_ptrue: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.h +; CHECK-NEXT: movprfx z0, z2 +; CHECK-NEXT: cnt z0.h, p0/m, z2.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_nxv8f16_ptrue: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.h +; CHECK-2p2-NEXT: cnt z0.h, p0/z, z2.h +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv8i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cnt.nxv8f16( %x, %pg, %y) + ret %0 +} + +define @test_svcnt_nxv8bf16_ptrue_u(double %z0, %x) { +; CHECK-LABEL: test_svcnt_nxv8bf16_ptrue_u: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.h +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cnt z0.h, p0/m, z1.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_nxv8bf16_ptrue_u: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.h +; CHECK-2p2-NEXT: cnt z0.h, p0/z, z1.h +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv8i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cnt.nxv8bf16( poison, %pg, %x) + ret %0 +} + +define @test_svcnt_nxv8bf16_ptrue(double %z0, %x, %y) { +; CHECK-LABEL: test_svcnt_nxv8bf16_ptrue: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.h +; CHECK-NEXT: movprfx z0, z2 +; CHECK-NEXT: cnt z0.h, p0/m, z2.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_nxv8bf16_ptrue: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.h +; CHECK-2p2-NEXT: cnt z0.h, p0/z, z2.h +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv8i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cnt.nxv8bf16( %x, %pg, %y) + ret %0 +} + +define @test_svcnt_nxv4f32_ptrue_u(double %z0, %x) { +; CHECK-LABEL: test_svcnt_nxv4f32_ptrue_u: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.s +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cnt z0.s, p0/m, z1.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_nxv4f32_ptrue_u: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.s +; CHECK-2p2-NEXT: cnt z0.s, p0/z, z1.s +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv4i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cnt.nxv4f32( poison, %pg, %x) + ret %0 +} + +define @test_svcnt_nxv4f32_ptrue(double %z0, %x, %y) { +; CHECK-LABEL: test_svcnt_nxv4f32_ptrue: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.s +; CHECK-NEXT: movprfx z0, z2 +; CHECK-NEXT: cnt z0.s, p0/m, z2.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_nxv4f32_ptrue: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.s +; CHECK-2p2-NEXT: cnt z0.s, p0/z, z2.s +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv4i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cnt.nxv4f32( %x, %pg, %y) + ret %0 +} + +define @test_svcnt_nxv2f64_ptrue_u(double %z0, %x) { +; CHECK-LABEL: test_svcnt_nxv2f64_ptrue_u: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.d +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cnt z0.d, p0/m, z1.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_nxv2f64_ptrue_u: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.d +; CHECK-2p2-NEXT: cnt z0.d, p0/z, z1.d +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv2i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cnt.nxv2f64( poison, %pg, %x) + ret %0 +} + +define @test_svcnt_nxv2f64_ptrue(double %z0, %x, %y) { +; CHECK-LABEL: test_svcnt_nxv2f64_ptrue: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.d +; CHECK-NEXT: movprfx z0, z2 +; CHECK-NEXT: cnt z0.d, p0/m, z2.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnt_nxv2f64_ptrue: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.d +; CHECK-2p2-NEXT: cnt z0.d, p0/z, z2.d +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv2i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cnt.nxv2f64( %x, %pg, %y) + ret %0 +} + +define @test_svcnot_nxv16i8_ptrue_u(double %z0, %x) { +; CHECK-LABEL: test_svcnot_nxv16i8_ptrue_u: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.b +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cnot z0.b, p0/m, z1.b +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnot_nxv16i8_ptrue_u: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.b +; CHECK-2p2-NEXT: cnot z0.b, p0/z, z1.b +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv16i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cnot.nxv16i8( poison, %pg, %x) + ret %0 +} + +define @test_svcnot_nxv16i8_ptrue(double %z0, %x, %y) { +; CHECK-LABEL: test_svcnot_nxv16i8_ptrue: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.b +; CHECK-NEXT: movprfx z0, z2 +; CHECK-NEXT: cnot z0.b, p0/m, z2.b +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnot_nxv16i8_ptrue: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.b +; CHECK-2p2-NEXT: cnot z0.b, p0/z, z2.b +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv16i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cnot.nxv16i8( %x, %pg, %y) + ret %0 +} + +define @test_svcnot_nxv8i16_ptrue_u(double %z0, %x) { +; CHECK-LABEL: test_svcnot_nxv8i16_ptrue_u: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.h +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cnot z0.h, p0/m, z1.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnot_nxv8i16_ptrue_u: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.h +; CHECK-2p2-NEXT: cnot z0.h, p0/z, z1.h +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv8i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cnot.nxv8i16( poison, %pg, %x) + ret %0 +} + +define @test_svcnot_nxv8i16_ptrue(double %z0, %x, %y) { +; CHECK-LABEL: test_svcnot_nxv8i16_ptrue: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.h +; CHECK-NEXT: movprfx z0, z2 +; CHECK-NEXT: cnot z0.h, p0/m, z2.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnot_nxv8i16_ptrue: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.h +; CHECK-2p2-NEXT: cnot z0.h, p0/z, z2.h +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv8i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cnot.nxv8i16( %x, %pg, %y) + ret %0 +} + +define @test_svcnot_nxv4i32_ptrue_u(double %z0, %x) { +; CHECK-LABEL: test_svcnot_nxv4i32_ptrue_u: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.s +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cnot z0.s, p0/m, z1.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnot_nxv4i32_ptrue_u: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.s +; CHECK-2p2-NEXT: cnot z0.s, p0/z, z1.s +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv4i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cnot.nxv4i32( poison, %pg, %x) + ret %0 +} + +define @test_svcnot_nxv4i32_ptrue(double %z0, %x, %y) { +; CHECK-LABEL: test_svcnot_nxv4i32_ptrue: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.s +; CHECK-NEXT: movprfx z0, z2 +; CHECK-NEXT: cnot z0.s, p0/m, z2.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnot_nxv4i32_ptrue: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.s +; CHECK-2p2-NEXT: cnot z0.s, p0/z, z2.s +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv4i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cnot.nxv4i32( %x, %pg, %y) + ret %0 +} + +define @test_svcnot_nxv2i64_ptrue_u(double %z0, %x) { +; CHECK-LABEL: test_svcnot_nxv2i64_ptrue_u: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.d +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: cnot z0.d, p0/m, z1.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnot_nxv2i64_ptrue_u: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.d +; CHECK-2p2-NEXT: cnot z0.d, p0/z, z1.d +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv2i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cnot.nxv2i64( poison, %pg, %x) + ret %0 +} + +define @test_svcnot_nxv2i64_ptrue(double %z0, %x, %y) { +; CHECK-LABEL: test_svcnot_nxv2i64_ptrue: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.d +; CHECK-NEXT: movprfx z0, z2 +; CHECK-NEXT: cnot z0.d, p0/m, z2.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svcnot_nxv2i64_ptrue: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.d +; CHECK-2p2-NEXT: cnot z0.d, p0/z, z2.d +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv2i1(i32 31) + %0 = tail call @llvm.aarch64.sve.cnot.nxv2i64( %x, %pg, %y) + ret %0 +} + +define @test_svnot_nxv16i8_ptrue_u(double %z0, %x) { +; CHECK-LABEL: test_svnot_nxv16i8_ptrue_u: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.b +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: not z0.b, p0/m, z1.b +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svnot_nxv16i8_ptrue_u: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.b +; CHECK-2p2-NEXT: not z0.b, p0/z, z1.b +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv16i1(i32 31) + %0 = tail call @llvm.aarch64.sve.not.nxv16i8( poison, %pg, %x) + ret %0 +} + +define @test_svnot_nxv16i8_ptrue(double %z0, %x, %y) { +; CHECK-LABEL: test_svnot_nxv16i8_ptrue: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.b +; CHECK-NEXT: movprfx z0, z2 +; CHECK-NEXT: not z0.b, p0/m, z2.b +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svnot_nxv16i8_ptrue: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.b +; CHECK-2p2-NEXT: not z0.b, p0/z, z2.b +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv16i1(i32 31) + %0 = tail call @llvm.aarch64.sve.not.nxv16i8( %x, %pg, %y) + ret %0 +} + +define @test_svnot_nxv8i16_ptrue_u(double %z0, %x) { +; CHECK-LABEL: test_svnot_nxv8i16_ptrue_u: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.h +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: not z0.h, p0/m, z1.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svnot_nxv8i16_ptrue_u: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.h +; CHECK-2p2-NEXT: not z0.h, p0/z, z1.h +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv8i1(i32 31) + %0 = tail call @llvm.aarch64.sve.not.nxv8i16( poison, %pg, %x) + ret %0 +} + +define @test_svnot_nxv8i16_ptrue(double %z0, %x, %y) { +; CHECK-LABEL: test_svnot_nxv8i16_ptrue: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.h +; CHECK-NEXT: movprfx z0, z2 +; CHECK-NEXT: not z0.h, p0/m, z2.h +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svnot_nxv8i16_ptrue: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.h +; CHECK-2p2-NEXT: not z0.h, p0/z, z2.h +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv8i1(i32 31) + %0 = tail call @llvm.aarch64.sve.not.nxv8i16( %x, %pg, %y) + ret %0 +} + +define @test_svnot_nxv4i32_ptrue_u(double %z0, %x) { +; CHECK-LABEL: test_svnot_nxv4i32_ptrue_u: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.s +; CHECK-NEXT: movprfx z0, z1 +; CHECK-NEXT: not z0.s, p0/m, z1.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svnot_nxv4i32_ptrue_u: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.s +; CHECK-2p2-NEXT: not z0.s, p0/z, z1.s +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv4i1(i32 31) + %0 = tail call @llvm.aarch64.sve.not.nxv4i32( poison, %pg, %x) + ret %0 +} + +define @test_svnot_nxv4i32_ptrue(double %z0, %x, %y) { +; CHECK-LABEL: test_svnot_nxv4i32_ptrue: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.s +; CHECK-NEXT: movprfx z0, z2 +; CHECK-NEXT: not z0.s, p0/m, z2.s +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svnot_nxv4i32_ptrue: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.s +; CHECK-2p2-NEXT: not z0.s, p0/z, z2.s +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv4i1(i32 31) + %0 = tail call @llvm.aarch64.sve.not.nxv4i32( %x, %pg, %y) + ret %0 +} + +define @test_svnot_nxv2i64_ptrue_u(double %z0, %x) { +; CHECK-LABEL: test_svnot_nxv2i64_ptrue_u: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.d +; CHECK-NEXT: movprfx z0, z1 ; CHECK-NEXT: not z0.d, p0/m, z1.d ; CHECK-NEXT: ret ; -; CHECK-2p2-LABEL: test_svnot_u64_z: +; CHECK-2p2-LABEL: test_svnot_nxv2i64_ptrue_u: ; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.d ; CHECK-2p2-NEXT: not z0.d, p0/z, z1.d ; CHECK-2p2-NEXT: ret entry: - %0 = tail call @llvm.aarch64.sve.not.nxv2i64( zeroinitializer, %pg, %x) + %pg = call @llvm.aarch64.sve.ptrue.nxv2i1(i32 31) + %0 = tail call @llvm.aarch64.sve.not.nxv2i64( poison, %pg, %x) + ret %0 +} + +define @test_svnot_nxv2i64_ptrue(double %z0, %x, %y) { +; CHECK-LABEL: test_svnot_nxv2i64_ptrue: +; CHECK: // %bb.0: // %entry +; CHECK-NEXT: ptrue p0.d +; CHECK-NEXT: movprfx z0, z2 +; CHECK-NEXT: not z0.d, p0/m, z2.d +; CHECK-NEXT: ret +; +; CHECK-2p2-LABEL: test_svnot_nxv2i64_ptrue: +; CHECK-2p2: // %bb.0: // %entry +; CHECK-2p2-NEXT: ptrue p0.d +; CHECK-2p2-NEXT: not z0.d, p0/z, z2.d +; CHECK-2p2-NEXT: ret +entry: + %pg = call @llvm.aarch64.sve.ptrue.nxv2i1(i32 31) + %0 = tail call @llvm.aarch64.sve.not.nxv2i64( %x, %pg, %y) ret %0 }