Author: ShihPo Hung Date: 2021-01-21T18:38:49-08:00 New Revision: 96677503315e689fd3c8f5ef164d8fb9725d4bb3
URL: https://github.com/llvm/llvm-project/commit/96677503315e689fd3c8f5ef164d8fb9725d4bb3 DIFF: https://github.com/llvm/llvm-project/commit/96677503315e689fd3c8f5ef164d8fb9725d4bb3.diff LOG: [RISCV] Add intrinsics for RVV1.0 VFRSQRTE7 & VFRECE7 Reviewed By: craig.topper, frasercrmck Differential Revision: https://reviews.llvm.org/D95113 Added: llvm/test/CodeGen/RISCV/vfrece7-rv32.ll llvm/test/CodeGen/RISCV/vfrece7-rv64.ll llvm/test/CodeGen/RISCV/vfrsqrte7-rv32.ll llvm/test/CodeGen/RISCV/vfrsqrte7-rv64.ll Modified: llvm/include/llvm/IR/IntrinsicsRISCV.td llvm/lib/Target/RISCV/RISCVInstrInfoVPseudos.td Removed: ################################################################################ diff --git a/llvm/include/llvm/IR/IntrinsicsRISCV.td b/llvm/include/llvm/IR/IntrinsicsRISCV.td index 4b174b3edc31..407b27744477 100644 --- a/llvm/include/llvm/IR/IntrinsicsRISCV.td +++ b/llvm/include/llvm/IR/IntrinsicsRISCV.td @@ -865,6 +865,8 @@ let TargetPrefix = "riscv" in { defm vfwnmsac : RISCVTernaryWide; defm vfsqrt : RISCVUnaryAA; + defm vfrsqrte7 : RISCVUnaryAA; + defm vfrece7 : RISCVUnaryAA; defm vfmin : RISCVBinaryAAX; defm vfmax : RISCVBinaryAAX; diff --git a/llvm/lib/Target/RISCV/RISCVInstrInfoVPseudos.td b/llvm/lib/Target/RISCV/RISCVInstrInfoVPseudos.td index fb09110acc4c..d1a823be25b6 100644 --- a/llvm/lib/Target/RISCV/RISCVInstrInfoVPseudos.td +++ b/llvm/lib/Target/RISCV/RISCVInstrInfoVPseudos.td @@ -3212,6 +3212,16 @@ defm PseudoVFWNMSAC : VPseudoTernaryW_VV_VX</*IsFloat*/true>; //===----------------------------------------------------------------------===// defm PseudoVFSQRT : VPseudoUnaryV_V; +//===----------------------------------------------------------------------===// +// 14.9. Vector Floating-Point Reciprocal Square-Root Estimate Instruction +//===----------------------------------------------------------------------===// +defm PseudoVFRSQRTE7 : VPseudoUnaryV_V; + +//===----------------------------------------------------------------------===// +// 14.10. Vector Floating-Point Reciprocal Estimate Instruction +//===----------------------------------------------------------------------===// +defm PseudoVFRECE7 : VPseudoUnaryV_V; + //===----------------------------------------------------------------------===// // 14.11. Vector Floating-Point Min/Max Instructions //===----------------------------------------------------------------------===// @@ -3871,6 +3881,16 @@ defm "" : VPatTernaryW_VV_VX<"int_riscv_vfwnmsac", "PseudoVFWNMSAC", AllWidenabl //===----------------------------------------------------------------------===// defm "" : VPatUnaryV_V<"int_riscv_vfsqrt", "PseudoVFSQRT", AllFloatVectors>; +//===----------------------------------------------------------------------===// +// 14.9. Vector Floating-Point Reciprocal Square-Root Estimate Instruction +//===----------------------------------------------------------------------===// +defm "" : VPatUnaryV_V<"int_riscv_vfrsqrte7", "PseudoVFRSQRTE7", AllFloatVectors>; + +//===----------------------------------------------------------------------===// +// 14.10. Vector Floating-Point Reciprocal Estimate Instruction +//===----------------------------------------------------------------------===// +defm "" : VPatUnaryV_V<"int_riscv_vfrece7", "PseudoVFRECE7", AllFloatVectors>; + //===----------------------------------------------------------------------===// // 14.11. Vector Floating-Point Min/Max Instructions //===----------------------------------------------------------------------===// diff --git a/llvm/test/CodeGen/RISCV/vfrece7-rv32.ll b/llvm/test/CodeGen/RISCV/vfrece7-rv32.ll new file mode 100644 index 000000000000..7a810f10d47d --- /dev/null +++ b/llvm/test/CodeGen/RISCV/vfrece7-rv32.ll @@ -0,0 +1,602 @@ +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py +; RUN: llc -mtriple=riscv32 -mattr=+experimental-v,+d,+experimental-zfh -verify-machineinstrs \ +; RUN: --riscv-no-aliases < %s | FileCheck %s +declare <vscale x 1 x half> @llvm.riscv.vfrece7.nxv1f16( + <vscale x 1 x half>, + i32); + +define <vscale x 1 x half> @intrinsic_vfrece7_v_nxv1f16_nxv1f16(<vscale x 1 x half> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv1f16_nxv1f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,mf4,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 1 x half> @llvm.riscv.vfrece7.nxv1f16( + <vscale x 1 x half> %0, + i32 %1) + + ret <vscale x 1 x half> %a +} + +declare <vscale x 1 x half> @llvm.riscv.vfrece7.mask.nxv1f16( + <vscale x 1 x half>, + <vscale x 1 x half>, + <vscale x 1 x i1>, + i32); + +define <vscale x 1 x half> @intrinsic_vfrece7_mask_v_nxv1f16_nxv1f16(<vscale x 1 x i1> %0, <vscale x 1 x half> %1, <vscale x 1 x half> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv1f16_nxv1f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,mf4,tu,mu +; CHECK-NEXT: vfrece7.v v8, v9, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 1 x half> @llvm.riscv.vfrece7.mask.nxv1f16( + <vscale x 1 x half> %1, + <vscale x 1 x half> %2, + <vscale x 1 x i1> %0, + i32 %3) + + ret <vscale x 1 x half> %a +} + +declare <vscale x 2 x half> @llvm.riscv.vfrece7.nxv2f16( + <vscale x 2 x half>, + i32); + +define <vscale x 2 x half> @intrinsic_vfrece7_v_nxv2f16_nxv2f16(<vscale x 2 x half> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv2f16_nxv2f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,mf2,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 2 x half> @llvm.riscv.vfrece7.nxv2f16( + <vscale x 2 x half> %0, + i32 %1) + + ret <vscale x 2 x half> %a +} + +declare <vscale x 2 x half> @llvm.riscv.vfrece7.mask.nxv2f16( + <vscale x 2 x half>, + <vscale x 2 x half>, + <vscale x 2 x i1>, + i32); + +define <vscale x 2 x half> @intrinsic_vfrece7_mask_v_nxv2f16_nxv2f16(<vscale x 2 x i1> %0, <vscale x 2 x half> %1, <vscale x 2 x half> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv2f16_nxv2f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,mf2,tu,mu +; CHECK-NEXT: vfrece7.v v8, v9, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 2 x half> @llvm.riscv.vfrece7.mask.nxv2f16( + <vscale x 2 x half> %1, + <vscale x 2 x half> %2, + <vscale x 2 x i1> %0, + i32 %3) + + ret <vscale x 2 x half> %a +} + +declare <vscale x 4 x half> @llvm.riscv.vfrece7.nxv4f16( + <vscale x 4 x half>, + i32); + +define <vscale x 4 x half> @intrinsic_vfrece7_v_nxv4f16_nxv4f16(<vscale x 4 x half> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv4f16_nxv4f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m1,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 4 x half> @llvm.riscv.vfrece7.nxv4f16( + <vscale x 4 x half> %0, + i32 %1) + + ret <vscale x 4 x half> %a +} + +declare <vscale x 4 x half> @llvm.riscv.vfrece7.mask.nxv4f16( + <vscale x 4 x half>, + <vscale x 4 x half>, + <vscale x 4 x i1>, + i32); + +define <vscale x 4 x half> @intrinsic_vfrece7_mask_v_nxv4f16_nxv4f16(<vscale x 4 x i1> %0, <vscale x 4 x half> %1, <vscale x 4 x half> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv4f16_nxv4f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m1,tu,mu +; CHECK-NEXT: vfrece7.v v8, v9, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 4 x half> @llvm.riscv.vfrece7.mask.nxv4f16( + <vscale x 4 x half> %1, + <vscale x 4 x half> %2, + <vscale x 4 x i1> %0, + i32 %3) + + ret <vscale x 4 x half> %a +} + +declare <vscale x 8 x half> @llvm.riscv.vfrece7.nxv8f16( + <vscale x 8 x half>, + i32); + +define <vscale x 8 x half> @intrinsic_vfrece7_v_nxv8f16_nxv8f16(<vscale x 8 x half> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv8f16_nxv8f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m2,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 8 x half> @llvm.riscv.vfrece7.nxv8f16( + <vscale x 8 x half> %0, + i32 %1) + + ret <vscale x 8 x half> %a +} + +declare <vscale x 8 x half> @llvm.riscv.vfrece7.mask.nxv8f16( + <vscale x 8 x half>, + <vscale x 8 x half>, + <vscale x 8 x i1>, + i32); + +define <vscale x 8 x half> @intrinsic_vfrece7_mask_v_nxv8f16_nxv8f16(<vscale x 8 x i1> %0, <vscale x 8 x half> %1, <vscale x 8 x half> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv8f16_nxv8f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m2,tu,mu +; CHECK-NEXT: vfrece7.v v8, v10, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 8 x half> @llvm.riscv.vfrece7.mask.nxv8f16( + <vscale x 8 x half> %1, + <vscale x 8 x half> %2, + <vscale x 8 x i1> %0, + i32 %3) + + ret <vscale x 8 x half> %a +} + +declare <vscale x 16 x half> @llvm.riscv.vfrece7.nxv16f16( + <vscale x 16 x half>, + i32); + +define <vscale x 16 x half> @intrinsic_vfrece7_v_nxv16f16_nxv16f16(<vscale x 16 x half> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv16f16_nxv16f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m4,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 16 x half> @llvm.riscv.vfrece7.nxv16f16( + <vscale x 16 x half> %0, + i32 %1) + + ret <vscale x 16 x half> %a +} + +declare <vscale x 16 x half> @llvm.riscv.vfrece7.mask.nxv16f16( + <vscale x 16 x half>, + <vscale x 16 x half>, + <vscale x 16 x i1>, + i32); + +define <vscale x 16 x half> @intrinsic_vfrece7_mask_v_nxv16f16_nxv16f16(<vscale x 16 x i1> %0, <vscale x 16 x half> %1, <vscale x 16 x half> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv16f16_nxv16f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m4,tu,mu +; CHECK-NEXT: vfrece7.v v8, v12, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 16 x half> @llvm.riscv.vfrece7.mask.nxv16f16( + <vscale x 16 x half> %1, + <vscale x 16 x half> %2, + <vscale x 16 x i1> %0, + i32 %3) + + ret <vscale x 16 x half> %a +} + +declare <vscale x 32 x half> @llvm.riscv.vfrece7.nxv32f16( + <vscale x 32 x half>, + i32); + +define <vscale x 32 x half> @intrinsic_vfrece7_v_nxv32f16_nxv32f16(<vscale x 32 x half> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv32f16_nxv32f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m8,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 32 x half> @llvm.riscv.vfrece7.nxv32f16( + <vscale x 32 x half> %0, + i32 %1) + + ret <vscale x 32 x half> %a +} + +declare <vscale x 32 x half> @llvm.riscv.vfrece7.mask.nxv32f16( + <vscale x 32 x half>, + <vscale x 32 x half>, + <vscale x 32 x i1>, + i32); + +define <vscale x 32 x half> @intrinsic_vfrece7_mask_v_nxv32f16_nxv32f16(<vscale x 32 x i1> %0, <vscale x 32 x half> %1, <vscale x 32 x half> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv32f16_nxv32f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m8,tu,mu +; CHECK-NEXT: vfrece7.v v8, v16, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 32 x half> @llvm.riscv.vfrece7.mask.nxv32f16( + <vscale x 32 x half> %1, + <vscale x 32 x half> %2, + <vscale x 32 x i1> %0, + i32 %3) + + ret <vscale x 32 x half> %a +} + +declare <vscale x 1 x float> @llvm.riscv.vfrece7.nxv1f32( + <vscale x 1 x float>, + i32); + +define <vscale x 1 x float> @intrinsic_vfrece7_v_nxv1f32_nxv1f32(<vscale x 1 x float> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv1f32_nxv1f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,mf2,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 1 x float> @llvm.riscv.vfrece7.nxv1f32( + <vscale x 1 x float> %0, + i32 %1) + + ret <vscale x 1 x float> %a +} + +declare <vscale x 1 x float> @llvm.riscv.vfrece7.mask.nxv1f32( + <vscale x 1 x float>, + <vscale x 1 x float>, + <vscale x 1 x i1>, + i32); + +define <vscale x 1 x float> @intrinsic_vfrece7_mask_v_nxv1f32_nxv1f32(<vscale x 1 x i1> %0, <vscale x 1 x float> %1, <vscale x 1 x float> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv1f32_nxv1f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,mf2,tu,mu +; CHECK-NEXT: vfrece7.v v8, v9, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 1 x float> @llvm.riscv.vfrece7.mask.nxv1f32( + <vscale x 1 x float> %1, + <vscale x 1 x float> %2, + <vscale x 1 x i1> %0, + i32 %3) + + ret <vscale x 1 x float> %a +} + +declare <vscale x 2 x float> @llvm.riscv.vfrece7.nxv2f32( + <vscale x 2 x float>, + i32); + +define <vscale x 2 x float> @intrinsic_vfrece7_v_nxv2f32_nxv2f32(<vscale x 2 x float> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv2f32_nxv2f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m1,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 2 x float> @llvm.riscv.vfrece7.nxv2f32( + <vscale x 2 x float> %0, + i32 %1) + + ret <vscale x 2 x float> %a +} + +declare <vscale x 2 x float> @llvm.riscv.vfrece7.mask.nxv2f32( + <vscale x 2 x float>, + <vscale x 2 x float>, + <vscale x 2 x i1>, + i32); + +define <vscale x 2 x float> @intrinsic_vfrece7_mask_v_nxv2f32_nxv2f32(<vscale x 2 x i1> %0, <vscale x 2 x float> %1, <vscale x 2 x float> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv2f32_nxv2f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m1,tu,mu +; CHECK-NEXT: vfrece7.v v8, v9, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 2 x float> @llvm.riscv.vfrece7.mask.nxv2f32( + <vscale x 2 x float> %1, + <vscale x 2 x float> %2, + <vscale x 2 x i1> %0, + i32 %3) + + ret <vscale x 2 x float> %a +} + +declare <vscale x 4 x float> @llvm.riscv.vfrece7.nxv4f32( + <vscale x 4 x float>, + i32); + +define <vscale x 4 x float> @intrinsic_vfrece7_v_nxv4f32_nxv4f32(<vscale x 4 x float> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv4f32_nxv4f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m2,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 4 x float> @llvm.riscv.vfrece7.nxv4f32( + <vscale x 4 x float> %0, + i32 %1) + + ret <vscale x 4 x float> %a +} + +declare <vscale x 4 x float> @llvm.riscv.vfrece7.mask.nxv4f32( + <vscale x 4 x float>, + <vscale x 4 x float>, + <vscale x 4 x i1>, + i32); + +define <vscale x 4 x float> @intrinsic_vfrece7_mask_v_nxv4f32_nxv4f32(<vscale x 4 x i1> %0, <vscale x 4 x float> %1, <vscale x 4 x float> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv4f32_nxv4f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m2,tu,mu +; CHECK-NEXT: vfrece7.v v8, v10, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 4 x float> @llvm.riscv.vfrece7.mask.nxv4f32( + <vscale x 4 x float> %1, + <vscale x 4 x float> %2, + <vscale x 4 x i1> %0, + i32 %3) + + ret <vscale x 4 x float> %a +} + +declare <vscale x 8 x float> @llvm.riscv.vfrece7.nxv8f32( + <vscale x 8 x float>, + i32); + +define <vscale x 8 x float> @intrinsic_vfrece7_v_nxv8f32_nxv8f32(<vscale x 8 x float> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv8f32_nxv8f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m4,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 8 x float> @llvm.riscv.vfrece7.nxv8f32( + <vscale x 8 x float> %0, + i32 %1) + + ret <vscale x 8 x float> %a +} + +declare <vscale x 8 x float> @llvm.riscv.vfrece7.mask.nxv8f32( + <vscale x 8 x float>, + <vscale x 8 x float>, + <vscale x 8 x i1>, + i32); + +define <vscale x 8 x float> @intrinsic_vfrece7_mask_v_nxv8f32_nxv8f32(<vscale x 8 x i1> %0, <vscale x 8 x float> %1, <vscale x 8 x float> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv8f32_nxv8f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m4,tu,mu +; CHECK-NEXT: vfrece7.v v8, v12, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 8 x float> @llvm.riscv.vfrece7.mask.nxv8f32( + <vscale x 8 x float> %1, + <vscale x 8 x float> %2, + <vscale x 8 x i1> %0, + i32 %3) + + ret <vscale x 8 x float> %a +} + +declare <vscale x 16 x float> @llvm.riscv.vfrece7.nxv16f32( + <vscale x 16 x float>, + i32); + +define <vscale x 16 x float> @intrinsic_vfrece7_v_nxv16f32_nxv16f32(<vscale x 16 x float> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv16f32_nxv16f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m8,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 16 x float> @llvm.riscv.vfrece7.nxv16f32( + <vscale x 16 x float> %0, + i32 %1) + + ret <vscale x 16 x float> %a +} + +declare <vscale x 16 x float> @llvm.riscv.vfrece7.mask.nxv16f32( + <vscale x 16 x float>, + <vscale x 16 x float>, + <vscale x 16 x i1>, + i32); + +define <vscale x 16 x float> @intrinsic_vfrece7_mask_v_nxv16f32_nxv16f32(<vscale x 16 x i1> %0, <vscale x 16 x float> %1, <vscale x 16 x float> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv16f32_nxv16f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m8,tu,mu +; CHECK-NEXT: vfrece7.v v8, v16, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 16 x float> @llvm.riscv.vfrece7.mask.nxv16f32( + <vscale x 16 x float> %1, + <vscale x 16 x float> %2, + <vscale x 16 x i1> %0, + i32 %3) + + ret <vscale x 16 x float> %a +} + +declare <vscale x 1 x double> @llvm.riscv.vfrece7.nxv1f64( + <vscale x 1 x double>, + i32); + +define <vscale x 1 x double> @intrinsic_vfrece7_v_nxv1f64_nxv1f64(<vscale x 1 x double> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv1f64_nxv1f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m1,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 1 x double> @llvm.riscv.vfrece7.nxv1f64( + <vscale x 1 x double> %0, + i32 %1) + + ret <vscale x 1 x double> %a +} + +declare <vscale x 1 x double> @llvm.riscv.vfrece7.mask.nxv1f64( + <vscale x 1 x double>, + <vscale x 1 x double>, + <vscale x 1 x i1>, + i32); + +define <vscale x 1 x double> @intrinsic_vfrece7_mask_v_nxv1f64_nxv1f64(<vscale x 1 x i1> %0, <vscale x 1 x double> %1, <vscale x 1 x double> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv1f64_nxv1f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m1,tu,mu +; CHECK-NEXT: vfrece7.v v8, v9, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 1 x double> @llvm.riscv.vfrece7.mask.nxv1f64( + <vscale x 1 x double> %1, + <vscale x 1 x double> %2, + <vscale x 1 x i1> %0, + i32 %3) + + ret <vscale x 1 x double> %a +} + +declare <vscale x 2 x double> @llvm.riscv.vfrece7.nxv2f64( + <vscale x 2 x double>, + i32); + +define <vscale x 2 x double> @intrinsic_vfrece7_v_nxv2f64_nxv2f64(<vscale x 2 x double> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv2f64_nxv2f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m2,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 2 x double> @llvm.riscv.vfrece7.nxv2f64( + <vscale x 2 x double> %0, + i32 %1) + + ret <vscale x 2 x double> %a +} + +declare <vscale x 2 x double> @llvm.riscv.vfrece7.mask.nxv2f64( + <vscale x 2 x double>, + <vscale x 2 x double>, + <vscale x 2 x i1>, + i32); + +define <vscale x 2 x double> @intrinsic_vfrece7_mask_v_nxv2f64_nxv2f64(<vscale x 2 x i1> %0, <vscale x 2 x double> %1, <vscale x 2 x double> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv2f64_nxv2f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m2,tu,mu +; CHECK-NEXT: vfrece7.v v8, v10, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 2 x double> @llvm.riscv.vfrece7.mask.nxv2f64( + <vscale x 2 x double> %1, + <vscale x 2 x double> %2, + <vscale x 2 x i1> %0, + i32 %3) + + ret <vscale x 2 x double> %a +} + +declare <vscale x 4 x double> @llvm.riscv.vfrece7.nxv4f64( + <vscale x 4 x double>, + i32); + +define <vscale x 4 x double> @intrinsic_vfrece7_v_nxv4f64_nxv4f64(<vscale x 4 x double> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv4f64_nxv4f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m4,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 4 x double> @llvm.riscv.vfrece7.nxv4f64( + <vscale x 4 x double> %0, + i32 %1) + + ret <vscale x 4 x double> %a +} + +declare <vscale x 4 x double> @llvm.riscv.vfrece7.mask.nxv4f64( + <vscale x 4 x double>, + <vscale x 4 x double>, + <vscale x 4 x i1>, + i32); + +define <vscale x 4 x double> @intrinsic_vfrece7_mask_v_nxv4f64_nxv4f64(<vscale x 4 x i1> %0, <vscale x 4 x double> %1, <vscale x 4 x double> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv4f64_nxv4f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m4,tu,mu +; CHECK-NEXT: vfrece7.v v8, v12, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 4 x double> @llvm.riscv.vfrece7.mask.nxv4f64( + <vscale x 4 x double> %1, + <vscale x 4 x double> %2, + <vscale x 4 x i1> %0, + i32 %3) + + ret <vscale x 4 x double> %a +} + +declare <vscale x 8 x double> @llvm.riscv.vfrece7.nxv8f64( + <vscale x 8 x double>, + i32); + +define <vscale x 8 x double> @intrinsic_vfrece7_v_nxv8f64_nxv8f64(<vscale x 8 x double> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv8f64_nxv8f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m8,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 8 x double> @llvm.riscv.vfrece7.nxv8f64( + <vscale x 8 x double> %0, + i32 %1) + + ret <vscale x 8 x double> %a +} + +declare <vscale x 8 x double> @llvm.riscv.vfrece7.mask.nxv8f64( + <vscale x 8 x double>, + <vscale x 8 x double>, + <vscale x 8 x i1>, + i32); + +define <vscale x 8 x double> @intrinsic_vfrece7_mask_v_nxv8f64_nxv8f64(<vscale x 8 x i1> %0, <vscale x 8 x double> %1, <vscale x 8 x double> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv8f64_nxv8f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m8,tu,mu +; CHECK-NEXT: vfrece7.v v8, v16, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 8 x double> @llvm.riscv.vfrece7.mask.nxv8f64( + <vscale x 8 x double> %1, + <vscale x 8 x double> %2, + <vscale x 8 x i1> %0, + i32 %3) + + ret <vscale x 8 x double> %a +} diff --git a/llvm/test/CodeGen/RISCV/vfrece7-rv64.ll b/llvm/test/CodeGen/RISCV/vfrece7-rv64.ll new file mode 100644 index 000000000000..3af3fe4078c5 --- /dev/null +++ b/llvm/test/CodeGen/RISCV/vfrece7-rv64.ll @@ -0,0 +1,602 @@ +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py +; RUN: llc -mtriple=riscv64 -mattr=+experimental-v,+d,+experimental-zfh -verify-machineinstrs \ +; RUN: --riscv-no-aliases < %s | FileCheck %s +declare <vscale x 1 x half> @llvm.riscv.vfrece7.nxv1f16( + <vscale x 1 x half>, + i64); + +define <vscale x 1 x half> @intrinsic_vfrece7_v_nxv1f16_nxv1f16(<vscale x 1 x half> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv1f16_nxv1f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,mf4,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 1 x half> @llvm.riscv.vfrece7.nxv1f16( + <vscale x 1 x half> %0, + i64 %1) + + ret <vscale x 1 x half> %a +} + +declare <vscale x 1 x half> @llvm.riscv.vfrece7.mask.nxv1f16( + <vscale x 1 x half>, + <vscale x 1 x half>, + <vscale x 1 x i1>, + i64); + +define <vscale x 1 x half> @intrinsic_vfrece7_mask_v_nxv1f16_nxv1f16(<vscale x 1 x i1> %0, <vscale x 1 x half> %1, <vscale x 1 x half> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv1f16_nxv1f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,mf4,tu,mu +; CHECK-NEXT: vfrece7.v v8, v9, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 1 x half> @llvm.riscv.vfrece7.mask.nxv1f16( + <vscale x 1 x half> %1, + <vscale x 1 x half> %2, + <vscale x 1 x i1> %0, + i64 %3) + + ret <vscale x 1 x half> %a +} + +declare <vscale x 2 x half> @llvm.riscv.vfrece7.nxv2f16( + <vscale x 2 x half>, + i64); + +define <vscale x 2 x half> @intrinsic_vfrece7_v_nxv2f16_nxv2f16(<vscale x 2 x half> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv2f16_nxv2f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,mf2,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 2 x half> @llvm.riscv.vfrece7.nxv2f16( + <vscale x 2 x half> %0, + i64 %1) + + ret <vscale x 2 x half> %a +} + +declare <vscale x 2 x half> @llvm.riscv.vfrece7.mask.nxv2f16( + <vscale x 2 x half>, + <vscale x 2 x half>, + <vscale x 2 x i1>, + i64); + +define <vscale x 2 x half> @intrinsic_vfrece7_mask_v_nxv2f16_nxv2f16(<vscale x 2 x i1> %0, <vscale x 2 x half> %1, <vscale x 2 x half> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv2f16_nxv2f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,mf2,tu,mu +; CHECK-NEXT: vfrece7.v v8, v9, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 2 x half> @llvm.riscv.vfrece7.mask.nxv2f16( + <vscale x 2 x half> %1, + <vscale x 2 x half> %2, + <vscale x 2 x i1> %0, + i64 %3) + + ret <vscale x 2 x half> %a +} + +declare <vscale x 4 x half> @llvm.riscv.vfrece7.nxv4f16( + <vscale x 4 x half>, + i64); + +define <vscale x 4 x half> @intrinsic_vfrece7_v_nxv4f16_nxv4f16(<vscale x 4 x half> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv4f16_nxv4f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m1,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 4 x half> @llvm.riscv.vfrece7.nxv4f16( + <vscale x 4 x half> %0, + i64 %1) + + ret <vscale x 4 x half> %a +} + +declare <vscale x 4 x half> @llvm.riscv.vfrece7.mask.nxv4f16( + <vscale x 4 x half>, + <vscale x 4 x half>, + <vscale x 4 x i1>, + i64); + +define <vscale x 4 x half> @intrinsic_vfrece7_mask_v_nxv4f16_nxv4f16(<vscale x 4 x i1> %0, <vscale x 4 x half> %1, <vscale x 4 x half> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv4f16_nxv4f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m1,tu,mu +; CHECK-NEXT: vfrece7.v v8, v9, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 4 x half> @llvm.riscv.vfrece7.mask.nxv4f16( + <vscale x 4 x half> %1, + <vscale x 4 x half> %2, + <vscale x 4 x i1> %0, + i64 %3) + + ret <vscale x 4 x half> %a +} + +declare <vscale x 8 x half> @llvm.riscv.vfrece7.nxv8f16( + <vscale x 8 x half>, + i64); + +define <vscale x 8 x half> @intrinsic_vfrece7_v_nxv8f16_nxv8f16(<vscale x 8 x half> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv8f16_nxv8f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m2,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 8 x half> @llvm.riscv.vfrece7.nxv8f16( + <vscale x 8 x half> %0, + i64 %1) + + ret <vscale x 8 x half> %a +} + +declare <vscale x 8 x half> @llvm.riscv.vfrece7.mask.nxv8f16( + <vscale x 8 x half>, + <vscale x 8 x half>, + <vscale x 8 x i1>, + i64); + +define <vscale x 8 x half> @intrinsic_vfrece7_mask_v_nxv8f16_nxv8f16(<vscale x 8 x i1> %0, <vscale x 8 x half> %1, <vscale x 8 x half> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv8f16_nxv8f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m2,tu,mu +; CHECK-NEXT: vfrece7.v v8, v10, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 8 x half> @llvm.riscv.vfrece7.mask.nxv8f16( + <vscale x 8 x half> %1, + <vscale x 8 x half> %2, + <vscale x 8 x i1> %0, + i64 %3) + + ret <vscale x 8 x half> %a +} + +declare <vscale x 16 x half> @llvm.riscv.vfrece7.nxv16f16( + <vscale x 16 x half>, + i64); + +define <vscale x 16 x half> @intrinsic_vfrece7_v_nxv16f16_nxv16f16(<vscale x 16 x half> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv16f16_nxv16f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m4,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 16 x half> @llvm.riscv.vfrece7.nxv16f16( + <vscale x 16 x half> %0, + i64 %1) + + ret <vscale x 16 x half> %a +} + +declare <vscale x 16 x half> @llvm.riscv.vfrece7.mask.nxv16f16( + <vscale x 16 x half>, + <vscale x 16 x half>, + <vscale x 16 x i1>, + i64); + +define <vscale x 16 x half> @intrinsic_vfrece7_mask_v_nxv16f16_nxv16f16(<vscale x 16 x i1> %0, <vscale x 16 x half> %1, <vscale x 16 x half> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv16f16_nxv16f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m4,tu,mu +; CHECK-NEXT: vfrece7.v v8, v12, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 16 x half> @llvm.riscv.vfrece7.mask.nxv16f16( + <vscale x 16 x half> %1, + <vscale x 16 x half> %2, + <vscale x 16 x i1> %0, + i64 %3) + + ret <vscale x 16 x half> %a +} + +declare <vscale x 32 x half> @llvm.riscv.vfrece7.nxv32f16( + <vscale x 32 x half>, + i64); + +define <vscale x 32 x half> @intrinsic_vfrece7_v_nxv32f16_nxv32f16(<vscale x 32 x half> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv32f16_nxv32f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m8,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 32 x half> @llvm.riscv.vfrece7.nxv32f16( + <vscale x 32 x half> %0, + i64 %1) + + ret <vscale x 32 x half> %a +} + +declare <vscale x 32 x half> @llvm.riscv.vfrece7.mask.nxv32f16( + <vscale x 32 x half>, + <vscale x 32 x half>, + <vscale x 32 x i1>, + i64); + +define <vscale x 32 x half> @intrinsic_vfrece7_mask_v_nxv32f16_nxv32f16(<vscale x 32 x i1> %0, <vscale x 32 x half> %1, <vscale x 32 x half> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv32f16_nxv32f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m8,tu,mu +; CHECK-NEXT: vfrece7.v v8, v16, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 32 x half> @llvm.riscv.vfrece7.mask.nxv32f16( + <vscale x 32 x half> %1, + <vscale x 32 x half> %2, + <vscale x 32 x i1> %0, + i64 %3) + + ret <vscale x 32 x half> %a +} + +declare <vscale x 1 x float> @llvm.riscv.vfrece7.nxv1f32( + <vscale x 1 x float>, + i64); + +define <vscale x 1 x float> @intrinsic_vfrece7_v_nxv1f32_nxv1f32(<vscale x 1 x float> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv1f32_nxv1f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,mf2,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 1 x float> @llvm.riscv.vfrece7.nxv1f32( + <vscale x 1 x float> %0, + i64 %1) + + ret <vscale x 1 x float> %a +} + +declare <vscale x 1 x float> @llvm.riscv.vfrece7.mask.nxv1f32( + <vscale x 1 x float>, + <vscale x 1 x float>, + <vscale x 1 x i1>, + i64); + +define <vscale x 1 x float> @intrinsic_vfrece7_mask_v_nxv1f32_nxv1f32(<vscale x 1 x i1> %0, <vscale x 1 x float> %1, <vscale x 1 x float> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv1f32_nxv1f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,mf2,tu,mu +; CHECK-NEXT: vfrece7.v v8, v9, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 1 x float> @llvm.riscv.vfrece7.mask.nxv1f32( + <vscale x 1 x float> %1, + <vscale x 1 x float> %2, + <vscale x 1 x i1> %0, + i64 %3) + + ret <vscale x 1 x float> %a +} + +declare <vscale x 2 x float> @llvm.riscv.vfrece7.nxv2f32( + <vscale x 2 x float>, + i64); + +define <vscale x 2 x float> @intrinsic_vfrece7_v_nxv2f32_nxv2f32(<vscale x 2 x float> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv2f32_nxv2f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m1,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 2 x float> @llvm.riscv.vfrece7.nxv2f32( + <vscale x 2 x float> %0, + i64 %1) + + ret <vscale x 2 x float> %a +} + +declare <vscale x 2 x float> @llvm.riscv.vfrece7.mask.nxv2f32( + <vscale x 2 x float>, + <vscale x 2 x float>, + <vscale x 2 x i1>, + i64); + +define <vscale x 2 x float> @intrinsic_vfrece7_mask_v_nxv2f32_nxv2f32(<vscale x 2 x i1> %0, <vscale x 2 x float> %1, <vscale x 2 x float> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv2f32_nxv2f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m1,tu,mu +; CHECK-NEXT: vfrece7.v v8, v9, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 2 x float> @llvm.riscv.vfrece7.mask.nxv2f32( + <vscale x 2 x float> %1, + <vscale x 2 x float> %2, + <vscale x 2 x i1> %0, + i64 %3) + + ret <vscale x 2 x float> %a +} + +declare <vscale x 4 x float> @llvm.riscv.vfrece7.nxv4f32( + <vscale x 4 x float>, + i64); + +define <vscale x 4 x float> @intrinsic_vfrece7_v_nxv4f32_nxv4f32(<vscale x 4 x float> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv4f32_nxv4f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m2,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 4 x float> @llvm.riscv.vfrece7.nxv4f32( + <vscale x 4 x float> %0, + i64 %1) + + ret <vscale x 4 x float> %a +} + +declare <vscale x 4 x float> @llvm.riscv.vfrece7.mask.nxv4f32( + <vscale x 4 x float>, + <vscale x 4 x float>, + <vscale x 4 x i1>, + i64); + +define <vscale x 4 x float> @intrinsic_vfrece7_mask_v_nxv4f32_nxv4f32(<vscale x 4 x i1> %0, <vscale x 4 x float> %1, <vscale x 4 x float> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv4f32_nxv4f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m2,tu,mu +; CHECK-NEXT: vfrece7.v v8, v10, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 4 x float> @llvm.riscv.vfrece7.mask.nxv4f32( + <vscale x 4 x float> %1, + <vscale x 4 x float> %2, + <vscale x 4 x i1> %0, + i64 %3) + + ret <vscale x 4 x float> %a +} + +declare <vscale x 8 x float> @llvm.riscv.vfrece7.nxv8f32( + <vscale x 8 x float>, + i64); + +define <vscale x 8 x float> @intrinsic_vfrece7_v_nxv8f32_nxv8f32(<vscale x 8 x float> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv8f32_nxv8f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m4,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 8 x float> @llvm.riscv.vfrece7.nxv8f32( + <vscale x 8 x float> %0, + i64 %1) + + ret <vscale x 8 x float> %a +} + +declare <vscale x 8 x float> @llvm.riscv.vfrece7.mask.nxv8f32( + <vscale x 8 x float>, + <vscale x 8 x float>, + <vscale x 8 x i1>, + i64); + +define <vscale x 8 x float> @intrinsic_vfrece7_mask_v_nxv8f32_nxv8f32(<vscale x 8 x i1> %0, <vscale x 8 x float> %1, <vscale x 8 x float> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv8f32_nxv8f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m4,tu,mu +; CHECK-NEXT: vfrece7.v v8, v12, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 8 x float> @llvm.riscv.vfrece7.mask.nxv8f32( + <vscale x 8 x float> %1, + <vscale x 8 x float> %2, + <vscale x 8 x i1> %0, + i64 %3) + + ret <vscale x 8 x float> %a +} + +declare <vscale x 16 x float> @llvm.riscv.vfrece7.nxv16f32( + <vscale x 16 x float>, + i64); + +define <vscale x 16 x float> @intrinsic_vfrece7_v_nxv16f32_nxv16f32(<vscale x 16 x float> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv16f32_nxv16f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m8,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 16 x float> @llvm.riscv.vfrece7.nxv16f32( + <vscale x 16 x float> %0, + i64 %1) + + ret <vscale x 16 x float> %a +} + +declare <vscale x 16 x float> @llvm.riscv.vfrece7.mask.nxv16f32( + <vscale x 16 x float>, + <vscale x 16 x float>, + <vscale x 16 x i1>, + i64); + +define <vscale x 16 x float> @intrinsic_vfrece7_mask_v_nxv16f32_nxv16f32(<vscale x 16 x i1> %0, <vscale x 16 x float> %1, <vscale x 16 x float> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv16f32_nxv16f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m8,tu,mu +; CHECK-NEXT: vfrece7.v v8, v16, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 16 x float> @llvm.riscv.vfrece7.mask.nxv16f32( + <vscale x 16 x float> %1, + <vscale x 16 x float> %2, + <vscale x 16 x i1> %0, + i64 %3) + + ret <vscale x 16 x float> %a +} + +declare <vscale x 1 x double> @llvm.riscv.vfrece7.nxv1f64( + <vscale x 1 x double>, + i64); + +define <vscale x 1 x double> @intrinsic_vfrece7_v_nxv1f64_nxv1f64(<vscale x 1 x double> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv1f64_nxv1f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m1,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 1 x double> @llvm.riscv.vfrece7.nxv1f64( + <vscale x 1 x double> %0, + i64 %1) + + ret <vscale x 1 x double> %a +} + +declare <vscale x 1 x double> @llvm.riscv.vfrece7.mask.nxv1f64( + <vscale x 1 x double>, + <vscale x 1 x double>, + <vscale x 1 x i1>, + i64); + +define <vscale x 1 x double> @intrinsic_vfrece7_mask_v_nxv1f64_nxv1f64(<vscale x 1 x i1> %0, <vscale x 1 x double> %1, <vscale x 1 x double> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv1f64_nxv1f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m1,tu,mu +; CHECK-NEXT: vfrece7.v v8, v9, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 1 x double> @llvm.riscv.vfrece7.mask.nxv1f64( + <vscale x 1 x double> %1, + <vscale x 1 x double> %2, + <vscale x 1 x i1> %0, + i64 %3) + + ret <vscale x 1 x double> %a +} + +declare <vscale x 2 x double> @llvm.riscv.vfrece7.nxv2f64( + <vscale x 2 x double>, + i64); + +define <vscale x 2 x double> @intrinsic_vfrece7_v_nxv2f64_nxv2f64(<vscale x 2 x double> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv2f64_nxv2f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m2,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 2 x double> @llvm.riscv.vfrece7.nxv2f64( + <vscale x 2 x double> %0, + i64 %1) + + ret <vscale x 2 x double> %a +} + +declare <vscale x 2 x double> @llvm.riscv.vfrece7.mask.nxv2f64( + <vscale x 2 x double>, + <vscale x 2 x double>, + <vscale x 2 x i1>, + i64); + +define <vscale x 2 x double> @intrinsic_vfrece7_mask_v_nxv2f64_nxv2f64(<vscale x 2 x i1> %0, <vscale x 2 x double> %1, <vscale x 2 x double> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv2f64_nxv2f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m2,tu,mu +; CHECK-NEXT: vfrece7.v v8, v10, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 2 x double> @llvm.riscv.vfrece7.mask.nxv2f64( + <vscale x 2 x double> %1, + <vscale x 2 x double> %2, + <vscale x 2 x i1> %0, + i64 %3) + + ret <vscale x 2 x double> %a +} + +declare <vscale x 4 x double> @llvm.riscv.vfrece7.nxv4f64( + <vscale x 4 x double>, + i64); + +define <vscale x 4 x double> @intrinsic_vfrece7_v_nxv4f64_nxv4f64(<vscale x 4 x double> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv4f64_nxv4f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m4,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 4 x double> @llvm.riscv.vfrece7.nxv4f64( + <vscale x 4 x double> %0, + i64 %1) + + ret <vscale x 4 x double> %a +} + +declare <vscale x 4 x double> @llvm.riscv.vfrece7.mask.nxv4f64( + <vscale x 4 x double>, + <vscale x 4 x double>, + <vscale x 4 x i1>, + i64); + +define <vscale x 4 x double> @intrinsic_vfrece7_mask_v_nxv4f64_nxv4f64(<vscale x 4 x i1> %0, <vscale x 4 x double> %1, <vscale x 4 x double> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv4f64_nxv4f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m4,tu,mu +; CHECK-NEXT: vfrece7.v v8, v12, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 4 x double> @llvm.riscv.vfrece7.mask.nxv4f64( + <vscale x 4 x double> %1, + <vscale x 4 x double> %2, + <vscale x 4 x i1> %0, + i64 %3) + + ret <vscale x 4 x double> %a +} + +declare <vscale x 8 x double> @llvm.riscv.vfrece7.nxv8f64( + <vscale x 8 x double>, + i64); + +define <vscale x 8 x double> @intrinsic_vfrece7_v_nxv8f64_nxv8f64(<vscale x 8 x double> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_v_nxv8f64_nxv8f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m8,ta,mu +; CHECK-NEXT: vfrece7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 8 x double> @llvm.riscv.vfrece7.nxv8f64( + <vscale x 8 x double> %0, + i64 %1) + + ret <vscale x 8 x double> %a +} + +declare <vscale x 8 x double> @llvm.riscv.vfrece7.mask.nxv8f64( + <vscale x 8 x double>, + <vscale x 8 x double>, + <vscale x 8 x i1>, + i64); + +define <vscale x 8 x double> @intrinsic_vfrece7_mask_v_nxv8f64_nxv8f64(<vscale x 8 x i1> %0, <vscale x 8 x double> %1, <vscale x 8 x double> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrece7_mask_v_nxv8f64_nxv8f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m8,tu,mu +; CHECK-NEXT: vfrece7.v v8, v16, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 8 x double> @llvm.riscv.vfrece7.mask.nxv8f64( + <vscale x 8 x double> %1, + <vscale x 8 x double> %2, + <vscale x 8 x i1> %0, + i64 %3) + + ret <vscale x 8 x double> %a +} diff --git a/llvm/test/CodeGen/RISCV/vfrsqrte7-rv32.ll b/llvm/test/CodeGen/RISCV/vfrsqrte7-rv32.ll new file mode 100644 index 000000000000..083b411ef3c8 --- /dev/null +++ b/llvm/test/CodeGen/RISCV/vfrsqrte7-rv32.ll @@ -0,0 +1,602 @@ +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py +; RUN: llc -mtriple=riscv32 -mattr=+experimental-v,+d,+experimental-zfh -verify-machineinstrs \ +; RUN: --riscv-no-aliases < %s | FileCheck %s +declare <vscale x 1 x half> @llvm.riscv.vfrsqrte7.nxv1f16( + <vscale x 1 x half>, + i32); + +define <vscale x 1 x half> @intrinsic_vfrsqrte7_v_nxv1f16_nxv1f16(<vscale x 1 x half> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv1f16_nxv1f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,mf4,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 1 x half> @llvm.riscv.vfrsqrte7.nxv1f16( + <vscale x 1 x half> %0, + i32 %1) + + ret <vscale x 1 x half> %a +} + +declare <vscale x 1 x half> @llvm.riscv.vfrsqrte7.mask.nxv1f16( + <vscale x 1 x half>, + <vscale x 1 x half>, + <vscale x 1 x i1>, + i32); + +define <vscale x 1 x half> @intrinsic_vfrsqrte7_mask_v_nxv1f16_nxv1f16(<vscale x 1 x i1> %0, <vscale x 1 x half> %1, <vscale x 1 x half> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv1f16_nxv1f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,mf4,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v9, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 1 x half> @llvm.riscv.vfrsqrte7.mask.nxv1f16( + <vscale x 1 x half> %1, + <vscale x 1 x half> %2, + <vscale x 1 x i1> %0, + i32 %3) + + ret <vscale x 1 x half> %a +} + +declare <vscale x 2 x half> @llvm.riscv.vfrsqrte7.nxv2f16( + <vscale x 2 x half>, + i32); + +define <vscale x 2 x half> @intrinsic_vfrsqrte7_v_nxv2f16_nxv2f16(<vscale x 2 x half> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv2f16_nxv2f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,mf2,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 2 x half> @llvm.riscv.vfrsqrte7.nxv2f16( + <vscale x 2 x half> %0, + i32 %1) + + ret <vscale x 2 x half> %a +} + +declare <vscale x 2 x half> @llvm.riscv.vfrsqrte7.mask.nxv2f16( + <vscale x 2 x half>, + <vscale x 2 x half>, + <vscale x 2 x i1>, + i32); + +define <vscale x 2 x half> @intrinsic_vfrsqrte7_mask_v_nxv2f16_nxv2f16(<vscale x 2 x i1> %0, <vscale x 2 x half> %1, <vscale x 2 x half> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv2f16_nxv2f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,mf2,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v9, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 2 x half> @llvm.riscv.vfrsqrte7.mask.nxv2f16( + <vscale x 2 x half> %1, + <vscale x 2 x half> %2, + <vscale x 2 x i1> %0, + i32 %3) + + ret <vscale x 2 x half> %a +} + +declare <vscale x 4 x half> @llvm.riscv.vfrsqrte7.nxv4f16( + <vscale x 4 x half>, + i32); + +define <vscale x 4 x half> @intrinsic_vfrsqrte7_v_nxv4f16_nxv4f16(<vscale x 4 x half> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv4f16_nxv4f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m1,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 4 x half> @llvm.riscv.vfrsqrte7.nxv4f16( + <vscale x 4 x half> %0, + i32 %1) + + ret <vscale x 4 x half> %a +} + +declare <vscale x 4 x half> @llvm.riscv.vfrsqrte7.mask.nxv4f16( + <vscale x 4 x half>, + <vscale x 4 x half>, + <vscale x 4 x i1>, + i32); + +define <vscale x 4 x half> @intrinsic_vfrsqrte7_mask_v_nxv4f16_nxv4f16(<vscale x 4 x i1> %0, <vscale x 4 x half> %1, <vscale x 4 x half> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv4f16_nxv4f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m1,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v9, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 4 x half> @llvm.riscv.vfrsqrte7.mask.nxv4f16( + <vscale x 4 x half> %1, + <vscale x 4 x half> %2, + <vscale x 4 x i1> %0, + i32 %3) + + ret <vscale x 4 x half> %a +} + +declare <vscale x 8 x half> @llvm.riscv.vfrsqrte7.nxv8f16( + <vscale x 8 x half>, + i32); + +define <vscale x 8 x half> @intrinsic_vfrsqrte7_v_nxv8f16_nxv8f16(<vscale x 8 x half> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv8f16_nxv8f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m2,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 8 x half> @llvm.riscv.vfrsqrte7.nxv8f16( + <vscale x 8 x half> %0, + i32 %1) + + ret <vscale x 8 x half> %a +} + +declare <vscale x 8 x half> @llvm.riscv.vfrsqrte7.mask.nxv8f16( + <vscale x 8 x half>, + <vscale x 8 x half>, + <vscale x 8 x i1>, + i32); + +define <vscale x 8 x half> @intrinsic_vfrsqrte7_mask_v_nxv8f16_nxv8f16(<vscale x 8 x i1> %0, <vscale x 8 x half> %1, <vscale x 8 x half> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv8f16_nxv8f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m2,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v10, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 8 x half> @llvm.riscv.vfrsqrte7.mask.nxv8f16( + <vscale x 8 x half> %1, + <vscale x 8 x half> %2, + <vscale x 8 x i1> %0, + i32 %3) + + ret <vscale x 8 x half> %a +} + +declare <vscale x 16 x half> @llvm.riscv.vfrsqrte7.nxv16f16( + <vscale x 16 x half>, + i32); + +define <vscale x 16 x half> @intrinsic_vfrsqrte7_v_nxv16f16_nxv16f16(<vscale x 16 x half> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv16f16_nxv16f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m4,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 16 x half> @llvm.riscv.vfrsqrte7.nxv16f16( + <vscale x 16 x half> %0, + i32 %1) + + ret <vscale x 16 x half> %a +} + +declare <vscale x 16 x half> @llvm.riscv.vfrsqrte7.mask.nxv16f16( + <vscale x 16 x half>, + <vscale x 16 x half>, + <vscale x 16 x i1>, + i32); + +define <vscale x 16 x half> @intrinsic_vfrsqrte7_mask_v_nxv16f16_nxv16f16(<vscale x 16 x i1> %0, <vscale x 16 x half> %1, <vscale x 16 x half> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv16f16_nxv16f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m4,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v12, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 16 x half> @llvm.riscv.vfrsqrte7.mask.nxv16f16( + <vscale x 16 x half> %1, + <vscale x 16 x half> %2, + <vscale x 16 x i1> %0, + i32 %3) + + ret <vscale x 16 x half> %a +} + +declare <vscale x 32 x half> @llvm.riscv.vfrsqrte7.nxv32f16( + <vscale x 32 x half>, + i32); + +define <vscale x 32 x half> @intrinsic_vfrsqrte7_v_nxv32f16_nxv32f16(<vscale x 32 x half> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv32f16_nxv32f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m8,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 32 x half> @llvm.riscv.vfrsqrte7.nxv32f16( + <vscale x 32 x half> %0, + i32 %1) + + ret <vscale x 32 x half> %a +} + +declare <vscale x 32 x half> @llvm.riscv.vfrsqrte7.mask.nxv32f16( + <vscale x 32 x half>, + <vscale x 32 x half>, + <vscale x 32 x i1>, + i32); + +define <vscale x 32 x half> @intrinsic_vfrsqrte7_mask_v_nxv32f16_nxv32f16(<vscale x 32 x i1> %0, <vscale x 32 x half> %1, <vscale x 32 x half> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv32f16_nxv32f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m8,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v16, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 32 x half> @llvm.riscv.vfrsqrte7.mask.nxv32f16( + <vscale x 32 x half> %1, + <vscale x 32 x half> %2, + <vscale x 32 x i1> %0, + i32 %3) + + ret <vscale x 32 x half> %a +} + +declare <vscale x 1 x float> @llvm.riscv.vfrsqrte7.nxv1f32( + <vscale x 1 x float>, + i32); + +define <vscale x 1 x float> @intrinsic_vfrsqrte7_v_nxv1f32_nxv1f32(<vscale x 1 x float> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv1f32_nxv1f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,mf2,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 1 x float> @llvm.riscv.vfrsqrte7.nxv1f32( + <vscale x 1 x float> %0, + i32 %1) + + ret <vscale x 1 x float> %a +} + +declare <vscale x 1 x float> @llvm.riscv.vfrsqrte7.mask.nxv1f32( + <vscale x 1 x float>, + <vscale x 1 x float>, + <vscale x 1 x i1>, + i32); + +define <vscale x 1 x float> @intrinsic_vfrsqrte7_mask_v_nxv1f32_nxv1f32(<vscale x 1 x i1> %0, <vscale x 1 x float> %1, <vscale x 1 x float> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv1f32_nxv1f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,mf2,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v9, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 1 x float> @llvm.riscv.vfrsqrte7.mask.nxv1f32( + <vscale x 1 x float> %1, + <vscale x 1 x float> %2, + <vscale x 1 x i1> %0, + i32 %3) + + ret <vscale x 1 x float> %a +} + +declare <vscale x 2 x float> @llvm.riscv.vfrsqrte7.nxv2f32( + <vscale x 2 x float>, + i32); + +define <vscale x 2 x float> @intrinsic_vfrsqrte7_v_nxv2f32_nxv2f32(<vscale x 2 x float> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv2f32_nxv2f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m1,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 2 x float> @llvm.riscv.vfrsqrte7.nxv2f32( + <vscale x 2 x float> %0, + i32 %1) + + ret <vscale x 2 x float> %a +} + +declare <vscale x 2 x float> @llvm.riscv.vfrsqrte7.mask.nxv2f32( + <vscale x 2 x float>, + <vscale x 2 x float>, + <vscale x 2 x i1>, + i32); + +define <vscale x 2 x float> @intrinsic_vfrsqrte7_mask_v_nxv2f32_nxv2f32(<vscale x 2 x i1> %0, <vscale x 2 x float> %1, <vscale x 2 x float> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv2f32_nxv2f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m1,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v9, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 2 x float> @llvm.riscv.vfrsqrte7.mask.nxv2f32( + <vscale x 2 x float> %1, + <vscale x 2 x float> %2, + <vscale x 2 x i1> %0, + i32 %3) + + ret <vscale x 2 x float> %a +} + +declare <vscale x 4 x float> @llvm.riscv.vfrsqrte7.nxv4f32( + <vscale x 4 x float>, + i32); + +define <vscale x 4 x float> @intrinsic_vfrsqrte7_v_nxv4f32_nxv4f32(<vscale x 4 x float> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv4f32_nxv4f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m2,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 4 x float> @llvm.riscv.vfrsqrte7.nxv4f32( + <vscale x 4 x float> %0, + i32 %1) + + ret <vscale x 4 x float> %a +} + +declare <vscale x 4 x float> @llvm.riscv.vfrsqrte7.mask.nxv4f32( + <vscale x 4 x float>, + <vscale x 4 x float>, + <vscale x 4 x i1>, + i32); + +define <vscale x 4 x float> @intrinsic_vfrsqrte7_mask_v_nxv4f32_nxv4f32(<vscale x 4 x i1> %0, <vscale x 4 x float> %1, <vscale x 4 x float> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv4f32_nxv4f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m2,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v10, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 4 x float> @llvm.riscv.vfrsqrte7.mask.nxv4f32( + <vscale x 4 x float> %1, + <vscale x 4 x float> %2, + <vscale x 4 x i1> %0, + i32 %3) + + ret <vscale x 4 x float> %a +} + +declare <vscale x 8 x float> @llvm.riscv.vfrsqrte7.nxv8f32( + <vscale x 8 x float>, + i32); + +define <vscale x 8 x float> @intrinsic_vfrsqrte7_v_nxv8f32_nxv8f32(<vscale x 8 x float> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv8f32_nxv8f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m4,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 8 x float> @llvm.riscv.vfrsqrte7.nxv8f32( + <vscale x 8 x float> %0, + i32 %1) + + ret <vscale x 8 x float> %a +} + +declare <vscale x 8 x float> @llvm.riscv.vfrsqrte7.mask.nxv8f32( + <vscale x 8 x float>, + <vscale x 8 x float>, + <vscale x 8 x i1>, + i32); + +define <vscale x 8 x float> @intrinsic_vfrsqrte7_mask_v_nxv8f32_nxv8f32(<vscale x 8 x i1> %0, <vscale x 8 x float> %1, <vscale x 8 x float> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv8f32_nxv8f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m4,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v12, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 8 x float> @llvm.riscv.vfrsqrte7.mask.nxv8f32( + <vscale x 8 x float> %1, + <vscale x 8 x float> %2, + <vscale x 8 x i1> %0, + i32 %3) + + ret <vscale x 8 x float> %a +} + +declare <vscale x 16 x float> @llvm.riscv.vfrsqrte7.nxv16f32( + <vscale x 16 x float>, + i32); + +define <vscale x 16 x float> @intrinsic_vfrsqrte7_v_nxv16f32_nxv16f32(<vscale x 16 x float> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv16f32_nxv16f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m8,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 16 x float> @llvm.riscv.vfrsqrte7.nxv16f32( + <vscale x 16 x float> %0, + i32 %1) + + ret <vscale x 16 x float> %a +} + +declare <vscale x 16 x float> @llvm.riscv.vfrsqrte7.mask.nxv16f32( + <vscale x 16 x float>, + <vscale x 16 x float>, + <vscale x 16 x i1>, + i32); + +define <vscale x 16 x float> @intrinsic_vfrsqrte7_mask_v_nxv16f32_nxv16f32(<vscale x 16 x i1> %0, <vscale x 16 x float> %1, <vscale x 16 x float> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv16f32_nxv16f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m8,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v16, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 16 x float> @llvm.riscv.vfrsqrte7.mask.nxv16f32( + <vscale x 16 x float> %1, + <vscale x 16 x float> %2, + <vscale x 16 x i1> %0, + i32 %3) + + ret <vscale x 16 x float> %a +} + +declare <vscale x 1 x double> @llvm.riscv.vfrsqrte7.nxv1f64( + <vscale x 1 x double>, + i32); + +define <vscale x 1 x double> @intrinsic_vfrsqrte7_v_nxv1f64_nxv1f64(<vscale x 1 x double> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv1f64_nxv1f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m1,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 1 x double> @llvm.riscv.vfrsqrte7.nxv1f64( + <vscale x 1 x double> %0, + i32 %1) + + ret <vscale x 1 x double> %a +} + +declare <vscale x 1 x double> @llvm.riscv.vfrsqrte7.mask.nxv1f64( + <vscale x 1 x double>, + <vscale x 1 x double>, + <vscale x 1 x i1>, + i32); + +define <vscale x 1 x double> @intrinsic_vfrsqrte7_mask_v_nxv1f64_nxv1f64(<vscale x 1 x i1> %0, <vscale x 1 x double> %1, <vscale x 1 x double> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv1f64_nxv1f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m1,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v9, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 1 x double> @llvm.riscv.vfrsqrte7.mask.nxv1f64( + <vscale x 1 x double> %1, + <vscale x 1 x double> %2, + <vscale x 1 x i1> %0, + i32 %3) + + ret <vscale x 1 x double> %a +} + +declare <vscale x 2 x double> @llvm.riscv.vfrsqrte7.nxv2f64( + <vscale x 2 x double>, + i32); + +define <vscale x 2 x double> @intrinsic_vfrsqrte7_v_nxv2f64_nxv2f64(<vscale x 2 x double> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv2f64_nxv2f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m2,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 2 x double> @llvm.riscv.vfrsqrte7.nxv2f64( + <vscale x 2 x double> %0, + i32 %1) + + ret <vscale x 2 x double> %a +} + +declare <vscale x 2 x double> @llvm.riscv.vfrsqrte7.mask.nxv2f64( + <vscale x 2 x double>, + <vscale x 2 x double>, + <vscale x 2 x i1>, + i32); + +define <vscale x 2 x double> @intrinsic_vfrsqrte7_mask_v_nxv2f64_nxv2f64(<vscale x 2 x i1> %0, <vscale x 2 x double> %1, <vscale x 2 x double> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv2f64_nxv2f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m2,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v10, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 2 x double> @llvm.riscv.vfrsqrte7.mask.nxv2f64( + <vscale x 2 x double> %1, + <vscale x 2 x double> %2, + <vscale x 2 x i1> %0, + i32 %3) + + ret <vscale x 2 x double> %a +} + +declare <vscale x 4 x double> @llvm.riscv.vfrsqrte7.nxv4f64( + <vscale x 4 x double>, + i32); + +define <vscale x 4 x double> @intrinsic_vfrsqrte7_v_nxv4f64_nxv4f64(<vscale x 4 x double> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv4f64_nxv4f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m4,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 4 x double> @llvm.riscv.vfrsqrte7.nxv4f64( + <vscale x 4 x double> %0, + i32 %1) + + ret <vscale x 4 x double> %a +} + +declare <vscale x 4 x double> @llvm.riscv.vfrsqrte7.mask.nxv4f64( + <vscale x 4 x double>, + <vscale x 4 x double>, + <vscale x 4 x i1>, + i32); + +define <vscale x 4 x double> @intrinsic_vfrsqrte7_mask_v_nxv4f64_nxv4f64(<vscale x 4 x i1> %0, <vscale x 4 x double> %1, <vscale x 4 x double> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv4f64_nxv4f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m4,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v12, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 4 x double> @llvm.riscv.vfrsqrte7.mask.nxv4f64( + <vscale x 4 x double> %1, + <vscale x 4 x double> %2, + <vscale x 4 x i1> %0, + i32 %3) + + ret <vscale x 4 x double> %a +} + +declare <vscale x 8 x double> @llvm.riscv.vfrsqrte7.nxv8f64( + <vscale x 8 x double>, + i32); + +define <vscale x 8 x double> @intrinsic_vfrsqrte7_v_nxv8f64_nxv8f64(<vscale x 8 x double> %0, i32 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv8f64_nxv8f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m8,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 8 x double> @llvm.riscv.vfrsqrte7.nxv8f64( + <vscale x 8 x double> %0, + i32 %1) + + ret <vscale x 8 x double> %a +} + +declare <vscale x 8 x double> @llvm.riscv.vfrsqrte7.mask.nxv8f64( + <vscale x 8 x double>, + <vscale x 8 x double>, + <vscale x 8 x i1>, + i32); + +define <vscale x 8 x double> @intrinsic_vfrsqrte7_mask_v_nxv8f64_nxv8f64(<vscale x 8 x i1> %0, <vscale x 8 x double> %1, <vscale x 8 x double> %2, i32 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv8f64_nxv8f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m8,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v16, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 8 x double> @llvm.riscv.vfrsqrte7.mask.nxv8f64( + <vscale x 8 x double> %1, + <vscale x 8 x double> %2, + <vscale x 8 x i1> %0, + i32 %3) + + ret <vscale x 8 x double> %a +} diff --git a/llvm/test/CodeGen/RISCV/vfrsqrte7-rv64.ll b/llvm/test/CodeGen/RISCV/vfrsqrte7-rv64.ll new file mode 100644 index 000000000000..d0f4c9c4ac58 --- /dev/null +++ b/llvm/test/CodeGen/RISCV/vfrsqrte7-rv64.ll @@ -0,0 +1,602 @@ +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py +; RUN: llc -mtriple=riscv64 -mattr=+experimental-v,+d,+experimental-zfh -verify-machineinstrs \ +; RUN: --riscv-no-aliases < %s | FileCheck %s +declare <vscale x 1 x half> @llvm.riscv.vfrsqrte7.nxv1f16( + <vscale x 1 x half>, + i64); + +define <vscale x 1 x half> @intrinsic_vfrsqrte7_v_nxv1f16_nxv1f16(<vscale x 1 x half> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv1f16_nxv1f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,mf4,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 1 x half> @llvm.riscv.vfrsqrte7.nxv1f16( + <vscale x 1 x half> %0, + i64 %1) + + ret <vscale x 1 x half> %a +} + +declare <vscale x 1 x half> @llvm.riscv.vfrsqrte7.mask.nxv1f16( + <vscale x 1 x half>, + <vscale x 1 x half>, + <vscale x 1 x i1>, + i64); + +define <vscale x 1 x half> @intrinsic_vfrsqrte7_mask_v_nxv1f16_nxv1f16(<vscale x 1 x i1> %0, <vscale x 1 x half> %1, <vscale x 1 x half> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv1f16_nxv1f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,mf4,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v9, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 1 x half> @llvm.riscv.vfrsqrte7.mask.nxv1f16( + <vscale x 1 x half> %1, + <vscale x 1 x half> %2, + <vscale x 1 x i1> %0, + i64 %3) + + ret <vscale x 1 x half> %a +} + +declare <vscale x 2 x half> @llvm.riscv.vfrsqrte7.nxv2f16( + <vscale x 2 x half>, + i64); + +define <vscale x 2 x half> @intrinsic_vfrsqrte7_v_nxv2f16_nxv2f16(<vscale x 2 x half> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv2f16_nxv2f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,mf2,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 2 x half> @llvm.riscv.vfrsqrte7.nxv2f16( + <vscale x 2 x half> %0, + i64 %1) + + ret <vscale x 2 x half> %a +} + +declare <vscale x 2 x half> @llvm.riscv.vfrsqrte7.mask.nxv2f16( + <vscale x 2 x half>, + <vscale x 2 x half>, + <vscale x 2 x i1>, + i64); + +define <vscale x 2 x half> @intrinsic_vfrsqrte7_mask_v_nxv2f16_nxv2f16(<vscale x 2 x i1> %0, <vscale x 2 x half> %1, <vscale x 2 x half> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv2f16_nxv2f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,mf2,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v9, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 2 x half> @llvm.riscv.vfrsqrte7.mask.nxv2f16( + <vscale x 2 x half> %1, + <vscale x 2 x half> %2, + <vscale x 2 x i1> %0, + i64 %3) + + ret <vscale x 2 x half> %a +} + +declare <vscale x 4 x half> @llvm.riscv.vfrsqrte7.nxv4f16( + <vscale x 4 x half>, + i64); + +define <vscale x 4 x half> @intrinsic_vfrsqrte7_v_nxv4f16_nxv4f16(<vscale x 4 x half> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv4f16_nxv4f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m1,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 4 x half> @llvm.riscv.vfrsqrte7.nxv4f16( + <vscale x 4 x half> %0, + i64 %1) + + ret <vscale x 4 x half> %a +} + +declare <vscale x 4 x half> @llvm.riscv.vfrsqrte7.mask.nxv4f16( + <vscale x 4 x half>, + <vscale x 4 x half>, + <vscale x 4 x i1>, + i64); + +define <vscale x 4 x half> @intrinsic_vfrsqrte7_mask_v_nxv4f16_nxv4f16(<vscale x 4 x i1> %0, <vscale x 4 x half> %1, <vscale x 4 x half> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv4f16_nxv4f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m1,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v9, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 4 x half> @llvm.riscv.vfrsqrte7.mask.nxv4f16( + <vscale x 4 x half> %1, + <vscale x 4 x half> %2, + <vscale x 4 x i1> %0, + i64 %3) + + ret <vscale x 4 x half> %a +} + +declare <vscale x 8 x half> @llvm.riscv.vfrsqrte7.nxv8f16( + <vscale x 8 x half>, + i64); + +define <vscale x 8 x half> @intrinsic_vfrsqrte7_v_nxv8f16_nxv8f16(<vscale x 8 x half> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv8f16_nxv8f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m2,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 8 x half> @llvm.riscv.vfrsqrte7.nxv8f16( + <vscale x 8 x half> %0, + i64 %1) + + ret <vscale x 8 x half> %a +} + +declare <vscale x 8 x half> @llvm.riscv.vfrsqrte7.mask.nxv8f16( + <vscale x 8 x half>, + <vscale x 8 x half>, + <vscale x 8 x i1>, + i64); + +define <vscale x 8 x half> @intrinsic_vfrsqrte7_mask_v_nxv8f16_nxv8f16(<vscale x 8 x i1> %0, <vscale x 8 x half> %1, <vscale x 8 x half> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv8f16_nxv8f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m2,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v10, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 8 x half> @llvm.riscv.vfrsqrte7.mask.nxv8f16( + <vscale x 8 x half> %1, + <vscale x 8 x half> %2, + <vscale x 8 x i1> %0, + i64 %3) + + ret <vscale x 8 x half> %a +} + +declare <vscale x 16 x half> @llvm.riscv.vfrsqrte7.nxv16f16( + <vscale x 16 x half>, + i64); + +define <vscale x 16 x half> @intrinsic_vfrsqrte7_v_nxv16f16_nxv16f16(<vscale x 16 x half> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv16f16_nxv16f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m4,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 16 x half> @llvm.riscv.vfrsqrte7.nxv16f16( + <vscale x 16 x half> %0, + i64 %1) + + ret <vscale x 16 x half> %a +} + +declare <vscale x 16 x half> @llvm.riscv.vfrsqrte7.mask.nxv16f16( + <vscale x 16 x half>, + <vscale x 16 x half>, + <vscale x 16 x i1>, + i64); + +define <vscale x 16 x half> @intrinsic_vfrsqrte7_mask_v_nxv16f16_nxv16f16(<vscale x 16 x i1> %0, <vscale x 16 x half> %1, <vscale x 16 x half> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv16f16_nxv16f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m4,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v12, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 16 x half> @llvm.riscv.vfrsqrte7.mask.nxv16f16( + <vscale x 16 x half> %1, + <vscale x 16 x half> %2, + <vscale x 16 x i1> %0, + i64 %3) + + ret <vscale x 16 x half> %a +} + +declare <vscale x 32 x half> @llvm.riscv.vfrsqrte7.nxv32f16( + <vscale x 32 x half>, + i64); + +define <vscale x 32 x half> @intrinsic_vfrsqrte7_v_nxv32f16_nxv32f16(<vscale x 32 x half> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv32f16_nxv32f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m8,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 32 x half> @llvm.riscv.vfrsqrte7.nxv32f16( + <vscale x 32 x half> %0, + i64 %1) + + ret <vscale x 32 x half> %a +} + +declare <vscale x 32 x half> @llvm.riscv.vfrsqrte7.mask.nxv32f16( + <vscale x 32 x half>, + <vscale x 32 x half>, + <vscale x 32 x i1>, + i64); + +define <vscale x 32 x half> @intrinsic_vfrsqrte7_mask_v_nxv32f16_nxv32f16(<vscale x 32 x i1> %0, <vscale x 32 x half> %1, <vscale x 32 x half> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv32f16_nxv32f16: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e16,m8,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v16, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 32 x half> @llvm.riscv.vfrsqrte7.mask.nxv32f16( + <vscale x 32 x half> %1, + <vscale x 32 x half> %2, + <vscale x 32 x i1> %0, + i64 %3) + + ret <vscale x 32 x half> %a +} + +declare <vscale x 1 x float> @llvm.riscv.vfrsqrte7.nxv1f32( + <vscale x 1 x float>, + i64); + +define <vscale x 1 x float> @intrinsic_vfrsqrte7_v_nxv1f32_nxv1f32(<vscale x 1 x float> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv1f32_nxv1f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,mf2,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 1 x float> @llvm.riscv.vfrsqrte7.nxv1f32( + <vscale x 1 x float> %0, + i64 %1) + + ret <vscale x 1 x float> %a +} + +declare <vscale x 1 x float> @llvm.riscv.vfrsqrte7.mask.nxv1f32( + <vscale x 1 x float>, + <vscale x 1 x float>, + <vscale x 1 x i1>, + i64); + +define <vscale x 1 x float> @intrinsic_vfrsqrte7_mask_v_nxv1f32_nxv1f32(<vscale x 1 x i1> %0, <vscale x 1 x float> %1, <vscale x 1 x float> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv1f32_nxv1f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,mf2,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v9, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 1 x float> @llvm.riscv.vfrsqrte7.mask.nxv1f32( + <vscale x 1 x float> %1, + <vscale x 1 x float> %2, + <vscale x 1 x i1> %0, + i64 %3) + + ret <vscale x 1 x float> %a +} + +declare <vscale x 2 x float> @llvm.riscv.vfrsqrte7.nxv2f32( + <vscale x 2 x float>, + i64); + +define <vscale x 2 x float> @intrinsic_vfrsqrte7_v_nxv2f32_nxv2f32(<vscale x 2 x float> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv2f32_nxv2f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m1,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 2 x float> @llvm.riscv.vfrsqrte7.nxv2f32( + <vscale x 2 x float> %0, + i64 %1) + + ret <vscale x 2 x float> %a +} + +declare <vscale x 2 x float> @llvm.riscv.vfrsqrte7.mask.nxv2f32( + <vscale x 2 x float>, + <vscale x 2 x float>, + <vscale x 2 x i1>, + i64); + +define <vscale x 2 x float> @intrinsic_vfrsqrte7_mask_v_nxv2f32_nxv2f32(<vscale x 2 x i1> %0, <vscale x 2 x float> %1, <vscale x 2 x float> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv2f32_nxv2f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m1,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v9, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 2 x float> @llvm.riscv.vfrsqrte7.mask.nxv2f32( + <vscale x 2 x float> %1, + <vscale x 2 x float> %2, + <vscale x 2 x i1> %0, + i64 %3) + + ret <vscale x 2 x float> %a +} + +declare <vscale x 4 x float> @llvm.riscv.vfrsqrte7.nxv4f32( + <vscale x 4 x float>, + i64); + +define <vscale x 4 x float> @intrinsic_vfrsqrte7_v_nxv4f32_nxv4f32(<vscale x 4 x float> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv4f32_nxv4f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m2,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 4 x float> @llvm.riscv.vfrsqrte7.nxv4f32( + <vscale x 4 x float> %0, + i64 %1) + + ret <vscale x 4 x float> %a +} + +declare <vscale x 4 x float> @llvm.riscv.vfrsqrte7.mask.nxv4f32( + <vscale x 4 x float>, + <vscale x 4 x float>, + <vscale x 4 x i1>, + i64); + +define <vscale x 4 x float> @intrinsic_vfrsqrte7_mask_v_nxv4f32_nxv4f32(<vscale x 4 x i1> %0, <vscale x 4 x float> %1, <vscale x 4 x float> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv4f32_nxv4f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m2,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v10, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 4 x float> @llvm.riscv.vfrsqrte7.mask.nxv4f32( + <vscale x 4 x float> %1, + <vscale x 4 x float> %2, + <vscale x 4 x i1> %0, + i64 %3) + + ret <vscale x 4 x float> %a +} + +declare <vscale x 8 x float> @llvm.riscv.vfrsqrte7.nxv8f32( + <vscale x 8 x float>, + i64); + +define <vscale x 8 x float> @intrinsic_vfrsqrte7_v_nxv8f32_nxv8f32(<vscale x 8 x float> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv8f32_nxv8f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m4,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 8 x float> @llvm.riscv.vfrsqrte7.nxv8f32( + <vscale x 8 x float> %0, + i64 %1) + + ret <vscale x 8 x float> %a +} + +declare <vscale x 8 x float> @llvm.riscv.vfrsqrte7.mask.nxv8f32( + <vscale x 8 x float>, + <vscale x 8 x float>, + <vscale x 8 x i1>, + i64); + +define <vscale x 8 x float> @intrinsic_vfrsqrte7_mask_v_nxv8f32_nxv8f32(<vscale x 8 x i1> %0, <vscale x 8 x float> %1, <vscale x 8 x float> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv8f32_nxv8f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m4,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v12, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 8 x float> @llvm.riscv.vfrsqrte7.mask.nxv8f32( + <vscale x 8 x float> %1, + <vscale x 8 x float> %2, + <vscale x 8 x i1> %0, + i64 %3) + + ret <vscale x 8 x float> %a +} + +declare <vscale x 16 x float> @llvm.riscv.vfrsqrte7.nxv16f32( + <vscale x 16 x float>, + i64); + +define <vscale x 16 x float> @intrinsic_vfrsqrte7_v_nxv16f32_nxv16f32(<vscale x 16 x float> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv16f32_nxv16f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m8,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 16 x float> @llvm.riscv.vfrsqrte7.nxv16f32( + <vscale x 16 x float> %0, + i64 %1) + + ret <vscale x 16 x float> %a +} + +declare <vscale x 16 x float> @llvm.riscv.vfrsqrte7.mask.nxv16f32( + <vscale x 16 x float>, + <vscale x 16 x float>, + <vscale x 16 x i1>, + i64); + +define <vscale x 16 x float> @intrinsic_vfrsqrte7_mask_v_nxv16f32_nxv16f32(<vscale x 16 x i1> %0, <vscale x 16 x float> %1, <vscale x 16 x float> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv16f32_nxv16f32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e32,m8,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v16, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 16 x float> @llvm.riscv.vfrsqrte7.mask.nxv16f32( + <vscale x 16 x float> %1, + <vscale x 16 x float> %2, + <vscale x 16 x i1> %0, + i64 %3) + + ret <vscale x 16 x float> %a +} + +declare <vscale x 1 x double> @llvm.riscv.vfrsqrte7.nxv1f64( + <vscale x 1 x double>, + i64); + +define <vscale x 1 x double> @intrinsic_vfrsqrte7_v_nxv1f64_nxv1f64(<vscale x 1 x double> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv1f64_nxv1f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m1,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 1 x double> @llvm.riscv.vfrsqrte7.nxv1f64( + <vscale x 1 x double> %0, + i64 %1) + + ret <vscale x 1 x double> %a +} + +declare <vscale x 1 x double> @llvm.riscv.vfrsqrte7.mask.nxv1f64( + <vscale x 1 x double>, + <vscale x 1 x double>, + <vscale x 1 x i1>, + i64); + +define <vscale x 1 x double> @intrinsic_vfrsqrte7_mask_v_nxv1f64_nxv1f64(<vscale x 1 x i1> %0, <vscale x 1 x double> %1, <vscale x 1 x double> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv1f64_nxv1f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m1,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v9, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 1 x double> @llvm.riscv.vfrsqrte7.mask.nxv1f64( + <vscale x 1 x double> %1, + <vscale x 1 x double> %2, + <vscale x 1 x i1> %0, + i64 %3) + + ret <vscale x 1 x double> %a +} + +declare <vscale x 2 x double> @llvm.riscv.vfrsqrte7.nxv2f64( + <vscale x 2 x double>, + i64); + +define <vscale x 2 x double> @intrinsic_vfrsqrte7_v_nxv2f64_nxv2f64(<vscale x 2 x double> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv2f64_nxv2f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m2,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 2 x double> @llvm.riscv.vfrsqrte7.nxv2f64( + <vscale x 2 x double> %0, + i64 %1) + + ret <vscale x 2 x double> %a +} + +declare <vscale x 2 x double> @llvm.riscv.vfrsqrte7.mask.nxv2f64( + <vscale x 2 x double>, + <vscale x 2 x double>, + <vscale x 2 x i1>, + i64); + +define <vscale x 2 x double> @intrinsic_vfrsqrte7_mask_v_nxv2f64_nxv2f64(<vscale x 2 x i1> %0, <vscale x 2 x double> %1, <vscale x 2 x double> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv2f64_nxv2f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m2,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v10, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 2 x double> @llvm.riscv.vfrsqrte7.mask.nxv2f64( + <vscale x 2 x double> %1, + <vscale x 2 x double> %2, + <vscale x 2 x i1> %0, + i64 %3) + + ret <vscale x 2 x double> %a +} + +declare <vscale x 4 x double> @llvm.riscv.vfrsqrte7.nxv4f64( + <vscale x 4 x double>, + i64); + +define <vscale x 4 x double> @intrinsic_vfrsqrte7_v_nxv4f64_nxv4f64(<vscale x 4 x double> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv4f64_nxv4f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m4,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 4 x double> @llvm.riscv.vfrsqrte7.nxv4f64( + <vscale x 4 x double> %0, + i64 %1) + + ret <vscale x 4 x double> %a +} + +declare <vscale x 4 x double> @llvm.riscv.vfrsqrte7.mask.nxv4f64( + <vscale x 4 x double>, + <vscale x 4 x double>, + <vscale x 4 x i1>, + i64); + +define <vscale x 4 x double> @intrinsic_vfrsqrte7_mask_v_nxv4f64_nxv4f64(<vscale x 4 x i1> %0, <vscale x 4 x double> %1, <vscale x 4 x double> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv4f64_nxv4f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m4,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v12, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 4 x double> @llvm.riscv.vfrsqrte7.mask.nxv4f64( + <vscale x 4 x double> %1, + <vscale x 4 x double> %2, + <vscale x 4 x i1> %0, + i64 %3) + + ret <vscale x 4 x double> %a +} + +declare <vscale x 8 x double> @llvm.riscv.vfrsqrte7.nxv8f64( + <vscale x 8 x double>, + i64); + +define <vscale x 8 x double> @intrinsic_vfrsqrte7_v_nxv8f64_nxv8f64(<vscale x 8 x double> %0, i64 %1) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_v_nxv8f64_nxv8f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m8,ta,mu +; CHECK-NEXT: vfrsqrte7.v v8, v8 +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 8 x double> @llvm.riscv.vfrsqrte7.nxv8f64( + <vscale x 8 x double> %0, + i64 %1) + + ret <vscale x 8 x double> %a +} + +declare <vscale x 8 x double> @llvm.riscv.vfrsqrte7.mask.nxv8f64( + <vscale x 8 x double>, + <vscale x 8 x double>, + <vscale x 8 x i1>, + i64); + +define <vscale x 8 x double> @intrinsic_vfrsqrte7_mask_v_nxv8f64_nxv8f64(<vscale x 8 x i1> %0, <vscale x 8 x double> %1, <vscale x 8 x double> %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vfrsqrte7_mask_v_nxv8f64_nxv8f64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a0, a0, e64,m8,tu,mu +; CHECK-NEXT: vfrsqrte7.v v8, v16, v0.t +; CHECK-NEXT: jalr zero, 0(ra) +entry: + %a = call <vscale x 8 x double> @llvm.riscv.vfrsqrte7.mask.nxv8f64( + <vscale x 8 x double> %1, + <vscale x 8 x double> %2, + <vscale x 8 x i1> %0, + i64 %3) + + ret <vscale x 8 x double> %a +} _______________________________________________ llvm-branch-commits mailing list llvm-branch-commits@lists.llvm.org https://lists.llvm.org/cgi-bin/mailman/listinfo/llvm-branch-commits