On Thu, 2021-06-17 at 10:19 -0500, Bill Schmidt via Gcc-patches wrote:
> 2021-06-17 Bill Schmidt <[email protected]>
>
Hi,
> gcc/
> * config/rs6000/rs6000-call.c (rs6000_expand_new_builtin): New
> forward decl.
> (rs6000_invalid_new_builtin): New stub function.
> (rs6000_expand_builtin): Call rs6000_expand_new_builtin.
> (rs6000_expand_ldst_mask): New stub function.
> (new_cpu_expand_builtin): Likewise.
> (elemrev_icode): Likewise.
> (ldv_expand_builtin): Likewise.
> (lxvrse_expand_builtin): Likewise.
> (lxvrze_expand_builtin): Likewise.
> (stv_expand_builtin): Likewise.
> (new_mma_expand_builtin): Likewise.
> (new_htm_expand_builtin): Likewise.
> (rs6000_expand_new_builtin): New function.
> ---
> gcc/config/rs6000/rs6000-call.c | 526 ++++++++++++++++++++++++++++++++
> 1 file changed, 526 insertions(+)
>
> diff --git a/gcc/config/rs6000/rs6000-call.c b/gcc/config/rs6000/rs6000-call.c
> index 52df3d165e1..8693836cd5a 100644
> --- a/gcc/config/rs6000/rs6000-call.c
> +++ b/gcc/config/rs6000/rs6000-call.c
> @@ -190,6 +190,7 @@ static tree builtin_function_type (machine_mode,
> machine_mode,
> static void rs6000_common_init_builtins (void);
> static void htm_init_builtins (void);
> static void mma_init_builtins (void);
> +static rtx rs6000_expand_new_builtin (tree, rtx, rtx, machine_mode, int);
> static bool rs6000_gimple_fold_new_builtin (gimple_stmt_iterator *gsi);
Ok if the forward declaration is necessary. I think we still have a
todo to rearrange and eliminate some of those, IMO not worth holding up this
patch series for that.
>
>
> @@ -11664,6 +11665,14 @@ rs6000_invalid_builtin (enum rs6000_builtins fncode)
> error ("%qs is not supported with the current options", name);
> }
>
> +/* Raise an error message for a builtin function that is called without the
> + appropriate target options being set. */
> +
> +static void
> +rs6000_invalid_new_builtin (enum rs6000_gen_builtins fncode)
> +{
> +}
> +
> /* Target hook for early folding of built-ins, shamelessly stolen
> from ia64.c. */
>
> @@ -14193,6 +14202,9 @@ rs6000_expand_builtin (tree exp, rtx target, rtx
> subtarget ATTRIBUTE_UNUSED,
> machine_mode mode ATTRIBUTE_UNUSED,
> int ignore ATTRIBUTE_UNUSED)
> {
> + if (new_builtins_are_live)
> + return rs6000_expand_new_builtin (exp, target, subtarget, mode, ignore);
> +
> tree fndecl = TREE_OPERAND (CALL_EXPR_FN (exp), 0);
> enum rs6000_builtins fcode
> = (enum rs6000_builtins) DECL_MD_FUNCTION_CODE (fndecl);
> @@ -14485,6 +14497,520 @@ rs6000_expand_builtin (tree exp, rtx target, rtx
> subtarget ATTRIBUTE_UNUSED,
> gcc_unreachable ();
> }
>
> +/* Expand ALTIVEC_BUILTIN_MASK_FOR_LOAD. */
> +rtx
> +rs6000_expand_ldst_mask (rtx target, tree arg0)
> + {
> + return target;
> + }
> +
> +/* Expand the CPU builtin in FCODE and store the result in TARGET. */
> +static rtx
> +new_cpu_expand_builtin (enum rs6000_gen_builtins fcode,
> + tree exp ATTRIBUTE_UNUSED, rtx target)
> +{
> + return target;
> +}
> +
> +static insn_code
> +elemrev_icode (rs6000_gen_builtins fcode)
> +{
> + return (insn_code) 0;
> +}
> +
> +static rtx
> +ldv_expand_builtin (rtx target, insn_code icode, rtx *op, machine_mode tmode)
> +{
> + return target;
> +}
> +
> +static rtx
> +lxvrse_expand_builtin (rtx target, insn_code icode, rtx *op,
> + machine_mode tmode, machine_mode smode)
> +{
> + return target;
> +}
> +
> +static rtx
> +lxvrze_expand_builtin (rtx target, insn_code icode, rtx *op,
> + machine_mode tmode, machine_mode smode)
> +{
> + return target;
> +}
> +
> +static rtx
> +stv_expand_builtin (insn_code icode, rtx *op,
> + machine_mode tmode, machine_mode smode)
> +{
> + return NULL_RTX;
> +}
> +
> +/* Expand the MMA built-in in EXP. */
> +static rtx
> +new_mma_expand_builtin (tree exp, rtx target, insn_code icode,
> + rs6000_gen_builtins fcode)
> +{
> + return target;
> +}
> +
> +/* Expand the HTM builtin in EXP and store the result in TARGET. */
> +static rtx
> +new_htm_expand_builtin (bifdata *bifaddr, rs6000_gen_builtins fcode,
> + tree exp, rtx target)
> +{
> + return const0_rtx;
> +}
> +
> +/* Expand an expression EXP that calls a built-in function,
> + with result going to TARGET if that's convenient
> + (and in mode MODE if that's convenient).
> + SUBTARGET may be used as the target for computing one of EXP's operands.
> + IGNORE is nonzero if the value is to be ignored.
> + Use the new builtin infrastructure. */
> +static rtx
> +rs6000_expand_new_builtin (tree exp, rtx target,
> + rtx subtarget ATTRIBUTE_UNUSED,
> + machine_mode ignore_mode ATTRIBUTE_UNUSED,
> + int ignore ATTRIBUTE_UNUSED)
> +{
> + tree fndecl = TREE_OPERAND (CALL_EXPR_FN (exp), 0);
> + enum rs6000_gen_builtins fcode
> + = (enum rs6000_gen_builtins) DECL_MD_FUNCTION_CODE (fndecl);
> + size_t uns_fcode = (size_t)fcode;
> + enum insn_code icode = rs6000_builtin_info_x[uns_fcode].icode;
> +
> + /* We have two different modes (KFmode, TFmode) that are the IEEE 128-bit
> + floating point type, depending on whether long double is the IBM
> extended
> + double (KFmode) or long double is IEEE 128-bit (TFmode). It is simpler
> if
> + we only define one variant of the built-in function, and switch the code
> + when defining it, rather than defining two built-ins and using the
> + overload table in rs6000-c.c to switch between the two. If we don't
> have
> + the proper assembler, don't do this switch because CODE_FOR_*kf* and
> + CODE_FOR_*tf* will be CODE_FOR_nothing. */
ok
> + if (FLOAT128_IEEE_P (TFmode))
> + switch (icode)
> + {
> + default:
> + break;
> +
> + case CODE_FOR_sqrtkf2_odd: icode = CODE_FOR_sqrttf2_odd; break;
> + case CODE_FOR_trunckfdf2_odd: icode = CODE_FOR_trunctfdf2_odd; break;
> + case CODE_FOR_addkf3_odd: icode = CODE_FOR_addtf3_odd;
> break;
> + case CODE_FOR_subkf3_odd: icode = CODE_FOR_subtf3_odd;
> break;
> + case CODE_FOR_mulkf3_odd: icode = CODE_FOR_multf3_odd;
> break;
> + case CODE_FOR_divkf3_odd: icode = CODE_FOR_divtf3_odd;
> break;
> + case CODE_FOR_fmakf4_odd: icode = CODE_FOR_fmatf4_odd;
> break;
> + case CODE_FOR_xsxexpqp_kf: icode = CODE_FOR_xsxexpqp_tf; break;
> + case CODE_FOR_xsxsigqp_kf: icode = CODE_FOR_xsxsigqp_tf; break;
> + case CODE_FOR_xststdcnegqp_kf: icode = CODE_FOR_xststdcnegqp_tf; break;
> + case CODE_FOR_xsiexpqp_kf: icode = CODE_FOR_xsiexpqp_tf; break;
> + case CODE_FOR_xsiexpqpf_kf: icode = CODE_FOR_xsiexpqpf_tf; break;
> + case CODE_FOR_xststdcqp_kf: icode = CODE_FOR_xststdcqp_tf; break;
> +
> + case CODE_FOR_xscmpexpqp_eq_kf:
> + icode = CODE_FOR_xscmpexpqp_eq_tf;
> + break;
> +
> + case CODE_FOR_xscmpexpqp_lt_kf:
> + icode = CODE_FOR_xscmpexpqp_lt_tf;
> + break;
> +
> + case CODE_FOR_xscmpexpqp_gt_kf:
> + icode = CODE_FOR_xscmpexpqp_gt_tf;
> + break;
> +
> + case CODE_FOR_xscmpexpqp_unordered_kf:
> + icode = CODE_FOR_xscmpexpqp_unordered_tf;
> + break;
> + }
> +
> + bifdata *bifaddr = &rs6000_builtin_info_x[uns_fcode];
> +
> + /* In case of "#pragma target" changes, we initialize all builtins
> + but check for actual availability during expand time. For
> + invalid builtins, generate a normal call. */
> + switch (bifaddr->enable)
> + {
> + default:
> + gcc_unreachable ();
> + case ENB_ALWAYS:
> + break;
> + case ENB_P5:
> + if (!TARGET_POPCNTB)
> + {
> + rs6000_invalid_new_builtin (fcode);
> + return expand_call (exp, target, ignore);
> + }
> + break;
> + case ENB_P6:
> + if (!TARGET_CMPB)
> + {
> + rs6000_invalid_new_builtin (fcode);
> + return expand_call (exp, target, ignore);
> + }
> + break;
> + case ENB_ALTIVEC:
> + if (!TARGET_ALTIVEC)
> + {
> + rs6000_invalid_new_builtin (fcode);
> + return expand_call (exp, target, ignore);
> + }
> + break;
> + case ENB_CELL:
> + if (!TARGET_ALTIVEC || rs6000_cpu != PROCESSOR_CELL)
> + {
> + rs6000_invalid_new_builtin (fcode);
> + return expand_call (exp, target, ignore);
> + }
> + break;
> + case ENB_VSX:
> + if (!TARGET_VSX)
> + {
> + rs6000_invalid_new_builtin (fcode);
> + return expand_call (exp, target, ignore);
> + }
> + break;
> + case ENB_P7:
> + if (!TARGET_POPCNTD)
> + {
> + rs6000_invalid_new_builtin (fcode);
> + return expand_call (exp, target, ignore);
> + }
> + break;
> + case ENB_P7_64:
> + if (!TARGET_POPCNTD || !TARGET_POWERPC64)
> + {
> + rs6000_invalid_new_builtin (fcode);
> + return expand_call (exp, target, ignore);
> + }
> + break;
> + case ENB_P8:
> + if (!TARGET_DIRECT_MOVE)
> + {
> + rs6000_invalid_new_builtin (fcode);
> + return expand_call (exp, target, ignore);
> + }
> + break;
> + case ENB_P8V:
> + if (!TARGET_P8_VECTOR)
> + {
> + rs6000_invalid_new_builtin (fcode);
> + return expand_call (exp, target, ignore);
> + }
> + break;
> + case ENB_P9:
> + if (!TARGET_MODULO)
> + {
> + rs6000_invalid_new_builtin (fcode);
> + return expand_call (exp, target, ignore);
> + }
> + break;
> + case ENB_P9_64:
> + if (!TARGET_MODULO || !TARGET_POWERPC64)
> + {
> + rs6000_invalid_new_builtin (fcode);
> + return expand_call (exp, target, ignore);
> + }
> + break;
> + case ENB_P9V:
> + if (!TARGET_P9_VECTOR)
> + {
> + rs6000_invalid_new_builtin (fcode);
> + return expand_call (exp, target, ignore);
> + }
> + break;
> + case ENB_IEEE128_HW:
> + if (!TARGET_FLOAT128_HW)
> + {
> + rs6000_invalid_new_builtin (fcode);
> + return expand_call (exp, target, ignore);
> + }
> + break;
> + case ENB_DFP:
> + if (!TARGET_DFP)
> + {
> + rs6000_invalid_new_builtin (fcode);
> + return expand_call (exp, target, ignore);
> + }
> + break;
> + case ENB_CRYPTO:
> + if (!TARGET_CRYPTO)
> + {
> + rs6000_invalid_new_builtin (fcode);
> + return expand_call (exp, target, ignore);
> + }
> + break;
> + case ENB_HTM:
> + if (!TARGET_HTM)
> + {
> + rs6000_invalid_new_builtin (fcode);
> + return expand_call (exp, target, ignore);
> + }
> + break;
> + case ENB_P10:
> + if (!TARGET_POWER10)
> + {
> + rs6000_invalid_new_builtin (fcode);
> + return expand_call (exp, target, ignore);
> + }
> + break;
> + case ENB_P10_64:
> + if (!TARGET_POWER10 || !TARGET_POWERPC64)
> + {
> + rs6000_invalid_new_builtin (fcode);
> + return expand_call (exp, target, ignore);
> + }
> + break;
> + case ENB_MMA:
> + if (!TARGET_MMA)
> + {
> + rs6000_invalid_new_builtin (fcode);
> + return expand_call (exp, target, ignore);
> + }
> + break;
> + };
> +
> + if (bif_is_nosoft (*bifaddr)
> + && rs6000_isa_flags & OPTION_MASK_SOFT_FLOAT)
> + {
> + error ("%<%s%> not supported with %<-msoft-float%>",
> + bifaddr->bifname);
> + return const0_rtx;
> + }
> +
> + if (bif_is_no32bit (*bifaddr) && TARGET_32BIT)
> + fatal_error (input_location,
> + "%<%s%> is not supported in 32-bit mode",
> + bifaddr->bifname);
ok
> +
> + if (bif_is_cpu (*bifaddr))
> + return new_cpu_expand_builtin (fcode, exp, target);
> +
> + if (bif_is_init (*bifaddr))
> + return altivec_expand_vec_init_builtin (TREE_TYPE (exp), exp, target);
> +
> + if (bif_is_set (*bifaddr))
> + return altivec_expand_vec_set_builtin (exp);
> +
> + if (bif_is_extract (*bifaddr))
> + return altivec_expand_vec_ext_builtin (exp, target);
> +
> + if (bif_is_predicate (*bifaddr))
> + return altivec_expand_predicate_builtin (icode, exp, target);
> +
> + if (bif_is_htm (*bifaddr))
> + return new_htm_expand_builtin (bifaddr, fcode, exp, target);
> +
> + rtx pat;
> + const int MAX_BUILTIN_ARGS = 6;
> + tree arg[MAX_BUILTIN_ARGS];
> + rtx op[MAX_BUILTIN_ARGS];
> + machine_mode mode[MAX_BUILTIN_ARGS + 1];
> + bool void_func = TREE_TYPE (TREE_TYPE (fndecl)) == void_type_node;
> + int k;
> +
> + int nargs = bifaddr->nargs;
> + gcc_assert (nargs <= MAX_BUILTIN_ARGS);
> +
> + if (void_func)
> + k = 0;
> + else
> + {
> + k = 1;
> + mode[0] = insn_data[icode].operand[0].mode;
> + }
> +
> + for (int i = 0; i < nargs; i++)
> + {
> + arg[i] = CALL_EXPR_ARG (exp, i);
> + if (arg[i] == error_mark_node)
> + return const0_rtx;
> + STRIP_NOPS (arg[i]);
> + op[i] = expand_normal (arg[i]);
> + /* We have a couple of pesky patterns that don't specify the mode...
> */
> + if (!insn_data[icode].operand[i+k].mode)
> + mode[i+k] = TARGET_64BIT ? Pmode : SImode;
> + else
> + mode[i+k] = insn_data[icode].operand[i+k].mode;
> + }
> +
> + /* Check for restricted constant arguments. */
> + for (int i = 0; i < 2; i++)
> + {
> + switch (bifaddr->restr[i])
> + {
> + default:
> + case RES_NONE:
> + break;
> + case RES_BITS:
> + {
> + size_t mask = (1 << bifaddr->restr_val1[i]) - 1;
> + tree restr_arg = arg[bifaddr->restr_opnd[i] - 1];
> + STRIP_NOPS (restr_arg);
> + if (TREE_CODE (restr_arg) != INTEGER_CST
> + || TREE_INT_CST_LOW (restr_arg) & ~mask)
> + {
> + error ("argument %d must be a %d-bit unsigned literal",
> + bifaddr->restr_opnd[i], bifaddr->restr_val1[i]);
> + return CONST0_RTX (mode[0]);
> + }
> + break;
> + }
ok
> + case RES_RANGE:
> + {
> + tree restr_arg = arg[bifaddr->restr_opnd[i] - 1];
> + STRIP_NOPS (restr_arg);
> + if (TREE_CODE (restr_arg) != INTEGER_CST
> + || !IN_RANGE (tree_to_shwi (restr_arg),
> + bifaddr->restr_val1[i],
> + bifaddr->restr_val2[i]))
> + {
> + error ("argument %d must be a literal between %d and %d,"
> + " inclusive",
> + bifaddr->restr_opnd[i], bifaddr->restr_val1[i],
> + bifaddr->restr_val2[i]);
> + return CONST0_RTX (mode[0]);
> + }
> + break;
> + }
ok
> + case RES_VAR_RANGE:
> + {
> + tree restr_arg = arg[bifaddr->restr_opnd[i] - 1];
> + STRIP_NOPS (restr_arg);
> + if (TREE_CODE (restr_arg) == INTEGER_CST
> + && !IN_RANGE (tree_to_shwi (restr_arg),
> + bifaddr->restr_val1[i],
> + bifaddr->restr_val2[i]))
> + {
> + error ("argument %d must be a variable or a literal "
> + "between %d and %d, inclusive",
> + bifaddr->restr_opnd[i], bifaddr->restr_val1[i],
> + bifaddr->restr_val2[i]);
> + return CONST0_RTX (mode[0]);
> + }
> + break;
> + }
ok
> + case RES_VALUES:
> + {
> + tree restr_arg = arg[bifaddr->restr_opnd[i] - 1];
> + STRIP_NOPS (restr_arg);
> + if (TREE_CODE (restr_arg) != INTEGER_CST
> + || (tree_to_shwi (restr_arg) != bifaddr->restr_val1[i]
> + && tree_to_shwi (restr_arg) != bifaddr->restr_val2[i]))
> + {
> + error ("argument %d must be either a literal %d or a "
> + "literal %d",
> + bifaddr->restr_opnd[i], bifaddr->restr_val1[i],
> + bifaddr->restr_val2[i]);
> + return CONST0_RTX (mode[0]);
> + }
> + break;
> + }
ok
> + }
> + }
> +
> + if (bif_is_ldstmask (*bifaddr))
> + return rs6000_expand_ldst_mask (target, arg[0]);
> +
> + if (bif_is_stvec (*bifaddr))
> + {
> + if (bif_is_reve (*bifaddr))
> + icode = elemrev_icode (fcode);
> + return stv_expand_builtin (icode, op, mode[0], mode[1]);
> + }
> +
> + if (bif_is_ldvec (*bifaddr))
> + {
> + if (bif_is_reve (*bifaddr))
> + icode = elemrev_icode (fcode);
> + return ldv_expand_builtin (target, icode, op, mode[0]);
> + }
> +
> + if (bif_is_lxvrse (*bifaddr))
> + return lxvrse_expand_builtin (target, icode, op, mode[0], mode[1]);
> +
> + if (bif_is_lxvrze (*bifaddr))
> + return lxvrze_expand_builtin (target, icode, op, mode[0], mode[1]);
> +
> + if (bif_is_mma (*bifaddr))
> + return new_mma_expand_builtin (exp, target, icode, fcode);
> +
> + if (fcode == RS6000_BIF_PACK_IF
> + && TARGET_LONG_DOUBLE_128 && !TARGET_IEEEQUAD)
> + {
> + icode = CODE_FOR_packtf;
> + fcode = RS6000_BIF_PACK_TF;
> + uns_fcode = (size_t)fcode;
> + }
> + else if (fcode == RS6000_BIF_UNPACK_IF
> + && TARGET_LONG_DOUBLE_128 && !TARGET_IEEEQUAD)
> + {
> + icode = CODE_FOR_unpacktf;
> + fcode = RS6000_BIF_UNPACK_TF;
> + uns_fcode = (size_t)fcode;
> + }
> +
> + if (TREE_TYPE (TREE_TYPE (fndecl)) == void_type_node)
> + target = NULL_RTX;
> + else if (target == 0
> + || GET_MODE (target) != mode[0]
> + || !(*insn_data[icode].operand[0].predicate) (target, mode[0]))
> + target = gen_reg_rtx (mode[0]);
> +
> + for (int i = 0; i < nargs; i++)
> + if (! (*insn_data[icode].operand[i+k].predicate) (op[i], mode[i+k]))
> + op[i] = copy_to_mode_reg (mode[i+k], op[i]);
> +
> + switch (nargs)
> + {
> + default:
> + gcc_assert (MAX_BUILTIN_ARGS == 6);
> + gcc_unreachable ();
> + case 0:
> + pat = (void_func
> + ? GEN_FCN (icode) ()
> + : GEN_FCN (icode) (target));
> + break;
> + case 1:
> + pat = (void_func
> + ? GEN_FCN (icode) (op[0])
> + : GEN_FCN (icode) (target, op[0]));
> + break;
> + case 2:
> + pat = (void_func
> + ? GEN_FCN (icode) (op[0], op[1])
> + : GEN_FCN (icode) (target, op[0], op[1]));
> + break;
> + case 3:
> + pat = (void_func
> + ? GEN_FCN (icode) (op[0], op[1], op[2])
> + : GEN_FCN (icode) (target, op[0], op[1], op[2]));
> + break;
> + case 4:
> + pat = (void_func
> + ? GEN_FCN (icode) (op[0], op[1], op[2], op[3])
> + : GEN_FCN (icode) (target, op[0], op[1], op[2], op[3]));
> + break;
> + case 5:
> + pat = (void_func
> + ? GEN_FCN (icode) (op[0], op[1], op[2], op[3], op[4])
> + : GEN_FCN (icode) (target, op[0], op[1], op[2], op[3], op[4]));
> + break;
> + case 6:
> + pat = (void_func
> + ? GEN_FCN (icode) (op[0], op[1], op[2], op[3], op[4], op[5])
> + : GEN_FCN (icode) (target, op[0], op[1],
> + op[2], op[3], op[4], op[5]));
> + break;
> + }
> +
> + if (!pat)
> + return 0;
> + emit_insn (pat);
> +
> + return target;
> +}
ok,
lgtm
thanks
-Will
> +
> /* Create a builtin vector type with a name. Taking care not to give
> the canonical type a name. */
>