The following changes the unsigned group_size argument to a poly_uint64
one to avoid too much special-casing in callers for VLA vectors when
passing down the effective maximum desirable vector size to vector
type query routines. The intent is to be able to pass down
the vectorization factor (times the SLP group size) eventually.
* tree-vectorizer.h (get_vectype_for_scalar_type,
get_mask_type_for_scalar_type, vect_get_vector_types_for_stmt):
Change group_size argument to poly_uint64 type.
(vect_get_mask_type_for_stmt): Remove prototype for no longer
existing function.
* tree-vect-stmts.cc (get_vectype_for_scalar_type): Change
group_size argument to poly_uint64.
(get_mask_type_for_scalar_type): Likewise.
(vect_get_vector_types_for_stmt): Likewise.
---
gcc/tree-vect-stmts.cc | 25 ++++++++++++++-----------
gcc/tree-vectorizer.h | 7 +++----
2 files changed, 17 insertions(+), 15 deletions(-)
diff --git a/gcc/tree-vect-stmts.cc b/gcc/tree-vect-stmts.cc
index 88401a2a00b..a5e26b746fb 100644
--- a/gcc/tree-vect-stmts.cc
+++ b/gcc/tree-vect-stmts.cc
@@ -13297,14 +13297,14 @@ get_related_vectype_for_scalar_type (machine_mode
prevailing_mode,
tree
get_vectype_for_scalar_type (vec_info *vinfo, tree scalar_type,
- unsigned int group_size)
+ poly_uint64 group_size)
{
/* For BB vectorization, we should always have a group size once we've
constructed the SLP tree; the only valid uses of zero GROUP_SIZEs
are tentative requests during things like early data reference
analysis and pattern recognition. */
if (is_a <bb_vec_info> (vinfo))
- gcc_assert (vinfo->slp_instances.is_empty () || group_size != 0);
+ gcc_assert (vinfo->slp_instances.is_empty () || known_ne (group_size, 0));
else
group_size = 0;
@@ -13320,9 +13320,11 @@ get_vectype_for_scalar_type (vec_info *vinfo, tree
scalar_type,
/* If the natural choice of vector type doesn't satisfy GROUP_SIZE,
try again with an explicit number of elements. */
+ uint64_t cst_group_size;
if (vectype
- && group_size
- && maybe_ge (TYPE_VECTOR_SUBPARTS (vectype), group_size))
+ && group_size.is_constant (&cst_group_size)
+ && cst_group_size != 0
+ && maybe_ge (TYPE_VECTOR_SUBPARTS (vectype), cst_group_size))
{
/* Start with the biggest number of units that fits within
GROUP_SIZE and halve it until we find a valid vector type.
@@ -13336,7 +13338,7 @@ get_vectype_for_scalar_type (vec_info *vinfo, tree
scalar_type,
even though the group is not a multiple of that vector size.
The BB vectorizer will then try to carve up the group into
smaller pieces. */
- unsigned int nunits = 1 << floor_log2 (group_size);
+ unsigned int nunits = 1 << floor_log2 (cst_group_size);
do
{
vectype = get_related_vectype_for_scalar_type (vinfo->vector_mode,
@@ -13372,7 +13374,7 @@ get_vectype_for_scalar_type (vec_info *vinfo, tree
scalar_type, slp_tree node)
tree
get_mask_type_for_scalar_type (vec_info *vinfo, tree scalar_type,
- unsigned int group_size)
+ poly_uint64 group_size)
{
tree vectype = get_vectype_for_scalar_type (vinfo, scalar_type, group_size);
@@ -14243,7 +14245,7 @@ opt_result
vect_get_vector_types_for_stmt (vec_info *vinfo, stmt_vec_info stmt_info,
tree *stmt_vectype_out,
tree *nunits_vectype_out,
- unsigned int group_size)
+ poly_uint64 group_size)
{
gimple *stmt = stmt_info->stmt;
@@ -14252,7 +14254,7 @@ vect_get_vector_types_for_stmt (vec_info *vinfo,
stmt_vec_info stmt_info,
are tentative requests during things like early data reference
analysis and pattern recognition. */
if (is_a <bb_vec_info> (vinfo))
- gcc_assert (vinfo->slp_instances.is_empty () || group_size != 0);
+ gcc_assert (vinfo->slp_instances.is_empty () || known_ne (group_size, 0));
else
group_size = 0;
@@ -14281,7 +14283,7 @@ vect_get_vector_types_for_stmt (vec_info *vinfo,
stmt_vec_info stmt_info,
tree vectype;
tree scalar_type = NULL_TREE;
- if (group_size == 0 && STMT_VINFO_VECTYPE (stmt_info))
+ if (known_eq (group_size, 0U) && STMT_VINFO_VECTYPE (stmt_info))
{
vectype = STMT_VINFO_VECTYPE (stmt_info);
if (dump_enabled_p ())
@@ -14310,10 +14312,11 @@ vect_get_vector_types_for_stmt (vec_info *vinfo,
stmt_vec_info stmt_info,
if (dump_enabled_p ())
{
- if (group_size)
+ if (known_ne (group_size, 0U))
dump_printf_loc (MSG_NOTE, vect_location,
"get vectype for scalar type (group size %d):"
- " %T\n", group_size, scalar_type);
+ " %T\n", (int)constant_lower_bound (group_size),
+ scalar_type);
else
dump_printf_loc (MSG_NOTE, vect_location,
"get vectype for scalar type: %T\n", scalar_type);
diff --git a/gcc/tree-vectorizer.h b/gcc/tree-vectorizer.h
index a2bab8676af..95240504d18 100644
--- a/gcc/tree-vectorizer.h
+++ b/gcc/tree-vectorizer.h
@@ -2204,9 +2204,9 @@ extern edge vec_init_loop_exit_info (class loop *);
/* In tree-vect-stmts.cc. */
extern tree get_related_vectype_for_scalar_type (machine_mode, tree,
poly_uint64 = 0);
-extern tree get_vectype_for_scalar_type (vec_info *, tree, unsigned int = 0);
+extern tree get_vectype_for_scalar_type (vec_info *, tree, poly_uint64 = 0);
extern tree get_vectype_for_scalar_type (vec_info *, tree, slp_tree);
-extern tree get_mask_type_for_scalar_type (vec_info *, tree, unsigned int = 0);
+extern tree get_mask_type_for_scalar_type (vec_info *, tree, poly_uint64 = 0);
extern tree get_mask_type_for_scalar_type (vec_info *, tree, slp_tree);
extern tree get_same_sized_vectype (tree, tree);
extern bool vect_chooses_same_modes_p (vec_info *, machine_mode);
@@ -2295,8 +2295,7 @@ extern tree vect_gen_while (gimple_seq *, tree, tree,
tree,
extern tree vect_gen_while_not (gimple_seq *, tree, tree, tree);
extern opt_result vect_get_vector_types_for_stmt (vec_info *,
stmt_vec_info, tree *,
- tree *, unsigned int = 0);
-extern opt_tree vect_get_mask_type_for_stmt (stmt_vec_info, unsigned int = 0);
+ tree *, poly_uint64 = 0);
/* In tree-vect-data-refs.cc. */
extern bool vect_can_force_dr_alignment_p (const_tree, poly_uint64);
--
2.35.3