The following makes us always use VMAT_STRIDED_SLP for negative
stride multi-element accesses. That handles falling back to
single element accesses transparently.
Bootstrap and regtest running on x86_64-unknown-linux-gnu.
* tree-vect-stmts.cc (get_load_store_type): Use VMAT_STRIDED_SLP
for negative stride accesses when VMAT_CONTIGUOUS_REVERSE
isn't applicable.
---
gcc/tree-vect-stmts.cc | 14 +++++---------
1 file changed, 5 insertions(+), 9 deletions(-)
diff --git a/gcc/tree-vect-stmts.cc b/gcc/tree-vect-stmts.cc
index 6168ac07b40..76c07b406b0 100644
--- a/gcc/tree-vect-stmts.cc
+++ b/gcc/tree-vect-stmts.cc
@@ -2026,6 +2026,7 @@ get_load_store_type (vec_info *vinfo, stmt_vec_info
stmt_info,
without permutation. */
if (! SLP_TREE_LOAD_PERMUTATION (slp_node).exists ())
first_dr_info = STMT_VINFO_DR_INFO (SLP_TREE_SCALAR_STMTS (slp_node)[0]);
+
if (STMT_VINFO_STRIDED_P (first_stmt_info))
/* Try to use consecutive accesses of as many elements as possible,
separated by the stride, until we have a complete vector.
@@ -2089,15 +2090,10 @@ get_load_store_type (vec_info *vinfo, stmt_vec_info
stmt_info,
(vinfo, stmt_info, vectype, vls_type, 1,
&neg_ldst_offset);
else
- {
- /* Try to use consecutive accesses of DR_GROUP_SIZE elements,
- separated by the stride, until we have a complete vector.
- Fall back to scalar accesses if that isn't possible. */
- if (multiple_p (nunits, group_size))
- *memory_access_type = VMAT_STRIDED_SLP;
- else
- *memory_access_type = VMAT_ELEMENTWISE;
- }
+ /* We can fall back to VMAT_STRIDED_SLP since that does
+ not care whether the stride between the group instances
+ is positive or negative. */
+ *memory_access_type = VMAT_STRIDED_SLP;
}
else if (cmp == 0 && loop_vinfo)
{
--
2.51.0