On Thu, 11 Jul 2019, Jan Hubicka wrote:

> Hi,
> this patch makes nonoverlapping_component_refs_since_match_p to accept
> paths with non-trivial MEM_REFs and TMRs assuming that they have same
> semantics.

Hmm.  We'll never get any TARGET_MEM_REFs wrapped with
handled-components so I wonder if it makes sense to handle it in
nonoverlapping_component_refs_since_match_p at all.

> Bootstrapped/regtested x86_64-linux, OK?
> 
> Honza
> 
>       * tree-ssa-alias.c (same_tmr_indexing_p): Break out from ...
>       (indirect_refs_may_alias_p): ... here.
>       (nonoverlapping_component_refs_since_match_p): Support also non-trivial
>       mem refs in the access paths.
> Index: testsuite/gcc.dg/tree-ssa/alias-access-path-9.c
> ===================================================================
> --- testsuite/gcc.dg/tree-ssa/alias-access-path-9.c   (nonexistent)
> +++ testsuite/gcc.dg/tree-ssa/alias-access-path-9.c   (working copy)
> @@ -0,0 +1,44 @@
> +/* { dg-do compile } */
> +/* { dg-options "-O2 -fdump-tree-fre1" } */
> +
> +/* This testcase tests nonoverlapping_component_refs_since_match_p in 
> presence
> +   of non-trivial mem-refs.  */
> +struct a {int a,b;};
> +struct b {struct a a[10];};
> +struct c {int c; struct b b;} c, *cptr;
> +
> +void
> +set_a(struct a *a, int p)
> +{
> +  a->a=p;
> +}
> +void
> +set_b(struct a *a, int p)
> +{
> +  a->b=p;
> +}
> +int
> +get_a(struct a *a)
> +{
> +  return a->a;
> +}
> +
> +int
> +test(int i, int j)
> +{
> +  struct b *bptr = &c.b;
> +  set_a (&bptr->a[i], 123);
> +  set_b (&bptr->a[j], 124);
> +  return get_a (&bptr->a[i]);
> +}
> +
> +int
> +test2(int i, int j)
> +{
> +  struct b *bptr = &cptr->b;
> +  set_a (&bptr->a[i], 125);
> +  set_b (&bptr->a[j], 126);
> +  return get_a (&bptr->a[i]);
> +}
> +/* { dg-final { scan-tree-dump-times "return 123" 1 "fre1"} } */
> +/* { dg-final { scan-tree-dump-times "return 125" 1 "fre1"} } */
> Index: tree-ssa-alias.c
> ===================================================================
> --- tree-ssa-alias.c  (revision 273322)
> +++ tree-ssa-alias.c  (working copy)
> @@ -1216,6 +1216,25 @@ nonoverlapping_component_refs_p_1 (const
>    return -1;
>  }
>  
> +/* Return if TARGET_MEM_REFS base1 and base2 have same offsets.  */
> +
> +static bool
> +same_tmr_indexing_p (tree base1, tree base2)
> +{
> +  return ((TMR_STEP (base1) == TMR_STEP (base2)
> +       || (TMR_STEP (base1) && TMR_STEP (base2)
> +           && operand_equal_p (TMR_STEP (base1),
> +                               TMR_STEP (base2), 0)))
> +       && (TMR_INDEX (base1) == TMR_INDEX (base2)
> +           || (TMR_INDEX (base1) && TMR_INDEX (base2)
> +               && operand_equal_p (TMR_INDEX (base1),
> +                                   TMR_INDEX (base2), 0)))
> +       && (TMR_INDEX2 (base1) == TMR_INDEX2 (base2)
> +           || (TMR_INDEX2 (base1) && TMR_INDEX2 (base2)
> +               && operand_equal_p (TMR_INDEX2 (base1),
> +                                   TMR_INDEX2 (base2), 0))));
> +}
> +
>  /* Try to disambiguate REF1 and REF2 under the assumption that MATCH1 and
>     MATCH2 either point to the same address or are disjoint.
>     MATCH1 and MATCH2 are assumed to be ref in the access path of REF1 and 
> REF2
> @@ -1265,20 +1284,6 @@ nonoverlapping_component_refs_since_matc
>          component_refs1.safe_push (ref1);
>        ref1 = TREE_OPERAND (ref1, 0);
>      }
> -  if (TREE_CODE (ref1) == MEM_REF && ref1 != match1)
> -    {
> -      if (!integer_zerop (TREE_OPERAND (ref1, 1)))
> -     {
> -       ++alias_stats.nonoverlapping_component_refs_since_match_p_may_alias;
> -       return -1;
> -     }
> -    }
> -  /* TODO: Handle TARGET_MEM_REF later.  */
> -  if (TREE_CODE (ref1) == TARGET_MEM_REF && ref1 != match1)
> -    {
> -      ++alias_stats.nonoverlapping_component_refs_since_match_p_may_alias;
> -      return -1;
> -    }
>  
>    /* Create the stack of handled components for REF2.  */
>    while (handled_component_p (ref2) && ref2 != match2)
> @@ -1290,15 +1295,39 @@ nonoverlapping_component_refs_since_matc
>          component_refs2.safe_push (ref2);
>        ref2 = TREE_OPERAND (ref2, 0);
>      }
> -  if (TREE_CODE (ref2) == MEM_REF && ref2 != match2)
> +
> +  bool mem_ref1 = TREE_CODE (ref1) == MEM_REF && ref1 != match1;
> +  bool mem_ref2 = TREE_CODE (ref2) == MEM_REF && ref2 != match2;
> +
> +  /* If only one of access path starts with MEM_REF check that offset is 0
> +     so the addresses stays the same after stripping it.
> +     TODO: In this case we may walk the other access path until we get same
> +     offset.
> +
> +     If both starts with MEM_REF, offset has to be same.  */
> +  if ((mem_ref1 && !mem_ref2 && !integer_zerop (TREE_OPERAND (ref1, 1)))
> +      || (mem_ref2 && !mem_ref1 && !integer_zerop (TREE_OPERAND (ref2, 1)))
> +      || (mem_ref1 && mem_ref2
> +       && !tree_int_cst_equal (TREE_OPERAND (ref1, 1),
> +                               TREE_OPERAND (ref2, 1))))
>      {
> -      if (!integer_zerop (TREE_OPERAND (ref2, 1)))
> -     {
> -       ++alias_stats.nonoverlapping_component_refs_since_match_p_may_alias;
> -       return -1;
> -     }
> +      ++alias_stats.nonoverlapping_component_refs_since_match_p_may_alias;
> +      return -1;
>      }
> -  if (TREE_CODE (ref2) == TARGET_MEM_REF && ref2 != match2)
> +
> +  bool target_mem_ref1 = TREE_CODE (ref1) == TARGET_MEM_REF && ref1 != 
> match1;
> +  bool target_mem_ref2 = TREE_CODE (ref2) == TARGET_MEM_REF && ref2 != 
> match2;
> +
> +  if ((target_mem_ref1 && !target_mem_ref2
> +       && (TMR_INDEX (ref1) || TMR_INDEX2 (ref1)
> +        || !integer_zerop (TREE_OPERAND (ref1, 1))))
> +      || (target_mem_ref2 && !target_mem_ref1
> +          && (TMR_INDEX (ref2) || TMR_INDEX2 (ref2)
> +           || !integer_zerop (TREE_OPERAND (ref2, 1))))
> +      || (target_mem_ref1 && target_mem_ref2
> +       && (!same_tmr_indexing_p (ref1, ref2)
> +           || !tree_int_cst_equal (TREE_OPERAND (ref1, 1),
> +                                   TREE_OPERAND (ref2, 1)))))
>      {
>        ++alias_stats.nonoverlapping_component_refs_since_match_p_may_alias;
>        return -1;
> @@ -1776,18 +1805,7 @@ indirect_refs_may_alias_p (tree ref1 ATT
>              || (!TMR_INDEX (base2) && !TMR_INDEX2 (base2))))
>         || (TREE_CODE (base1) == TARGET_MEM_REF
>             && TREE_CODE (base2) == TARGET_MEM_REF
> -           && (TMR_STEP (base1) == TMR_STEP (base2)
> -               || (TMR_STEP (base1) && TMR_STEP (base2)
> -                   && operand_equal_p (TMR_STEP (base1),
> -                                       TMR_STEP (base2), 0)))
> -           && (TMR_INDEX (base1) == TMR_INDEX (base2)
> -               || (TMR_INDEX (base1) && TMR_INDEX (base2)
> -                   && operand_equal_p (TMR_INDEX (base1),
> -                                       TMR_INDEX (base2), 0)))
> -           && (TMR_INDEX2 (base1) == TMR_INDEX2 (base2)
> -               || (TMR_INDEX2 (base1) && TMR_INDEX2 (base2)
> -                   && operand_equal_p (TMR_INDEX2 (base1),
> -                                       TMR_INDEX2 (base2), 0))))))
> +           && same_tmr_indexing_p (base1, base2))))
>      {
>        poly_offset_int moff1 = mem_ref_offset (base1) << LOG2_BITS_PER_UNIT;
>        poly_offset_int moff2 = mem_ref_offset (base2) << LOG2_BITS_PER_UNIT;
> 

-- 
Richard Biener <rguent...@suse.de>
SUSE Linux GmbH, Maxfeldstrasse 5, 90409 Nuernberg, Germany;
GF: Felix Imendörffer, Mary Higgins, Sri Rasiah; HRB 21284 (AG Nürnberg)

Reply via email to