apinski--- via Gcc-patches <gcc-patches@gcc.gnu.org> writes:
> From: Andrew Pinski <apin...@marvell.com>
>
> This patch adds simple folding of __builtin_aarch64_im_lane_boundsi where
> we are not going to error out. It fixes the problem by the removal
> of the function from the IR.
>
> OK? Bootstrapped and tested on aarch64-linux-gnu with no regressions.
>
> gcc/ChangeLog:
>
>       PR target/95969
>       * config/aarch64/aarch64-builtins.c (aarch64_fold_builtin_lane_check):
>       New function.
>       (aarch64_general_fold_builtin): Handle AARCH64_SIMD_BUILTIN_LANE_CHECK.
>       (aarch64_general_gimple_fold_builtin): Likewise.
>
> gcc/testsuite/ChangeLog:
>
>       PR target/95969
>       * gcc.target/aarch64/lane-bound-1.c: New test.
>       * gcc.target/aarch64/lane-bound-2.c: New test.

OK, thanks.  Sorry for the slow reply, was away last week.

Richard

> ---
>  gcc/config/aarch64/aarch64-builtins.c         | 35 +++++++++++++++++++
>  .../gcc.target/aarch64/lane-bound-1.c         | 14 ++++++++
>  .../gcc.target/aarch64/lane-bound-2.c         | 10 ++++++
>  3 files changed, 59 insertions(+)
>  create mode 100644 gcc/testsuite/gcc.target/aarch64/lane-bound-1.c
>  create mode 100644 gcc/testsuite/gcc.target/aarch64/lane-bound-2.c
>
> diff --git a/gcc/config/aarch64/aarch64-builtins.c 
> b/gcc/config/aarch64/aarch64-builtins.c
> index eef9fc0f444..119f67d4e4c 100644
> --- a/gcc/config/aarch64/aarch64-builtins.c
> +++ b/gcc/config/aarch64/aarch64-builtins.c
> @@ -29,6 +29,7 @@
>  #include "rtl.h"
>  #include "tree.h"
>  #include "gimple.h"
> +#include "ssa.h"
>  #include "memmodel.h"
>  #include "tm_p.h"
>  #include "expmed.h"
> @@ -2333,6 +2334,27 @@ aarch64_general_builtin_rsqrt (unsigned int fn)
>    return NULL_TREE;
>  }
>  
> +/* Return true if the lane check can be removed as there is no
> +   error going to be emitted.  */
> +static bool
> +aarch64_fold_builtin_lane_check (tree arg0, tree arg1, tree arg2)
> +{
> +  if (TREE_CODE (arg0) != INTEGER_CST)
> +    return false;
> +  if (TREE_CODE (arg1) != INTEGER_CST)
> +    return false;
> +  if (TREE_CODE (arg2) != INTEGER_CST)
> +    return false;
> +
> +  auto totalsize = wi::to_widest (arg0);
> +  auto elementsize = wi::to_widest (arg1);
> +  if (totalsize == 0 || elementsize == 0)
> +    return false;
> +  auto lane = wi::to_widest (arg2);
> +  auto high = wi::udiv_trunc (totalsize, elementsize);
> +  return wi::ltu_p (lane, high);
> +}
> +
>  #undef VAR1
>  #define VAR1(T, N, MAP, FLAG, A) \
>    case AARCH64_SIMD_BUILTIN_##T##_##N##A:
> @@ -2353,6 +2375,11 @@ aarch64_general_fold_builtin (unsigned int fcode, tree 
> type,
>        VAR1 (UNOP, floatv4si, 2, ALL, v4sf)
>        VAR1 (UNOP, floatv2di, 2, ALL, v2df)
>       return fold_build1 (FLOAT_EXPR, type, args[0]);
> +      case AARCH64_SIMD_BUILTIN_LANE_CHECK:
> +     gcc_assert (n_args == 3);
> +     if (aarch64_fold_builtin_lane_check (args[0], args[1], args[2]))
> +       return void_node;
> +     break;
>        default:
>       break;
>      }
> @@ -2440,6 +2467,14 @@ aarch64_general_gimple_fold_builtin (unsigned int 
> fcode, gcall *stmt)
>           }
>         break;
>       }
> +    case AARCH64_SIMD_BUILTIN_LANE_CHECK:
> +      if (aarch64_fold_builtin_lane_check (args[0], args[1], args[2]))
> +     {
> +       unlink_stmt_vdef (stmt);
> +       release_defs (stmt);
> +       new_stmt = gimple_build_nop ();
> +     }
> +      break;
>      default:
>        break;
>      }
> diff --git a/gcc/testsuite/gcc.target/aarch64/lane-bound-1.c 
> b/gcc/testsuite/gcc.target/aarch64/lane-bound-1.c
> new file mode 100644
> index 00000000000..bbbe679fd80
> --- /dev/null
> +++ b/gcc/testsuite/gcc.target/aarch64/lane-bound-1.c
> @@ -0,0 +1,14 @@
> +/* { dg-do compile } */
> +/* { dg-options "-O2 -fdump-tree-optimized" } */
> +#include <arm_neon.h>
> +
> +void
> +f (float32x4_t **ptr)
> +{
> +  float32x4_t res = vsetq_lane_f32 (0.0f, **ptr, 0);
> +  **ptr = res;
> +}
> +/* GCC should be able to remove the call to 
> "__builtin_aarch64_im_lane_boundsi"
> +   and optimize out the second load from *ptr.  */
> +/* { dg-final { scan-tree-dump-times "__builtin_aarch64_im_lane_boundsi" 0 
> "optimized" } } */
> +/* { dg-final { scan-tree-dump-times " = \\\*ptr_" 1 "optimized" } } */
> diff --git a/gcc/testsuite/gcc.target/aarch64/lane-bound-2.c 
> b/gcc/testsuite/gcc.target/aarch64/lane-bound-2.c
> new file mode 100644
> index 00000000000..923c94687c6
> --- /dev/null
> +++ b/gcc/testsuite/gcc.target/aarch64/lane-bound-2.c
> @@ -0,0 +1,10 @@
> +/* { dg-do compile } */
> +/* { dg-options "-O2 -fdump-tree-original" } */
> +void
> +f (void)
> +{
> +  __builtin_aarch64_im_lane_boundsi (16, 4, 0);
> +  __builtin_aarch64_im_lane_boundsi (8, 8, 0);
> +}
> +/* GCC should be able to optimize these out before gimplification. */
> +/* { dg-final { scan-tree-dump-times "__builtin_aarch64_im_lane_boundsi" 0 
> "original" } } */

Reply via email to