https://gcc.gnu.org/bugzilla/show_bug.cgi?id=95341

            Bug ID: 95341
           Summary: Poor vector_size decomposition when SVE is enabled
           Product: gcc
           Version: 11.0
            Status: UNCONFIRMED
          Severity: normal
          Priority: P3
         Component: target
          Assignee: unassigned at gcc dot gnu.org
          Reporter: rsandifo at gcc dot gnu.org
  Target Milestone: ---
            Target: aarch64*-*-*

Compiling the following with -O2 -march=armv8-a:

  typedef unsigned int uint_vec __attribute__((vector_size(32)));
  uint_vec f1(uint_vec x, uint_vec y) { return x + y; }

generates nice Advanced SIMD code:

        ld1     {v4.16b - v5.16b}, [x0]
        ld1     {v2.16b - v3.16b}, [x1]
        add     v0.4s, v4.4s, v2.4s
        add     v1.4s, v5.4s, v3.4s
        st1     {v0.16b - v1.16b}, [x8]
        ret

But compiling with -march=armv8.2-a+sve generates extremely bad
scalar code, so bad that I'll spare people's eyes by not quoting
it here.

I haven't yet analysed this or checked whether it's a regression.

Reply via email to