https://gcc.gnu.org/bugzilla/show_bug.cgi?id=69909
--- Comment #6 from Jakub Jelinek <jakub at gcc dot gnu.org> --- Indeed, the *.optimized dump looks good, the late gimple DCE just removed really dead stmts. Slightly cleaned up testcase: typedef unsigned V __attribute__ ((vector_size (32))); typedef __int128 T; typedef __int128 U __attribute__ ((vector_size (32))); __attribute__((noinline, noclone)) T foo (T a, V b, V c, V d, V e, U f) { d[6] ^= 0x10; f -= (U) d; f[1] |= f[1] << (a & 127); c ^= d; return b[7] + c[2] + c[2] + d[6] + e[2] + f[1]; } int main () { T x = foo (1, (V) { 9, 2, 5, 8, 1, 2, 9, 3 }, (V) { 1, 2, 3, 4, 5, 6, 7, 8 }, (V) { 4, 1, 2, 9, 8, 3, 5, 2 }, (V) { 3, 6, 1, 3, 2, 9, 4, 8 }, (U) { 3, 5 }); if (((unsigned long long) (x >> 64) != 0xffffffffffffffffULL || (unsigned long long) x != 0xfffffffe0000001aULL) && ((unsigned long long) (x >> 64) != 0xfffffffffffffffdULL || (unsigned long long) x != 0xffffffff00000022ULL)) __builtin_abort (); return 0; }