Hello! Attached patch fixes invalid atomic_storedi_fpu predicate
(match_operand:DI 1 "register_operand" "x,m,?r"). This predicate is not compatible with memory constraint. While at this, I also changed expanders to accept memory operands that can be handled by *_fpu instructions. To avoid unwanted pattern matching, UNSPEC_MOVA has been replaced with UNSPEC_LDA and UNSPEC_STA. Also, the move from uninitalized operand in atomic_store_fpu was fixed in case of memory input operand with xmm temporary register. 2015-04-01 Uros Bizjak <ubiz...@gmail.com> * config/i386/sync.md (UNSPEC_MOVA): Remove. (atomic_load<mode>): Change operand 0 predicate to nonimmediate_operand and fix up the destination when needed. Use UNSPEC_LDA. (atomic_loaddi_fpu): Use UNSPEC_LDA. (atomic_store<mode>): Change operand 1 predicate to nonimmendate_operand and move the source to register when needed. Use UNSPEC_STA. (atomic_store<mode>_1): Use UNSPEC_STA. (atomic_storedi_fpu): Change operand 1 to nonimmediate_operand. Fix moves from memory operand. Use UNSPEC_STA. The patch was bootstraped and regression tested on x86_64-linux-gnu {,-m32} and committed to mainline SVN. Uros.
Index: config/i386/sync.md =================================================================== --- config/i386/sync.md (revision 221807) +++ config/i386/sync.md (working copy) @@ -21,7 +21,8 @@ UNSPEC_LFENCE UNSPEC_SFENCE UNSPEC_MFENCE - UNSPEC_MOVA ; For __atomic support + + ;; __atomic support UNSPEC_LDA UNSPEC_STA ]) @@ -140,10 +141,10 @@ ]) (define_expand "atomic_load<mode>" - [(set (match_operand:ATOMIC 0 "register_operand") + [(set (match_operand:ATOMIC 0 "nonimmediate_operand") (unspec:ATOMIC [(match_operand:ATOMIC 1 "memory_operand") (match_operand:SI 2 "const_int_operand")] - UNSPEC_MOVA))] + UNSPEC_LDA))] "" { /* For DImode on 32-bit, we can use the FPU to perform the load. */ @@ -152,14 +153,25 @@ (operands[0], operands[1], assign_386_stack_local (DImode, SLOT_TEMP))); else - emit_move_insn (operands[0], operands[1]); + { + rtx dst = operands[0]; + + if (MEM_P (dst)) + dst = gen_reg_rtx (<MODE>mode); + + emit_move_insn (dst, operands[1]); + + /* Fix up the destination if needed. */ + if (dst != operands[0]) + emit_move_insn (operands[0], dst); + } DONE; }) (define_insn_and_split "atomic_loaddi_fpu" [(set (match_operand:DI 0 "nonimmediate_operand" "=x,m,?r") (unspec:DI [(match_operand:DI 1 "memory_operand" "m,m,m")] - UNSPEC_MOVA)) + UNSPEC_LDA)) (clobber (match_operand:DI 2 "memory_operand" "=X,X,m")) (clobber (match_scratch:DF 3 "=X,xf,xf"))] "!TARGET_64BIT && (TARGET_80387 || TARGET_SSE)" @@ -197,9 +209,9 @@ (define_expand "atomic_store<mode>" [(set (match_operand:ATOMIC 0 "memory_operand") - (unspec:ATOMIC [(match_operand:ATOMIC 1 "register_operand") + (unspec:ATOMIC [(match_operand:ATOMIC 1 "nonimmediate_operand") (match_operand:SI 2 "const_int_operand")] - UNSPEC_MOVA))] + UNSPEC_STA))] "" { enum memmodel model = (enum memmodel) (INTVAL (operands[2]) & MEMMODEL_MASK); @@ -215,6 +227,8 @@ } else { + operands[1] = force_reg (<MODE>mode, operands[1]); + /* For seq-cst stores, when we lack MFENCE, use XCHG. */ if (model == MEMMODEL_SEQ_CST && !(TARGET_64BIT || TARGET_SSE2)) { @@ -238,14 +252,14 @@ [(set (match_operand:SWI 0 "memory_operand" "=m") (unspec:SWI [(match_operand:SWI 1 "<nonmemory_operand>" "<r><i>") (match_operand:SI 2 "const_int_operand")] - UNSPEC_MOVA))] + UNSPEC_STA))] "" "%K2mov{<imodesuffix>}\t{%1, %0|%0, %1}") (define_insn_and_split "atomic_storedi_fpu" [(set (match_operand:DI 0 "memory_operand" "=m,m,m") - (unspec:DI [(match_operand:DI 1 "register_operand" "x,m,?r")] - UNSPEC_MOVA)) + (unspec:DI [(match_operand:DI 1 "nonimmediate_operand" "x,m,?r")] + UNSPEC_STA)) (clobber (match_operand:DI 2 "memory_operand" "=X,X,m")) (clobber (match_scratch:DF 3 "=X,xf,xf"))] "!TARGET_64BIT && (TARGET_80387 || TARGET_SSE)" @@ -273,7 +287,7 @@ else { adjust_reg_mode (tmp, DImode); - emit_move_insn (tmp, mem); + emit_move_insn (tmp, src); src = tmp; } }