http://gcc.gnu.org/bugzilla/show_bug.cgi?id=59128
ktkachov at gcc dot gnu.org changed:
What |Removed |Added
----------------------------------------------------------------------------
Status|UNCONFIRMED |RESOLVED
CC| |ktkachov at gcc dot gnu.org
Resolution|--- |INVALID
Status|UNCONFIRMED |RESOLVED
Resolution|--- |INVALID
--- Comment #1 from ktkachov at gcc dot gnu.org ---
That's expected behaviour.
ALPHA2 expands to 10.*10.
f47/ALPHA2 is then 100.0 / 10.0 * 10.0
The * and / operators bind from left to right, therefore this is evaluated as
(100.0 / 10.0) * 10.0 = 100.0
That's why it's usually good practice to put parentheses in your #defines:
#define ALPHA2 ((ALPHA) * (ALPHA))
--- Comment #2 from Marc Glisse <glisse at gcc dot gnu.org> ---
> #define ALPHA = 10.
No = there.
> #define ALPHA2 ALPHA*ALPHA
You forgot parentheses.
This has nothing to do with gcc. Look at the output of gcc ZED3.c -E and try to
understand why your code is wrong.