On Fri, Apr 12, 2019 at 11:18 PM H.J. Lu <hjl.to...@gmail.com> wrote:
>
> On Fri, Apr 12, 2019 at 3:19 AM Uros Bizjak <ubiz...@gmail.com> wrote:
> >
> > On Fri, Apr 12, 2019 at 11:03 AM Hongtao Liu <crazy...@gmail.com> wrote:
> > >
> > > On Fri, Apr 12, 2019 at 3:30 PM Uros Bizjak <ubiz...@gmail.com> wrote:
> > > >
> > > > On Fri, Apr 12, 2019 at 9:09 AM Liu, Hongtao <hongtao....@intel.com> 
> > > > wrote:
> > > > >
> > > > > Hi :
> > > > >     This patch is about to enable support for bfloat16 which will be 
> > > > > in Future Cooper Lake, Please refer to 
> > > > > https://software.intel.com/en-us/download/intel-architecture-instruction-set-extensions-programming-reference
> > > > > for more details about BF16.
> > > > >
> > > > > There are 3 instructions for AVX512BF16: VCVTNE2PS2BF16, 
> > > > > VCVTNEPS2BF16 and DPBF16PS instructions, which are Vector Neural 
> > > > > Network Instructions supporting:
> > > > >
> > > > > -       VCVTNE2PS2BF16: Convert Two Packed Single Data to One Packed 
> > > > > BF16 Data.
> > > > > -       VCVTNEPS2BF16: Convert Packed Single Data to Packed BF16 Data.
> > > > > -       VDPBF16PS: Dot Product of BF16 Pairs Accumulated into Packed 
> > > > > Single Precision.
> > > > >
> > > > > Since only BF16 intrinsics are supported, we treat it as HI for 
> > > > > simplicity.
> > > >
> > > > I think it was a mistake declaring cvtps2ph and cvtph2ps using HImode
> > > > instead of HFmode. Is there a compelling reason not to introduce
> > > > corresponding bf16_format supporting infrastructure and declare these
> > > > intrinsics using half-binary (HBmode ?) mode instead?
> > > >
> > > > Uros.
> > >
> > > Bfloat16 isn't IEEE standard which we want to reserve HFmode for.
> >
> > True.
> >
> > > The IEEE 754 standard specifies a binary16 as having the following format:
> > > Sign bit: 1 bit
> > > Exponent width: 5 bits
> > > Significand precision: 11 bits (10 explicitly stored)
> > >
> > > Bfloat16 has the following format:
> > > Sign bit: 1 bit
> > > Exponent width: 8 bits
> > > Significand precision: 8 bits (7 explicitly stored), as opposed to 24
> > > bits in a classical single-precision floating-point format
> >
> > This is why I proposed to introduce HBmode (and corresponding
> > bfloat16_format) to distingush between ieee HFmode and BFmode.
> >
>
> Unless there is BF16 language level support,  HBmode has no advantage
> over HImode.   We can add HBmode when we gain BF16 language support.
>
> --
> H.J.

Any other comments, I'll merge this to trunk?

-- 
BR,
Hongtao

Reply via email to