Hi Yann,

On Wed, Sep 12, 2018 at 11:50:00AM +0200, Yann Droneaud wrote:
> Hi,
> 
> Le mardi 11 septembre 2018 à 20:05 -0700, Eric Biggers a écrit :
> > From: Eric Biggers <ebigg...@google.com>
> > 
> > In commit 9f480faec58c ("crypto: chacha20 - Fix keystream alignment for
> > chacha20_block()"), I had missed that chacha20_block() can be called
> > directly on the buffer passed to get_random_bytes(), which can have any
> > alignment.  So, while my commit didn't break anything, it didn't fully
> > solve the alignment problems.
> > 
> > Revert my solution and just update chacha20_block() to use
> > put_unaligned_le32(), so the output buffer need not be aligned.
> > This is simpler, and on many CPUs it's the same speed.
> > 
> > But, I kept the 'tmp' buffers in extract_crng_user() and
> > _get_random_bytes() 4-byte aligned, since that alignment is actually
> > needed for _crng_backtrack_protect() too.
> > 
> > Reported-by: Stephan Müller <smuel...@chronox.de>
> > Cc: Theodore Ts'o <ty...@mit.edu>
> > Signed-off-by: Eric Biggers <ebigg...@google.com>
> > ---
> >  crypto/chacha20_generic.c |  7 ++++---
> >  drivers/char/random.c     | 24 ++++++++++++------------
> >  include/crypto/chacha20.h |  3 +--
> >  lib/chacha20.c            |  6 +++---
> >  4 files changed, 20 insertions(+), 20 deletions(-)
> > 
> > diff --git a/crypto/chacha20_generic.c b/crypto/chacha20_generic.c
> > index e451c3cb6a56..3ae96587caf9 100644
> > --- a/crypto/chacha20_generic.c
> > +++ b/crypto/chacha20_generic.c
> > @@ -18,20 +18,21 @@
> >  static void chacha20_docrypt(u32 *state, u8 *dst, const u8 *src,
> >                          unsigned int bytes)
> >  {
> > -   u32 stream[CHACHA20_BLOCK_WORDS];
> > +   /* aligned to potentially speed up crypto_xor() */
> > +   u8 stream[CHACHA20_BLOCK_SIZE] __aligned(sizeof(long));
> >  
> >     if (dst != src)
> >             memcpy(dst, src, bytes);
> >  
> >     while (bytes >= CHACHA20_BLOCK_SIZE) {
> >             chacha20_block(state, stream);
> > -           crypto_xor(dst, (const u8 *)stream, CHACHA20_BLOCK_SIZE);
> > +           crypto_xor(dst, stream, CHACHA20_BLOCK_SIZE);
> >             bytes -= CHACHA20_BLOCK_SIZE;
> >             dst += CHACHA20_BLOCK_SIZE;
> >     }
> >     if (bytes) {
> >             chacha20_block(state, stream);
> > -           crypto_xor(dst, (const u8 *)stream, bytes);
> > +           crypto_xor(dst, stream, bytes);
> >     }
> >  }
> >  
> > diff --git a/drivers/char/random.c b/drivers/char/random.c
> > index bf5f99fc36f1..d22d967c50f0 100644
> > --- a/drivers/char/random.c
> > +++ b/drivers/char/random.c
> > @@ -1003,7 +1003,7 @@ static void extract_crng(__u32 
> > out[CHACHA20_BLOCK_WORDS])
> >   * enough) to mutate the CRNG key to provide backtracking protection.
> >   */
> >  static void _crng_backtrack_protect(struct crng_state *crng,
> > -                               __u32 tmp[CHACHA20_BLOCK_WORDS], int used)
> > +                               __u8 tmp[CHACHA20_BLOCK_SIZE], int used)
> >  {
> >     unsigned long   flags;
> >     __u32           *s, *d;
> > @@ -1015,14 +1015,14 @@ static void _crng_backtrack_protect(struct 
> > crng_state *crng,
> >             used = 0;
> >     }
> >     spin_lock_irqsave(&crng->lock, flags);
> > -   s = &tmp[used / sizeof(__u32)];
> > +   s = (__u32 *) &tmp[used];
> 
> This introduces a alignment issue: tmp is not aligned for __u32, but is
> dereferenced as such later.
> 
> >     d = &crng->state[4];
> >     for (i=0; i < 8; i++)
> >             *d++ ^= *s++;
> >     spin_unlock_irqrestore(&crng->lock, flags);
> >  }
> >  
> 

I explained this in the patch; the callers ensure the buffer is aligned.

- Eric

Reply via email to