Re: [PATCH v2] crypto: algapi - make crypto_xor() and crypto_inc() alignment agnostic

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 4 February 2017 at 21:20, Eric Biggers <ebiggers3@xxxxxxxxx> wrote:
> Hi Ard,
>
> On Thu, Feb 02, 2017 at 03:56:28PM +0000, Ard Biesheuvel wrote:
>> +     const int size = sizeof(unsigned long);
>> +     int delta = ((unsigned long)dst ^ (unsigned long)src) & (size - 1);
>> +     int misalign = 0;
>> +
>> +     if (!IS_ENABLED(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && delta) {
>> +             misalign = 1 << __ffs(delta);
>> +
>> +             /*
>> +              * If we care about alignment, process as many bytes as
>> +              * needed to advance dst and src to values whose alignments
>> +              * equal their relative misalignment. This will allow us to
>> +              * process the remainder of the input using optimal strides.
>> +              */
>> +             while (((unsigned long)dst & (misalign - 1)) && len > 0) {
>> +                     *dst++ ^= *src++;
>> +                     len--;
>> +             }
>> +     }
>>
>> +     while (len >= size && !misalign) {
>> +             *(unsigned long *)dst ^= *(unsigned long *)src;
>> +             dst += size;
>> +             src += size;
>> +             len -= size;
>> +     }
>>
>
> Unfortunately this is still broken, for two different reasons.  First, if the
> pointers have the same relative misalignment, then 'delta' and 'misalign' will
> be set to 0 and long accesses will be used, even though the pointers may
> actually be misaligned, e.g. 0x80000001 and 0x90000001.

In this case, the initial loop should run 7 times, but it obviously does not :-(

>  Second, if the pointers
> have a relative misalignent that is not a power-of-2, then 'misalign' will be
> set to the wrong value.  For example, with delta=3, it's actually only safe to
> do byte accesses, but the code will set misalign=2 and do u16 accesses.
>

As you realised, delta == 3 results in misalign == 1, which I think
does the right thing (module the initial loop which is incorrect)

> I kind of liked the version with put_unaligned/get_unaligned (and it seems to
> perform okay on MIPS, though not on ARM which is probably more important).

The reason I don't like the _unaligned() accessors is because they
take the performance hit regardless of whether the pointer is aligned
or not.

>  But
> if the various cases with !CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS are going to
> be handled/optimized I think they will just need to be separated out, maybe
> something like this:
>
> void crypto_xor(u8 *dst, const u8 *src, unsigned int len)
> {
> #ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS
>         unsigned long delta;
>
>         if (((unsigned long)dst | (unsigned long)src | len) %
>             sizeof(unsigned long) == 0) {
>                 /* fast path: everything is aligned, including len */
>                 while (len >= sizeof(unsigned long)) {
>                         *(unsigned long *)dst ^= *(unsigned long *)src;
>                         dst += sizeof(unsigned long);
>                         src += sizeof(unsigned long);
>                         len -= sizeof(unsigned long);
>                 }
>                 return;
>         }
>
>         /* handle relative misalignment */
>         delta = (unsigned long)dst ^ (unsigned long)src;
>         if (delta % sizeof(unsigned long)) {
>
>                 /* 1-byte relative misalignment: do byte accesses */
>                 if (delta & 1) {
>                         while (len--)
>                                 *dst++ ^= *src++;
>                         return;
>                 }
>
>                 /* 2-byte relative misalignment: do u16 accesses */
>                 if ((delta & 2) || sizeof(unsigned long) == 4) {
>                         if ((unsigned long)dst % 2 && len) {
>                                 *dst++ ^= *src++;
>                                 len--;
>                         }
>                         while (len >= 2) {
>                                 *(u16 *)dst ^= *(u16 *)src;
>                                 dst += 2, src += 2, len -= 2;
>                         }
>                         if (len)
>                                 *dst ^= *src;
>                         return;
>                 }
>
>                 /* 4-byte relative misalignment: do u32 accesses */
>                 while ((unsigned long)dst % 4 && len) {
>                         *dst++ ^= *src++;
>                         len--;
>                 }
>                 while (len >= 4) {
>                         *(u32 *)dst ^= *(u32 *)src;
>                         dst += 4, src += 4, len -= 4;
>                 }
>                 while (len--)
>                         *dst++ ^= *src++;
>                 return;
>         }
>
>         /* handle absolute misalignment */
>         while ((unsigned long)dst % sizeof(unsigned long) && len) {
>                 *dst++ ^= *src++;
>                 len--;
>         }
> #endif /* !CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS */
>
>         while (len >= sizeof(unsigned long)) {
>                 *(unsigned long *)dst ^= *(unsigned long *)src;
>                 dst += sizeof(unsigned long);
>                 src += sizeof(unsigned long);
>                 len -= sizeof(unsigned long);
>         }
>
>         while (len--)
>                 *dst++ ^= *src++;
> }



[Index of Archives]     [Kernel]     [Gnu Classpath]     [Gnu Crypto]     [DM Crypt]     [Netfilter]     [Bugtraq]

  Powered by Linux