Lines Matching refs:alignmask
97 addr = (u8 *)ALIGN((unsigned long)walk->buffer, walk->alignmask + 1);
193 data = PTR_ALIGN(&p->buffer[0], walk->alignmask + 1);
227 unsigned alignmask = walk->alignmask;
252 /* Minimum size to align p->buffer by alignmask. */
253 n += alignmask & ~a;
256 n += (bsize - 1) & ~(alignmask | a);
273 walk->dst.virt.addr = PTR_ALIGN(buffer, alignmask + 1);
366 if (unlikely((walk->in.offset | walk->out.offset) & walk->alignmask)) {
399 unsigned alignmask = walk->alignmask;
406 aligned_bs = ALIGN(bs, alignmask + 1);
408 /* Minimum size to align buffer by alignmask. */
409 size = alignmask & ~a;
417 size += (bs - 1) & ~(alignmask | a);
424 iv = PTR_ALIGN(walk->buffer, alignmask + 1);
437 if (unlikely(((unsigned long)walk->iv & walk->alignmask))) {
471 walk->alignmask = crypto_skcipher_alignmask(tfm);
542 walk->alignmask = crypto_aead_alignmask(tfm);
581 unsigned long alignmask = crypto_skcipher_alignmask(tfm);
587 absize = keylen + alignmask;
592 alignbuffer = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1);
603 unsigned long alignmask = crypto_skcipher_alignmask(tfm);
609 if ((unsigned long)key & alignmask)
924 * alignmask, and priority are set from the underlying cipher but can be