Lines Matching refs:alignmask
116 addr = (u8 *)ALIGN((unsigned long)walk->buffer, walk->alignmask + 1);
212 data = PTR_ALIGN(&p->buffer[0], walk->alignmask + 1);
246 unsigned alignmask = walk->alignmask;
271 /* Minimum size to align p->buffer by alignmask. */
272 n += alignmask & ~a;
275 n += (bsize - 1) & ~(alignmask | a);
292 walk->dst.virt.addr = PTR_ALIGN(buffer, alignmask + 1);
385 if (unlikely((walk->in.offset | walk->out.offset) & walk->alignmask)) {
418 unsigned alignmask = walk->alignmask;
425 aligned_bs = ALIGN(bs, alignmask + 1);
427 /* Minimum size to align buffer by alignmask. */
428 size = alignmask & ~a;
436 size += (bs - 1) & ~(alignmask | a);
443 iv = PTR_ALIGN(walk->buffer, alignmask + 1);
456 if (unlikely(((unsigned long)walk->iv & walk->alignmask))) {
490 walk->alignmask = crypto_skcipher_alignmask(tfm);
555 walk->alignmask = crypto_aead_alignmask(tfm);
594 unsigned long alignmask = crypto_skcipher_alignmask(tfm);
600 absize = keylen + alignmask;
605 alignbuffer = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1);
616 unsigned long alignmask = crypto_skcipher_alignmask(tfm);
622 if ((unsigned long)key & alignmask)
970 * alignmask, and priority are set from the underlying cipher but can be