Lines Matching defs:cache
61 u8 cache[HASH_CACHE_SIZE] __aligned(sizeof(u32));
271 memcpy(sreq->cache, sreq->state,
302 memcpy(sreq->cache, sreq->cache_next, cache_len);
331 * fit into full cache blocks, cache it for the next send call.
336 * is a multiple of a block, cache the last one for now.
369 req->cache + cache_len,
373 memset(req->cache + cache_len + skip, 0, extra);
376 req->cache[cache_len + skip] = 0x80;
379 u32 *cache = (void *)req->cache;
384 cache[i] ^= swab32(x);
392 crypto_xor(req->cache, (const u8 *)req->state, AES_BLOCK_SIZE);
398 req->cache_dma = dma_map_single(priv->dev, req->cache,
659 /* safexcel_ahash_cache: cache data until at least one request can be sent to
674 * block size), cache the data until we have enough.
678 req->cache + cache_len,
683 /* We couldn't cache all the data */
749 /* Add request to the cache if it fits */
755 /* If not all data could fit into the cache, go process the excess.
836 /* generate pad block in the cache */
838 memset(req->cache, 0, req->block_sz);
840 req->cache[0] = 0x80;
844 req->cache[req->block_sz-8] = (req->block_sz << 3) &
846 req->cache[req->block_sz-7] = (req->block_sz >> 5);
849 req->cache[req->block_sz-2] = (req->block_sz >> 5);
850 req->cache[req->block_sz-1] = (req->block_sz << 3) &
890 memcpy(export->cache, req->cache, HASH_CACHE_SIZE);
910 memcpy(req->cache, export->cache, HASH_CACHE_SIZE);