Lines Matching defs:cache
63 u8 cache[HASH_CACHE_SIZE] __aligned(sizeof(u32));
273 memcpy(sreq->cache, sreq->state,
304 memcpy(sreq->cache, sreq->cache_next, cache_len);
333 * fit into full cache blocks, cache it for the next send call.
338 * is a multiple of a block, cache the last one for now.
371 req->cache + cache_len,
375 memset(req->cache + cache_len + skip, 0, extra);
378 req->cache[cache_len + skip] = 0x80;
381 u32 *cache = (void *)req->cache;
386 cache[i] ^= swab32(x);
394 crypto_xor(req->cache, (const u8 *)req->state, AES_BLOCK_SIZE);
400 req->cache_dma = dma_map_single(priv->dev, req->cache,
661 /* safexcel_ahash_cache: cache data until at least one request can be sent to
676 * block size), cache the data until we have enough.
680 req->cache + cache_len,
685 /* We couldn't cache all the data */
751 /* Add request to the cache if it fits */
757 /* If not all data could fit into the cache, go process the excess.
838 /* generate pad block in the cache */
840 memset(req->cache, 0, req->block_sz);
842 req->cache[0] = 0x80;
846 req->cache[req->block_sz-8] = (req->block_sz << 3) &
848 req->cache[req->block_sz-7] = (req->block_sz >> 5);
851 req->cache[req->block_sz-2] = (req->block_sz >> 5);
852 req->cache[req->block_sz-1] = (req->block_sz << 3) &
892 memcpy(export->cache, req->cache, HASH_CACHE_SIZE);
912 memcpy(req->cache, export->cache, HASH_CACHE_SIZE);