Lines Matching refs:rctx
35 struct qce_sha_reqctx *rctx = ahash_request_ctx(req);
47 dma_unmap_sg(qce->dev, req->src, rctx->src_nents, DMA_TO_DEVICE);
48 dma_unmap_sg(qce->dev, &rctx->result_sg, 1, DMA_FROM_DEVICE);
50 memcpy(rctx->digest, result->auth_iv, digestsize);
54 rctx->byte_count[0] = cpu_to_be32(result->auth_byte_count[0]);
55 rctx->byte_count[1] = cpu_to_be32(result->auth_byte_count[1]);
61 req->src = rctx->src_orig;
62 req->nbytes = rctx->nbytes_orig;
63 rctx->last_blk = false;
64 rctx->first_blk = false;
72 struct qce_sha_reqctx *rctx = ahash_request_ctx(req);
76 unsigned long flags = rctx->flags;
80 rctx->authkey = ctx->authkey;
81 rctx->authklen = QCE_SHA_HMAC_KEY_SIZE;
83 rctx->authkey = ctx->authkey;
84 rctx->authklen = AES_KEYSIZE_128;
87 rctx->src_nents = sg_nents_for_len(req->src, req->nbytes);
88 if (rctx->src_nents < 0) {
90 return rctx->src_nents;
93 ret = dma_map_sg(qce->dev, req->src, rctx->src_nents, DMA_TO_DEVICE);
97 sg_init_one(&rctx->result_sg, qce->dma.result_buf, QCE_RESULT_BUF_SZ);
99 ret = dma_map_sg(qce->dev, &rctx->result_sg, 1, DMA_FROM_DEVICE);
103 ret = qce_dma_prep_sgs(&qce->dma, req->src, rctx->src_nents,
104 &rctx->result_sg, 1, qce_ahash_done, async_req);
119 dma_unmap_sg(qce->dev, &rctx->result_sg, 1, DMA_FROM_DEVICE);
121 dma_unmap_sg(qce->dev, req->src, rctx->src_nents, DMA_TO_DEVICE);
127 struct qce_sha_reqctx *rctx = ahash_request_ctx(req);
131 memset(rctx, 0, sizeof(*rctx));
132 rctx->first_blk = true;
133 rctx->last_blk = false;
134 rctx->flags = tmpl->alg_flags;
135 memcpy(rctx->digest, std_iv, sizeof(rctx->digest));
143 struct qce_sha_reqctx *rctx = ahash_request_ctx(req);
144 unsigned long flags = rctx->flags;
152 out_state->count = rctx->count;
154 rctx->digest, digestsize);
155 memcpy(out_state->buffer, rctx->buf, blocksize);
159 out_state->count = rctx->count;
161 rctx->digest, digestsize);
162 memcpy(out_state->buf, rctx->buf, blocksize);
174 struct qce_sha_reqctx *rctx = ahash_request_ctx(req);
180 rctx->count = in_count;
181 memcpy(rctx->buf, buffer, blocksize);
184 rctx->first_blk = 1;
186 rctx->first_blk = 0;
196 rctx->byte_count[0] = (__force __be32)(count & ~SHA_PADDING_MASK);
197 rctx->byte_count[1] = (__force __be32)(count >> 32);
198 qce_cpu_to_be32p_array((__be32 *)rctx->digest, (const u8 *)state,
200 rctx->buflen = (unsigned int)(in_count & (blocksize - 1));
207 struct qce_sha_reqctx *rctx;
216 rctx = ahash_request_ctx(req);
217 flags = rctx->flags;
238 struct qce_sha_reqctx *rctx = ahash_request_ctx(req);
248 rctx->count += req->nbytes;
251 total = req->nbytes + rctx->buflen;
254 scatterwalk_map_and_copy(rctx->buf + rctx->buflen, req->src,
256 rctx->buflen += req->nbytes;
261 rctx->src_orig = req->src;
262 rctx->nbytes_orig = req->nbytes;
268 if (rctx->buflen)
269 memcpy(rctx->tmpbuf, rctx->buf, rctx->buflen);
275 scatterwalk_map_and_copy(rctx->buf, req->src, src_offset,
282 len = rctx->buflen;
296 if (rctx->buflen) {
297 sg_init_table(rctx->sg, 2);
298 sg_set_buf(rctx->sg, rctx->tmpbuf, rctx->buflen);
299 sg_chain(rctx->sg, 2, req->src);
300 req->src = rctx->sg;
304 rctx->buflen = hash_later;
311 struct qce_sha_reqctx *rctx = ahash_request_ctx(req);
315 if (!rctx->buflen) {
322 rctx->last_blk = true;
324 rctx->src_orig = req->src;
325 rctx->nbytes_orig = req->nbytes;
327 memcpy(rctx->tmpbuf, rctx->buf, rctx->buflen);
328 sg_init_one(rctx->sg, rctx->tmpbuf, rctx->buflen);
330 req->src = rctx->sg;
331 req->nbytes = rctx->buflen;
338 struct qce_sha_reqctx *rctx = ahash_request_ctx(req);
347 rctx->src_orig = req->src;
348 rctx->nbytes_orig = req->nbytes;
349 rctx->first_blk = true;
350 rctx->last_blk = true;
352 if (!rctx->nbytes_orig) {