Lines Matching defs:creq
61 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req);
64 dma_unmap_sg(cesa_dev->dev, req->dst, creq->dst_nents,
66 dma_unmap_sg(cesa_dev->dev, req->src, creq->src_nents,
69 dma_unmap_sg(cesa_dev->dev, req->src, creq->src_nents,
72 mv_cesa_dma_cleanup(&creq->base);
77 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req);
79 if (mv_cesa_req_get_type(&creq->base) == CESA_DMA_REQ)
85 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req);
86 struct mv_cesa_skcipher_std_req *sreq = &creq->std;
87 struct mv_cesa_engine *engine = creq->base.engine;
97 len = mv_cesa_sg_copy_to_sram(engine, req->src, creq->src_nents,
126 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req);
127 struct mv_cesa_skcipher_std_req *sreq = &creq->std;
128 struct mv_cesa_engine *engine = creq->base.engine;
131 len = mv_cesa_sg_copy_from_sram(engine, req->dst, creq->dst_nents,
146 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(skreq);
147 struct mv_cesa_req *basereq = &creq->base;
158 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(skreq);
160 if (mv_cesa_req_get_type(&creq->base) == CESA_DMA_REQ)
161 mv_cesa_dma_step(&creq->base);
169 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req);
170 struct mv_cesa_req *basereq = &creq->base;
178 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req);
179 struct mv_cesa_skcipher_std_req *sreq = &creq->std;
189 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(skreq);
191 creq->base.engine = engine;
193 if (mv_cesa_req_get_type(&creq->base) == CESA_DMA_REQ)
211 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(skreq);
212 struct mv_cesa_engine *engine = creq->base.engine;
218 if (mv_cesa_req_get_type(&creq->base) == CESA_DMA_REQ) {
221 basereq = &creq->base;
315 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req);
318 struct mv_cesa_req *basereq = &creq->base;
327 ret = dma_map_sg(cesa_dev->dev, req->src, creq->src_nents,
332 ret = dma_map_sg(cesa_dev->dev, req->dst, creq->dst_nents,
339 ret = dma_map_sg(cesa_dev->dev, req->src, creq->src_nents,
396 dma_unmap_sg(cesa_dev->dev, req->dst, creq->dst_nents,
400 dma_unmap_sg(cesa_dev->dev, req->src, creq->src_nents,
410 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req);
411 struct mv_cesa_skcipher_std_req *sreq = &creq->std;
412 struct mv_cesa_req *basereq = &creq->base;
425 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req);
433 creq->src_nents = sg_nents_for_len(req->src, req->cryptlen);
434 if (creq->src_nents < 0) {
436 return creq->src_nents;
438 creq->dst_nents = sg_nents_for_len(req->dst, req->cryptlen);
439 if (creq->dst_nents < 0) {
441 return creq->dst_nents;
459 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req);
469 ret = mv_cesa_queue_req(&req->base, &creq->base);