Lines Matching defs:creq
61 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req);
64 dma_unmap_sg(cesa_dev->dev, req->dst, creq->dst_nents,
66 dma_unmap_sg(cesa_dev->dev, req->src, creq->src_nents,
69 dma_unmap_sg(cesa_dev->dev, req->src, creq->src_nents,
72 mv_cesa_dma_cleanup(&creq->base);
77 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req);
79 if (mv_cesa_req_get_type(&creq->base) == CESA_DMA_REQ)
85 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req);
86 struct mv_cesa_skcipher_std_req *sreq = &creq->std;
87 struct mv_cesa_engine *engine = creq->base.engine;
94 len = sg_pcopy_to_buffer(req->src, creq->src_nents,
119 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req);
120 struct mv_cesa_skcipher_std_req *sreq = &creq->std;
121 struct mv_cesa_engine *engine = creq->base.engine;
124 len = sg_pcopy_from_buffer(req->dst, creq->dst_nents,
139 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(skreq);
140 struct mv_cesa_req *basereq = &creq->base;
151 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(skreq);
153 if (mv_cesa_req_get_type(&creq->base) == CESA_DMA_REQ)
154 mv_cesa_dma_step(&creq->base);
162 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req);
163 struct mv_cesa_req *basereq = &creq->base;
171 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req);
172 struct mv_cesa_skcipher_std_req *sreq = &creq->std;
182 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(skreq);
184 creq->base.engine = engine;
186 if (mv_cesa_req_get_type(&creq->base) == CESA_DMA_REQ)
204 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(skreq);
205 struct mv_cesa_engine *engine = creq->base.engine;
211 if (mv_cesa_req_get_type(&creq->base) == CESA_DMA_REQ) {
214 basereq = &creq->base;
305 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req);
308 struct mv_cesa_req *basereq = &creq->base;
317 ret = dma_map_sg(cesa_dev->dev, req->src, creq->src_nents,
322 ret = dma_map_sg(cesa_dev->dev, req->dst, creq->dst_nents,
329 ret = dma_map_sg(cesa_dev->dev, req->src, creq->src_nents,
386 dma_unmap_sg(cesa_dev->dev, req->dst, creq->dst_nents,
390 dma_unmap_sg(cesa_dev->dev, req->src, creq->src_nents,
400 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req);
401 struct mv_cesa_skcipher_std_req *sreq = &creq->std;
402 struct mv_cesa_req *basereq = &creq->base;
415 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req);
423 creq->src_nents = sg_nents_for_len(req->src, req->cryptlen);
424 if (creq->src_nents < 0) {
426 return creq->src_nents;
428 creq->dst_nents = sg_nents_for_len(req->dst, req->cryptlen);
429 if (creq->dst_nents < 0) {
431 return creq->dst_nents;
449 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req);
459 ret = mv_cesa_queue_req(&req->base, &creq->base);