Lines Matching refs:src
121 struct scatterlist src[2];
129 struct scatterlist src[2];
166 struct scatterlist src[2];
212 struct atmel_aes_dma src;
633 struct scatterlist *src,
643 sg_copy_to_buffer(src, sg_nents(src), dd->buf, len);
710 struct scatterlist *src,
718 dd->src.sg = src;
722 src_aligned = atmel_aes_check_aligned(dd, src, len, &dd->src);
723 if (src == dst)
734 sg_copy_to_buffer(src, sg_nents(src), dd->buf, len);
735 dd->src.sg = &dd->aligned_sg;
736 dd->src.nents = 1;
737 dd->src.remainder = 0;
750 if (dd->src.sg == dd->dst.sg) {
751 dd->src.sg_len = dma_map_sg(dd->dev, dd->src.sg, dd->src.nents,
753 dd->dst.sg_len = dd->src.sg_len;
754 if (!dd->src.sg_len)
757 dd->src.sg_len = dma_map_sg(dd->dev, dd->src.sg, dd->src.nents,
759 if (!dd->src.sg_len)
765 dma_unmap_sg(dd->dev, dd->src.sg, dd->src.nents,
776 if (dd->src.sg == dd->dst.sg) {
777 dma_unmap_sg(dd->dev, dd->src.sg, dd->src.nents,
780 if (dd->src.sg != &dd->aligned_sg)
781 atmel_aes_restore_sg(&dd->src);
789 dma_unmap_sg(dd->dev, dd->src.sg, dd->src.nents,
792 if (dd->src.sg != &dd->aligned_sg)
793 atmel_aes_restore_sg(&dd->src);
820 dma = &dd->src;
853 struct scatterlist *src,
889 err = atmel_aes_map(dd, src, dst, len);
990 return atmel_aes_dma_start(dd, req->src, req->dst,
994 return atmel_aes_cpu_start(dd, req->src, req->dst, req->cryptlen,
1002 struct scatterlist *src, *dst;
1031 src = scatterwalk_ffwd(ctx->src, req->src, ctx->offset);
1032 dst = ((req->src == req->dst) ? src :
1047 return atmel_aes_dma_start(dd, src, dst, datalen,
1050 return atmel_aes_cpu_start(dd, src, dst, datalen,
1082 skcipher_request_set_crypt(&rctx->fallback_req, req->src, req->dst,
1147 scatterwalk_map_and_copy(rctx->lastc, req->src,
1591 sg_copy_to_buffer(req->src, sg_nents(req->src), dd->buf, req->assoclen);
1604 struct scatterlist *src, *dst;
1625 /* Prepare src and dst scatter lists to transfer cipher/plain texts */
1626 src = scatterwalk_ffwd(ctx->src, req->src, req->assoclen);
1627 dst = ((req->src == req->dst) ? src :
1639 return atmel_aes_dma_start(dd, src, dst, ctx->textlen,
1643 return atmel_aes_cpu_start(dd, src, dst, ctx->textlen,
1713 scatterwalk_map_and_copy(itag, req->src, offset, authsize, 0);
1862 return atmel_aes_dma_start(dd, req->src, req->dst,
1866 return atmel_aes_cpu_start(dd, req->src, req->dst, req->cryptlen,
2007 req->src, req->assoclen,
2018 struct scatterlist *src, *dst;
2027 /* Prepare src and dst scatter-lists to transfer cipher/plain texts. */
2028 src = scatterwalk_ffwd(rctx->src, req->src, req->assoclen);
2029 dst = src;
2031 if (req->src != req->dst)
2051 return atmel_aes_dma_start(dd, src, dst, rctx->textlen,
2087 scatterwalk_map_and_copy(idigest, req->src, offs, authsize, 0);
2334 dd->src.chan = dma_request_chan(dd->dev, "tx");
2335 if (IS_ERR(dd->src.chan)) {
2336 ret = PTR_ERR(dd->src.chan);
2349 dma_release_channel(dd->src.chan);
2358 dma_release_channel(dd->src.chan);
2630 dma_chan_name(aes_dd->src.chan),