/kernel/linux/linux-6.6/drivers/crypto/aspeed/ |
H A D | aspeed-hace-hash.c | 78 struct aspeed_sham_reqctx *rctx) in aspeed_ahash_fill_padding() 83 AHASH_DBG(hace_dev, "rctx flags:0x%x\n", (u32)rctx->flags); in aspeed_ahash_fill_padding() 85 switch (rctx->flags & SHA_FLAGS_MASK) { in aspeed_ahash_fill_padding() 89 bits[0] = cpu_to_be64(rctx->digcnt[0] << 3); in aspeed_ahash_fill_padding() 90 index = rctx->bufcnt & 0x3f; in aspeed_ahash_fill_padding() 92 *(rctx->buffer + rctx->bufcnt) = 0x80; in aspeed_ahash_fill_padding() 93 memset(rctx->buffer + rctx in aspeed_ahash_fill_padding() 77 aspeed_ahash_fill_padding(struct aspeed_hace_dev *hace_dev, struct aspeed_sham_reqctx *rctx) aspeed_ahash_fill_padding() argument 119 struct aspeed_sham_reqctx *rctx = ahash_request_ctx(req); aspeed_ahash_dma_prepare() local 168 struct aspeed_sham_reqctx *rctx = ahash_request_ctx(req); aspeed_ahash_dma_prepare_sg() local 293 struct aspeed_sham_reqctx *rctx = ahash_request_ctx(req); aspeed_ahash_transfer() local 316 struct aspeed_sham_reqctx *rctx = ahash_request_ctx(req); aspeed_hace_ahash_trigger() local 350 struct aspeed_sham_reqctx *rctx = ahash_request_ctx(req); aspeed_ahash_hmac_resume() local 409 struct aspeed_sham_reqctx *rctx = ahash_request_ctx(req); aspeed_ahash_req_final() local 457 struct aspeed_sham_reqctx *rctx = ahash_request_ctx(req); aspeed_ahash_update_resume_sg() local 488 struct aspeed_sham_reqctx *rctx = ahash_request_ctx(req); aspeed_ahash_update_resume() local 505 struct aspeed_sham_reqctx *rctx = ahash_request_ctx(req); aspeed_ahash_req_update() local 536 struct aspeed_sham_reqctx *rctx = ahash_request_ctx(req); aspeed_ahash_do_request() local 583 struct aspeed_sham_reqctx *rctx = ahash_request_ctx(req); aspeed_sham_update() local 624 struct aspeed_sham_reqctx *rctx = ahash_request_ctx(req); aspeed_sham_final() local 638 struct aspeed_sham_reqctx *rctx = ahash_request_ctx(req); aspeed_sham_finup() local 663 struct aspeed_sham_reqctx *rctx = ahash_request_ctx(req); aspeed_sham_init() local 835 struct aspeed_sham_reqctx *rctx = ahash_request_ctx(req); aspeed_sham_export() local 844 struct aspeed_sham_reqctx *rctx = ahash_request_ctx(req); aspeed_sham_import() local [all...] |
/kernel/linux/linux-6.6/drivers/crypto/qce/ |
H A D | aead.c | 27 struct qce_aead_reqctx *rctx = aead_request_ctx_dma(req); in qce_aead_done() local 49 dma_unmap_sg(qce->dev, rctx->src_sg, rctx->src_nents, dir_src); in qce_aead_done() 51 dma_unmap_sg(qce->dev, rctx->dst_sg, rctx->dst_nents, dir_dst); in qce_aead_done() 53 if (IS_CCM(rctx->flags)) { in qce_aead_done() 55 sg_free_table(&rctx->src_tbl); in qce_aead_done() 57 sg_free_table(&rctx->dst_tbl); in qce_aead_done() 59 if (!(IS_DECRYPT(rctx->flags) && !diff_dst)) in qce_aead_done() 60 sg_free_table(&rctx in qce_aead_done() 95 struct qce_aead_reqctx *rctx = aead_request_ctx_dma(req); qce_aead_prepare_result_buf() local 106 struct qce_aead_reqctx *rctx = aead_request_ctx_dma(req); qce_aead_prepare_ccm_result_buf() local 115 struct qce_aead_reqctx *rctx = aead_request_ctx_dma(req); qce_aead_prepare_dst_buf() local 189 struct qce_aead_reqctx *rctx = aead_request_ctx_dma(req); qce_aead_ccm_prepare_buf_assoclen() local 303 struct qce_aead_reqctx *rctx = aead_request_ctx_dma(req); qce_aead_prepare_buf() local 331 struct qce_aead_reqctx *rctx = aead_request_ctx_dma(req); qce_aead_ccm_prepare_buf() local 361 qce_aead_create_ccm_nonce(struct qce_aead_reqctx *rctx, struct qce_aead_ctx *ctx) qce_aead_create_ccm_nonce() argument 411 struct qce_aead_reqctx *rctx = aead_request_ctx_dma(req); qce_aead_async_req_handle() local 505 struct qce_aead_reqctx *rctx = aead_request_ctx_dma(req); qce_aead_crypt() local [all...] |
H A D | sha.c | 41 struct qce_sha_reqctx *rctx = ahash_request_ctx_dma(req); in qce_ahash_done() local 53 dma_unmap_sg(qce->dev, req->src, rctx->src_nents, DMA_TO_DEVICE); in qce_ahash_done() 54 dma_unmap_sg(qce->dev, &rctx->result_sg, 1, DMA_FROM_DEVICE); in qce_ahash_done() 56 memcpy(rctx->digest, result->auth_iv, digestsize); in qce_ahash_done() 57 if (req->result && rctx->last_blk) in qce_ahash_done() 60 rctx->byte_count[0] = cpu_to_be32(result->auth_byte_count[0]); in qce_ahash_done() 61 rctx->byte_count[1] = cpu_to_be32(result->auth_byte_count[1]); in qce_ahash_done() 67 req->src = rctx->src_orig; in qce_ahash_done() 68 req->nbytes = rctx->nbytes_orig; in qce_ahash_done() 69 rctx in qce_ahash_done() 78 struct qce_sha_reqctx *rctx = ahash_request_ctx_dma(req); qce_ahash_async_req_handle() local 135 struct qce_sha_reqctx *rctx = ahash_request_ctx_dma(req); qce_ahash_init() local 150 struct qce_sha_reqctx *rctx = ahash_request_ctx_dma(req); qce_ahash_export() local 167 struct qce_sha_reqctx *rctx = ahash_request_ctx_dma(req); qce_ahash_import() local 186 struct qce_sha_reqctx *rctx = ahash_request_ctx_dma(req); qce_ahash_update() local 278 struct qce_sha_reqctx *rctx = ahash_request_ctx_dma(req); qce_ahash_final() local 305 struct qce_sha_reqctx *rctx = ahash_request_ctx_dma(req); qce_ahash_digest() local [all...] |
H A D | skcipher.c | 31 struct qce_cipher_reqctx *rctx = skcipher_request_ctx(req); in qce_skcipher_done() local 50 dma_unmap_sg(qce->dev, rctx->src_sg, rctx->src_nents, dir_src); in qce_skcipher_done() 51 dma_unmap_sg(qce->dev, rctx->dst_sg, rctx->dst_nents, dir_dst); in qce_skcipher_done() 53 sg_free_table(&rctx->dst_tbl); in qce_skcipher_done() 59 memcpy(rctx->iv, result_buf->encr_cntr_iv, rctx->ivsize); in qce_skcipher_done() 67 struct qce_cipher_reqctx *rctx = skcipher_request_ctx(req); in qce_skcipher_async_req_handle() local 77 rctx in qce_skcipher_async_req_handle() 260 struct qce_cipher_reqctx *rctx = skcipher_request_ctx(req); qce_skcipher_crypt() local [all...] |
/kernel/linux/linux-6.6/drivers/crypto/intel/keembay/ |
H A D | keembay-ocs-hcu-core.c | 115 static inline unsigned int kmb_get_total_data(struct ocs_hcu_rctx *rctx) in kmb_get_total_data() argument 117 return rctx->sg_data_total + rctx->buf_cnt; in kmb_get_total_data() 121 static int flush_sg_to_ocs_buffer(struct ocs_hcu_rctx *rctx) in flush_sg_to_ocs_buffer() argument 125 if (rctx->sg_data_total > (sizeof(rctx->buffer) - rctx->buf_cnt)) { in flush_sg_to_ocs_buffer() 130 while (rctx->sg_data_total) { in flush_sg_to_ocs_buffer() 131 if (!rctx->sg) { in flush_sg_to_ocs_buffer() 139 if (rctx in flush_sg_to_ocs_buffer() 187 kmb_ocs_hcu_dma_cleanup(struct ahash_request *req, struct ocs_hcu_rctx *rctx) kmb_ocs_hcu_dma_cleanup() argument 229 struct ocs_hcu_rctx *rctx = ahash_request_ctx_dma(req); kmb_ocs_dma_prepare() local 359 struct ocs_hcu_rctx *rctx = ahash_request_ctx_dma(req); kmb_ocs_hcu_secure_cleanup() local 377 struct ocs_hcu_rctx *rctx = ahash_request_ctx_dma(req); prepare_ipad() local 417 struct ocs_hcu_rctx *rctx = ahash_request_ctx_dma(req); kmb_ocs_hcu_do_one_request() local 564 struct ocs_hcu_rctx *rctx = ahash_request_ctx_dma(req); kmb_ocs_hcu_init() local 617 struct ocs_hcu_rctx *rctx = ahash_request_ctx_dma(req); kmb_ocs_hcu_update() local 653 struct ocs_hcu_rctx *rctx = ahash_request_ctx_dma(req); kmb_ocs_hcu_fin_common() local 690 struct ocs_hcu_rctx *rctx = ahash_request_ctx_dma(req); kmb_ocs_hcu_final() local 701 struct ocs_hcu_rctx *rctx = ahash_request_ctx_dma(req); kmb_ocs_hcu_finup() local 729 struct ocs_hcu_rctx *rctx = ahash_request_ctx_dma(req); kmb_ocs_hcu_export() local 739 struct ocs_hcu_rctx *rctx = ahash_request_ctx_dma(req); kmb_ocs_hcu_import() local [all...] |
/kernel/linux/linux-5.10/drivers/crypto/ccp/ |
H A D | ccp-crypto-aes-cmac.c | 28 struct ccp_aes_cmac_req_ctx *rctx = ahash_request_ctx(req); in ccp_aes_cmac_complete() local 34 if (rctx->hash_rem) { in ccp_aes_cmac_complete() 36 unsigned int offset = rctx->nbytes - rctx->hash_rem; in ccp_aes_cmac_complete() 38 scatterwalk_map_and_copy(rctx->buf, rctx->src, in ccp_aes_cmac_complete() 39 offset, rctx->hash_rem, 0); in ccp_aes_cmac_complete() 40 rctx->buf_count = rctx->hash_rem; in ccp_aes_cmac_complete() 42 rctx in ccp_aes_cmac_complete() 60 struct ccp_aes_cmac_req_ctx *rctx = ahash_request_ctx(req); ccp_do_cmac_update() local 185 struct ccp_aes_cmac_req_ctx *rctx = ahash_request_ctx(req); ccp_aes_cmac_init() local 222 struct ccp_aes_cmac_req_ctx *rctx = ahash_request_ctx(req); ccp_aes_cmac_export() local 241 struct ccp_aes_cmac_req_ctx *rctx = ahash_request_ctx(req); ccp_aes_cmac_import() local [all...] |
H A D | ccp-crypto-sha.c | 30 struct ccp_sha_req_ctx *rctx = ahash_request_ctx(req); in ccp_sha_complete() local 36 if (rctx->hash_rem) { in ccp_sha_complete() 38 unsigned int offset = rctx->nbytes - rctx->hash_rem; in ccp_sha_complete() 40 scatterwalk_map_and_copy(rctx->buf, rctx->src, in ccp_sha_complete() 41 offset, rctx->hash_rem, 0); in ccp_sha_complete() 42 rctx->buf_count = rctx->hash_rem; in ccp_sha_complete() 44 rctx in ccp_sha_complete() 62 struct ccp_sha_req_ctx *rctx = ahash_request_ctx(req); ccp_do_sha_update() local 185 struct ccp_sha_req_ctx *rctx = ahash_request_ctx(req); ccp_sha_init() local 233 struct ccp_sha_req_ctx *rctx = ahash_request_ctx(req); ccp_sha_export() local 254 struct ccp_sha_req_ctx *rctx = ahash_request_ctx(req); ccp_sha_import() local [all...] |
/kernel/linux/linux-6.6/drivers/crypto/ccp/ |
H A D | ccp-crypto-aes-cmac.c | 28 struct ccp_aes_cmac_req_ctx *rctx = ahash_request_ctx_dma(req); in ccp_aes_cmac_complete() local 34 if (rctx->hash_rem) { in ccp_aes_cmac_complete() 36 unsigned int offset = rctx->nbytes - rctx->hash_rem; in ccp_aes_cmac_complete() 38 scatterwalk_map_and_copy(rctx->buf, rctx->src, in ccp_aes_cmac_complete() 39 offset, rctx->hash_rem, 0); in ccp_aes_cmac_complete() 40 rctx->buf_count = rctx->hash_rem; in ccp_aes_cmac_complete() 42 rctx in ccp_aes_cmac_complete() 60 struct ccp_aes_cmac_req_ctx *rctx = ahash_request_ctx_dma(req); ccp_do_cmac_update() local 185 struct ccp_aes_cmac_req_ctx *rctx = ahash_request_ctx_dma(req); ccp_aes_cmac_init() local 222 struct ccp_aes_cmac_req_ctx *rctx = ahash_request_ctx_dma(req); ccp_aes_cmac_export() local 241 struct ccp_aes_cmac_req_ctx *rctx = ahash_request_ctx_dma(req); ccp_aes_cmac_import() local [all...] |
H A D | ccp-crypto-sha.c | 31 struct ccp_sha_req_ctx *rctx = ahash_request_ctx_dma(req); in ccp_sha_complete() local 37 if (rctx->hash_rem) { in ccp_sha_complete() 39 unsigned int offset = rctx->nbytes - rctx->hash_rem; in ccp_sha_complete() 41 scatterwalk_map_and_copy(rctx->buf, rctx->src, in ccp_sha_complete() 42 offset, rctx->hash_rem, 0); in ccp_sha_complete() 43 rctx->buf_count = rctx->hash_rem; in ccp_sha_complete() 45 rctx in ccp_sha_complete() 63 struct ccp_sha_req_ctx *rctx = ahash_request_ctx_dma(req); ccp_do_sha_update() local 186 struct ccp_sha_req_ctx *rctx = ahash_request_ctx_dma(req); ccp_sha_init() local 234 struct ccp_sha_req_ctx *rctx = ahash_request_ctx_dma(req); ccp_sha_export() local 255 struct ccp_sha_req_ctx *rctx = ahash_request_ctx_dma(req); ccp_sha_import() local [all...] |
/kernel/linux/linux-5.10/drivers/crypto/qce/ |
H A D | sha.c | 35 struct qce_sha_reqctx *rctx = ahash_request_ctx(req); in qce_ahash_done() local 47 dma_unmap_sg(qce->dev, req->src, rctx->src_nents, DMA_TO_DEVICE); in qce_ahash_done() 48 dma_unmap_sg(qce->dev, &rctx->result_sg, 1, DMA_FROM_DEVICE); in qce_ahash_done() 50 memcpy(rctx->digest, result->auth_iv, digestsize); in qce_ahash_done() 54 rctx->byte_count[0] = cpu_to_be32(result->auth_byte_count[0]); in qce_ahash_done() 55 rctx->byte_count[1] = cpu_to_be32(result->auth_byte_count[1]); in qce_ahash_done() 61 req->src = rctx->src_orig; in qce_ahash_done() 62 req->nbytes = rctx->nbytes_orig; in qce_ahash_done() 63 rctx->last_blk = false; in qce_ahash_done() 64 rctx in qce_ahash_done() 72 struct qce_sha_reqctx *rctx = ahash_request_ctx(req); qce_ahash_async_req_handle() local 127 struct qce_sha_reqctx *rctx = ahash_request_ctx(req); qce_ahash_init() local 143 struct qce_sha_reqctx *rctx = ahash_request_ctx(req); qce_ahash_export() local 174 struct qce_sha_reqctx *rctx = ahash_request_ctx(req); qce_import_common() local 207 struct qce_sha_reqctx *rctx; qce_ahash_import() local 238 struct qce_sha_reqctx *rctx = ahash_request_ctx(req); qce_ahash_update() local 311 struct qce_sha_reqctx *rctx = ahash_request_ctx(req); qce_ahash_final() local 338 struct qce_sha_reqctx *rctx = ahash_request_ctx(req); qce_ahash_digest() local [all...] |
H A D | skcipher.c | 30 struct qce_cipher_reqctx *rctx = skcipher_request_ctx(req); in qce_skcipher_done() local 49 dma_unmap_sg(qce->dev, rctx->src_sg, rctx->src_nents, dir_src); in qce_skcipher_done() 50 dma_unmap_sg(qce->dev, rctx->dst_sg, rctx->dst_nents, dir_dst); in qce_skcipher_done() 52 sg_free_table(&rctx->dst_tbl); in qce_skcipher_done() 58 memcpy(rctx->iv, result_buf->encr_cntr_iv, rctx->ivsize); in qce_skcipher_done() 66 struct qce_cipher_reqctx *rctx = skcipher_request_ctx(req); in qce_skcipher_async_req_handle() local 76 rctx in qce_skcipher_async_req_handle() 227 struct qce_cipher_reqctx *rctx = skcipher_request_ctx(req); qce_skcipher_crypt() local [all...] |
/kernel/linux/linux-6.6/drivers/crypto/starfive/ |
H A D | jh7110-rsa.c | 93 struct starfive_cryp_request_ctx *rctx = ctx->rctx; in starfive_rsa_montgomery_form() local 94 int count = rctx->total / sizeof(u32) - 1; in starfive_rsa_montgomery_form() 100 rctx->csr.pka.v = 0; in starfive_rsa_montgomery_form() 102 writel(rctx->csr.pka.v, cryp->base + STARFIVE_PKA_CACR_OFFSET); in starfive_rsa_montgomery_form() 108 rctx->csr.pka.v = 0; in starfive_rsa_montgomery_form() 109 rctx->csr.pka.cln_done = 1; in starfive_rsa_montgomery_form() 110 rctx->csr.pka.opsize = opsize; in starfive_rsa_montgomery_form() 111 rctx->csr.pka.exposize = opsize; in starfive_rsa_montgomery_form() 112 rctx in starfive_rsa_montgomery_form() 194 struct starfive_cryp_request_ctx *rctx = ctx->rctx; starfive_rsa_cpu_start() local 275 struct starfive_cryp_request_ctx *rctx = ctx->rctx; starfive_rsa_enc_core() local 312 struct starfive_cryp_request_ctx *rctx = akcipher_request_ctx(req); starfive_rsa_enc() local 344 struct starfive_cryp_request_ctx *rctx = akcipher_request_ctx(req); starfive_rsa_dec() local [all...] |
H A D | jh7110-hash.c | 59 struct starfive_cryp_request_ctx *rctx = ctx->rctx; in starfive_hash_hmac_key() local 67 rctx->csr.hash.hmac = 1; in starfive_hash_hmac_key() 68 rctx->csr.hash.key_flag = 1; in starfive_hash_hmac_key() 70 writel(rctx->csr.hash.v, cryp->base + STARFIVE_HASH_SHACSR); in starfive_hash_hmac_key() 90 struct starfive_cryp_request_ctx *rctx = ctx->rctx; in starfive_hash_start() local 96 dma_unmap_sg(cryp->dev, rctx->in_sg, rctx->in_sg_len, DMA_TO_DEVICE); in starfive_hash_start() 115 struct starfive_cryp_request_ctx *rctx in starfive_hash_xmit_dma() local 170 struct starfive_cryp_request_ctx *rctx = ctx->rctx; starfive_hash_xmit() local 200 struct starfive_cryp_request_ctx *rctx = ahash_request_ctx(req); starfive_hash_copy_hash() local 274 struct starfive_cryp_request_ctx *rctx = ahash_request_ctx(req); starfive_hash_init() local 290 struct starfive_cryp_request_ctx *rctx = ahash_request_ctx(req); starfive_hash_update() local 307 struct starfive_cryp_request_ctx *rctx = ahash_request_ctx(req); starfive_hash_final() local 324 struct starfive_cryp_request_ctx *rctx = ahash_request_ctx(req); starfive_hash_finup() local 341 struct starfive_cryp_request_ctx *rctx = ahash_request_ctx(req); starfive_hash_digest_fb() local 359 struct starfive_cryp_request_ctx *rctx = ahash_request_ctx(req); starfive_hash_digest() local 380 struct starfive_cryp_request_ctx *rctx = ahash_request_ctx(req); starfive_hash_export() local 394 struct starfive_cryp_request_ctx *rctx = ahash_request_ctx(req); starfive_hash_import() local [all...] |
/kernel/linux/linux-5.10/drivers/crypto/cavium/nitrox/ |
H A D | nitrox_aead.c | 152 static int nitrox_set_creq(struct nitrox_aead_rctx *rctx) in nitrox_set_creq() argument 154 struct se_crypto_request *creq = &rctx->nkreq.creq; in nitrox_set_creq() 158 creq->flags = rctx->flags; in nitrox_set_creq() 159 creq->gfp = (rctx->flags & CRYPTO_TFM_REQ_MAY_SLEEP) ? GFP_KERNEL : in nitrox_set_creq() 164 creq->ctrl.s.arg = rctx->ctrl_arg; in nitrox_set_creq() 166 creq->gph.param0 = cpu_to_be16(rctx->cryptlen); in nitrox_set_creq() 167 creq->gph.param1 = cpu_to_be16(rctx->cryptlen + rctx->assoclen); in nitrox_set_creq() 168 creq->gph.param2 = cpu_to_be16(rctx->ivsize + rctx in nitrox_set_creq() 194 struct nitrox_aead_rctx *rctx = aead_request_ctx(areq); nitrox_aead_callback() local 218 struct nitrox_aead_rctx *rctx = aead_request_ctx(areq); nitrox_aes_gcm_enc() local 252 struct nitrox_aead_rctx *rctx = aead_request_ctx(areq); nitrox_aes_gcm_dec() local 396 struct nitrox_rfc4106_rctx *rctx = aead_request_ctx(areq); nitrox_rfc4106_set_aead_rctx_sglist() local 428 struct nitrox_rfc4106_rctx *rctx = aead_request_ctx(areq); nitrox_rfc4106_callback() local 445 struct nitrox_rfc4106_rctx *rctx = aead_request_ctx(areq); nitrox_rfc4106_enc() local 477 struct nitrox_rfc4106_rctx *rctx = aead_request_ctx(areq); nitrox_rfc4106_dec() local [all...] |
/kernel/linux/linux-6.6/drivers/crypto/cavium/nitrox/ |
H A D | nitrox_aead.c | 151 static int nitrox_set_creq(struct nitrox_aead_rctx *rctx) in nitrox_set_creq() argument 153 struct se_crypto_request *creq = &rctx->nkreq.creq; in nitrox_set_creq() 157 creq->flags = rctx->flags; in nitrox_set_creq() 158 creq->gfp = (rctx->flags & CRYPTO_TFM_REQ_MAY_SLEEP) ? GFP_KERNEL : in nitrox_set_creq() 163 creq->ctrl.s.arg = rctx->ctrl_arg; in nitrox_set_creq() 165 creq->gph.param0 = cpu_to_be16(rctx->cryptlen); in nitrox_set_creq() 166 creq->gph.param1 = cpu_to_be16(rctx->cryptlen + rctx->assoclen); in nitrox_set_creq() 167 creq->gph.param2 = cpu_to_be16(rctx->ivsize + rctx in nitrox_set_creq() 193 struct nitrox_aead_rctx *rctx = aead_request_ctx(areq); nitrox_aead_callback() local 217 struct nitrox_aead_rctx *rctx = aead_request_ctx(areq); nitrox_aes_gcm_enc() local 251 struct nitrox_aead_rctx *rctx = aead_request_ctx(areq); nitrox_aes_gcm_dec() local 395 struct nitrox_rfc4106_rctx *rctx = aead_request_ctx_dma(areq); nitrox_rfc4106_set_aead_rctx_sglist() local 427 struct nitrox_rfc4106_rctx *rctx = aead_request_ctx_dma(areq); nitrox_rfc4106_callback() local 444 struct nitrox_rfc4106_rctx *rctx = aead_request_ctx_dma(areq); nitrox_rfc4106_enc() local 476 struct nitrox_rfc4106_rctx *rctx = aead_request_ctx(areq); nitrox_rfc4106_dec() local [all...] |
/kernel/linux/linux-6.6/crypto/ |
H A D | chacha20poly1305.c | 74 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); in async_done_continue() local 76 rctx->flags &= ~CRYPTO_TFM_REQ_MAY_SLEEP; in async_done_continue() 97 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); in poly_verify_tag() local 98 u8 tag[sizeof(rctx->tag)]; in poly_verify_tag() 101 req->assoclen + rctx->cryptlen, in poly_verify_tag() 103 if (crypto_memneq(tag, rctx->tag, sizeof(tag))) in poly_verify_tag() 110 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); in poly_copy_tag() local 112 scatterwalk_map_and_copy(rctx->tag, req->dst, in poly_copy_tag() 113 req->assoclen + rctx->cryptlen, in poly_copy_tag() 114 sizeof(rctx in poly_copy_tag() 126 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); chacha_decrypt() local 156 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); poly_tail_continue() local 173 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); poly_tail() local 202 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); poly_cipherpad() local 231 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); poly_cipher() local 261 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); poly_adpad() local 290 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); poly_ad() local 314 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); poly_setkey() local 340 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); poly_init() local 364 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); poly_genkey() local 402 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); chacha_encrypt() local 432 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); chachapoly_encrypt() local 454 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); chachapoly_decrypt() local [all...] |
/kernel/linux/linux-5.10/drivers/crypto/allwinner/sun8i-ss/ |
H A D | sun8i-ss-hash.c | 75 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); in sun8i_ss_hash_init() local 79 memset(rctx, 0, sizeof(struct sun8i_ss_hash_reqctx)); in sun8i_ss_hash_init() 81 ahash_request_set_tfm(&rctx->fallback_req, tfmctx->fallback_tfm); in sun8i_ss_hash_init() 82 rctx->fallback_req.base.flags = areq->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP; in sun8i_ss_hash_init() 84 return crypto_ahash_init(&rctx->fallback_req); in sun8i_ss_hash_init() 89 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); in sun8i_ss_hash_export() local 93 ahash_request_set_tfm(&rctx->fallback_req, tfmctx->fallback_tfm); in sun8i_ss_hash_export() 94 rctx->fallback_req.base.flags = areq->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP; in sun8i_ss_hash_export() 96 return crypto_ahash_export(&rctx->fallback_req, out); in sun8i_ss_hash_export() 101 struct sun8i_ss_hash_reqctx *rctx in sun8i_ss_hash_import() local 113 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); sun8i_ss_hash_final() local 136 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); sun8i_ss_hash_update() local 151 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); sun8i_ss_hash_finup() local 176 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); sun8i_ss_hash_digest_fb() local 199 sun8i_ss_run_hash_task(struct sun8i_ss_dev *ss, struct sun8i_ss_hash_reqctx *rctx, const char *name) sun8i_ss_run_hash_task() argument 286 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); sun8i_ss_hash_digest() local 323 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); sun8i_ss_hash_run() local [all...] |
/kernel/linux/linux-6.6/drivers/crypto/allwinner/sun8i-ss/ |
H A D | sun8i-ss-hash.c | 163 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); in sun8i_ss_hash_init() local 167 memset(rctx, 0, sizeof(struct sun8i_ss_hash_reqctx)); in sun8i_ss_hash_init() 169 ahash_request_set_tfm(&rctx->fallback_req, tfmctx->fallback_tfm); in sun8i_ss_hash_init() 170 rctx->fallback_req.base.flags = areq->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP; in sun8i_ss_hash_init() 172 return crypto_ahash_init(&rctx->fallback_req); in sun8i_ss_hash_init() 177 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); in sun8i_ss_hash_export() local 181 ahash_request_set_tfm(&rctx->fallback_req, tfmctx->fallback_tfm); in sun8i_ss_hash_export() 182 rctx->fallback_req.base.flags = areq->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP; in sun8i_ss_hash_export() 184 return crypto_ahash_export(&rctx->fallback_req, out); in sun8i_ss_hash_export() 189 struct sun8i_ss_hash_reqctx *rctx in sun8i_ss_hash_import() local 201 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); sun8i_ss_hash_final() local 227 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); sun8i_ss_hash_update() local 242 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); sun8i_ss_hash_finup() local 271 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); sun8i_ss_hash_digest_fb() local 298 sun8i_ss_run_hash_task(struct sun8i_ss_dev *ss, struct sun8i_ss_hash_reqctx *rctx, const char *name) sun8i_ss_run_hash_task() argument 409 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); sun8i_ss_hash_digest() local 495 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); sun8i_ss_hash_run() local [all...] |
/kernel/linux/linux-6.6/drivers/crypto/rockchip/ |
H A D | rk3288_crypto_ahash.c | 46 struct rk_ahash_rctx *rctx = ahash_request_ctx(areq); in rk_ahash_digest_fb() local 54 ahash_request_set_tfm(&rctx->fallback_req, tfmctx->fallback_tfm); in rk_ahash_digest_fb() 55 rctx->fallback_req.base.flags = areq->base.flags & in rk_ahash_digest_fb() 58 rctx->fallback_req.nbytes = areq->nbytes; in rk_ahash_digest_fb() 59 rctx->fallback_req.src = areq->src; in rk_ahash_digest_fb() 60 rctx->fallback_req.result = areq->result; in rk_ahash_digest_fb() 62 return crypto_ahash_digest(&rctx->fallback_req); in rk_ahash_digest_fb() 90 struct rk_ahash_rctx *rctx = ahash_request_ctx(req); in rk_ahash_reg_init() local 110 CRYPTO_WRITE(dev, RK_CRYPTO_HASH_CTRL, rctx->mode | in rk_ahash_reg_init() 122 struct rk_ahash_rctx *rctx in rk_ahash_init() local 135 struct rk_ahash_rctx *rctx = ahash_request_ctx(req); rk_ahash_update() local 150 struct rk_ahash_rctx *rctx = ahash_request_ctx(req); rk_ahash_final() local 164 struct rk_ahash_rctx *rctx = ahash_request_ctx(req); rk_ahash_finup() local 181 struct rk_ahash_rctx *rctx = ahash_request_ctx(req); rk_ahash_import() local 194 struct rk_ahash_rctx *rctx = ahash_request_ctx(req); rk_ahash_export() local 207 struct rk_ahash_rctx *rctx = ahash_request_ctx(req); rk_ahash_digest() local 236 struct rk_ahash_rctx *rctx = ahash_request_ctx(areq); rk_hash_prepare() local 252 struct rk_ahash_rctx *rctx = ahash_request_ctx(areq); rk_hash_unprepare() local 262 struct rk_ahash_rctx *rctx = ahash_request_ctx(areq); rk_hash_run() local [all...] |
/kernel/linux/linux-5.10/drivers/crypto/bcm/ |
H A D | cipher.c | 117 * @rctx: crypto request context 134 struct iproc_reqctx_s *rctx, in spu_skcipher_rx_sg_create() 140 struct iproc_ctx_s *ctx = rctx->ctx; in spu_skcipher_rx_sg_create() 144 rctx->gfp); in spu_skcipher_rx_sg_create() 151 sg_set_buf(sg++, rctx->msg_buf.spu_resp_hdr, ctx->spu_resp_hdr_len); in spu_skcipher_rx_sg_create() 156 sg_set_buf(sg++, rctx->msg_buf.c.supdt_tweak, in spu_skcipher_rx_sg_create() 160 datalen = spu_msg_sg_add(&sg, &rctx->dst_sg, &rctx->dst_skip, in spu_skcipher_rx_sg_create() 161 rctx->dst_nents, chunksize); in spu_skcipher_rx_sg_create() 169 sg_set_buf(sg++, rctx in spu_skcipher_rx_sg_create() 133 spu_skcipher_rx_sg_create(struct brcm_message *mssg, struct iproc_reqctx_s *rctx, u8 rx_frag_num, unsigned int chunksize, u32 stat_pad_len) spu_skcipher_rx_sg_create() argument 197 spu_skcipher_tx_sg_create(struct brcm_message *mssg, struct iproc_reqctx_s *rctx, u8 tx_frag_num, unsigned int chunksize, u32 pad_len) spu_skcipher_tx_sg_create() argument 299 handle_skcipher_req(struct iproc_reqctx_s *rctx) handle_skcipher_req() argument 470 handle_skcipher_resp(struct iproc_reqctx_s *rctx) handle_skcipher_resp() argument 527 spu_ahash_rx_sg_create(struct brcm_message *mssg, struct iproc_reqctx_s *rctx, u8 rx_frag_num, unsigned int digestsize, u32 stat_pad_len) spu_ahash_rx_sg_create() argument 579 spu_ahash_tx_sg_create(struct brcm_message *mssg, struct iproc_reqctx_s *rctx, u8 tx_frag_num, u32 spu_hdr_len, unsigned int hash_carry_len, unsigned int new_data_len, u32 pad_len) spu_ahash_tx_sg_create() argument 654 handle_ahash_req(struct iproc_reqctx_s *rctx) handle_ahash_req() argument 945 ahash_req_done(struct iproc_reqctx_s *rctx) ahash_req_done() argument 995 handle_ahash_resp(struct iproc_reqctx_s *rctx) handle_ahash_resp() argument 1044 spu_aead_rx_sg_create(struct brcm_message *mssg, struct aead_request *req, struct iproc_reqctx_s *rctx, u8 rx_frag_num, unsigned int assoc_len, u32 ret_iv_len, unsigned int resp_len, unsigned int digestsize, u32 stat_pad_len) spu_aead_rx_sg_create() argument 1164 spu_aead_tx_sg_create(struct brcm_message *mssg, struct iproc_reqctx_s *rctx, u8 tx_frag_num, u32 spu_hdr_len, struct scatterlist *assoc, unsigned int assoc_len, int assoc_nents, unsigned int aead_iv_len, unsigned int chunksize, u32 aad_pad_len, u32 pad_len, bool incl_icv) spu_aead_tx_sg_create() argument 1261 handle_aead_req(struct iproc_reqctx_s *rctx) handle_aead_req() argument 1531 handle_aead_resp(struct iproc_reqctx_s *rctx) handle_aead_resp() argument 1591 spu_chunk_cleanup(struct iproc_reqctx_s *rctx) spu_chunk_cleanup() argument 1609 finish_req(struct iproc_reqctx_s *rctx, int err) finish_req() argument 1631 struct iproc_reqctx_s *rctx; spu_rx_callback() local 1717 struct iproc_reqctx_s *rctx = skcipher_request_ctx(req); skcipher_enqueue() local 1907 struct iproc_reqctx_s *rctx = ahash_request_ctx(req); ahash_enqueue() local 1964 struct iproc_reqctx_s *rctx = ahash_request_ctx(req); __ahash_init() local 2080 struct iproc_reqctx_s *rctx = ahash_request_ctx(req); __ahash_update() local 2138 struct iproc_reqctx_s *rctx = ahash_request_ctx(req); __ahash_final() local 2175 struct iproc_reqctx_s *rctx = ahash_request_ctx(req); __ahash_finup() local 2290 const struct iproc_reqctx_s *rctx = ahash_request_ctx(req); ahash_export() local 2305 struct iproc_reqctx_s *rctx = ahash_request_ctx(req); ahash_import() local 2420 struct iproc_reqctx_s *rctx = ahash_request_ctx(req); ahash_hmac_init() local 2470 struct iproc_reqctx_s *rctx = ahash_request_ctx(req); ahash_hmac_digest() local 2508 struct iproc_reqctx_s *rctx = aead_request_ctx(req); aead_need_fallback() local 2579 struct iproc_reqctx_s *rctx = aead_request_ctx(req); aead_complete() local 2596 struct iproc_reqctx_s *rctx = aead_request_ctx(req); aead_do_fallback() local 2639 struct iproc_reqctx_s *rctx = aead_request_ctx(req); aead_enqueue() local [all...] |
/kernel/linux/linux-5.10/drivers/crypto/rockchip/ |
H A D | rk3288_crypto_ahash.c | 39 struct rk_ahash_rctx *rctx = ahash_request_ctx(areq); in rk_ahash_digest_fb() local 43 ahash_request_set_tfm(&rctx->fallback_req, tfmctx->fallback_tfm); in rk_ahash_digest_fb() 44 rctx->fallback_req.base.flags = areq->base.flags & in rk_ahash_digest_fb() 47 rctx->fallback_req.nbytes = areq->nbytes; in rk_ahash_digest_fb() 48 rctx->fallback_req.src = areq->src; in rk_ahash_digest_fb() 49 rctx->fallback_req.result = areq->result; in rk_ahash_digest_fb() 51 return crypto_ahash_digest(&rctx->fallback_req); in rk_ahash_digest_fb() 78 struct rk_ahash_rctx *rctx = ahash_request_ctx(req); in rk_ahash_reg_init() local 101 CRYPTO_WRITE(dev, RK_CRYPTO_HASH_CTRL, rctx->mode | in rk_ahash_reg_init() 113 struct rk_ahash_rctx *rctx in rk_ahash_init() local 126 struct rk_ahash_rctx *rctx = ahash_request_ctx(req); rk_ahash_update() local 141 struct rk_ahash_rctx *rctx = ahash_request_ctx(req); rk_ahash_final() local 155 struct rk_ahash_rctx *rctx = ahash_request_ctx(req); rk_ahash_finup() local 172 struct rk_ahash_rctx *rctx = ahash_request_ctx(req); rk_ahash_import() local 185 struct rk_ahash_rctx *rctx = ahash_request_ctx(req); rk_ahash_export() local 222 struct rk_ahash_rctx *rctx = ahash_request_ctx(areq); rk_hash_prepare() local 239 struct rk_ahash_rctx *rctx = ahash_request_ctx(areq); rk_hash_unprepare() local 250 struct rk_ahash_rctx *rctx = ahash_request_ctx(areq); rk_hash_run() local [all...] |
/kernel/linux/linux-6.6/drivers/crypto/bcm/ |
H A D | cipher.c | 117 * @rctx: crypto request context 134 struct iproc_reqctx_s *rctx, in spu_skcipher_rx_sg_create() 140 struct iproc_ctx_s *ctx = rctx->ctx; in spu_skcipher_rx_sg_create() 144 rctx->gfp); in spu_skcipher_rx_sg_create() 151 sg_set_buf(sg++, rctx->msg_buf.spu_resp_hdr, ctx->spu_resp_hdr_len); in spu_skcipher_rx_sg_create() 156 sg_set_buf(sg++, rctx->msg_buf.c.supdt_tweak, in spu_skcipher_rx_sg_create() 160 datalen = spu_msg_sg_add(&sg, &rctx->dst_sg, &rctx->dst_skip, in spu_skcipher_rx_sg_create() 161 rctx->dst_nents, chunksize); in spu_skcipher_rx_sg_create() 169 sg_set_buf(sg++, rctx in spu_skcipher_rx_sg_create() 133 spu_skcipher_rx_sg_create(struct brcm_message *mssg, struct iproc_reqctx_s *rctx, u8 rx_frag_num, unsigned int chunksize, u32 stat_pad_len) spu_skcipher_rx_sg_create() argument 197 spu_skcipher_tx_sg_create(struct brcm_message *mssg, struct iproc_reqctx_s *rctx, u8 tx_frag_num, unsigned int chunksize, u32 pad_len) spu_skcipher_tx_sg_create() argument 299 handle_skcipher_req(struct iproc_reqctx_s *rctx) handle_skcipher_req() argument 470 handle_skcipher_resp(struct iproc_reqctx_s *rctx) handle_skcipher_resp() argument 525 spu_ahash_rx_sg_create(struct brcm_message *mssg, struct iproc_reqctx_s *rctx, u8 rx_frag_num, unsigned int digestsize, u32 stat_pad_len) spu_ahash_rx_sg_create() argument 577 spu_ahash_tx_sg_create(struct brcm_message *mssg, struct iproc_reqctx_s *rctx, u8 tx_frag_num, u32 spu_hdr_len, unsigned int hash_carry_len, unsigned int new_data_len, u32 pad_len) spu_ahash_tx_sg_create() argument 652 handle_ahash_req(struct iproc_reqctx_s *rctx) handle_ahash_req() argument 943 ahash_req_done(struct iproc_reqctx_s *rctx) ahash_req_done() argument 993 handle_ahash_resp(struct iproc_reqctx_s *rctx) handle_ahash_resp() argument 1041 spu_aead_rx_sg_create(struct brcm_message *mssg, struct aead_request *req, struct iproc_reqctx_s *rctx, u8 rx_frag_num, unsigned int assoc_len, u32 ret_iv_len, unsigned int resp_len, unsigned int digestsize, u32 stat_pad_len) spu_aead_rx_sg_create() argument 1161 spu_aead_tx_sg_create(struct brcm_message *mssg, struct iproc_reqctx_s *rctx, u8 tx_frag_num, u32 spu_hdr_len, struct scatterlist *assoc, unsigned int assoc_len, int assoc_nents, unsigned int aead_iv_len, unsigned int chunksize, u32 aad_pad_len, u32 pad_len, bool incl_icv) spu_aead_tx_sg_create() argument 1258 handle_aead_req(struct iproc_reqctx_s *rctx) handle_aead_req() argument 1528 handle_aead_resp(struct iproc_reqctx_s *rctx) handle_aead_resp() argument 1588 spu_chunk_cleanup(struct iproc_reqctx_s *rctx) spu_chunk_cleanup() argument 1606 finish_req(struct iproc_reqctx_s *rctx, int err) finish_req() argument 1628 struct iproc_reqctx_s *rctx; spu_rx_callback() local 1714 struct iproc_reqctx_s *rctx = skcipher_request_ctx(req); skcipher_enqueue() local 1904 struct iproc_reqctx_s *rctx = ahash_request_ctx(req); ahash_enqueue() local 1961 struct iproc_reqctx_s *rctx = ahash_request_ctx(req); __ahash_init() local 2077 struct iproc_reqctx_s *rctx = ahash_request_ctx(req); __ahash_update() local 2135 struct iproc_reqctx_s *rctx = ahash_request_ctx(req); __ahash_final() local 2172 struct iproc_reqctx_s *rctx = ahash_request_ctx(req); __ahash_finup() local 2287 const struct iproc_reqctx_s *rctx = ahash_request_ctx(req); ahash_export() local 2302 struct iproc_reqctx_s *rctx = ahash_request_ctx(req); ahash_import() local 2418 struct iproc_reqctx_s *rctx = ahash_request_ctx(req); ahash_hmac_init() local 2468 struct iproc_reqctx_s *rctx = ahash_request_ctx(req); ahash_hmac_digest() local 2506 struct iproc_reqctx_s *rctx = aead_request_ctx(req); aead_need_fallback() local 2577 struct iproc_reqctx_s *rctx = aead_request_ctx(req); aead_do_fallback() local 2600 struct iproc_reqctx_s *rctx = aead_request_ctx(req); aead_enqueue() local [all...] |
/kernel/linux/linux-5.10/crypto/ |
H A D | chacha20poly1305.c | 74 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); in async_done_continue() local 76 rctx->flags &= ~CRYPTO_TFM_REQ_MAY_SLEEP; in async_done_continue() 97 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); in poly_verify_tag() local 98 u8 tag[sizeof(rctx->tag)]; in poly_verify_tag() 101 req->assoclen + rctx->cryptlen, in poly_verify_tag() 103 if (crypto_memneq(tag, rctx->tag, sizeof(tag))) in poly_verify_tag() 110 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); in poly_copy_tag() local 112 scatterwalk_map_and_copy(rctx->tag, req->dst, in poly_copy_tag() 113 req->assoclen + rctx->cryptlen, in poly_copy_tag() 114 sizeof(rctx in poly_copy_tag() 126 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); chacha_decrypt() local 156 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); poly_tail_continue() local 173 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); poly_tail() local 202 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); poly_cipherpad() local 231 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); poly_cipher() local 261 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); poly_adpad() local 290 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); poly_ad() local 314 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); poly_setkey() local 340 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); poly_init() local 364 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); poly_genkey() local 402 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); chacha_encrypt() local 432 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); chachapoly_encrypt() local 454 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); chachapoly_decrypt() local [all...] |
/kernel/linux/linux-5.10/drivers/crypto/allwinner/sun8i-ce/ |
H A D | sun8i-ce-cipher.c | 57 struct sun8i_cipher_req_ctx *rctx = skcipher_request_ctx(areq); in sun8i_ce_cipher_fallback() local 67 skcipher_request_set_tfm(&rctx->fallback_req, op->fallback_tfm); in sun8i_ce_cipher_fallback() 68 skcipher_request_set_callback(&rctx->fallback_req, areq->base.flags, in sun8i_ce_cipher_fallback() 70 skcipher_request_set_crypt(&rctx->fallback_req, areq->src, areq->dst, in sun8i_ce_cipher_fallback() 72 if (rctx->op_dir & CE_DECRYPTION) in sun8i_ce_cipher_fallback() 73 err = crypto_skcipher_decrypt(&rctx->fallback_req); in sun8i_ce_cipher_fallback() 75 err = crypto_skcipher_encrypt(&rctx->fallback_req); in sun8i_ce_cipher_fallback() 85 struct sun8i_cipher_req_ctx *rctx = skcipher_request_ctx(areq); in sun8i_ce_cipher_prepare() local 103 rctx->op_dir, areq->iv, crypto_skcipher_ivsize(tfm), in sun8i_ce_cipher_prepare() 110 flow = rctx in sun8i_ce_cipher_prepare() 279 struct sun8i_cipher_req_ctx *rctx = skcipher_request_ctx(breq); sun8i_ce_cipher_run() local 296 struct sun8i_cipher_req_ctx *rctx = skcipher_request_ctx(areq); sun8i_ce_cipher_unprepare() local 340 struct sun8i_cipher_req_ctx *rctx = skcipher_request_ctx(areq); sun8i_ce_skdecrypt() local 359 struct sun8i_cipher_req_ctx *rctx = skcipher_request_ctx(areq); sun8i_ce_skencrypt() local [all...] |
/kernel/linux/linux-5.10/drivers/crypto/ |
H A D | sahara.c | 551 struct sahara_aes_reqctx *rctx = skcipher_request_ctx(req); in sahara_aes_cbc_update_iv() local 555 if (rctx->mode & FLAGS_ENCRYPT) { in sahara_aes_cbc_update_iv() 559 memcpy(req->iv, rctx->iv_out, ivsize); in sahara_aes_cbc_update_iv() 568 struct sahara_aes_reqctx *rctx; in sahara_aes_process() local 582 rctx = skcipher_request_ctx(req); in sahara_aes_process() 584 rctx->mode &= FLAGS_MODE_MASK; in sahara_aes_process() 585 dev->flags = (dev->flags & ~FLAGS_MODE_MASK) | rctx->mode; in sahara_aes_process() 594 rctx->iv_out, ivsize, in sahara_aes_process() 654 struct sahara_aes_reqctx *rctx = skcipher_request_ctx(req); in sahara_aes_fallback() local 658 skcipher_request_set_tfm(&rctx in sahara_aes_fallback() 674 struct sahara_aes_reqctx *rctx = skcipher_request_ctx(req); sahara_aes_crypt() local 751 sahara_sha_init_hdr(struct sahara_dev *dev, struct sahara_sha_reqctx *rctx) sahara_sha_init_hdr() argument 774 sahara_sha_hw_links_create(struct sahara_dev *dev, struct sahara_sha_reqctx *rctx, int start) sahara_sha_hw_links_create() argument 817 sahara_sha_hw_data_descriptor_create(struct sahara_dev *dev, struct sahara_sha_reqctx *rctx, struct ahash_request *req, int index) sahara_sha_hw_data_descriptor_create() argument 870 sahara_sha_hw_context_descriptor_create(struct sahara_dev *dev, struct sahara_sha_reqctx *rctx, struct ahash_request *req, int index) sahara_sha_hw_context_descriptor_create() argument 892 struct sahara_sha_reqctx *rctx = ahash_request_ctx(req); sahara_sha_prepare_request() local 950 struct sahara_sha_reqctx *rctx = ahash_request_ctx(req); sahara_sha_process() local 1049 struct sahara_sha_reqctx *rctx = ahash_request_ctx(req); sahara_sha_enqueue() local 1075 struct sahara_sha_reqctx *rctx = ahash_request_ctx(req); sahara_sha_init() local 1123 struct sahara_sha_reqctx *rctx = ahash_request_ctx(req); sahara_sha_export() local 1132 struct sahara_sha_reqctx *rctx = ahash_request_ctx(req); sahara_sha_import() local [all...] |