/kernel/linux/linux-5.10/drivers/crypto/nx/ |
H A D | nx-aes-ccm.c | 133 static int generate_b0(u8 *iv, unsigned int assoclen, unsigned int authsize, in generate_b0() argument 148 if (assoclen) in generate_b0() 161 unsigned int assoclen, in generate_pat() 182 * assoclen is an unsigned int, thus it cannot hold a length in generate_pat() 188 if (!assoclen) { in generate_pat() 190 } else if (assoclen <= 14) { in generate_pat() 196 iauth_len = assoclen; in generate_pat() 197 } else if (assoclen <= 65280) { in generate_pat() 211 rc = generate_b0(iv, assoclen, authsize, nbytes, b0); in generate_pat() 221 if (assoclen < in generate_pat() 156 generate_pat(u8 *iv, struct aead_request *req, struct nx_crypto_ctx *nx_ctx, unsigned int authsize, unsigned int nbytes, unsigned int assoclen, u8 *out) generate_pat() argument 329 ccm_nx_decrypt(struct aead_request *req, u8 *iv, unsigned int assoclen) ccm_nx_decrypt() argument 407 ccm_nx_encrypt(struct aead_request *req, u8 *iv, unsigned int assoclen) ccm_nx_encrypt() argument [all...] |
H A D | nx-aes-gcm.c | 102 unsigned int assoclen) in nx_gca() 108 unsigned int nbytes = assoclen; in nx_gca() 159 atomic64_add(assoclen, &(nx_ctx->stats->aes_bytes)); in nx_gca() 169 static int gmac(struct aead_request *req, const u8 *iv, unsigned int assoclen) in gmac() argument 176 unsigned int nbytes = assoclen; in gmac() 231 atomic64_add(assoclen, &(nx_ctx->stats->aes_bytes)); in gmac() 308 unsigned int assoclen) in gcm_aes_nx_crypt() 325 if (assoclen == 0) in gcm_aes_nx_crypt() 328 rc = gmac(req, rctx->iv, assoclen); in gcm_aes_nx_crypt() 336 csbcpb->cpb.aes_gcm.bit_length_aad = assoclen * in gcm_aes_nx_crypt() 99 nx_gca(struct nx_crypto_ctx *nx_ctx, struct aead_request *req, u8 *out, unsigned int assoclen) nx_gca() argument 307 gcm_aes_nx_crypt(struct aead_request *req, int enc, unsigned int assoclen) gcm_aes_nx_crypt() argument [all...] |
/kernel/linux/linux-6.6/drivers/crypto/nx/ |
H A D | nx-aes-ccm.c | 133 static int generate_b0(u8 *iv, unsigned int assoclen, unsigned int authsize, in generate_b0() argument 147 if (assoclen) in generate_b0() 158 unsigned int assoclen, in generate_pat() 179 * assoclen is an unsigned int, thus it cannot hold a length in generate_pat() 185 if (!assoclen) { in generate_pat() 187 } else if (assoclen <= 14) { in generate_pat() 193 iauth_len = assoclen; in generate_pat() 194 } else if (assoclen <= 65280) { in generate_pat() 208 rc = generate_b0(iv, assoclen, authsize, nbytes, b0); in generate_pat() 218 if (assoclen < in generate_pat() 153 generate_pat(u8 *iv, struct aead_request *req, struct nx_crypto_ctx *nx_ctx, unsigned int authsize, unsigned int nbytes, unsigned int assoclen, u8 *out) generate_pat() argument 326 ccm_nx_decrypt(struct aead_request *req, u8 *iv, unsigned int assoclen) ccm_nx_decrypt() argument 404 ccm_nx_encrypt(struct aead_request *req, u8 *iv, unsigned int assoclen) ccm_nx_encrypt() argument [all...] |
H A D | nx-aes-gcm.c | 102 unsigned int assoclen) in nx_gca() 108 unsigned int nbytes = assoclen; in nx_gca() 159 atomic64_add(assoclen, &(nx_ctx->stats->aes_bytes)); in nx_gca() 169 static int gmac(struct aead_request *req, const u8 *iv, unsigned int assoclen) in gmac() argument 176 unsigned int nbytes = assoclen; in gmac() 231 atomic64_add(assoclen, &(nx_ctx->stats->aes_bytes)); in gmac() 308 unsigned int assoclen) in gcm_aes_nx_crypt() 325 if (assoclen == 0) in gcm_aes_nx_crypt() 328 rc = gmac(req, rctx->iv, assoclen); in gcm_aes_nx_crypt() 336 csbcpb->cpb.aes_gcm.bit_length_aad = assoclen * in gcm_aes_nx_crypt() 99 nx_gca(struct nx_crypto_ctx *nx_ctx, struct aead_request *req, u8 *out, unsigned int assoclen) nx_gca() argument 307 gcm_aes_nx_crypt(struct aead_request *req, int enc, unsigned int assoclen) gcm_aes_nx_crypt() argument [all...] |
/kernel/linux/linux-5.10/crypto/ |
H A D | authencesn.c | 96 unsigned int assoclen = req->assoclen; in crypto_authenc_esn_genicv_tail() local 103 scatterwalk_map_and_copy(tmp + 1, dst, assoclen + cryptlen, 4, 0); in crypto_authenc_esn_genicv_tail() 106 scatterwalk_map_and_copy(hash, dst, assoclen + cryptlen, authsize, 1); in crypto_authenc_esn_genicv_tail() 130 unsigned int assoclen = req->assoclen; in crypto_authenc_esn_genicv() local 141 scatterwalk_map_and_copy(tmp + 1, dst, assoclen + cryptlen, 4, 1); in crypto_authenc_esn_genicv() 147 ahash_request_set_crypt(ahreq, dst, hash, assoclen + cryptlen); in crypto_authenc_esn_genicv() 189 unsigned int assoclen = req->assoclen; in crypto_authenc_esn_encrypt() local 232 unsigned int assoclen = req->assoclen; crypto_authenc_esn_decrypt_tail() local 280 unsigned int assoclen = req->assoclen; crypto_authenc_esn_decrypt() local [all...] |
H A D | aegis128-core.c | 72 u64 assoclen, u64 cryptlen); 282 unsigned int assoclen) in crypto_aegis128_process_ad() 289 while (assoclen != 0) { in crypto_aegis128_process_ad() 290 unsigned int size = scatterwalk_clamp(&walk, assoclen); in crypto_aegis128_process_ad() 313 assoclen -= size; in crypto_aegis128_process_ad() 316 scatterwalk_done(&walk, 0, assoclen); in crypto_aegis128_process_ad() 350 u64 assoclen, u64 cryptlen) in crypto_aegis128_final() 352 u64 assocbits = assoclen * 8; in crypto_aegis128_final() 405 crypto_aegis128_process_ad(&state, req->src, req->assoclen); in crypto_aegis128_encrypt() 408 crypto_aegis128_final_simd(&state, &tag, req->assoclen, in crypto_aegis128_encrypt() 280 crypto_aegis128_process_ad(struct aegis_state *state, struct scatterlist *sg_src, unsigned int assoclen) crypto_aegis128_process_ad() argument 348 crypto_aegis128_final(struct aegis_state *state, union aegis_block *tag_xor, u64 assoclen, u64 cryptlen) crypto_aegis128_final() argument [all...] |
H A D | ccm.c | 142 if (req->assoclen) in format_input() 174 unsigned int assoclen = req->assoclen; in crypto_ccm_auth() local 189 if (assoclen) { in crypto_ccm_auth() 190 ilen = format_adata(idata, assoclen); in crypto_ccm_auth() 200 ahash_request_set_crypt(ahreq, sg, NULL, assoclen + ilen + 16); in crypto_ccm_auth() 209 ilen = 16 - (assoclen + ilen) % 16; in crypto_ccm_auth() 235 req->assoclen + req->cryptlen, in crypto_ccm_encrypt_done() 269 sg = scatterwalk_ffwd(pctx->src + 1, req->src, req->assoclen); in crypto_ccm_init_crypt() 276 sg = scatterwalk_ffwd(pctx->dst + 1, req->dst, req->assoclen); in crypto_ccm_init_crypt() [all...] |
H A D | echainiv.c | 51 req->assoclen + req->cryptlen, in echainiv_encrypt() 63 aead_request_set_ad(subreq, req->assoclen); in echainiv_encrypt() 69 scatterwalk_map_and_copy(info, req->dst, req->assoclen, ivsize, 1); in echainiv_encrypt() 105 aead_request_set_ad(subreq, req->assoclen + ivsize); in echainiv_decrypt() 107 scatterwalk_map_and_copy(req->iv, req->src, req->assoclen, ivsize, 0); in echainiv_decrypt()
|
H A D | chacha20poly1305.c | 38 __le64 assoclen; member 61 unsigned int assoclen; member 101 req->assoclen + rctx->cryptlen, in poly_verify_tag() 113 req->assoclen + rctx->cryptlen, in poly_copy_tag() 136 src = scatterwalk_ffwd(rctx->src, req->src, req->assoclen); in chacha_decrypt() 139 dst = scatterwalk_ffwd(rctx->dst, req->dst, req->assoclen); in chacha_decrypt() 177 preq->tail.assoclen = cpu_to_le64(rctx->assoclen); in poly_tail() 239 crypt = scatterwalk_ffwd(rctx->src, crypt, req->assoclen); in poly_cipher() 266 padlen = -rctx->assoclen in poly_adpad() [all...] |
/kernel/linux/linux-6.6/crypto/ |
H A D | authencesn.c | 96 unsigned int assoclen = req->assoclen; in crypto_authenc_esn_genicv_tail() local 103 scatterwalk_map_and_copy(tmp + 1, dst, assoclen + cryptlen, 4, 0); in crypto_authenc_esn_genicv_tail() 106 scatterwalk_map_and_copy(hash, dst, assoclen + cryptlen, authsize, 1); in crypto_authenc_esn_genicv_tail() 129 unsigned int assoclen = req->assoclen; in crypto_authenc_esn_genicv() local 140 scatterwalk_map_and_copy(tmp + 1, dst, assoclen + cryptlen, 4, 1); in crypto_authenc_esn_genicv() 146 ahash_request_set_crypt(ahreq, dst, hash, assoclen + cryptlen); in crypto_authenc_esn_genicv() 187 unsigned int assoclen = req->assoclen; in crypto_authenc_esn_encrypt() local 230 unsigned int assoclen = req->assoclen; crypto_authenc_esn_decrypt_tail() local 277 unsigned int assoclen = req->assoclen; crypto_authenc_esn_decrypt() local [all...] |
H A D | aegis128-core.c | 278 unsigned int assoclen, in crypto_aegis128_process_ad() 286 while (assoclen != 0) { in crypto_aegis128_process_ad() 287 unsigned int size = scatterwalk_clamp(&walk, assoclen); in crypto_aegis128_process_ad() 310 assoclen -= size; in crypto_aegis128_process_ad() 313 scatterwalk_done(&walk, 0, assoclen); in crypto_aegis128_process_ad() 346 u64 assoclen, u64 cryptlen) in crypto_aegis128_final() 348 u64 assocbits = assoclen * 8; in crypto_aegis128_final() 400 crypto_aegis128_process_ad(&state, req->src, req->assoclen, false); in crypto_aegis128_encrypt_generic() 403 crypto_aegis128_final(&state, &tag, req->assoclen, cryptlen); in crypto_aegis128_encrypt_generic() 405 scatterwalk_map_and_copy(tag.bytes, req->dst, req->assoclen in crypto_aegis128_encrypt_generic() 276 crypto_aegis128_process_ad(struct aegis_state *state, struct scatterlist *sg_src, unsigned int assoclen, bool do_simd) crypto_aegis128_process_ad() argument 344 crypto_aegis128_final(struct aegis_state *state, union aegis_block *tag_xor, u64 assoclen, u64 cryptlen) crypto_aegis128_final() argument [all...] |
H A D | chacha20poly1305.c | 38 __le64 assoclen; member 61 unsigned int assoclen; member 101 req->assoclen + rctx->cryptlen, in poly_verify_tag() 113 req->assoclen + rctx->cryptlen, in poly_copy_tag() 136 src = scatterwalk_ffwd(rctx->src, req->src, req->assoclen); in chacha_decrypt() 139 dst = scatterwalk_ffwd(rctx->dst, req->dst, req->assoclen); in chacha_decrypt() 177 preq->tail.assoclen = cpu_to_le64(rctx->assoclen); in poly_tail() 239 crypt = scatterwalk_ffwd(rctx->src, crypt, req->assoclen); in poly_cipher() 266 padlen = -rctx->assoclen in poly_adpad() [all...] |
H A D | ccm.c | 143 if (req->assoclen) in format_input() 175 unsigned int assoclen = req->assoclen; in crypto_ccm_auth() local 190 if (assoclen) { in crypto_ccm_auth() 191 ilen = format_adata(idata, assoclen); in crypto_ccm_auth() 201 ahash_request_set_crypt(ahreq, sg, NULL, assoclen + ilen + 16); in crypto_ccm_auth() 210 ilen = 16 - (assoclen + ilen) % 16; in crypto_ccm_auth() 236 req->assoclen + req->cryptlen, in crypto_ccm_encrypt_done() 270 sg = scatterwalk_ffwd(pctx->src + 1, req->src, req->assoclen); in crypto_ccm_init_crypt() 277 sg = scatterwalk_ffwd(pctx->dst + 1, req->dst, req->assoclen); in crypto_ccm_init_crypt() [all...] |
H A D | echainiv.c | 51 req->assoclen + req->cryptlen, in echainiv_encrypt() 63 aead_request_set_ad(subreq, req->assoclen); in echainiv_encrypt() 69 scatterwalk_map_and_copy(info, req->dst, req->assoclen, ivsize, 1); in echainiv_encrypt() 105 aead_request_set_ad(subreq, req->assoclen + ivsize); in echainiv_decrypt() 107 scatterwalk_map_and_copy(req->iv, req->src, req->assoclen, ivsize, 0); in echainiv_decrypt()
|
/kernel/linux/linux-6.6/drivers/crypto/qce/ |
H A D | aead.c | 54 if (req->assoclen) { in qce_aead_done() 71 totallen = req->cryptlen + req->assoclen; in qce_aead_done() 80 totallen = req->cryptlen + req->assoclen - ctx->authsize; in qce_aead_done() 120 unsigned int assoclen = req->assoclen; in qce_aead_prepare_dst_buf() local 124 totallen = rctx->cryptlen + assoclen; in qce_aead_prepare_dst_buf() 141 if (IS_CCM(rctx->flags) && assoclen) { in qce_aead_prepare_dst_buf() 143 msg_sg = scatterwalk_ffwd(__sg, req->dst, assoclen); in qce_aead_prepare_dst_buf() 146 rctx->assoclen); in qce_aead_prepare_dst_buf() 157 totallen = rctx->cryptlen + rctx->assoclen; in qce_aead_prepare_dst_buf() 191 unsigned int assoclen = rctx->assoclen; qce_aead_ccm_prepare_buf_assoclen() local [all...] |
/kernel/linux/linux-5.10/drivers/crypto/ |
H A D | omap-aes-gcm.c | 61 dd->aead_req->assoclen, dd->total, in omap_aes_gcm_done_task() 67 dd->total + dd->aead_req->assoclen, in omap_aes_gcm_done_task() 91 int alen, clen, cryptlen, assoclen, ret; in omap_aes_gcm_copy_buffers() local 98 assoclen = req->assoclen; in omap_aes_gcm_copy_buffers() 102 assoclen -= 8; in omap_aes_gcm_copy_buffers() 107 alen = ALIGN(assoclen, AES_BLOCK_SIZE); in omap_aes_gcm_copy_buffers() 110 nsg = !!(assoclen && cryptlen); in omap_aes_gcm_copy_buffers() 115 if (assoclen) { in omap_aes_gcm_copy_buffers() 117 ret = omap_crypto_align_sg(&tmp, assoclen, in omap_aes_gcm_copy_buffers() 245 int err, assoclen; omap_aes_gcm_crypt() local [all...] |
/kernel/linux/linux-6.6/drivers/crypto/ |
H A D | omap-aes-gcm.c | 64 dd->aead_req->assoclen, dd->total, in omap_aes_gcm_done_task() 70 dd->total + dd->aead_req->assoclen, in omap_aes_gcm_done_task() 94 int alen, clen, cryptlen, assoclen, ret; in omap_aes_gcm_copy_buffers() local 101 assoclen = req->assoclen; in omap_aes_gcm_copy_buffers() 105 assoclen -= 8; in omap_aes_gcm_copy_buffers() 110 alen = ALIGN(assoclen, AES_BLOCK_SIZE); in omap_aes_gcm_copy_buffers() 113 nsg = !!(assoclen && cryptlen); in omap_aes_gcm_copy_buffers() 118 if (assoclen) { in omap_aes_gcm_copy_buffers() 120 ret = omap_crypto_align_sg(&tmp, assoclen, in omap_aes_gcm_copy_buffers() 246 int err, assoclen; omap_aes_gcm_crypt() local [all...] |
/kernel/linux/linux-5.10/drivers/crypto/cavium/nitrox/ |
H A D | nitrox_aead.c | 167 creq->gph.param1 = cpu_to_be16(rctx->cryptlen + rctx->assoclen); in nitrox_set_creq() 168 creq->gph.param2 = cpu_to_be16(rctx->ivsize + rctx->assoclen); in nitrox_set_creq() 206 static inline bool nitrox_aes_gcm_assoclen_supported(unsigned int assoclen) in nitrox_aes_gcm_assoclen_supported() argument 208 if (assoclen <= 512) in nitrox_aes_gcm_assoclen_supported() 223 if (!nitrox_aes_gcm_assoclen_supported(areq->assoclen)) in nitrox_aes_gcm_enc() 229 rctx->assoclen = areq->assoclen; in nitrox_aes_gcm_enc() 230 rctx->srclen = areq->assoclen + areq->cryptlen; in nitrox_aes_gcm_enc() 257 if (!nitrox_aes_gcm_assoclen_supported(areq->assoclen)) in nitrox_aes_gcm_dec() 263 rctx->assoclen in nitrox_aes_gcm_dec() 398 unsigned int assoclen = areq->assoclen - GCM_RFC4106_IV_SIZE; nitrox_rfc4106_set_aead_rctx_sglist() local [all...] |
/kernel/linux/linux-6.6/drivers/crypto/cavium/nitrox/ |
H A D | nitrox_aead.c | 166 creq->gph.param1 = cpu_to_be16(rctx->cryptlen + rctx->assoclen); in nitrox_set_creq() 167 creq->gph.param2 = cpu_to_be16(rctx->ivsize + rctx->assoclen); in nitrox_set_creq() 205 static inline bool nitrox_aes_gcm_assoclen_supported(unsigned int assoclen) in nitrox_aes_gcm_assoclen_supported() argument 207 if (assoclen <= 512) in nitrox_aes_gcm_assoclen_supported() 222 if (!nitrox_aes_gcm_assoclen_supported(areq->assoclen)) in nitrox_aes_gcm_enc() 228 rctx->assoclen = areq->assoclen; in nitrox_aes_gcm_enc() 229 rctx->srclen = areq->assoclen + areq->cryptlen; in nitrox_aes_gcm_enc() 256 if (!nitrox_aes_gcm_assoclen_supported(areq->assoclen)) in nitrox_aes_gcm_dec() 262 rctx->assoclen in nitrox_aes_gcm_dec() 397 unsigned int assoclen = areq->assoclen - GCM_RFC4106_IV_SIZE; nitrox_rfc4106_set_aead_rctx_sglist() local [all...] |
/kernel/linux/linux-6.6/arch/arm64/crypto/ |
H A D | sm4-ce-ccm-glue.c | 75 if (req->assoclen) in ccm_format_input() 96 u32 assoclen = req->assoclen; in ccm_calculate_auth_mac() local 100 if (assoclen < 0xff00) { in ccm_calculate_auth_mac() 101 aadlen.l = cpu_to_be16(assoclen); in ccm_calculate_auth_mac() 105 put_unaligned_be32(assoclen, &aadlen.h); in ccm_calculate_auth_mac() 115 u32 n = scatterwalk_clamp(&walk, assoclen); in ccm_calculate_auth_mac() 120 n = scatterwalk_clamp(&walk, assoclen); in ccm_calculate_auth_mac() 124 assoclen -= n; in ccm_calculate_auth_mac() 158 scatterwalk_done(&walk, 0, assoclen); in ccm_calculate_auth_mac() [all...] |
H A D | sm4-ce-gcm-glue.c | 78 u32 assoclen = req->assoclen; in gcm_calculate_auth_mac() local 85 u32 n = scatterwalk_clamp(&walk, assoclen); in gcm_calculate_auth_mac() 90 n = scatterwalk_clamp(&walk, assoclen); in gcm_calculate_auth_mac() 94 assoclen -= n; in gcm_calculate_auth_mac() 127 scatterwalk_done(&walk, 0, assoclen); in gcm_calculate_auth_mac() 128 } while (assoclen); in gcm_calculate_auth_mac() 151 lengths.a = cpu_to_be64(req->assoclen * 8); in gcm_crypt() 159 if (req->assoclen) in gcm_crypt() 211 scatterwalk_map_and_copy(ghash, req->dst, req->assoclen in gcm_encrypt() [all...] |
H A D | ghash-ce-glue.c | 335 static int gcm_encrypt(struct aead_request *req, char *iv, int assoclen) in gcm_encrypt() argument 347 lengths.a = cpu_to_be64(assoclen * 8); in gcm_encrypt() 350 if (assoclen) in gcm_encrypt() 351 gcm_calculate_auth_mac(req, dg, assoclen); in gcm_encrypt() 392 scatterwalk_map_and_copy(tag, req->dst, req->assoclen + req->cryptlen, in gcm_encrypt() 398 static int gcm_decrypt(struct aead_request *req, char *iv, int assoclen) in gcm_decrypt() argument 413 lengths.a = cpu_to_be64(assoclen * 8); in gcm_decrypt() 416 if (assoclen) in gcm_decrypt() 417 gcm_calculate_auth_mac(req, dg, assoclen); in gcm_decrypt() 422 req->assoclen in gcm_decrypt() [all...] |
/kernel/linux/linux-6.6/arch/powerpc/crypto/ |
H A D | aes-gcm-p10-glue.c | 114 struct Hash_ctx *hash, u8 *assoc, unsigned int assoclen) in gcmp10_init() 135 gctx->aadLen = assoclen; in gcmp10_init() 137 if (assoclen) in gcmp10_init() 138 set_aad(gctx, hash, assoc, assoclen); in gcmp10_init() 213 unsigned int assoclen = req->assoclen; in p10_aes_gcm_crypt() local 228 if (req->src->length >= assoclen && req->src->length) { in p10_aes_gcm_crypt() 236 assocmem = kmalloc(assoclen, flags); in p10_aes_gcm_crypt() 241 scatterwalk_map_and_copy(assoc, req->src, 0, assoclen, 0); in p10_aes_gcm_crypt() 245 gcmp10_init(gctx, iv, (unsigned char *) &ctx->enc_key, hash, assoc, assoclen); in p10_aes_gcm_crypt() 113 gcmp10_init(struct gcm_ctx *gctx, u8 *iv, unsigned char *rdkey, struct Hash_ctx *hash, u8 *assoc, unsigned int assoclen) gcmp10_init() argument [all...] |
/kernel/linux/linux-5.10/arch/x86/crypto/ |
H A D | aegis128-aesni-glue.c | 45 unsigned int assoclen); 71 unsigned int assoclen) in crypto_aegis128_aesni_process_ad() 78 while (assoclen != 0) { in crypto_aegis128_aesni_process_ad() 79 unsigned int size = scatterwalk_clamp(&walk, assoclen); in crypto_aegis128_aesni_process_ad() 104 assoclen -= size; in crypto_aegis128_aesni_process_ad() 108 scatterwalk_done(&walk, 0, assoclen); in crypto_aegis128_aesni_process_ad() 180 crypto_aegis128_aesni_process_ad(&state, req->src, req->assoclen); in crypto_aegis128_aesni_crypt() 182 crypto_aegis128_aesni_final(&state, tag_xor, req->assoclen, cryptlen); in crypto_aegis128_aesni_crypt() 203 req->assoclen + cryptlen, authsize, 1); in crypto_aegis128_aesni_encrypt() 223 req->assoclen in crypto_aegis128_aesni_decrypt() 69 crypto_aegis128_aesni_process_ad( struct aegis_state *state, struct scatterlist *sg_src, unsigned int assoclen) crypto_aegis128_aesni_process_ad() argument [all...] |
/kernel/linux/linux-6.6/arch/x86/crypto/ |
H A D | aegis128-aesni-glue.c | 45 unsigned int assoclen); 71 unsigned int assoclen) in crypto_aegis128_aesni_process_ad() 78 while (assoclen != 0) { in crypto_aegis128_aesni_process_ad() 79 unsigned int size = scatterwalk_clamp(&walk, assoclen); in crypto_aegis128_aesni_process_ad() 104 assoclen -= size; in crypto_aegis128_aesni_process_ad() 108 scatterwalk_done(&walk, 0, assoclen); in crypto_aegis128_aesni_process_ad() 180 crypto_aegis128_aesni_process_ad(&state, req->src, req->assoclen); in crypto_aegis128_aesni_crypt() 182 crypto_aegis128_aesni_final(&state, tag_xor, req->assoclen, cryptlen); in crypto_aegis128_aesni_crypt() 203 req->assoclen + cryptlen, authsize, 1); in crypto_aegis128_aesni_encrypt() 223 req->assoclen in crypto_aegis128_aesni_decrypt() 69 crypto_aegis128_aesni_process_ad( struct aegis_state *state, struct scatterlist *sg_src, unsigned int assoclen) crypto_aegis128_aesni_process_ad() argument [all...] |