/kernel/linux/linux-5.10/drivers/crypto/marvell/cesa/ |
H A D | cipher.c | 413 struct mv_cesa_op_ctx *tmpl) in mv_cesa_skcipher_req_init() 434 mv_cesa_update_op_cfg(tmpl, CESA_SA_DESC_CFG_OP_CRYPT_ONLY, in mv_cesa_skcipher_req_init() 438 ret = mv_cesa_skcipher_dma_req_init(req, tmpl); in mv_cesa_skcipher_req_init() 440 ret = mv_cesa_skcipher_std_req_init(req, tmpl); in mv_cesa_skcipher_req_init() 446 struct mv_cesa_op_ctx *tmpl) in mv_cesa_skcipher_queue_req() 452 ret = mv_cesa_skcipher_req_init(req, tmpl); in mv_cesa_skcipher_queue_req() 468 struct mv_cesa_op_ctx *tmpl) in mv_cesa_des_op() 472 mv_cesa_update_op_cfg(tmpl, CESA_SA_DESC_CFG_CRYPTM_DES, in mv_cesa_des_op() 475 memcpy(tmpl->ctx.skcipher.key, ctx->key, DES_KEY_SIZE); in mv_cesa_des_op() 477 return mv_cesa_skcipher_queue_req(req, tmpl); in mv_cesa_des_op() 412 mv_cesa_skcipher_req_init(struct skcipher_request *req, struct mv_cesa_op_ctx *tmpl) mv_cesa_skcipher_req_init() argument 445 mv_cesa_skcipher_queue_req(struct skcipher_request *req, struct mv_cesa_op_ctx *tmpl) mv_cesa_skcipher_queue_req() argument 467 mv_cesa_des_op(struct skcipher_request *req, struct mv_cesa_op_ctx *tmpl) mv_cesa_des_op() argument 482 struct mv_cesa_op_ctx tmpl; mv_cesa_ecb_des_encrypt() local 493 struct mv_cesa_op_ctx tmpl; mv_cesa_ecb_des_decrypt() local 523 mv_cesa_cbc_des_op(struct skcipher_request *req, struct mv_cesa_op_ctx *tmpl) mv_cesa_cbc_des_op() argument 536 struct mv_cesa_op_ctx tmpl; mv_cesa_cbc_des_encrypt() local 545 struct mv_cesa_op_ctx tmpl; mv_cesa_cbc_des_decrypt() local 574 mv_cesa_des3_op(struct skcipher_request *req, struct mv_cesa_op_ctx *tmpl) mv_cesa_des3_op() argument 589 struct mv_cesa_op_ctx tmpl; mv_cesa_ecb_des3_ede_encrypt() local 601 struct mv_cesa_op_ctx tmpl; mv_cesa_ecb_des3_ede_decrypt() local 632 mv_cesa_cbc_des3_op(struct skcipher_request *req, struct mv_cesa_op_ctx *tmpl) mv_cesa_cbc_des3_op() argument 642 struct mv_cesa_op_ctx tmpl; mv_cesa_cbc_des3_ede_encrypt() local 654 struct mv_cesa_op_ctx tmpl; mv_cesa_cbc_des3_ede_decrypt() local 686 mv_cesa_aes_op(struct skcipher_request *req, struct mv_cesa_op_ctx *tmpl) mv_cesa_aes_op() argument 718 struct mv_cesa_op_ctx tmpl; mv_cesa_ecb_aes_encrypt() local 729 struct mv_cesa_op_ctx tmpl; mv_cesa_ecb_aes_decrypt() local 759 mv_cesa_cbc_aes_op(struct skcipher_request *req, struct mv_cesa_op_ctx *tmpl) mv_cesa_cbc_aes_op() argument 771 struct mv_cesa_op_ctx tmpl; mv_cesa_cbc_aes_encrypt() local 780 struct mv_cesa_op_ctx tmpl; mv_cesa_cbc_aes_decrypt() local [all...] |
H A D | hash.c | 417 struct mv_cesa_op_ctx *tmpl, bool algo_le) in mv_cesa_ahash_init() 422 mv_cesa_update_op_cfg(tmpl, in mv_cesa_ahash_init() 427 mv_cesa_set_mac_op_total_len(tmpl, 0); in mv_cesa_ahash_init() 428 mv_cesa_set_mac_op_frag_len(tmpl, 0); in mv_cesa_ahash_init() 429 creq->op_tmpl = *tmpl; in mv_cesa_ahash_init() 469 struct mv_cesa_op_ctx *tmpl, unsigned int frag_len, in mv_cesa_dma_add_frag() 475 op = mv_cesa_dma_add_op(chain, tmpl, false, flags); in mv_cesa_dma_add_frag() 487 if (mv_cesa_mac_op_is_first_frag(tmpl)) in mv_cesa_dma_add_frag() 488 mv_cesa_update_op_cfg(tmpl, in mv_cesa_dma_add_frag() 795 struct mv_cesa_op_ctx *tmpl in mv_cesa_ahash_final() local 416 mv_cesa_ahash_init(struct ahash_request *req, struct mv_cesa_op_ctx *tmpl, bool algo_le) mv_cesa_ahash_init() argument 468 mv_cesa_dma_add_frag(struct mv_cesa_tdma_chain *chain, struct mv_cesa_op_ctx *tmpl, unsigned int frag_len, gfp_t flags) mv_cesa_dma_add_frag() argument 807 struct mv_cesa_op_ctx *tmpl = &creq->op_tmpl; mv_cesa_ahash_finup() local 871 struct mv_cesa_op_ctx tmpl = { }; mv_cesa_md5_init() local 941 struct mv_cesa_op_ctx tmpl = { }; mv_cesa_sha1_init() local 1012 struct mv_cesa_op_ctx tmpl = { }; mv_cesa_sha256_init() local 1250 struct mv_cesa_op_ctx tmpl = { }; mv_cesa_ahmac_md5_init() local 1321 struct mv_cesa_op_ctx tmpl = { }; mv_cesa_ahmac_sha1_init() local 1412 struct mv_cesa_op_ctx tmpl = { }; mv_cesa_ahmac_sha256_init() local [all...] |
/kernel/linux/linux-6.6/drivers/crypto/marvell/cesa/ |
H A D | cipher.c | 423 struct mv_cesa_op_ctx *tmpl) in mv_cesa_skcipher_req_init() 444 mv_cesa_update_op_cfg(tmpl, CESA_SA_DESC_CFG_OP_CRYPT_ONLY, in mv_cesa_skcipher_req_init() 448 ret = mv_cesa_skcipher_dma_req_init(req, tmpl); in mv_cesa_skcipher_req_init() 450 ret = mv_cesa_skcipher_std_req_init(req, tmpl); in mv_cesa_skcipher_req_init() 456 struct mv_cesa_op_ctx *tmpl) in mv_cesa_skcipher_queue_req() 462 ret = mv_cesa_skcipher_req_init(req, tmpl); in mv_cesa_skcipher_queue_req() 478 struct mv_cesa_op_ctx *tmpl) in mv_cesa_des_op() 482 mv_cesa_update_op_cfg(tmpl, CESA_SA_DESC_CFG_CRYPTM_DES, in mv_cesa_des_op() 485 memcpy(tmpl->ctx.skcipher.key, ctx->key, DES_KEY_SIZE); in mv_cesa_des_op() 487 return mv_cesa_skcipher_queue_req(req, tmpl); in mv_cesa_des_op() 422 mv_cesa_skcipher_req_init(struct skcipher_request *req, struct mv_cesa_op_ctx *tmpl) mv_cesa_skcipher_req_init() argument 455 mv_cesa_skcipher_queue_req(struct skcipher_request *req, struct mv_cesa_op_ctx *tmpl) mv_cesa_skcipher_queue_req() argument 477 mv_cesa_des_op(struct skcipher_request *req, struct mv_cesa_op_ctx *tmpl) mv_cesa_des_op() argument 492 struct mv_cesa_op_ctx tmpl; mv_cesa_ecb_des_encrypt() local 503 struct mv_cesa_op_ctx tmpl; mv_cesa_ecb_des_decrypt() local 533 mv_cesa_cbc_des_op(struct skcipher_request *req, struct mv_cesa_op_ctx *tmpl) mv_cesa_cbc_des_op() argument 546 struct mv_cesa_op_ctx tmpl; mv_cesa_cbc_des_encrypt() local 555 struct mv_cesa_op_ctx tmpl; mv_cesa_cbc_des_decrypt() local 584 mv_cesa_des3_op(struct skcipher_request *req, struct mv_cesa_op_ctx *tmpl) mv_cesa_des3_op() argument 599 struct mv_cesa_op_ctx tmpl; mv_cesa_ecb_des3_ede_encrypt() local 611 struct mv_cesa_op_ctx tmpl; mv_cesa_ecb_des3_ede_decrypt() local 642 mv_cesa_cbc_des3_op(struct skcipher_request *req, struct mv_cesa_op_ctx *tmpl) mv_cesa_cbc_des3_op() argument 652 struct mv_cesa_op_ctx tmpl; mv_cesa_cbc_des3_ede_encrypt() local 664 struct mv_cesa_op_ctx tmpl; mv_cesa_cbc_des3_ede_decrypt() local 696 mv_cesa_aes_op(struct skcipher_request *req, struct mv_cesa_op_ctx *tmpl) mv_cesa_aes_op() argument 728 struct mv_cesa_op_ctx tmpl; mv_cesa_ecb_aes_encrypt() local 739 struct mv_cesa_op_ctx tmpl; mv_cesa_ecb_aes_decrypt() local 769 mv_cesa_cbc_aes_op(struct skcipher_request *req, struct mv_cesa_op_ctx *tmpl) mv_cesa_cbc_aes_op() argument 781 struct mv_cesa_op_ctx tmpl; mv_cesa_cbc_aes_encrypt() local 790 struct mv_cesa_op_ctx tmpl; mv_cesa_cbc_aes_decrypt() local [all...] |
H A D | hash.c | 441 struct mv_cesa_op_ctx *tmpl, bool algo_le) in mv_cesa_ahash_init() 446 mv_cesa_update_op_cfg(tmpl, in mv_cesa_ahash_init() 451 mv_cesa_set_mac_op_total_len(tmpl, 0); in mv_cesa_ahash_init() 452 mv_cesa_set_mac_op_frag_len(tmpl, 0); in mv_cesa_ahash_init() 453 creq->op_tmpl = *tmpl; in mv_cesa_ahash_init() 493 struct mv_cesa_op_ctx *tmpl, unsigned int frag_len, in mv_cesa_dma_add_frag() 499 op = mv_cesa_dma_add_op(chain, tmpl, false, flags); in mv_cesa_dma_add_frag() 511 if (mv_cesa_mac_op_is_first_frag(tmpl)) in mv_cesa_dma_add_frag() 512 mv_cesa_update_op_cfg(tmpl, in mv_cesa_dma_add_frag() 819 struct mv_cesa_op_ctx *tmpl in mv_cesa_ahash_final() local 440 mv_cesa_ahash_init(struct ahash_request *req, struct mv_cesa_op_ctx *tmpl, bool algo_le) mv_cesa_ahash_init() argument 492 mv_cesa_dma_add_frag(struct mv_cesa_tdma_chain *chain, struct mv_cesa_op_ctx *tmpl, unsigned int frag_len, gfp_t flags) mv_cesa_dma_add_frag() argument 831 struct mv_cesa_op_ctx *tmpl = &creq->op_tmpl; mv_cesa_ahash_finup() local 895 struct mv_cesa_op_ctx tmpl = { }; mv_cesa_md5_init() local 965 struct mv_cesa_op_ctx tmpl = { }; mv_cesa_sha1_init() local 1036 struct mv_cesa_op_ctx tmpl = { }; mv_cesa_sha256_init() local 1249 struct mv_cesa_op_ctx tmpl = { }; mv_cesa_ahmac_md5_init() local 1320 struct mv_cesa_op_ctx tmpl = { }; mv_cesa_ahmac_sha1_init() local 1411 struct mv_cesa_op_ctx tmpl = { }; mv_cesa_ahmac_sha256_init() local [all...] |
/kernel/linux/linux-5.10/drivers/crypto/qce/ |
H A D | sha.c | 36 struct qce_alg_template *tmpl = to_ahash_tmpl(async_req->tfm); in qce_ahash_done() local 37 struct qce_device *qce = tmpl->qce; in qce_ahash_done() 66 qce->async_req_done(tmpl->qce, error); in qce_ahash_done() 74 struct qce_alg_template *tmpl = to_ahash_tmpl(async_req->tfm); in qce_ahash_async_req_handle() local 75 struct qce_device *qce = tmpl->qce; in qce_ahash_async_req_handle() 110 ret = qce_start(async_req, tmpl->crypto_alg_type, 0, 0); in qce_ahash_async_req_handle() 128 struct qce_alg_template *tmpl = to_ahash_tmpl(req->base.tfm); in qce_ahash_init() local 129 const u32 *std_iv = tmpl->std_iv; in qce_ahash_init() 134 rctx->flags = tmpl->alg_flags; in qce_ahash_init() 239 struct qce_alg_template *tmpl in qce_ahash_update() local 312 struct qce_alg_template *tmpl = to_ahash_tmpl(req->base.tfm); qce_ahash_final() local 339 struct qce_alg_template *tmpl = to_ahash_tmpl(req->base.tfm); qce_ahash_digest() local 488 struct qce_alg_template *tmpl; qce_ahash_register_one() local 548 struct qce_alg_template *tmpl, *n; qce_ahash_unregister() local [all...] |
H A D | skcipher.c | 31 struct qce_alg_template *tmpl = to_cipher_tmpl(crypto_skcipher_reqtfm(req)); in qce_skcipher_done() local 32 struct qce_device *qce = tmpl->qce; in qce_skcipher_done() 59 qce->async_req_done(tmpl->qce, error); in qce_skcipher_done() 68 struct qce_alg_template *tmpl = to_cipher_tmpl(crypto_skcipher_reqtfm(req)); in qce_skcipher_async_req_handle() local 69 struct qce_device *qce = tmpl->qce; in qce_skcipher_async_req_handle() 151 ret = qce_start(async_req, tmpl->crypto_alg_type, req->cryptlen, 0); in qce_skcipher_async_req_handle() 228 struct qce_alg_template *tmpl = to_cipher_tmpl(tfm); in qce_skcipher_crypt() local 232 rctx->flags = tmpl->alg_flags; in qce_skcipher_crypt() 256 return tmpl->qce->async_req_enqueue(tmpl in qce_skcipher_crypt() 388 struct qce_alg_template *tmpl; qce_skcipher_register_one() local 448 struct qce_alg_template *tmpl, *n; qce_skcipher_unregister() local [all...] |
/kernel/linux/linux-6.6/drivers/crypto/qce/ |
H A D | sha.c | 42 struct qce_alg_template *tmpl = to_ahash_tmpl(async_req->tfm); in qce_ahash_done() local 43 struct qce_device *qce = tmpl->qce; in qce_ahash_done() 72 qce->async_req_done(tmpl->qce, error); in qce_ahash_done() 80 struct qce_alg_template *tmpl = to_ahash_tmpl(async_req->tfm); in qce_ahash_async_req_handle() local 81 struct qce_device *qce = tmpl->qce; in qce_ahash_async_req_handle() 118 ret = qce_start(async_req, tmpl->crypto_alg_type); in qce_ahash_async_req_handle() 136 struct qce_alg_template *tmpl = to_ahash_tmpl(req->base.tfm); in qce_ahash_init() local 137 const u32 *std_iv = tmpl->std_iv; in qce_ahash_init() 142 rctx->flags = tmpl->alg_flags; in qce_ahash_init() 187 struct qce_alg_template *tmpl in qce_ahash_update() local 279 struct qce_alg_template *tmpl = to_ahash_tmpl(req->base.tfm); qce_ahash_final() local 306 struct qce_alg_template *tmpl = to_ahash_tmpl(req->base.tfm); qce_ahash_digest() local 455 struct qce_alg_template *tmpl; qce_ahash_register_one() local 515 struct qce_alg_template *tmpl, *n; qce_ahash_unregister() local [all...] |
H A D | skcipher.c | 32 struct qce_alg_template *tmpl = to_cipher_tmpl(crypto_skcipher_reqtfm(req)); in qce_skcipher_done() local 33 struct qce_device *qce = tmpl->qce; in qce_skcipher_done() 60 qce->async_req_done(tmpl->qce, error); in qce_skcipher_done() 69 struct qce_alg_template *tmpl = to_cipher_tmpl(crypto_skcipher_reqtfm(req)); in qce_skcipher_async_req_handle() local 70 struct qce_device *qce = tmpl->qce; in qce_skcipher_async_req_handle() 152 ret = qce_start(async_req, tmpl->crypto_alg_type); in qce_skcipher_async_req_handle() 261 struct qce_alg_template *tmpl = to_cipher_tmpl(tfm); in qce_skcipher_crypt() local 266 rctx->flags = tmpl->alg_flags; in qce_skcipher_crypt() 307 return tmpl->qce->async_req_enqueue(tmpl in qce_skcipher_crypt() 439 struct qce_alg_template *tmpl; qce_skcipher_register_one() local 499 struct qce_alg_template *tmpl, *n; qce_skcipher_unregister() local [all...] |
H A D | aead.c | 29 struct qce_alg_template *tmpl = to_aead_tmpl(crypto_aead_reqtfm(req)); in qce_aead_done() local 30 struct qce_device *qce = tmpl->qce; in qce_aead_done() 96 struct qce_alg_template *tmpl = to_aead_tmpl(crypto_aead_reqtfm(req)); in qce_aead_prepare_result_buf() local 97 struct qce_device *qce = tmpl->qce; in qce_aead_prepare_result_buf() 116 struct qce_alg_template *tmpl = to_aead_tmpl(crypto_aead_reqtfm(req)); in qce_aead_prepare_dst_buf() local 117 struct qce_device *qce = tmpl->qce; in qce_aead_prepare_dst_buf() 304 struct qce_alg_template *tmpl = to_aead_tmpl(crypto_aead_reqtfm(req)); in qce_aead_prepare_buf() local 305 struct qce_device *qce = tmpl->qce; in qce_aead_prepare_buf() 414 struct qce_alg_template *tmpl = to_aead_tmpl(crypto_aead_reqtfm(req)); in qce_aead_async_req_handle() local 415 struct qce_device *qce = tmpl in qce_aead_async_req_handle() 507 struct qce_alg_template *tmpl = to_aead_tmpl(tfm); qce_aead_crypt() local 761 struct qce_alg_template *tmpl; qce_aead_register_one() local 817 struct qce_alg_template *tmpl, *n; qce_aead_unregister() local [all...] |
/kernel/linux/linux-5.10/crypto/ |
H A D | xor.c | 58 static void __init do_xor_register(struct xor_block_template *tmpl) in do_xor_register() argument 60 tmpl->next = template_list; in do_xor_register() 61 template_list = tmpl; in do_xor_register() 83 do_xor_speed(struct xor_block_template *tmpl, void *b1, void *b2) in do_xor_speed() argument 89 tmpl->next = template_list; in do_xor_speed() 90 template_list = tmpl; in do_xor_speed() 99 tmpl->do_2(BENCH_SIZE, b1, b2); in do_xor_speed() 113 tmpl->speed = speed; in do_xor_speed() 115 pr_info(" %-16s: %5d MB/sec\n", tmpl->name, speed); in do_xor_speed()
|
H A D | algboss.c | 60 struct crypto_template *tmpl; in cryptomgr_probe() local 63 tmpl = crypto_lookup_template(param->template); in cryptomgr_probe() 64 if (!tmpl) in cryptomgr_probe() 68 err = tmpl->create(tmpl, param->tb); in cryptomgr_probe() 71 crypto_tmpl_put(tmpl); in cryptomgr_probe()
|
/kernel/linux/linux-6.6/crypto/ |
H A D | xor.c | 58 static void __init do_xor_register(struct xor_block_template *tmpl) in do_xor_register() argument 60 tmpl->next = template_list; in do_xor_register() 61 template_list = tmpl; in do_xor_register() 83 do_xor_speed(struct xor_block_template *tmpl, void *b1, void *b2) in do_xor_speed() argument 89 tmpl->next = template_list; in do_xor_speed() 90 template_list = tmpl; in do_xor_speed() 99 tmpl->do_2(BENCH_SIZE, b1, b2); in do_xor_speed() 113 tmpl->speed = speed; in do_xor_speed() 115 pr_info(" %-16s: %5d MB/sec\n", tmpl->name, speed); in do_xor_speed()
|
H A D | algapi.c | 82 struct crypto_template *tmpl = inst->tmpl; in crypto_destroy_instance_workfn() local 85 crypto_tmpl_put(tmpl); in crypto_destroy_instance_workfn() 134 struct crypto_template *tmpl = inst->tmpl; in crypto_remove_instance() local 141 if (!tmpl || !crypto_tmpl_get(tmpl)) in crypto_remove_instance() 546 int crypto_register_template(struct crypto_template *tmpl) in crypto_register_template() argument 553 crypto_check_module_sig(tmpl->module); in crypto_register_template() 556 if (q == tmpl) in crypto_register_template() 586 crypto_unregister_template(struct crypto_template *tmpl) crypto_unregister_template() argument 626 struct crypto_template *q, *tmpl = NULL; __crypto_lookup_template() local 650 crypto_register_instance(struct crypto_template *tmpl, struct crypto_instance *inst) crypto_register_instance() argument [all...] |
H A D | dh.c | 551 struct crypto_template *tmpl, struct rtattr **tb, in __dh_safe_prime_create() 588 tmpl->name, &dh_alg->base); in __dh_safe_prime_create() 604 err = kpp_register_instance(tmpl, inst); in __dh_safe_prime_create() 832 static int dh_ffdhe2048_create(struct crypto_template *tmpl, in dh_ffdhe2048_create() argument 835 return __dh_safe_prime_create(tmpl, tb, &ffdhe2048_prime); in dh_ffdhe2048_create() 838 static int dh_ffdhe3072_create(struct crypto_template *tmpl, in dh_ffdhe3072_create() argument 841 return __dh_safe_prime_create(tmpl, tb, &ffdhe3072_prime); in dh_ffdhe3072_create() 844 static int dh_ffdhe4096_create(struct crypto_template *tmpl, in dh_ffdhe4096_create() argument 847 return __dh_safe_prime_create(tmpl, tb, &ffdhe4096_prime); in dh_ffdhe4096_create() 850 static int dh_ffdhe6144_create(struct crypto_template *tmpl, in dh_ffdhe6144_create() argument 550 __dh_safe_prime_create( struct crypto_template *tmpl, struct rtattr **tb, const struct dh_safe_prime *safe_prime) __dh_safe_prime_create() argument 856 dh_ffdhe8192_create(struct crypto_template *tmpl, struct rtattr **tb) dh_ffdhe8192_create() argument [all...] |
/kernel/linux/linux-5.10/drivers/scsi/fnic/ |
H A D | cq_exch_desc.h | 28 u8 tmpl; member 62 u16 tmpl; member 95 u16 *tmpl, in cq_fcp_rq_desc_dec() 116 *tmpl = desc_ptr->tmpl & CQ_FCP_RQ_DESC_TMPL_MASK; in cq_fcp_rq_desc_dec() 141 u16 tmpl; member 169 u16 *tmpl, in cq_sgl_desc_dec() 178 *tmpl = desc_ptr->tmpl & CQ_SGL_TMPL_MASK; in cq_sgl_desc_dec() 86 cq_fcp_rq_desc_dec(struct cq_fcp_rq_desc *desc_ptr, u8 *type, u8 *color, u16 *q_number, u16 *completed_index, u8 *eop, u8 *sop, u8 *fck, u16 *exchange_id, u16 *tmpl, u32 *bytes_written, u8 *sof, u8 *eof, u8 *ingress_port, u8 *packet_err, u8 *fcoe_err, u8 *fcs_ok, u8 *vlan_stripped, u16 *vlan) cq_fcp_rq_desc_dec() argument 162 cq_sgl_desc_dec(struct cq_sgl_desc *desc_ptr, u8 *type, u8 *color, u16 *q_number, u16 *exchange_id, u32 *active_burst_offset, u32 *tot_data_bytes, u16 *tmpl, u8 *sgl_err) cq_sgl_desc_dec() argument
|
/kernel/linux/linux-6.6/drivers/scsi/fnic/ |
H A D | cq_exch_desc.h | 16 u8 tmpl; member 50 u16 tmpl; member 83 u16 *tmpl, in cq_fcp_rq_desc_dec() 104 *tmpl = desc_ptr->tmpl & CQ_FCP_RQ_DESC_TMPL_MASK; in cq_fcp_rq_desc_dec() 129 u16 tmpl; member 157 u16 *tmpl, in cq_sgl_desc_dec() 166 *tmpl = desc_ptr->tmpl & CQ_SGL_TMPL_MASK; in cq_sgl_desc_dec() 74 cq_fcp_rq_desc_dec(struct cq_fcp_rq_desc *desc_ptr, u8 *type, u8 *color, u16 *q_number, u16 *completed_index, u8 *eop, u8 *sop, u8 *fck, u16 *exchange_id, u16 *tmpl, u32 *bytes_written, u8 *sof, u8 *eof, u8 *ingress_port, u8 *packet_err, u8 *fcoe_err, u8 *fcs_ok, u8 *vlan_stripped, u16 *vlan) cq_fcp_rq_desc_dec() argument 150 cq_sgl_desc_dec(struct cq_sgl_desc *desc_ptr, u8 *type, u8 *color, u16 *q_number, u16 *exchange_id, u32 *active_burst_offset, u32 *tot_data_bytes, u16 *tmpl, u8 *sgl_err) cq_sgl_desc_dec() argument
|
/kernel/linux/linux-5.10/include/net/netfilter/ |
H A D | nf_conntrack_zones.h | 29 nf_ct_zone_tmpl(const struct nf_conn *tmpl, const struct sk_buff *skb, in nf_ct_zone_tmpl() argument 33 if (!tmpl) in nf_ct_zone_tmpl() 36 if (tmpl->zone.flags & NF_CT_FLAG_MARK) in nf_ct_zone_tmpl() 37 return nf_ct_zone_init(tmp, skb->mark, tmpl->zone.dir, 0); in nf_ct_zone_tmpl() 39 return nf_ct_zone(tmpl); in nf_ct_zone_tmpl()
|
H A D | nf_conntrack_synproxy.h | 33 const struct nf_conn *tmpl) in nf_ct_add_synproxy() 36 if (tmpl && nfct_synproxy(tmpl)) { in nf_ct_add_synproxy() 32 nf_ct_add_synproxy(struct nf_conn *ct, const struct nf_conn *tmpl) nf_ct_add_synproxy() argument
|
/kernel/linux/linux-6.6/include/net/netfilter/ |
H A D | nf_conntrack_zones.h | 29 nf_ct_zone_tmpl(const struct nf_conn *tmpl, const struct sk_buff *skb, in nf_ct_zone_tmpl() argument 33 if (!tmpl) in nf_ct_zone_tmpl() 36 if (tmpl->zone.flags & NF_CT_FLAG_MARK) in nf_ct_zone_tmpl() 37 return nf_ct_zone_init(tmp, skb->mark, tmpl->zone.dir, 0); in nf_ct_zone_tmpl() 39 return nf_ct_zone(tmpl); in nf_ct_zone_tmpl()
|
H A D | nf_conntrack_synproxy.h | 33 const struct nf_conn *tmpl) in nf_ct_add_synproxy() 36 if (tmpl && nfct_synproxy(tmpl)) { in nf_ct_add_synproxy() 32 nf_ct_add_synproxy(struct nf_conn *ct, const struct nf_conn *tmpl) nf_ct_add_synproxy() argument
|
/kernel/linux/linux-5.10/drivers/s390/net/ |
H A D | fsm.c | 18 int nr_events, const fsm_node *tmpl, int tmpl_len, gfp_t order) in init_fsm() 57 if ((tmpl[i].cond_state >= nr_states) || in init_fsm() 58 (tmpl[i].cond_event >= nr_events) ) { in init_fsm() 61 name, i, (long)tmpl[i].cond_state, (long)f->nr_states, in init_fsm() 62 (long)tmpl[i].cond_event, (long)f->nr_events); in init_fsm() 66 m[nr_states * tmpl[i].cond_event + tmpl[i].cond_state] = in init_fsm() 67 tmpl[i].function; in init_fsm() 17 init_fsm(char *name, const char **state_names, const char **event_names, int nr_states, int nr_events, const fsm_node *tmpl, int tmpl_len, gfp_t order) init_fsm() argument
|
/kernel/linux/linux-6.6/drivers/s390/net/ |
H A D | fsm.c | 18 int nr_events, const fsm_node *tmpl, int tmpl_len, gfp_t order) in init_fsm() 57 if ((tmpl[i].cond_state >= nr_states) || in init_fsm() 58 (tmpl[i].cond_event >= nr_events) ) { in init_fsm() 61 name, i, (long)tmpl[i].cond_state, (long)f->nr_states, in init_fsm() 62 (long)tmpl[i].cond_event, (long)f->nr_events); in init_fsm() 66 m[nr_states * tmpl[i].cond_event + tmpl[i].cond_state] = in init_fsm() 67 tmpl[i].function; in init_fsm() 17 init_fsm(char *name, const char **state_names, const char **event_names, int nr_states, int nr_events, const fsm_node *tmpl, int tmpl_len, gfp_t order) init_fsm() argument
|
/kernel/linux/linux-5.10/net/netfilter/ |
H A D | nf_conntrack_core.c | 545 struct nf_conn *tmpl, *p; in nf_ct_tmpl_alloc() local 548 tmpl = kzalloc(sizeof(*tmpl) + NFCT_INFOMASK, flags); in nf_ct_tmpl_alloc() 549 if (!tmpl) in nf_ct_tmpl_alloc() 552 p = tmpl; in nf_ct_tmpl_alloc() 553 tmpl = (struct nf_conn *)NFCT_ALIGN((unsigned long)p); in nf_ct_tmpl_alloc() 554 if (tmpl != p) { in nf_ct_tmpl_alloc() 555 tmpl = (struct nf_conn *)NFCT_ALIGN((unsigned long)p); in nf_ct_tmpl_alloc() 556 tmpl->proto.tmpl_padto = (char *)tmpl in nf_ct_tmpl_alloc() 573 nf_ct_tmpl_free(struct nf_conn *tmpl) nf_ct_tmpl_free() argument 1538 init_conntrack(struct net *net, struct nf_conn *tmpl, const struct nf_conntrack_tuple *tuple, struct sk_buff *skb, unsigned int dataoff, u32 hash) init_conntrack() argument 1630 resolve_normal_ct(struct nf_conn *tmpl, struct sk_buff *skb, unsigned int dataoff, u_int8_t protonum, const struct nf_hook_state *state) resolve_normal_ct() argument 1693 nf_conntrack_handle_icmp(struct nf_conn *tmpl, struct sk_buff *skb, unsigned int dataoff, u8 protonum, const struct nf_hook_state *state) nf_conntrack_handle_icmp() argument 1777 struct nf_conn *ct, *tmpl; nf_conntrack_in() local [all...] |
H A D | nft_dynset.c | 20 struct nft_set_ext_tmpl tmpl; member 42 elem = nft_set_elem_init(set, &priv->tmpl, in nft_dynset_new() 197 nft_set_ext_prepare(&priv->tmpl); in nft_dynset_init() 198 nft_set_ext_add_length(&priv->tmpl, NFT_SET_EXT_KEY, set->klen); in nft_dynset_init() 200 nft_set_ext_add_length(&priv->tmpl, NFT_SET_EXT_DATA, set->dlen); in nft_dynset_init() 202 nft_set_ext_add_length(&priv->tmpl, NFT_SET_EXT_EXPR, in nft_dynset_init() 206 nft_set_ext_add(&priv->tmpl, NFT_SET_EXT_TIMEOUT); in nft_dynset_init() 207 nft_set_ext_add(&priv->tmpl, NFT_SET_EXT_EXPIRATION); in nft_dynset_init()
|
/kernel/linux/linux-6.6/net/netfilter/ |
H A D | nf_conntrack_core.c | 525 struct nf_conn *tmpl, *p; in nf_ct_tmpl_alloc() local 528 tmpl = kzalloc(sizeof(*tmpl) + NFCT_INFOMASK, flags); in nf_ct_tmpl_alloc() 529 if (!tmpl) in nf_ct_tmpl_alloc() 532 p = tmpl; in nf_ct_tmpl_alloc() 533 tmpl = (struct nf_conn *)NFCT_ALIGN((unsigned long)p); in nf_ct_tmpl_alloc() 534 if (tmpl != p) { in nf_ct_tmpl_alloc() 535 tmpl = (struct nf_conn *)NFCT_ALIGN((unsigned long)p); in nf_ct_tmpl_alloc() 536 tmpl->proto.tmpl_padto = (char *)tmpl in nf_ct_tmpl_alloc() 553 nf_ct_tmpl_free(struct nf_conn *tmpl) nf_ct_tmpl_free() argument 1703 init_conntrack(struct net *net, struct nf_conn *tmpl, const struct nf_conntrack_tuple *tuple, struct sk_buff *skb, unsigned int dataoff, u32 hash) init_conntrack() argument 1808 resolve_normal_ct(struct nf_conn *tmpl, struct sk_buff *skb, unsigned int dataoff, u_int8_t protonum, const struct nf_hook_state *state) resolve_normal_ct() argument 1879 nf_conntrack_handle_icmp(struct nf_conn *tmpl, struct sk_buff *skb, unsigned int dataoff, u8 protonum, const struct nf_hook_state *state) nf_conntrack_handle_icmp() argument 1963 struct nf_conn *ct, *tmpl; nf_conntrack_in() local [all...] |