/kernel/linux/linux-5.10/drivers/soc/fsl/qe/ |
H A D | qe.c | 29 #include <soc/fsl/qe/immap_qe.h> 30 #include <soc/fsl/qe/qe.h> 53 struct device_node *qe; in qe_get_device_node() local 56 * Newer device trees have an "fsl,qe" compatible property for the QE in qe_get_device_node() 59 qe = of_find_compatible_node(NULL, NULL, "fsl,qe"); in qe_get_device_node() 60 if (qe) in qe_get_device_node() 61 return qe; in qe_get_device_node() 62 return of_find_node_by_type(NULL, "qe"); in qe_get_device_node() 67 struct device_node *qe; get_qe_base() local 161 struct device_node *qe; qe_get_brg_clk() local 303 struct device_node *qe; qe_snums_init() local 569 struct device_node *qe; qe_get_firmware_info() local 619 struct device_node *qe; qe_get_num_of_risc() local [all...] |
/kernel/linux/linux-6.6/drivers/soc/fsl/qe/ |
H A D | qe.c | 30 #include <soc/fsl/qe/immap_qe.h> 31 #include <soc/fsl/qe/qe.h> 54 struct device_node *qe; in qe_get_device_node() local 57 * Newer device trees have an "fsl,qe" compatible property for the QE in qe_get_device_node() 60 qe = of_find_compatible_node(NULL, NULL, "fsl,qe"); in qe_get_device_node() 61 if (qe) in qe_get_device_node() 62 return qe; in qe_get_device_node() 63 return of_find_node_by_type(NULL, "qe"); in qe_get_device_node() 68 struct device_node *qe; get_qe_base() local 162 struct device_node *qe; qe_get_brg_clk() local 304 struct device_node *qe; qe_snums_init() local 570 struct device_node *qe; qe_get_firmware_info() local 620 struct device_node *qe; qe_get_num_of_risc() local [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/chelsio/cxgb4/ |
H A D | sched.c | 82 struct sched_queue_entry *qe; in t4_sched_bind_unbind_op() local 84 qe = (struct sched_queue_entry *)arg; in t4_sched_bind_unbind_op() 92 fw_class = bind ? qe->param.class : FW_SCHED_CLS_NONE; in t4_sched_bind_unbind_op() 93 fw_param = (fw_mnem | FW_PARAMS_PARAM_YZ_V(qe->cntxt_id)); in t4_sched_bind_unbind_op() 137 struct sched_queue_entry *qe; in t4_sched_entry_lookup() local 139 list_for_each_entry(qe, &e->entry_list, list) { in t4_sched_entry_lookup() 140 if (qe->cntxt_id == val) { in t4_sched_entry_lookup() 141 found = qe; in t4_sched_entry_lookup() 173 struct sched_queue_entry *qe = NULL; in cxgb4_sched_queue_lookup() local 181 qe in cxgb4_sched_queue_lookup() 187 struct sched_queue_entry *qe = NULL; t4_sched_queue_unbind() local 218 struct sched_queue_entry *qe = NULL; t4_sched_queue_bind() local 333 struct sched_queue_entry *qe; t4_sched_class_unbind_all() local 361 struct ch_sched_queue *qe = (struct ch_sched_queue *)arg; t4_sched_class_bind_unbind_op() local 410 struct ch_sched_queue *qe = (struct ch_sched_queue *)arg; cxgb4_sched_class_bind() local 457 struct ch_sched_queue *qe = (struct ch_sched_queue *)arg; cxgb4_sched_class_unbind() local [all...] |
H A D | cxgb4_tc_matchall.c | 18 struct ch_sched_queue qe; in cxgb4_matchall_egress_validate() local 66 memset(&qe, 0, sizeof(qe)); in cxgb4_matchall_egress_validate() 67 qe.queue = i; in cxgb4_matchall_egress_validate() 69 e = cxgb4_sched_queue_lookup(dev, &qe); in cxgb4_matchall_egress_validate() 83 struct ch_sched_queue qe; in cxgb4_matchall_tc_bind_queues() local 88 qe.queue = i; in cxgb4_matchall_tc_bind_queues() 89 qe.class = tc; in cxgb4_matchall_tc_bind_queues() 90 ret = cxgb4_sched_class_bind(dev, &qe, SCHED_QUEUE); in cxgb4_matchall_tc_bind_queues() 99 qe in cxgb4_matchall_tc_bind_queues() 110 struct ch_sched_queue qe; cxgb4_matchall_tc_unbind_queues() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/chelsio/cxgb4/ |
H A D | sched.c | 82 struct sched_queue_entry *qe; in t4_sched_bind_unbind_op() local 84 qe = (struct sched_queue_entry *)arg; in t4_sched_bind_unbind_op() 92 fw_class = bind ? qe->param.class : FW_SCHED_CLS_NONE; in t4_sched_bind_unbind_op() 93 fw_param = (fw_mnem | FW_PARAMS_PARAM_YZ_V(qe->cntxt_id)); in t4_sched_bind_unbind_op() 137 struct sched_queue_entry *qe; in t4_sched_entry_lookup() local 139 list_for_each_entry(qe, &e->entry_list, list) { in t4_sched_entry_lookup() 140 if (qe->cntxt_id == val) { in t4_sched_entry_lookup() 141 found = qe; in t4_sched_entry_lookup() 173 struct sched_queue_entry *qe = NULL; in cxgb4_sched_queue_lookup() local 181 qe in cxgb4_sched_queue_lookup() 187 struct sched_queue_entry *qe = NULL; t4_sched_queue_unbind() local 218 struct sched_queue_entry *qe = NULL; t4_sched_queue_bind() local 333 struct sched_queue_entry *qe; t4_sched_class_unbind_all() local 361 struct ch_sched_queue *qe = (struct ch_sched_queue *)arg; t4_sched_class_bind_unbind_op() local 410 struct ch_sched_queue *qe = (struct ch_sched_queue *)arg; cxgb4_sched_class_bind() local 457 struct ch_sched_queue *qe = (struct ch_sched_queue *)arg; cxgb4_sched_class_unbind() local [all...] |
H A D | cxgb4_tc_matchall.c | 58 struct ch_sched_queue qe; in cxgb4_matchall_egress_validate() local 110 memset(&qe, 0, sizeof(qe)); in cxgb4_matchall_egress_validate() 111 qe.queue = i; in cxgb4_matchall_egress_validate() 113 e = cxgb4_sched_queue_lookup(dev, &qe); in cxgb4_matchall_egress_validate() 127 struct ch_sched_queue qe; in cxgb4_matchall_tc_bind_queues() local 132 qe.queue = i; in cxgb4_matchall_tc_bind_queues() 133 qe.class = tc; in cxgb4_matchall_tc_bind_queues() 134 ret = cxgb4_sched_class_bind(dev, &qe, SCHED_QUEUE); in cxgb4_matchall_tc_bind_queues() 143 qe in cxgb4_matchall_tc_bind_queues() 154 struct ch_sched_queue qe; cxgb4_matchall_tc_unbind_queues() local [all...] |
/kernel/linux/linux-5.10/security/integrity/ima/ |
H A D | ima_queue.c | 51 struct ima_queue_entry *qe, *ret = NULL; in ima_lookup_digest_entry() local 57 hlist_for_each_entry_rcu(qe, &ima_htable.queue[key], hnext) { in ima_lookup_digest_entry() 58 rc = memcmp(qe->entry->digests[ima_hash_algo_idx].digest, in ima_lookup_digest_entry() 60 if ((rc == 0) && (qe->entry->pcr == pcr)) { in ima_lookup_digest_entry() 61 ret = qe; in ima_lookup_digest_entry() 96 struct ima_queue_entry *qe; in ima_add_digest_entry() local 99 qe = kmalloc(sizeof(*qe), GFP_KERNEL); in ima_add_digest_entry() 100 if (qe == NULL) { in ima_add_digest_entry() 104 qe in ima_add_digest_entry() [all...] |
H A D | ima_fs.c | 80 struct ima_queue_entry *qe; in ima_measurements_start() local 84 list_for_each_entry_rcu(qe, &ima_measurements, later) { in ima_measurements_start() 87 return qe; in ima_measurements_start() 96 struct ima_queue_entry *qe = v; in ima_measurements_next() local 102 qe = list_entry_rcu(qe->later.next, struct ima_queue_entry, later); in ima_measurements_next() 106 return (&qe->later == &ima_measurements) ? NULL : qe; in ima_measurements_next() 130 struct ima_queue_entry *qe = v; in ima_measurements_show() local 138 e = qe in ima_measurements_show() 220 struct ima_queue_entry *qe = v; ima_ascii_measurements_show() local [all...] |
/kernel/linux/linux-6.6/security/integrity/ima/ |
H A D | ima_queue.c | 51 struct ima_queue_entry *qe, *ret = NULL; in ima_lookup_digest_entry() local 57 hlist_for_each_entry_rcu(qe, &ima_htable.queue[key], hnext) { in ima_lookup_digest_entry() 58 rc = memcmp(qe->entry->digests[ima_hash_algo_idx].digest, in ima_lookup_digest_entry() 60 if ((rc == 0) && (qe->entry->pcr == pcr)) { in ima_lookup_digest_entry() 61 ret = qe; in ima_lookup_digest_entry() 96 struct ima_queue_entry *qe; in ima_add_digest_entry() local 99 qe = kmalloc(sizeof(*qe), GFP_KERNEL); in ima_add_digest_entry() 100 if (qe == NULL) { in ima_add_digest_entry() 104 qe in ima_add_digest_entry() [all...] |
H A D | ima_fs.c | 80 struct ima_queue_entry *qe; in ima_measurements_start() local 84 list_for_each_entry_rcu(qe, &ima_measurements, later) { in ima_measurements_start() 87 return qe; in ima_measurements_start() 96 struct ima_queue_entry *qe = v; in ima_measurements_next() local 102 qe = list_entry_rcu(qe->later.next, struct ima_queue_entry, later); in ima_measurements_next() 106 return (&qe->later == &ima_measurements) ? NULL : qe; in ima_measurements_next() 130 struct ima_queue_entry *qe = v; in ima_measurements_show() local 138 e = qe in ima_measurements_show() 220 struct ima_queue_entry *qe = v; ima_ascii_measurements_show() local [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/sun/ |
H A D | sunqe.c | 465 * so we just run through each qe and check to see who is signaling 655 /* Program the qe with the new filter value. */ in qe_set_multicast() 837 struct sunqe *qe; in qec_ether_init() local 849 qe = netdev_priv(dev); in qec_ether_init() 856 qe->channel = i; in qec_ether_init() 857 spin_lock_init(&qe->lock); in qec_ether_init() 863 qecp->qes[qe->channel] = qe; in qec_ether_init() 864 qe->dev = dev; in qec_ether_init() 865 qe in qec_ether_init() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/sun/ |
H A D | sunqe.c | 465 * so we just run through each qe and check to see who is signaling 655 /* Program the qe with the new filter value. */ in qe_set_multicast() 837 struct sunqe *qe; in qec_ether_init() local 849 qe = netdev_priv(dev); in qec_ether_init() 856 qe->channel = i; in qec_ether_init() 857 spin_lock_init(&qe->lock); in qec_ether_init() 863 qecp->qes[qe->channel] = qe; in qec_ether_init() 864 qe->dev = dev; in qec_ether_init() 865 qe in qec_ether_init() [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/brocade/bna/ |
H A D | bna_tx_rx.c | 352 list_for_each_entry(mac, &rxf->mcast_active_q, qe) in bna_rxf_mcmac_get() 356 list_for_each_entry(mac, &rxf->mcast_pending_del_q, qe) in bna_rxf_mcmac_get() 368 list_for_each_entry(mchandle, &rxf->mcast_handle_q, qe) in bna_rxf_mchandle_get() 387 list_add_tail(&mchandle->qe, &rxf->mcast_handle_q); in bna_rxf_mchandle_attach() 410 list_del(&mchandle->qe); in bna_rxf_mcast_del() 427 struct bna_mac, qe); in bna_rxf_mcast_cfg_apply() 429 list_move_tail(&mac->qe, bna_mcam_mod_del_q(rxf->rx->bna)); in bna_rxf_mcast_cfg_apply() 437 struct bna_mac, qe); in bna_rxf_mcast_cfg_apply() 438 list_move_tail(&mac->qe, &rxf->mcast_active_q); in bna_rxf_mcast_cfg_apply() 475 struct bna_mac, qe); in bna_rxf_mcast_cfg_reset() 1820 struct list_head *qe; bna_rx_put() local 2489 struct list_head *qe; bna_rx_destroy() local 3218 struct list_head *qe; bna_tx_free() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/brocade/bna/ |
H A D | bna_tx_rx.c | 352 list_for_each_entry(mac, &rxf->mcast_active_q, qe) in bna_rxf_mcmac_get() 356 list_for_each_entry(mac, &rxf->mcast_pending_del_q, qe) in bna_rxf_mcmac_get() 368 list_for_each_entry(mchandle, &rxf->mcast_handle_q, qe) in bna_rxf_mchandle_get() 387 list_add_tail(&mchandle->qe, &rxf->mcast_handle_q); in bna_rxf_mchandle_attach() 410 list_del(&mchandle->qe); in bna_rxf_mcast_del() 427 struct bna_mac, qe); in bna_rxf_mcast_cfg_apply() 429 list_move_tail(&mac->qe, bna_mcam_mod_del_q(rxf->rx->bna)); in bna_rxf_mcast_cfg_apply() 437 struct bna_mac, qe); in bna_rxf_mcast_cfg_apply() 438 list_move_tail(&mac->qe, &rxf->mcast_active_q); in bna_rxf_mcast_cfg_apply() 475 struct bna_mac, qe); in bna_rxf_mcast_cfg_reset() 1820 struct list_head *qe; bna_rx_put() local 2489 struct list_head *qe; bna_rx_destroy() local 3218 struct list_head *qe; bna_tx_free() local [all...] |
/kernel/linux/linux-5.10/drivers/scsi/arm/ |
H A D | msgqueue.c | 55 msgq->qe = NULL; in msgqueue_initialise() 82 struct msgqueue_entry *mq = msgq->qe; in msgqueue_msglength() 85 for (mq = msgq->qe; mq; mq = mq->next) in msgqueue_msglength() 102 for (mq = msgq->qe; mq && msgno; mq = mq->next, msgno--); in msgqueue_getmsg() 133 mqp = &msgq->qe; in msgqueue_addmsg() 152 for (mq = msgq->qe; mq; mq = mqnext) { in msgqueue_flush() 156 msgq->qe = NULL; in msgqueue_flush()
|
/kernel/linux/linux-6.6/drivers/scsi/arm/ |
H A D | msgqueue.c | 55 msgq->qe = NULL; in msgqueue_initialise() 82 struct msgqueue_entry *mq = msgq->qe; in msgqueue_msglength() 85 for (mq = msgq->qe; mq; mq = mq->next) in msgqueue_msglength() 102 for (mq = msgq->qe; mq && msgno; mq = mq->next, msgno--); in msgqueue_getmsg() 133 mqp = &msgq->qe; in msgqueue_addmsg() 152 for (mq = msgq->qe; mq; mq = mqnext) { in msgqueue_flush() 156 msgq->qe = NULL; in msgqueue_flush()
|
/kernel/linux/linux-5.10/drivers/atm/ |
H A D | firestream.c | 583 static void submit_qentry (struct fs_dev *dev, struct queue *q, struct FS_QENTRY *qe) in submit_qentry() argument 589 still the same as the value passed as qe... -- REW */ in submit_qentry() 599 if (qe != cqe) { in submit_qentry() 600 fs_dprintk (FS_DEBUG_TXQ, "q mismatch! %p %p\n", qe, cqe); in submit_qentry() 630 struct FS_QENTRY *qe; in submit_queue() local 632 qe = get_qentry (dev, q); in submit_queue() 633 qe->cmd = cmd; in submit_queue() 634 qe->p0 = p1; in submit_queue() 635 qe->p1 = p2; in submit_queue() 636 qe in submit_queue() 669 struct FS_QENTRY *qe; process_return_queue() local 696 struct FS_QENTRY *qe; process_txdone_queue() local 765 struct FS_QENTRY *qe; process_incoming() local 1468 struct FS_BPENTRY *qe, *ne; top_off_fp() local [all...] |
/kernel/linux/linux-5.10/include/soc/fsl/qe/ |
H A D | qe_tdm.h | 16 #include <soc/fsl/qe/immap_qe.h> 17 #include <soc/fsl/qe/qe.h> 19 #include <soc/fsl/qe/ucc.h> 20 #include <soc/fsl/qe/ucc_fast.h>
|
H A D | ucc.h | 14 #include <soc/fsl/qe/immap_qe.h> 15 #include <soc/fsl/qe/qe.h>
|
/kernel/linux/linux-6.6/include/soc/fsl/qe/ |
H A D | qe_tdm.h | 16 #include <soc/fsl/qe/immap_qe.h> 17 #include <soc/fsl/qe/qe.h> 19 #include <soc/fsl/qe/ucc.h> 20 #include <soc/fsl/qe/ucc_fast.h>
|
H A D | ucc.h | 14 #include <soc/fsl/qe/immap_qe.h> 15 #include <soc/fsl/qe/qe.h>
|
/kernel/linux/linux-5.10/drivers/scsi/bfa/ |
H A D | bfa_fcpim.c | 25 list_add_tail(&(__itnim)->qe, &(__itnim)->fcpim->itnim_q) 29 list_del(&(__itnim)->qe); \ 86 list_del(&(__ioim)->qe); \ 87 list_add_tail(&(__ioim)->qe, &(__ioim)->fcpim->ioim_comp_q); \ 338 struct list_head *qe, *qen; in bfa_fcpim_iocdisable() local 343 list_for_each_safe(qe, qen, &fcpim->itnim_q) { in bfa_fcpim_iocdisable() 344 itnim = (struct bfa_itnim_s *) qe; in bfa_fcpim_iocdisable() 425 struct list_head *qe, *qen; in bfa_fcpim_port_iostats() local 430 list_for_each_safe(qe, qen, &fcpim->itnim_q) { in bfa_fcpim_port_iostats() 431 itnim = (struct bfa_itnim_s *) qe; in bfa_fcpim_port_iostats() 467 struct list_head *qe, *qen; bfa_fcpim_profile_on() local 969 struct list_head *qe, *qen; bfa_itnim_iocdisable_cleanup() local 1015 struct list_head *qe, *qen; bfa_itnim_cleanup() local 1211 struct list_head *qe, *qen; bfa_itnim_delayed_comp() local 3307 struct list_head *qe, *qen; bfa_tskim_gather_ios() local 3360 struct list_head *qe, *qen; bfa_tskim_cleanup_ios() local 3458 struct list_head *qe, *qen; bfa_tskim_iocdisable_ios() local 3614 struct list_head *qe; bfa_tskim_res_recfg() local 3736 struct list_head *qe; bfa_fcp_res_recfg() local [all...] |
/kernel/linux/linux-6.6/drivers/scsi/bfa/ |
H A D | bfa_fcpim.c | 25 list_add_tail(&(__itnim)->qe, &(__itnim)->fcpim->itnim_q) 29 list_del(&(__itnim)->qe); \ 86 list_del(&(__ioim)->qe); \ 87 list_add_tail(&(__ioim)->qe, &(__ioim)->fcpim->ioim_comp_q); \ 338 struct list_head *qe, *qen; in bfa_fcpim_iocdisable() local 343 list_for_each_safe(qe, qen, &fcpim->itnim_q) { in bfa_fcpim_iocdisable() 344 itnim = (struct bfa_itnim_s *) qe; in bfa_fcpim_iocdisable() 425 struct list_head *qe, *qen; in bfa_fcpim_port_iostats() local 430 list_for_each_safe(qe, qen, &fcpim->itnim_q) { in bfa_fcpim_port_iostats() 431 itnim = (struct bfa_itnim_s *) qe; in bfa_fcpim_port_iostats() 467 struct list_head *qe, *qen; bfa_fcpim_profile_on() local 969 struct list_head *qe, *qen; bfa_itnim_iocdisable_cleanup() local 1015 struct list_head *qe, *qen; bfa_itnim_cleanup() local 1211 struct list_head *qe, *qen; bfa_itnim_delayed_comp() local 3307 struct list_head *qe, *qen; bfa_tskim_gather_ios() local 3360 struct list_head *qe, *qen; bfa_tskim_cleanup_ios() local 3458 struct list_head *qe, *qen; bfa_tskim_iocdisable_ios() local 3614 struct list_head *qe; bfa_tskim_res_recfg() local 3736 struct list_head *qe; bfa_fcp_res_recfg() local [all...] |
/kernel/linux/linux-5.10/drivers/net/wan/ |
H A D | fsl_ucc_hdlc.h | 13 #include <soc/fsl/qe/immap_qe.h> 14 #include <soc/fsl/qe/qe.h> 16 #include <soc/fsl/qe/ucc.h> 17 #include <soc/fsl/qe/ucc_fast.h>
|
/kernel/linux/linux-6.6/drivers/net/wan/ |
H A D | fsl_ucc_hdlc.h | 13 #include <soc/fsl/qe/immap_qe.h> 14 #include <soc/fsl/qe/qe.h> 16 #include <soc/fsl/qe/ucc.h> 17 #include <soc/fsl/qe/ucc_fast.h>
|