/kernel/linux/linux-6.6/drivers/infiniband/hw/mlx5/ |
H A D | qpc.c | 13 struct mlx5_core_dct *dct); 93 struct mlx5_core_dct *dct; in dct_event_notifier() local 97 qpn = be32_to_cpu(eqe->data.dct.dctn) & 0xFFFFFF; in dct_event_notifier() 99 dct = xa_load(&dev->qp_table.dct_xa, qpn); in dct_event_notifier() 100 if (dct) in dct_event_notifier() 101 complete(&dct->drained); in dct_event_notifier() 196 struct mlx5_core_dct *dct) in _mlx5_core_destroy_dct() 199 struct mlx5_core_qp *qp = &dct->mqp; in _mlx5_core_destroy_dct() 207 int mlx5_core_create_dct(struct mlx5_ib_dev *dev, struct mlx5_core_dct *dct, in mlx5_core_create_dct() argument 210 struct mlx5_core_qp *qp = &dct in mlx5_core_create_dct() 195 _mlx5_core_destroy_dct(struct mlx5_ib_dev *dev, struct mlx5_core_dct *dct) _mlx5_core_destroy_dct() argument 264 mlx5_core_drain_dct(struct mlx5_ib_dev *dev, struct mlx5_core_dct *dct) mlx5_core_drain_dct() argument 276 mlx5_core_destroy_dct(struct mlx5_ib_dev *dev, struct mlx5_core_dct *dct) mlx5_core_destroy_dct() argument 535 mlx5_core_dct_query(struct mlx5_ib_dev *dev, struct mlx5_core_dct *dct, u32 *out, int outlen) mlx5_core_dct_query() argument [all...] |
H A D | qp.h | 31 int mlx5_core_destroy_dct(struct mlx5_ib_dev *dev, struct mlx5_core_dct *dct); 34 int mlx5_core_dct_query(struct mlx5_ib_dev *dev, struct mlx5_core_dct *dct,
|
H A D | qp.c | 2741 qp->dct.in = kzalloc(MLX5_ST_SZ_BYTES(create_dct_in), GFP_KERNEL); in create_dct() 2742 if (!qp->dct.in) in create_dct() 2745 MLX5_SET(create_dct_in, qp->dct.in, uid, to_mpd(pd)->uid); in create_dct() 2746 dctc = MLX5_ADDR_OF(create_dct_in, qp->dct.in, dct_context_entry); in create_dct() 2769 if (attr->qp_type == IB_QPT_DRIVER && !MLX5_CAP_GEN(dev->mdev, dct)) in check_qp_type() 3195 err = mlx5_core_destroy_dct(dev, &mqp->dct.mdct); in mlx5_ib_destroy_dct() 3202 kfree(mqp->dct.in); in mlx5_ib_destroy_dct() 4508 dctc = MLX5_ADDR_OF(create_dct_in, qp->dct.in, dct_context_entry); in mlx5_ib_modify_dct() 4581 err = mlx5_core_create_dct(dev, &qp->dct.mdct, qp->dct in mlx5_ib_modify_dct() 5001 struct mlx5_core_dct *dct = &mqp->dct.mdct; mlx5_ib_dct_query_qp() local [all...] |
H A D | devx.c | 653 qp->dct.mdct.mqp.qpn) == obj_id; in devx_is_valid_obj_id() 2435 obj_id = be32_to_cpu(eqe->data.dct.dctn) & 0xffffff; in devx_get_obj_id_from_event()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx5/ |
H A D | qpc.c | 13 struct mlx5_core_dct *dct); 96 struct mlx5_core_dct *dct; in rsc_event_notifier() local 105 rsn = be32_to_cpu(eqe->data.dct.dctn) & 0xffffff; in rsc_event_notifier() 140 dct = (struct mlx5_core_dct *)common; in rsc_event_notifier() 142 complete(&dct->drained); in rsc_event_notifier() 190 struct mlx5_core_dct *dct, bool need_cleanup) in _mlx5_core_destroy_dct() 193 struct mlx5_core_qp *qp = &dct->mqp; in _mlx5_core_destroy_dct() 196 err = mlx5_core_drain_dct(dev, dct); in _mlx5_core_destroy_dct() 203 wait_for_completion(&dct->drained); in _mlx5_core_destroy_dct() 206 destroy_resource_common(dev, &dct in _mlx5_core_destroy_dct() 189 _mlx5_core_destroy_dct(struct mlx5_ib_dev *dev, struct mlx5_core_dct *dct, bool need_cleanup) _mlx5_core_destroy_dct() argument 214 mlx5_core_create_dct(struct mlx5_ib_dev *dev, struct mlx5_core_dct *dct, u32 *in, int inlen, u32 *out, int outlen) mlx5_core_create_dct() argument 271 mlx5_core_drain_dct(struct mlx5_ib_dev *dev, struct mlx5_core_dct *dct) mlx5_core_drain_dct() argument 283 mlx5_core_destroy_dct(struct mlx5_ib_dev *dev, struct mlx5_core_dct *dct) mlx5_core_destroy_dct() argument 509 mlx5_core_dct_query(struct mlx5_ib_dev *dev, struct mlx5_core_dct *dct, u32 *out, int outlen) mlx5_core_dct_query() argument [all...] |
H A D | qp.h | 21 int mlx5_core_destroy_dct(struct mlx5_ib_dev *dev, struct mlx5_core_dct *dct); 24 int mlx5_core_dct_query(struct mlx5_ib_dev *dev, struct mlx5_core_dct *dct,
|
H A D | qp.c | 2414 qp->dct.in = kzalloc(MLX5_ST_SZ_BYTES(create_dct_in), GFP_KERNEL); in create_dct() 2415 if (!qp->dct.in) in create_dct() 2418 MLX5_SET(create_dct_in, qp->dct.in, uid, to_mpd(pd)->uid); in create_dct() 2419 dctc = MLX5_ADDR_OF(create_dct_in, qp->dct.in, dct_context_entry); in create_dct() 2443 if (attr->qp_type == IB_QPT_DRIVER && !MLX5_CAP_GEN(dev->mdev, dct)) in check_qp_type() 2857 err = mlx5_core_destroy_dct(dev, &mqp->dct.mdct); in mlx5_ib_destroy_dct() 2864 kfree(mqp->dct.in); in mlx5_ib_destroy_dct() 4176 dctc = MLX5_ADDR_OF(create_dct_in, qp->dct.in, dct_context_entry); in mlx5_ib_modify_dct() 4249 err = mlx5_core_create_dct(dev, &qp->dct.mdct, qp->dct in mlx5_ib_modify_dct() 4674 struct mlx5_core_dct *dct = &mqp->dct.mdct; mlx5_ib_dct_query_qp() local [all...] |
H A D | devx.c | 582 qp->dct.mdct.mqp.qpn) == obj_id; in devx_is_valid_obj_id() 2268 obj_id = be32_to_cpu(eqe->data.dct.dctn) & 0xffffff; in devx_get_obj_id_from_event()
|
H A D | mlx5_ib.h | 409 struct mlx5_ib_dct dct; member
|
/kernel/linux/linux-6.6/drivers/edac/ |
H A D | amd64_edac.c | 102 static void f15h_select_dct(struct amd64_pvt *pvt, u8 dct) in f15h_select_dct() argument 108 reg |= dct; in f15h_select_dct() 126 static inline int amd64_read_dct_pci_cfg(struct amd64_pvt *pvt, u8 dct, in amd64_read_dct_pci_cfg() argument 131 if (dct || offset >= 0x100) in amd64_read_dct_pci_cfg() 136 if (dct) { in amd64_read_dct_pci_cfg() 154 dct = (dct && pvt->model == 0x30) ? 3 : dct; in amd64_read_dct_pci_cfg() 155 f15h_select_dct(pvt, dct); in amd64_read_dct_pci_cfg() 159 if (dct) in amd64_read_dct_pci_cfg() 372 get_cs_base_and_mask(struct amd64_pvt *pvt, int csrow, u8 dct, u64 *base, u64 *mask) get_cs_base_and_mask() argument 2111 k8_dbam_to_chip_select(struct amd64_pvt *pvt, u8 dct, unsigned cs_mode, int cs_mask_nr) k8_dbam_to_chip_select() argument 2215 f10_dbam_to_chip_select(struct amd64_pvt *pvt, u8 dct, unsigned cs_mode, int cs_mask_nr) f10_dbam_to_chip_select() argument 2231 f15_dbam_to_chip_select(struct amd64_pvt *pvt, u8 dct, unsigned cs_mode, int cs_mask_nr) f15_dbam_to_chip_select() argument 2240 f15_m60h_dbam_to_chip_select(struct amd64_pvt *pvt, u8 dct, unsigned cs_mode, int cs_mask_nr) f15_m60h_dbam_to_chip_select() argument 2273 f16_dbam_to_chip_select(struct amd64_pvt *pvt, u8 dct, unsigned cs_mode, int cs_mask_nr) f16_dbam_to_chip_select() argument 2445 f10_process_possible_spare(struct amd64_pvt *pvt, u8 dct, int csrow) f10_process_possible_spare() argument 2470 f1x_lookup_addr_in_dct(u64 in_addr, u8 nid, u8 dct) f1x_lookup_addr_in_dct() argument 3289 dct_get_csrow_nr_pages(struct amd64_pvt *pvt, u8 dct, int csrow_nr) dct_get_csrow_nr_pages() argument 3307 umc_get_csrow_nr_pages(struct amd64_pvt *pvt, u8 dct, int csrow_nr_orig) umc_get_csrow_nr_pages() argument 3822 gpu_get_csrow_nr_pages(struct amd64_pvt *pvt, u8 dct, int csrow_nr) gpu_get_csrow_nr_pages() argument 4224 int cs = 0, dct = 0; instance_has_memory() local [all...] |
H A D | amd64_edac.h | 166 #define csrow_enabled(i, dct, pvt) ((pvt)->csels[(dct)].csbases[(i)] & DCSB_CS_ENABLE) 167 #define csrow_sec_enabled(i, dct, pvt) ((pvt)->csels[(dct)].csbases_sec[(i)] & DCSB_CS_ENABLE) 467 int (*dbam_to_cs)(struct amd64_pvt *pvt, u8 dct,
|
/kernel/linux/linux-5.10/drivers/edac/ |
H A D | amd64_edac.c | 89 static void f15h_select_dct(struct amd64_pvt *pvt, u8 dct) in f15h_select_dct() argument 95 reg |= dct; in f15h_select_dct() 113 static inline int amd64_read_dct_pci_cfg(struct amd64_pvt *pvt, u8 dct, in amd64_read_dct_pci_cfg() argument 118 if (dct || offset >= 0x100) in amd64_read_dct_pci_cfg() 123 if (dct) { in amd64_read_dct_pci_cfg() 141 dct = (dct && pvt->model == 0x30) ? 3 : dct; in amd64_read_dct_pci_cfg() 142 f15h_select_dct(pvt, dct); in amd64_read_dct_pci_cfg() 146 if (dct) in amd64_read_dct_pci_cfg() 385 get_cs_base_and_mask(struct amd64_pvt *pvt, int csrow, u8 dct, u64 *base, u64 *mask) get_cs_base_and_mask() argument 1356 k8_dbam_to_chip_select(struct amd64_pvt *pvt, u8 dct, unsigned cs_mode, int cs_mask_nr) k8_dbam_to_chip_select() argument 1523 f10_dbam_to_chip_select(struct amd64_pvt *pvt, u8 dct, unsigned cs_mode, int cs_mask_nr) f10_dbam_to_chip_select() argument 1539 f15_dbam_to_chip_select(struct amd64_pvt *pvt, u8 dct, unsigned cs_mode, int cs_mask_nr) f15_dbam_to_chip_select() argument 1548 f15_m60h_dbam_to_chip_select(struct amd64_pvt *pvt, u8 dct, unsigned cs_mode, int cs_mask_nr) f15_m60h_dbam_to_chip_select() argument 1581 f16_dbam_to_chip_select(struct amd64_pvt *pvt, u8 dct, unsigned cs_mode, int cs_mask_nr) f16_dbam_to_chip_select() argument 1814 f10_process_possible_spare(struct amd64_pvt *pvt, u8 dct, int csrow) f10_process_possible_spare() argument 1839 f1x_lookup_addr_in_dct(u64 in_addr, u8 nid, u8 dct) f1x_lookup_addr_in_dct() argument 2928 get_csrow_nr_pages(struct amd64_pvt *pvt, u8 dct, int csrow_nr_orig) get_csrow_nr_pages() argument 3542 int cs = 0, dct = 0; instance_has_memory() local [all...] |
H A D | amd64_edac.h | 178 #define csrow_enabled(i, dct, pvt) ((pvt)->csels[(dct)].csbases[(i)] & DCSB_CS_ENABLE) 179 #define csrow_sec_enabled(i, dct, pvt) ((pvt)->csels[(dct)].csbases_sec[(i)] & DCSB_CS_ENABLE) 481 int (*dbam_to_cs) (struct amd64_pvt *pvt, u8 dct,
|
/kernel/linux/linux-5.10/drivers/soc/fsl/dpio/ |
H A D | qbman-portal.h | 208 enum qbman_pull_type_e dct); 210 enum qbman_pull_type_e dct);
|
H A D | qbman-portal.c | 1058 * @dct: the dequeue command type 1061 enum qbman_pull_type_e dct) in qbman_pull_desc_set_wq() 1063 d->verb |= dct << QB_VDQCR_VERB_DCT_SHIFT; in qbman_pull_desc_set_wq() 1072 * @dct: the dequeue command type 1075 enum qbman_pull_type_e dct) in qbman_pull_desc_set_channel() 1077 d->verb |= dct << QB_VDQCR_VERB_DCT_SHIFT; in qbman_pull_desc_set_channel() 1060 qbman_pull_desc_set_wq(struct qbman_pull_desc *d, u32 wqid, enum qbman_pull_type_e dct) qbman_pull_desc_set_wq() argument 1074 qbman_pull_desc_set_channel(struct qbman_pull_desc *d, u32 chid, enum qbman_pull_type_e dct) qbman_pull_desc_set_channel() argument
|
/kernel/linux/linux-6.6/drivers/soc/fsl/dpio/ |
H A D | qbman-portal.h | 215 enum qbman_pull_type_e dct); 217 enum qbman_pull_type_e dct);
|
H A D | qbman-portal.c | 1065 * @dct: the dequeue command type 1068 enum qbman_pull_type_e dct) in qbman_pull_desc_set_wq() 1070 d->verb |= dct << QB_VDQCR_VERB_DCT_SHIFT; in qbman_pull_desc_set_wq() 1080 * @dct: the dequeue command type 1083 enum qbman_pull_type_e dct) in qbman_pull_desc_set_channel() 1085 d->verb |= dct << QB_VDQCR_VERB_DCT_SHIFT; in qbman_pull_desc_set_channel() 1067 qbman_pull_desc_set_wq(struct qbman_pull_desc *d, u32 wqid, enum qbman_pull_type_e dct) qbman_pull_desc_set_wq() argument 1082 qbman_pull_desc_set_channel(struct qbman_pull_desc *d, u32 chid, enum qbman_pull_type_e dct) qbman_pull_desc_set_channel() argument
|
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | main.c | 552 if (MLX5_CAP_GEN_MAX(dev, dct)) in handle_hca_cap() 553 MLX5_SET(cmd_hca_cap, set_hca_cap, dct, 1); in handle_hca_cap()
|
H A D | eq.c | 580 if (MLX5_CAP_GEN_MAX(dev, dct)) in gather_async_events_mask()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | main.c | 622 if (MLX5_CAP_GEN_MAX(dev, dct)) in handle_hca_cap() 623 MLX5_SET(cmd_hca_cap, set_hca_cap, dct, 1); in handle_hca_cap()
|
H A D | eq.c | 571 if (MLX5_CAP_GEN_MAX(dev, dct)) in gather_async_events_mask()
|
/kernel/linux/linux-5.10/arch/ia64/include/asm/ |
H A D | pal.h | 1293 dct :4, /* Date cache tracking */ member
|
/kernel/linux/linux-6.6/arch/ia64/include/asm/ |
H A D | pal.h | 1294 dct :4, /* Date cache tracking */ member
|
/kernel/linux/linux-5.10/include/linux/mlx5/ |
H A D | device.h | 735 struct mlx5_eqe_dct dct; member
|
/kernel/linux/linux-6.6/include/linux/mlx5/ |
H A D | device.h | 751 struct mlx5_eqe_dct dct; member
|