Home
last modified time | relevance | path

Searched refs:xa_load (Results 1 - 25 of 227) sorted by relevance

12345678910

/kernel/linux/linux-5.10/lib/
H A Dtest_xarray.c63 XA_BUG_ON(xa, xa_load(xa, index) != NULL); in xa_erase_index()
153 void *entry = xa_load(xa, j); in check_xa_load()
164 void *entry = xa_load(xa, j); in check_xa_load()
344 XA_BUG_ON(xa, xa_load(xa, 1) != NULL); in check_xa_shrink()
349 XA_BUG_ON(xa, xa_load(xa, 0) != xa_mk_value(0)); in check_xa_shrink()
356 XA_BUG_ON(xa, xa_load(xa, max) != xa_mk_value(0)); in check_xa_shrink()
357 XA_BUG_ON(xa, xa_load(xa, max + 1) != NULL); in check_xa_shrink()
366 XA_BUG_ON(xa, xa_load(xa, max + 1) != NULL); in check_xa_shrink()
379 XA_BUG_ON(xa, xa_load(xa, i - 1) != NULL); in check_insert()
380 XA_BUG_ON(xa, xa_load(x in check_insert()
[all...]
/kernel/linux/linux-6.6/lib/
H A Dtest_xarray.c63 XA_BUG_ON(xa, xa_load(xa, index) != NULL); in xa_erase_index()
153 void *entry = xa_load(xa, j); in check_xa_load()
164 void *entry = xa_load(xa, j); in check_xa_load()
344 XA_BUG_ON(xa, xa_load(xa, 1) != NULL); in check_xa_shrink()
349 XA_BUG_ON(xa, xa_load(xa, 0) != xa_mk_value(0)); in check_xa_shrink()
356 XA_BUG_ON(xa, xa_load(xa, max) != xa_mk_value(0)); in check_xa_shrink()
357 XA_BUG_ON(xa, xa_load(xa, max + 1) != NULL); in check_xa_shrink()
366 XA_BUG_ON(xa, xa_load(xa, max + 1) != NULL); in check_xa_shrink()
379 XA_BUG_ON(xa, xa_load(xa, i - 1) != NULL); in check_insert()
380 XA_BUG_ON(xa, xa_load(x in check_insert()
[all...]
/kernel/linux/linux-5.10/drivers/iommu/
H A Dioasid.c274 ioasid_data = xa_load(&active_allocator->xa, ioasid); in ioasid_set_data()
357 ioasid_data = xa_load(&active_allocator->xa, ioasid); in ioasid_free()
398 ioasid_data = xa_load(&idata->xa, ioasid); in ioasid_find()
/kernel/linux/linux-6.6/drivers/gpu/drm/tegra/
H A Duapi.c170 context = xa_load(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_close()
200 context = xa_load(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_map()
281 context = xa_load(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_unmap()
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/steering/
H A Ddr_domain.c81 recalc_cs_ft = xa_load(&dmn->csum_fts_xa, vport_num); in mlx5dr_domain_get_recalc_cs_ft_addr()
327 vport_caps = xa_load(&caps->vports.vports_caps_xa, vport); in mlx5dr_domain_get_vport_cap()
568 peer = xa_load(&dmn->peer_dmn_xa, peer_vhca_id); in mlx5dr_domain_set_peer()
574 peer = xa_load(&dmn->peer_dmn_xa, peer_vhca_id); in mlx5dr_domain_set_peer()
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/en/
H A Dmapping.c108 mi = xa_load(&ctx->xarray, index); in mapping_remove()
131 mi = xa_load(&ctx->xarray, index); in mapping_find()
/kernel/linux/linux-5.10/drivers/infiniband/hw/cxgb4/
H A Dev.c127 qhp = xa_load(&dev->qps, CQE_QPID(err_cqe)); in c4iw_ev_dispatch()
229 chp = xa_load(&dev->cqs, qid); in c4iw_ev_handler()
/kernel/linux/linux-6.6/drivers/infiniband/hw/cxgb4/
H A Dev.c127 qhp = xa_load(&dev->qps, CQE_QPID(err_cqe)); in c4iw_ev_dispatch()
228 chp = xa_load(&dev->cqs, qid); in c4iw_ev_handler()
/kernel/linux/linux-6.6/drivers/infiniband/sw/rxe/
H A Drxe_mr.c255 page = xa_load(&mr->page_list, index); in rxe_mr_copy_xarray()
449 page = xa_load(&mr->page_list, index); in rxe_flush_pmem_iova()
498 page = xa_load(&mr->page_list, index); in rxe_mr_do_atomic_op()
556 page = xa_load(&mr->page_list, index); in rxe_mr_do_atomic_write()
/kernel/linux/linux-6.6/drivers/firmware/arm_scmi/
H A Dperf.c528 opp = xa_load(&dom->opps_by_lvl, min_perf); in scmi_perf_limits_set()
536 opp = xa_load(&dom->opps_by_lvl, max_perf); in scmi_perf_limits_set()
607 opp = xa_load(&dom->opps_by_idx, *min_perf); in scmi_perf_limits_get()
613 opp = xa_load(&dom->opps_by_idx, *max_perf); in scmi_perf_limits_get()
674 opp = xa_load(&dom->opps_by_lvl, level); in scmi_perf_level_set()
737 opp = xa_load(&dom->opps_by_idx, *level); in scmi_perf_level_get()
910 opp = xa_load(&dom->opps_by_idx, level); in scmi_dvfs_freq_get()
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/
H A Dpagealloc.c87 root = xa_load(&dev->priv.page_root_xa, function); in page_root_per_function()
158 root = xa_load(&dev->priv.page_root_xa, function); in find_fw_page()
240 root = xa_load(&dev->priv.page_root_xa, fwp->function); in free_fwp()
410 root = xa_load(&dev->priv.page_root_xa, function); in release_all_pages()
470 root = xa_load(&dev->priv.page_root_xa, get_function(func_id, ec_function)); in reclaim_pages_cmd()
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/
H A Dpagealloc.c106 root = xa_load(&dev->priv.page_root_xa, function); in page_root_per_function()
177 root = xa_load(&dev->priv.page_root_xa, function); in find_fw_page()
259 root = xa_load(&dev->priv.page_root_xa, fwp->function); in free_fwp()
443 root = xa_load(&dev->priv.page_root_xa, function); in release_all_pages()
501 root = xa_load(&dev->priv.page_root_xa, get_function(func_id, ec_function)); in reclaim_pages_cmd()
H A Deq.c823 irq = xa_load(&table->comp_irqs, vecidx); in comp_irq_release_pci()
890 irq = xa_load(&table->comp_irqs, vecidx); in comp_irq_release_sf()
1037 irq = xa_load(&table->comp_irqs, vecidx); in create_comp_eq()
1077 eq = xa_load(&table->comp_eqs, vecidx); in mlx5_comp_eqn_get()
1108 eq = xa_load(&table->comp_eqs, vector); in mlx5_comp_irqn_get()
1125 eq = xa_load(&table->comp_eqs, vector); in mlx5_comp_irq_get_affinity_mask()
/kernel/linux/linux-5.10/arch/arm64/mm/
H A Dmteswap.c51 void *tags = xa_load(&mte_pages, entry.val); in mte_restore_tags()
/kernel/linux/linux-6.6/arch/arm64/mm/
H A Dmteswap.c52 void *tags = xa_load(&mte_pages, entry.val); in mte_restore_tags()
/kernel/linux/linux-6.6/security/apparmor/
H A Dsecid.c61 return xa_load(&aa_secids, secid); in aa_secid_to_label()
/kernel/linux/linux-6.6/mm/
H A Dreadahead.c230 struct folio *folio = xa_load(&mapping->i_pages, index + i); in page_cache_ra_unbounded()
791 struct folio *folio = xa_load(&mapping->i_pages, index); in readahead_expand()
818 struct folio *folio = xa_load(&mapping->i_pages, index); in readahead_expand()
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/sf/dev/
H A Ddev.c176 sf_dev = xa_load(&table->devices, sf_index); in mlx5_sf_dev_state_change_handler()
254 if (!xa_load(&table->devices, i)) in mlx5_sf_dev_add_active_work()
/kernel/linux/linux-6.6/fs/smb/server/mgmt/
H A Duser_session.c138 entry = xa_load(&sess->rpc_handle_list, id); in ksmbd_session_rpc_method()
258 sess = xa_load(&conn->sessions, id); in ksmbd_session_lookup()
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/en/
H A Dmapping.c116 mi = xa_load(&ctx->xarray, index); in mapping_remove()
139 mi = xa_load(&ctx->xarray, index); in mapping_find()
/kernel/linux/linux-6.6/drivers/infiniband/hw/erdma/
H A Derdma_verbs.h275 return (struct erdma_qp *)xa_load(&dev->qp_xa, id); in find_qp_by_qpn()
280 return (struct erdma_cq *)xa_load(&dev->cq_xa, id); in find_cq_by_cqn()
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx5/
H A Ddevx.c1285 event = xa_load(&dev->devx_event_table.event_xa, in devx_cleanup_subscription()
1289 xa_val_level2 = xa_load(&event->object_ids, sub->xa_key_level2); in devx_cleanup_subscription()
1351 event = xa_load(&table->event_xa, MLX5_EVENT_TYPE_COMP); in devx_cq_comp()
1355 obj_event = xa_load(&event->object_ids, obj_id); in devx_cq_comp()
1761 event = xa_load(&devx_event_table->event_xa, key_level1); in subscribe_event_xa_dealloc()
1764 xa_val_level2 = xa_load(&event->object_ids, in subscribe_event_xa_dealloc()
1783 event = xa_load(&devx_event_table->event_xa, key_level1); in subscribe_event_xa_alloc()
1805 obj_event = xa_load(&event->object_ids, key_level2); in subscribe_event_xa_alloc()
2037 event = xa_load(&devx_event_table->event_xa, in MLX5_IB_METHOD_DEVX_SUBSCRIBE_EVENT()
2047 obj_event = xa_load( in MLX5_IB_METHOD_DEVX_SUBSCRIBE_EVENT()
[all...]
/kernel/linux/linux-6.6/drivers/infiniband/hw/mlx5/
H A Ddevx.c1374 event = xa_load(&dev->devx_event_table.event_xa, in devx_cleanup_subscription()
1378 xa_val_level2 = xa_load(&event->object_ids, sub->xa_key_level2); in devx_cleanup_subscription()
1439 event = xa_load(&table->event_xa, MLX5_EVENT_TYPE_COMP); in devx_cq_comp()
1443 obj_event = xa_load(&event->object_ids, obj_id); in devx_cq_comp()
1866 event = xa_load(&devx_event_table->event_xa, key_level1); in subscribe_event_xa_dealloc()
1869 xa_val_level2 = xa_load(&event->object_ids, in subscribe_event_xa_dealloc()
1888 event = xa_load(&devx_event_table->event_xa, key_level1); in subscribe_event_xa_alloc()
1910 obj_event = xa_load(&event->object_ids, key_level2); in subscribe_event_xa_alloc()
2140 event = xa_load(&devx_event_table->event_xa, in MLX5_IB_METHOD_DEVX_SUBSCRIBE_EVENT()
2150 obj_event = xa_load( in MLX5_IB_METHOD_DEVX_SUBSCRIBE_EVENT()
[all...]
/kernel/linux/linux-6.6/drivers/net/ethernet/intel/ice/
H A Dice_eswitch_br.c312 port = xa_load(&bridge->ports, vsi_idx); in ice_esw_br_port_vlan_lookup()
318 vlan = xa_load(&port->vlans, vid); in ice_esw_br_port_vlan_lookup()
739 port = xa_load(&bridge->ports, vsi_idx); in ice_eswitch_br_port_vlan_add()
750 vlan = xa_load(&port->vlans, vid); in ice_eswitch_br_port_vlan_add()
774 port = xa_load(&bridge->ports, vsi_idx); in ice_eswitch_br_port_vlan_del()
778 vlan = xa_load(&port->vlans, vid); in ice_eswitch_br_port_vlan_del()
/kernel/linux/linux-5.10/drivers/infiniband/core/
H A Ddevice.c311 device = xa_load(&devices, index); in ib_device_get_by_index()
430 struct ib_client *client = xa_load(&clients, index); in ib_device_rename()
636 if (xa_load(&devices, device->index) == device) in ib_dealloc_device()
731 client_data = xa_load(&device->client_data, client_id); in remove_client_context()
733 client = xa_load(&clients, client_id); in remove_client_context()
903 cdev = xa_load(&device->compat_devs, rnet->id); in add_one_compat_dev()
1713 if (xa_load(&clients, highest_client_id - 1)) in remove_client_id()
1838 struct ib_client *client = xa_load(&clients, index); in __ib_get_client_nl_info()
2511 struct ib_client *client = xa_load(&clients, index); in ib_get_net_dev_by_params()

Completed in 29 milliseconds

12345678910