Home
last modified time | relevance | path

Searched refs:tbl (Results 1 - 25 of 574) sorted by relevance

12345678910>>...23

/kernel/linux/linux-5.10/fs/nfs/
H A Dnfs4session.c27 static void nfs4_init_slot_table(struct nfs4_slot_table *tbl, const char *queue) in nfs4_init_slot_table() argument
29 tbl->highest_used_slotid = NFS4_NO_SLOT; in nfs4_init_slot_table()
30 spin_lock_init(&tbl->slot_tbl_lock); in nfs4_init_slot_table()
31 rpc_init_priority_wait_queue(&tbl->slot_tbl_waitq, queue); in nfs4_init_slot_table()
32 init_waitqueue_head(&tbl->slot_waitq); in nfs4_init_slot_table()
33 init_completion(&tbl->complete); in nfs4_init_slot_table()
39 static void nfs4_shrink_slot_table(struct nfs4_slot_table *tbl, u32 newsize) in nfs4_shrink_slot_table() argument
42 if (newsize >= tbl->max_slots) in nfs4_shrink_slot_table()
45 p = &tbl->slots; in nfs4_shrink_slot_table()
53 tbl in nfs4_shrink_slot_table()
62 nfs4_slot_tbl_drain_complete(struct nfs4_slot_table *tbl) nfs4_slot_tbl_drain_complete() argument
83 nfs4_free_slot(struct nfs4_slot_table *tbl, struct nfs4_slot *slot) nfs4_free_slot() argument
104 nfs4_new_slot(struct nfs4_slot_table *tbl, u32 slotid, u32 seq_init, gfp_t gfp_mask) nfs4_new_slot() argument
120 nfs4_find_or_create_slot(struct nfs4_slot_table *tbl, u32 slotid, u32 seq_init, gfp_t gfp_mask) nfs4_find_or_create_slot() argument
142 nfs4_lock_slot(struct nfs4_slot_table *tbl, struct nfs4_slot *slot) nfs4_lock_slot() argument
159 nfs4_try_to_lock_slot(struct nfs4_slot_table *tbl, struct nfs4_slot *slot) nfs4_try_to_lock_slot() argument
172 nfs4_lookup_slot(struct nfs4_slot_table *tbl, u32 slotid) nfs4_lookup_slot() argument
201 nfs4_slot_seqid_in_use(struct nfs4_slot_table *tbl, u32 slotid, u32 seq_nr) nfs4_slot_seqid_in_use() argument
222 nfs4_slot_wait_on_seqid(struct nfs4_slot_table *tbl, u32 slotid, u32 seq_nr, unsigned long timeout) nfs4_slot_wait_on_seqid() argument
242 nfs4_alloc_slot(struct nfs4_slot_table *tbl) nfs4_alloc_slot() argument
262 nfs4_grow_slot_table(struct nfs4_slot_table *tbl, u32 max_reqs, u32 ivalue) nfs4_grow_slot_table() argument
272 nfs4_reset_slot_table(struct nfs4_slot_table *tbl, u32 server_highest_slotid, u32 ivalue) nfs4_reset_slot_table() argument
297 nfs4_realloc_slot_table(struct nfs4_slot_table *tbl, u32 max_reqs, u32 ivalue) nfs4_realloc_slot_table() argument
326 nfs4_release_slot_table(struct nfs4_slot_table *tbl) nfs4_release_slot_table() argument
336 nfs4_shutdown_slot_table(struct nfs4_slot_table *tbl) nfs4_shutdown_slot_table() argument
350 nfs4_setup_slot_table(struct nfs4_slot_table *tbl, unsigned int max_reqs, const char *queue) nfs4_setup_slot_table() argument
362 struct nfs4_slot_table *tbl = slot->table; nfs41_assign_slot() local
375 __nfs41_wake_and_assign_slot(struct nfs4_slot_table *tbl, struct nfs4_slot *slot) __nfs41_wake_and_assign_slot() argument
383 nfs41_wake_and_assign_slot(struct nfs4_slot_table *tbl, struct nfs4_slot *slot) nfs41_wake_and_assign_slot() argument
391 nfs41_try_wake_next_slot_table_entry(struct nfs4_slot_table *tbl) nfs41_try_wake_next_slot_table_entry() argument
403 nfs41_wake_slot_table(struct nfs4_slot_table *tbl) nfs41_wake_slot_table() argument
413 nfs41_set_max_slotid_locked(struct nfs4_slot_table *tbl, u32 target_highest_slotid) nfs41_set_max_slotid_locked() argument
428 nfs41_set_target_slotid_locked(struct nfs4_slot_table *tbl, u32 target_highest_slotid) nfs41_set_target_slotid_locked() argument
437 nfs41_set_target_slotid(struct nfs4_slot_table *tbl, u32 target_highest_slotid) nfs41_set_target_slotid() argument
448 nfs41_set_server_slotid_locked(struct nfs4_slot_table *tbl, u32 highest_slotid) nfs41_set_server_slotid_locked() argument
489 nfs41_is_outlier_target_slotid(struct nfs4_slot_table *tbl, u32 new_target) nfs41_is_outlier_target_slotid() argument
510 nfs41_update_target_slotid(struct nfs4_slot_table *tbl, struct nfs4_slot *slot, struct nfs4_sequence_res *res) nfs41_update_target_slotid() argument
534 struct nfs4_slot_table *tbl; nfs4_setup_session_slot_tables() local
[all...]
/kernel/linux/linux-6.6/fs/nfs/
H A Dnfs4session.c27 static void nfs4_init_slot_table(struct nfs4_slot_table *tbl, const char *queue) in nfs4_init_slot_table() argument
29 tbl->highest_used_slotid = NFS4_NO_SLOT; in nfs4_init_slot_table()
30 spin_lock_init(&tbl->slot_tbl_lock); in nfs4_init_slot_table()
31 rpc_init_priority_wait_queue(&tbl->slot_tbl_waitq, queue); in nfs4_init_slot_table()
32 init_waitqueue_head(&tbl->slot_waitq); in nfs4_init_slot_table()
33 init_completion(&tbl->complete); in nfs4_init_slot_table()
39 static void nfs4_shrink_slot_table(struct nfs4_slot_table *tbl, u32 newsize) in nfs4_shrink_slot_table() argument
42 if (newsize >= tbl->max_slots) in nfs4_shrink_slot_table()
45 p = &tbl->slots; in nfs4_shrink_slot_table()
53 tbl in nfs4_shrink_slot_table()
62 nfs4_slot_tbl_drain_complete(struct nfs4_slot_table *tbl) nfs4_slot_tbl_drain_complete() argument
83 nfs4_free_slot(struct nfs4_slot_table *tbl, struct nfs4_slot *slot) nfs4_free_slot() argument
104 nfs4_new_slot(struct nfs4_slot_table *tbl, u32 slotid, u32 seq_init, gfp_t gfp_mask) nfs4_new_slot() argument
120 nfs4_find_or_create_slot(struct nfs4_slot_table *tbl, u32 slotid, u32 seq_init, gfp_t gfp_mask) nfs4_find_or_create_slot() argument
142 nfs4_lock_slot(struct nfs4_slot_table *tbl, struct nfs4_slot *slot) nfs4_lock_slot() argument
159 nfs4_try_to_lock_slot(struct nfs4_slot_table *tbl, struct nfs4_slot *slot) nfs4_try_to_lock_slot() argument
172 nfs4_lookup_slot(struct nfs4_slot_table *tbl, u32 slotid) nfs4_lookup_slot() argument
201 nfs4_slot_seqid_in_use(struct nfs4_slot_table *tbl, u32 slotid, u32 seq_nr) nfs4_slot_seqid_in_use() argument
222 nfs4_slot_wait_on_seqid(struct nfs4_slot_table *tbl, u32 slotid, u32 seq_nr, unsigned long timeout) nfs4_slot_wait_on_seqid() argument
242 nfs4_alloc_slot(struct nfs4_slot_table *tbl) nfs4_alloc_slot() argument
262 nfs4_grow_slot_table(struct nfs4_slot_table *tbl, u32 max_reqs, u32 ivalue) nfs4_grow_slot_table() argument
272 nfs4_reset_slot_table(struct nfs4_slot_table *tbl, u32 server_highest_slotid, u32 ivalue) nfs4_reset_slot_table() argument
297 nfs4_realloc_slot_table(struct nfs4_slot_table *tbl, u32 max_reqs, u32 ivalue) nfs4_realloc_slot_table() argument
326 nfs4_release_slot_table(struct nfs4_slot_table *tbl) nfs4_release_slot_table() argument
336 nfs4_shutdown_slot_table(struct nfs4_slot_table *tbl) nfs4_shutdown_slot_table() argument
350 nfs4_setup_slot_table(struct nfs4_slot_table *tbl, unsigned int max_reqs, const char *queue) nfs4_setup_slot_table() argument
362 struct nfs4_slot_table *tbl = slot->table; nfs41_assign_slot() local
375 __nfs41_wake_and_assign_slot(struct nfs4_slot_table *tbl, struct nfs4_slot *slot) __nfs41_wake_and_assign_slot() argument
383 nfs41_wake_and_assign_slot(struct nfs4_slot_table *tbl, struct nfs4_slot *slot) nfs41_wake_and_assign_slot() argument
391 nfs41_try_wake_next_slot_table_entry(struct nfs4_slot_table *tbl) nfs41_try_wake_next_slot_table_entry() argument
403 nfs41_wake_slot_table(struct nfs4_slot_table *tbl) nfs41_wake_slot_table() argument
413 nfs41_set_max_slotid_locked(struct nfs4_slot_table *tbl, u32 target_highest_slotid) nfs41_set_max_slotid_locked() argument
428 nfs41_set_target_slotid_locked(struct nfs4_slot_table *tbl, u32 target_highest_slotid) nfs41_set_target_slotid_locked() argument
437 nfs41_set_target_slotid(struct nfs4_slot_table *tbl, u32 target_highest_slotid) nfs41_set_target_slotid() argument
448 nfs41_set_server_slotid_locked(struct nfs4_slot_table *tbl, u32 highest_slotid) nfs41_set_server_slotid_locked() argument
489 nfs41_is_outlier_target_slotid(struct nfs4_slot_table *tbl, u32 new_target) nfs41_is_outlier_target_slotid() argument
510 nfs41_update_target_slotid(struct nfs4_slot_table *tbl, struct nfs4_slot *slot, struct nfs4_sequence_res *res) nfs41_update_target_slotid() argument
538 struct nfs4_slot_table *tbl; nfs4_setup_session_slot_tables() local
[all...]
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/steering/
H A Ddr_table.c6 int mlx5dr_table_set_miss_action(struct mlx5dr_table *tbl, in mlx5dr_table_set_miss_action() argument
17 mlx5dr_domain_lock(tbl->dmn); in mlx5dr_table_set_miss_action()
19 if (!list_empty(&tbl->matcher_list)) in mlx5dr_table_set_miss_action()
20 last_matcher = list_last_entry(&tbl->matcher_list, in mlx5dr_table_set_miss_action()
24 if (tbl->dmn->type == MLX5DR_DOMAIN_TYPE_NIC_RX || in mlx5dr_table_set_miss_action()
25 tbl->dmn->type == MLX5DR_DOMAIN_TYPE_FDB) { in mlx5dr_table_set_miss_action()
29 last_htbl = tbl->rx.s_anchor; in mlx5dr_table_set_miss_action()
31 tbl->rx.default_icm_addr = action ? in mlx5dr_table_set_miss_action()
32 action->dest_tbl.tbl->rx.s_anchor->chunk->icm_addr : in mlx5dr_table_set_miss_action()
33 tbl in mlx5dr_table_set_miss_action()
93 dr_table_uninit_fdb(struct mlx5dr_table *tbl) dr_table_uninit_fdb() argument
99 dr_table_uninit(struct mlx5dr_table *tbl) dr_table_uninit() argument
158 dr_table_init_fdb(struct mlx5dr_table *tbl) dr_table_init_fdb() argument
177 dr_table_init(struct mlx5dr_table *tbl) dr_table_init() argument
212 dr_table_destroy_sw_owned_tbl(struct mlx5dr_table *tbl) dr_table_destroy_sw_owned_tbl() argument
219 dr_table_create_sw_owned_tbl(struct mlx5dr_table *tbl) dr_table_create_sw_owned_tbl() argument
250 struct mlx5dr_table *tbl; mlx5dr_table_create() local
283 mlx5dr_table_destroy(struct mlx5dr_table *tbl) mlx5dr_table_destroy() argument
305 mlx5dr_table_get_id(struct mlx5dr_table *tbl) mlx5dr_table_get_id() argument
[all...]
/kernel/linux/linux-5.10/arch/powerpc/kernel/
H A Diommu.c175 struct iommu_table *tbl, in iommu_range_alloc()
208 pool_nr = raw_cpu_read(iommu_pool_hash) & (tbl->nr_pools - 1); in iommu_range_alloc()
211 pool = &(tbl->large_pool); in iommu_range_alloc()
213 pool = &(tbl->pools[pool_nr]); in iommu_range_alloc()
233 if (limit + tbl->it_offset > mask) { in iommu_range_alloc()
234 limit = mask - tbl->it_offset + 1; in iommu_range_alloc()
241 pool = &(tbl->pools[0]); in iommu_range_alloc()
249 n = iommu_area_alloc(tbl->it_map, limit, start, npages, tbl->it_offset, in iommu_range_alloc()
250 dma_get_seg_boundary_nr_pages(dev, tbl in iommu_range_alloc()
174 iommu_range_alloc(struct device *dev, struct iommu_table *tbl, unsigned long npages, unsigned long *handle, unsigned long mask, unsigned int align_order) iommu_range_alloc() argument
297 iommu_alloc(struct device *dev, struct iommu_table *tbl, void *page, unsigned int npages, enum dma_data_direction direction, unsigned long mask, unsigned int align_order, unsigned long attrs) iommu_alloc() argument
340 iommu_free_check(struct iommu_table *tbl, dma_addr_t dma_addr, unsigned int npages) iommu_free_check() argument
368 get_pool(struct iommu_table *tbl, unsigned long entry) get_pool() argument
387 __iommu_free(struct iommu_table *tbl, dma_addr_t dma_addr, unsigned int npages) __iommu_free() argument
409 iommu_free(struct iommu_table *tbl, dma_addr_t dma_addr, unsigned int npages) iommu_free() argument
422 ppc_iommu_map_sg(struct device *dev, struct iommu_table *tbl, struct scatterlist *sglist, int nelems, unsigned long mask, enum dma_data_direction direction, unsigned long attrs) ppc_iommu_map_sg() argument
567 ppc_iommu_unmap_sg(struct iommu_table *tbl, struct scatterlist *sglist, int nelems, enum dma_data_direction direction, unsigned long attrs) ppc_iommu_unmap_sg() argument
599 iommu_table_clear(struct iommu_table *tbl) iommu_table_clear() argument
640 iommu_table_reserve_pages(struct iommu_table *tbl, unsigned long res_start, unsigned long res_end) iommu_table_reserve_pages() argument
667 iommu_table_release_pages(struct iommu_table *tbl) iommu_table_release_pages() argument
686 iommu_init_table(struct iommu_table *tbl, int nid, unsigned long res_start, unsigned long res_end) iommu_init_table() argument
746 struct iommu_table *tbl; iommu_table_free() local
775 iommu_tce_table_get(struct iommu_table *tbl) iommu_tce_table_get() argument
784 iommu_tce_table_put(struct iommu_table *tbl) iommu_tce_table_put() argument
798 iommu_map_page(struct device *dev, struct iommu_table *tbl, struct page *page, unsigned long offset, size_t size, unsigned long mask, enum dma_data_direction direction, unsigned long attrs) iommu_map_page() argument
837 iommu_unmap_page(struct iommu_table *tbl, dma_addr_t dma_handle, size_t size, enum dma_data_direction direction, unsigned long attrs) iommu_unmap_page() argument
856 iommu_alloc_coherent(struct device *dev, struct iommu_table *tbl, size_t size, dma_addr_t *dma_handle, unsigned long mask, gfp_t flag, int node) iommu_alloc_coherent() argument
903 iommu_free_coherent(struct iommu_table *tbl, size_t size, void *vaddr, dma_addr_t dma_handle) iommu_free_coherent() argument
978 iommu_flush_tce(struct iommu_table *tbl) iommu_flush_tce() argument
1020 iommu_tce_xchg_no_kill(struct mm_struct *mm, struct iommu_table *tbl, unsigned long entry, unsigned long *hpa, enum dma_data_direction *direction) iommu_tce_xchg_no_kill() argument
1039 iommu_tce_kill(struct iommu_table *tbl, unsigned long entry, unsigned long pages) iommu_tce_kill() argument
1047 iommu_take_ownership(struct iommu_table *tbl) iommu_take_ownership() argument
1086 iommu_release_ownership(struct iommu_table *tbl) iommu_release_ownership() argument
[all...]
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/steering/
H A Ddr_table.c28 action->dest_tbl->tbl->rx.s_anchor->chunk : in dr_table_set_miss_action_nic()
29 action->dest_tbl->tbl->tx.s_anchor->chunk; in dr_table_set_miss_action_nic()
46 int mlx5dr_table_set_miss_action(struct mlx5dr_table *tbl, in mlx5dr_table_set_miss_action() argument
54 mlx5dr_domain_lock(tbl->dmn); in mlx5dr_table_set_miss_action()
56 if (tbl->dmn->type == MLX5DR_DOMAIN_TYPE_NIC_RX || in mlx5dr_table_set_miss_action()
57 tbl->dmn->type == MLX5DR_DOMAIN_TYPE_FDB) { in mlx5dr_table_set_miss_action()
58 ret = dr_table_set_miss_action_nic(tbl->dmn, &tbl->rx, action); in mlx5dr_table_set_miss_action()
63 if (tbl->dmn->type == MLX5DR_DOMAIN_TYPE_NIC_TX || in mlx5dr_table_set_miss_action()
64 tbl in mlx5dr_table_set_miss_action()
92 dr_table_uninit_fdb(struct mlx5dr_table *tbl) dr_table_uninit_fdb() argument
98 dr_table_uninit(struct mlx5dr_table *tbl) dr_table_uninit() argument
159 dr_table_init_fdb(struct mlx5dr_table *tbl) dr_table_init_fdb() argument
178 dr_table_init(struct mlx5dr_table *tbl) dr_table_init() argument
213 dr_table_destroy_sw_owned_tbl(struct mlx5dr_table *tbl) dr_table_destroy_sw_owned_tbl() argument
220 dr_table_create_sw_owned_tbl(struct mlx5dr_table *tbl, u16 uid) dr_table_create_sw_owned_tbl() argument
253 struct mlx5dr_table *tbl; mlx5dr_table_create() local
288 mlx5dr_table_destroy(struct mlx5dr_table *tbl) mlx5dr_table_destroy() argument
311 mlx5dr_table_get_id(struct mlx5dr_table *tbl) mlx5dr_table_get_id() argument
[all...]
/kernel/linux/linux-6.6/arch/powerpc/kernel/
H A Diommu.c45 struct iommu_table *tbl = data; in iommu_debugfs_weight_get() local
46 *val = bitmap_weight(tbl->it_map, tbl->it_size); in iommu_debugfs_weight_get()
51 static void iommu_debugfs_add(struct iommu_table *tbl) in iommu_debugfs_add() argument
56 sprintf(name, "%08lx", tbl->it_index); in iommu_debugfs_add()
59 debugfs_create_file_unsafe("weight", 0400, liobn_entry, tbl, &iommu_debugfs_fops_weight); in iommu_debugfs_add()
60 debugfs_create_ulong("it_size", 0400, liobn_entry, &tbl->it_size); in iommu_debugfs_add()
61 debugfs_create_ulong("it_page_shift", 0400, liobn_entry, &tbl->it_page_shift); in iommu_debugfs_add()
62 debugfs_create_ulong("it_reserved_start", 0400, liobn_entry, &tbl->it_reserved_start); in iommu_debugfs_add()
63 debugfs_create_ulong("it_reserved_end", 0400, liobn_entry, &tbl in iommu_debugfs_add()
68 iommu_debugfs_del(struct iommu_table *tbl) iommu_debugfs_del() argument
76 iommu_debugfs_add(struct iommu_table *tbl) iommu_debugfs_add() argument
77 iommu_debugfs_del(struct iommu_table *tbl) iommu_debugfs_del() argument
213 iommu_range_alloc(struct device *dev, struct iommu_table *tbl, unsigned long npages, unsigned long *handle, unsigned long mask, unsigned int align_order) iommu_range_alloc() argument
345 iommu_alloc(struct device *dev, struct iommu_table *tbl, void *page, unsigned int npages, enum dma_data_direction direction, unsigned long mask, unsigned int align_order, unsigned long attrs) iommu_alloc() argument
388 iommu_free_check(struct iommu_table *tbl, dma_addr_t dma_addr, unsigned int npages) iommu_free_check() argument
416 get_pool(struct iommu_table *tbl, unsigned long entry) get_pool() argument
435 __iommu_free(struct iommu_table *tbl, dma_addr_t dma_addr, unsigned int npages) __iommu_free() argument
457 iommu_free(struct iommu_table *tbl, dma_addr_t dma_addr, unsigned int npages) iommu_free() argument
470 ppc_iommu_map_sg(struct device *dev, struct iommu_table *tbl, struct scatterlist *sglist, int nelems, unsigned long mask, enum dma_data_direction direction, unsigned long attrs) ppc_iommu_map_sg() argument
613 ppc_iommu_unmap_sg(struct iommu_table *tbl, struct scatterlist *sglist, int nelems, enum dma_data_direction direction, unsigned long attrs) ppc_iommu_unmap_sg() argument
645 iommu_table_clear(struct iommu_table *tbl) iommu_table_clear() argument
686 iommu_table_reserve_pages(struct iommu_table *tbl, unsigned long res_start, unsigned long res_end) iommu_table_reserve_pages() argument
724 iommu_init_table(struct iommu_table *tbl, int nid, unsigned long res_start, unsigned long res_end) iommu_init_table() argument
781 iommu_table_in_use(struct iommu_table *tbl) iommu_table_in_use() argument
804 struct iommu_table *tbl; iommu_table_free() local
829 iommu_tce_table_get(struct iommu_table *tbl) iommu_tce_table_get() argument
838 iommu_tce_table_put(struct iommu_table *tbl) iommu_tce_table_put() argument
852 iommu_map_page(struct device *dev, struct iommu_table *tbl, struct page *page, unsigned long offset, size_t size, unsigned long mask, enum dma_data_direction direction, unsigned long attrs) iommu_map_page() argument
891 iommu_unmap_page(struct iommu_table *tbl, dma_addr_t dma_handle, size_t size, enum dma_data_direction direction, unsigned long attrs) iommu_unmap_page() argument
910 iommu_alloc_coherent(struct device *dev, struct iommu_table *tbl, size_t size, dma_addr_t *dma_handle, unsigned long mask, gfp_t flag, int node) iommu_alloc_coherent() argument
960 iommu_free_coherent(struct iommu_table *tbl, size_t size, void *vaddr, dma_addr_t dma_handle) iommu_free_coherent() argument
1035 iommu_flush_tce(struct iommu_table *tbl) iommu_flush_tce() argument
1077 iommu_tce_xchg_no_kill(struct mm_struct *mm, struct iommu_table *tbl, unsigned long entry, unsigned long *hpa, enum dma_data_direction *direction) iommu_tce_xchg_no_kill() argument
1096 iommu_tce_kill(struct iommu_table *tbl, unsigned long entry, unsigned long pages) iommu_tce_kill() argument
1105 iommu_take_ownership(struct iommu_table *tbl) iommu_take_ownership() argument
1138 iommu_release_ownership(struct iommu_table *tbl) iommu_release_ownership() argument
1209 struct iommu_table *tbl = table_group->tables[0]; spapr_tce_create_table() local
1223 spapr_tce_set_window(struct iommu_table_group *table_group, int num, struct iommu_table *tbl) spapr_tce_set_window() argument
1239 struct iommu_table *tbl = table_group->tables[i]; spapr_tce_take_ownership() local
1260 struct iommu_table *tbl = table_group->tables[i]; spapr_tce_release_ownership() local
[all...]
/kernel/linux/linux-5.10/drivers/infiniband/hw/vmw_pvrdma/
H A Dpvrdma_doorbell.c56 struct pvrdma_id_table *tbl = &dev->uar_table.tbl; in pvrdma_uar_table_init() local
61 tbl->last = 0; in pvrdma_uar_table_init()
62 tbl->top = 0; in pvrdma_uar_table_init()
63 tbl->max = num; in pvrdma_uar_table_init()
64 tbl->mask = mask; in pvrdma_uar_table_init()
65 spin_lock_init(&tbl->lock); in pvrdma_uar_table_init()
66 tbl->table = kcalloc(BITS_TO_LONGS(num), sizeof(long), GFP_KERNEL); in pvrdma_uar_table_init()
67 if (!tbl->table) in pvrdma_uar_table_init()
71 set_bit(0, tbl in pvrdma_uar_table_init()
78 struct pvrdma_id_table *tbl = &dev->uar_table.tbl; pvrdma_uar_table_cleanup() local
85 struct pvrdma_id_table *tbl; pvrdma_uar_alloc() local
117 struct pvrdma_id_table *tbl = &dev->uar_table.tbl; pvrdma_uar_free() local
[all...]
/kernel/linux/linux-6.6/drivers/infiniband/hw/vmw_pvrdma/
H A Dpvrdma_doorbell.c56 struct pvrdma_id_table *tbl = &dev->uar_table.tbl; in pvrdma_uar_table_init() local
61 tbl->last = 0; in pvrdma_uar_table_init()
62 tbl->top = 0; in pvrdma_uar_table_init()
63 tbl->max = num; in pvrdma_uar_table_init()
64 tbl->mask = mask; in pvrdma_uar_table_init()
65 spin_lock_init(&tbl->lock); in pvrdma_uar_table_init()
66 tbl->table = bitmap_zalloc(num, GFP_KERNEL); in pvrdma_uar_table_init()
67 if (!tbl->table) in pvrdma_uar_table_init()
71 __set_bit(0, tbl in pvrdma_uar_table_init()
78 struct pvrdma_id_table *tbl = &dev->uar_table.tbl; pvrdma_uar_table_cleanup() local
85 struct pvrdma_id_table *tbl; pvrdma_uar_alloc() local
117 struct pvrdma_id_table *tbl = &dev->uar_table.tbl; pvrdma_uar_free() local
[all...]
/kernel/linux/linux-5.10/include/linux/
H A Drhashtable.h119 static inline unsigned int rht_bucket_index(const struct bucket_table *tbl, in rht_bucket_index() argument
122 return hash & (tbl->size - 1); in rht_bucket_index()
156 struct rhashtable *ht, const struct bucket_table *tbl, in rht_key_hashfn()
159 unsigned int hash = rht_key_get_hash(ht, key, params, tbl->hash_rnd); in rht_key_hashfn()
161 return rht_bucket_index(tbl, hash); in rht_key_hashfn()
165 struct rhashtable *ht, const struct bucket_table *tbl, in rht_head_hashfn()
171 rht_bucket_index(tbl, params.obj_hashfn(ptr, params.key_len ?: in rht_head_hashfn()
173 tbl->hash_rnd)) : in rht_head_hashfn()
174 rht_key_hashfn(ht, tbl, ptr + params.key_offset, params); in rht_head_hashfn()
180 * @tbl
155 rht_key_hashfn( struct rhashtable *ht, const struct bucket_table *tbl, const void *key, const struct rhashtable_params params) rht_key_hashfn() argument
164 rht_head_hashfn( struct rhashtable *ht, const struct bucket_table *tbl, const struct rhash_head *he, const struct rhashtable_params params) rht_head_hashfn() argument
182 rht_grow_above_75(const struct rhashtable *ht, const struct bucket_table *tbl) rht_grow_above_75() argument
195 rht_shrink_below_30(const struct rhashtable *ht, const struct bucket_table *tbl) rht_shrink_below_30() argument
208 rht_grow_above_100(const struct rhashtable *ht, const struct bucket_table *tbl) rht_grow_above_100() argument
220 rht_grow_above_max(const struct rhashtable *ht, const struct bucket_table *tbl) rht_grow_above_max() argument
235 lockdep_rht_bucket_is_held(const struct bucket_table *tbl, u32 hash) lockdep_rht_bucket_is_held() argument
286 rht_bucket( const struct bucket_table *tbl, unsigned int hash) rht_bucket() argument
293 rht_bucket_var( struct bucket_table *tbl, unsigned int hash) rht_bucket_var() argument
300 rht_bucket_insert( struct rhashtable *ht, struct bucket_table *tbl, unsigned int hash) rht_bucket_insert() argument
326 rht_lock(struct bucket_table *tbl, struct rhash_lock_head __rcu **bkt) rht_lock() argument
334 rht_lock_nested(struct bucket_table *tbl, struct rhash_lock_head __rcu **bucket, unsigned int subclass) rht_lock_nested() argument
343 rht_unlock(struct bucket_table *tbl, struct rhash_lock_head __rcu **bkt) rht_unlock() argument
372 rht_ptr( struct rhash_lock_head __rcu *const *bkt, struct bucket_table *tbl, unsigned int hash) rht_ptr() argument
394 rht_assign_unlock(struct bucket_table *tbl, struct rhash_lock_head __rcu **bkt, struct rhash_head *obj) rht_assign_unlock() argument
590 struct bucket_table *tbl; __rhashtable_lookup() local
707 struct bucket_table *tbl; __rhashtable_insert_fast() local
986 __rhashtable_remove_fast_one( struct rhashtable *ht, struct bucket_table *tbl, struct rhash_head *obj, const struct rhashtable_params params, bool rhlist) __rhashtable_remove_fast_one() argument
1071 struct bucket_table *tbl; __rhashtable_remove_fast() local
1138 __rhashtable_replace_fast( struct rhashtable *ht, struct bucket_table *tbl, struct rhash_head *obj_old, struct rhash_head *obj_new, const struct rhashtable_params params) __rhashtable_replace_fast() argument
1205 struct bucket_table *tbl; rhashtable_replace_fast() local
[all...]
/kernel/linux/linux-6.6/include/linux/
H A Drhashtable.h119 static inline unsigned int rht_bucket_index(const struct bucket_table *tbl, in rht_bucket_index() argument
122 return hash & (tbl->size - 1); in rht_bucket_index()
156 struct rhashtable *ht, const struct bucket_table *tbl, in rht_key_hashfn()
159 unsigned int hash = rht_key_get_hash(ht, key, params, tbl->hash_rnd); in rht_key_hashfn()
161 return rht_bucket_index(tbl, hash); in rht_key_hashfn()
165 struct rhashtable *ht, const struct bucket_table *tbl, in rht_head_hashfn()
171 rht_bucket_index(tbl, params.obj_hashfn(ptr, params.key_len ?: in rht_head_hashfn()
173 tbl->hash_rnd)) : in rht_head_hashfn()
174 rht_key_hashfn(ht, tbl, ptr + params.key_offset, params); in rht_head_hashfn()
180 * @tbl
155 rht_key_hashfn( struct rhashtable *ht, const struct bucket_table *tbl, const void *key, const struct rhashtable_params params) rht_key_hashfn() argument
164 rht_head_hashfn( struct rhashtable *ht, const struct bucket_table *tbl, const struct rhash_head *he, const struct rhashtable_params params) rht_head_hashfn() argument
182 rht_grow_above_75(const struct rhashtable *ht, const struct bucket_table *tbl) rht_grow_above_75() argument
195 rht_shrink_below_30(const struct rhashtable *ht, const struct bucket_table *tbl) rht_shrink_below_30() argument
208 rht_grow_above_100(const struct rhashtable *ht, const struct bucket_table *tbl) rht_grow_above_100() argument
220 rht_grow_above_max(const struct rhashtable *ht, const struct bucket_table *tbl) rht_grow_above_max() argument
235 lockdep_rht_bucket_is_held(const struct bucket_table *tbl, u32 hash) lockdep_rht_bucket_is_held() argument
286 rht_bucket( const struct bucket_table *tbl, unsigned int hash) rht_bucket() argument
293 rht_bucket_var( struct bucket_table *tbl, unsigned int hash) rht_bucket_var() argument
300 rht_bucket_insert( struct rhashtable *ht, struct bucket_table *tbl, unsigned int hash) rht_bucket_insert() argument
326 rht_lock(struct bucket_table *tbl, struct rhash_lock_head __rcu **bkt) rht_lock() argument
337 rht_lock_nested(struct bucket_table *tbl, struct rhash_lock_head __rcu **bucket, unsigned int subclass) rht_lock_nested() argument
349 rht_unlock(struct bucket_table *tbl, struct rhash_lock_head __rcu **bkt, unsigned long flags) rht_unlock() argument
379 rht_ptr( struct rhash_lock_head __rcu *const *bkt, struct bucket_table *tbl, unsigned int hash) rht_ptr() argument
401 rht_assign_unlock(struct bucket_table *tbl, struct rhash_lock_head __rcu **bkt, struct rhash_head *obj, unsigned long flags) rht_assign_unlock() argument
598 struct bucket_table *tbl; __rhashtable_lookup() local
715 struct bucket_table *tbl; __rhashtable_insert_fast() local
995 __rhashtable_remove_fast_one( struct rhashtable *ht, struct bucket_table *tbl, struct rhash_head *obj, const struct rhashtable_params params, bool rhlist) __rhashtable_remove_fast_one() argument
1081 struct bucket_table *tbl; __rhashtable_remove_fast() local
1148 __rhashtable_replace_fast( struct rhashtable *ht, struct bucket_table *tbl, struct rhash_head *obj_old, struct rhash_head *obj_new, const struct rhashtable_params params) __rhashtable_replace_fast() argument
1216 struct bucket_table *tbl; rhashtable_replace_fast() local
[all...]
/kernel/linux/linux-5.10/tools/perf/util/
H A Dsyscalltbl.c63 static int syscalltbl__init_native(struct syscalltbl *tbl) in syscalltbl__init_native() argument
72 entries = tbl->syscalls.entries = malloc(sizeof(struct syscall) * nr_entries); in syscalltbl__init_native()
73 if (tbl->syscalls.entries == NULL) in syscalltbl__init_native()
84 qsort(tbl->syscalls.entries, nr_entries, sizeof(struct syscall), syscallcmp); in syscalltbl__init_native()
85 tbl->syscalls.nr_entries = nr_entries; in syscalltbl__init_native()
86 tbl->syscalls.max_id = syscalltbl_native_max_id; in syscalltbl__init_native()
92 struct syscalltbl *tbl = malloc(sizeof(*tbl)); in syscalltbl__new() local
93 if (tbl) { in syscalltbl__new()
94 if (syscalltbl__init_native(tbl)) { in syscalltbl__new()
102 syscalltbl__delete(struct syscalltbl *tbl) syscalltbl__delete() argument
113 syscalltbl__id(struct syscalltbl *tbl, const char *name) syscalltbl__id() argument
122 syscalltbl__strglobmatch_next(struct syscalltbl *tbl, const char *syscall_glob, int *idx) syscalltbl__strglobmatch_next() argument
137 syscalltbl__strglobmatch_first(struct syscalltbl *tbl, const char *syscall_glob, int *idx) syscalltbl__strglobmatch_first() argument
149 struct syscalltbl *tbl = zalloc(sizeof(*tbl)); syscalltbl__new() local
155 syscalltbl__delete(struct syscalltbl *tbl) syscalltbl__delete() argument
160 syscalltbl__name(const struct syscalltbl *tbl, int id) syscalltbl__name() argument
165 syscalltbl__id(struct syscalltbl *tbl, const char *name) syscalltbl__id() argument
176 syscalltbl__strglobmatch_first(struct syscalltbl *tbl, const char *syscall_glob, int *idx) syscalltbl__strglobmatch_first() argument
[all...]
/kernel/linux/linux-6.6/tools/perf/util/
H A Dsyscalltbl.c67 static int syscalltbl__init_native(struct syscalltbl *tbl) in syscalltbl__init_native() argument
76 entries = tbl->syscalls.entries = malloc(sizeof(struct syscall) * nr_entries); in syscalltbl__init_native()
77 if (tbl->syscalls.entries == NULL) in syscalltbl__init_native()
88 qsort(tbl->syscalls.entries, nr_entries, sizeof(struct syscall), syscallcmp); in syscalltbl__init_native()
89 tbl->syscalls.nr_entries = nr_entries; in syscalltbl__init_native()
90 tbl->syscalls.max_id = syscalltbl_native_max_id; in syscalltbl__init_native()
96 struct syscalltbl *tbl = malloc(sizeof(*tbl)); in syscalltbl__new() local
97 if (tbl) { in syscalltbl__new()
98 if (syscalltbl__init_native(tbl)) { in syscalltbl__new()
106 syscalltbl__delete(struct syscalltbl *tbl) syscalltbl__delete() argument
117 syscalltbl__id(struct syscalltbl *tbl, const char *name) syscalltbl__id() argument
126 syscalltbl__strglobmatch_next(struct syscalltbl *tbl, const char *syscall_glob, int *idx) syscalltbl__strglobmatch_next() argument
141 syscalltbl__strglobmatch_first(struct syscalltbl *tbl, const char *syscall_glob, int *idx) syscalltbl__strglobmatch_first() argument
153 struct syscalltbl *tbl = zalloc(sizeof(*tbl)); syscalltbl__new() local
159 syscalltbl__delete(struct syscalltbl *tbl) syscalltbl__delete() argument
164 syscalltbl__name(const struct syscalltbl *tbl, int id) syscalltbl__name() argument
169 syscalltbl__id(struct syscalltbl *tbl, const char *name) syscalltbl__id() argument
180 syscalltbl__strglobmatch_first(struct syscalltbl *tbl, const char *syscall_glob, int *idx) syscalltbl__strglobmatch_first() argument
[all...]
/kernel/linux/linux-5.10/drivers/net/wireless/intel/iwlwifi/dvm/
H A Drs.c436 static s32 get_expected_tpt(struct iwl_scale_tbl_info *tbl, int rs_index) in get_expected_tpt() argument
438 if (tbl->expected_tpt) in get_expected_tpt()
439 return tbl->expected_tpt[rs_index]; in get_expected_tpt()
450 static int rs_collect_tx_data(struct iwl_scale_tbl_info *tbl, in rs_collect_tx_data() argument
461 window = &(tbl->win[scale_index]); in rs_collect_tx_data()
464 tpt = get_expected_tpt(tbl, scale_index); in rs_collect_tx_data()
529 struct iwl_scale_tbl_info *tbl, in rate_n_flags_from_tbl()
534 if (is_legacy(tbl->lq_type)) { in rate_n_flags_from_tbl()
539 } else if (is_Ht(tbl->lq_type)) { in rate_n_flags_from_tbl()
546 if (is_siso(tbl in rate_n_flags_from_tbl()
528 rate_n_flags_from_tbl(struct iwl_priv *priv, struct iwl_scale_tbl_info *tbl, int index, u8 use_green) rate_n_flags_from_tbl() argument
584 rs_get_tbl_info_from_mcs(const u32 rate_n_flags, enum nl80211_band band, struct iwl_scale_tbl_info *tbl, int *rate_idx) rs_get_tbl_info_from_mcs() argument
650 rs_toggle_antenna(u32 valid_ant, u32 *rate_n_flags, struct iwl_scale_tbl_info *tbl) rs_toggle_antenna() argument
774 rs_get_lower_rate(struct iwl_lq_sta *lq_sta, struct iwl_scale_tbl_info *tbl, u8 scale_index, u8 ht_possible) rs_get_lower_rate() argument
847 struct iwl_scale_tbl_info *tbl; rs_bt_update_lq() local
1086 rs_set_expected_tpt_table(struct iwl_lq_sta *lq_sta, struct iwl_scale_tbl_info *tbl) rs_set_expected_tpt_table() argument
1142 rs_get_best_rate(struct iwl_priv *priv, struct iwl_lq_sta *lq_sta, struct iwl_scale_tbl_info *tbl, u16 rate_mask, s8 index) rs_get_best_rate() argument
1235 rs_switch_to_mimo2(struct iwl_priv *priv, struct iwl_lq_sta *lq_sta, struct ieee80211_conf *conf, struct ieee80211_sta *sta, struct iwl_scale_tbl_info *tbl, int index) rs_switch_to_mimo2() argument
1290 rs_switch_to_mimo3(struct iwl_priv *priv, struct iwl_lq_sta *lq_sta, struct ieee80211_conf *conf, struct ieee80211_sta *sta, struct iwl_scale_tbl_info *tbl, int index) rs_switch_to_mimo3() argument
1346 rs_switch_to_siso(struct iwl_priv *priv, struct iwl_lq_sta *lq_sta, struct ieee80211_conf *conf, struct ieee80211_sta *sta, struct iwl_scale_tbl_info *tbl, int index) rs_switch_to_siso() argument
1401 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); rs_move_legacy_other() local
1574 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); rs_move_siso_to_other() local
1744 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); rs_move_mimo2_to_other() local
1912 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); rs_move_mimo3_to_other() local
2085 struct iwl_scale_tbl_info *tbl; rs_stay_in_table() local
2162 rs_update_rate_tbl(struct iwl_priv *priv, struct iwl_rxon_context *ctx, struct iwl_lq_sta *lq_sta, struct iwl_scale_tbl_info *tbl, int index, u8 is_green) rs_update_rate_tbl() argument
2200 struct iwl_scale_tbl_info *tbl, *tbl1; rs_rate_scale_perform() local
2633 struct iwl_scale_tbl_info *tbl; rs_initialize_lq() local
3099 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); rs_sta_dbgfs_scale_table_read() local
3224 struct iwl_scale_tbl_info *tbl = &lq_sta->lq_info[lq_sta->active_tbl]; rs_sta_dbgfs_rate_scale_data_read() local
[all...]
/kernel/linux/linux-6.6/drivers/net/wireless/intel/iwlwifi/dvm/
H A Drs.c414 static s32 get_expected_tpt(struct iwl_scale_tbl_info *tbl, int rs_index) in get_expected_tpt() argument
416 if (tbl->expected_tpt) in get_expected_tpt()
417 return tbl->expected_tpt[rs_index]; in get_expected_tpt()
428 static int rs_collect_tx_data(struct iwl_scale_tbl_info *tbl, in rs_collect_tx_data() argument
439 window = &(tbl->win[scale_index]); in rs_collect_tx_data()
442 tpt = get_expected_tpt(tbl, scale_index); in rs_collect_tx_data()
507 struct iwl_scale_tbl_info *tbl, in rate_n_flags_from_tbl()
512 if (is_legacy(tbl->lq_type)) { in rate_n_flags_from_tbl()
517 } else if (is_Ht(tbl->lq_type)) { in rate_n_flags_from_tbl()
524 if (is_siso(tbl in rate_n_flags_from_tbl()
506 rate_n_flags_from_tbl(struct iwl_priv *priv, struct iwl_scale_tbl_info *tbl, int index, u8 use_green) rate_n_flags_from_tbl() argument
562 rs_get_tbl_info_from_mcs(const u32 rate_n_flags, enum nl80211_band band, struct iwl_scale_tbl_info *tbl, int *rate_idx) rs_get_tbl_info_from_mcs() argument
628 rs_toggle_antenna(u32 valid_ant, u32 *rate_n_flags, struct iwl_scale_tbl_info *tbl) rs_toggle_antenna() argument
752 rs_get_lower_rate(struct iwl_lq_sta *lq_sta, struct iwl_scale_tbl_info *tbl, u8 scale_index, u8 ht_possible) rs_get_lower_rate() argument
825 struct iwl_scale_tbl_info *tbl; rs_bt_update_lq() local
1064 rs_set_expected_tpt_table(struct iwl_lq_sta *lq_sta, struct iwl_scale_tbl_info *tbl) rs_set_expected_tpt_table() argument
1120 rs_get_best_rate(struct iwl_priv *priv, struct iwl_lq_sta *lq_sta, struct iwl_scale_tbl_info *tbl, u16 rate_mask, s8 index) rs_get_best_rate() argument
1213 rs_switch_to_mimo2(struct iwl_priv *priv, struct iwl_lq_sta *lq_sta, struct ieee80211_conf *conf, struct ieee80211_sta *sta, struct iwl_scale_tbl_info *tbl, int index) rs_switch_to_mimo2() argument
1268 rs_switch_to_mimo3(struct iwl_priv *priv, struct iwl_lq_sta *lq_sta, struct ieee80211_conf *conf, struct ieee80211_sta *sta, struct iwl_scale_tbl_info *tbl, int index) rs_switch_to_mimo3() argument
1324 rs_switch_to_siso(struct iwl_priv *priv, struct iwl_lq_sta *lq_sta, struct ieee80211_conf *conf, struct ieee80211_sta *sta, struct iwl_scale_tbl_info *tbl, int index) rs_switch_to_siso() argument
1379 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); rs_move_legacy_other() local
1552 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); rs_move_siso_to_other() local
1722 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); rs_move_mimo2_to_other() local
1890 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); rs_move_mimo3_to_other() local
2063 struct iwl_scale_tbl_info *tbl; rs_stay_in_table() local
2140 rs_update_rate_tbl(struct iwl_priv *priv, struct iwl_rxon_context *ctx, struct iwl_lq_sta *lq_sta, struct iwl_scale_tbl_info *tbl, int index, u8 is_green) rs_update_rate_tbl() argument
2178 struct iwl_scale_tbl_info *tbl, *tbl1; rs_rate_scale_perform() local
2611 struct iwl_scale_tbl_info *tbl; rs_initialize_lq() local
3094 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); rs_sta_dbgfs_scale_table_read() local
3222 struct iwl_scale_tbl_info *tbl = &lq_sta->lq_info[lq_sta->active_tbl]; rs_sta_dbgfs_rate_scale_data_read() local
[all...]
/kernel/linux/linux-5.10/arch/powerpc/platforms/powernv/
H A Dpci-ioda-tce.c48 void pnv_pci_setup_iommu_table(struct iommu_table *tbl, in pnv_pci_setup_iommu_table() argument
52 tbl->it_blocksize = 16; in pnv_pci_setup_iommu_table()
53 tbl->it_base = (unsigned long)tce_mem; in pnv_pci_setup_iommu_table()
54 tbl->it_page_shift = page_shift; in pnv_pci_setup_iommu_table()
55 tbl->it_offset = dma_offset >> tbl->it_page_shift; in pnv_pci_setup_iommu_table()
56 tbl->it_index = 0; in pnv_pci_setup_iommu_table()
57 tbl->it_size = tce_size >> 3; in pnv_pci_setup_iommu_table()
58 tbl->it_busno = 0; in pnv_pci_setup_iommu_table()
59 tbl in pnv_pci_setup_iommu_table()
83 pnv_tce(struct iommu_table *tbl, bool user, long idx, bool alloc) pnv_tce() argument
124 pnv_tce_build(struct iommu_table *tbl, long index, long npages, unsigned long uaddr, enum dma_data_direction direction, unsigned long attrs) pnv_tce_build() argument
147 pnv_tce_xchg(struct iommu_table *tbl, long index, unsigned long *hpa, enum dma_data_direction *direction, bool alloc) pnv_tce_xchg() argument
182 pnv_tce_useraddrptr(struct iommu_table *tbl, long index, bool alloc) pnv_tce_useraddrptr() argument
191 pnv_tce_free(struct iommu_table *tbl, long index, long npages) pnv_tce_free() argument
207 pnv_tce_get(struct iommu_table *tbl, long index) pnv_tce_get() argument
241 pnv_pci_ioda2_table_free_pages(struct iommu_table *tbl) pnv_pci_ioda2_table_free_pages() argument
291 pnv_pci_ioda2_table_alloc_pages(int nid, __u64 bus_offset, __u32 page_shift, __u64 window_size, __u32 levels, bool alloc_userspace_copy, struct iommu_table *tbl) pnv_pci_ioda2_table_alloc_pages() argument
371 pnv_pci_unlink_table_and_group(struct iommu_table *tbl, struct iommu_table_group *table_group) pnv_pci_unlink_table_and_group() argument
411 pnv_pci_link_table_and_group(int node, int num, struct iommu_table *tbl, struct iommu_table_group *table_group) pnv_pci_link_table_and_group() argument
[all...]
/kernel/linux/linux-6.6/arch/powerpc/platforms/powernv/
H A Dpci-ioda-tce.c48 void pnv_pci_setup_iommu_table(struct iommu_table *tbl, in pnv_pci_setup_iommu_table() argument
52 tbl->it_blocksize = 16; in pnv_pci_setup_iommu_table()
53 tbl->it_base = (unsigned long)tce_mem; in pnv_pci_setup_iommu_table()
54 tbl->it_page_shift = page_shift; in pnv_pci_setup_iommu_table()
55 tbl->it_offset = dma_offset >> tbl->it_page_shift; in pnv_pci_setup_iommu_table()
56 tbl->it_index = 0; in pnv_pci_setup_iommu_table()
57 tbl->it_size = tce_size >> 3; in pnv_pci_setup_iommu_table()
58 tbl->it_busno = 0; in pnv_pci_setup_iommu_table()
59 tbl in pnv_pci_setup_iommu_table()
83 pnv_tce(struct iommu_table *tbl, bool user, long idx, bool alloc) pnv_tce() argument
124 pnv_tce_build(struct iommu_table *tbl, long index, long npages, unsigned long uaddr, enum dma_data_direction direction, unsigned long attrs) pnv_tce_build() argument
147 pnv_tce_xchg(struct iommu_table *tbl, long index, unsigned long *hpa, enum dma_data_direction *direction) pnv_tce_xchg() argument
181 pnv_tce_useraddrptr(struct iommu_table *tbl, long index, bool alloc) pnv_tce_useraddrptr() argument
190 pnv_tce_free(struct iommu_table *tbl, long index, long npages) pnv_tce_free() argument
206 pnv_tce_get(struct iommu_table *tbl, long index) pnv_tce_get() argument
240 pnv_pci_ioda2_table_free_pages(struct iommu_table *tbl) pnv_pci_ioda2_table_free_pages() argument
290 pnv_pci_ioda2_table_alloc_pages(int nid, __u64 bus_offset, __u32 page_shift, __u64 window_size, __u32 levels, bool alloc_userspace_copy, struct iommu_table *tbl) pnv_pci_ioda2_table_alloc_pages() argument
370 pnv_pci_unlink_table_and_group(struct iommu_table *tbl, struct iommu_table_group *table_group) pnv_pci_unlink_table_and_group() argument
410 pnv_pci_link_table_and_group(int node, int num, struct iommu_table *tbl, struct iommu_table_group *table_group) pnv_pci_link_table_and_group() argument
[all...]
/kernel/linux/linux-5.10/lib/
H A Drhashtable.c38 const struct bucket_table *tbl, in head_hashfn()
41 return rht_head_hashfn(ht, tbl, he, ht->p); in head_hashfn()
53 int lockdep_rht_bucket_is_held(const struct bucket_table *tbl, u32 hash) in lockdep_rht_bucket_is_held() argument
57 if (unlikely(tbl->nest)) in lockdep_rht_bucket_is_held()
59 return bit_spin_is_locked(0, (unsigned long *)&tbl->buckets[hash]); in lockdep_rht_bucket_is_held()
67 const struct bucket_table *tbl) in nested_table_top()
70 * because it's set at the same time as tbl->nest. in nested_table_top()
72 return (void *)rcu_dereference_protected(tbl->buckets[0], 1); in nested_table_top()
94 static void nested_bucket_table_free(const struct bucket_table *tbl) in nested_bucket_table_free() argument
96 unsigned int size = tbl in nested_bucket_table_free()
37 head_hashfn(struct rhashtable *ht, const struct bucket_table *tbl, const struct rhash_head *he) head_hashfn() argument
66 nested_table_top( const struct bucket_table *tbl) nested_table_top() argument
109 bucket_table_free(const struct bucket_table *tbl) bucket_table_free() argument
152 struct bucket_table *tbl; nested_bucket_table_alloc() local
179 struct bucket_table *tbl = NULL; bucket_table_alloc() local
211 rhashtable_last_table(struct rhashtable *ht, struct bucket_table *tbl) rhashtable_last_table() argument
410 struct bucket_table *tbl; rht_deferred_worker() local
439 rhashtable_insert_rehash(struct rhashtable *ht, struct bucket_table *tbl) rhashtable_insert_rehash() argument
487 rhashtable_lookup_one(struct rhashtable *ht, struct rhash_lock_head __rcu **bkt, struct bucket_table *tbl, unsigned int hash, const void *key, struct rhash_head *obj) rhashtable_lookup_one() argument
538 rhashtable_insert_one( struct rhashtable *ht, struct rhash_lock_head __rcu **bkt, struct bucket_table *tbl, unsigned int hash, struct rhash_head *obj, void *data) rhashtable_insert_one() argument
591 struct bucket_table *tbl; rhashtable_try_insert() local
789 struct bucket_table *tbl = iter->walker.tbl; __rhashtable_walk_find_next() local
934 struct bucket_table *tbl = iter->walker.tbl; __releases() local
1018 struct bucket_table *tbl; rhashtable_init() local
1134 struct bucket_table *tbl, *next_tbl; rhashtable_free_and_destroy() local
1174 __rht_bucket_nested( const struct bucket_table *tbl, unsigned int hash) __rht_bucket_nested() argument
1203 rht_bucket_nested( const struct bucket_table *tbl, unsigned int hash) rht_bucket_nested() argument
1214 rht_bucket_nested_insert( struct rhashtable *ht, struct bucket_table *tbl, unsigned int hash) rht_bucket_nested_insert() argument
[all...]
/kernel/linux/linux-6.6/lib/
H A Drhashtable.c38 const struct bucket_table *tbl, in head_hashfn()
41 return rht_head_hashfn(ht, tbl, he, ht->p); in head_hashfn()
53 int lockdep_rht_bucket_is_held(const struct bucket_table *tbl, u32 hash) in lockdep_rht_bucket_is_held() argument
57 if (unlikely(tbl->nest)) in lockdep_rht_bucket_is_held()
59 return bit_spin_is_locked(0, (unsigned long *)&tbl->buckets[hash]); in lockdep_rht_bucket_is_held()
67 const struct bucket_table *tbl) in nested_table_top()
70 * because it's set at the same time as tbl->nest. in nested_table_top()
72 return (void *)rcu_dereference_protected(tbl->buckets[0], 1); in nested_table_top()
94 static void nested_bucket_table_free(const struct bucket_table *tbl) in nested_bucket_table_free() argument
96 unsigned int size = tbl in nested_bucket_table_free()
37 head_hashfn(struct rhashtable *ht, const struct bucket_table *tbl, const struct rhash_head *he) head_hashfn() argument
66 nested_table_top( const struct bucket_table *tbl) nested_table_top() argument
109 bucket_table_free(const struct bucket_table *tbl) bucket_table_free() argument
152 struct bucket_table *tbl; nested_bucket_table_alloc() local
179 struct bucket_table *tbl = NULL; bucket_table_alloc() local
211 rhashtable_last_table(struct rhashtable *ht, struct bucket_table *tbl) rhashtable_last_table() argument
413 struct bucket_table *tbl; rht_deferred_worker() local
442 rhashtable_insert_rehash(struct rhashtable *ht, struct bucket_table *tbl) rhashtable_insert_rehash() argument
490 rhashtable_lookup_one(struct rhashtable *ht, struct rhash_lock_head __rcu **bkt, struct bucket_table *tbl, unsigned int hash, const void *key, struct rhash_head *obj) rhashtable_lookup_one() argument
541 rhashtable_insert_one( struct rhashtable *ht, struct rhash_lock_head __rcu **bkt, struct bucket_table *tbl, unsigned int hash, struct rhash_head *obj, void *data) rhashtable_insert_one() argument
594 struct bucket_table *tbl; rhashtable_try_insert() local
793 struct bucket_table *tbl = iter->walker.tbl; __rhashtable_walk_find_next() local
938 struct bucket_table *tbl = iter->walker.tbl; __releases() local
1022 struct bucket_table *tbl; rhashtable_init() local
1138 struct bucket_table *tbl, *next_tbl; rhashtable_free_and_destroy() local
1178 __rht_bucket_nested( const struct bucket_table *tbl, unsigned int hash) __rht_bucket_nested() argument
1207 rht_bucket_nested( const struct bucket_table *tbl, unsigned int hash) rht_bucket_nested() argument
1218 rht_bucket_nested_insert( struct rhashtable *ht, struct bucket_table *tbl, unsigned int hash) rht_bucket_nested_insert() argument
[all...]
/kernel/linux/linux-5.10/net/core/
H A Dneighbour.c58 static int pneigh_ifdown_and_unlock(struct neigh_table *tbl,
66 Neighbour hash table buckets are protected with rwlock tbl->lock.
124 atomic_dec(&n->tbl->gc_entries); in neigh_mark_dead()
132 write_lock_bh(&n->tbl->lock); in neigh_update_gc_list()
147 atomic_dec(&n->tbl->gc_entries); in neigh_update_gc_list()
150 list_add_tail(&n->gc_list, &n->tbl->gc_list); in neigh_update_gc_list()
151 atomic_inc(&n->tbl->gc_entries); in neigh_update_gc_list()
156 write_unlock_bh(&n->tbl->lock); in neigh_update_gc_list()
182 struct neigh_table *tbl) in neigh_del()
191 lockdep_is_held(&tbl in neigh_del()
181 neigh_del(struct neighbour *n, struct neighbour __rcu **np, struct neigh_table *tbl) neigh_del() argument
202 neigh_remove_one(struct neighbour *ndel, struct neigh_table *tbl) neigh_remove_one() argument
225 neigh_forced_gc(struct neigh_table *tbl) neigh_forced_gc() argument
326 neigh_flush_dev(struct neigh_table *tbl, struct net_device *dev, bool skip_perm) neigh_flush_dev() argument
380 neigh_changeaddr(struct neigh_table *tbl, struct net_device *dev) neigh_changeaddr() argument
388 __neigh_ifdown(struct neigh_table *tbl, struct net_device *dev, bool skip_perm) __neigh_ifdown() argument
400 neigh_carrier_down(struct neigh_table *tbl, struct net_device *dev) neigh_carrier_down() argument
407 neigh_ifdown(struct neigh_table *tbl, struct net_device *dev) neigh_ifdown() argument
414 neigh_alloc(struct neigh_table *tbl, struct net_device *dev, u8 flags, bool exempt_from_gc) neigh_alloc() argument
521 neigh_hash_grow(struct neigh_table *tbl, unsigned long new_shift) neigh_hash_grow() argument
562 neigh_lookup(struct neigh_table *tbl, const void *pkey, struct net_device *dev) neigh_lookup() argument
583 ___neigh_create(struct neigh_table *tbl, const void *pkey, struct net_device *dev, u8 flags, bool exempt_from_gc, bool want_ref) ___neigh_create() argument
677 __neigh_create(struct neigh_table *tbl, const void *pkey, struct net_device *dev, bool want_ref) __neigh_create() argument
710 __pneigh_lookup(struct neigh_table *tbl, struct net *net, const void *pkey, struct net_device *dev) __pneigh_lookup() argument
721 pneigh_lookup(struct neigh_table *tbl, struct net *net, const void *pkey, struct net_device *dev, int creat) pneigh_lookup() argument
767 pneigh_delete(struct neigh_table *tbl, struct net *net, const void *pkey, struct net_device *dev) pneigh_delete() argument
793 pneigh_ifdown_and_unlock(struct neigh_table *tbl, struct net_device *dev) pneigh_ifdown_and_unlock() argument
895 struct neigh_table *tbl = container_of(work, struct neigh_table, gc_work.work); neigh_periodic_work() local
1456 neigh_event_ns(struct neigh_table *tbl, u8 *lladdr, void *saddr, struct net_device *dev) neigh_event_ns() argument
1555 struct neigh_table *tbl = from_timer(tbl, t, proxy_timer); neigh_proxy_process() local
1587 pneigh_enqueue(struct neigh_table *tbl, struct neigh_parms *p, struct sk_buff *skb) pneigh_enqueue() argument
1616 lookup_neigh_parms(struct neigh_table *tbl, struct net *net, int ifindex) lookup_neigh_parms() argument
1630 neigh_parms_alloc(struct net_device *dev, struct neigh_table *tbl) neigh_parms_alloc() argument
1672 neigh_parms_release(struct neigh_table *tbl, struct neigh_parms *parms) neigh_parms_release() argument
1695 neigh_table_init(int index, struct neigh_table *tbl) neigh_table_init() argument
1747 neigh_table_clear(int index, struct neigh_table *tbl) neigh_table_clear() argument
1776 struct neigh_table *tbl = NULL; neigh_find_table() local
1817 struct neigh_table *tbl; neigh_delete() local
1884 struct neigh_table *tbl; neigh_add() local
2054 neightbl_fill_info(struct sk_buff *skb, struct neigh_table *tbl, u32 pid, u32 seq, int type, int flags) neightbl_fill_info() argument
2144 neightbl_fill_param_info(struct sk_buff *skb, struct neigh_table *tbl, struct neigh_parms *parms, u32 pid, u32 seq, int type, unsigned int flags) neightbl_fill_param_info() argument
2207 struct neigh_table *tbl; neightbl_set() local
2396 struct neigh_table *tbl; neightbl_dump_info() local
2506 pneigh_fill_info(struct sk_buff *skb, struct pneigh_entry *pn, u32 pid, u32 seq, int type, unsigned int flags, struct neigh_table *tbl) pneigh_fill_info() argument
2573 neigh_dump_table(struct neigh_table *tbl, struct sk_buff *skb, struct netlink_callback *cb, struct neigh_dump_filter *filter) neigh_dump_table() argument
2620 pneigh_dump_table(struct neigh_table *tbl, struct sk_buff *skb, struct netlink_callback *cb, struct neigh_dump_filter *filter) pneigh_dump_table() argument
2730 struct neigh_table *tbl; neigh_dump_info() local
2772 neigh_valid_get_req(const struct nlmsghdr *nlh, struct neigh_table **tbl, void **dst, int *dev_idx, u8 *ndm_flags, struct netlink_ext_ack *extack) neigh_valid_get_req() argument
2870 pneigh_get_reply(struct net *net, struct pneigh_entry *neigh, u32 pid, u32 seq, struct neigh_table *tbl) pneigh_get_reply() argument
2896 struct neigh_table *tbl = NULL; neigh_get() local
2952 neigh_for_each(struct neigh_table *tbl, void (*cb)(struct neighbour *, void *), void *cookie) neigh_for_each() argument
2975 __neigh_for_each_release(struct neigh_table *tbl, int (*cb)(struct neighbour *)) __neigh_for_each_release() argument
3014 struct neigh_table *tbl; neigh_xmit() local
3161 struct neigh_table *tbl = state->tbl; pneigh_get_first() local
3184 struct neigh_table *tbl = state->tbl; pneigh_get_next() local
3284 struct neigh_table *tbl = state->tbl; __releases() local
3295 struct neigh_table *tbl = PDE_DATA(file_inode(seq->file)); neigh_stat_seq_start() local
3312 struct neigh_table *tbl = PDE_DATA(file_inode(seq->file)); neigh_stat_seq_next() local
3332 struct neigh_table *tbl = PDE_DATA(file_inode(seq->file)); neigh_stat_seq_show() local
3668 struct neigh_table *tbl = p->tbl; neigh_sysctl_register() local
[all...]
/kernel/linux/linux-5.10/net/mac80211/
H A Dmesh_pathtbl.c18 static void mesh_path_free_rcu(struct mesh_table *tbl, struct mesh_path *mpath);
45 struct mesh_table *tbl = tblptr; in mesh_path_rht_free() local
47 mesh_path_free_rcu(tbl, mpath); in mesh_path_rht_free()
50 static void mesh_table_init(struct mesh_table *tbl) in mesh_table_init() argument
52 INIT_HLIST_HEAD(&tbl->known_gates); in mesh_table_init()
53 INIT_HLIST_HEAD(&tbl->walk_head); in mesh_table_init()
54 atomic_set(&tbl->entries, 0); in mesh_table_init()
55 spin_lock_init(&tbl->gates_lock); in mesh_table_init()
56 spin_lock_init(&tbl->walk_lock); in mesh_table_init()
61 WARN_ON(rhashtable_init(&tbl in mesh_table_init()
64 mesh_table_free(struct mesh_table *tbl) mesh_table_free() argument
206 mpath_lookup(struct mesh_table *tbl, const u8 *dst, struct ieee80211_sub_if_data *sdata) mpath_lookup() argument
243 __mesh_path_lookup_by_idx(struct mesh_table *tbl, int idx) __mesh_path_lookup_by_idx() argument
300 struct mesh_table *tbl; mesh_path_add_gate() local
335 mesh_gate_del(struct mesh_table *tbl, struct mesh_path *mpath) mesh_gate_del() argument
396 struct mesh_table *tbl; mesh_path_add() local
438 struct mesh_table *tbl; mpp_path_add() local
484 struct mesh_table *tbl = &sdata->u.mesh.mesh_paths; mesh_plink_broken() local
506 mesh_path_free_rcu(struct mesh_table *tbl, struct mesh_path *mpath) mesh_path_free_rcu() argument
522 __mesh_path_del(struct mesh_table *tbl, struct mesh_path *mpath) __mesh_path_del() argument
543 struct mesh_table *tbl = &sdata->u.mesh.mesh_paths; mesh_path_flush_by_nexthop() local
558 struct mesh_table *tbl = &sdata->u.mesh.mpp_paths; mpp_flush_by_proxy() local
570 table_flush_by_iface(struct mesh_table *tbl) table_flush_by_iface() argument
605 table_path_del(struct mesh_table *tbl, struct ieee80211_sub_if_data *sdata, const u8 *addr) table_path_del() argument
672 struct mesh_table *tbl; mesh_path_send_to_gates() local
777 mesh_path_tbl_expire(struct ieee80211_sub_if_data *sdata, struct mesh_table *tbl) mesh_path_tbl_expire() argument
[all...]
/kernel/linux/linux-5.10/net/netfilter/ipvs/
H A Dip_vs_lblc.c9 * *lock(tbl->lock) ==> *lock(&tbl->lock)
10 * Wensong Zhang : fixed the uninitialized tbl->lock bug
168 ip_vs_lblc_hash(struct ip_vs_lblc_table *tbl, struct ip_vs_lblc_entry *en) in ip_vs_lblc_hash() argument
172 hlist_add_head_rcu(&en->list, &tbl->bucket[hash]); in ip_vs_lblc_hash()
173 atomic_inc(&tbl->entries); in ip_vs_lblc_hash()
179 ip_vs_lblc_get(int af, struct ip_vs_lblc_table *tbl, in ip_vs_lblc_get() argument
185 hlist_for_each_entry_rcu(en, &tbl->bucket[hash], list) in ip_vs_lblc_get()
198 ip_vs_lblc_new(struct ip_vs_lblc_table *tbl, const union nf_inet_addr *daddr, in ip_vs_lblc_new() argument
203 en = ip_vs_lblc_get(af, tbl, dadd in ip_vs_lblc_new()
231 struct ip_vs_lblc_table *tbl = svc->sched_data; ip_vs_lblc_flush() local
258 struct ip_vs_lblc_table *tbl = svc->sched_data; ip_vs_lblc_full_check() local
296 struct ip_vs_lblc_table *tbl = from_timer(tbl, t, periodic_timer); ip_vs_lblc_check_expire() local
346 struct ip_vs_lblc_table *tbl; ip_vs_lblc_init_svc() local
384 struct ip_vs_lblc_table *tbl = svc->sched_data; ip_vs_lblc_done_svc() local
484 struct ip_vs_lblc_table *tbl = svc->sched_data; ip_vs_lblc_schedule() local
[all...]
/kernel/linux/linux-6.6/net/netfilter/ipvs/
H A Dip_vs_lblc.c9 * *lock(tbl->lock) ==> *lock(&tbl->lock)
10 * Wensong Zhang : fixed the uninitialized tbl->lock bug
168 ip_vs_lblc_hash(struct ip_vs_lblc_table *tbl, struct ip_vs_lblc_entry *en) in ip_vs_lblc_hash() argument
172 hlist_add_head_rcu(&en->list, &tbl->bucket[hash]); in ip_vs_lblc_hash()
173 atomic_inc(&tbl->entries); in ip_vs_lblc_hash()
179 ip_vs_lblc_get(int af, struct ip_vs_lblc_table *tbl, in ip_vs_lblc_get() argument
185 hlist_for_each_entry_rcu(en, &tbl->bucket[hash], list) in ip_vs_lblc_get()
198 ip_vs_lblc_new(struct ip_vs_lblc_table *tbl, const union nf_inet_addr *daddr, in ip_vs_lblc_new() argument
203 en = ip_vs_lblc_get(af, tbl, dadd in ip_vs_lblc_new()
231 struct ip_vs_lblc_table *tbl = svc->sched_data; ip_vs_lblc_flush() local
258 struct ip_vs_lblc_table *tbl = svc->sched_data; ip_vs_lblc_full_check() local
296 struct ip_vs_lblc_table *tbl = from_timer(tbl, t, periodic_timer); ip_vs_lblc_check_expire() local
346 struct ip_vs_lblc_table *tbl; ip_vs_lblc_init_svc() local
384 struct ip_vs_lblc_table *tbl = svc->sched_data; ip_vs_lblc_done_svc() local
484 struct ip_vs_lblc_table *tbl = svc->sched_data; ip_vs_lblc_schedule() local
[all...]
/kernel/linux/linux-5.10/arch/powerpc/include/asm/
H A Diommu.h40 int (*set)(struct iommu_table *tbl,
51 int (*xchg_no_kill)(struct iommu_table *tbl,
57 void (*tce_kill)(struct iommu_table *tbl,
62 __be64 *(*useraddrptr)(struct iommu_table *tbl, long index, bool alloc);
64 void (*clear)(struct iommu_table *tbl,
67 unsigned long (*get)(struct iommu_table *tbl, long index);
68 void (*flush)(struct iommu_table *tbl);
69 void (*free)(struct iommu_table *tbl);
119 #define IOMMU_TABLE_USERSPACE_ENTRY_RO(tbl, entry) \
120 ((tbl)
126 get_iommu_order(unsigned long size, struct iommu_table *tbl) get_iommu_order() argument
[all...]
/kernel/linux/linux-6.6/arch/powerpc/include/asm/
H A Diommu.h43 int (*set)(struct iommu_table *tbl,
54 int (*xchg_no_kill)(struct iommu_table *tbl,
59 void (*tce_kill)(struct iommu_table *tbl,
63 __be64 *(*useraddrptr)(struct iommu_table *tbl, long index, bool alloc);
65 void (*clear)(struct iommu_table *tbl,
68 unsigned long (*get)(struct iommu_table *tbl, long index);
69 void (*flush)(struct iommu_table *tbl);
70 void (*free)(struct iommu_table *tbl);
120 #define IOMMU_TABLE_USERSPACE_ENTRY_RO(tbl, entry) \
121 ((tbl)
127 get_iommu_order(unsigned long size, struct iommu_table *tbl) get_iommu_order() argument
[all...]
/kernel/linux/linux-5.10/drivers/net/wireless/marvell/mwifiex/
H A D11n_rxreorder.c110 struct mwifiex_rx_reorder_tbl *tbl, in mwifiex_11n_dispatch_pkt_until_start_win()
120 pkt_to_send = (start_win > tbl->start_win) ? in mwifiex_11n_dispatch_pkt_until_start_win()
121 min((start_win - tbl->start_win), tbl->win_size) : in mwifiex_11n_dispatch_pkt_until_start_win()
122 tbl->win_size; in mwifiex_11n_dispatch_pkt_until_start_win()
125 if (tbl->rx_reorder_ptr[i]) { in mwifiex_11n_dispatch_pkt_until_start_win()
126 skb = tbl->rx_reorder_ptr[i]; in mwifiex_11n_dispatch_pkt_until_start_win()
128 tbl->rx_reorder_ptr[i] = NULL; in mwifiex_11n_dispatch_pkt_until_start_win()
136 for (i = 0; i < tbl->win_size - pkt_to_send; ++i) { in mwifiex_11n_dispatch_pkt_until_start_win()
137 tbl in mwifiex_11n_dispatch_pkt_until_start_win()
109 mwifiex_11n_dispatch_pkt_until_start_win(struct mwifiex_private *priv, struct mwifiex_rx_reorder_tbl *tbl, int start_win) mwifiex_11n_dispatch_pkt_until_start_win() argument
157 mwifiex_11n_scan_and_dispatch(struct mwifiex_private *priv, struct mwifiex_rx_reorder_tbl *tbl) mwifiex_11n_scan_and_dispatch() argument
201 mwifiex_del_rx_reorder_entry(struct mwifiex_private *priv, struct mwifiex_rx_reorder_tbl *tbl) mwifiex_del_rx_reorder_entry() argument
244 struct mwifiex_rx_reorder_tbl *tbl; mwifiex_11n_get_rx_reorder_tbl() local
263 struct mwifiex_rx_reorder_tbl *tbl, *tmp; mwifiex_11n_del_rx_reorder_tbl_by_ta() local
345 struct mwifiex_rx_reorder_tbl *tbl, *new_node; mwifiex_11n_create_rx_reorder_tbl() local
422 mwifiex_11n_rxreorder_timer_restart(struct mwifiex_rx_reorder_tbl *tbl) mwifiex_11n_rxreorder_timer_restart() argument
563 struct mwifiex_rx_reorder_tbl *tbl; mwifiex_11n_rx_reorder_pkt() local
675 struct mwifiex_rx_reorder_tbl *tbl; mwifiex_del_ba_tbl() local
729 struct mwifiex_rx_reorder_tbl *tbl; mwifiex_ret_11n_addba_resp() local
820 struct mwifiex_rx_reorder_tbl *tbl; mwifiex_update_rxreor_flags() local
[all...]

Completed in 24 milliseconds

12345678910>>...23