/kernel/linux/linux-5.10/fs/nfs/ |
H A D | nfs4session.c | 27 static void nfs4_init_slot_table(struct nfs4_slot_table *tbl, const char *queue) in nfs4_init_slot_table() argument 29 tbl->highest_used_slotid = NFS4_NO_SLOT; in nfs4_init_slot_table() 30 spin_lock_init(&tbl->slot_tbl_lock); in nfs4_init_slot_table() 31 rpc_init_priority_wait_queue(&tbl->slot_tbl_waitq, queue); in nfs4_init_slot_table() 32 init_waitqueue_head(&tbl->slot_waitq); in nfs4_init_slot_table() 33 init_completion(&tbl->complete); in nfs4_init_slot_table() 39 static void nfs4_shrink_slot_table(struct nfs4_slot_table *tbl, u32 newsize) in nfs4_shrink_slot_table() argument 42 if (newsize >= tbl->max_slots) in nfs4_shrink_slot_table() 45 p = &tbl->slots; in nfs4_shrink_slot_table() 53 tbl in nfs4_shrink_slot_table() 62 nfs4_slot_tbl_drain_complete(struct nfs4_slot_table *tbl) nfs4_slot_tbl_drain_complete() argument 83 nfs4_free_slot(struct nfs4_slot_table *tbl, struct nfs4_slot *slot) nfs4_free_slot() argument 104 nfs4_new_slot(struct nfs4_slot_table *tbl, u32 slotid, u32 seq_init, gfp_t gfp_mask) nfs4_new_slot() argument 120 nfs4_find_or_create_slot(struct nfs4_slot_table *tbl, u32 slotid, u32 seq_init, gfp_t gfp_mask) nfs4_find_or_create_slot() argument 142 nfs4_lock_slot(struct nfs4_slot_table *tbl, struct nfs4_slot *slot) nfs4_lock_slot() argument 159 nfs4_try_to_lock_slot(struct nfs4_slot_table *tbl, struct nfs4_slot *slot) nfs4_try_to_lock_slot() argument 172 nfs4_lookup_slot(struct nfs4_slot_table *tbl, u32 slotid) nfs4_lookup_slot() argument 201 nfs4_slot_seqid_in_use(struct nfs4_slot_table *tbl, u32 slotid, u32 seq_nr) nfs4_slot_seqid_in_use() argument 222 nfs4_slot_wait_on_seqid(struct nfs4_slot_table *tbl, u32 slotid, u32 seq_nr, unsigned long timeout) nfs4_slot_wait_on_seqid() argument 242 nfs4_alloc_slot(struct nfs4_slot_table *tbl) nfs4_alloc_slot() argument 262 nfs4_grow_slot_table(struct nfs4_slot_table *tbl, u32 max_reqs, u32 ivalue) nfs4_grow_slot_table() argument 272 nfs4_reset_slot_table(struct nfs4_slot_table *tbl, u32 server_highest_slotid, u32 ivalue) nfs4_reset_slot_table() argument 297 nfs4_realloc_slot_table(struct nfs4_slot_table *tbl, u32 max_reqs, u32 ivalue) nfs4_realloc_slot_table() argument 326 nfs4_release_slot_table(struct nfs4_slot_table *tbl) nfs4_release_slot_table() argument 336 nfs4_shutdown_slot_table(struct nfs4_slot_table *tbl) nfs4_shutdown_slot_table() argument 350 nfs4_setup_slot_table(struct nfs4_slot_table *tbl, unsigned int max_reqs, const char *queue) nfs4_setup_slot_table() argument 362 struct nfs4_slot_table *tbl = slot->table; nfs41_assign_slot() local 375 __nfs41_wake_and_assign_slot(struct nfs4_slot_table *tbl, struct nfs4_slot *slot) __nfs41_wake_and_assign_slot() argument 383 nfs41_wake_and_assign_slot(struct nfs4_slot_table *tbl, struct nfs4_slot *slot) nfs41_wake_and_assign_slot() argument 391 nfs41_try_wake_next_slot_table_entry(struct nfs4_slot_table *tbl) nfs41_try_wake_next_slot_table_entry() argument 403 nfs41_wake_slot_table(struct nfs4_slot_table *tbl) nfs41_wake_slot_table() argument 413 nfs41_set_max_slotid_locked(struct nfs4_slot_table *tbl, u32 target_highest_slotid) nfs41_set_max_slotid_locked() argument 428 nfs41_set_target_slotid_locked(struct nfs4_slot_table *tbl, u32 target_highest_slotid) nfs41_set_target_slotid_locked() argument 437 nfs41_set_target_slotid(struct nfs4_slot_table *tbl, u32 target_highest_slotid) nfs41_set_target_slotid() argument 448 nfs41_set_server_slotid_locked(struct nfs4_slot_table *tbl, u32 highest_slotid) nfs41_set_server_slotid_locked() argument 489 nfs41_is_outlier_target_slotid(struct nfs4_slot_table *tbl, u32 new_target) nfs41_is_outlier_target_slotid() argument 510 nfs41_update_target_slotid(struct nfs4_slot_table *tbl, struct nfs4_slot *slot, struct nfs4_sequence_res *res) nfs41_update_target_slotid() argument 534 struct nfs4_slot_table *tbl; nfs4_setup_session_slot_tables() local [all...] |
/kernel/linux/linux-6.6/fs/nfs/ |
H A D | nfs4session.c | 27 static void nfs4_init_slot_table(struct nfs4_slot_table *tbl, const char *queue) in nfs4_init_slot_table() argument 29 tbl->highest_used_slotid = NFS4_NO_SLOT; in nfs4_init_slot_table() 30 spin_lock_init(&tbl->slot_tbl_lock); in nfs4_init_slot_table() 31 rpc_init_priority_wait_queue(&tbl->slot_tbl_waitq, queue); in nfs4_init_slot_table() 32 init_waitqueue_head(&tbl->slot_waitq); in nfs4_init_slot_table() 33 init_completion(&tbl->complete); in nfs4_init_slot_table() 39 static void nfs4_shrink_slot_table(struct nfs4_slot_table *tbl, u32 newsize) in nfs4_shrink_slot_table() argument 42 if (newsize >= tbl->max_slots) in nfs4_shrink_slot_table() 45 p = &tbl->slots; in nfs4_shrink_slot_table() 53 tbl in nfs4_shrink_slot_table() 62 nfs4_slot_tbl_drain_complete(struct nfs4_slot_table *tbl) nfs4_slot_tbl_drain_complete() argument 83 nfs4_free_slot(struct nfs4_slot_table *tbl, struct nfs4_slot *slot) nfs4_free_slot() argument 104 nfs4_new_slot(struct nfs4_slot_table *tbl, u32 slotid, u32 seq_init, gfp_t gfp_mask) nfs4_new_slot() argument 120 nfs4_find_or_create_slot(struct nfs4_slot_table *tbl, u32 slotid, u32 seq_init, gfp_t gfp_mask) nfs4_find_or_create_slot() argument 142 nfs4_lock_slot(struct nfs4_slot_table *tbl, struct nfs4_slot *slot) nfs4_lock_slot() argument 159 nfs4_try_to_lock_slot(struct nfs4_slot_table *tbl, struct nfs4_slot *slot) nfs4_try_to_lock_slot() argument 172 nfs4_lookup_slot(struct nfs4_slot_table *tbl, u32 slotid) nfs4_lookup_slot() argument 201 nfs4_slot_seqid_in_use(struct nfs4_slot_table *tbl, u32 slotid, u32 seq_nr) nfs4_slot_seqid_in_use() argument 222 nfs4_slot_wait_on_seqid(struct nfs4_slot_table *tbl, u32 slotid, u32 seq_nr, unsigned long timeout) nfs4_slot_wait_on_seqid() argument 242 nfs4_alloc_slot(struct nfs4_slot_table *tbl) nfs4_alloc_slot() argument 262 nfs4_grow_slot_table(struct nfs4_slot_table *tbl, u32 max_reqs, u32 ivalue) nfs4_grow_slot_table() argument 272 nfs4_reset_slot_table(struct nfs4_slot_table *tbl, u32 server_highest_slotid, u32 ivalue) nfs4_reset_slot_table() argument 297 nfs4_realloc_slot_table(struct nfs4_slot_table *tbl, u32 max_reqs, u32 ivalue) nfs4_realloc_slot_table() argument 326 nfs4_release_slot_table(struct nfs4_slot_table *tbl) nfs4_release_slot_table() argument 336 nfs4_shutdown_slot_table(struct nfs4_slot_table *tbl) nfs4_shutdown_slot_table() argument 350 nfs4_setup_slot_table(struct nfs4_slot_table *tbl, unsigned int max_reqs, const char *queue) nfs4_setup_slot_table() argument 362 struct nfs4_slot_table *tbl = slot->table; nfs41_assign_slot() local 375 __nfs41_wake_and_assign_slot(struct nfs4_slot_table *tbl, struct nfs4_slot *slot) __nfs41_wake_and_assign_slot() argument 383 nfs41_wake_and_assign_slot(struct nfs4_slot_table *tbl, struct nfs4_slot *slot) nfs41_wake_and_assign_slot() argument 391 nfs41_try_wake_next_slot_table_entry(struct nfs4_slot_table *tbl) nfs41_try_wake_next_slot_table_entry() argument 403 nfs41_wake_slot_table(struct nfs4_slot_table *tbl) nfs41_wake_slot_table() argument 413 nfs41_set_max_slotid_locked(struct nfs4_slot_table *tbl, u32 target_highest_slotid) nfs41_set_max_slotid_locked() argument 428 nfs41_set_target_slotid_locked(struct nfs4_slot_table *tbl, u32 target_highest_slotid) nfs41_set_target_slotid_locked() argument 437 nfs41_set_target_slotid(struct nfs4_slot_table *tbl, u32 target_highest_slotid) nfs41_set_target_slotid() argument 448 nfs41_set_server_slotid_locked(struct nfs4_slot_table *tbl, u32 highest_slotid) nfs41_set_server_slotid_locked() argument 489 nfs41_is_outlier_target_slotid(struct nfs4_slot_table *tbl, u32 new_target) nfs41_is_outlier_target_slotid() argument 510 nfs41_update_target_slotid(struct nfs4_slot_table *tbl, struct nfs4_slot *slot, struct nfs4_sequence_res *res) nfs41_update_target_slotid() argument 538 struct nfs4_slot_table *tbl; nfs4_setup_session_slot_tables() local [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/steering/ |
H A D | dr_table.c | 6 int mlx5dr_table_set_miss_action(struct mlx5dr_table *tbl, in mlx5dr_table_set_miss_action() argument 17 mlx5dr_domain_lock(tbl->dmn); in mlx5dr_table_set_miss_action() 19 if (!list_empty(&tbl->matcher_list)) in mlx5dr_table_set_miss_action() 20 last_matcher = list_last_entry(&tbl->matcher_list, in mlx5dr_table_set_miss_action() 24 if (tbl->dmn->type == MLX5DR_DOMAIN_TYPE_NIC_RX || in mlx5dr_table_set_miss_action() 25 tbl->dmn->type == MLX5DR_DOMAIN_TYPE_FDB) { in mlx5dr_table_set_miss_action() 29 last_htbl = tbl->rx.s_anchor; in mlx5dr_table_set_miss_action() 31 tbl->rx.default_icm_addr = action ? in mlx5dr_table_set_miss_action() 32 action->dest_tbl.tbl->rx.s_anchor->chunk->icm_addr : in mlx5dr_table_set_miss_action() 33 tbl in mlx5dr_table_set_miss_action() 93 dr_table_uninit_fdb(struct mlx5dr_table *tbl) dr_table_uninit_fdb() argument 99 dr_table_uninit(struct mlx5dr_table *tbl) dr_table_uninit() argument 158 dr_table_init_fdb(struct mlx5dr_table *tbl) dr_table_init_fdb() argument 177 dr_table_init(struct mlx5dr_table *tbl) dr_table_init() argument 212 dr_table_destroy_sw_owned_tbl(struct mlx5dr_table *tbl) dr_table_destroy_sw_owned_tbl() argument 219 dr_table_create_sw_owned_tbl(struct mlx5dr_table *tbl) dr_table_create_sw_owned_tbl() argument 250 struct mlx5dr_table *tbl; mlx5dr_table_create() local 283 mlx5dr_table_destroy(struct mlx5dr_table *tbl) mlx5dr_table_destroy() argument 305 mlx5dr_table_get_id(struct mlx5dr_table *tbl) mlx5dr_table_get_id() argument [all...] |
/kernel/linux/linux-5.10/arch/powerpc/kernel/ |
H A D | iommu.c | 175 struct iommu_table *tbl, in iommu_range_alloc() 208 pool_nr = raw_cpu_read(iommu_pool_hash) & (tbl->nr_pools - 1); in iommu_range_alloc() 211 pool = &(tbl->large_pool); in iommu_range_alloc() 213 pool = &(tbl->pools[pool_nr]); in iommu_range_alloc() 233 if (limit + tbl->it_offset > mask) { in iommu_range_alloc() 234 limit = mask - tbl->it_offset + 1; in iommu_range_alloc() 241 pool = &(tbl->pools[0]); in iommu_range_alloc() 249 n = iommu_area_alloc(tbl->it_map, limit, start, npages, tbl->it_offset, in iommu_range_alloc() 250 dma_get_seg_boundary_nr_pages(dev, tbl in iommu_range_alloc() 174 iommu_range_alloc(struct device *dev, struct iommu_table *tbl, unsigned long npages, unsigned long *handle, unsigned long mask, unsigned int align_order) iommu_range_alloc() argument 297 iommu_alloc(struct device *dev, struct iommu_table *tbl, void *page, unsigned int npages, enum dma_data_direction direction, unsigned long mask, unsigned int align_order, unsigned long attrs) iommu_alloc() argument 340 iommu_free_check(struct iommu_table *tbl, dma_addr_t dma_addr, unsigned int npages) iommu_free_check() argument 368 get_pool(struct iommu_table *tbl, unsigned long entry) get_pool() argument 387 __iommu_free(struct iommu_table *tbl, dma_addr_t dma_addr, unsigned int npages) __iommu_free() argument 409 iommu_free(struct iommu_table *tbl, dma_addr_t dma_addr, unsigned int npages) iommu_free() argument 422 ppc_iommu_map_sg(struct device *dev, struct iommu_table *tbl, struct scatterlist *sglist, int nelems, unsigned long mask, enum dma_data_direction direction, unsigned long attrs) ppc_iommu_map_sg() argument 567 ppc_iommu_unmap_sg(struct iommu_table *tbl, struct scatterlist *sglist, int nelems, enum dma_data_direction direction, unsigned long attrs) ppc_iommu_unmap_sg() argument 599 iommu_table_clear(struct iommu_table *tbl) iommu_table_clear() argument 640 iommu_table_reserve_pages(struct iommu_table *tbl, unsigned long res_start, unsigned long res_end) iommu_table_reserve_pages() argument 667 iommu_table_release_pages(struct iommu_table *tbl) iommu_table_release_pages() argument 686 iommu_init_table(struct iommu_table *tbl, int nid, unsigned long res_start, unsigned long res_end) iommu_init_table() argument 746 struct iommu_table *tbl; iommu_table_free() local 775 iommu_tce_table_get(struct iommu_table *tbl) iommu_tce_table_get() argument 784 iommu_tce_table_put(struct iommu_table *tbl) iommu_tce_table_put() argument 798 iommu_map_page(struct device *dev, struct iommu_table *tbl, struct page *page, unsigned long offset, size_t size, unsigned long mask, enum dma_data_direction direction, unsigned long attrs) iommu_map_page() argument 837 iommu_unmap_page(struct iommu_table *tbl, dma_addr_t dma_handle, size_t size, enum dma_data_direction direction, unsigned long attrs) iommu_unmap_page() argument 856 iommu_alloc_coherent(struct device *dev, struct iommu_table *tbl, size_t size, dma_addr_t *dma_handle, unsigned long mask, gfp_t flag, int node) iommu_alloc_coherent() argument 903 iommu_free_coherent(struct iommu_table *tbl, size_t size, void *vaddr, dma_addr_t dma_handle) iommu_free_coherent() argument 978 iommu_flush_tce(struct iommu_table *tbl) iommu_flush_tce() argument 1020 iommu_tce_xchg_no_kill(struct mm_struct *mm, struct iommu_table *tbl, unsigned long entry, unsigned long *hpa, enum dma_data_direction *direction) iommu_tce_xchg_no_kill() argument 1039 iommu_tce_kill(struct iommu_table *tbl, unsigned long entry, unsigned long pages) iommu_tce_kill() argument 1047 iommu_take_ownership(struct iommu_table *tbl) iommu_take_ownership() argument 1086 iommu_release_ownership(struct iommu_table *tbl) iommu_release_ownership() argument [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/steering/ |
H A D | dr_table.c | 28 action->dest_tbl->tbl->rx.s_anchor->chunk : in dr_table_set_miss_action_nic() 29 action->dest_tbl->tbl->tx.s_anchor->chunk; in dr_table_set_miss_action_nic() 46 int mlx5dr_table_set_miss_action(struct mlx5dr_table *tbl, in mlx5dr_table_set_miss_action() argument 54 mlx5dr_domain_lock(tbl->dmn); in mlx5dr_table_set_miss_action() 56 if (tbl->dmn->type == MLX5DR_DOMAIN_TYPE_NIC_RX || in mlx5dr_table_set_miss_action() 57 tbl->dmn->type == MLX5DR_DOMAIN_TYPE_FDB) { in mlx5dr_table_set_miss_action() 58 ret = dr_table_set_miss_action_nic(tbl->dmn, &tbl->rx, action); in mlx5dr_table_set_miss_action() 63 if (tbl->dmn->type == MLX5DR_DOMAIN_TYPE_NIC_TX || in mlx5dr_table_set_miss_action() 64 tbl in mlx5dr_table_set_miss_action() 92 dr_table_uninit_fdb(struct mlx5dr_table *tbl) dr_table_uninit_fdb() argument 98 dr_table_uninit(struct mlx5dr_table *tbl) dr_table_uninit() argument 159 dr_table_init_fdb(struct mlx5dr_table *tbl) dr_table_init_fdb() argument 178 dr_table_init(struct mlx5dr_table *tbl) dr_table_init() argument 213 dr_table_destroy_sw_owned_tbl(struct mlx5dr_table *tbl) dr_table_destroy_sw_owned_tbl() argument 220 dr_table_create_sw_owned_tbl(struct mlx5dr_table *tbl, u16 uid) dr_table_create_sw_owned_tbl() argument 253 struct mlx5dr_table *tbl; mlx5dr_table_create() local 288 mlx5dr_table_destroy(struct mlx5dr_table *tbl) mlx5dr_table_destroy() argument 311 mlx5dr_table_get_id(struct mlx5dr_table *tbl) mlx5dr_table_get_id() argument [all...] |
/kernel/linux/linux-6.6/arch/powerpc/kernel/ |
H A D | iommu.c | 45 struct iommu_table *tbl = data; in iommu_debugfs_weight_get() local 46 *val = bitmap_weight(tbl->it_map, tbl->it_size); in iommu_debugfs_weight_get() 51 static void iommu_debugfs_add(struct iommu_table *tbl) in iommu_debugfs_add() argument 56 sprintf(name, "%08lx", tbl->it_index); in iommu_debugfs_add() 59 debugfs_create_file_unsafe("weight", 0400, liobn_entry, tbl, &iommu_debugfs_fops_weight); in iommu_debugfs_add() 60 debugfs_create_ulong("it_size", 0400, liobn_entry, &tbl->it_size); in iommu_debugfs_add() 61 debugfs_create_ulong("it_page_shift", 0400, liobn_entry, &tbl->it_page_shift); in iommu_debugfs_add() 62 debugfs_create_ulong("it_reserved_start", 0400, liobn_entry, &tbl->it_reserved_start); in iommu_debugfs_add() 63 debugfs_create_ulong("it_reserved_end", 0400, liobn_entry, &tbl in iommu_debugfs_add() 68 iommu_debugfs_del(struct iommu_table *tbl) iommu_debugfs_del() argument 76 iommu_debugfs_add(struct iommu_table *tbl) iommu_debugfs_add() argument 77 iommu_debugfs_del(struct iommu_table *tbl) iommu_debugfs_del() argument 213 iommu_range_alloc(struct device *dev, struct iommu_table *tbl, unsigned long npages, unsigned long *handle, unsigned long mask, unsigned int align_order) iommu_range_alloc() argument 345 iommu_alloc(struct device *dev, struct iommu_table *tbl, void *page, unsigned int npages, enum dma_data_direction direction, unsigned long mask, unsigned int align_order, unsigned long attrs) iommu_alloc() argument 388 iommu_free_check(struct iommu_table *tbl, dma_addr_t dma_addr, unsigned int npages) iommu_free_check() argument 416 get_pool(struct iommu_table *tbl, unsigned long entry) get_pool() argument 435 __iommu_free(struct iommu_table *tbl, dma_addr_t dma_addr, unsigned int npages) __iommu_free() argument 457 iommu_free(struct iommu_table *tbl, dma_addr_t dma_addr, unsigned int npages) iommu_free() argument 470 ppc_iommu_map_sg(struct device *dev, struct iommu_table *tbl, struct scatterlist *sglist, int nelems, unsigned long mask, enum dma_data_direction direction, unsigned long attrs) ppc_iommu_map_sg() argument 613 ppc_iommu_unmap_sg(struct iommu_table *tbl, struct scatterlist *sglist, int nelems, enum dma_data_direction direction, unsigned long attrs) ppc_iommu_unmap_sg() argument 645 iommu_table_clear(struct iommu_table *tbl) iommu_table_clear() argument 686 iommu_table_reserve_pages(struct iommu_table *tbl, unsigned long res_start, unsigned long res_end) iommu_table_reserve_pages() argument 724 iommu_init_table(struct iommu_table *tbl, int nid, unsigned long res_start, unsigned long res_end) iommu_init_table() argument 781 iommu_table_in_use(struct iommu_table *tbl) iommu_table_in_use() argument 804 struct iommu_table *tbl; iommu_table_free() local 829 iommu_tce_table_get(struct iommu_table *tbl) iommu_tce_table_get() argument 838 iommu_tce_table_put(struct iommu_table *tbl) iommu_tce_table_put() argument 852 iommu_map_page(struct device *dev, struct iommu_table *tbl, struct page *page, unsigned long offset, size_t size, unsigned long mask, enum dma_data_direction direction, unsigned long attrs) iommu_map_page() argument 891 iommu_unmap_page(struct iommu_table *tbl, dma_addr_t dma_handle, size_t size, enum dma_data_direction direction, unsigned long attrs) iommu_unmap_page() argument 910 iommu_alloc_coherent(struct device *dev, struct iommu_table *tbl, size_t size, dma_addr_t *dma_handle, unsigned long mask, gfp_t flag, int node) iommu_alloc_coherent() argument 960 iommu_free_coherent(struct iommu_table *tbl, size_t size, void *vaddr, dma_addr_t dma_handle) iommu_free_coherent() argument 1035 iommu_flush_tce(struct iommu_table *tbl) iommu_flush_tce() argument 1077 iommu_tce_xchg_no_kill(struct mm_struct *mm, struct iommu_table *tbl, unsigned long entry, unsigned long *hpa, enum dma_data_direction *direction) iommu_tce_xchg_no_kill() argument 1096 iommu_tce_kill(struct iommu_table *tbl, unsigned long entry, unsigned long pages) iommu_tce_kill() argument 1105 iommu_take_ownership(struct iommu_table *tbl) iommu_take_ownership() argument 1138 iommu_release_ownership(struct iommu_table *tbl) iommu_release_ownership() argument 1209 struct iommu_table *tbl = table_group->tables[0]; spapr_tce_create_table() local 1223 spapr_tce_set_window(struct iommu_table_group *table_group, int num, struct iommu_table *tbl) spapr_tce_set_window() argument 1239 struct iommu_table *tbl = table_group->tables[i]; spapr_tce_take_ownership() local 1260 struct iommu_table *tbl = table_group->tables[i]; spapr_tce_release_ownership() local [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/hw/vmw_pvrdma/ |
H A D | pvrdma_doorbell.c | 56 struct pvrdma_id_table *tbl = &dev->uar_table.tbl; in pvrdma_uar_table_init() local 61 tbl->last = 0; in pvrdma_uar_table_init() 62 tbl->top = 0; in pvrdma_uar_table_init() 63 tbl->max = num; in pvrdma_uar_table_init() 64 tbl->mask = mask; in pvrdma_uar_table_init() 65 spin_lock_init(&tbl->lock); in pvrdma_uar_table_init() 66 tbl->table = kcalloc(BITS_TO_LONGS(num), sizeof(long), GFP_KERNEL); in pvrdma_uar_table_init() 67 if (!tbl->table) in pvrdma_uar_table_init() 71 set_bit(0, tbl in pvrdma_uar_table_init() 78 struct pvrdma_id_table *tbl = &dev->uar_table.tbl; pvrdma_uar_table_cleanup() local 85 struct pvrdma_id_table *tbl; pvrdma_uar_alloc() local 117 struct pvrdma_id_table *tbl = &dev->uar_table.tbl; pvrdma_uar_free() local [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/hw/vmw_pvrdma/ |
H A D | pvrdma_doorbell.c | 56 struct pvrdma_id_table *tbl = &dev->uar_table.tbl; in pvrdma_uar_table_init() local 61 tbl->last = 0; in pvrdma_uar_table_init() 62 tbl->top = 0; in pvrdma_uar_table_init() 63 tbl->max = num; in pvrdma_uar_table_init() 64 tbl->mask = mask; in pvrdma_uar_table_init() 65 spin_lock_init(&tbl->lock); in pvrdma_uar_table_init() 66 tbl->table = bitmap_zalloc(num, GFP_KERNEL); in pvrdma_uar_table_init() 67 if (!tbl->table) in pvrdma_uar_table_init() 71 __set_bit(0, tbl in pvrdma_uar_table_init() 78 struct pvrdma_id_table *tbl = &dev->uar_table.tbl; pvrdma_uar_table_cleanup() local 85 struct pvrdma_id_table *tbl; pvrdma_uar_alloc() local 117 struct pvrdma_id_table *tbl = &dev->uar_table.tbl; pvrdma_uar_free() local [all...] |
/kernel/linux/linux-5.10/include/linux/ |
H A D | rhashtable.h | 119 static inline unsigned int rht_bucket_index(const struct bucket_table *tbl, in rht_bucket_index() argument 122 return hash & (tbl->size - 1); in rht_bucket_index() 156 struct rhashtable *ht, const struct bucket_table *tbl, in rht_key_hashfn() 159 unsigned int hash = rht_key_get_hash(ht, key, params, tbl->hash_rnd); in rht_key_hashfn() 161 return rht_bucket_index(tbl, hash); in rht_key_hashfn() 165 struct rhashtable *ht, const struct bucket_table *tbl, in rht_head_hashfn() 171 rht_bucket_index(tbl, params.obj_hashfn(ptr, params.key_len ?: in rht_head_hashfn() 173 tbl->hash_rnd)) : in rht_head_hashfn() 174 rht_key_hashfn(ht, tbl, ptr + params.key_offset, params); in rht_head_hashfn() 180 * @tbl 155 rht_key_hashfn( struct rhashtable *ht, const struct bucket_table *tbl, const void *key, const struct rhashtable_params params) rht_key_hashfn() argument 164 rht_head_hashfn( struct rhashtable *ht, const struct bucket_table *tbl, const struct rhash_head *he, const struct rhashtable_params params) rht_head_hashfn() argument 182 rht_grow_above_75(const struct rhashtable *ht, const struct bucket_table *tbl) rht_grow_above_75() argument 195 rht_shrink_below_30(const struct rhashtable *ht, const struct bucket_table *tbl) rht_shrink_below_30() argument 208 rht_grow_above_100(const struct rhashtable *ht, const struct bucket_table *tbl) rht_grow_above_100() argument 220 rht_grow_above_max(const struct rhashtable *ht, const struct bucket_table *tbl) rht_grow_above_max() argument 235 lockdep_rht_bucket_is_held(const struct bucket_table *tbl, u32 hash) lockdep_rht_bucket_is_held() argument 286 rht_bucket( const struct bucket_table *tbl, unsigned int hash) rht_bucket() argument 293 rht_bucket_var( struct bucket_table *tbl, unsigned int hash) rht_bucket_var() argument 300 rht_bucket_insert( struct rhashtable *ht, struct bucket_table *tbl, unsigned int hash) rht_bucket_insert() argument 326 rht_lock(struct bucket_table *tbl, struct rhash_lock_head __rcu **bkt) rht_lock() argument 334 rht_lock_nested(struct bucket_table *tbl, struct rhash_lock_head __rcu **bucket, unsigned int subclass) rht_lock_nested() argument 343 rht_unlock(struct bucket_table *tbl, struct rhash_lock_head __rcu **bkt) rht_unlock() argument 372 rht_ptr( struct rhash_lock_head __rcu *const *bkt, struct bucket_table *tbl, unsigned int hash) rht_ptr() argument 394 rht_assign_unlock(struct bucket_table *tbl, struct rhash_lock_head __rcu **bkt, struct rhash_head *obj) rht_assign_unlock() argument 590 struct bucket_table *tbl; __rhashtable_lookup() local 707 struct bucket_table *tbl; __rhashtable_insert_fast() local 986 __rhashtable_remove_fast_one( struct rhashtable *ht, struct bucket_table *tbl, struct rhash_head *obj, const struct rhashtable_params params, bool rhlist) __rhashtable_remove_fast_one() argument 1071 struct bucket_table *tbl; __rhashtable_remove_fast() local 1138 __rhashtable_replace_fast( struct rhashtable *ht, struct bucket_table *tbl, struct rhash_head *obj_old, struct rhash_head *obj_new, const struct rhashtable_params params) __rhashtable_replace_fast() argument 1205 struct bucket_table *tbl; rhashtable_replace_fast() local [all...] |
/kernel/linux/linux-6.6/include/linux/ |
H A D | rhashtable.h | 119 static inline unsigned int rht_bucket_index(const struct bucket_table *tbl, in rht_bucket_index() argument 122 return hash & (tbl->size - 1); in rht_bucket_index() 156 struct rhashtable *ht, const struct bucket_table *tbl, in rht_key_hashfn() 159 unsigned int hash = rht_key_get_hash(ht, key, params, tbl->hash_rnd); in rht_key_hashfn() 161 return rht_bucket_index(tbl, hash); in rht_key_hashfn() 165 struct rhashtable *ht, const struct bucket_table *tbl, in rht_head_hashfn() 171 rht_bucket_index(tbl, params.obj_hashfn(ptr, params.key_len ?: in rht_head_hashfn() 173 tbl->hash_rnd)) : in rht_head_hashfn() 174 rht_key_hashfn(ht, tbl, ptr + params.key_offset, params); in rht_head_hashfn() 180 * @tbl 155 rht_key_hashfn( struct rhashtable *ht, const struct bucket_table *tbl, const void *key, const struct rhashtable_params params) rht_key_hashfn() argument 164 rht_head_hashfn( struct rhashtable *ht, const struct bucket_table *tbl, const struct rhash_head *he, const struct rhashtable_params params) rht_head_hashfn() argument 182 rht_grow_above_75(const struct rhashtable *ht, const struct bucket_table *tbl) rht_grow_above_75() argument 195 rht_shrink_below_30(const struct rhashtable *ht, const struct bucket_table *tbl) rht_shrink_below_30() argument 208 rht_grow_above_100(const struct rhashtable *ht, const struct bucket_table *tbl) rht_grow_above_100() argument 220 rht_grow_above_max(const struct rhashtable *ht, const struct bucket_table *tbl) rht_grow_above_max() argument 235 lockdep_rht_bucket_is_held(const struct bucket_table *tbl, u32 hash) lockdep_rht_bucket_is_held() argument 286 rht_bucket( const struct bucket_table *tbl, unsigned int hash) rht_bucket() argument 293 rht_bucket_var( struct bucket_table *tbl, unsigned int hash) rht_bucket_var() argument 300 rht_bucket_insert( struct rhashtable *ht, struct bucket_table *tbl, unsigned int hash) rht_bucket_insert() argument 326 rht_lock(struct bucket_table *tbl, struct rhash_lock_head __rcu **bkt) rht_lock() argument 337 rht_lock_nested(struct bucket_table *tbl, struct rhash_lock_head __rcu **bucket, unsigned int subclass) rht_lock_nested() argument 349 rht_unlock(struct bucket_table *tbl, struct rhash_lock_head __rcu **bkt, unsigned long flags) rht_unlock() argument 379 rht_ptr( struct rhash_lock_head __rcu *const *bkt, struct bucket_table *tbl, unsigned int hash) rht_ptr() argument 401 rht_assign_unlock(struct bucket_table *tbl, struct rhash_lock_head __rcu **bkt, struct rhash_head *obj, unsigned long flags) rht_assign_unlock() argument 598 struct bucket_table *tbl; __rhashtable_lookup() local 715 struct bucket_table *tbl; __rhashtable_insert_fast() local 995 __rhashtable_remove_fast_one( struct rhashtable *ht, struct bucket_table *tbl, struct rhash_head *obj, const struct rhashtable_params params, bool rhlist) __rhashtable_remove_fast_one() argument 1081 struct bucket_table *tbl; __rhashtable_remove_fast() local 1148 __rhashtable_replace_fast( struct rhashtable *ht, struct bucket_table *tbl, struct rhash_head *obj_old, struct rhash_head *obj_new, const struct rhashtable_params params) __rhashtable_replace_fast() argument 1216 struct bucket_table *tbl; rhashtable_replace_fast() local [all...] |
/kernel/linux/linux-5.10/tools/perf/util/ |
H A D | syscalltbl.c | 63 static int syscalltbl__init_native(struct syscalltbl *tbl) in syscalltbl__init_native() argument 72 entries = tbl->syscalls.entries = malloc(sizeof(struct syscall) * nr_entries); in syscalltbl__init_native() 73 if (tbl->syscalls.entries == NULL) in syscalltbl__init_native() 84 qsort(tbl->syscalls.entries, nr_entries, sizeof(struct syscall), syscallcmp); in syscalltbl__init_native() 85 tbl->syscalls.nr_entries = nr_entries; in syscalltbl__init_native() 86 tbl->syscalls.max_id = syscalltbl_native_max_id; in syscalltbl__init_native() 92 struct syscalltbl *tbl = malloc(sizeof(*tbl)); in syscalltbl__new() local 93 if (tbl) { in syscalltbl__new() 94 if (syscalltbl__init_native(tbl)) { in syscalltbl__new() 102 syscalltbl__delete(struct syscalltbl *tbl) syscalltbl__delete() argument 113 syscalltbl__id(struct syscalltbl *tbl, const char *name) syscalltbl__id() argument 122 syscalltbl__strglobmatch_next(struct syscalltbl *tbl, const char *syscall_glob, int *idx) syscalltbl__strglobmatch_next() argument 137 syscalltbl__strglobmatch_first(struct syscalltbl *tbl, const char *syscall_glob, int *idx) syscalltbl__strglobmatch_first() argument 149 struct syscalltbl *tbl = zalloc(sizeof(*tbl)); syscalltbl__new() local 155 syscalltbl__delete(struct syscalltbl *tbl) syscalltbl__delete() argument 160 syscalltbl__name(const struct syscalltbl *tbl, int id) syscalltbl__name() argument 165 syscalltbl__id(struct syscalltbl *tbl, const char *name) syscalltbl__id() argument 176 syscalltbl__strglobmatch_first(struct syscalltbl *tbl, const char *syscall_glob, int *idx) syscalltbl__strglobmatch_first() argument [all...] |
/kernel/linux/linux-6.6/tools/perf/util/ |
H A D | syscalltbl.c | 67 static int syscalltbl__init_native(struct syscalltbl *tbl) in syscalltbl__init_native() argument 76 entries = tbl->syscalls.entries = malloc(sizeof(struct syscall) * nr_entries); in syscalltbl__init_native() 77 if (tbl->syscalls.entries == NULL) in syscalltbl__init_native() 88 qsort(tbl->syscalls.entries, nr_entries, sizeof(struct syscall), syscallcmp); in syscalltbl__init_native() 89 tbl->syscalls.nr_entries = nr_entries; in syscalltbl__init_native() 90 tbl->syscalls.max_id = syscalltbl_native_max_id; in syscalltbl__init_native() 96 struct syscalltbl *tbl = malloc(sizeof(*tbl)); in syscalltbl__new() local 97 if (tbl) { in syscalltbl__new() 98 if (syscalltbl__init_native(tbl)) { in syscalltbl__new() 106 syscalltbl__delete(struct syscalltbl *tbl) syscalltbl__delete() argument 117 syscalltbl__id(struct syscalltbl *tbl, const char *name) syscalltbl__id() argument 126 syscalltbl__strglobmatch_next(struct syscalltbl *tbl, const char *syscall_glob, int *idx) syscalltbl__strglobmatch_next() argument 141 syscalltbl__strglobmatch_first(struct syscalltbl *tbl, const char *syscall_glob, int *idx) syscalltbl__strglobmatch_first() argument 153 struct syscalltbl *tbl = zalloc(sizeof(*tbl)); syscalltbl__new() local 159 syscalltbl__delete(struct syscalltbl *tbl) syscalltbl__delete() argument 164 syscalltbl__name(const struct syscalltbl *tbl, int id) syscalltbl__name() argument 169 syscalltbl__id(struct syscalltbl *tbl, const char *name) syscalltbl__id() argument 180 syscalltbl__strglobmatch_first(struct syscalltbl *tbl, const char *syscall_glob, int *idx) syscalltbl__strglobmatch_first() argument [all...] |
/kernel/linux/linux-5.10/drivers/net/wireless/intel/iwlwifi/dvm/ |
H A D | rs.c | 436 static s32 get_expected_tpt(struct iwl_scale_tbl_info *tbl, int rs_index) in get_expected_tpt() argument 438 if (tbl->expected_tpt) in get_expected_tpt() 439 return tbl->expected_tpt[rs_index]; in get_expected_tpt() 450 static int rs_collect_tx_data(struct iwl_scale_tbl_info *tbl, in rs_collect_tx_data() argument 461 window = &(tbl->win[scale_index]); in rs_collect_tx_data() 464 tpt = get_expected_tpt(tbl, scale_index); in rs_collect_tx_data() 529 struct iwl_scale_tbl_info *tbl, in rate_n_flags_from_tbl() 534 if (is_legacy(tbl->lq_type)) { in rate_n_flags_from_tbl() 539 } else if (is_Ht(tbl->lq_type)) { in rate_n_flags_from_tbl() 546 if (is_siso(tbl in rate_n_flags_from_tbl() 528 rate_n_flags_from_tbl(struct iwl_priv *priv, struct iwl_scale_tbl_info *tbl, int index, u8 use_green) rate_n_flags_from_tbl() argument 584 rs_get_tbl_info_from_mcs(const u32 rate_n_flags, enum nl80211_band band, struct iwl_scale_tbl_info *tbl, int *rate_idx) rs_get_tbl_info_from_mcs() argument 650 rs_toggle_antenna(u32 valid_ant, u32 *rate_n_flags, struct iwl_scale_tbl_info *tbl) rs_toggle_antenna() argument 774 rs_get_lower_rate(struct iwl_lq_sta *lq_sta, struct iwl_scale_tbl_info *tbl, u8 scale_index, u8 ht_possible) rs_get_lower_rate() argument 847 struct iwl_scale_tbl_info *tbl; rs_bt_update_lq() local 1086 rs_set_expected_tpt_table(struct iwl_lq_sta *lq_sta, struct iwl_scale_tbl_info *tbl) rs_set_expected_tpt_table() argument 1142 rs_get_best_rate(struct iwl_priv *priv, struct iwl_lq_sta *lq_sta, struct iwl_scale_tbl_info *tbl, u16 rate_mask, s8 index) rs_get_best_rate() argument 1235 rs_switch_to_mimo2(struct iwl_priv *priv, struct iwl_lq_sta *lq_sta, struct ieee80211_conf *conf, struct ieee80211_sta *sta, struct iwl_scale_tbl_info *tbl, int index) rs_switch_to_mimo2() argument 1290 rs_switch_to_mimo3(struct iwl_priv *priv, struct iwl_lq_sta *lq_sta, struct ieee80211_conf *conf, struct ieee80211_sta *sta, struct iwl_scale_tbl_info *tbl, int index) rs_switch_to_mimo3() argument 1346 rs_switch_to_siso(struct iwl_priv *priv, struct iwl_lq_sta *lq_sta, struct ieee80211_conf *conf, struct ieee80211_sta *sta, struct iwl_scale_tbl_info *tbl, int index) rs_switch_to_siso() argument 1401 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); rs_move_legacy_other() local 1574 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); rs_move_siso_to_other() local 1744 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); rs_move_mimo2_to_other() local 1912 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); rs_move_mimo3_to_other() local 2085 struct iwl_scale_tbl_info *tbl; rs_stay_in_table() local 2162 rs_update_rate_tbl(struct iwl_priv *priv, struct iwl_rxon_context *ctx, struct iwl_lq_sta *lq_sta, struct iwl_scale_tbl_info *tbl, int index, u8 is_green) rs_update_rate_tbl() argument 2200 struct iwl_scale_tbl_info *tbl, *tbl1; rs_rate_scale_perform() local 2633 struct iwl_scale_tbl_info *tbl; rs_initialize_lq() local 3099 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); rs_sta_dbgfs_scale_table_read() local 3224 struct iwl_scale_tbl_info *tbl = &lq_sta->lq_info[lq_sta->active_tbl]; rs_sta_dbgfs_rate_scale_data_read() local [all...] |
/kernel/linux/linux-6.6/drivers/net/wireless/intel/iwlwifi/dvm/ |
H A D | rs.c | 414 static s32 get_expected_tpt(struct iwl_scale_tbl_info *tbl, int rs_index) in get_expected_tpt() argument 416 if (tbl->expected_tpt) in get_expected_tpt() 417 return tbl->expected_tpt[rs_index]; in get_expected_tpt() 428 static int rs_collect_tx_data(struct iwl_scale_tbl_info *tbl, in rs_collect_tx_data() argument 439 window = &(tbl->win[scale_index]); in rs_collect_tx_data() 442 tpt = get_expected_tpt(tbl, scale_index); in rs_collect_tx_data() 507 struct iwl_scale_tbl_info *tbl, in rate_n_flags_from_tbl() 512 if (is_legacy(tbl->lq_type)) { in rate_n_flags_from_tbl() 517 } else if (is_Ht(tbl->lq_type)) { in rate_n_flags_from_tbl() 524 if (is_siso(tbl in rate_n_flags_from_tbl() 506 rate_n_flags_from_tbl(struct iwl_priv *priv, struct iwl_scale_tbl_info *tbl, int index, u8 use_green) rate_n_flags_from_tbl() argument 562 rs_get_tbl_info_from_mcs(const u32 rate_n_flags, enum nl80211_band band, struct iwl_scale_tbl_info *tbl, int *rate_idx) rs_get_tbl_info_from_mcs() argument 628 rs_toggle_antenna(u32 valid_ant, u32 *rate_n_flags, struct iwl_scale_tbl_info *tbl) rs_toggle_antenna() argument 752 rs_get_lower_rate(struct iwl_lq_sta *lq_sta, struct iwl_scale_tbl_info *tbl, u8 scale_index, u8 ht_possible) rs_get_lower_rate() argument 825 struct iwl_scale_tbl_info *tbl; rs_bt_update_lq() local 1064 rs_set_expected_tpt_table(struct iwl_lq_sta *lq_sta, struct iwl_scale_tbl_info *tbl) rs_set_expected_tpt_table() argument 1120 rs_get_best_rate(struct iwl_priv *priv, struct iwl_lq_sta *lq_sta, struct iwl_scale_tbl_info *tbl, u16 rate_mask, s8 index) rs_get_best_rate() argument 1213 rs_switch_to_mimo2(struct iwl_priv *priv, struct iwl_lq_sta *lq_sta, struct ieee80211_conf *conf, struct ieee80211_sta *sta, struct iwl_scale_tbl_info *tbl, int index) rs_switch_to_mimo2() argument 1268 rs_switch_to_mimo3(struct iwl_priv *priv, struct iwl_lq_sta *lq_sta, struct ieee80211_conf *conf, struct ieee80211_sta *sta, struct iwl_scale_tbl_info *tbl, int index) rs_switch_to_mimo3() argument 1324 rs_switch_to_siso(struct iwl_priv *priv, struct iwl_lq_sta *lq_sta, struct ieee80211_conf *conf, struct ieee80211_sta *sta, struct iwl_scale_tbl_info *tbl, int index) rs_switch_to_siso() argument 1379 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); rs_move_legacy_other() local 1552 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); rs_move_siso_to_other() local 1722 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); rs_move_mimo2_to_other() local 1890 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); rs_move_mimo3_to_other() local 2063 struct iwl_scale_tbl_info *tbl; rs_stay_in_table() local 2140 rs_update_rate_tbl(struct iwl_priv *priv, struct iwl_rxon_context *ctx, struct iwl_lq_sta *lq_sta, struct iwl_scale_tbl_info *tbl, int index, u8 is_green) rs_update_rate_tbl() argument 2178 struct iwl_scale_tbl_info *tbl, *tbl1; rs_rate_scale_perform() local 2611 struct iwl_scale_tbl_info *tbl; rs_initialize_lq() local 3094 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); rs_sta_dbgfs_scale_table_read() local 3222 struct iwl_scale_tbl_info *tbl = &lq_sta->lq_info[lq_sta->active_tbl]; rs_sta_dbgfs_rate_scale_data_read() local [all...] |
/kernel/linux/linux-5.10/arch/powerpc/platforms/powernv/ |
H A D | pci-ioda-tce.c | 48 void pnv_pci_setup_iommu_table(struct iommu_table *tbl, in pnv_pci_setup_iommu_table() argument 52 tbl->it_blocksize = 16; in pnv_pci_setup_iommu_table() 53 tbl->it_base = (unsigned long)tce_mem; in pnv_pci_setup_iommu_table() 54 tbl->it_page_shift = page_shift; in pnv_pci_setup_iommu_table() 55 tbl->it_offset = dma_offset >> tbl->it_page_shift; in pnv_pci_setup_iommu_table() 56 tbl->it_index = 0; in pnv_pci_setup_iommu_table() 57 tbl->it_size = tce_size >> 3; in pnv_pci_setup_iommu_table() 58 tbl->it_busno = 0; in pnv_pci_setup_iommu_table() 59 tbl in pnv_pci_setup_iommu_table() 83 pnv_tce(struct iommu_table *tbl, bool user, long idx, bool alloc) pnv_tce() argument 124 pnv_tce_build(struct iommu_table *tbl, long index, long npages, unsigned long uaddr, enum dma_data_direction direction, unsigned long attrs) pnv_tce_build() argument 147 pnv_tce_xchg(struct iommu_table *tbl, long index, unsigned long *hpa, enum dma_data_direction *direction, bool alloc) pnv_tce_xchg() argument 182 pnv_tce_useraddrptr(struct iommu_table *tbl, long index, bool alloc) pnv_tce_useraddrptr() argument 191 pnv_tce_free(struct iommu_table *tbl, long index, long npages) pnv_tce_free() argument 207 pnv_tce_get(struct iommu_table *tbl, long index) pnv_tce_get() argument 241 pnv_pci_ioda2_table_free_pages(struct iommu_table *tbl) pnv_pci_ioda2_table_free_pages() argument 291 pnv_pci_ioda2_table_alloc_pages(int nid, __u64 bus_offset, __u32 page_shift, __u64 window_size, __u32 levels, bool alloc_userspace_copy, struct iommu_table *tbl) pnv_pci_ioda2_table_alloc_pages() argument 371 pnv_pci_unlink_table_and_group(struct iommu_table *tbl, struct iommu_table_group *table_group) pnv_pci_unlink_table_and_group() argument 411 pnv_pci_link_table_and_group(int node, int num, struct iommu_table *tbl, struct iommu_table_group *table_group) pnv_pci_link_table_and_group() argument [all...] |
/kernel/linux/linux-6.6/arch/powerpc/platforms/powernv/ |
H A D | pci-ioda-tce.c | 48 void pnv_pci_setup_iommu_table(struct iommu_table *tbl, in pnv_pci_setup_iommu_table() argument 52 tbl->it_blocksize = 16; in pnv_pci_setup_iommu_table() 53 tbl->it_base = (unsigned long)tce_mem; in pnv_pci_setup_iommu_table() 54 tbl->it_page_shift = page_shift; in pnv_pci_setup_iommu_table() 55 tbl->it_offset = dma_offset >> tbl->it_page_shift; in pnv_pci_setup_iommu_table() 56 tbl->it_index = 0; in pnv_pci_setup_iommu_table() 57 tbl->it_size = tce_size >> 3; in pnv_pci_setup_iommu_table() 58 tbl->it_busno = 0; in pnv_pci_setup_iommu_table() 59 tbl in pnv_pci_setup_iommu_table() 83 pnv_tce(struct iommu_table *tbl, bool user, long idx, bool alloc) pnv_tce() argument 124 pnv_tce_build(struct iommu_table *tbl, long index, long npages, unsigned long uaddr, enum dma_data_direction direction, unsigned long attrs) pnv_tce_build() argument 147 pnv_tce_xchg(struct iommu_table *tbl, long index, unsigned long *hpa, enum dma_data_direction *direction) pnv_tce_xchg() argument 181 pnv_tce_useraddrptr(struct iommu_table *tbl, long index, bool alloc) pnv_tce_useraddrptr() argument 190 pnv_tce_free(struct iommu_table *tbl, long index, long npages) pnv_tce_free() argument 206 pnv_tce_get(struct iommu_table *tbl, long index) pnv_tce_get() argument 240 pnv_pci_ioda2_table_free_pages(struct iommu_table *tbl) pnv_pci_ioda2_table_free_pages() argument 290 pnv_pci_ioda2_table_alloc_pages(int nid, __u64 bus_offset, __u32 page_shift, __u64 window_size, __u32 levels, bool alloc_userspace_copy, struct iommu_table *tbl) pnv_pci_ioda2_table_alloc_pages() argument 370 pnv_pci_unlink_table_and_group(struct iommu_table *tbl, struct iommu_table_group *table_group) pnv_pci_unlink_table_and_group() argument 410 pnv_pci_link_table_and_group(int node, int num, struct iommu_table *tbl, struct iommu_table_group *table_group) pnv_pci_link_table_and_group() argument [all...] |
/kernel/linux/linux-5.10/lib/ |
H A D | rhashtable.c | 38 const struct bucket_table *tbl, in head_hashfn() 41 return rht_head_hashfn(ht, tbl, he, ht->p); in head_hashfn() 53 int lockdep_rht_bucket_is_held(const struct bucket_table *tbl, u32 hash) in lockdep_rht_bucket_is_held() argument 57 if (unlikely(tbl->nest)) in lockdep_rht_bucket_is_held() 59 return bit_spin_is_locked(0, (unsigned long *)&tbl->buckets[hash]); in lockdep_rht_bucket_is_held() 67 const struct bucket_table *tbl) in nested_table_top() 70 * because it's set at the same time as tbl->nest. in nested_table_top() 72 return (void *)rcu_dereference_protected(tbl->buckets[0], 1); in nested_table_top() 94 static void nested_bucket_table_free(const struct bucket_table *tbl) in nested_bucket_table_free() argument 96 unsigned int size = tbl in nested_bucket_table_free() 37 head_hashfn(struct rhashtable *ht, const struct bucket_table *tbl, const struct rhash_head *he) head_hashfn() argument 66 nested_table_top( const struct bucket_table *tbl) nested_table_top() argument 109 bucket_table_free(const struct bucket_table *tbl) bucket_table_free() argument 152 struct bucket_table *tbl; nested_bucket_table_alloc() local 179 struct bucket_table *tbl = NULL; bucket_table_alloc() local 211 rhashtable_last_table(struct rhashtable *ht, struct bucket_table *tbl) rhashtable_last_table() argument 410 struct bucket_table *tbl; rht_deferred_worker() local 439 rhashtable_insert_rehash(struct rhashtable *ht, struct bucket_table *tbl) rhashtable_insert_rehash() argument 487 rhashtable_lookup_one(struct rhashtable *ht, struct rhash_lock_head __rcu **bkt, struct bucket_table *tbl, unsigned int hash, const void *key, struct rhash_head *obj) rhashtable_lookup_one() argument 538 rhashtable_insert_one( struct rhashtable *ht, struct rhash_lock_head __rcu **bkt, struct bucket_table *tbl, unsigned int hash, struct rhash_head *obj, void *data) rhashtable_insert_one() argument 591 struct bucket_table *tbl; rhashtable_try_insert() local 789 struct bucket_table *tbl = iter->walker.tbl; __rhashtable_walk_find_next() local 934 struct bucket_table *tbl = iter->walker.tbl; __releases() local 1018 struct bucket_table *tbl; rhashtable_init() local 1134 struct bucket_table *tbl, *next_tbl; rhashtable_free_and_destroy() local 1174 __rht_bucket_nested( const struct bucket_table *tbl, unsigned int hash) __rht_bucket_nested() argument 1203 rht_bucket_nested( const struct bucket_table *tbl, unsigned int hash) rht_bucket_nested() argument 1214 rht_bucket_nested_insert( struct rhashtable *ht, struct bucket_table *tbl, unsigned int hash) rht_bucket_nested_insert() argument [all...] |
/kernel/linux/linux-6.6/lib/ |
H A D | rhashtable.c | 38 const struct bucket_table *tbl, in head_hashfn() 41 return rht_head_hashfn(ht, tbl, he, ht->p); in head_hashfn() 53 int lockdep_rht_bucket_is_held(const struct bucket_table *tbl, u32 hash) in lockdep_rht_bucket_is_held() argument 57 if (unlikely(tbl->nest)) in lockdep_rht_bucket_is_held() 59 return bit_spin_is_locked(0, (unsigned long *)&tbl->buckets[hash]); in lockdep_rht_bucket_is_held() 67 const struct bucket_table *tbl) in nested_table_top() 70 * because it's set at the same time as tbl->nest. in nested_table_top() 72 return (void *)rcu_dereference_protected(tbl->buckets[0], 1); in nested_table_top() 94 static void nested_bucket_table_free(const struct bucket_table *tbl) in nested_bucket_table_free() argument 96 unsigned int size = tbl in nested_bucket_table_free() 37 head_hashfn(struct rhashtable *ht, const struct bucket_table *tbl, const struct rhash_head *he) head_hashfn() argument 66 nested_table_top( const struct bucket_table *tbl) nested_table_top() argument 109 bucket_table_free(const struct bucket_table *tbl) bucket_table_free() argument 152 struct bucket_table *tbl; nested_bucket_table_alloc() local 179 struct bucket_table *tbl = NULL; bucket_table_alloc() local 211 rhashtable_last_table(struct rhashtable *ht, struct bucket_table *tbl) rhashtable_last_table() argument 413 struct bucket_table *tbl; rht_deferred_worker() local 442 rhashtable_insert_rehash(struct rhashtable *ht, struct bucket_table *tbl) rhashtable_insert_rehash() argument 490 rhashtable_lookup_one(struct rhashtable *ht, struct rhash_lock_head __rcu **bkt, struct bucket_table *tbl, unsigned int hash, const void *key, struct rhash_head *obj) rhashtable_lookup_one() argument 541 rhashtable_insert_one( struct rhashtable *ht, struct rhash_lock_head __rcu **bkt, struct bucket_table *tbl, unsigned int hash, struct rhash_head *obj, void *data) rhashtable_insert_one() argument 594 struct bucket_table *tbl; rhashtable_try_insert() local 793 struct bucket_table *tbl = iter->walker.tbl; __rhashtable_walk_find_next() local 938 struct bucket_table *tbl = iter->walker.tbl; __releases() local 1022 struct bucket_table *tbl; rhashtable_init() local 1138 struct bucket_table *tbl, *next_tbl; rhashtable_free_and_destroy() local 1178 __rht_bucket_nested( const struct bucket_table *tbl, unsigned int hash) __rht_bucket_nested() argument 1207 rht_bucket_nested( const struct bucket_table *tbl, unsigned int hash) rht_bucket_nested() argument 1218 rht_bucket_nested_insert( struct rhashtable *ht, struct bucket_table *tbl, unsigned int hash) rht_bucket_nested_insert() argument [all...] |
/kernel/linux/linux-5.10/net/core/ |
H A D | neighbour.c | 58 static int pneigh_ifdown_and_unlock(struct neigh_table *tbl, 66 Neighbour hash table buckets are protected with rwlock tbl->lock. 124 atomic_dec(&n->tbl->gc_entries); in neigh_mark_dead() 132 write_lock_bh(&n->tbl->lock); in neigh_update_gc_list() 147 atomic_dec(&n->tbl->gc_entries); in neigh_update_gc_list() 150 list_add_tail(&n->gc_list, &n->tbl->gc_list); in neigh_update_gc_list() 151 atomic_inc(&n->tbl->gc_entries); in neigh_update_gc_list() 156 write_unlock_bh(&n->tbl->lock); in neigh_update_gc_list() 182 struct neigh_table *tbl) in neigh_del() 191 lockdep_is_held(&tbl in neigh_del() 181 neigh_del(struct neighbour *n, struct neighbour __rcu **np, struct neigh_table *tbl) neigh_del() argument 202 neigh_remove_one(struct neighbour *ndel, struct neigh_table *tbl) neigh_remove_one() argument 225 neigh_forced_gc(struct neigh_table *tbl) neigh_forced_gc() argument 326 neigh_flush_dev(struct neigh_table *tbl, struct net_device *dev, bool skip_perm) neigh_flush_dev() argument 380 neigh_changeaddr(struct neigh_table *tbl, struct net_device *dev) neigh_changeaddr() argument 388 __neigh_ifdown(struct neigh_table *tbl, struct net_device *dev, bool skip_perm) __neigh_ifdown() argument 400 neigh_carrier_down(struct neigh_table *tbl, struct net_device *dev) neigh_carrier_down() argument 407 neigh_ifdown(struct neigh_table *tbl, struct net_device *dev) neigh_ifdown() argument 414 neigh_alloc(struct neigh_table *tbl, struct net_device *dev, u8 flags, bool exempt_from_gc) neigh_alloc() argument 521 neigh_hash_grow(struct neigh_table *tbl, unsigned long new_shift) neigh_hash_grow() argument 562 neigh_lookup(struct neigh_table *tbl, const void *pkey, struct net_device *dev) neigh_lookup() argument 583 ___neigh_create(struct neigh_table *tbl, const void *pkey, struct net_device *dev, u8 flags, bool exempt_from_gc, bool want_ref) ___neigh_create() argument 677 __neigh_create(struct neigh_table *tbl, const void *pkey, struct net_device *dev, bool want_ref) __neigh_create() argument 710 __pneigh_lookup(struct neigh_table *tbl, struct net *net, const void *pkey, struct net_device *dev) __pneigh_lookup() argument 721 pneigh_lookup(struct neigh_table *tbl, struct net *net, const void *pkey, struct net_device *dev, int creat) pneigh_lookup() argument 767 pneigh_delete(struct neigh_table *tbl, struct net *net, const void *pkey, struct net_device *dev) pneigh_delete() argument 793 pneigh_ifdown_and_unlock(struct neigh_table *tbl, struct net_device *dev) pneigh_ifdown_and_unlock() argument 895 struct neigh_table *tbl = container_of(work, struct neigh_table, gc_work.work); neigh_periodic_work() local 1456 neigh_event_ns(struct neigh_table *tbl, u8 *lladdr, void *saddr, struct net_device *dev) neigh_event_ns() argument 1555 struct neigh_table *tbl = from_timer(tbl, t, proxy_timer); neigh_proxy_process() local 1587 pneigh_enqueue(struct neigh_table *tbl, struct neigh_parms *p, struct sk_buff *skb) pneigh_enqueue() argument 1616 lookup_neigh_parms(struct neigh_table *tbl, struct net *net, int ifindex) lookup_neigh_parms() argument 1630 neigh_parms_alloc(struct net_device *dev, struct neigh_table *tbl) neigh_parms_alloc() argument 1672 neigh_parms_release(struct neigh_table *tbl, struct neigh_parms *parms) neigh_parms_release() argument 1695 neigh_table_init(int index, struct neigh_table *tbl) neigh_table_init() argument 1747 neigh_table_clear(int index, struct neigh_table *tbl) neigh_table_clear() argument 1776 struct neigh_table *tbl = NULL; neigh_find_table() local 1817 struct neigh_table *tbl; neigh_delete() local 1884 struct neigh_table *tbl; neigh_add() local 2054 neightbl_fill_info(struct sk_buff *skb, struct neigh_table *tbl, u32 pid, u32 seq, int type, int flags) neightbl_fill_info() argument 2144 neightbl_fill_param_info(struct sk_buff *skb, struct neigh_table *tbl, struct neigh_parms *parms, u32 pid, u32 seq, int type, unsigned int flags) neightbl_fill_param_info() argument 2207 struct neigh_table *tbl; neightbl_set() local 2396 struct neigh_table *tbl; neightbl_dump_info() local 2506 pneigh_fill_info(struct sk_buff *skb, struct pneigh_entry *pn, u32 pid, u32 seq, int type, unsigned int flags, struct neigh_table *tbl) pneigh_fill_info() argument 2573 neigh_dump_table(struct neigh_table *tbl, struct sk_buff *skb, struct netlink_callback *cb, struct neigh_dump_filter *filter) neigh_dump_table() argument 2620 pneigh_dump_table(struct neigh_table *tbl, struct sk_buff *skb, struct netlink_callback *cb, struct neigh_dump_filter *filter) pneigh_dump_table() argument 2730 struct neigh_table *tbl; neigh_dump_info() local 2772 neigh_valid_get_req(const struct nlmsghdr *nlh, struct neigh_table **tbl, void **dst, int *dev_idx, u8 *ndm_flags, struct netlink_ext_ack *extack) neigh_valid_get_req() argument 2870 pneigh_get_reply(struct net *net, struct pneigh_entry *neigh, u32 pid, u32 seq, struct neigh_table *tbl) pneigh_get_reply() argument 2896 struct neigh_table *tbl = NULL; neigh_get() local 2952 neigh_for_each(struct neigh_table *tbl, void (*cb)(struct neighbour *, void *), void *cookie) neigh_for_each() argument 2975 __neigh_for_each_release(struct neigh_table *tbl, int (*cb)(struct neighbour *)) __neigh_for_each_release() argument 3014 struct neigh_table *tbl; neigh_xmit() local 3161 struct neigh_table *tbl = state->tbl; pneigh_get_first() local 3184 struct neigh_table *tbl = state->tbl; pneigh_get_next() local 3284 struct neigh_table *tbl = state->tbl; __releases() local 3295 struct neigh_table *tbl = PDE_DATA(file_inode(seq->file)); neigh_stat_seq_start() local 3312 struct neigh_table *tbl = PDE_DATA(file_inode(seq->file)); neigh_stat_seq_next() local 3332 struct neigh_table *tbl = PDE_DATA(file_inode(seq->file)); neigh_stat_seq_show() local 3668 struct neigh_table *tbl = p->tbl; neigh_sysctl_register() local [all...] |
/kernel/linux/linux-5.10/net/mac80211/ |
H A D | mesh_pathtbl.c | 18 static void mesh_path_free_rcu(struct mesh_table *tbl, struct mesh_path *mpath); 45 struct mesh_table *tbl = tblptr; in mesh_path_rht_free() local 47 mesh_path_free_rcu(tbl, mpath); in mesh_path_rht_free() 50 static void mesh_table_init(struct mesh_table *tbl) in mesh_table_init() argument 52 INIT_HLIST_HEAD(&tbl->known_gates); in mesh_table_init() 53 INIT_HLIST_HEAD(&tbl->walk_head); in mesh_table_init() 54 atomic_set(&tbl->entries, 0); in mesh_table_init() 55 spin_lock_init(&tbl->gates_lock); in mesh_table_init() 56 spin_lock_init(&tbl->walk_lock); in mesh_table_init() 61 WARN_ON(rhashtable_init(&tbl in mesh_table_init() 64 mesh_table_free(struct mesh_table *tbl) mesh_table_free() argument 206 mpath_lookup(struct mesh_table *tbl, const u8 *dst, struct ieee80211_sub_if_data *sdata) mpath_lookup() argument 243 __mesh_path_lookup_by_idx(struct mesh_table *tbl, int idx) __mesh_path_lookup_by_idx() argument 300 struct mesh_table *tbl; mesh_path_add_gate() local 335 mesh_gate_del(struct mesh_table *tbl, struct mesh_path *mpath) mesh_gate_del() argument 396 struct mesh_table *tbl; mesh_path_add() local 438 struct mesh_table *tbl; mpp_path_add() local 484 struct mesh_table *tbl = &sdata->u.mesh.mesh_paths; mesh_plink_broken() local 506 mesh_path_free_rcu(struct mesh_table *tbl, struct mesh_path *mpath) mesh_path_free_rcu() argument 522 __mesh_path_del(struct mesh_table *tbl, struct mesh_path *mpath) __mesh_path_del() argument 543 struct mesh_table *tbl = &sdata->u.mesh.mesh_paths; mesh_path_flush_by_nexthop() local 558 struct mesh_table *tbl = &sdata->u.mesh.mpp_paths; mpp_flush_by_proxy() local 570 table_flush_by_iface(struct mesh_table *tbl) table_flush_by_iface() argument 605 table_path_del(struct mesh_table *tbl, struct ieee80211_sub_if_data *sdata, const u8 *addr) table_path_del() argument 672 struct mesh_table *tbl; mesh_path_send_to_gates() local 777 mesh_path_tbl_expire(struct ieee80211_sub_if_data *sdata, struct mesh_table *tbl) mesh_path_tbl_expire() argument [all...] |
/kernel/linux/linux-5.10/net/netfilter/ipvs/ |
H A D | ip_vs_lblc.c | 9 * *lock(tbl->lock) ==> *lock(&tbl->lock) 10 * Wensong Zhang : fixed the uninitialized tbl->lock bug 168 ip_vs_lblc_hash(struct ip_vs_lblc_table *tbl, struct ip_vs_lblc_entry *en) in ip_vs_lblc_hash() argument 172 hlist_add_head_rcu(&en->list, &tbl->bucket[hash]); in ip_vs_lblc_hash() 173 atomic_inc(&tbl->entries); in ip_vs_lblc_hash() 179 ip_vs_lblc_get(int af, struct ip_vs_lblc_table *tbl, in ip_vs_lblc_get() argument 185 hlist_for_each_entry_rcu(en, &tbl->bucket[hash], list) in ip_vs_lblc_get() 198 ip_vs_lblc_new(struct ip_vs_lblc_table *tbl, const union nf_inet_addr *daddr, in ip_vs_lblc_new() argument 203 en = ip_vs_lblc_get(af, tbl, dadd in ip_vs_lblc_new() 231 struct ip_vs_lblc_table *tbl = svc->sched_data; ip_vs_lblc_flush() local 258 struct ip_vs_lblc_table *tbl = svc->sched_data; ip_vs_lblc_full_check() local 296 struct ip_vs_lblc_table *tbl = from_timer(tbl, t, periodic_timer); ip_vs_lblc_check_expire() local 346 struct ip_vs_lblc_table *tbl; ip_vs_lblc_init_svc() local 384 struct ip_vs_lblc_table *tbl = svc->sched_data; ip_vs_lblc_done_svc() local 484 struct ip_vs_lblc_table *tbl = svc->sched_data; ip_vs_lblc_schedule() local [all...] |
/kernel/linux/linux-6.6/net/netfilter/ipvs/ |
H A D | ip_vs_lblc.c | 9 * *lock(tbl->lock) ==> *lock(&tbl->lock) 10 * Wensong Zhang : fixed the uninitialized tbl->lock bug 168 ip_vs_lblc_hash(struct ip_vs_lblc_table *tbl, struct ip_vs_lblc_entry *en) in ip_vs_lblc_hash() argument 172 hlist_add_head_rcu(&en->list, &tbl->bucket[hash]); in ip_vs_lblc_hash() 173 atomic_inc(&tbl->entries); in ip_vs_lblc_hash() 179 ip_vs_lblc_get(int af, struct ip_vs_lblc_table *tbl, in ip_vs_lblc_get() argument 185 hlist_for_each_entry_rcu(en, &tbl->bucket[hash], list) in ip_vs_lblc_get() 198 ip_vs_lblc_new(struct ip_vs_lblc_table *tbl, const union nf_inet_addr *daddr, in ip_vs_lblc_new() argument 203 en = ip_vs_lblc_get(af, tbl, dadd in ip_vs_lblc_new() 231 struct ip_vs_lblc_table *tbl = svc->sched_data; ip_vs_lblc_flush() local 258 struct ip_vs_lblc_table *tbl = svc->sched_data; ip_vs_lblc_full_check() local 296 struct ip_vs_lblc_table *tbl = from_timer(tbl, t, periodic_timer); ip_vs_lblc_check_expire() local 346 struct ip_vs_lblc_table *tbl; ip_vs_lblc_init_svc() local 384 struct ip_vs_lblc_table *tbl = svc->sched_data; ip_vs_lblc_done_svc() local 484 struct ip_vs_lblc_table *tbl = svc->sched_data; ip_vs_lblc_schedule() local [all...] |
/kernel/linux/linux-5.10/arch/powerpc/include/asm/ |
H A D | iommu.h | 40 int (*set)(struct iommu_table *tbl, 51 int (*xchg_no_kill)(struct iommu_table *tbl, 57 void (*tce_kill)(struct iommu_table *tbl, 62 __be64 *(*useraddrptr)(struct iommu_table *tbl, long index, bool alloc); 64 void (*clear)(struct iommu_table *tbl, 67 unsigned long (*get)(struct iommu_table *tbl, long index); 68 void (*flush)(struct iommu_table *tbl); 69 void (*free)(struct iommu_table *tbl); 119 #define IOMMU_TABLE_USERSPACE_ENTRY_RO(tbl, entry) \ 120 ((tbl) 126 get_iommu_order(unsigned long size, struct iommu_table *tbl) get_iommu_order() argument [all...] |
/kernel/linux/linux-6.6/arch/powerpc/include/asm/ |
H A D | iommu.h | 43 int (*set)(struct iommu_table *tbl, 54 int (*xchg_no_kill)(struct iommu_table *tbl, 59 void (*tce_kill)(struct iommu_table *tbl, 63 __be64 *(*useraddrptr)(struct iommu_table *tbl, long index, bool alloc); 65 void (*clear)(struct iommu_table *tbl, 68 unsigned long (*get)(struct iommu_table *tbl, long index); 69 void (*flush)(struct iommu_table *tbl); 70 void (*free)(struct iommu_table *tbl); 120 #define IOMMU_TABLE_USERSPACE_ENTRY_RO(tbl, entry) \ 121 ((tbl) 127 get_iommu_order(unsigned long size, struct iommu_table *tbl) get_iommu_order() argument [all...] |
/kernel/linux/linux-5.10/drivers/net/wireless/marvell/mwifiex/ |
H A D | 11n_rxreorder.c | 110 struct mwifiex_rx_reorder_tbl *tbl, in mwifiex_11n_dispatch_pkt_until_start_win() 120 pkt_to_send = (start_win > tbl->start_win) ? in mwifiex_11n_dispatch_pkt_until_start_win() 121 min((start_win - tbl->start_win), tbl->win_size) : in mwifiex_11n_dispatch_pkt_until_start_win() 122 tbl->win_size; in mwifiex_11n_dispatch_pkt_until_start_win() 125 if (tbl->rx_reorder_ptr[i]) { in mwifiex_11n_dispatch_pkt_until_start_win() 126 skb = tbl->rx_reorder_ptr[i]; in mwifiex_11n_dispatch_pkt_until_start_win() 128 tbl->rx_reorder_ptr[i] = NULL; in mwifiex_11n_dispatch_pkt_until_start_win() 136 for (i = 0; i < tbl->win_size - pkt_to_send; ++i) { in mwifiex_11n_dispatch_pkt_until_start_win() 137 tbl in mwifiex_11n_dispatch_pkt_until_start_win() 109 mwifiex_11n_dispatch_pkt_until_start_win(struct mwifiex_private *priv, struct mwifiex_rx_reorder_tbl *tbl, int start_win) mwifiex_11n_dispatch_pkt_until_start_win() argument 157 mwifiex_11n_scan_and_dispatch(struct mwifiex_private *priv, struct mwifiex_rx_reorder_tbl *tbl) mwifiex_11n_scan_and_dispatch() argument 201 mwifiex_del_rx_reorder_entry(struct mwifiex_private *priv, struct mwifiex_rx_reorder_tbl *tbl) mwifiex_del_rx_reorder_entry() argument 244 struct mwifiex_rx_reorder_tbl *tbl; mwifiex_11n_get_rx_reorder_tbl() local 263 struct mwifiex_rx_reorder_tbl *tbl, *tmp; mwifiex_11n_del_rx_reorder_tbl_by_ta() local 345 struct mwifiex_rx_reorder_tbl *tbl, *new_node; mwifiex_11n_create_rx_reorder_tbl() local 422 mwifiex_11n_rxreorder_timer_restart(struct mwifiex_rx_reorder_tbl *tbl) mwifiex_11n_rxreorder_timer_restart() argument 563 struct mwifiex_rx_reorder_tbl *tbl; mwifiex_11n_rx_reorder_pkt() local 675 struct mwifiex_rx_reorder_tbl *tbl; mwifiex_del_ba_tbl() local 729 struct mwifiex_rx_reorder_tbl *tbl; mwifiex_ret_11n_addba_resp() local 820 struct mwifiex_rx_reorder_tbl *tbl; mwifiex_update_rxreor_flags() local [all...] |