/kernel/linux/linux-5.10/mm/ |
H A D | swap_slots.c | 3 * Manage cache of swap slots to be used for and returned from 10 * We allocate the swap slots from the global pool and put 17 * lock. We do not reuse the returned slots directly but 19 * allows the slots to coaellesce and reduce fragmentation. 25 * The swap slots cache is protected by a mutex instead of 26 * a spin lock as when we search for slots with scan_swap_map, 42 /* Serialize swap slots cache enable/disable operations */ 117 swp_entry_t *slots, *slots_ret; in alloc_swap_slot_cache() local 124 slots = kvcalloc(SWAP_SLOTS_CACHE_SIZE, sizeof(swp_entry_t), in alloc_swap_slot_cache() 126 if (!slots) in alloc_swap_slot_cache() 173 swp_entry_t *slots = NULL; drain_slots_cache_cpu() local [all...] |
H A D | z3fold.c | 90 * be enough slots to hold all possible variants 106 * @slots: pointer to the structure holding buddy slots 120 struct z3fold_buddy_slots *slots; member 213 struct z3fold_buddy_slots *slots; in alloc_slots() local 215 slots = kmem_cache_zalloc(pool->c_handle, in alloc_slots() 218 if (slots) { in alloc_slots() 220 kmemleak_not_leak(slots); in alloc_slots() 221 slots->pool = (unsigned long)pool; in alloc_slots() 222 rwlock_init(&slots in alloc_slots() 260 struct z3fold_buddy_slots *slots; __get_z3fold_header() local 308 struct z3fold_buddy_slots *slots; free_handle() local 401 struct z3fold_buddy_slots *slots; init_z3fold_page() local 455 __encode_handle(struct z3fold_header *zhdr, struct z3fold_buddy_slots *slots, enum buddy bud) __encode_handle() argument 489 struct z3fold_buddy_slots *slots = handle_to_slots(handle); handle_to_chunks() local 506 struct z3fold_buddy_slots *slots = handle_to_slots(handle); handle_to_buddy() local 1335 struct z3fold_buddy_slots slots __attribute__((aligned(SLOTS_ALIGN))); z3fold_reclaim_page() local 1465 struct z3fold_buddy_slots *slots = zhdr->slots; z3fold_reclaim_page() local [all...] |
/kernel/linux/linux-6.6/mm/ |
H A D | swap_slots.c | 3 * Manage cache of swap slots to be used for and returned from 10 * We allocate the swap slots from the global pool and put 17 * lock. We do not reuse the returned slots directly but 19 * allows the slots to coalesce and reduce fragmentation. 25 * The swap slots cache is protected by a mutex instead of 26 * a spin lock as when we search for slots with scan_swap_map, 43 /* Serialize swap slots cache enable/disable operations */ 116 swp_entry_t *slots, *slots_ret; in alloc_swap_slot_cache() local 123 slots = kvcalloc(SWAP_SLOTS_CACHE_SIZE, sizeof(swp_entry_t), in alloc_swap_slot_cache() 125 if (!slots) in alloc_swap_slot_cache() 172 swp_entry_t *slots = NULL; drain_slots_cache_cpu() local [all...] |
H A D | z3fold.c | 83 * be enough slots to hold all possible variants 99 * @slots: pointer to the structure holding buddy slots 113 struct z3fold_buddy_slots *slots; member 197 struct z3fold_buddy_slots *slots = kmem_cache_zalloc(pool->c_handle, in alloc_slots() local 200 if (slots) { in alloc_slots() 202 kmemleak_not_leak(slots); in alloc_slots() 203 slots->pool = (unsigned long)pool; in alloc_slots() 204 rwlock_init(&slots->lock); in alloc_slots() 207 return slots; in alloc_slots() 241 struct z3fold_buddy_slots *slots; get_z3fold_header() local 281 struct z3fold_buddy_slots *slots; free_handle() local 323 struct z3fold_buddy_slots *slots; init_z3fold_page() local 370 __encode_handle(struct z3fold_header *zhdr, struct z3fold_buddy_slots *slots, enum buddy bud) __encode_handle() argument 404 struct z3fold_buddy_slots *slots = handle_to_slots(handle); handle_to_chunks() local 421 struct z3fold_buddy_slots *slots = handle_to_slots(handle); handle_to_buddy() local [all...] |
/kernel/linux/linux-5.10/sound/pci/ac97/ |
H A D | ac97_pcm.c | 295 * slots 7+8 in snd_ac97_set_rate() 312 unsigned short slots = 0; in get_pslots() local 326 slots |= (1<<AC97_SLOT_PCM_LEFT)|(1<<AC97_SLOT_PCM_RIGHT); in get_pslots() 328 slots |= (1<<AC97_SLOT_PCM_SLEFT)|(1<<AC97_SLOT_PCM_SRIGHT); in get_pslots() 330 slots |= (1<<AC97_SLOT_PCM_CENTER)|(1<<AC97_SLOT_LFE); in get_pslots() 343 slots |= (1<<AC97_SLOT_PCM_SLEFT)|(1<<AC97_SLOT_PCM_SRIGHT); in get_pslots() 345 slots |= (1<<AC97_SLOT_PCM_CENTER)|(1<<AC97_SLOT_LFE); in get_pslots() 355 slots |= (1<<AC97_SLOT_PCM_CENTER)|(1<<AC97_SLOT_LFE); in get_pslots() 361 return slots; in get_pslots() 363 unsigned short slots; in get_pslots() local 384 unsigned short slots; get_cslots() local 393 get_rates(struct ac97_pcm *pcm, unsigned int cidx, unsigned short slots, int dbl) get_rates() argument 440 unsigned short tmp, slots; snd_ac97_pcm_assign() local 555 snd_ac97_pcm_open(struct ac97_pcm *pcm, unsigned int rate, enum ac97_pcm_cfg cfg, unsigned short slots) snd_ac97_pcm_open() argument 649 unsigned short slots = pcm->aslots; snd_ac97_pcm_close() local [all...] |
/kernel/linux/linux-6.6/sound/pci/ac97/ |
H A D | ac97_pcm.c | 295 * slots 7+8 in snd_ac97_set_rate() 312 unsigned short slots = 0; in get_pslots() local 326 slots |= (1<<AC97_SLOT_PCM_LEFT)|(1<<AC97_SLOT_PCM_RIGHT); in get_pslots() 328 slots |= (1<<AC97_SLOT_PCM_SLEFT)|(1<<AC97_SLOT_PCM_SRIGHT); in get_pslots() 330 slots |= (1<<AC97_SLOT_PCM_CENTER)|(1<<AC97_SLOT_LFE); in get_pslots() 343 slots |= (1<<AC97_SLOT_PCM_SLEFT)|(1<<AC97_SLOT_PCM_SRIGHT); in get_pslots() 345 slots |= (1<<AC97_SLOT_PCM_CENTER)|(1<<AC97_SLOT_LFE); in get_pslots() 355 slots |= (1<<AC97_SLOT_PCM_CENTER)|(1<<AC97_SLOT_LFE); in get_pslots() 361 return slots; in get_pslots() 363 unsigned short slots; in get_pslots() local 384 unsigned short slots; get_cslots() local 393 get_rates(struct ac97_pcm *pcm, unsigned int cidx, unsigned short slots, int dbl) get_rates() argument 440 unsigned short tmp, slots; snd_ac97_pcm_assign() local 555 snd_ac97_pcm_open(struct ac97_pcm *pcm, unsigned int rate, enum ac97_pcm_cfg cfg, unsigned short slots) snd_ac97_pcm_open() argument 649 unsigned short slots = pcm->aslots; snd_ac97_pcm_close() local [all...] |
/kernel/linux/linux-6.6/tools/testing/selftests/kvm/lib/ |
H A D | memstress.c | 126 uint64_t vcpu_memory_bytes, int slots, in memstress_create_vm() 155 TEST_ASSERT(guest_num_pages % slots == 0, in memstress_create_vm() 156 "Guest memory cannot be evenly divided into %d slots.", in memstress_create_vm() 157 slots); in memstress_create_vm() 207 /* Add extra memory slots for testing */ in memstress_create_vm() 208 for (i = 0; i < slots; i++) { in memstress_create_vm() 209 uint64_t region_pages = guest_num_pages / slots; in memstress_create_vm() 328 static void toggle_dirty_logging(struct kvm_vm *vm, int slots, bool enable) in toggle_dirty_logging() argument 332 for (i = 0; i < slots; i++) { in toggle_dirty_logging() 340 void memstress_enable_dirty_logging(struct kvm_vm *vm, int slots) in memstress_enable_dirty_logging() argument 125 memstress_create_vm(enum vm_guest_mode mode, int nr_vcpus, uint64_t vcpu_memory_bytes, int slots, enum vm_mem_backing_src_type backing_src, bool partition_vcpu_memory_access) memstress_create_vm() argument 345 memstress_disable_dirty_logging(struct kvm_vm *vm, int slots) memstress_disable_dirty_logging() argument 350 memstress_get_dirty_log(struct kvm_vm *vm, unsigned long *bitmaps[], int slots) memstress_get_dirty_log() argument 361 memstress_clear_dirty_log(struct kvm_vm *vm, unsigned long *bitmaps[], int slots, uint64_t pages_per_slot) memstress_clear_dirty_log() argument 373 memstress_alloc_bitmaps(int slots, uint64_t pages_per_slot) memstress_alloc_bitmaps() argument 389 memstress_free_bitmaps(unsigned long *bitmaps[], int slots) memstress_free_bitmaps() argument [all...] |
/kernel/linux/linux-6.6/drivers/input/ |
H A D | input-mt.c | 25 * input_mt_init_slots() - initialize MT input slots 27 * @num_slots: number of slots used by the device 37 * reinitialize with a different number of slots. 50 mt = kzalloc(struct_size(mt, slots, num_slots), GFP_KERNEL); in input_mt_init_slots() 89 /* Mark slots as 'inactive' */ in input_mt_init_slots() 91 input_mt_set_value(&mt->slots[i], ABS_MT_TRACKING_ID, -1); in input_mt_init_slots() 93 /* Mark slots as 'unused' */ in input_mt_init_slots() 105 * input_mt_destroy_slots() - frees the MT slots of the input device 106 * @dev: input device with allocated MT slots 109 * automatically free the MT slots whe 426 input_mt_set_slots(struct input_mt *mt, int *slots, int num_pos) input_mt_set_slots() argument 480 input_mt_assign_slots(struct input_dev *dev, int *slots, const struct input_mt_pos *pos, int num_pos, int dmax) input_mt_assign_slots() argument [all...] |
/kernel/linux/linux-5.10/drivers/input/ |
H A D | input-mt.c | 24 * input_mt_init_slots() - initialize MT input slots 26 * @num_slots: number of slots used by the device 36 * reinitialize with a different number of slots. 49 mt = kzalloc(struct_size(mt, slots, num_slots), GFP_KERNEL); in input_mt_init_slots() 88 /* Mark slots as 'inactive' */ in input_mt_init_slots() 90 input_mt_set_value(&mt->slots[i], ABS_MT_TRACKING_ID, -1); in input_mt_init_slots() 92 /* Mark slots as 'unused' */ in input_mt_init_slots() 104 * input_mt_destroy_slots() - frees the MT slots of the input device 105 * @dev: input device with allocated MT slots 108 * automatically free the MT slots whe 386 input_mt_set_slots(struct input_mt *mt, int *slots, int num_pos) input_mt_set_slots() argument 440 input_mt_assign_slots(struct input_dev *dev, int *slots, const struct input_mt_pos *pos, int num_pos, int dmax) input_mt_assign_slots() argument [all...] |
/kernel/linux/linux-6.6/fs/btrfs/ |
H A D | file-item.c | 184 item = btrfs_item_ptr(leaf, path->slots[0], in btrfs_insert_hole_extent() 228 if (path->slots[0] == 0) in btrfs_lookup_csum() 230 path->slots[0]--; in btrfs_lookup_csum() 231 btrfs_item_key_to_cpu(leaf, &found_key, path->slots[0]); in btrfs_lookup_csum() 237 csums_in_item = btrfs_item_size(leaf, path->slots[0]); in btrfs_lookup_csum() 247 item = btrfs_item_ptr(leaf, path->slots[0], struct btrfs_csum_item); in btrfs_lookup_csum() 301 item = btrfs_item_ptr(path->nodes[0], path->slots[0], in search_csum_tree() 303 btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); in search_csum_tree() 304 itemsize = btrfs_item_size(path->nodes[0], path->slots[0]); in search_csum_tree() 321 btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[ in search_csum_tree() [all...] |
H A D | inode-item.c | 102 return btrfs_find_name_in_ext_backref(path->nodes[0], path->slots[0], in btrfs_lookup_inode_extref() 142 extref = btrfs_find_name_in_ext_backref(path->nodes[0], path->slots[0], in btrfs_del_inode_extref() 151 item_size = btrfs_item_size(leaf, path->slots[0]); in btrfs_del_inode_extref() 165 item_start = btrfs_item_ptr_offset(leaf, path->slots[0]); in btrfs_del_inode_extref() 211 ref = btrfs_find_name_in_backref(path->nodes[0], path->slots[0], name); in btrfs_del_inode_ref() 218 item_size = btrfs_item_size(leaf, path->slots[0]); in btrfs_del_inode_ref() 229 item_start = btrfs_item_ptr_offset(leaf, path->slots[0]); in btrfs_del_inode_ref() 280 path->slots[0], in btrfs_insert_inode_extref() 292 ptr = (unsigned long)btrfs_item_ptr(leaf, path->slots[0], char); in btrfs_insert_inode_extref() 293 ptr += btrfs_item_size(leaf, path->slots[ in btrfs_insert_inode_extref() [all...] |
/kernel/linux/linux-5.10/arch/ia64/scripts/ |
H A D | unwcheck.py | 27 def check_func (func, slots, rlen_sum): 28 if slots != rlen_sum: 32 print("ERROR: %s: %lu slots, total region length = %lu" % (func, slots, rlen_sum)) 38 slots = 0 variable 43 check_func(func, slots, rlen_sum) 48 slots = 3 * (end - start) / 16 55 check_func(func, slots, rlen_sum)
|
/kernel/linux/linux-6.6/arch/ia64/scripts/ |
H A D | unwcheck.py | 27 def check_func (func, slots, rlen_sum): 28 if slots != rlen_sum: 32 print("ERROR: %s: %lu slots, total region length = %lu" % (func, slots, rlen_sum)) 38 slots = 0 variable 43 check_func(func, slots, rlen_sum) 48 slots = 3 * (end - start) / 16 55 check_func(func, slots, rlen_sum)
|
/kernel/linux/linux-5.10/lib/ |
H A D | assoc_array.c | 52 ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */ in assoc_array_subtree_iterate() 81 ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */ in assoc_array_subtree_iterate() 209 ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */ in assoc_array_walk() 324 ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */ in assoc_array_find() 382 struct assoc_array_ptr *ptr = node->slots[slot]; in assoc_array_destroy_subtree() 462 edit->leaf_p = &new_n0->slots[0]; in assoc_array_insert_in_empty_tree() 505 ptr = node->slots[i]; in assoc_array_insert_into_terminal_node() 514 edit->leaf_p = &node->slots[i]; in assoc_array_insert_into_terminal_node() 515 edit->dead_leaf = node->slots[i]; in assoc_array_insert_into_terminal_node() 526 edit->leaf_p = &node->slots[free_slo in assoc_array_insert_into_terminal_node() [all...] |
/kernel/linux/linux-6.6/lib/ |
H A D | assoc_array.c | 52 ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */ in assoc_array_subtree_iterate() 81 ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */ in assoc_array_subtree_iterate() 209 ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */ in assoc_array_walk() 324 ptr = READ_ONCE(node->slots[slot]); /* Address dependency. */ in assoc_array_find() 382 struct assoc_array_ptr *ptr = node->slots[slot]; in assoc_array_destroy_subtree() 462 edit->leaf_p = &new_n0->slots[0]; in assoc_array_insert_in_empty_tree() 505 ptr = node->slots[i]; in assoc_array_insert_into_terminal_node() 514 edit->leaf_p = &node->slots[i]; in assoc_array_insert_into_terminal_node() 515 edit->dead_leaf = node->slots[i]; in assoc_array_insert_into_terminal_node() 526 edit->leaf_p = &node->slots[free_slo in assoc_array_insert_into_terminal_node() [all...] |
H A D | maple_tree.c | 12 * Each node type has a number of slots for entries and a number of slots for 23 * The following illustrates the layout of a range64 nodes slots and pivots. 777 * ma_slots() - Get a pointer to the maple node slots. 781 * Return: A pointer to the maple node slots 810 void __rcu **slots, unsigned char offset) in mt_slot() 812 return rcu_dereference_check(slots[offset], mt_locked(mt)); in mt_slot() 815 static inline void *mt_slot_locked(struct maple_tree *mt, void __rcu **slots, in mt_slot_locked() argument 818 return rcu_dereference_protected(slots[offset], mt_write_locked(mt)); in mt_slot_locked() 823 * @slots 809 mt_slot(const struct maple_tree *mt, void __rcu **slots, unsigned char offset) mt_slot() argument 828 mas_slot_locked(struct ma_state *mas, void __rcu **slots, unsigned char offset) mas_slot_locked() argument 842 mas_slot(struct ma_state *mas, void __rcu **slots, unsigned char offset) mas_slot() argument 915 void __rcu **slots; mt_clear_meta() local 1037 void __rcu **slots; mas_descend() local 1251 void **slots = NULL; mas_alloc_nodes() local 1498 void __rcu **slots; mas_leaf_max_gap() local 1709 void __rcu **slots = ma_slots(node, type); mas_adopt_children() local 1732 void __rcu **slots; global() variable 1778 void __rcu **slots; global() variable 1948 void __rcu **slots; mas_mab_cp() local 2029 void __rcu **slots = ma_slots(node, mt); mab_mas_cp() local 2790 void __rcu **slots; mtree_range_walk() local 3055 void __rcu **l_slots, **slots; mas_destroy_rebalance() local 3511 void __rcu **slots; mas_root_expand() local 3731 void __rcu **slots; mtree_lookup_walk() local 3781 void __rcu **slots; mas_new_root() local 4014 void __rcu **slots = wr_mas->slots; mas_wr_slot_store() local 4137 void __rcu **slots; mas_wr_append() local 4363 void __rcu **slots; mas_prev_node() local 4443 void __rcu **slots; mas_prev_slot() local 4528 void __rcu **slots; mas_next_node() local 4606 void __rcu **slots; mas_next_slot() local 4727 void __rcu **slots; mas_rev_awalk() local 4814 void __rcu **slots; mas_anode_descend() local 5140 mte_dead_leaves(struct maple_enode *enode, struct maple_tree *mt, void __rcu **slots) mte_dead_leaves() argument 5174 void __rcu **slots = NULL; mte_dead_walk() local 5197 void __rcu **slots; mt_free_walk() local 5240 void __rcu **slots = NULL; mte_destroy_descend() local 5268 void __rcu **slots; mt_destroy_walk() local 7020 void __rcu **slots; mas_validate_parent_slot() local 7052 void __rcu **slots = ma_slots(mte_to_node(mas->node), type); mas_validate_child_slot() local 7098 void __rcu **slots = ma_slots(mte_to_node(mas->node), type); mas_validate_limits() local 7165 void __rcu **slots; mt_validate_nulls() local [all...] |
/kernel/linux/linux-6.6/tools/testing/selftests/hid/tests/ |
H A D | test_multitouch.py | 153 def event(self, slots, global_data=None, contact_count=None, incr_scantime=True): 157 # make sure we have only the required number of available slots 158 slots = slots[: self.max_contacts] 163 global_data.contactcount = len(slots) 168 while len(slots): 170 application=self.cur_application, data=slots, global_data=global_data 261 slots=None, 282 if slots is None: 283 slots [all...] |
/kernel/linux/linux-5.10/drivers/misc/mei/ |
H A D | interrupt.c | 226 int slots; in mei_cl_irq_disconnect_rsp() local 230 slots = mei_hbuf_empty_slots(dev); in mei_cl_irq_disconnect_rsp() 231 if (slots < 0) in mei_cl_irq_disconnect_rsp() 234 if ((u32)slots < msg_slots) in mei_cl_irq_disconnect_rsp() 258 int slots; in mei_cl_irq_read() local 265 slots = mei_hbuf_empty_slots(dev); in mei_cl_irq_read() 266 if (slots < 0) in mei_cl_irq_read() 269 if ((u32)slots < msg_slots) in mei_cl_irq_read() 323 * @slots: slots t 327 mei_irq_read_handler(struct mei_device *dev, struct list_head *cmpl_list, s32 *slots) mei_irq_read_handler() argument 483 s32 slots; mei_irq_write_handler() local [all...] |
H A D | dma-ring.c | 125 * @offset: offset in slots. 126 * @n: number of slots to copy. 145 * @offset: offset in slots. 146 * @n: number of slots to copy. 171 u32 rd_idx, rem, slots; in mei_dma_ring_read() local 183 slots = mei_data2slots(len); in mei_dma_ring_read() 189 if (rd_idx + slots > dbuf_depth) { in mei_dma_ring_read() 191 rem = slots - (dbuf_depth - rd_idx); in mei_dma_ring_read() 194 rem = slots; in mei_dma_ring_read() 199 WRITE_ONCE(ctrl->dbuf_rd_idx, ctrl->dbuf_rd_idx + slots); in mei_dma_ring_read() 248 u32 wr_idx, rem, slots; mei_dma_ring_write() local [all...] |
/kernel/linux/linux-6.6/drivers/misc/mei/ |
H A D | interrupt.c | 259 int slots; in mei_cl_irq_disconnect_rsp() local 263 slots = mei_hbuf_empty_slots(dev); in mei_cl_irq_disconnect_rsp() 264 if (slots < 0) in mei_cl_irq_disconnect_rsp() 267 if ((u32)slots < msg_slots) in mei_cl_irq_disconnect_rsp() 291 int slots; in mei_cl_irq_read() local 298 slots = mei_hbuf_empty_slots(dev); in mei_cl_irq_read() 299 if (slots < 0) in mei_cl_irq_read() 302 if ((u32)slots < msg_slots) in mei_cl_irq_read() 356 * @slots: slots t 360 mei_irq_read_handler(struct mei_device *dev, struct list_head *cmpl_list, s32 *slots) mei_irq_read_handler() argument 513 s32 slots; mei_irq_write_handler() local [all...] |
H A D | dma-ring.c | 125 * @offset: offset in slots. 126 * @n: number of slots to copy. 145 * @offset: offset in slots. 146 * @n: number of slots to copy. 171 u32 rd_idx, rem, slots; in mei_dma_ring_read() local 183 slots = mei_data2slots(len); in mei_dma_ring_read() 189 if (rd_idx + slots > dbuf_depth) { in mei_dma_ring_read() 191 rem = slots - (dbuf_depth - rd_idx); in mei_dma_ring_read() 194 rem = slots; in mei_dma_ring_read() 199 WRITE_ONCE(ctrl->dbuf_rd_idx, ctrl->dbuf_rd_idx + slots); in mei_dma_ring_read() 248 u32 wr_idx, rem, slots; mei_dma_ring_write() local [all...] |
/kernel/linux/linux-6.6/tools/testing/selftests/kvm/include/ |
H A D | memstress.h | 59 uint64_t vcpu_memory_bytes, int slots, 75 void memstress_enable_dirty_logging(struct kvm_vm *vm, int slots); 76 void memstress_disable_dirty_logging(struct kvm_vm *vm, int slots); 77 void memstress_get_dirty_log(struct kvm_vm *vm, unsigned long *bitmaps[], int slots); 79 int slots, uint64_t pages_per_slot); 80 unsigned long **memstress_alloc_bitmaps(int slots, uint64_t pages_per_slot); 81 void memstress_free_bitmaps(unsigned long *bitmaps[], int slots);
|
/kernel/linux/linux-5.10/arch/arm64/kernel/ |
H A D | hw_breakpoint.c | 178 * @slots: pointer to array of slots 179 * @max_slots: max number of slots 188 static int hw_breakpoint_slot_setup(struct perf_event **slots, int max_slots, in hw_breakpoint_slot_setup() argument 196 slot = &slots[i]; in hw_breakpoint_slot_setup() 226 struct perf_event **slots; in hw_breakpoint_control() local 236 slots = this_cpu_ptr(bp_on_reg); in hw_breakpoint_control() 243 slots = this_cpu_ptr(wp_on_reg); in hw_breakpoint_control() 248 i = hw_breakpoint_slot_setup(slots, max_slots, bp, ops); in hw_breakpoint_control() 584 struct perf_event **slots; in toggle_bp_registers() local 626 struct perf_event *bp, **slots; breakpoint_handler() local 761 struct perf_event *wp, **slots; watchpoint_handler() local 944 struct perf_event **slots; hw_breakpoint_reset() local [all...] |
/kernel/linux/linux-6.6/arch/arm64/kernel/ |
H A D | hw_breakpoint.c | 178 * @slots: pointer to array of slots 179 * @max_slots: max number of slots 188 static int hw_breakpoint_slot_setup(struct perf_event **slots, int max_slots, in hw_breakpoint_slot_setup() argument 196 slot = &slots[i]; in hw_breakpoint_slot_setup() 226 struct perf_event **slots; in hw_breakpoint_control() local 236 slots = this_cpu_ptr(bp_on_reg); in hw_breakpoint_control() 243 slots = this_cpu_ptr(wp_on_reg); in hw_breakpoint_control() 248 i = hw_breakpoint_slot_setup(slots, max_slots, bp, ops); in hw_breakpoint_control() 584 struct perf_event **slots; in toggle_bp_registers() local 626 struct perf_event *bp, **slots; breakpoint_handler() local 761 struct perf_event *wp, **slots; watchpoint_handler() local 944 struct perf_event **slots; hw_breakpoint_reset() local [all...] |
/kernel/linux/linux-5.10/fs/btrfs/ |
H A D | file-item.c | 152 item = btrfs_item_ptr(leaf, path->slots[0], in btrfs_insert_file_extent() 196 if (path->slots[0] == 0) in btrfs_lookup_csum() 198 path->slots[0]--; in btrfs_lookup_csum() 199 btrfs_item_key_to_cpu(leaf, &found_key, path->slots[0]); in btrfs_lookup_csum() 205 csums_in_item = btrfs_item_size_nr(leaf, path->slots[0]); in btrfs_lookup_csum() 215 item = btrfs_item_ptr(leaf, path->slots[0], struct btrfs_csum_item); in btrfs_lookup_csum() 353 path->slots[0]); in btrfs_lookup_bio_sums() 357 path->slots[0]); in btrfs_lookup_bio_sums() 361 item = btrfs_item_ptr(path->nodes[0], path->slots[0], in btrfs_lookup_bio_sums() 431 if (ret > 0 && path->slots[ in btrfs_lookup_csums_range() [all...] |