/kernel/linux/linux-6.6/kernel/futex/ |
H A D | waitwake.c | 146 struct futex_q *this, *next; in futex_wake() local 166 plist_for_each_entry_safe(this, next, &hb->chain, list) { in futex_wake() 243 struct futex_q *this, *next; in futex_wake_op() local 286 plist_for_each_entry_safe(this, next, &hb1->chain, list) { in futex_wake_op() 300 plist_for_each_entry_safe(this, next, &hb2->chain, list) { in futex_wake_op() 408 * each futex on the list before dealing with the next one to avoid in futex_wait_multiple_setup() 412 * of the next key, because it calls get_user_pages, which can sleep. in futex_wait_multiple_setup() 446 * next futex. Queue each futex at this moment so hb can in futex_wait_multiple_setup()
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/irdma/ |
H A D | main.h | 419 * @next: next free id 423 u32 *req_rsrc_num, u32 *next) in irdma_alloc_rsrc() 429 rsrc_num = find_next_zero_bit(rsrc_array, max_rsrc, *next); in irdma_alloc_rsrc() 441 *next = rsrc_num + 1; in irdma_alloc_rsrc() 442 if (*next == max_rsrc) in irdma_alloc_rsrc() 443 *next = 0; in irdma_alloc_rsrc() 421 irdma_alloc_rsrc(struct irdma_pci_f *rf, unsigned long *rsrc_array, u32 max_rsrc, u32 *req_rsrc_num, u32 *next) irdma_alloc_rsrc() argument
|
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | dev.c | 600 struct device *next; in mlx5_get_next_dev() local 605 next = bus_find_device(&pci_bus_type, NULL, dev, match); in mlx5_get_next_dev() 606 if (!next) in mlx5_get_next_dev() 609 put_device(next); in mlx5_get_next_dev() 610 return pci_get_drvdata(to_pci_dev(next)); in mlx5_get_next_dev()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/ |
H A D | drm_suballoc.c | 24 * next paragraph) shall be included in all copies or substantial portions 131 if (sa_manager->hole->next == &sa_manager->olist) in drm_suballoc_try_free() 134 sa = list_entry(sa_manager->hole->next, struct drm_suballoc, olist); in drm_suballoc_try_free() 157 if (hole->next != &sa_manager->olist) in drm_suballoc_hole_eoffset() 158 return list_entry(hole->next, struct drm_suballoc, olist)->soffset; in drm_suballoc_hole_eoffset() 233 if (sa_manager->hole->next == &sa_manager->olist) { in drm_suballoc_next_hole()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/i915/selftests/ |
H A D | i915_gem_evict.c | 11 * The above copyright notice and this permission notice (including the next 385 struct reserved *next; in igt_evict_contexts() member 439 r->next = reserved; in igt_evict_contexts() 522 struct reserved *next = reserved->next; in igt_evict_contexts() local 527 reserved = next; in igt_evict_contexts()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | vmmgf100.c | 38 if (map->ctag && !(map->next & (1ULL << 44))) { in gf100_vmm_pgt_pte() 45 base += map->next; in gf100_vmm_pgt_pte() 52 data += map->next; in gf100_vmm_pgt_pte() 253 map->next = (1 << page->shift) >> 8; in gf100_vmm_valid() 305 map->next |= 1ULL << 44; in gf100_vmm_valid()
|
/kernel/linux/linux-6.6/drivers/media/common/saa7146/ |
H A D | saa7146_vbi.c | 135 static void saa7146_set_vbi_capture(struct saa7146_dev *dev, struct saa7146_buf *buf, struct saa7146_buf *next) in saa7146_set_vbi_capture() argument 207 struct saa7146_buf *next) in buffer_activate() 211 DEB_VBI("dev:%p, buf:%p, next:%p\n", dev, buf, next); in buffer_activate() 212 saa7146_set_vbi_capture(dev,buf,next); in buffer_activate() 302 buf = list_entry(dq->queue.next, struct saa7146_buf, list); in return_buffers() 205 buffer_activate(struct saa7146_dev *dev, struct saa7146_buf *buf, struct saa7146_buf *next) buffer_activate() argument
|
/kernel/linux/linux-6.6/tools/net/ynl/generated/ |
H A D | fou-user.c | 289 struct fou_get_list *next = rsp; in fou_get_list_free() local 291 while ((void *)next != YNL_LIST_END) { in fou_get_list_free() 292 rsp = next; in fou_get_list_free() 293 next = rsp->next; in fou_get_list_free()
|
/kernel/linux/linux-6.6/net/sctp/ |
H A D | bind_addr.c | 286 goto next; in sctp_raw_to_bind_addrs() 293 next: in sctp_raw_to_bind_addrs() 349 goto next; in sctp_bind_addrs_check() 354 next: in sctp_bind_addrs_check() 553 addr = list_entry(bp->address_list.next, in sctp_is_ep_boundall()
|
/kernel/linux/linux-6.6/kernel/trace/ |
H A D | fgraph.c | 453 struct task_struct *next, in ftrace_graph_probe_sched_switch() 471 if (!next->ftrace_timestamp) in ftrace_graph_probe_sched_switch() 475 * Update all the counters in next to make up for the in ftrace_graph_probe_sched_switch() 476 * time next was sleeping. in ftrace_graph_probe_sched_switch() 478 timestamp -= next->ftrace_timestamp; in ftrace_graph_probe_sched_switch() 480 for (index = next->curr_ret_stack; index >= 0; index--) in ftrace_graph_probe_sched_switch() 481 next->ret_stack[index].calltime += timestamp; in ftrace_graph_probe_sched_switch() 451 ftrace_graph_probe_sched_switch(void *ignore, bool preempt, struct task_struct *prev, struct task_struct *next, unsigned int prev_state) ftrace_graph_probe_sched_switch() argument
|
H A D | ring_buffer.c | 330 local_t write; /* index for next write */ 331 unsigned read; /* index for next read */ 731 struct list_head *next; in verify_event() local 747 next = rb_list_head(page->list.next); in verify_event() 748 page = list_entry(next, struct buffer_page, list); in verify_event() 1006 * empty. The flag only causes the next event to run in ring_buffer_wait() 1114 * is set, the next event will wake the task up, but we can get stuck in ring_buffer_poll_wait() 1204 * ASCII art, the reader sets its old page to point to the next 1220 * head->list->prev->next bi [all...] |
/kernel/linux/linux-6.6/net/netfilter/ipset/ |
H A D | ip_set_hash_netiface.c | 129 hash_netiface4_data_next(struct hash_netiface4_elem *next, in hash_netiface4_data_next() argument 132 next->ip = d->ip; in hash_netiface4_data_next() 261 ip = ntohl(h->next.ip); in hash_netiface4_uadt() 266 hash_netiface4_data_next(&h->next, &e); in hash_netiface4_uadt() 363 hash_netiface6_data_next(struct hash_netiface6_elem *next, in hash_netiface6_data_next() argument
|
H A D | ip_set_hash_net.c | 102 hash_net4_data_next(struct hash_net4_elem *next, in hash_net4_data_next() argument 105 next->ip = d->ip; in hash_net4_data_next() 193 ip = ntohl(h->next.ip); in hash_net4_uadt() 198 hash_net4_data_next(&h->next, &e); in hash_net4_uadt() 273 hash_net6_data_next(struct hash_net6_elem *next, in hash_net6_data_next() argument
|
/foundation/filemanagement/storage_service/services/storage_daemon/mtpfs/src/ |
H A D | mtpfs_libmtp.cpp | 30 f = f->next; in LIBMTPFreeFilesAndFolders()
|
/kernel/linux/linux-5.10/arch/csky/include/asm/ |
H A D | spinlock_types.h | 18 u16 next; member
|
/kernel/linux/linux-5.10/arch/sparc/include/asm/ |
H A D | cpudata_32.h | 25 int next; member
|
/kernel/linux/linux-5.10/drivers/net/appletalk/ |
H A D | ipddp.h | 21 struct ipddp_route *next; member
|
/kernel/linux/linux-6.6/arch/sparc/include/asm/ |
H A D | cpudata_32.h | 25 int next; member
|
/kernel/linux/linux-5.10/include/linux/ |
H A D | if_tunnel.h | 15 for (pos = rcu_dereference(start); pos; pos = rcu_dereference(pos->next))
|
/kernel/linux/linux-6.6/drivers/net/appletalk/ |
H A D | ipddp.h | 21 struct ipddp_route *next; member
|
/kernel/linux/linux-6.6/include/linux/ |
H A D | if_tunnel.h | 15 for (pos = rcu_dereference(start); pos; pos = rcu_dereference(pos->next))
|
/kernel/linux/linux-6.6/kernel/sched/ |
H A D | rt.c | 159 rt_rq->highest_prio.next = MAX_RT_PRIO-1; in init_rt_rq() 454 if (p->prio < rq->rt.highest_prio.next) in enqueue_pushable_task() 455 rq->rt.highest_prio.next = p->prio; in enqueue_pushable_task() 466 rq->rt.highest_prio.next = p->prio; in dequeue_pushable_task() 468 rq->rt.highest_prio.next = MAX_RT_PRIO-1; in dequeue_pushable_task() 564 tg = list_entry_rcu(tg->list.next, in next_task_group() 768 goto next; in do_balance_runtime() 786 next: in do_balance_runtime() 1754 * - current will be preempted on the next reschedule in check_preempt_curr_rt() 1795 struct sched_rt_entity *next in pick_next_rt_entity() local 2425 int next; rto_next_cpu() local 2814 unsigned long next; watchdog() local [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/ulp/rtrs/ |
H A D | rtrs-srv.c | 286 wr->wr.next = &rwr.wr; in rdma_write_sg() 287 rwr.wr.next = &inv_wr; in rdma_write_sg() 288 inv_wr.next = &imm_wr.wr; in rdma_write_sg() 290 wr->wr.next = &rwr.wr; in rdma_write_sg() 291 rwr.wr.next = &imm_wr.wr; in rdma_write_sg() 293 wr->wr.next = &inv_wr; in rdma_write_sg() 294 inv_wr.next = &imm_wr.wr; in rdma_write_sg() 296 wr->wr.next = &imm_wr.wr; in rdma_write_sg() 314 imm_wr.wr.next = NULL; in rdma_write_sg() 414 inv_wr.next in send_io_resp_imm() [all...] |
/kernel/linux/linux-5.10/drivers/mmc/host/ |
H A D | mmci.c | 506 static int mmci_prep_data(struct mmci_host *host, struct mmc_data *data, bool next) in mmci_prep_data() argument 513 err = host->ops->prep_data(host, data, next); in mmci_prep_data() 515 if (next && !err) in mmci_prep_data() 568 * to fire next DMA request. When that happens, MMCI will in mmci_dma_start() 979 bool next) in mmci_dmae_prep_data() 987 if (next) in mmci_dmae_prep_data() 989 /* Check if next job is already prepared. */ in mmci_dmae_prep_data() 1019 struct mmci_dmae_next *next = &dmae->next_data; in mmci_dmae_get_next_data() local 1024 WARN_ON(!data->host_cookie && (next->desc || next in mmci_dmae_get_next_data() 977 mmci_dmae_prep_data(struct mmci_host *host, struct mmc_data *data, bool next) mmci_dmae_prep_data() argument 1044 struct mmci_dmae_next *next = &dmae->next_data; mmci_dmae_unprep_data() local [all...] |
/kernel/linux/linux-5.10/kernel/futex/ |
H A D | core.c | 853 struct list_head *next, *head = &curr->pi_state_list; in exit_pi_state_list() local 867 next = head->next; in exit_pi_state_list() 868 pi_state = list_entry(next, struct futex_pi_state, list); in exit_pi_state_list() 897 if (head->next != next) { in exit_pi_state_list() 1528 * We pass it to the next owner. The WAITERS bit is always kept in wake_futex_pi() 1603 struct futex_q *this, *next; in futex_wake() local 1623 plist_for_each_entry_safe(this, next, &hb->chain, list) { in futex_wake() 1701 struct futex_q *this, *next; in futex_wake_op() local 1938 struct futex_q *this, *next; futex_requeue() local [all...] |