/kernel/linux/linux-5.10/lib/ |
H A D | ts_fsm.c | 137 struct ts_fsm_token *cur = NULL, *next; in fsm_find() local 169 next = &fsm->tokens[tok_idx + 1]; in fsm_find() 171 next = NULL; in fsm_find() 199 if (next == NULL) in fsm_find() 205 while (!match_token(next, data[block_idx])) { in fsm_find() 222 while (!match_token(next, data[block_idx])) { in fsm_find()
|
/kernel/linux/linux-5.10/lib/mpi/ |
H A D | mpih-mul.c | 399 if (!ctx->next) { in mpihelp_mul_karatsuba_case() 400 ctx->next = kzalloc(sizeof *ctx, GFP_KERNEL); in mpihelp_mul_karatsuba_case() 401 if (!ctx->next) in mpihelp_mul_karatsuba_case() 407 ctx->next) < 0) in mpihelp_mul_karatsuba_case() 426 for (ctx = ctx->next; ctx; ctx = ctx2) { in mpihelp_release_karatsuba_ctx() 427 ctx2 = ctx->next; in mpihelp_release_karatsuba_ctx()
|
/kernel/linux/linux-5.10/kernel/gcov/ |
H A D | fs.c | 106 /* seq_file.next() implementation for gcov data files. */ 135 .next = gcov_seq_next, 568 char *next; in add_node() local 577 for (curr = filename; (next = strchr(curr, '/')); curr = next + 1) { in add_node() 578 if (curr == next) in add_node() 580 *next = 0; in add_node()
|
/kernel/linux/linux-5.10/net/core/ |
H A D | flow_offload.c | 409 struct flow_block_cb *this, *next; in __flow_block_indr_cleanup() local 411 list_for_each_entry_safe(this, next, &flow_block_indr_list, indr.list) { in __flow_block_indr_cleanup() 420 struct flow_block_cb *this, *next; in flow_block_indr_notify() local 422 list_for_each_entry_safe(this, next, cleanup_list, indr.list) { in flow_block_indr_notify() 431 struct flow_indr_dev *this, *next, *indr_dev = NULL; in flow_indr_dev_unregister() local 435 list_for_each_entry_safe(this, next, &flow_block_indr_dev_list, list) { in flow_indr_dev_unregister()
|
/kernel/linux/linux-5.10/net/ipv6/ |
H A D | ip6_input.c | 85 struct sk_buff *skb, *next; in ip6_sublist_rcv_finish() local 87 list_for_each_entry_safe(skb, next, head, list) { in ip6_sublist_rcv_finish() 112 struct sk_buff *skb, *next, *hint = NULL; in ip6_list_rcv_finish() local 117 list_for_each_entry_safe(skb, next, head, list) { in ip6_list_rcv_finish() 320 struct sk_buff *skb, *next; in ipv6_list_rcv() local 324 list_for_each_entry_safe(skb, next, head, list) { in ipv6_list_rcv()
|
/kernel/linux/linux-5.10/net/netfilter/ipset/ |
H A D | ip_set_hash_netnet.c | 121 hash_netnet4_data_next(struct hash_netnet4_elem *next, in hash_netnet4_data_next() argument 124 next->ipcmp = d->ipcmp; in hash_netnet4_data_next() 248 ip = ntohl(h->next.ip[0]); in hash_netnet4_uadt() 249 ip2 = ntohl(h->next.ip[1]); in hash_netnet4_uadt() 261 hash_netnet4_data_next(&h->next, &e); in hash_netnet4_uadt() 357 hash_netnet6_data_next(struct hash_netnet6_elem *next, in hash_netnet6_data_next() argument
|
/kernel/linux/linux-5.10/net/netfilter/ipvs/ |
H A D | ip_vs_lblc.c | 233 struct hlist_node *next; in ip_vs_lblc_flush() local 239 hlist_for_each_entry_safe(en, next, &tbl->bucket[i], list) { in ip_vs_lblc_flush() 260 struct hlist_node *next; in ip_vs_lblc_full_check() local 268 hlist_for_each_entry_safe(en, next, &tbl->bucket[j], list) { in ip_vs_lblc_full_check() 302 struct hlist_node *next; in ip_vs_lblc_check_expire() local 324 hlist_for_each_entry_safe(en, next, &tbl->bucket[j], list) { in ip_vs_lblc_check_expire()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/ |
H A D | nv50.c | 204 if (likely(iobj->lru.next) && iobj->map) { in nv50_instobj_release() 248 if (likely(iobj->lru.next)) in nv50_instobj_acquire() 274 if (likely(iobj->lru.next)) { in nv50_instobj_boot() 276 iobj->lru.next = NULL; in nv50_instobj_boot() 302 iobj->lru.next = NULL; /* Exclude from eviction. */ in nv50_instobj_bar2() 324 if (likely(iobj->lru.next)) in nv50_instobj_dtor()
|
/kernel/linux/linux-5.10/drivers/media/test-drivers/vivid/ |
H A D | vivid-kthread-out.c | 63 vid_out_buf = list_entry(dev->vid_out_active.next, in vivid_thread_vid_out_tick() 70 vbi_out_buf = list_entry(dev->vbi_out_active.next, in vivid_thread_vid_out_tick() 75 meta_out_buf = list_entry(dev->meta_out_active.next, in vivid_thread_vid_out_tick() 225 * Calculate when that next buffer is supposed to start in vivid_thread_vid_out() 304 buf = list_entry(dev->vid_out_active.next, in vivid_stop_generating_vid_out() 319 buf = list_entry(dev->vbi_out_active.next, in vivid_stop_generating_vid_out() 334 buf = list_entry(dev->meta_out_active.next, in vivid_stop_generating_vid_out()
|
/kernel/linux/linux-5.10/drivers/parport/ |
H A D | daisy.c | 34 struct daisydev *next; member 60 for (p = &topology; *p && (*p)->devnum<devnum; p = &(*p)->next) in add_dev() 62 newdev->next = *p; in add_dev() 212 p = &dev->next; in parport_daisy_fini() 215 *p = dev->next; in parport_daisy_fini() 252 p = p->next; in parport_open() 488 last_dev from next device or if last_dev does not in assign_addrs()
|
/kernel/linux/linux-6.6/drivers/mtd/ |
H A D | mtd_blkdevs.c | 454 struct mtd_blktrans_dev *dev, *next; in blktrans_notify_remove() local 457 list_for_each_entry_safe(dev, next, &tr->devs, list) in blktrans_notify_remove() 486 if (!blktrans_notifier.list.next) in register_mtd_blktrans() 514 struct mtd_blktrans_dev *dev, *next; in deregister_mtd_blktrans() local 521 list_for_each_entry_safe(dev, next, &tr->devs, list) in deregister_mtd_blktrans() 535 if (blktrans_notifier.list.next) in mtd_blktrans_exit()
|
/kernel/linux/linux-6.6/drivers/misc/mei/ |
H A D | interrupt.c | 31 struct mei_cl_cb *cb, *next; in mei_irq_compl_handler() local 34 list_for_each_entry_safe(cb, next, cmpl_list, list) { in mei_irq_compl_handler() 512 struct mei_cl_cb *cb, *next; in mei_irq_write_handler() local 530 list_for_each_entry_safe(cb, next, &dev->write_waiting_list, list) { in mei_irq_write_handler() 541 list_for_each_entry_safe(cb, next, &dev->ctrl_wr_list, list) { in mei_irq_write_handler() 595 list_for_each_entry_safe(cb, next, &dev->write_list, list) { in mei_irq_write_handler()
|
/kernel/linux/linux-6.6/drivers/media/test-drivers/vivid/ |
H A D | vivid-kthread-out.c | 64 vid_out_buf = list_entry(dev->vid_out_active.next, in vivid_thread_vid_out_tick() 71 vbi_out_buf = list_entry(dev->vbi_out_active.next, in vivid_thread_vid_out_tick() 76 meta_out_buf = list_entry(dev->meta_out_active.next, in vivid_thread_vid_out_tick() 227 * Calculate when that next buffer is supposed to start in vivid_thread_vid_out() 308 buf = list_entry(dev->vid_out_active.next, in vivid_stop_generating_vid_out() 323 buf = list_entry(dev->vbi_out_active.next, in vivid_stop_generating_vid_out() 338 buf = list_entry(dev->meta_out_active.next, in vivid_stop_generating_vid_out()
|
/kernel/linux/linux-6.6/fs/netfs/ |
H A D | io.c | 172 struct netfs_io_subrequest *subreq, *next, *p; in netfs_rreq_do_write_to_cache() local 194 next = list_next_entry(subreq, rreq_link); in netfs_rreq_do_write_to_cache() 195 if (next->start != subreq->start + subreq->len) in netfs_rreq_do_write_to_cache() 197 subreq->len += next->len; in netfs_rreq_do_write_to_cache() 198 list_del_init(&next->rreq_link); in netfs_rreq_do_write_to_cache() 199 netfs_put_subrequest(next, false, in netfs_rreq_do_write_to_cache() 324 * Assess the state of a read request and decide what to do next. 500 * Work out what sort of subrequest the next one will be. 566 * again and ask it to download the next piece. in netfs_rreq_submit_slice()
|
/kernel/linux/linux-6.6/fs/exfat/ |
H A D | fatent.c | 241 unsigned int clu, next; in exfat_find_last_cluster() local 244 next = p_chain->dir; in exfat_find_last_cluster() 246 *ret_clu = next + p_chain->size - 1; in exfat_find_last_cluster() 252 clu = next; in exfat_find_last_cluster() 253 if (exfat_ent_get(sb, clu, &next)) in exfat_find_last_cluster() 255 } while (next != EXFAT_EOF_CLUSTER); in exfat_find_last_cluster()
|
/kernel/linux/linux-6.6/fs/verity/ |
H A D | enable.c | 21 /* Hash a block, writing the result to the next level's pending block buffer. */ 26 struct block_buffer *next = cur + 1; in hash_one_block() local 34 if (WARN_ON_ONCE(next->is_root_hash && next->filled != 0)) in hash_one_block() 41 &next->data[next->filled]); in hash_one_block() 44 next->filled += params->digest_size; in hash_one_block()
|
/kernel/linux/linux-6.6/drivers/block/aoe/ |
H A D | aoedev.c | 328 for (d = devlist; d; d = d->next) { in flush() 359 for (d = devlist; d; d = d->next) { in flush() 377 *dd = d->next; in flush() 380 dd = &d->next; in flush() 448 for (d=devlist; d; d=d->next) in aoedev_by_aoeaddr() 490 d->next = devlist; in aoedev_by_aoeaddr()
|
/kernel/linux/linux-6.6/drivers/char/agp/ |
H A D | isoch.c | 40 struct list_head *pos, *tmp, *head = &list->list, *start = head->next; in agp_3_5_dev_list_sort() 53 pos = pos->next; in agp_3_5_dev_list_sort() 299 for (pos=head->next; cdev<ndevs; cdev++, pos=pos->next) { in agp_3_5_nonisochronous_node_enable() 455 for (pos=head->next; pos!=head; ) { in agp_3_5_enable() 458 pos = pos->next; in agp_3_5_enable()
|
/kernel/linux/linux-6.6/drivers/media/pci/tw686x/ |
H A D | tw686x-audio.c | 40 struct tw686x_audio_buf *next = NULL; in tw686x_audio_irq() local 54 next = list_first_entry(&ac->buf_list, in tw686x_audio_irq() 56 list_move_tail(&next->list, &ac->buf_list); in tw686x_audio_irq() 58 ac->curr_bufs[pb] = next; in tw686x_audio_irq() 74 reg_write(dev, reg, next->dma); in tw686x_audio_irq() 284 ss; ss = ss->next, i++) in tw686x_snd_pcm_init()
|
/kernel/linux/linux-6.6/drivers/parport/ |
H A D | daisy.c | 34 struct daisydev *next; member 60 for (p = &topology; *p && (*p)->devnum<devnum; p = &(*p)->next) in add_dev() 62 newdev->next = *p; in add_dev() 212 p = &dev->next; in parport_daisy_fini() 215 *p = dev->next; in parport_daisy_fini() 252 p = p->next; in parport_open() 488 last_dev from next device or if last_dev does not in assign_addrs()
|
/kernel/linux/linux-6.6/block/ |
H A D | t10-pi.c | 70 goto next; in t10_pi_verify() 83 goto next; in t10_pi_verify() 96 next: in t10_pi_verify() 331 goto next; in ext_pi_crc64_verify() 343 goto next; in ext_pi_crc64_verify() 355 next: in ext_pi_crc64_verify()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/ |
H A D | nv50.c | 204 if (likely(iobj->lru.next) && iobj->map) { in nv50_instobj_release() 245 if (likely(iobj->lru.next)) in nv50_instobj_acquire() 269 if (likely(iobj->lru.next)) { in nv50_instobj_boot() 271 iobj->lru.next = NULL; in nv50_instobj_boot() 297 iobj->lru.next = NULL; /* Exclude from eviction. */ in nv50_instobj_bar2() 319 if (likely(iobj->lru.next)) in nv50_instobj_dtor()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | vmmnv50.c | 35 u64 next = addr + map->type, data; in nv50_vmm_pgt_pte() local 48 data = next | (log2blk << 7); in nv50_vmm_pgt_pte() 49 next += pten * map->next; in nv50_vmm_pgt_pte() 242 map->next = 1 << page->shift; in nv50_vmm_valid() 313 map->next |= map->ctag; in nv50_vmm_valid()
|
/kernel/linux/linux-6.6/drivers/hid/bpf/ |
H A D | hid_bpf_jmp_table.c | 235 struct hid_bpf_prog_entry *next; in hid_bpf_release_progs() local 237 next = &jmp_table.entries[j]; in hid_bpf_release_progs() 239 if (test_bit(next->idx, jmp_table.enabled)) in hid_bpf_release_progs() 242 if (next->hdev == hdev && next->type == type) { in hid_bpf_release_progs() 248 next->hdev = NULL; in hid_bpf_release_progs() 440 /* take the next prog_entry slot */ in __hid_bpf_attach_prog()
|
/kernel/linux/linux-6.6/drivers/infiniband/sw/rdmavt/ |
H A D | cq.c | 33 u32 next; in rvt_cq_enter() local 56 next = 0; in rvt_cq_enter() 58 next = head + 1; in rvt_cq_enter() 61 if (unlikely(next == tail || cq->cq_full)) { in rvt_cq_enter() 95 RDMA_WRITE_UAPI_ATOMIC(u_wc->head, next); in rvt_cq_enter() 98 k_wc->head = next; in rvt_cq_enter()
|