/foundation/communication/dsoftbus/core/transmission/trans_channel/manager/src/ |
H A D | trans_lane_manager.c | 257 TransLaneInfo *next = NULL; in TransLaneMgrDelLane() local 261 LIST_FOR_EACH_ENTRY_SAFE(laneItem, next, &(g_channelLaneList->list), TransLaneInfo, node) { in TransLaneMgrDelLane() 305 TransLaneInfo *next = NULL; in TransLaneMgrDeathCallback() local 306 LIST_FOR_EACH_ENTRY_SAFE(laneItem, next, &(g_channelLaneList->list), TransLaneInfo, node) { in TransLaneMgrDeathCallback() 547 SocketWithChannelInfo *next = NULL; in TransDeleteSocketChannelInfoBySession() local 548 LIST_FOR_EACH_ENTRY_SAFE(socketItem, next, &(g_socketChannelList->list), SocketWithChannelInfo, node) { in TransDeleteSocketChannelInfoBySession() 578 SocketWithChannelInfo *next = NULL; in TransDeleteSocketChannelInfoByChannel() local 579 LIST_FOR_EACH_ENTRY_SAFE(socketItem, next, &(g_socketChannelList->list), SocketWithChannelInfo, node) { in TransDeleteSocketChannelInfoByChannel() 608 SocketWithChannelInfo *next = NULL; in TransDeleteSocketChannelInfoByPid() local 609 LIST_FOR_EACH_ENTRY_SAFE(socketItem, next, in TransDeleteSocketChannelInfoByPid() [all...] |
/kernel/linux/linux-5.10/arch/powerpc/platforms/powernv/ |
H A D | opal.c | 734 /* HMI exception handler called in virtual mode when irqs are next enabled. */ 1159 struct opal_sg_list *next; in opal_vmalloc_to_sg_list() local 1161 next = kzalloc(PAGE_SIZE, GFP_KERNEL); in opal_vmalloc_to_sg_list() 1162 if (!next) in opal_vmalloc_to_sg_list() 1168 sg->next = cpu_to_be64(__pa(next)); in opal_vmalloc_to_sg_list() 1169 sg = next; in opal_vmalloc_to_sg_list() 1189 uint64_t next = be64_to_cpu(sg->next); in opal_free_sg_list() local 1193 if (next) in opal_free_sg_list() [all...] |
/kernel/linux/linux-6.6/arch/powerpc/platforms/powernv/ |
H A D | opal.c | 734 /* HMI exception handler called in virtual mode when irqs are next enabled. */ 1161 struct opal_sg_list *next; in opal_vmalloc_to_sg_list() local 1163 next = kzalloc(PAGE_SIZE, GFP_KERNEL); in opal_vmalloc_to_sg_list() 1164 if (!next) in opal_vmalloc_to_sg_list() 1170 sg->next = cpu_to_be64(__pa(next)); in opal_vmalloc_to_sg_list() 1171 sg = next; in opal_vmalloc_to_sg_list() 1191 uint64_t next = be64_to_cpu(sg->next); in opal_free_sg_list() local 1195 if (next) in opal_free_sg_list() [all...] |
/kernel/linux/linux-5.10/drivers/media/mc/ |
H A D | mc-entity.c | 249 graph->stack[graph->top].link = entity->links.next; in stack_push() 312 struct media_entity *next; in media_graph_walk_iter() local 318 link_top(graph) = link_top(graph)->next; in media_graph_walk_iter() 327 next = media_entity_other(entity, link); in media_graph_walk_iter() 330 if (media_entity_enum_test_and_set(&graph->ent_enum, next)) { in media_graph_walk_iter() 331 link_top(graph) = link_top(graph)->next; in media_graph_walk_iter() 334 next->name); in media_graph_walk_iter() 339 link_top(graph) = link_top(graph)->next; in media_graph_walk_iter() 340 stack_push(graph, next); in media_graph_walk_iter() 342 next in media_graph_walk_iter() [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/core/ |
H A D | rw.c | 80 reg->inv_wr.next = ®->reg_wr.wr; in rdma_rw_inv_key() 83 reg->inv_wr.next = NULL; in rdma_rw_inv_key() 151 prev->wr.wr.next = ®->inv_wr; in rdma_rw_init_mr_wrs() 153 prev->wr.wr.next = ®->reg_wr.wr; in rdma_rw_init_mr_wrs() 156 reg->reg_wr.wr.next = ®->wr.wr; in rdma_rw_init_mr_wrs() 181 prev->wr.wr.next = NULL; in rdma_rw_init_mr_wrs() 237 rdma_wr->wr.next = i + 1 < ctx->nr_ops ? in rdma_rw_init_map_wrs() 461 ctx->reg->reg_wr.wr.next = &rdma_wr->wr; in rdma_rw_ctx_signature_init() 522 if (ctx->reg[0].inv_wr.next) in rdma_rw_ctx_wrs() 541 last_wr->next in rdma_rw_ctx_wrs() [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/chelsio/cxgb4/ |
H A D | l2t.c | 61 struct l2t_entry *rover; /* starting point for next allocation */ 287 for (p = &d->l2tab[e->hash].first; *p; p = &(*p)->next) in alloc_l2e() 289 *p = e->next; in alloc_l2e() 290 e->next = NULL; in alloc_l2e() 329 for (p = &d->l2tab[e->hash].first; *p; p = &(*p)->next) in find_or_alloc_l2e() 331 *p = e->next; in find_or_alloc_l2e() 332 e->next = NULL; in find_or_alloc_l2e() 394 * Update an L2T entry that was previously used for the same next hop as neigh. 440 for (e = d->l2tab[hash].first; e; e = e->next) in cxgb4_l2t_get() 464 e->next in cxgb4_l2t_get() [all...] |
/kernel/linux/linux-5.10/fs/afs/ |
H A D | flock.c | 113 struct file_lock *p, *_p, *next = NULL; in afs_next_locker() local 131 /* Select the next locker to hand off to. */ in afs_next_locker() 132 if (next && in afs_next_locker() 133 (next->fl_type == F_WRLCK || p->fl_type == F_RDLCK)) in afs_next_locker() 135 next = p; in afs_next_locker() 141 if (next) { in afs_next_locker() 143 next->fl_u.afs.state = AFS_LOCK_YOUR_TRY; in afs_next_locker() 144 trace_afs_flock_op(vnode, next, afs_flock_op_wake); in afs_next_locker() 145 wake_up(&next->fl_wait); in afs_next_locker() 165 p = list_entry(vnode->pending_locks.next, in afs_kill_lockers_enoent() [all...] |
/kernel/linux/linux-5.10/fs/dlm/ |
H A D | debug_fs.c | 468 * move to the first rsb in the next non-empty bucket in table_seq_start() 504 struct rb_node *next; in table_seq_next() local 513 * move to the next rsb in the same bucket in table_seq_next() 518 next = rb_next(&rp->res_hashnode); in table_seq_next() 520 if (next) { in table_seq_next() 521 r = rb_entry(next, struct dlm_rsb, res_hashnode); in table_seq_next() 533 * move to the first rsb in the next non-empty bucket in table_seq_next() 552 next = rb_first(tree); in table_seq_next() 553 r = rb_entry(next, struct dlm_rsb, res_hashnode); in table_seq_next() 577 .next [all...] |
/kernel/linux/linux-5.10/kernel/bpf/ |
H A D | devmap.c | 227 struct hlist_node *next; in dev_map_free() local 231 hlist_for_each_entry_safe(dev, next, head, index_hlist) { in dev_map_free() 265 u32 *next = next_key; in dev_map_get_next_key() local 268 *next = 0; in dev_map_get_next_key() 274 *next = index + 1; in dev_map_get_next_key() 296 u32 idx, *next = next_key; in dev_map_hash_get_next_key() local 314 *next = next_dev->idx; in dev_map_hash_get_next_key() 329 *next = next_dev->idx; in dev_map_hash_get_next_key() 792 struct hlist_node *next; in dev_map_hash_remove_netdev() local 796 hlist_for_each_entry_safe(dev, next, hea in dev_map_hash_remove_netdev() [all...] |
/kernel/linux/linux-5.10/fs/ |
H A D | seq_file.c | 48 * element of sequence. @op->stop() shuts it down. @op->next() 49 * returns the next element of sequence. @op->show() prints element 50 * into the buffer. In case of error ->start() and ->next() return 120 p = m->op->next(m, p, &m->index); in traverse() 236 p = m->op->next(m, p, &m->index); in seq_read_iter() 262 p = m->op->next(m, p, &m->index); in seq_read_iter() 264 pr_info_ratelimited("buggy .next function %ps did not update position index\n", in seq_read_iter() 265 m->op->next); in seq_read_iter() 268 if (!p || IS_ERR(p)) // no next record for us in seq_read_iter() 569 op->next in single_open() [all...] |
/kernel/linux/linux-5.10/drivers/staging/android/ |
H A D | ashmem.c | 319 struct ashmem_range *range, *next; in ashmem_release() local 322 list_for_each_entry_safe(range, next, &asma->unpinned_list, unpinned) in ashmem_release() 678 struct ashmem_range *range, *next; in ashmem_pin() local 687 list_for_each_entry_safe(range, next, &asma->unpinned_list, unpinned) { in ashmem_pin() 752 struct ashmem_range *range, *next; in ashmem_unpin() local 766 list_for_each_entry_safe(range, next, &asma->unpinned_list, unpinned) { in ashmem_unpin() 897 struct ashmem_range *range, *next; in ashmem_shrink_by_id() local 903 list_for_each_entry_safe(range, next, &ashmem_lru_list, lru) { in ashmem_shrink_by_id() 941 struct ashmem_range *range, *next; in is_ashmem_unpin() local 949 list_for_each_entry_safe(range, next, in is_ashmem_unpin() [all...] |
/kernel/linux/linux-6.6/fs/afs/ |
H A D | flock.c | 113 struct file_lock *p, *_p, *next = NULL; in afs_next_locker() local 131 /* Select the next locker to hand off to. */ in afs_next_locker() 132 if (next && in afs_next_locker() 133 (next->fl_type == F_WRLCK || p->fl_type == F_RDLCK)) in afs_next_locker() 135 next = p; in afs_next_locker() 141 if (next) { in afs_next_locker() 143 next->fl_u.afs.state = AFS_LOCK_YOUR_TRY; in afs_next_locker() 144 trace_afs_flock_op(vnode, next, afs_flock_op_wake); in afs_next_locker() 145 wake_up(&next->fl_wait); in afs_next_locker() 165 p = list_entry(vnode->pending_locks.next, in afs_kill_lockers_enoent() [all...] |
/kernel/linux/linux-6.6/fs/nfs/ |
H A D | direct.c | 278 struct nfs_page *req = nfs_list_entry(hdr->pages.next); in nfs_direct_read_completion() 299 req = nfs_list_entry(head->next); in nfs_read_sync_pgio_error() 561 req = nfs_list_entry(reqs.next); in nfs_direct_write_reschedule() 583 req = nfs_list_entry(reqs.next); in nfs_direct_write_reschedule() 621 req = nfs_list_entry(data->pages.next); in nfs_direct_commit_complete() 694 req = nfs_list_entry(reqs.next); in nfs_direct_write_clear_reqs() 732 struct nfs_page *req = nfs_list_entry(hdr->pages.next); in nfs_direct_write_completion() 756 req = nfs_list_entry(hdr->pages.next); in nfs_direct_write_completion() 782 req = nfs_list_entry(head->next); in nfs_write_sync_pgio_error() 803 req = nfs_list_entry(hdr->pages.next); in nfs_direct_write_reschedule_io() [all...] |
/kernel/linux/linux-6.6/fs/btrfs/ |
H A D | delayed-ref.c | 369 * is given, the next bigger entry is returned if no exact match is found. 443 struct btrfs_delayed_ref_node *next; in merge_ref() local 450 next = rb_entry(node, struct btrfs_delayed_ref_node, ref_node); in merge_ref() 452 if (seq && next->seq >= seq) in merge_ref() 454 if (comp_refs(ref, next, false)) in merge_ref() 457 if (ref->action == next->action) { in merge_ref() 458 mod = next->ref_mod; in merge_ref() 460 if (ref->ref_mod < next->ref_mod) { in merge_ref() 461 swap(ref, next); in merge_ref() 464 mod = -next in merge_ref() [all...] |
/kernel/linux/linux-6.6/fs/ |
H A D | seq_file.c | 48 * element of sequence. @op->stop() shuts it down. @op->next() 49 * returns the next element of sequence. @op->show() prints element 50 * into the buffer. In case of error ->start() and ->next() return 120 p = m->op->next(m, p, &m->index); in traverse() 236 p = m->op->next(m, p, &m->index); in seq_read_iter() 262 p = m->op->next(m, p, &m->index); in seq_read_iter() 264 pr_info_ratelimited("buggy .next function %ps did not update position index\n", in seq_read_iter() 265 m->op->next); in seq_read_iter() 268 if (!p || IS_ERR(p)) // no next record for us in seq_read_iter() 580 op->next in single_open() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/chelsio/cxgb4/ |
H A D | l2t.c | 61 struct l2t_entry *rover; /* starting point for next allocation */ 287 for (p = &d->l2tab[e->hash].first; *p; p = &(*p)->next) in alloc_l2e() 289 *p = e->next; in alloc_l2e() 290 e->next = NULL; in alloc_l2e() 329 for (p = &d->l2tab[e->hash].first; *p; p = &(*p)->next) in find_or_alloc_l2e() 331 *p = e->next; in find_or_alloc_l2e() 332 e->next = NULL; in find_or_alloc_l2e() 394 * Update an L2T entry that was previously used for the same next hop as neigh. 440 for (e = d->l2tab[hash].first; e; e = e->next) in cxgb4_l2t_get() 464 e->next in cxgb4_l2t_get() [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/core/ |
H A D | rw.c | 81 reg->inv_wr.next = ®->reg_wr.wr; in rdma_rw_inv_key() 84 reg->inv_wr.next = NULL; in rdma_rw_inv_key() 152 prev->wr.wr.next = ®->inv_wr; in rdma_rw_init_mr_wrs() 154 prev->wr.wr.next = ®->reg_wr.wr; in rdma_rw_init_mr_wrs() 157 reg->reg_wr.wr.next = ®->wr.wr; in rdma_rw_init_mr_wrs() 182 prev->wr.wr.next = NULL; in rdma_rw_init_mr_wrs() 238 rdma_wr->wr.next = i + 1 < ctx->nr_ops ? in rdma_rw_init_map_wrs() 451 ctx->reg->reg_wr.wr.next = &rdma_wr->wr; in rdma_rw_ctx_signature_init() 512 if (ctx->reg[0].inv_wr.next) in rdma_rw_ctx_wrs() 531 last_wr->next in rdma_rw_ctx_wrs() [all...] |
/kernel/linux/linux-6.6/scripts/kconfig/ |
H A D | confdata.c | 896 else if (menu->next != NULL) { in conf_write_defconfig() 897 menu = menu->next; in conf_write_defconfig() 900 if (menu->next != NULL) { in conf_write_defconfig() 901 menu = menu->next; in conf_write_defconfig() 960 goto next; in conf_write() 971 goto next; in conf_write() 980 next: in conf_write() 993 if (menu->next) { in conf_write() 994 menu = menu->next; in conf_write() 1053 for (file = file_list; file; file = file->next) in conf_write_autoconf_cmd() [all...] |
/third_party/ffmpeg/libavformat/ |
H A D | rtpdec.c | 144 * @return the next registered rtp dynamic protocol handler 455 pkt = pkt->next; in find_missing_packets() 783 RTPPacket *next = s->queue->next; in ff_rtp_reset_packet_queue() local 786 s->queue = next; in ff_rtp_reset_packet_queue() 803 cur = &(*cur)->next; in enqueue_packet() 813 packet->next = *cur; in enqueue_packet() 833 RTPPacket *next; in rtp_parse_queued_packet() local 849 next = s->queue->next; in rtp_parse_queued_packet() [all...] |
/third_party/curl/lib/ |
H A D | socks.c | 126 nread = Curl_conn_cf_recv(cf->next, data, buf, buffersize, &err); in Curl_blockread_all() 219 nwritten = Curl_conn_cf_send(cf->next, data, (char *)sx->outp, in socks_state_send() 250 nread = Curl_conn_cf_recv(cf->next, data, (char *)sx->outp, in socks_state_recv() 1123 the next magic steps. If 'done' isn't set TRUE, it is not done yet and 1143 result = cf->next->cft->do_connect(cf->next, data, blocking, done); in socks_proxy_cf_connect() 1215 DEBUGASSERT(cf->next); in socks_proxy_cf_close() 1218 cf->next->cft->do_close(cf->next, data); in socks_proxy_cf_close() 1241 cf->next in socks_cf_get_host() [all...] |
/third_party/FreeBSD/contrib/gdtoa/ |
H A D | misc.c | 61 freelist[k] = rv->next; 102 v->next = freelist[v->k]; 399 p5->next = 0; 404 p5->next = 0; 415 if ((p51 = p5->next) == 0) { 418 if (!(p51 = p5->next)) { 419 p51 = p5->next = mult(p5,p5); 420 p51->next = 0; 424 p51 = p5->next = mult(p5,p5); 425 p51->next [all...] |
/third_party/FreeBSD/sbin/fsck_msdosfs/ |
H A D | dir.c | 117 freede = de->next; in newDosDirEntry() 124 de->next = freede; in freeDosDirEntry() 142 freedt = dt->next; in newDirTodo() 149 dt->next = freedt; in freeDirTodo() 272 np = p->next; in finishDosDirSection() 281 if (!(nd = d->next)) in finishDosDirSection() 794 dirent.next = dir->child; in readDosDirSection() 982 n->next = pendingDirectories; in readDosDirSection() 1045 struct dirTodoNode *n = pendingDirectories->next; in handleDirTree() 1086 for (lostDir = rootDir->child; lostDir; lostDir = lostDir->next) { in reconnect() [all...] |
/third_party/jinja2/ |
H A D | ext.py | 356 lineno = next(parser.stream).lineno 390 next(parser.stream) 426 next(parser.stream) 439 next(parser.stream) 442 next(parser.stream) 482 """Parse until the next block tag with a given name.""" 489 next(parser.stream) 491 next(parser.stream) 497 next(parser.stream) 597 node = nodes.ExprStmt(lineno=next(parse [all...] |
/third_party/node/deps/v8/src/objects/ |
H A D | code.cc | 151 for (RelocIterator it(*this, mode_mask); !it.done(); it.next()) { in ClearEmbeddedObjects() 159 for (RelocIterator it(*this, RelocInfo::kApplyMask); !it.done(); it.next()) { in Relocate() 191 for (RelocIterator it(*this, reloc_info, mode_mask); !it.done(); it.next()) { in RelocateFromDesc() 315 for (RelocIterator it(*this, kModeMask); !it.done(); it.next()) { in IsIsolateIndependent() 362 Object next; in Next() local 364 // Get next code in the linked list. in Next() 365 next = current_code_.next_code_link(); in Next() 367 // Linked list of code exhausted. Get list of next context. in Next() 368 next = next_context_.OptimizedCodeListHead(); in Next() 378 next in Next() [all...] |
/third_party/skia/src/gpu/ |
H A D | GrDrawOpAtlas.cpp | 98 , fGenID(fGenerationCounter->next()) in Plot() 198 fGenID = fGenerationCounter->next(); in resetRects() 229 , fAtlasGeneration(fGenerationCounter->next()) in GrDrawOpAtlas() 256 fAtlasGeneration = fGenerationCounter->next(); in processEviction() 293 for (Plot* plot = plotIter.get(); plot; plot = plotIter.next()) { in uploadToPage() 439 plotIter.next(); in compactRadicals() 479 plotIter.next(); in compact() 526 plotIter.next(); in compact() 563 plotIter.next(); in compact() 593 plotIter.next(); in compact() [all...] |