/foundation/bundlemanager/bundle_framework/common/utils/src/ |
H A D | bundle_file_util.cpp | 180 struct dirent *entry = nullptr; in GetHapFilesFromBundlePath() local 181 while ((entry = readdir(dir)) != nullptr) { in GetHapFilesFromBundlePath() 182 if (strcmp(entry->d_name, ".") == 0 || strcmp(entry->d_name, "..") == 0) { in GetHapFilesFromBundlePath() 185 const std::string hapFilePath = bundlePath + entry->d_name; in GetHapFilesFromBundlePath()
|
/kernel/linux/linux-5.10/arch/loongarch/kernel/ |
H A D | machine_kexec.c | 262 unsigned long entry, *ptr; in machine_kexec() local 279 for (ptr = &image->head; (entry = *ptr) && !(entry & IND_DONE); in machine_kexec() 280 ptr = (entry & IND_INDIRECTION) ? in machine_kexec() 281 phys_to_virt(entry & PAGE_MASK) : ptr + 1) { in machine_kexec()
|
/kernel/linux/linux-5.10/arch/mips/mm/ |
H A D | tlb-r3k.c | 42 static void local_flush_tlb_from(int entry) in local_flush_tlb_from() argument 48 while (entry < current_cpu_data.tlbsize) { in local_flush_tlb_from() 49 write_c0_index(entry << 8); in local_flush_tlb_from() 50 write_c0_entryhi((entry | 0x80000) << 12); in local_flush_tlb_from() 51 entry++; /* BARRIER */ in local_flush_tlb_from() 232 printk("[tlbwired<entry lo0 %8x, hi %8x\n, pagemask %8x>]\n", in add_wired_entry() 256 printk("[tlbwired<entry lo0 %8x, hi %8x\n>]\n", in add_wired_entry()
|
/kernel/linux/linux-5.10/arch/x86/kernel/ |
H A D | idt.c | 150 /* Must be page-aligned because the real IDT is used in the cpu entry area */ 271 * entry area, so that the "sidt" instruction will not leak the in idt_map_in_cea() 286 void *entry; in idt_setup_apic_and_irq_gates() local 291 entry = irq_entries_start + 8 * (i - FIRST_EXTERNAL_VECTOR); in idt_setup_apic_and_irq_gates() 292 set_intr_gate(i, entry); in idt_setup_apic_and_irq_gates() 302 entry = spurious_entries_start + 8 * (i - FIRST_SYSTEM_VECTOR); in idt_setup_apic_and_irq_gates() 303 set_intr_gate(i, entry); in idt_setup_apic_and_irq_gates() 306 /* Map IDT into CPU entry area and reload it. */ in idt_setup_apic_and_irq_gates()
|
/kernel/linux/linux-6.6/arch/loongarch/kernel/ |
H A D | machine_kexec.c | 259 unsigned long entry, *ptr; in machine_kexec() local 276 for (ptr = &image->head; (entry = *ptr) && !(entry & IND_DONE); in machine_kexec() 277 ptr = (entry & IND_INDIRECTION) ? in machine_kexec() 278 phys_to_virt(entry & PAGE_MASK) : ptr + 1) { in machine_kexec()
|
/kernel/linux/linux-5.10/kernel/rcu/ |
H A D | srcutiny.c | 39 INIT_LIST_HEAD(&ssp->srcu_work.entry); in init_srcu_struct_fields() 169 else if (list_empty(&ssp->srcu_work.entry)) in srcu_gp_start_if_needed() 170 list_add(&ssp->srcu_work.entry, &srcu_boot_list); in srcu_gp_start_if_needed() 268 struct srcu_struct, srcu_work.entry); in srcu_init() 269 list_del_init(&ssp->srcu_work.entry); in srcu_init()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/via/ |
H A D | via_mm.c | 214 struct via_memblock *entry, *next; in via_reclaim_buffers_locked() local 231 list_for_each_entry_safe(entry, next, &file_priv->obj_list, in via_reclaim_buffers_locked() 233 list_del(&entry->owner_list); in via_reclaim_buffers_locked() 234 drm_mm_remove_node(&entry->mm_node); in via_reclaim_buffers_locked() 235 kfree(entry); in via_reclaim_buffers_locked()
|
/kernel/linux/linux-6.6/drivers/android/ |
H A D | binder_internal.h | 119 #define binder_for_each_debugfs_entry(entry) \ 120 for ((entry) = binder_debugfs_entries; \ 121 (entry)->name; \ 122 (entry)++) 144 * @entry: node enqueued on list 150 struct list_head entry; member 302 * @node_entry: list entry for node->refs list in target node 503 * @fixup_entry: list entry
|
/kernel/linux/linux-6.6/fs/xfs/ |
H A D | xfs_attr_inactive.c | 90 struct xfs_attr_leaf_entry *entry; in xfs_attr3_leaf_inactive() local 101 entry = xfs_attr3_leaf_entryp(leaf); in xfs_attr3_leaf_inactive() 102 for (i = 0; i < ichdr.count; entry++, i++) { in xfs_attr3_leaf_inactive() 105 if (!entry->nameidx || (entry->flags & XFS_ATTR_LOCAL)) in xfs_attr3_leaf_inactive()
|
/kernel/linux/linux-6.6/include/trace/ |
H A D | trace_custom_events.h | 126 struct trace_custom_event_raw_##call __maybe_unused *entry; \ 148 struct trace_custom_event_raw_##call *entry; \ 156 entry = trace_event_buffer_reserve(&fbuffer, trace_file, \ 157 sizeof(*entry) + __data_size); \ 159 if (!entry) \
|
/kernel/linux/linux-6.6/arch/x86/kernel/ |
H A D | idt.c | 165 /* Must be page-aligned because the real IDT is used in the cpu entry area */ 262 * entry area, so that the "sidt" instruction will not leak the in idt_map_in_cea() 277 void *entry; in idt_setup_apic_and_irq_gates() local 282 entry = irq_entries_start + IDT_ALIGN * (i - FIRST_EXTERNAL_VECTOR); in idt_setup_apic_and_irq_gates() 283 set_intr_gate(i, entry); in idt_setup_apic_and_irq_gates() 293 entry = spurious_entries_start + IDT_ALIGN * (i - FIRST_SYSTEM_VECTOR); in idt_setup_apic_and_irq_gates() 294 set_intr_gate(i, entry); in idt_setup_apic_and_irq_gates() 297 /* Map IDT into CPU entry area and reload it. */ in idt_setup_apic_and_irq_gates()
|
/kernel/linux/linux-6.6/tools/testing/selftests/bpf/map_tests/ |
H A D | map_in_map_batch_ops.c | 53 /* Add entry into this created map in create_inner_maps() 88 int inner_map_fd, entry, err; in validate_fetch_results() local 91 for (entry = 0; entry < max_entries_fetched; ++entry) { in validate_fetch_results() 92 outer_map_value = fetched_values[entry]; in validate_fetch_results()
|
/kernel/linux/linux-6.6/kernel/rcu/ |
H A D | srcutiny.c | 39 INIT_LIST_HEAD(&ssp->srcu_work.entry); in init_srcu_struct_fields() 169 else if (list_empty(&ssp->srcu_work.entry)) in srcu_gp_start_if_needed() 170 list_add(&ssp->srcu_work.entry, &srcu_boot_list); in srcu_gp_start_if_needed() 280 struct srcu_struct, srcu_work.entry); in srcu_init() 281 list_del_init(&ssp->srcu_work.entry); in srcu_init()
|
/kernel/linux/linux-6.6/net/ethtool/ |
H A D | tunnels.c | 87 struct nlattr *ports, *table, *entry; in ethnl_tunnel_info_fill_reply() local 138 entry = nla_nest_start(skb, ETHTOOL_A_TUNNEL_UDP_TABLE_ENTRY); in ethnl_tunnel_info_fill_reply() 139 if (!entry) in ethnl_tunnel_info_fill_reply() 148 nla_nest_end(skb, entry); in ethnl_tunnel_info_fill_reply() 157 nla_nest_cancel(skb, entry); in ethnl_tunnel_info_fill_reply()
|
/third_party/mesa3d/src/compiler/nir/ |
H A D | nir_lower_scratch.c | 162 struct set_entry *entry = _mesa_set_search(set, deref->var); in nir_lower_vars_to_scratch() local 163 if (!entry) in nir_lower_vars_to_scratch() 167 _mesa_set_remove(set, entry); in nir_lower_vars_to_scratch() 172 set_foreach(set, entry) { in nir_lower_vars_to_scratch() 173 nir_variable* var = (void*)entry->key; in nir_lower_vars_to_scratch()
|
/third_party/node/deps/v8/src/heap/cppgc/ |
H A D | object-allocator.cc | 191 // entry is checked. in RefillLinearAllocationBuffer() 209 const FreeList::Block entry = space.free_list().Allocate(size); in RefillLinearAllocationBufferFromFreeList() local 210 if (!entry.address) return false; in RefillLinearAllocationBufferFromFreeList() 213 auto& page = *NormalPage::From(BasePage::FromPayload(entry.address)); in RefillLinearAllocationBufferFromFreeList() 220 space, stats_collector_, static_cast<Address>(entry.address), entry.size); in RefillLinearAllocationBufferFromFreeList()
|
/third_party/node/deps/v8/src/utils/ |
H A D | identity-map.h | 23 T* entry; member 126 return {reinterpret_cast<V*>(raw.entry), raw.already_exists}; in FindOrInsert() 170 V* entry() const { in entry() function in v8::internal::IdentityMap::Iterator 174 V* operator*() { return entry(); } in operator *() 175 V* operator->() { return entry(); } in operator ->()
|
/third_party/skia/third_party/externals/sfntly/cpp/src/sfntly/table/bitmap/ |
H A D | ebdt_table.cc | 219 for (BitmapGlyphInfoMap::iterator entry = loca_map->begin(), in Initialize() 221 entry != entry_end; entry++) { in Initialize() 222 BitmapGlyphInfoPtr info = entry->second; in Initialize() 229 glyph_builder_map[entry->first] = glyph_builder; in Initialize()
|
/kernel/linux/linux-5.10/arch/nds32/kernel/ |
H A D | perf_event_cpu.c | 1290 struct perf_callchain_entry_ctx *entry = data; in callchain_trace() local 1292 perf_callchain_store(entry, fr->lp); in callchain_trace() 1301 user_backtrace(struct perf_callchain_entry_ctx *entry, unsigned long fp) in user_backtrace() argument 1320 perf_callchain_store(entry, lp); in user_backtrace() 1325 user_backtrace_opt_size(struct perf_callchain_entry_ctx *entry, in user_backtrace_opt_size() argument 1347 perf_callchain_store(entry, lp); in user_backtrace_opt_size() 1363 perf_callchain_user(struct perf_callchain_entry_ctx *entry, in perf_callchain_user() argument 1380 perf_callchain_store(entry, regs->ipc); in perf_callchain_user() 1385 if (entry->nr < PERF_MAX_STACK_DEPTH && in perf_callchain_user() 1436 perf_callchain_store(entry, l in perf_callchain_user() 1480 perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs) perf_callchain_kernel() argument [all...] |
/kernel/linux/linux-6.6/tools/tracing/latency/ |
H A D | latency-collector.c | 184 struct entry { struct 222 struct entry entries[QUEUE_SIZE]; 661 static int printstate_next_ticket(struct entry *req) in printstate_next_ticket() 673 void printstate_mark_req_completed(const struct entry *req) in printstate_mark_req_completed() 680 bool printstate_has_new_req_arrived(const struct entry *req) in printstate_has_new_req_arrived() 716 bool prev_req_won_race(const struct entry *req) in prev_req_won_race() 761 static int table_get_probability(const struct entry *req, in table_get_probability() 824 const struct entry *e) in queue_push_to_back() 830 static __always_inline struct entry queue_pop_from_front(struct queue *q) in queue_pop_from_front() 832 struct entry in queue_pop_from_front() [all...] |
/third_party/skia/third_party/externals/dawn/src/tests/unittests/validation/ |
H A D | BindGroupValidationTests.cpp | 385 // Chaining a struct that isn't an external texture binding entry is an error. in TEST_F() 971 wgpu::BindGroupLayoutEntry entry; in TEST_F() local 972 entry.binding = 0; in TEST_F() 973 entry.visibility = wgpu::ShaderStage::None; in TEST_F() 974 entry.buffer.type = wgpu::BufferBindingType::Uniform; in TEST_F() 977 descriptor.entries = &entry; in TEST_F() 1005 wgpu::BindGroupLayoutEntry entry; in TEST_F() member 1036 wgpu::BindGroupLayoutEntry entry = info.entry; in TEST_F() local 1037 entry in TEST_F() 1047 wgpu::BindGroupLayoutEntry entry = info.otherEntry; TEST_F() local 1056 wgpu::BindGroupLayoutEntry entry = info.entry; TEST_F() local 1066 wgpu::BindGroupLayoutEntry entry = info.entry; TEST_F() local 1081 wgpu::BindGroupLayoutEntry entry = info.entry; TEST_F() local 1100 wgpu::BindGroupLayoutEntry entry; TEST_F() member 1121 wgpu::BindGroupLayoutEntry entry = BGLEntryType(&utils::kExternalTextureBindingLayout); TEST_F() local 1128 wgpu::BindGroupLayoutEntry entry = info.entry; TEST_F() local 1139 wgpu::BindGroupLayoutEntry entry = info.otherEntry; TEST_F() local 1148 wgpu::BindGroupLayoutEntry entry = info.entry; TEST_F() local 1158 wgpu::BindGroupLayoutEntry entry = info.entry; TEST_F() local 1173 wgpu::BindGroupLayoutEntry entry = info.entry; TEST_F() local [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx4/ |
H A D | mcg.c | 144 * Add new entry to steering data structure. 192 * steering entry, as they should also receive the packets in new_steering_entry() 220 /* update the qps count and update the entry with all the promisc qps*/ in new_steering_entry() 238 /* update the data structures with existing steering entry */ 244 struct mlx4_steer_index *tmp_entry, *entry = NULL; in existing_steering_entry() local 259 entry = tmp_entry; in existing_steering_entry() 263 if (unlikely(!entry)) { in existing_steering_entry() 264 mlx4_warn(dev, "Steering entry at index %x is not registered\n", index); in existing_steering_entry() 269 * we need to add it as a duplicate to this entry in existing_steering_entry() 271 list_for_each_entry(dqp, &entry in existing_steering_entry() 293 struct mlx4_steer_index *tmp_entry, *entry = NULL; check_duplicate_entry() local 373 struct mlx4_steer_index *entry = NULL, *tmp_entry; can_remove_steering_entry() local 424 struct mlx4_steer_index *entry; add_promisc_qp() local 555 struct mlx4_steer_index *entry, *tmp_entry; remove_promisc_qp() local [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/samsung/sxgbe/ |
H A D | sxgbe_main.c | 748 unsigned int entry = tqueue->dirty_tx % tx_rsize; in sxgbe_tx_queue_clean() local 749 struct sk_buff *skb = tqueue->tx_skbuff[entry]; in sxgbe_tx_queue_clean() 752 p = tqueue->dma_tx + entry; in sxgbe_tx_queue_clean() 762 if (likely(tqueue->tx_skbuff_dma[entry])) { in sxgbe_tx_queue_clean() 764 tqueue->tx_skbuff_dma[entry], in sxgbe_tx_queue_clean() 767 tqueue->tx_skbuff_dma[entry] = 0; in sxgbe_tx_queue_clean() 772 tqueue->tx_skbuff[entry] = NULL; in sxgbe_tx_queue_clean() 1050 * sxgbe_open - open entry point of the driver 1053 * This function is the open entry point of the driver. 1199 * sxgbe_release - close entry poin 1274 unsigned int entry, frag_num; sxgbe_xmit() local 1443 unsigned int entry = priv->rxq[qnum]->dirty_rx % rxsize; sxgbe_rx_refill() local 1485 unsigned int entry = priv->rxq[qnum]->cur_rx; sxgbe_rx() local [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/smsc/ |
H A D | epic100.c | 915 /* Mark the last entry as wrapping the ring. */ in epic_init_ring() 948 int entry, free_count; in epic_start_xmit() local 958 /* Calculate the next Tx descriptor entry. */ in epic_start_xmit() 961 entry = ep->cur_tx % TX_RING_SIZE; in epic_start_xmit() 963 ep->tx_skbuff[entry] = skb; in epic_start_xmit() 964 ep->tx_ring[entry].bufaddr = dma_map_single(&ep->pci_dev->dev, in epic_start_xmit() 974 /* Leave room for an additional entry. */ in epic_start_xmit() 978 ep->tx_ring[entry].buflength = ctrl_word | skb->len; in epic_start_xmit() 979 ep->tx_ring[entry].txstatus = in epic_start_xmit() 993 skb->len, entry, ctrl_wor in epic_start_xmit() 1031 int entry = dirty_tx % TX_RING_SIZE; epic_tx() local 1142 int entry = ep->cur_rx % RX_RING_SIZE; epic_rx() local [all...] |
/kernel/linux/linux-5.10/drivers/xen/ |
H A D | xen-scsiback.c | 275 struct v2p_entry *entry = container_of(kref, struct v2p_entry, kref); in scsiback_free_translation_entry() local 276 struct scsiback_tpg *tpg = entry->tpg; in scsiback_free_translation_entry() 282 kfree(entry); in scsiback_free_translation_entry() 593 struct v2p_entry *entry; in scsiback_do_translation() local 598 list_for_each_entry(entry, head, l) { in scsiback_do_translation() 599 if ((entry->v.chn == v->chn) && in scsiback_do_translation() 600 (entry->v.tgt == v->tgt) && in scsiback_do_translation() 601 (entry->v.lun == v->lun)) { in scsiback_do_translation() 602 kref_get(&entry->kref); in scsiback_do_translation() 606 entry in scsiback_do_translation() 849 struct v2p_entry *entry; scsiback_chk_translation_entry() local 954 __scsiback_del_translation_entry(struct v2p_entry *entry) __scsiback_del_translation_entry() argument 966 struct v2p_entry *entry; scsiback_del_translation_entry() local 985 struct v2p_entry *entry; scsiback_do_add_lun() local 1181 struct v2p_entry *entry, *tmp; scsiback_release_translation_entry() local [all...] |