Home
last modified time | relevance | path

Searched refs:entry (Results 2351 - 2375 of 7646) sorted by relevance

1...<<919293949596979899100>>...306

/foundation/bundlemanager/bundle_framework/common/utils/src/
H A Dbundle_file_util.cpp180 struct dirent *entry = nullptr; in GetHapFilesFromBundlePath() local
181 while ((entry = readdir(dir)) != nullptr) { in GetHapFilesFromBundlePath()
182 if (strcmp(entry->d_name, ".") == 0 || strcmp(entry->d_name, "..") == 0) { in GetHapFilesFromBundlePath()
185 const std::string hapFilePath = bundlePath + entry->d_name; in GetHapFilesFromBundlePath()
/kernel/linux/linux-5.10/arch/loongarch/kernel/
H A Dmachine_kexec.c262 unsigned long entry, *ptr; in machine_kexec() local
279 for (ptr = &image->head; (entry = *ptr) && !(entry & IND_DONE); in machine_kexec()
280 ptr = (entry & IND_INDIRECTION) ? in machine_kexec()
281 phys_to_virt(entry & PAGE_MASK) : ptr + 1) { in machine_kexec()
/kernel/linux/linux-5.10/arch/mips/mm/
H A Dtlb-r3k.c42 static void local_flush_tlb_from(int entry) in local_flush_tlb_from() argument
48 while (entry < current_cpu_data.tlbsize) { in local_flush_tlb_from()
49 write_c0_index(entry << 8); in local_flush_tlb_from()
50 write_c0_entryhi((entry | 0x80000) << 12); in local_flush_tlb_from()
51 entry++; /* BARRIER */ in local_flush_tlb_from()
232 printk("[tlbwired<entry lo0 %8x, hi %8x\n, pagemask %8x>]\n", in add_wired_entry()
256 printk("[tlbwired<entry lo0 %8x, hi %8x\n>]\n", in add_wired_entry()
/kernel/linux/linux-5.10/arch/x86/kernel/
H A Didt.c150 /* Must be page-aligned because the real IDT is used in the cpu entry area */
271 * entry area, so that the "sidt" instruction will not leak the in idt_map_in_cea()
286 void *entry; in idt_setup_apic_and_irq_gates() local
291 entry = irq_entries_start + 8 * (i - FIRST_EXTERNAL_VECTOR); in idt_setup_apic_and_irq_gates()
292 set_intr_gate(i, entry); in idt_setup_apic_and_irq_gates()
302 entry = spurious_entries_start + 8 * (i - FIRST_SYSTEM_VECTOR); in idt_setup_apic_and_irq_gates()
303 set_intr_gate(i, entry); in idt_setup_apic_and_irq_gates()
306 /* Map IDT into CPU entry area and reload it. */ in idt_setup_apic_and_irq_gates()
/kernel/linux/linux-6.6/arch/loongarch/kernel/
H A Dmachine_kexec.c259 unsigned long entry, *ptr; in machine_kexec() local
276 for (ptr = &image->head; (entry = *ptr) && !(entry & IND_DONE); in machine_kexec()
277 ptr = (entry & IND_INDIRECTION) ? in machine_kexec()
278 phys_to_virt(entry & PAGE_MASK) : ptr + 1) { in machine_kexec()
/kernel/linux/linux-5.10/kernel/rcu/
H A Dsrcutiny.c39 INIT_LIST_HEAD(&ssp->srcu_work.entry); in init_srcu_struct_fields()
169 else if (list_empty(&ssp->srcu_work.entry)) in srcu_gp_start_if_needed()
170 list_add(&ssp->srcu_work.entry, &srcu_boot_list); in srcu_gp_start_if_needed()
268 struct srcu_struct, srcu_work.entry); in srcu_init()
269 list_del_init(&ssp->srcu_work.entry); in srcu_init()
/kernel/linux/linux-5.10/drivers/gpu/drm/via/
H A Dvia_mm.c214 struct via_memblock *entry, *next; in via_reclaim_buffers_locked() local
231 list_for_each_entry_safe(entry, next, &file_priv->obj_list, in via_reclaim_buffers_locked()
233 list_del(&entry->owner_list); in via_reclaim_buffers_locked()
234 drm_mm_remove_node(&entry->mm_node); in via_reclaim_buffers_locked()
235 kfree(entry); in via_reclaim_buffers_locked()
/kernel/linux/linux-6.6/drivers/android/
H A Dbinder_internal.h119 #define binder_for_each_debugfs_entry(entry) \
120 for ((entry) = binder_debugfs_entries; \
121 (entry)->name; \
122 (entry)++)
144 * @entry: node enqueued on list
150 struct list_head entry; member
302 * @node_entry: list entry for node->refs list in target node
503 * @fixup_entry: list entry
/kernel/linux/linux-6.6/fs/xfs/
H A Dxfs_attr_inactive.c90 struct xfs_attr_leaf_entry *entry; in xfs_attr3_leaf_inactive() local
101 entry = xfs_attr3_leaf_entryp(leaf); in xfs_attr3_leaf_inactive()
102 for (i = 0; i < ichdr.count; entry++, i++) { in xfs_attr3_leaf_inactive()
105 if (!entry->nameidx || (entry->flags & XFS_ATTR_LOCAL)) in xfs_attr3_leaf_inactive()
/kernel/linux/linux-6.6/include/trace/
H A Dtrace_custom_events.h126 struct trace_custom_event_raw_##call __maybe_unused *entry; \
148 struct trace_custom_event_raw_##call *entry; \
156 entry = trace_event_buffer_reserve(&fbuffer, trace_file, \
157 sizeof(*entry) + __data_size); \
159 if (!entry) \
/kernel/linux/linux-6.6/arch/x86/kernel/
H A Didt.c165 /* Must be page-aligned because the real IDT is used in the cpu entry area */
262 * entry area, so that the "sidt" instruction will not leak the in idt_map_in_cea()
277 void *entry; in idt_setup_apic_and_irq_gates() local
282 entry = irq_entries_start + IDT_ALIGN * (i - FIRST_EXTERNAL_VECTOR); in idt_setup_apic_and_irq_gates()
283 set_intr_gate(i, entry); in idt_setup_apic_and_irq_gates()
293 entry = spurious_entries_start + IDT_ALIGN * (i - FIRST_SYSTEM_VECTOR); in idt_setup_apic_and_irq_gates()
294 set_intr_gate(i, entry); in idt_setup_apic_and_irq_gates()
297 /* Map IDT into CPU entry area and reload it. */ in idt_setup_apic_and_irq_gates()
/kernel/linux/linux-6.6/tools/testing/selftests/bpf/map_tests/
H A Dmap_in_map_batch_ops.c53 /* Add entry into this created map in create_inner_maps()
88 int inner_map_fd, entry, err; in validate_fetch_results() local
91 for (entry = 0; entry < max_entries_fetched; ++entry) { in validate_fetch_results()
92 outer_map_value = fetched_values[entry]; in validate_fetch_results()
/kernel/linux/linux-6.6/kernel/rcu/
H A Dsrcutiny.c39 INIT_LIST_HEAD(&ssp->srcu_work.entry); in init_srcu_struct_fields()
169 else if (list_empty(&ssp->srcu_work.entry)) in srcu_gp_start_if_needed()
170 list_add(&ssp->srcu_work.entry, &srcu_boot_list); in srcu_gp_start_if_needed()
280 struct srcu_struct, srcu_work.entry); in srcu_init()
281 list_del_init(&ssp->srcu_work.entry); in srcu_init()
/kernel/linux/linux-6.6/net/ethtool/
H A Dtunnels.c87 struct nlattr *ports, *table, *entry; in ethnl_tunnel_info_fill_reply() local
138 entry = nla_nest_start(skb, ETHTOOL_A_TUNNEL_UDP_TABLE_ENTRY); in ethnl_tunnel_info_fill_reply()
139 if (!entry) in ethnl_tunnel_info_fill_reply()
148 nla_nest_end(skb, entry); in ethnl_tunnel_info_fill_reply()
157 nla_nest_cancel(skb, entry); in ethnl_tunnel_info_fill_reply()
/third_party/mesa3d/src/compiler/nir/
H A Dnir_lower_scratch.c162 struct set_entry *entry = _mesa_set_search(set, deref->var); in nir_lower_vars_to_scratch() local
163 if (!entry) in nir_lower_vars_to_scratch()
167 _mesa_set_remove(set, entry); in nir_lower_vars_to_scratch()
172 set_foreach(set, entry) { in nir_lower_vars_to_scratch()
173 nir_variable* var = (void*)entry->key; in nir_lower_vars_to_scratch()
/third_party/node/deps/v8/src/heap/cppgc/
H A Dobject-allocator.cc191 // entry is checked. in RefillLinearAllocationBuffer()
209 const FreeList::Block entry = space.free_list().Allocate(size); in RefillLinearAllocationBufferFromFreeList() local
210 if (!entry.address) return false; in RefillLinearAllocationBufferFromFreeList()
213 auto& page = *NormalPage::From(BasePage::FromPayload(entry.address)); in RefillLinearAllocationBufferFromFreeList()
220 space, stats_collector_, static_cast<Address>(entry.address), entry.size); in RefillLinearAllocationBufferFromFreeList()
/third_party/node/deps/v8/src/utils/
H A Didentity-map.h23 T* entry; member
126 return {reinterpret_cast<V*>(raw.entry), raw.already_exists}; in FindOrInsert()
170 V* entry() const { in entry() function in v8::internal::IdentityMap::Iterator
174 V* operator*() { return entry(); } in operator *()
175 V* operator->() { return entry(); } in operator ->()
/third_party/skia/third_party/externals/sfntly/cpp/src/sfntly/table/bitmap/
H A Debdt_table.cc219 for (BitmapGlyphInfoMap::iterator entry = loca_map->begin(), in Initialize()
221 entry != entry_end; entry++) { in Initialize()
222 BitmapGlyphInfoPtr info = entry->second; in Initialize()
229 glyph_builder_map[entry->first] = glyph_builder; in Initialize()
/kernel/linux/linux-5.10/arch/nds32/kernel/
H A Dperf_event_cpu.c1290 struct perf_callchain_entry_ctx *entry = data; in callchain_trace() local
1292 perf_callchain_store(entry, fr->lp); in callchain_trace()
1301 user_backtrace(struct perf_callchain_entry_ctx *entry, unsigned long fp) in user_backtrace() argument
1320 perf_callchain_store(entry, lp); in user_backtrace()
1325 user_backtrace_opt_size(struct perf_callchain_entry_ctx *entry, in user_backtrace_opt_size() argument
1347 perf_callchain_store(entry, lp); in user_backtrace_opt_size()
1363 perf_callchain_user(struct perf_callchain_entry_ctx *entry, in perf_callchain_user() argument
1380 perf_callchain_store(entry, regs->ipc); in perf_callchain_user()
1385 if (entry->nr < PERF_MAX_STACK_DEPTH && in perf_callchain_user()
1436 perf_callchain_store(entry, l in perf_callchain_user()
1480 perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs) perf_callchain_kernel() argument
[all...]
/kernel/linux/linux-6.6/tools/tracing/latency/
H A Dlatency-collector.c184 struct entry { struct
222 struct entry entries[QUEUE_SIZE];
661 static int printstate_next_ticket(struct entry *req) in printstate_next_ticket()
673 void printstate_mark_req_completed(const struct entry *req) in printstate_mark_req_completed()
680 bool printstate_has_new_req_arrived(const struct entry *req) in printstate_has_new_req_arrived()
716 bool prev_req_won_race(const struct entry *req) in prev_req_won_race()
761 static int table_get_probability(const struct entry *req, in table_get_probability()
824 const struct entry *e) in queue_push_to_back()
830 static __always_inline struct entry queue_pop_from_front(struct queue *q) in queue_pop_from_front()
832 struct entry in queue_pop_from_front()
[all...]
/third_party/skia/third_party/externals/dawn/src/tests/unittests/validation/
H A DBindGroupValidationTests.cpp385 // Chaining a struct that isn't an external texture binding entry is an error. in TEST_F()
971 wgpu::BindGroupLayoutEntry entry; in TEST_F() local
972 entry.binding = 0; in TEST_F()
973 entry.visibility = wgpu::ShaderStage::None; in TEST_F()
974 entry.buffer.type = wgpu::BufferBindingType::Uniform; in TEST_F()
977 descriptor.entries = &entry; in TEST_F()
1005 wgpu::BindGroupLayoutEntry entry; in TEST_F() member
1036 wgpu::BindGroupLayoutEntry entry = info.entry; in TEST_F() local
1037 entry in TEST_F()
1047 wgpu::BindGroupLayoutEntry entry = info.otherEntry; TEST_F() local
1056 wgpu::BindGroupLayoutEntry entry = info.entry; TEST_F() local
1066 wgpu::BindGroupLayoutEntry entry = info.entry; TEST_F() local
1081 wgpu::BindGroupLayoutEntry entry = info.entry; TEST_F() local
1100 wgpu::BindGroupLayoutEntry entry; TEST_F() member
1121 wgpu::BindGroupLayoutEntry entry = BGLEntryType(&utils::kExternalTextureBindingLayout); TEST_F() local
1128 wgpu::BindGroupLayoutEntry entry = info.entry; TEST_F() local
1139 wgpu::BindGroupLayoutEntry entry = info.otherEntry; TEST_F() local
1148 wgpu::BindGroupLayoutEntry entry = info.entry; TEST_F() local
1158 wgpu::BindGroupLayoutEntry entry = info.entry; TEST_F() local
1173 wgpu::BindGroupLayoutEntry entry = info.entry; TEST_F() local
[all...]
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx4/
H A Dmcg.c144 * Add new entry to steering data structure.
192 * steering entry, as they should also receive the packets in new_steering_entry()
220 /* update the qps count and update the entry with all the promisc qps*/ in new_steering_entry()
238 /* update the data structures with existing steering entry */
244 struct mlx4_steer_index *tmp_entry, *entry = NULL; in existing_steering_entry() local
259 entry = tmp_entry; in existing_steering_entry()
263 if (unlikely(!entry)) { in existing_steering_entry()
264 mlx4_warn(dev, "Steering entry at index %x is not registered\n", index); in existing_steering_entry()
269 * we need to add it as a duplicate to this entry in existing_steering_entry()
271 list_for_each_entry(dqp, &entry in existing_steering_entry()
293 struct mlx4_steer_index *tmp_entry, *entry = NULL; check_duplicate_entry() local
373 struct mlx4_steer_index *entry = NULL, *tmp_entry; can_remove_steering_entry() local
424 struct mlx4_steer_index *entry; add_promisc_qp() local
555 struct mlx4_steer_index *entry, *tmp_entry; remove_promisc_qp() local
[all...]
/kernel/linux/linux-5.10/drivers/net/ethernet/samsung/sxgbe/
H A Dsxgbe_main.c748 unsigned int entry = tqueue->dirty_tx % tx_rsize; in sxgbe_tx_queue_clean() local
749 struct sk_buff *skb = tqueue->tx_skbuff[entry]; in sxgbe_tx_queue_clean()
752 p = tqueue->dma_tx + entry; in sxgbe_tx_queue_clean()
762 if (likely(tqueue->tx_skbuff_dma[entry])) { in sxgbe_tx_queue_clean()
764 tqueue->tx_skbuff_dma[entry], in sxgbe_tx_queue_clean()
767 tqueue->tx_skbuff_dma[entry] = 0; in sxgbe_tx_queue_clean()
772 tqueue->tx_skbuff[entry] = NULL; in sxgbe_tx_queue_clean()
1050 * sxgbe_open - open entry point of the driver
1053 * This function is the open entry point of the driver.
1199 * sxgbe_release - close entry poin
1274 unsigned int entry, frag_num; sxgbe_xmit() local
1443 unsigned int entry = priv->rxq[qnum]->dirty_rx % rxsize; sxgbe_rx_refill() local
1485 unsigned int entry = priv->rxq[qnum]->cur_rx; sxgbe_rx() local
[all...]
/kernel/linux/linux-5.10/drivers/net/ethernet/smsc/
H A Depic100.c915 /* Mark the last entry as wrapping the ring. */ in epic_init_ring()
948 int entry, free_count; in epic_start_xmit() local
958 /* Calculate the next Tx descriptor entry. */ in epic_start_xmit()
961 entry = ep->cur_tx % TX_RING_SIZE; in epic_start_xmit()
963 ep->tx_skbuff[entry] = skb; in epic_start_xmit()
964 ep->tx_ring[entry].bufaddr = dma_map_single(&ep->pci_dev->dev, in epic_start_xmit()
974 /* Leave room for an additional entry. */ in epic_start_xmit()
978 ep->tx_ring[entry].buflength = ctrl_word | skb->len; in epic_start_xmit()
979 ep->tx_ring[entry].txstatus = in epic_start_xmit()
993 skb->len, entry, ctrl_wor in epic_start_xmit()
1031 int entry = dirty_tx % TX_RING_SIZE; epic_tx() local
1142 int entry = ep->cur_rx % RX_RING_SIZE; epic_rx() local
[all...]
/kernel/linux/linux-5.10/drivers/xen/
H A Dxen-scsiback.c275 struct v2p_entry *entry = container_of(kref, struct v2p_entry, kref); in scsiback_free_translation_entry() local
276 struct scsiback_tpg *tpg = entry->tpg; in scsiback_free_translation_entry()
282 kfree(entry); in scsiback_free_translation_entry()
593 struct v2p_entry *entry; in scsiback_do_translation() local
598 list_for_each_entry(entry, head, l) { in scsiback_do_translation()
599 if ((entry->v.chn == v->chn) && in scsiback_do_translation()
600 (entry->v.tgt == v->tgt) && in scsiback_do_translation()
601 (entry->v.lun == v->lun)) { in scsiback_do_translation()
602 kref_get(&entry->kref); in scsiback_do_translation()
606 entry in scsiback_do_translation()
849 struct v2p_entry *entry; scsiback_chk_translation_entry() local
954 __scsiback_del_translation_entry(struct v2p_entry *entry) __scsiback_del_translation_entry() argument
966 struct v2p_entry *entry; scsiback_del_translation_entry() local
985 struct v2p_entry *entry; scsiback_do_add_lun() local
1181 struct v2p_entry *entry, *tmp; scsiback_release_translation_entry() local
[all...]

Completed in 24 milliseconds

1...<<919293949596979899100>>...306