/kernel/linux/linux-6.6/tools/testing/selftests/bpf/ |
H A D | xsk.h | 108 __u32 entries = r->cached_prod - r->cached_cons; in xsk_cons_nb_avail() local 110 if (entries == 0) { in xsk_cons_nb_avail() 112 entries = r->cached_prod - r->cached_cons; in xsk_cons_nb_avail() 115 return (entries > nb) ? nb : entries; in xsk_cons_nb_avail() 144 __u32 entries = xsk_cons_nb_avail(cons, nb); in xsk_ring_cons__peek() local 146 if (entries > 0) { in xsk_ring_cons__peek() 148 cons->cached_cons += entries; in xsk_ring_cons__peek() 151 return entries; in xsk_ring_cons__peek() 162 * with the entries b in xsk_ring_cons__release() [all...] |
/foundation/distributeddatamgr/kv_store/frameworks/libs/distributeddb/syncer/src/device/multiver/ |
H A D | multi_ver_data_sync.cpp | 271 std::vector<MultiVerKvEntry *> entries; in AckRecvCallback() local 278 entries.push_back(item); in AckRecvCallback() 281 context->SetEntries(entries); in AckRecvCallback() 283 context->SetEntriesSize(static_cast<int>(entries.size())); in AckRecvCallback() 284 LOGD("MultiVerDataSync::AckRecvCallback src=%s{private}, entries num = %zu", in AckRecvCallback() 285 context->GetDeviceId().c_str(), entries.size()); in AckRecvCallback() 287 if (entries.size() > 0) { in AckRecvCallback() 288 entry = entries[0]; in AckRecvCallback() 302 int MultiVerDataSync::PutCommitData(const MultiVerCommitNode &commit, const std::vector<MultiVerKvEntry *> &entries, in PutCommitData() argument 305 return storagePtr_->PutCommitData(commit, entries, deviceNam in PutCommitData() 463 std::vector<std::vector<uint8_t>> entries; AckPacketSerialization() local 504 std::vector<std::vector<uint8_t>> entries; AckPacketDeSerialization() local 652 std::vector<std::vector<uint8_t>> entries; SendAckPacket() local 683 GetCommitData(const MultiVerCommitNode &commit, std::vector<MultiVerKvEntry *> &entries) GetCommitData() argument [all...] |
/kernel/linux/linux-5.10/arch/x86/kernel/ |
H A D | ldt.c | 76 * If page table isolation is enabled, ldt->entries in load_mm_ldt() 83 set_ldt(ldt->entries, ldt->nr_entries); in load_mm_ldt() 171 new_ldt->entries = __vmalloc(alloc_size, GFP_KERNEL_ACCOUNT | __GFP_ZERO); in alloc_ldt_struct() 173 new_ldt->entries = (void *)get_zeroed_page(GFP_KERNEL_ACCOUNT); in alloc_ldt_struct() 175 if (!new_ldt->entries) { in alloc_ldt_struct() 307 is_vmalloc = is_vmalloc_addr(ldt->entries); in map_ldt_struct() 313 const void *src = (char *)ldt->entries + offset; in map_ldt_struct() 410 paravirt_alloc_ldt(ldt->entries, ldt->nr_entries); in finalize_ldt_struct() 431 paravirt_free_ldt(ldt->entries, ldt->nr_entries); in free_ldt_struct() 433 vfree_atomic(ldt->entries); in free_ldt_struct() [all...] |
/kernel/linux/linux-6.6/arch/x86/kernel/ |
H A D | ldt.c | 76 * If page table isolation is enabled, ldt->entries in load_mm_ldt() 83 set_ldt(ldt->entries, ldt->nr_entries); in load_mm_ldt() 171 new_ldt->entries = __vmalloc(alloc_size, GFP_KERNEL_ACCOUNT | __GFP_ZERO); in alloc_ldt_struct() 173 new_ldt->entries = (void *)get_zeroed_page(GFP_KERNEL_ACCOUNT); in alloc_ldt_struct() 175 if (!new_ldt->entries) { in alloc_ldt_struct() 307 is_vmalloc = is_vmalloc_addr(ldt->entries); in map_ldt_struct() 313 const void *src = (char *)ldt->entries + offset; in map_ldt_struct() 418 paravirt_alloc_ldt(ldt->entries, ldt->nr_entries); in finalize_ldt_struct() 439 paravirt_free_ldt(ldt->entries, ldt->nr_entries); in free_ldt_struct() 441 vfree_atomic(ldt->entries); in free_ldt_struct() [all...] |
/kernel/linux/linux-6.6/drivers/acpi/apei/ |
H A D | erst.c | 371 ERST_TAB_ENTRY(erst_tab), erst_tab->entries); in erst_exec_ctx_init() 429 u64 *entries; member 478 u64 *entries; in __erst_record_id_cache_add_one() local 499 entries = erst_record_id_cache.entries; in __erst_record_id_cache_add_one() 501 if (entries[i] == id) in __erst_record_id_cache_add_one() 519 new_entries = kvmalloc_array(new_size, sizeof(entries[0]), in __erst_record_id_cache_add_one() 523 memcpy(new_entries, entries, in __erst_record_id_cache_add_one() 524 erst_record_id_cache.len * sizeof(entries[0])); in __erst_record_id_cache_add_one() 525 kvfree(entries); in __erst_record_id_cache_add_one() 543 u64 *entries; erst_get_record_id_next() local 588 u64 *entries; __erst_record_id_cache_compact() local 862 u64 *entries; erst_clear_cache() local 931 u64 *entries; erst_clear() local [all...] |
/kernel/linux/linux-6.6/drivers/net/dsa/sja1105/ |
H A D | sja1105_main.c | 75 vlan = priv->static_config.tables[BLK_IDX_VLAN_LOOKUP].entries; in sja1105_is_vlan_configured() 91 mac = priv->static_config.tables[BLK_IDX_MAC_CONFIG].entries; in sja1105_drop_untagged() 106 mac = priv->static_config.tables[BLK_IDX_MAC_CONFIG].entries; in sja1105_pvid_apply() 144 vlan = priv->static_config.tables[BLK_IDX_VLAN_LOOKUP].entries; in sja1105_commit_pvid() 208 kfree(table->entries); in sja1105_init_mac_settings() 212 table->entries = kcalloc(table->ops->max_entry_count, in sja1105_init_mac_settings() 214 if (!table->entries) in sja1105_init_mac_settings() 219 mac = table->entries; in sja1105_init_mac_settings() 229 * learning to ensure that only FDB entries belonging to the in sja1105_init_mac_settings() 259 kfree(table->entries); in sja1105_init_mii_settings() [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/amd/pm/powerplay/hwmgr/ |
H A D | processpptables.c | 396 dep_table->entries[i].clk = in get_clock_voltage_dependency_table() 397 ((unsigned long)table->entries[i].ucClockHigh << 16) | in get_clock_voltage_dependency_table() 398 le16_to_cpu(table->entries[i].usClockLow); in get_clock_voltage_dependency_table() 399 dep_table->entries[i].v = in get_clock_voltage_dependency_table() 400 (unsigned long)le16_to_cpu(table->entries[i].usVoltage); in get_clock_voltage_dependency_table() 423 clock_table->values[i] = (unsigned long)table->entries[i].clk; in get_valid_clk() 434 limits->sclk = ((unsigned long)table->entries[0].ucSclkHigh << 16) | in get_clock_voltage_limit() 435 le16_to_cpu(table->entries[0].usSclkLow); in get_clock_voltage_limit() 436 limits->mclk = ((unsigned long)table->entries[0].ucMclkHigh << 16) | in get_clock_voltage_limit() 437 le16_to_cpu(table->entries[ in get_clock_voltage_limit() [all...] |
/test/xts/acts/distributeddatamgr/distributedKVStoretest/distributedKVStorejstest/hap/entry/src/ohosTest/js/test/ |
H A D | DeviceKvStoreKVCallbackJsTest.js | 47 let entries = []; 56 entries.push(entry); 58 return entries; 1048 let entries = []; 1058 entries.push(entry); 1060 console.info('SUB_DDM_DKV_DEVICESTORE_PUTBATCH_0100 entries: ' + JSON.stringify(entries)); 1061 await kvStore.putBatch(entries, async function (err,data) { 1089 let entries = []; 1099 entries [all...] |
/third_party/node/test/parallel/ |
H A D | test-fs-opendir.js | 48 const entries = files.map(() => { 53 assert.deepStrictEqual(entries.map((d) => d.name), files); 54 assert.deepStrictEqual(entries.map((d) => d.path), files.map((name) => path.join(testDir, name))); 55 assert.deepStrictEqual(entries.map((d) => d.parentPath), Array(entries.length).fill(testDir)); 57 // dir.read should return null when no more entries exist 124 const entries = []; 129 entries.push(dirent.name); 133 assert.deepStrictEqual(files, entries.sort()); 135 // dir.read should return null when no more entries exis [all...] |
/foundation/graphic/graphic_3d/lume/LumeEngine/src/io/ |
H A D | proxy_directory.cpp | 66 std::unordered_set<IDirectory::Entry> entries; in GetEntries() local 69 entries.insert(entry); in GetEntries() 72 result.reserve(entries.size()); in GetEntries() 75 std::move(entries.begin(), entries.end(), std::back_inserter(result)); in GetEntries()
|
/foundation/bundlemanager/bundle_framework/interfaces/kits/js/zip/src/ |
H A D | zip.cpp | 139 std::list<FileAccessor::DirectoryContentEntry> entries; in Zip() local 141 entries.push_back(FileAccessor::DirectoryContentEntry(srcDir, true)); in Zip() 143 for (auto iter = entries.begin(); iter != entries.end(); ++iter) { in Zip() 144 if (iter != entries.begin() && ((!params.GetIncludeHiddenFiles() && IsHiddenFile(iter->path)) || in Zip() 148 if (iter != entries.begin()) { in Zip() 159 entries.insert(entries.end(), subEntries.begin(), subEntries.end()); in Zip() 180 std::list<FileAccessor::DirectoryContentEntry> &entries, in GetZipsAllRelativeFilesInner() 184 for (auto iter = entries in GetZipsAllRelativeFilesInner() 179 GetZipsAllRelativeFilesInner(const ZipParams ¶ms, const FilePath &iterPath, std::list<FileAccessor::DirectoryContentEntry> &entries, std::vector<std::pair<FilePath, FilePath>> &allRelativeFiles) GetZipsAllRelativeFilesInner() argument 207 std::list<FileAccessor::DirectoryContentEntry> entries; GetZipsAllRelativeFiles() local [all...] |
/foundation/distributeddatamgr/kv_store/frameworks/libs/distributeddb/test/moduletest/src/ |
H A D | distributeddb_nb_crud_power_test.cpp | 91 vector<Entry> entries; in RepeatExeCrud() local 93 EXPECT_EQ(g_nbDelegate->GetEntries({'p'}, entries), OK); in RepeatExeCrud() 94 entries.clear(); in RepeatExeCrud() 98 EXPECT_EQ(g_nbDelegate->GetEntries({'u'}, entries), OK); in RepeatExeCrud() 99 entries.clear(); in RepeatExeCrud() 102 EXPECT_EQ(g_nbDelegate->GetEntries(KEY_EMPTY, entries), OK); in RepeatExeCrud() 103 entries.clear(); in RepeatExeCrud()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/cisco/enic/ |
H A D | vnic_rq.h | 56 /* Break the vnic_rq_buf allocations into blocks of 32/64 entries */ 59 #define VNIC_RQ_BUF_BLK_ENTRIES(entries) \ 60 ((unsigned int)((entries < VNIC_RQ_BUF_DFLT_BLK_ENTRIES) ? \ 62 #define VNIC_RQ_BUF_BLK_SZ(entries) \ 63 (VNIC_RQ_BUF_BLK_ENTRIES(entries) * sizeof(struct vnic_rq_buf)) 64 #define VNIC_RQ_BUF_BLKS_NEEDED(entries) \ 65 DIV_ROUND_UP(entries, VNIC_RQ_BUF_BLK_ENTRIES(entries))
|
H A D | vnic_wq.h | 68 /* Break the vnic_wq_buf allocations into blocks of 32/64 entries */ 71 #define VNIC_WQ_BUF_BLK_ENTRIES(entries) \ 72 ((unsigned int)((entries < VNIC_WQ_BUF_DFLT_BLK_ENTRIES) ? \ 74 #define VNIC_WQ_BUF_BLK_SZ(entries) \ 75 (VNIC_WQ_BUF_BLK_ENTRIES(entries) * sizeof(struct vnic_wq_buf)) 76 #define VNIC_WQ_BUF_BLKS_NEEDED(entries) \ 77 DIV_ROUND_UP(entries, VNIC_WQ_BUF_BLK_ENTRIES(entries))
|
/kernel/linux/linux-6.6/drivers/net/ethernet/cisco/enic/ |
H A D | vnic_rq.h | 43 /* Break the vnic_rq_buf allocations into blocks of 32/64 entries */ 46 #define VNIC_RQ_BUF_BLK_ENTRIES(entries) \ 47 ((unsigned int)((entries < VNIC_RQ_BUF_DFLT_BLK_ENTRIES) ? \ 49 #define VNIC_RQ_BUF_BLK_SZ(entries) \ 50 (VNIC_RQ_BUF_BLK_ENTRIES(entries) * sizeof(struct vnic_rq_buf)) 51 #define VNIC_RQ_BUF_BLKS_NEEDED(entries) \ 52 DIV_ROUND_UP(entries, VNIC_RQ_BUF_BLK_ENTRIES(entries))
|
H A D | vnic_wq.h | 55 /* Break the vnic_wq_buf allocations into blocks of 32/64 entries */ 58 #define VNIC_WQ_BUF_BLK_ENTRIES(entries) \ 59 ((unsigned int)((entries < VNIC_WQ_BUF_DFLT_BLK_ENTRIES) ? \ 61 #define VNIC_WQ_BUF_BLK_SZ(entries) \ 62 (VNIC_WQ_BUF_BLK_ENTRIES(entries) * sizeof(struct vnic_wq_buf)) 63 #define VNIC_WQ_BUF_BLKS_NEEDED(entries) \ 64 DIV_ROUND_UP(entries, VNIC_WQ_BUF_BLK_ENTRIES(entries))
|
/kernel/linux/linux-5.10/sound/firewire/dice/ |
H A D | dice-presonus.c | 27 } *entry, entries[] = { in snd_dice_detect_presonus_formats() local 43 for (i = 0; i < ARRAY_SIZE(entries); ++i) { in snd_dice_detect_presonus_formats() 44 entry = entries + i; in snd_dice_detect_presonus_formats() 48 if (i == ARRAY_SIZE(entries)) in snd_dice_detect_presonus_formats()
|
/kernel/linux/linux-5.10/kernel/ |
H A D | backtracetest.c | 47 unsigned long entries[8]; in backtrace_test_saved() local 53 nr_entries = stack_trace_save(entries, ARRAY_SIZE(entries), 0); in backtrace_test_saved() 54 stack_trace_print(entries, nr_entries, 0); in backtrace_test_saved()
|
/kernel/linux/linux-6.6/kernel/ |
H A D | backtracetest.c | 47 unsigned long entries[8]; in backtrace_test_saved() local 53 nr_entries = stack_trace_save(entries, ARRAY_SIZE(entries), 0); in backtrace_test_saved() 54 stack_trace_print(entries, nr_entries, 0); in backtrace_test_saved()
|
/kernel/linux/linux-6.6/sound/firewire/dice/ |
H A D | dice-presonus.c | 25 } *entry, entries[] = { in snd_dice_detect_presonus_formats() local 41 for (i = 0; i < ARRAY_SIZE(entries); ++i) { in snd_dice_detect_presonus_formats() 42 entry = entries + i; in snd_dice_detect_presonus_formats() 46 if (i == ARRAY_SIZE(entries)) in snd_dice_detect_presonus_formats()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/subdev/mxm/ |
H A D | mxms.c | 107 u8 entries = 0; in mxms_foreach() local 124 entries = (ROM32(desc[0]) & 0x01f00000) >> 20; in mxms_foreach() 133 entries = (desc[1] & 0xf0) >> 4; in mxms_foreach() 141 entries = desc[1] & 0x07; in mxms_foreach() 162 for (i = 0; i < entries; i++, dump += recordlen) { in mxms_foreach() 174 desc += headerlen + (entries * recordlen); in mxms_foreach()
|
/kernel/linux/linux-5.10/drivers/scsi/snic/ |
H A D | vnic_wq.h | 60 /* Break the vnic_wq_buf allocations into blocks of 64 entries */ 63 #define VNIC_WQ_BUF_BLK_ENTRIES(entries) \ 64 ((unsigned int)(entries < VNIC_WQ_BUF_DFLT_BLK_ENTRIES) ? \ 68 #define VNIC_WQ_BUF_BLKS_NEEDED(entries) \ 69 DIV_ROUND_UP(entries, VNIC_WQ_BUF_DFLT_BLK_ENTRIES) 70 #define VNIC_WQ_BUF_BLKS_NEEDED(entries) \ 71 DIV_ROUND_UP(entries, VNIC_WQ_BUF_DFLT_BLK_ENTRIES)
|
/kernel/linux/linux-5.10/tools/lib/perf/include/internal/ |
H A D | evlist.h | 17 struct list_head entries; member 71 __perf_evlist__for_each_entry(&(evlist)->entries, evsel) 87 __perf_evlist__for_each_entry_reverse(&(evlist)->entries, evsel) 105 __perf_evlist__for_each_entry_safe(&(evlist)->entries, tmp, evsel) 109 return list_entry(evlist->entries.next, struct perf_evsel, node); in perf_evlist__first() 114 return list_entry(evlist->entries.prev, struct perf_evsel, node); in perf_evlist__last()
|
/kernel/linux/linux-6.6/drivers/scsi/snic/ |
H A D | vnic_wq.h | 46 /* Break the vnic_wq_buf allocations into blocks of 64 entries */ 49 #define VNIC_WQ_BUF_BLK_ENTRIES(entries) \ 50 ((unsigned int)(entries < VNIC_WQ_BUF_DFLT_BLK_ENTRIES) ? \ 54 #define VNIC_WQ_BUF_BLKS_NEEDED(entries) \ 55 DIV_ROUND_UP(entries, VNIC_WQ_BUF_DFLT_BLK_ENTRIES) 56 #define VNIC_WQ_BUF_BLKS_NEEDED(entries) \ 57 DIV_ROUND_UP(entries, VNIC_WQ_BUF_DFLT_BLK_ENTRIES)
|
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/nvkm/subdev/mxm/ |
H A D | mxms.c | 107 u8 entries = 0; in mxms_foreach() local 124 entries = (ROM32(desc[0]) & 0x01f00000) >> 20; in mxms_foreach() 133 entries = (desc[1] & 0xf0) >> 4; in mxms_foreach() 141 entries = desc[1] & 0x07; in mxms_foreach() 162 for (i = 0; i < entries; i++, dump += recordlen) { in mxms_foreach() 174 desc += headerlen + (entries * recordlen); in mxms_foreach()
|