/kernel/linux/linux-5.10/include/linux/ |
H A D | interconnect-provider.h | 31 * @num_nodes: number of nodes in this device 32 * @nodes: array of pointers to the nodes in this device 36 struct icc_node *nodes[]; member 47 * @nodes: internal list of the interconnect provider nodes 53 * @xlate: provider-specific callback for mapping nodes from phandle arguments 62 struct list_head nodes; member 82 * @num_links: number of links to other interconnect nodes 84 * @node_list: the list entry in the parent provider's "nodes" lis [all...] |
/kernel/linux/linux-6.6/fs/btrfs/ |
H A D | delayed-inode.h | 28 * Used for delayed nodes which is waiting to be dealt with by the 35 int nodes; /* for delayed nodes */ member 50 * Used to add the node into the prepare list, the nodes in this list 106 delayed_root->nodes = 0; in btrfs_init_delayed_root()
|
H A D | backref.c | 491 eb = path->nodes[level]; in add_all_parents() 508 eb = path->nodes[0]; in add_all_parents() 519 eb = path->nodes[0]; in add_all_parents() 608 * other tree nodes. This happens when qgroups does backref walks when in resolve_indirect_ref() 680 eb = path->nodes[level]; in resolve_indirect_ref() 687 eb = path->nodes[level]; in resolve_indirect_ref() 731 * New backrefs (i.e., for parent nodes) are added to the appropriate 1035 leaf = path->nodes[0]; in add_inline_refs() 1170 leaf = path->nodes[0]; in add_keyed_refs() 1301 * If the node at this level is shared, than all nodes belo in lookup_backref_shared_cache() [all...] |
H A D | defrag.c | 400 * leafs from path->nodes[1], so set lowest_level to 1 to avoid later in btrfs_defrag_leaves() 410 if (!path->nodes[1]) { in btrfs_defrag_leaves() 421 path->nodes[1], 0, in btrfs_defrag_leaves() 437 path->slots[1] = btrfs_header_nritems(path->nodes[1]); in btrfs_defrag_leaves() 474 * will not visit leaves and nodes with a generation smaller than given 515 if (path.slots[0] >= btrfs_header_nritems(path.nodes[0])) { in defrag_get_extent() 521 ASSERT(btrfs_header_nritems(path.nodes[0])); in defrag_get_extent() 522 path.slots[0] = btrfs_header_nritems(path.nodes[0]) - 1; in defrag_get_extent() 524 btrfs_item_key_to_cpu(path.nodes[0], &key, path.slots[0]); in defrag_get_extent() 532 btrfs_item_key_to_cpu(path.nodes[ in defrag_get_extent() [all...] |
H A D | verity.c | 150 btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); in drop_verity_items() 258 leaf = path->nodes[0]; in write_key_bytes() 336 leaf = path->nodes[0]; in read_key_bytes() 393 if (path->slots[0] >= btrfs_header_nritems(path->nodes[0])) { in read_key_bytes()
|
/kernel/linux/linux-6.6/include/linux/ |
H A D | interconnect-provider.h | 31 * @num_nodes: number of nodes in this device 32 * @nodes: array of pointers to the nodes in this device 36 struct icc_node *nodes[] __counted_by(num_nodes); 47 * @nodes: internal list of the interconnect provider nodes 53 * @xlate: provider-specific callback for mapping nodes from phandle arguments 62 struct list_head nodes; member 82 * @num_links: number of links to other interconnect nodes 84 * @node_list: the list entry in the parent provider's "nodes" lis [all...] |
/kernel/linux/linux-6.6/drivers/md/persistent-data/ |
H A D | dm-btree-internal.h | 68 struct dm_block *nodes[2]; member 81 struct dm_block *nodes[2]; member
|
H A D | dm-btree.c | 550 * Redistributes entries between two btree nodes to make them 579 * Redistribute entries between three nodes. Assumes the central 625 * Splits a node by creating a sibling node and shifting half the nodes 693 s->nodes[1] = left; in split_one_into_two() 696 s->nodes[1] = right; in split_one_into_two() 734 * Splits two nodes into three. This is more work, but results in fuller 735 * nodes, so saves metadata space. 801 s->nodes[1] = left; in split_two_into_three() 805 s->nodes[1] = middle; in split_two_into_three() 809 s->nodes[ in split_two_into_three() [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_xgmi.c | 256 * created which contains the hive ID and the list of nodes. 338 sprintf(buf + 3 * i, "%02x ", top->nodes[i].num_hops); in amdgpu_xgmi_show_num_hops() 353 sprintf(buf + 3 * i, "%02x ", top->nodes[i].num_links); in amdgpu_xgmi_show_num_links() 672 if (top->nodes[i].node_id == peer_adev->gmc.xgmi.node_id) in amdgpu_xgmi_get_hops_count() 673 return top->nodes[i].num_hops & num_hops_mask; in amdgpu_xgmi_get_hops_count() 684 if (top->nodes[i].node_id == peer_adev->gmc.xgmi.node_id) in amdgpu_xgmi_get_num_links() 685 return top->nodes[i].num_links; in amdgpu_xgmi_get_num_links() 768 top_info->nodes[count++].node_id = entry->node_id; in amdgpu_xgmi_add_device() 780 top_info->nodes[count - 1].node_id = in amdgpu_xgmi_add_device()
|
/kernel/linux/linux-5.10/include/trace/events/ |
H A D | bcache.h | 298 TP_PROTO(unsigned nodes), 299 TP_ARGS(nodes), 302 __field(unsigned, nodes ) 306 __entry->nodes = nodes; 309 TP_printk("coalesced %u nodes", __entry->nodes)
|
/kernel/linux/linux-6.6/tools/perf/util/ |
H A D | header.c | 702 struct numa_topology_node *n = &tp->nodes[i]; in write_numa_topology() 957 struct hybrid_topology_node *n = &tp->nodes[i]; in write_hybrid_topology() 1411 struct memory_node *nodes = NULL; in build_mem_topology() local 1437 reallocarray(nodes, cnt + 4, sizeof(*nodes)); in build_mem_topology() 1440 pr_err("Failed to write MEM_TOPOLOGY, size %zd nodes\n", size); in build_mem_topology() 1444 nodes = new_nodes; in build_mem_topology() 1447 ret = memory_node__read(&nodes[cnt], idx); in build_mem_topology() 1455 *nodesp = nodes; in build_mem_topology() 1456 qsort(nodes, cn in build_mem_topology() 1481 struct memory_node *nodes = NULL; write_mem_topology() local 2247 struct memory_node *nodes; print_mem_topology() local 2711 struct numa_node *nodes, *n; process_numa_topology() local 2975 struct memory_node *nodes; process_mem_topology() local 3070 struct hybrid_node *nodes, *n; process_hybrid_topology() local [all...] |
/kernel/linux/linux-5.10/fs/unicode/ |
H A D | mkutf8data.c | 113 * Internal nodes are one byte for the node itself, and up to three 128 * Due to the way utf8 works, there cannot be branching nodes with 129 * NEXTBYTE set, and moreover those nodes always have a righthand 424 int nodes, singletons, leaves; in tree_walk() local 426 nodes = singletons = leaves = 0; in tree_walk() 444 nodes += 1; in tree_walk() 485 printf("nodes %d leaves %d singletons %d\n", in tree_walk() 486 nodes, leaves, singletons); in tree_walk() 638 * Prune internal nodes. 641 * been collapsed. There are still internal nodes tha 1198 int nodes[4]; emit() local [all...] |
/kernel/linux/linux-6.6/fs/unicode/ |
H A D | mkutf8data.c | 113 * Internal nodes are one byte for the node itself, and up to three 128 * Due to the way utf8 works, there cannot be branching nodes with 129 * NEXTBYTE set, and moreover those nodes always have a righthand 424 int nodes, singletons, leaves; in tree_walk() local 426 nodes = singletons = leaves = 0; in tree_walk() 444 nodes += 1; in tree_walk() 485 printf("nodes %d leaves %d singletons %d\n", in tree_walk() 486 nodes, leaves, singletons); in tree_walk() 638 * Prune internal nodes. 641 * been collapsed. There are still internal nodes tha 1198 int nodes[4]; emit() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | ptp.c | 32 struct mlx5e_ptp_port_ts_cqe_tracker *nodes; member 41 struct mlx5e_ptp_port_ts_cqe_tracker *tracker = &list->nodes[metadata]; in mlx5e_ptp_port_ts_cqe_list_add() 53 struct mlx5e_ptp_port_ts_cqe_tracker *tracker = &list->nodes[metadata]; in mlx5e_ptp_port_ts_cqe_list_remove() 191 if (likely(pending_cqe_list->nodes[metadata_id].inuse)) { in mlx5e_ptp_handle_ts_cqe() 389 cqe_list->nodes = kvzalloc_node(array_size(db_sz, sizeof(*cqe_list->nodes)), in mlx5e_ptp_alloc_traffic_db() 391 if (!cqe_list->nodes) in mlx5e_ptp_alloc_traffic_db() 404 cqe_list->nodes[md].metadata_id = md; in mlx5e_ptp_alloc_traffic_db() 421 kvfree(cqe_list->nodes); in mlx5e_ptp_alloc_traffic_db() 443 kvfree(ptpsq->ts_cqe_pending_list->nodes); in mlx5e_ptp_free_traffic_db() [all...] |
/kernel/linux/linux-6.6/fs/ntfs3/ |
H A D | index.c | 661 struct indx_node *n = fnd->nodes[i]; in fnd_clear() 667 fnd->nodes[i] = NULL; in fnd_clear() 678 if (i < 0 || i >= ARRAY_SIZE(fnd->nodes)) in fnd_push() 680 fnd->nodes[i] = n; in fnd_push() 692 n = fnd->nodes[i]; in fnd_pop() 693 fnd->nodes[i] = NULL; in fnd_pop() 1230 n = fnd->nodes[level - 1]; in indx_find_sort() 1292 n = fnd->nodes[level - 1]; in indx_find_sort() 1368 n = fnd->nodes[fnd->level - 1]; in indx_find_raw() 1789 struct indx_node *n1 = fnd->nodes[leve in indx_insert_into_buffer() [all...] |
/kernel/linux/linux-5.10/Documentation/sphinx/ |
H A D | kernel_abi.py | 44 from docutils import nodes, statemachine namespace 142 node = nodes.section()
|
/kernel/linux/linux-5.10/tools/perf/util/ |
H A D | cputopo.h | 25 struct numa_topology_node nodes[]; member
|
H A D | cputopo.c | 324 tp = zalloc(sizeof(*tp) + sizeof(tp->nodes[0])*nr); in numa_topology__new() 331 if (load_numa_node(&tp->nodes[i], node_map->map[i])) { in numa_topology__new() 350 zfree(&tp->nodes[i].cpus); in numa_topology__delete()
|
/kernel/linux/linux-5.10/fs/ubifs/ |
H A D | scan.c | 13 * determining what nodes are in an eraseblock. The scan is used to replay the 138 INIT_LIST_HEAD(&sleb->nodes); in ubifs_start_scan() 150 * Note, we ignore integrity errors (EBASMSG) because all the nodes are in ubifs_start_scan() 205 * nodes. in ubifs_add_snod() 213 list_add_tail(&snod->list, &sleb->nodes); in ubifs_add_snod() 359 head = &sleb->nodes; in ubifs_scan_destroy()
|
/kernel/linux/linux-6.6/fs/ubifs/ |
H A D | scan.c | 13 * determining what nodes are in an eraseblock. The scan is used to replay the 138 INIT_LIST_HEAD(&sleb->nodes); in ubifs_start_scan() 150 * Note, we ignore integrity errors (EBASMSG) because all the nodes are in ubifs_start_scan() 205 * nodes. in ubifs_add_snod() 213 list_add_tail(&snod->list, &sleb->nodes); in ubifs_add_snod() 359 head = &sleb->nodes; in ubifs_scan_destroy()
|
/kernel/linux/linux-5.10/drivers/md/persistent-data/ |
H A D | dm-btree.c | 506 * Splits a node by creating a sibling node and shifting half the nodes 593 s->nodes[1] = left; in btree_split_sibling() 596 s->nodes[1] = right; in btree_split_sibling() 1002 struct cursor_node *n = c->nodes + c->depth - 1; in prefetch_values() 1017 struct cursor_node *n = c->nodes + c->depth - 1; in leaf_node() 1026 struct cursor_node *n = c->nodes + c->depth; in push_node() 1049 unlock_block(c->info, c->nodes[c->depth].b); in pop_node() 1061 n = c->nodes + c->depth - 1; in inc_or_backtrack() 1082 n = c->nodes + c->depth - 1; in find_leaf() 1154 struct cursor_node *n = c->nodes in dm_btree_cursor_get_value() [all...] |
/kernel/linux/linux-6.6/drivers/vfio/ |
H A D | vfio_main.c | 1001 struct interval_tree_node *nodes; in vfio_ioctl_device_feature_logging_start() local 1026 nodes = kmalloc_array(nnodes, sizeof(struct interval_tree_node), in vfio_ioctl_device_feature_logging_start() 1028 if (!nodes) in vfio_ioctl_device_feature_logging_start() 1048 nodes[i].start = range.iova; in vfio_ioctl_device_feature_logging_start() 1049 nodes[i].last = range.iova + range.length - 1; in vfio_ioctl_device_feature_logging_start() 1050 if (interval_tree_iter_first(&root, nodes[i].start, in vfio_ioctl_device_feature_logging_start() 1051 nodes[i].last)) { in vfio_ioctl_device_feature_logging_start() 1056 interval_tree_insert(nodes + i, &root); in vfio_ioctl_device_feature_logging_start() 1070 kfree(nodes); in vfio_ioctl_device_feature_logging_start()
|
/kernel/linux/linux-5.10/arch/powerpc/oprofile/cell/ |
H A D | spu_task_sync.c | 447 int nodes = 0; in number_of_online_nodes() local 450 if (tmp > nodes) in number_of_online_nodes() 451 nodes++; in number_of_online_nodes() 453 return nodes; in number_of_online_nodes()
|
/kernel/linux/linux-5.10/drivers/net/wireless/ath/ath9k/ |
H A D | dynack.c | 130 list_for_each_entry(an, &da->nodes, list) in ath_dynack_compute_ackto() 329 list_add_tail(&an->list, &da->nodes); in ath_dynack_node_init() 370 list_for_each_entry(an, &da->nodes, list) in ath_dynack_reset() 392 INIT_LIST_HEAD(&da->nodes); in ath_dynack_init()
|
/kernel/linux/linux-5.10/fs/btrfs/ |
H A D | xattr.c | 48 leaf = path->nodes[0]; in btrfs_getxattr() 141 btrfs_assert_tree_locked(path->nodes[0]); in btrfs_setxattr() 169 struct extent_buffer *leaf = path->nodes[0]; in btrfs_setxattr() 310 leaf = path->nodes[0]; in btrfs_listxattr()
|