Lines Matching refs:mi
126 struct numa_meminfo *mi)
139 if (mi->nr_blks >= NR_NODE_MEMBLKS) {
144 mi->blk[mi->nr_blks].start = start;
145 mi->blk[mi->nr_blks].end = end;
146 mi->blk[mi->nr_blks].nid = nid;
147 mi->nr_blks++;
154 * @mi: numa_meminfo to remove memblk from
156 * Remove @idx'th numa_memblk from @mi by shifting @mi->blk[] and
157 * decrementing @mi->nr_blks.
159 void __init numa_remove_memblk_from(int idx, struct numa_meminfo *mi)
161 mi->nr_blks--;
162 memmove(&mi->blk[idx], &mi->blk[idx + 1],
163 (mi->nr_blks - idx) * sizeof(mi->blk[0]));
230 * @mi: numa_meminfo to clean up
232 * Sanitize @mi by merging and removing unnecessary memblks. Also check for
238 int __init numa_cleanup_meminfo(struct numa_meminfo *mi)
245 for (i = 0; i < mi->nr_blks; i++) {
246 struct numa_memblk *bi = &mi->blk[i];
251 numa_move_tail_memblk(&numa_reserved_meminfo, i--, mi);
267 numa_remove_memblk_from(i--, mi);
271 for (i = 0; i < mi->nr_blks; i++) {
272 struct numa_memblk *bi = &mi->blk[i];
274 for (j = i + 1; j < mi->nr_blks; j++) {
275 struct numa_memblk *bj = &mi->blk[j];
304 for (k = 0; k < mi->nr_blks; k++) {
305 struct numa_memblk *bk = &mi->blk[k];
312 if (k < mi->nr_blks)
319 numa_remove_memblk_from(j--, mi);
324 for (i = mi->nr_blks; i < ARRAY_SIZE(mi->blk); i++) {
325 mi->blk[i].start = mi->blk[i].end = 0;
326 mi->blk[i].nid = NUMA_NO_NODE;
333 * Set nodes, which have memory in @mi, in *@nodemask.
336 const struct numa_meminfo *mi)
340 for (i = 0; i < ARRAY_SIZE(mi->blk); i++)
341 if (mi->blk[i].start != mi->blk[i].end &&
342 mi->blk[i].nid != NUMA_NO_NODE)
343 node_set(mi->blk[i].nid, *nodemask);
455 static bool __init numa_meminfo_cover_memory(const struct numa_meminfo *mi)
461 for (i = 0; i < mi->nr_blks; i++) {
462 u64 s = mi->blk[i].start >> PAGE_SHIFT;
463 u64 e = mi->blk[i].end >> PAGE_SHIFT;
465 numaram -= __absent_pages_in_range(mi->blk[i].nid, s, e);
548 static int __init numa_register_memblks(struct numa_meminfo *mi)
554 numa_nodemask_from_meminfo(&node_possible_map, mi);
558 for (i = 0; i < mi->nr_blks; i++) {
559 struct numa_memblk *mb = &mi->blk[i];
587 if (!numa_meminfo_cover_memory(mi))
595 for (i = 0; i < mi->nr_blks; i++) {
596 if (nid != mi->blk[i].nid)
598 start = min(mi->blk[i].start, start);
599 end = max(mi->blk[i].end, end);
917 static int meminfo_to_nid(struct numa_meminfo *mi, u64 start)
921 for (i = 0; i < mi->nr_blks; i++)
922 if (mi->blk[i].start <= start && mi->blk[i].end > start)
923 return mi->blk[i].nid;