Lines Matching defs:sbm

192 			 * sbm.mb_states.
201 } sbm;
214 /* One byte state per big block. See sbm.mb_states. */
223 * Mutex that protects the sbm.mb_count, sbm.mb_states,
224 * sbm.sb_states, bbm.bb_count, and bbm.bb_states
360 return (addr - mb_addr) / vm->sbm.sb_size;
436 const unsigned long idx = mb_id - vm->sbm.first_mb_id;
439 old_state = vm->sbm.mb_states[idx];
440 vm->sbm.mb_states[idx] = state;
442 BUG_ON(vm->sbm.mb_count[old_state] == 0);
443 vm->sbm.mb_count[old_state]--;
444 vm->sbm.mb_count[state]++;
453 const unsigned long idx = mb_id - vm->sbm.first_mb_id;
455 return vm->sbm.mb_states[idx];
463 int old_pages = PFN_UP(vm->sbm.next_mb_id - vm->sbm.first_mb_id);
464 int new_pages = PFN_UP(vm->sbm.next_mb_id - vm->sbm.first_mb_id + 1);
467 if (vm->sbm.mb_states && old_pages == new_pages)
475 if (vm->sbm.mb_states)
476 memcpy(new_array, vm->sbm.mb_states, old_pages * PAGE_SIZE);
477 vfree(vm->sbm.mb_states);
478 vm->sbm.mb_states = new_array;
485 for (_mb_id = _vm->sbm.first_mb_id; \
486 _mb_id < _vm->sbm.next_mb_id && _vm->sbm.mb_count[_state]; \
491 for (_mb_id = _vm->sbm.next_mb_id - 1; \
492 _mb_id >= _vm->sbm.first_mb_id && _vm->sbm.mb_count[_state]; \
503 return (mb_id - vm->sbm.first_mb_id) * vm->sbm.sbs_per_mb + sb_id;
517 __bitmap_set(vm->sbm.sb_states, bit, count);
531 __bitmap_clear(vm->sbm.sb_states, bit, count);
544 return test_bit(bit, vm->sbm.sb_states);
547 return find_next_zero_bit(vm->sbm.sb_states, bit + count, bit) >=
561 return find_next_bit(vm->sbm.sb_states, bit + count, bit) >=
566 * Find the first unplugged subblock. Returns vm->sbm.sbs_per_mb in case there is
574 return find_next_zero_bit(vm->sbm.sb_states,
575 bit + vm->sbm.sbs_per_mb, bit) - bit;
583 const unsigned long old_nb_mb = vm->sbm.next_mb_id - vm->sbm.first_mb_id;
584 const unsigned long old_nb_bits = old_nb_mb * vm->sbm.sbs_per_mb;
585 const unsigned long new_nb_bits = (old_nb_mb + 1) * vm->sbm.sbs_per_mb;
590 if (vm->sbm.sb_states && old_pages == new_pages)
598 if (vm->sbm.sb_states)
599 memcpy(new_bitmap, vm->sbm.sb_states, old_pages * PAGE_SIZE);
601 old_bitmap = vm->sbm.sb_states;
602 vm->sbm.sb_states = new_bitmap;
790 if (!virtio_mem_sbm_test_sb_unplugged(vm, mb_id, 0, vm->sbm.sbs_per_mb))
923 const unsigned long nr_pages = PFN_DOWN(vm->sbm.sb_size);
927 for (sb_id = 0; sb_id < vm->sbm.sbs_per_mb; sb_id++) {
931 sb_id * vm->sbm.sb_size);
939 const unsigned long nr_pages = PFN_DOWN(vm->sbm.sb_size);
943 for (sb_id = 0; sb_id < vm->sbm.sbs_per_mb; sb_id++) {
947 sb_id * vm->sbm.sb_size);
1310 order = ilog2(vm->sbm.sb_size) - PAGE_SHIFT;
1502 sb_id * vm->sbm.sb_size;
1503 const uint64_t size = count * vm->sbm.sb_size;
1520 sb_id * vm->sbm.sb_size;
1521 const uint64_t size = count * vm->sbm.sb_size;
1571 sb_id = vm->sbm.sbs_per_mb - 1;
1606 uint64_t nb_sb = vm->sbm.sbs_per_mb;
1619 if (vm->sbm.next_mb_id > vm->sbm.last_usable_mb_id)
1632 vm->sbm.mb_count[VIRTIO_MEM_SBM_MB_UNUSED]++;
1633 *mb_id = vm->sbm.next_mb_id++;
1646 const int count = min_t(int, *nb_sb, vm->sbm.sbs_per_mb);
1664 if (count == vm->sbm.sbs_per_mb)
1707 if (sb_id >= vm->sbm.sbs_per_mb)
1711 sb_id + count < vm->sbm.sbs_per_mb &&
1724 sb_id * vm->sbm.sb_size);
1725 nr_pages = PFN_DOWN(count * vm->sbm.sb_size);
1729 if (virtio_mem_sbm_test_sb_plugged(vm, mb_id, 0, vm->sbm.sbs_per_mb))
1742 uint64_t nb_sb = diff / vm->sbm.sb_size;
1922 if (!virtio_mem_sbm_test_sb_plugged(vm, mb_id, 0, vm->sbm.sbs_per_mb))
1928 if (virtio_mem_sbm_test_sb_unplugged(vm, mb_id, 0, vm->sbm.sbs_per_mb)) {
1955 const unsigned long nr_pages = PFN_DOWN(vm->sbm.sb_size) * count;
1961 sb_id * vm->sbm.sb_size);
2006 if (*nb_sb >= vm->sbm.sbs_per_mb &&
2007 virtio_mem_sbm_test_sb_plugged(vm, mb_id, 0, vm->sbm.sbs_per_mb)) {
2009 vm->sbm.sbs_per_mb);
2011 *nb_sb -= vm->sbm.sbs_per_mb;
2018 for (sb_id = vm->sbm.sbs_per_mb - 1; sb_id >= 0 && *nb_sb; sb_id--) {
2037 vm->sbm.have_unplugged_mb = 1;
2083 uint64_t nb_sb = diff / vm->sbm.sb_size;
2318 if (!vm->sbm.have_unplugged_mb)
2325 vm->sbm.have_unplugged_mb = false;
2337 vm->sbm.have_unplugged_mb = true;
2363 vm->sbm.last_usable_mb_id = virtio_mem_phys_to_mb_id(end_addr);
2365 vm->sbm.last_usable_mb_id--;
2438 if (!rc && vm->in_sbm && vm->sbm.have_unplugged_mb)
2542 vm->sbm.sb_size = sb_size;
2543 vm->sbm.sbs_per_mb = memory_block_size_bytes() /
2544 vm->sbm.sb_size;
2549 vm->sbm.first_mb_id = virtio_mem_phys_to_mb_id(addr);
2550 vm->sbm.next_mb_id = vm->sbm.first_mb_id;
2583 (unsigned long long)vm->sbm.sb_size);
2918 vfree(vm->sbm.mb_states);
2919 vfree(vm->sbm.sb_states);