/third_party/mesa3d/src/util/ |
H A D | vma.c | 59 util_vma_foreach_hole_safe(hole, heap) in util_vma_heap_finish() 60 free(hole); in util_vma_heap_finish() 68 util_vma_foreach_hole(hole, heap) { in util_vma_heap_validate() 69 assert(hole->offset > 0); in util_vma_heap_validate() 70 assert(hole->size > 0); in util_vma_heap_validate() 72 if (&hole->link == heap->holes.next) { in util_vma_heap_validate() 73 /* This must be the top-most hole. Assert that, if it overflows, it in util_vma_heap_validate() 76 assert(hole->size + hole->offset == 0 || in util_vma_heap_validate() 77 hole in util_vma_heap_validate() 95 util_vma_hole_alloc(struct util_vma_hole *hole, uint64_t offset, uint64_t size) util_vma_hole_alloc() argument 294 struct util_vma_hole *hole = calloc(1, sizeof(*hole)); util_vma_heap_free() local [all...] |
/third_party/libdrm/amdgpu/ |
H A D | amdgpu_vamgr.c | 64 struct amdgpu_bo_va_hole *hole, *tmp; in amdgpu_vamgr_deinit() local 65 LIST_FOR_EACH_ENTRY_SAFE(hole, tmp, &mgr->va_holes, list) { in amdgpu_vamgr_deinit() 66 list_del(&hole->list); in amdgpu_vamgr_deinit() 67 free(hole); in amdgpu_vamgr_deinit() 73 amdgpu_vamgr_subtract_hole(struct amdgpu_bo_va_hole *hole, uint64_t start_va, in amdgpu_vamgr_subtract_hole() argument 76 if (start_va > hole->offset && end_va - hole->offset < hole->size) { in amdgpu_vamgr_subtract_hole() 81 n->size = start_va - hole->offset; in amdgpu_vamgr_subtract_hole() 82 n->offset = hole in amdgpu_vamgr_subtract_hole() 105 struct amdgpu_bo_va_hole *hole, *n; amdgpu_vamgr_find_va() local 170 struct amdgpu_bo_va_hole *hole, *next; amdgpu_vamgr_free_va() local [all...] |
/third_party/rust/crates/regex/src/ |
H A D | compile.rs | 23 hole: Hole, 144 let mut dotstar_patch = Patch { hole: Hole::None, entry: 0 }; in compile_one() 155 self.fill(dotstar_patch.hole, patch.entry); in compile_one() 159 self.fill_to_next(patch.hole); in compile_one() 175 let mut dotstar_patch = Patch { hole: Hole::None, entry: 0 }; in compile_many() 182 self.fill_to_next(dotstar_patch.hole); in compile_many() 188 let Patch { hole, entry } = in compile_many() 190 self.fill_to_next(hole); in compile_many() 196 let Patch { hole, entry } = in compile_many() 199 self.fill_to_next(hole); in compile_many() [all...] |
/third_party/mesa3d/src/util/tests/vma/ |
H A D | vma_random_test.cpp | 118 for (const auto& hole : heap_holes) { in alloc() 120 (align_pages - (hole.start_page % align_pages)) % align_pages; in alloc() 121 assert(hole.num_pages < size_pages + hole_alignment_pages); in alloc() 130 allocation hole = *i; in alloc() local 132 assert(hole.start_page <= addr_page); in alloc() 133 assert(hole.num_pages >= size_pages + addr_page - hole.start_page); in alloc() 136 if (hole.start_page < a.start_page) { in alloc() 137 heap_holes.emplace(allocation{hole.start_page, in alloc() 138 a.start_page - hole in alloc() 219 auto& hole = *begin(heap_holes); empty() local [all...] |
/third_party/skia/src/core/ |
H A D | SkTSort.h | 22 * Sinks a hole from array[root] to leaf and then sifts the original array[root] element 121 T* hole = next; in SkTInsertionSort() local 123 *hole = std::move(*(hole - 1)); in SkTInsertionSort() 124 --hole; in SkTInsertionSort() 125 } while (left < hole && lessThan(insert, *(hole - 1))); in SkTInsertionSort() 126 *hole = std::move(insert); in SkTInsertionSort()
|
/third_party/mesa3d/src/gallium/winsys/radeon/drm/ |
H A D | radeon_drm_bo.c | 202 struct radeon_bo_va_hole *hole, *n; in radeon_bomgr_find_va() local 211 /* first look for a hole */ in radeon_bomgr_find_va() 212 LIST_FOR_EACH_ENTRY_SAFE(hole, n, &heap->holes, list) { in radeon_bomgr_find_va() 213 offset = hole->offset; in radeon_bomgr_find_va() 217 if (offset >= (hole->offset + hole->size)) { in radeon_bomgr_find_va() 220 if (!waste && hole->size == size) { in radeon_bomgr_find_va() 221 offset = hole->offset; in radeon_bomgr_find_va() 222 list_del(&hole->list); in radeon_bomgr_find_va() 223 FREE(hole); in radeon_bomgr_find_va() 287 struct radeon_bo_va_hole *hole = NULL; radeon_bomgr_free_va() local [all...] |
/third_party/python/Lib/turtledemo/ |
H A D | sorting_animate.py | 87 hole = i 88 while hole > 0 and shelf[i].size < shelf[hole - 1].size: 89 hole = hole - 1 90 shelf.insert(hole, shelf.pop(i))
|
/third_party/mesa3d/src/amd/vulkan/ |
H A D | radv_shader.c | 1369 remove_hole(struct radv_device *device, union radv_shader_arena_block *hole) in remove_hole() argument 1371 unsigned size_class = get_size_class(hole->size, false); in remove_hole() 1372 list_del(&hole->freelist); in remove_hole() 1378 add_hole(struct radv_device *device, union radv_shader_arena_block *hole) in add_hole() argument 1380 unsigned size_class = get_size_class(hole->size, false); in add_hole() 1381 list_addtail(&hole->freelist, &device->shader_free_lists[size_class]); in add_hole() 1423 /* Try to use an existing hole. Unless the shader is very large, this should only have to look in radv_alloc_shader_memory() 1432 list_for_each_entry(union radv_shader_arena_block, hole, in radv_alloc_shader_memory() 1435 if (hole->size < size) in radv_alloc_shader_memory() 1438 assert(hole in radv_alloc_shader_memory() 1471 union radv_shader_arena_block *alloc = NULL, *hole = NULL; radv_alloc_shader_memory() local 1536 union radv_shader_arena_block *hole = list_entry(head, union radv_shader_arena_block, list); get_hole() local 1548 union radv_shader_arena_block *hole = alloc; radv_free_shader_memory() local [all...] |
/third_party/node/deps/v8/src/objects/ |
H A D | ordered-hash-table.cc | 346 Object hole = ReadOnlyRoots(isolate).the_hole_value(); in Delete() local 348 table.set(index + i, hole); in Delete() 501 Object hole = ReadOnlyRoots(isolate).the_hole_value(); in DeleteEntry() local 503 table->SetEntry(entry, hole, hole, details); in DeleteEntry() 889 Object hole = ReadOnlyRoots(isolate).the_hole_value(); in Delete() local 891 table.SetDataEntry(entry.as_int(), j, hole); in Delete() 906 Object hole = ReadOnlyRoots(isolate).the_hole_value(); in DeleteEntry() local 908 table->SetEntry(entry, hole, hole, detail in DeleteEntry() [all...] |
/third_party/node/deps/openssl/openssl/ssl/statem/ |
H A D | statem_lib.c | 1762 * (creating a hole between TLSv1.3 and TLSv1.1) can still in check_for_downgrade() 2043 * @real_max: The highest version below the lowest compile time version hole 2044 * where that hole lies above at least one run-time enabled 2064 int hole; in ssl_get_min_max_version() local 2099 * (protocol version client is disabled at compile-time) is also a "hole". in ssl_get_min_max_version() 2101 * Our initial state is hole == 1, version == 0. That is, versions above in ssl_get_min_max_version() 2102 * the first version in the method table are disabled (a "hole" above in ssl_get_min_max_version() 2105 * Whenever "hole == 1", and we hit an enabled method, its version becomes in ssl_get_min_max_version() 2107 * method. We're no longer in a hole, so "hole" become in ssl_get_min_max_version() [all...] |
/third_party/openssl/ssl/statem/ |
H A D | statem_lib.c | 1727 * (creating a hole between TLSv1.3 and TLSv1.1) can still in check_for_downgrade() 2008 * @real_max: The highest version below the lowest compile time version hole 2009 * where that hole lies above at least one run-time enabled 2029 int hole; in ssl_get_min_max_version() local 2064 * (protocol version client is disabled at compile-time) is also a "hole". in ssl_get_min_max_version() 2066 * Our initial state is hole == 1, version == 0. That is, versions above in ssl_get_min_max_version() 2067 * the first version in the method table are disabled (a "hole" above in ssl_get_min_max_version() 2070 * Whenever "hole == 1", and we hit an enabled method, its version becomes in ssl_get_min_max_version() 2072 * method. We're no longer in a hole, so "hole" become in ssl_get_min_max_version() [all...] |
/third_party/rust/crates/memchr/bench/data/code/ |
H A D | rust-library.rs | 22823 // remaining hole. This is a good method. in insert_head() 22828 // hole. This method is very good. Benchmarks demonstrated slightly better in insert_head() 22834 // Intermediate state of the insertion process is always tracked by `hole`, which in insert_head() 22837 // 2. Fills the remaining hole in `v` in the end. in insert_head() 22841 // If `is_less` panics at any point during the process, `hole` will get dropped and in insert_head() 22842 // fill the hole in `v` with `tmp`, thus ensuring that `v` still holds every object it in insert_head() 22844 let mut hole = InsertionHole { src: &mut *tmp, dest: &mut v[1] }; in insert_head() variables 22852 hole.dest = &mut v[i]; in insert_head() 22854 // `hole` gets dropped and thus copies `tmp` into the remaining hole i in insert_head() 22905 let mut hole; merge() variables 44804 let mut hole = unsafe { Hole::new(&mut self.data, pos) }; sift_up() variables 44832 let mut hole = unsafe { Hole::new(&mut self.data, pos) }; sift_down_range() variables 44891 let mut hole = unsafe { Hole::new(&mut self.data, pos) }; sift_down_to_bottom() variables [all...] |