Lines Matching refs:bucket

153    /** Size of this bucket, in bytes. */
310 * This function finds the correct bucket fit for the input size.
387 * This uses either the bucket allocator for the given size, or the large
867 struct bo_cache_bucket *bucket,
874 if (!bucket)
881 list_for_each_entry_safe(struct iris_bo, cur, &bucket->head, head) {
1067 struct bo_cache_bucket *bucket = bucket_for_size(bufmgr, size, heap);
1077 /* Round the size up to the bucket size, or if we don't have caching
1081 bucket ? bucket->size : MAX2(ALIGN(size, page_size), page_size);
1101 bo = alloc_bo_from_cache(bufmgr, bucket, alignment, memzone, mmap_mode,
1106 bo = alloc_bo_from_cache(bufmgr, bucket, alignment, memzone, mmap_mode,
1129 bo->real.reusable = bucket && bufmgr->bo_reuse;
1397 struct bo_cache_bucket *bucket = &bufmgr->cache_bucket[i];
1399 list_for_each_entry_safe(struct iris_bo, bo, &bucket->head, head) {
1410 struct bo_cache_bucket *bucket = &bufmgr->local_cache_bucket[i];
1412 list_for_each_entry_safe(struct iris_bo, bo, &bucket->head, head) {
1423 struct bo_cache_bucket *bucket = &bufmgr->local_preferred_cache_bucket[i];
1425 list_for_each_entry_safe(struct iris_bo, bo, &bucket->head, head) {
1453 struct bo_cache_bucket *bucket;
1459 bucket = NULL;
1461 bucket = bucket_for_size(bufmgr, bo->size, bo->real.heap);
1463 if (bucket && iris_bo_madvise(bo, I915_MADV_DONTNEED)) {
1467 list_addtail(&bo->head, &bucket->head);
1764 struct bo_cache_bucket *bucket = &bufmgr->cache_bucket[i];
1766 list_for_each_entry_safe(struct iris_bo, bo, &bucket->head, head) {
1774 struct bo_cache_bucket *bucket = &bufmgr->local_cache_bucket[i];
1776 list_for_each_entry_safe(struct iris_bo, bo, &bucket->head, head) {
1784 struct bo_cache_bucket *bucket = &bufmgr->local_preferred_cache_bucket[i];
1786 list_for_each_entry_safe(struct iris_bo, bo, &bucket->head, head) {