Lines Matching refs:base
24 struct address_space *mapping = bo->base.base.filp->f_mapping;
25 struct device *dev = bo->base.base.dev->dev;
32 if (bo->heap_size >= bo->base.base.size)
35 new_size = min(new_size, bo->base.base.size);
37 dma_resv_lock(bo->base.base.resv, NULL);
39 if (bo->base.pages) {
40 pages = bo->base.pages;
42 pages = kvmalloc_array(bo->base.base.size >> PAGE_SHIFT,
45 dma_resv_unlock(bo->base.base.resv);
49 bo->base.pages = pages;
50 bo->base.pages_use_count = 1;
59 dma_resv_unlock(bo->base.base.resv);
65 dma_resv_unlock(bo->base.base.resv);
72 if (bo->base.sgt) {
73 dma_unmap_sgtable(dev, bo->base.sgt, DMA_BIDIRECTIONAL, 0);
74 sg_free_table(bo->base.sgt);
76 bo->base.sgt = kmalloc(sizeof(*bo->base.sgt), GFP_KERNEL);
77 if (!bo->base.sgt) {
87 *bo->base.sgt = sgt;
101 kfree(bo->base.sgt);
102 bo->base.sgt = NULL;
122 obj = &shmem->base;
160 drm_gem_shmem_free(&bo->base);
188 return drm_gem_shmem_pin(&bo->base);
198 return drm_gem_shmem_vmap(&bo->base, map);
208 return drm_gem_shmem_mmap(&bo->base, vma);
235 bo->base.map_wc = true;
236 bo->base.base.funcs = &lima_gem_funcs;
238 return &bo->base.base;
275 return drm_sched_job_add_implicit_dependencies(&task->base,
276 &bo->base.base,
288 err = drm_sched_job_add_syncobj_dependency(&submit->task->base, file,
373 drm_gem_object_put(&bos[i]->base.base);
394 drm_gem_object_put(&bos[i]->base.base);