/kernel/linux/linux-5.10/drivers/gpu/drm/radeon/ |
H A D | radeon_object.c | 101 void radeon_ttm_placement_from_domain(struct radeon_bo *rbo, u32 domain) in radeon_ttm_placement_from_domain() argument 105 rbo->placement.placement = rbo->placements; in radeon_ttm_placement_from_domain() 106 rbo->placement.busy_placement = rbo->placements; in radeon_ttm_placement_from_domain() 111 if ((rbo->flags & RADEON_GEM_NO_CPU_ACCESS) && in radeon_ttm_placement_from_domain() 112 rbo->rdev->mc.visible_vram_size < rbo->rdev->mc.real_vram_size) { in radeon_ttm_placement_from_domain() 113 rbo->placements[c].fpfn = in radeon_ttm_placement_from_domain() 114 rbo in radeon_ttm_placement_from_domain() 780 struct radeon_bo *rbo; radeon_bo_move_notify() local 801 struct radeon_bo *rbo; radeon_bo_fault_reserve_notify() local [all...] |
H A D | radeon_fb.c | 113 struct radeon_bo *rbo = gem_to_radeon_bo(gobj); in radeonfb_destroy_pinned_object() local 116 ret = radeon_bo_reserve(rbo, false); in radeonfb_destroy_pinned_object() 118 radeon_bo_kunmap(rbo); in radeonfb_destroy_pinned_object() 119 radeon_bo_unpin(rbo); in radeonfb_destroy_pinned_object() 120 radeon_bo_unreserve(rbo); in radeonfb_destroy_pinned_object() 132 struct radeon_bo *rbo = NULL; in radeonfb_create_pinned_object() local 158 rbo = gem_to_radeon_bo(gobj); in radeonfb_create_pinned_object() 176 ret = radeon_bo_set_tiling_flags(rbo, in radeonfb_create_pinned_object() 184 ret = radeon_bo_reserve(rbo, false); in radeonfb_create_pinned_object() 188 ret = radeon_bo_pin_restricted(rbo, RADEON_GEM_DOMAIN_VRA in radeonfb_create_pinned_object() 221 struct radeon_bo *rbo = NULL; radeonfb_create() local [all...] |
H A D | radeon_gem.c | 151 struct radeon_bo *rbo = gem_to_radeon_bo(obj); in radeon_gem_object_open() local 152 struct radeon_device *rdev = rbo->rdev; in radeon_gem_object_open() 163 r = radeon_bo_reserve(rbo, false); in radeon_gem_object_open() 168 bo_va = radeon_vm_bo_find(vm, rbo); in radeon_gem_object_open() 170 bo_va = radeon_vm_bo_add(rdev, vm, rbo); in radeon_gem_object_open() 174 radeon_bo_unreserve(rbo); in radeon_gem_object_open() 182 struct radeon_bo *rbo = gem_to_radeon_bo(obj); in radeon_gem_object_close() local 183 struct radeon_device *rdev = rbo->rdev; in radeon_gem_object_close() 194 r = radeon_bo_reserve(rbo, true); in radeon_gem_object_close() 200 bo_va = radeon_vm_bo_find(vm, rbo); in radeon_gem_object_close() 521 struct radeon_bo *rbo; radeon_gem_get_tiling_ioctl() local 608 struct radeon_bo *rbo; radeon_gem_va_ioctl() local 784 struct radeon_bo *rbo; radeon_debugfs_gem_info() local [all...] |
H A D | radeon_ttm.c | 95 struct radeon_bo *rbo; in radeon_evict_flags() local 104 rbo = container_of(bo, struct radeon_bo, tbo); in radeon_evict_flags() 107 if (rbo->rdev->ring[radeon_copy_ring_index(rbo->rdev)].ready == false) in radeon_evict_flags() 108 radeon_ttm_placement_from_domain(rbo, RADEON_GEM_DOMAIN_CPU); in radeon_evict_flags() 109 else if (rbo->rdev->mc.visible_vram_size < rbo->rdev->mc.real_vram_size && in radeon_evict_flags() 110 bo->mem.start < (rbo->rdev->mc.visible_vram_size >> PAGE_SHIFT)) { in radeon_evict_flags() 111 unsigned fpfn = rbo->rdev->mc.visible_vram_size >> PAGE_SHIFT; in radeon_evict_flags() 119 radeon_ttm_placement_from_domain(rbo, RADEON_GEM_DOMAIN_VRA in radeon_evict_flags() 144 struct radeon_bo *rbo = container_of(bo, struct radeon_bo, tbo); radeon_verify_access() local 303 struct radeon_bo *rbo; radeon_bo_move() local [all...] |
H A D | radeon_legacy_crtc.c | 382 struct radeon_bo *rbo; in radeon_crtc_do_set_base() local 425 rbo = gem_to_radeon_bo(obj); in radeon_crtc_do_set_base() 427 r = radeon_bo_reserve(rbo, false); in radeon_crtc_do_set_base() 431 r = radeon_bo_pin_restricted(rbo, RADEON_GEM_DOMAIN_VRAM, 1 << 27, in radeon_crtc_do_set_base() 434 radeon_bo_unreserve(rbo); in radeon_crtc_do_set_base() 454 nsize = radeon_bo_size(rbo); in radeon_crtc_do_set_base() 464 radeon_bo_get_tiling_flags(rbo, &tiling_flags, NULL); in radeon_crtc_do_set_base() 465 radeon_bo_unreserve(rbo); in radeon_crtc_do_set_base() 559 rbo = gem_to_radeon_bo(fb->obj[0]); in radeon_crtc_do_set_base() 560 r = radeon_bo_reserve(rbo, fals in radeon_crtc_do_set_base() 1093 struct radeon_bo *rbo; radeon_crtc_disable() local [all...] |
H A D | radeon_uvd.c | 302 void radeon_uvd_force_into_uvd_segment(struct radeon_bo *rbo, in radeon_uvd_force_into_uvd_segment() argument 307 for (i = 0; i < rbo->placement.num_placement; ++i) { in radeon_uvd_force_into_uvd_segment() 308 rbo->placements[i].fpfn = 0 >> PAGE_SHIFT; in radeon_uvd_force_into_uvd_segment() 309 rbo->placements[i].lpfn = (256 * 1024 * 1024) >> PAGE_SHIFT; in radeon_uvd_force_into_uvd_segment() 317 if (rbo->placement.num_placement > 1) in radeon_uvd_force_into_uvd_segment() 321 rbo->placements[1] = rbo->placements[0]; in radeon_uvd_force_into_uvd_segment() 322 rbo->placements[1].fpfn += (256 * 1024 * 1024) >> PAGE_SHIFT; in radeon_uvd_force_into_uvd_segment() 323 rbo->placements[1].lpfn += (256 * 1024 * 1024) >> PAGE_SHIFT; in radeon_uvd_force_into_uvd_segment() 324 rbo in radeon_uvd_force_into_uvd_segment() [all...] |
H A D | atombios_crtc.c | 1152 struct radeon_bo *rbo; in dce4_crtc_do_set_base() local 1177 rbo = gem_to_radeon_bo(obj); in dce4_crtc_do_set_base() 1178 r = radeon_bo_reserve(rbo, false); in dce4_crtc_do_set_base() 1183 fb_location = radeon_bo_gpu_offset(rbo); in dce4_crtc_do_set_base() 1185 r = radeon_bo_pin(rbo, RADEON_GEM_DOMAIN_VRAM, &fb_location); in dce4_crtc_do_set_base() 1187 radeon_bo_unreserve(rbo); in dce4_crtc_do_set_base() 1192 radeon_bo_get_tiling_flags(rbo, &tiling_flags, NULL); in dce4_crtc_do_set_base() 1193 radeon_bo_unreserve(rbo); in dce4_crtc_do_set_base() 1451 rbo = gem_to_radeon_bo(fb->obj[0]); in dce4_crtc_do_set_base() 1452 r = radeon_bo_reserve(rbo, fals in dce4_crtc_do_set_base() 1473 struct radeon_bo *rbo; avivo_crtc_do_set_base() local 2167 struct radeon_bo *rbo; atombios_crtc_disable() local [all...] |
H A D | radeon.h | 1694 void radeon_uvd_force_into_uvd_segment(struct radeon_bo *rbo, 2817 extern void radeon_ttm_placement_from_domain(struct radeon_bo *rbo, u32 domain);
|
/kernel/linux/linux-6.6/drivers/gpu/drm/radeon/ |
H A D | radeon_object.c | 76 void radeon_ttm_placement_from_domain(struct radeon_bo *rbo, u32 domain) in radeon_ttm_placement_from_domain() argument 80 rbo->placement.placement = rbo->placements; in radeon_ttm_placement_from_domain() 81 rbo->placement.busy_placement = rbo->placements; in radeon_ttm_placement_from_domain() 86 if ((rbo->flags & RADEON_GEM_NO_CPU_ACCESS) && in radeon_ttm_placement_from_domain() 87 rbo->rdev->mc.visible_vram_size < rbo->rdev->mc.real_vram_size) { in radeon_ttm_placement_from_domain() 88 rbo->placements[c].fpfn = in radeon_ttm_placement_from_domain() 89 rbo in radeon_ttm_placement_from_domain() 714 struct radeon_bo *rbo; radeon_bo_move_notify() local 728 struct radeon_bo *rbo; radeon_bo_fault_reserve_notify() local [all...] |
H A D | radeon_fbdev.c | 43 struct radeon_bo *rbo = gem_to_radeon_bo(gobj); in radeon_fbdev_destroy_pinned_object() local 46 ret = radeon_bo_reserve(rbo, false); in radeon_fbdev_destroy_pinned_object() 48 radeon_bo_kunmap(rbo); in radeon_fbdev_destroy_pinned_object() 49 radeon_bo_unpin(rbo); in radeon_fbdev_destroy_pinned_object() 50 radeon_bo_unreserve(rbo); in radeon_fbdev_destroy_pinned_object() 62 struct radeon_bo *rbo = NULL; in radeon_fbdev_create_pinned_object() local 88 rbo = gem_to_radeon_bo(gobj); in radeon_fbdev_create_pinned_object() 107 ret = radeon_bo_set_tiling_flags(rbo, in radeon_fbdev_create_pinned_object() 114 ret = radeon_bo_reserve(rbo, false); in radeon_fbdev_create_pinned_object() 118 ret = radeon_bo_pin_restricted(rbo, RADEON_GEM_DOMAIN_VRA in radeon_fbdev_create_pinned_object() 212 struct radeon_bo *rbo; radeon_fbdev_fb_helper_fb_probe() local [all...] |
H A D | radeon_gem.c | 200 struct radeon_bo *rbo = gem_to_radeon_bo(obj); in radeon_gem_object_open() local 201 struct radeon_device *rdev = rbo->rdev; in radeon_gem_object_open() 212 r = radeon_bo_reserve(rbo, false); in radeon_gem_object_open() 217 bo_va = radeon_vm_bo_find(vm, rbo); in radeon_gem_object_open() 219 bo_va = radeon_vm_bo_add(rdev, vm, rbo); in radeon_gem_object_open() 223 radeon_bo_unreserve(rbo); in radeon_gem_object_open() 231 struct radeon_bo *rbo = gem_to_radeon_bo(obj); in radeon_gem_object_close() local 232 struct radeon_device *rdev = rbo->rdev; in radeon_gem_object_close() 243 r = radeon_bo_reserve(rbo, true); in radeon_gem_object_close() 249 bo_va = radeon_vm_bo_find(vm, rbo); in radeon_gem_object_close() 580 struct radeon_bo *rbo; radeon_gem_get_tiling_ioctl() local 667 struct radeon_bo *rbo; radeon_gem_va_ioctl() local 865 struct radeon_bo *rbo; radeon_debugfs_gem_info_show() local [all...] |
H A D | radeon_ttm.c | 91 struct radeon_bo *rbo; in radeon_evict_flags() local 100 rbo = container_of(bo, struct radeon_bo, tbo); in radeon_evict_flags() 103 if (rbo->rdev->ring[radeon_copy_ring_index(rbo->rdev)].ready == false) in radeon_evict_flags() 104 radeon_ttm_placement_from_domain(rbo, RADEON_GEM_DOMAIN_CPU); in radeon_evict_flags() 105 else if (rbo->rdev->mc.visible_vram_size < rbo->rdev->mc.real_vram_size && in radeon_evict_flags() 106 bo->resource->start < (rbo->rdev->mc.visible_vram_size >> PAGE_SHIFT)) { in radeon_evict_flags() 107 unsigned fpfn = rbo->rdev->mc.visible_vram_size >> PAGE_SHIFT; in radeon_evict_flags() 115 radeon_ttm_placement_from_domain(rbo, RADEON_GEM_DOMAIN_VRA in radeon_evict_flags() 200 struct radeon_bo *rbo; radeon_bo_move() local 493 struct radeon_bo *rbo; radeon_ttm_tt_create() local [all...] |
H A D | radeon_legacy_crtc.c | 382 struct radeon_bo *rbo; in radeon_crtc_do_set_base() local 425 rbo = gem_to_radeon_bo(obj); in radeon_crtc_do_set_base() 427 r = radeon_bo_reserve(rbo, false); in radeon_crtc_do_set_base() 431 r = radeon_bo_pin_restricted(rbo, RADEON_GEM_DOMAIN_VRAM, 1 << 27, in radeon_crtc_do_set_base() 434 radeon_bo_unreserve(rbo); in radeon_crtc_do_set_base() 454 nsize = radeon_bo_size(rbo); in radeon_crtc_do_set_base() 464 radeon_bo_get_tiling_flags(rbo, &tiling_flags, NULL); in radeon_crtc_do_set_base() 465 radeon_bo_unreserve(rbo); in radeon_crtc_do_set_base() 559 rbo = gem_to_radeon_bo(fb->obj[0]); in radeon_crtc_do_set_base() 560 r = radeon_bo_reserve(rbo, fals in radeon_crtc_do_set_base() 1093 struct radeon_bo *rbo; radeon_crtc_disable() local [all...] |
H A D | radeon_uvd.c | 304 void radeon_uvd_force_into_uvd_segment(struct radeon_bo *rbo, in radeon_uvd_force_into_uvd_segment() argument 309 for (i = 0; i < rbo->placement.num_placement; ++i) { in radeon_uvd_force_into_uvd_segment() 310 rbo->placements[i].fpfn = 0 >> PAGE_SHIFT; in radeon_uvd_force_into_uvd_segment() 311 rbo->placements[i].lpfn = (256 * 1024 * 1024) >> PAGE_SHIFT; in radeon_uvd_force_into_uvd_segment() 319 if (rbo->placement.num_placement > 1) in radeon_uvd_force_into_uvd_segment() 323 rbo->placements[1] = rbo->placements[0]; in radeon_uvd_force_into_uvd_segment() 324 rbo->placements[1].fpfn += (256 * 1024 * 1024) >> PAGE_SHIFT; in radeon_uvd_force_into_uvd_segment() 325 rbo->placements[1].lpfn += (256 * 1024 * 1024) >> PAGE_SHIFT; in radeon_uvd_force_into_uvd_segment() 326 rbo in radeon_uvd_force_into_uvd_segment() [all...] |
H A D | atombios_crtc.c | 1143 struct radeon_bo *rbo; in dce4_crtc_do_set_base() local 1167 rbo = gem_to_radeon_bo(obj); in dce4_crtc_do_set_base() 1168 r = radeon_bo_reserve(rbo, false); in dce4_crtc_do_set_base() 1173 fb_location = radeon_bo_gpu_offset(rbo); in dce4_crtc_do_set_base() 1175 r = radeon_bo_pin(rbo, RADEON_GEM_DOMAIN_VRAM, &fb_location); in dce4_crtc_do_set_base() 1177 radeon_bo_unreserve(rbo); in dce4_crtc_do_set_base() 1182 radeon_bo_get_tiling_flags(rbo, &tiling_flags, NULL); in dce4_crtc_do_set_base() 1183 radeon_bo_unreserve(rbo); in dce4_crtc_do_set_base() 1441 rbo = gem_to_radeon_bo(fb->obj[0]); in dce4_crtc_do_set_base() 1442 r = radeon_bo_reserve(rbo, fals in dce4_crtc_do_set_base() 1463 struct radeon_bo *rbo; avivo_crtc_do_set_base() local 2155 struct radeon_bo *rbo; atombios_crtc_disable() local [all...] |
H A D | radeon.h | 1687 void radeon_uvd_force_into_uvd_segment(struct radeon_bo *rbo, 2803 extern void radeon_ttm_placement_from_domain(struct radeon_bo *rbo, u32 domain);
|
/kernel/linux/linux-6.6/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_vkms.c | 307 struct amdgpu_bo *rbo; in amdgpu_vkms_prepare_fb() local 317 rbo = gem_to_amdgpu_bo(obj); in amdgpu_vkms_prepare_fb() 318 adev = amdgpu_ttm_adev(rbo->tbo.bdev); in amdgpu_vkms_prepare_fb() 320 r = amdgpu_bo_reserve(rbo, true); in amdgpu_vkms_prepare_fb() 326 r = dma_resv_reserve_fences(rbo->tbo.base.resv, 1); in amdgpu_vkms_prepare_fb() 333 domain = amdgpu_display_supported_domains(adev, rbo->flags); in amdgpu_vkms_prepare_fb() 337 r = amdgpu_bo_pin(rbo, domain); in amdgpu_vkms_prepare_fb() 344 r = amdgpu_ttm_alloc_gart(&rbo->tbo); in amdgpu_vkms_prepare_fb() 346 DRM_ERROR("%p bind failed\n", rbo); in amdgpu_vkms_prepare_fb() 350 amdgpu_bo_unreserve(rbo); in amdgpu_vkms_prepare_fb() 369 struct amdgpu_bo *rbo; amdgpu_vkms_cleanup_fb() local [all...] |
H A D | amdgpu_display.c | 680 struct amdgpu_bo *rbo; in extract_render_dcc_offset() local 685 rbo = gem_to_amdgpu_bo(obj); in extract_render_dcc_offset() 686 r = amdgpu_bo_reserve(rbo, false); in extract_render_dcc_offset() 695 r = amdgpu_bo_get_metadata(rbo, metadata, sizeof(metadata), &size, NULL); in extract_render_dcc_offset() 696 amdgpu_bo_unreserve(rbo); in extract_render_dcc_offset() 1104 struct amdgpu_bo *rbo; in amdgpu_display_get_fb_info() local 1113 rbo = gem_to_amdgpu_bo(amdgpu_fb->base.obj[0]); in amdgpu_display_get_fb_info() 1114 r = amdgpu_bo_reserve(rbo, false); in amdgpu_display_get_fb_info() 1124 amdgpu_bo_get_tiling_flags(rbo, tiling_flags); in amdgpu_display_get_fb_info() 1127 *tmz_surface = amdgpu_bo_encrypted(rbo); in amdgpu_display_get_fb_info() [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/amd/display/amdgpu_dm/ |
H A D | amdgpu_dm_plane.c | 840 struct amdgpu_bo *rbo; in dm_plane_helper_prepare_fb() local 852 rbo = gem_to_amdgpu_bo(obj); in dm_plane_helper_prepare_fb() 853 adev = amdgpu_ttm_adev(rbo->tbo.bdev); in dm_plane_helper_prepare_fb() 855 r = amdgpu_bo_reserve(rbo, true); in dm_plane_helper_prepare_fb() 861 r = dma_resv_reserve_fences(rbo->tbo.base.resv, 1); in dm_plane_helper_prepare_fb() 868 domain = amdgpu_display_supported_domains(adev, rbo->flags); in dm_plane_helper_prepare_fb() 872 r = amdgpu_bo_pin(rbo, domain); in dm_plane_helper_prepare_fb() 879 r = amdgpu_ttm_alloc_gart(&rbo->tbo); in dm_plane_helper_prepare_fb() 881 DRM_ERROR("%p bind failed\n", rbo); in dm_plane_helper_prepare_fb() 889 amdgpu_bo_unreserve(rbo); in dm_plane_helper_prepare_fb() 933 struct amdgpu_bo *rbo; dm_plane_helper_cleanup_fb() local [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/qxl/ |
H A D | qxl_release.c | 323 struct qxl_bo **rbo) in qxl_alloc_release_reserved() 349 if (rbo) in qxl_alloc_release_reserved() 350 *rbo = NULL; in qxl_alloc_release_reserved() 375 if (rbo) in qxl_alloc_release_reserved() 376 *rbo = bo; in qxl_alloc_release_reserved() 321 qxl_alloc_release_reserved(struct qxl_device *qdev, unsigned long size, int type, struct qxl_release **release, struct qxl_bo **rbo) qxl_alloc_release_reserved() argument
|
H A D | qxl_drv.h | 405 struct qxl_bo **rbo);
|
/kernel/linux/linux-6.6/drivers/gpu/drm/qxl/ |
H A D | qxl_release.c | 284 struct qxl_bo **rbo) in qxl_alloc_release_reserved() 310 if (rbo) in qxl_alloc_release_reserved() 311 *rbo = NULL; in qxl_alloc_release_reserved() 341 if (rbo) in qxl_alloc_release_reserved() 342 *rbo = bo; in qxl_alloc_release_reserved() 282 qxl_alloc_release_reserved(struct qxl_device *qdev, unsigned long size, int type, struct qxl_release **release, struct qxl_bo **rbo) qxl_alloc_release_reserved() argument
|
H A D | qxl_drv.h | 382 struct qxl_bo **rbo);
|
/kernel/linux/linux-5.10/drivers/gpu/drm/amd/display/amdgpu_dm/ |
H A D | amdgpu_dm.c | 3810 struct amdgpu_bo *rbo; 3819 rbo = gem_to_amdgpu_bo(amdgpu_fb->base.obj[0]); 3820 r = amdgpu_bo_reserve(rbo, false); 3830 amdgpu_bo_get_tiling_flags(rbo, tiling_flags); 3833 *tmz_surface = amdgpu_bo_encrypted(rbo); 3835 amdgpu_bo_unreserve(rbo); 5900 struct amdgpu_bo *rbo; 5915 rbo = gem_to_amdgpu_bo(obj); 5916 adev = amdgpu_ttm_adev(rbo->tbo.bdev); 5919 tv.bo = &rbo [all...] |