/kernel/linux/linux-6.6/drivers/gpu/drm/radeon/ |
H A D | radeon_ib.c | 51 * @ib: IB object returned 59 struct radeon_ib *ib, struct radeon_vm *vm, in radeon_ib_get() 64 r = radeon_sa_bo_new(&rdev->ring_tmp_bo, &ib->sa_bo, size, 256); in radeon_ib_get() 70 radeon_sync_create(&ib->sync); in radeon_ib_get() 72 ib->ring = ring; in radeon_ib_get() 73 ib->fence = NULL; in radeon_ib_get() 74 ib->ptr = radeon_sa_bo_cpu_addr(ib->sa_bo); in radeon_ib_get() 75 ib->vm = vm; in radeon_ib_get() 77 /* ib poo in radeon_ib_get() 58 radeon_ib_get(struct radeon_device *rdev, int ring, struct radeon_ib *ib, struct radeon_vm *vm, unsigned size) radeon_ib_get() argument 97 radeon_ib_free(struct radeon_device *rdev, struct radeon_ib *ib) radeon_ib_free() argument 125 radeon_ib_schedule(struct radeon_device *rdev, struct radeon_ib *ib, struct radeon_ib *const_ib, bool hdp_flush) radeon_ib_schedule() argument [all...] |
H A D | si_dma.c | 61 * @ib: indirect buffer to fill with commands 69 struct radeon_ib *ib, in si_dma_vm_copy_pages() 78 ib->ptr[ib->length_dw++] = DMA_PACKET(DMA_PACKET_COPY, in si_dma_vm_copy_pages() 80 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in si_dma_vm_copy_pages() 81 ib->ptr[ib->length_dw++] = lower_32_bits(src); in si_dma_vm_copy_pages() 82 ib->ptr[ib in si_dma_vm_copy_pages() 68 si_dma_vm_copy_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t src, unsigned count) si_dma_vm_copy_pages() argument 104 si_dma_vm_write_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint32_t flags) si_dma_vm_write_pages() argument 151 si_dma_vm_set_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint32_t flags) si_dma_vm_set_pages() argument [all...] |
H A D | radeon_vce.c | 349 struct radeon_ib ib; in radeon_vce_get_create_msg() local 353 r = radeon_ib_get(rdev, ring, &ib, NULL, ib_size_dw * 4); in radeon_vce_get_create_msg() 355 DRM_ERROR("radeon: failed to get ib (%d).\n", r); in radeon_vce_get_create_msg() 359 dummy = ib.gpu_addr + 1024; in radeon_vce_get_create_msg() 362 ib.length_dw = 0; in radeon_vce_get_create_msg() 363 ib.ptr[ib.length_dw++] = cpu_to_le32(0x0000000c); /* len */ in radeon_vce_get_create_msg() 364 ib.ptr[ib.length_dw++] = cpu_to_le32(0x00000001); /* session cmd */ in radeon_vce_get_create_msg() 365 ib in radeon_vce_get_create_msg() 416 struct radeon_ib ib; radeon_vce_get_destroy_msg() local 718 radeon_vce_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib) radeon_vce_ib_execute() argument [all...] |
H A D | ni_dma.c | 117 * @ib: IB object to schedule 122 struct radeon_ib *ib) in cayman_dma_ring_ib_execute() 124 struct radeon_ring *ring = &rdev->ring[ib->ring]; in cayman_dma_ring_ib_execute() 125 unsigned vm_id = ib->vm ? ib->vm->ids[ib->ring].id : 0; in cayman_dma_ring_ib_execute() 144 radeon_ring_write(ring, (ib->gpu_addr & 0xFFFFFFE0)); in cayman_dma_ring_ib_execute() 145 radeon_ring_write(ring, (ib->length_dw << 12) | (upper_32_bits(ib->gpu_addr) & 0xFF)); in cayman_dma_ring_ib_execute() 307 * @ib 121 cayman_dma_ring_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib) cayman_dma_ring_ib_execute() argument 314 cayman_dma_vm_copy_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t src, unsigned count) cayman_dma_vm_copy_pages() argument 352 cayman_dma_vm_write_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint32_t flags) cayman_dma_vm_write_pages() argument 400 cayman_dma_vm_set_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint32_t flags) cayman_dma_vm_set_pages() argument 442 cayman_dma_vm_pad_ib(struct radeon_ib *ib) cayman_dma_vm_pad_ib() argument [all...] |
H A D | cik_sdma.c | 128 * @ib: IB object to schedule 133 struct radeon_ib *ib) in cik_sdma_ring_ib_execute() 135 struct radeon_ring *ring = &rdev->ring[ib->ring]; in cik_sdma_ring_ib_execute() 136 u32 extra_bits = (ib->vm ? ib->vm->ids[ib->ring].id : 0) & 0xf; in cik_sdma_ring_ib_execute() 154 radeon_ring_write(ring, ib->gpu_addr & 0xffffffe0); /* base must be 32 byte aligned */ in cik_sdma_ring_ib_execute() 155 radeon_ring_write(ring, upper_32_bits(ib->gpu_addr)); in cik_sdma_ring_ib_execute() 156 radeon_ring_write(ring, ib->length_dw); in cik_sdma_ring_ib_execute() 703 struct radeon_ib ib; in cik_sdma_ib_test() local 132 cik_sdma_ring_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib) cik_sdma_ring_ib_execute() argument 802 cik_sdma_vm_copy_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t src, unsigned count) cik_sdma_vm_copy_pages() argument 840 cik_sdma_vm_write_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint32_t flags) cik_sdma_vm_write_pages() argument 889 cik_sdma_vm_set_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint32_t flags) cik_sdma_vm_set_pages() argument 932 cik_sdma_vm_pad_ib(struct radeon_ib *ib) cik_sdma_vm_pad_ib() argument [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_vce.c | 442 struct amdgpu_ib *ib; in amdgpu_vce_get_create_msg() local 463 ib = &job->ibs[0]; in amdgpu_vce_get_create_msg() 468 ib->length_dw = 0; in amdgpu_vce_get_create_msg() 469 ib->ptr[ib->length_dw++] = 0x0000000c; /* len */ in amdgpu_vce_get_create_msg() 470 ib->ptr[ib->length_dw++] = 0x00000001; /* session cmd */ in amdgpu_vce_get_create_msg() 471 ib->ptr[ib->length_dw++] = handle; in amdgpu_vce_get_create_msg() 474 ib in amdgpu_vce_get_create_msg() 534 struct amdgpu_ib *ib; amdgpu_vce_get_destroy_msg() local 598 amdgpu_vce_validate_bo(struct amdgpu_cs_parser *p, struct amdgpu_ib *ib, int lo, int hi, unsigned int size, int32_t index) amdgpu_vce_validate_bo() argument 649 amdgpu_vce_cs_reloc(struct amdgpu_cs_parser *p, struct amdgpu_ib *ib, int lo, int hi, unsigned int size, uint32_t index) amdgpu_vce_cs_reloc() argument 735 amdgpu_vce_ring_parse_cs(struct amdgpu_cs_parser *p, struct amdgpu_job *job, struct amdgpu_ib *ib) amdgpu_vce_ring_parse_cs() argument 973 amdgpu_vce_ring_parse_cs_vm(struct amdgpu_cs_parser *p, struct amdgpu_job *job, struct amdgpu_ib *ib) amdgpu_vce_ring_parse_cs_vm() argument 1067 amdgpu_vce_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) amdgpu_vce_ring_emit_ib() argument [all...] |
H A D | amdgpu_vcn.c | 534 struct amdgpu_ib *ib; in amdgpu_vcn_dec_send_msg() local 543 ib = &job->ibs[0]; in amdgpu_vcn_dec_send_msg() 544 ib->ptr[0] = PACKET0(adev->vcn.internal.data0, 0); in amdgpu_vcn_dec_send_msg() 545 ib->ptr[1] = addr; in amdgpu_vcn_dec_send_msg() 546 ib->ptr[2] = PACKET0(adev->vcn.internal.data1, 0); in amdgpu_vcn_dec_send_msg() 547 ib->ptr[3] = addr >> 32; in amdgpu_vcn_dec_send_msg() 548 ib->ptr[4] = PACKET0(adev->vcn.internal.cmd, 0); in amdgpu_vcn_dec_send_msg() 549 ib->ptr[5] = 0; in amdgpu_vcn_dec_send_msg() 551 ib->ptr[i] = PACKET0(adev->vcn.internal.nop, 0); in amdgpu_vcn_dec_send_msg() 552 ib in amdgpu_vcn_dec_send_msg() 575 amdgpu_vcn_dec_get_create_msg(struct amdgpu_ring *ring, uint32_t handle, struct amdgpu_ib *ib) amdgpu_vcn_dec_get_create_msg() argument 610 amdgpu_vcn_dec_get_destroy_msg(struct amdgpu_ring *ring, uint32_t handle, struct amdgpu_ib *ib) amdgpu_vcn_dec_get_destroy_msg() argument 640 struct amdgpu_ib ib; amdgpu_vcn_dec_ring_test_ib() local 669 amdgpu_vcn_unified_ring_ib_header(struct amdgpu_ib *ib, uint32_t ib_pack_in_dw, bool enc) amdgpu_vcn_unified_ring_ib_header() argument 708 struct amdgpu_ib *ib; amdgpu_vcn_dec_sw_send_msg() local 772 struct amdgpu_ib ib; amdgpu_vcn_dec_sw_ring_test_ib() local 838 struct amdgpu_ib *ib; amdgpu_vcn_enc_get_create_msg() local 905 struct amdgpu_ib *ib; amdgpu_vcn_enc_get_destroy_msg() local 970 struct amdgpu_ib ib; amdgpu_vcn_enc_ring_test_ib() local [all...] |
H A D | si_dma.c | 64 struct amdgpu_ib *ib, in si_dma_ring_emit_ib() 74 amdgpu_ring_write(ring, (ib->gpu_addr & 0xFFFFFFE0)); in si_dma_ring_emit_ib() 75 amdgpu_ring_write(ring, (ib->length_dw << 12) | (upper_32_bits(ib->gpu_addr) & 0xFF)); in si_dma_ring_emit_ib() 251 struct amdgpu_ib ib; in si_dma_ring_test_ib() local 265 memset(&ib, 0, sizeof(ib)); in si_dma_ring_test_ib() 267 AMDGPU_IB_POOL_DIRECT, &ib); in si_dma_ring_test_ib() 271 ib.ptr[0] = DMA_PACKET(DMA_PACKET_WRITE, 0, 0, 0, 1); in si_dma_ring_test_ib() 272 ib in si_dma_ring_test_ib() 62 si_dma_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) si_dma_ring_emit_ib() argument 311 si_dma_vm_copy_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t src, unsigned count) si_dma_vm_copy_pte() argument 336 si_dma_vm_write_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t value, unsigned count, uint32_t incr) si_dma_vm_write_pte() argument 364 si_dma_vm_set_pte_pde(struct amdgpu_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint64_t flags) si_dma_vm_set_pte_pde() argument 405 si_dma_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) si_dma_ring_pad_ib() argument 775 si_dma_emit_copy_buffer(struct amdgpu_ib *ib, uint64_t src_offset, uint64_t dst_offset, uint32_t byte_count, bool tmz) si_dma_emit_copy_buffer() argument 799 si_dma_emit_fill_buffer(struct amdgpu_ib *ib, uint32_t src_data, uint64_t dst_offset, uint32_t byte_count) si_dma_emit_fill_buffer() argument [all...] |
H A D | sdma_v3_0.c | 415 * @ib: IB object to schedule 422 struct amdgpu_ib *ib, in sdma_v3_0_ring_emit_ib() 433 amdgpu_ring_write(ring, lower_32_bits(ib->gpu_addr) & 0xffffffe0); in sdma_v3_0_ring_emit_ib() 434 amdgpu_ring_write(ring, upper_32_bits(ib->gpu_addr)); in sdma_v3_0_ring_emit_ib() 435 amdgpu_ring_write(ring, ib->length_dw); in sdma_v3_0_ring_emit_ib() 863 struct amdgpu_ib ib; in sdma_v3_0_ring_test_ib() local 877 memset(&ib, 0, sizeof(ib)); in sdma_v3_0_ring_test_ib() 879 AMDGPU_IB_POOL_DIRECT, &ib); in sdma_v3_0_ring_test_ib() 883 ib in sdma_v3_0_ring_test_ib() 420 sdma_v3_0_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) sdma_v3_0_ring_emit_ib() argument 928 sdma_v3_0_vm_copy_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t src, unsigned count) sdma_v3_0_vm_copy_pte() argument 955 sdma_v3_0_vm_write_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t value, unsigned count, uint32_t incr) sdma_v3_0_vm_write_pte() argument 985 sdma_v3_0_vm_set_pte_pde(struct amdgpu_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint64_t flags) sdma_v3_0_vm_set_pte_pde() argument 1009 sdma_v3_0_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) sdma_v3_0_ring_pad_ib() argument 1631 sdma_v3_0_emit_copy_buffer(struct amdgpu_ib *ib, uint64_t src_offset, uint64_t dst_offset, uint32_t byte_count, bool tmz) sdma_v3_0_emit_copy_buffer() argument 1657 sdma_v3_0_emit_fill_buffer(struct amdgpu_ib *ib, uint32_t src_data, uint64_t dst_offset, uint32_t byte_count) sdma_v3_0_emit_fill_buffer() argument [all...] |
H A D | sdma_v2_4.c | 241 * @ib: IB object to schedule 248 struct amdgpu_ib *ib, in sdma_v2_4_ring_emit_ib() 259 amdgpu_ring_write(ring, lower_32_bits(ib->gpu_addr) & 0xffffffe0); in sdma_v2_4_ring_emit_ib() 260 amdgpu_ring_write(ring, upper_32_bits(ib->gpu_addr)); in sdma_v2_4_ring_emit_ib() 261 amdgpu_ring_write(ring, ib->length_dw); in sdma_v2_4_ring_emit_ib() 591 struct amdgpu_ib ib; in sdma_v2_4_ring_test_ib() local 605 memset(&ib, 0, sizeof(ib)); in sdma_v2_4_ring_test_ib() 607 AMDGPU_IB_POOL_DIRECT, &ib); in sdma_v2_4_ring_test_ib() 611 ib in sdma_v2_4_ring_test_ib() 246 sdma_v2_4_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) sdma_v2_4_ring_emit_ib() argument 657 sdma_v2_4_vm_copy_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t src, unsigned count) sdma_v2_4_vm_copy_pte() argument 684 sdma_v2_4_vm_write_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t value, unsigned count, uint32_t incr) sdma_v2_4_vm_write_pte() argument 714 sdma_v2_4_vm_set_pte_pde(struct amdgpu_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint64_t flags) sdma_v2_4_vm_set_pte_pde() argument 738 sdma_v2_4_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) sdma_v2_4_ring_pad_ib() argument 1193 sdma_v2_4_emit_copy_buffer(struct amdgpu_ib *ib, uint64_t src_offset, uint64_t dst_offset, uint32_t byte_count, bool tmz) sdma_v2_4_emit_copy_buffer() argument 1219 sdma_v2_4_emit_fill_buffer(struct amdgpu_ib *ib, uint32_t src_data, uint64_t dst_offset, uint32_t byte_count) sdma_v2_4_emit_fill_buffer() argument [all...] |
H A D | cik_sdma.c | 215 * @ib: IB object to schedule 222 struct amdgpu_ib *ib, in cik_sdma_ring_emit_ib() 232 amdgpu_ring_write(ring, ib->gpu_addr & 0xffffffe0); /* base must be 32 byte aligned */ in cik_sdma_ring_emit_ib() 233 amdgpu_ring_write(ring, upper_32_bits(ib->gpu_addr) & 0xffffffff); in cik_sdma_ring_emit_ib() 234 amdgpu_ring_write(ring, ib->length_dw); in cik_sdma_ring_emit_ib() 658 struct amdgpu_ib ib; in cik_sdma_ring_test_ib() local 672 memset(&ib, 0, sizeof(ib)); in cik_sdma_ring_test_ib() 674 AMDGPU_IB_POOL_DIRECT, &ib); in cik_sdma_ring_test_ib() 678 ib in cik_sdma_ring_test_ib() 220 cik_sdma_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) cik_sdma_ring_emit_ib() argument 720 cik_sdma_vm_copy_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t src, unsigned count) cik_sdma_vm_copy_pte() argument 747 cik_sdma_vm_write_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t value, unsigned count, uint32_t incr) cik_sdma_vm_write_pte() argument 777 cik_sdma_vm_set_pte_pde(struct amdgpu_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint64_t flags) cik_sdma_vm_set_pte_pde() argument 801 cik_sdma_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) cik_sdma_ring_pad_ib() argument 1305 cik_sdma_emit_copy_buffer(struct amdgpu_ib *ib, uint64_t src_offset, uint64_t dst_offset, uint32_t byte_count, bool tmz) cik_sdma_emit_copy_buffer() argument 1330 cik_sdma_emit_fill_buffer(struct amdgpu_ib *ib, uint32_t src_data, uint64_t dst_offset, uint32_t byte_count) cik_sdma_emit_fill_buffer() argument [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/radeon/ |
H A D | si_dma.c | 62 * @ib: indirect buffer to fill with commands 70 struct radeon_ib *ib, in si_dma_vm_copy_pages() 79 ib->ptr[ib->length_dw++] = DMA_PACKET(DMA_PACKET_COPY, in si_dma_vm_copy_pages() 81 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in si_dma_vm_copy_pages() 82 ib->ptr[ib->length_dw++] = lower_32_bits(src); in si_dma_vm_copy_pages() 83 ib->ptr[ib in si_dma_vm_copy_pages() 69 si_dma_vm_copy_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t src, unsigned count) si_dma_vm_copy_pages() argument 105 si_dma_vm_write_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint32_t flags) si_dma_vm_write_pages() argument 152 si_dma_vm_set_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint32_t flags) si_dma_vm_set_pages() argument [all...] |
H A D | radeon_ib.c | 51 * @ib: IB object returned 59 struct radeon_ib *ib, struct radeon_vm *vm, in radeon_ib_get() 64 r = radeon_sa_bo_new(rdev, &rdev->ring_tmp_bo, &ib->sa_bo, size, 256); in radeon_ib_get() 70 radeon_sync_create(&ib->sync); in radeon_ib_get() 72 ib->ring = ring; in radeon_ib_get() 73 ib->fence = NULL; in radeon_ib_get() 74 ib->ptr = radeon_sa_bo_cpu_addr(ib->sa_bo); in radeon_ib_get() 75 ib->vm = vm; in radeon_ib_get() 77 /* ib poo in radeon_ib_get() 58 radeon_ib_get(struct radeon_device *rdev, int ring, struct radeon_ib *ib, struct radeon_vm *vm, unsigned size) radeon_ib_get() argument 97 radeon_ib_free(struct radeon_device *rdev, struct radeon_ib *ib) radeon_ib_free() argument 125 radeon_ib_schedule(struct radeon_device *rdev, struct radeon_ib *ib, struct radeon_ib *const_ib, bool hdp_flush) radeon_ib_schedule() argument [all...] |
H A D | radeon_vce.c | 349 struct radeon_ib ib; in radeon_vce_get_create_msg() local 353 r = radeon_ib_get(rdev, ring, &ib, NULL, ib_size_dw * 4); in radeon_vce_get_create_msg() 355 DRM_ERROR("radeon: failed to get ib (%d).\n", r); in radeon_vce_get_create_msg() 359 dummy = ib.gpu_addr + 1024; in radeon_vce_get_create_msg() 362 ib.length_dw = 0; in radeon_vce_get_create_msg() 363 ib.ptr[ib.length_dw++] = cpu_to_le32(0x0000000c); /* len */ in radeon_vce_get_create_msg() 364 ib.ptr[ib.length_dw++] = cpu_to_le32(0x00000001); /* session cmd */ in radeon_vce_get_create_msg() 365 ib in radeon_vce_get_create_msg() 416 struct radeon_ib ib; radeon_vce_get_destroy_msg() local 718 radeon_vce_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib) radeon_vce_ib_execute() argument [all...] |
H A D | ni_dma.c | 118 * @ib: IB object to schedule 123 struct radeon_ib *ib) in cayman_dma_ring_ib_execute() 125 struct radeon_ring *ring = &rdev->ring[ib->ring]; in cayman_dma_ring_ib_execute() 126 unsigned vm_id = ib->vm ? ib->vm->ids[ib->ring].id : 0; in cayman_dma_ring_ib_execute() 145 radeon_ring_write(ring, (ib->gpu_addr & 0xFFFFFFE0)); in cayman_dma_ring_ib_execute() 146 radeon_ring_write(ring, (ib->length_dw << 12) | (upper_32_bits(ib->gpu_addr) & 0xFF)); in cayman_dma_ring_ib_execute() 308 * @ib 122 cayman_dma_ring_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib) cayman_dma_ring_ib_execute() argument 315 cayman_dma_vm_copy_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t src, unsigned count) cayman_dma_vm_copy_pages() argument 353 cayman_dma_vm_write_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint32_t flags) cayman_dma_vm_write_pages() argument 401 cayman_dma_vm_set_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint32_t flags) cayman_dma_vm_set_pages() argument 443 cayman_dma_vm_pad_ib(struct radeon_ib *ib) cayman_dma_vm_pad_ib() argument [all...] |
H A D | cik_sdma.c | 129 * @ib: IB object to schedule 134 struct radeon_ib *ib) in cik_sdma_ring_ib_execute() 136 struct radeon_ring *ring = &rdev->ring[ib->ring]; in cik_sdma_ring_ib_execute() 137 u32 extra_bits = (ib->vm ? ib->vm->ids[ib->ring].id : 0) & 0xf; in cik_sdma_ring_ib_execute() 155 radeon_ring_write(ring, ib->gpu_addr & 0xffffffe0); /* base must be 32 byte aligned */ in cik_sdma_ring_ib_execute() 156 radeon_ring_write(ring, upper_32_bits(ib->gpu_addr)); in cik_sdma_ring_ib_execute() 157 radeon_ring_write(ring, ib->length_dw); in cik_sdma_ring_ib_execute() 704 struct radeon_ib ib; in cik_sdma_ib_test() local 133 cik_sdma_ring_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib) cik_sdma_ring_ib_execute() argument 803 cik_sdma_vm_copy_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t src, unsigned count) cik_sdma_vm_copy_pages() argument 841 cik_sdma_vm_write_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint32_t flags) cik_sdma_vm_write_pages() argument 890 cik_sdma_vm_set_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint32_t flags) cik_sdma_vm_set_pages() argument 933 cik_sdma_vm_pad_ib(struct radeon_ib *ib) cik_sdma_vm_pad_ib() argument [all...] |
/kernel/linux/linux-5.10/arch/s390/include/asm/ |
H A D | idals.h | 120 struct idal_buffer *ib; in idal_buffer_alloc() local 125 ib = kmalloc(struct_size(ib, data, nr_ptrs), GFP_DMA | GFP_KERNEL); in idal_buffer_alloc() 126 if (ib == NULL) in idal_buffer_alloc() 128 ib->size = size; in idal_buffer_alloc() 129 ib->page_order = page_order; in idal_buffer_alloc() 132 ib->data[i] = ib->data[i-1] + IDA_BLOCK_SIZE; in idal_buffer_alloc() 135 ib->data[i] = (void *) in idal_buffer_alloc() 137 if (ib in idal_buffer_alloc() 155 idal_buffer_free(struct idal_buffer *ib) idal_buffer_free() argument 170 __idal_buffer_is_needed(struct idal_buffer *ib) __idal_buffer_is_needed() argument 180 idal_buffer_set_cda(struct idal_buffer *ib, struct ccw1 *ccw) idal_buffer_set_cda() argument 196 idal_buffer_to_user(struct idal_buffer *ib, void __user *to, size_t count) idal_buffer_to_user() argument 216 idal_buffer_from_user(struct idal_buffer *ib, const void __user *from, size_t count) idal_buffer_from_user() argument [all...] |
/kernel/linux/linux-6.6/arch/s390/include/asm/ |
H A D | idals.h | 132 struct idal_buffer *ib; in idal_buffer_alloc() local 137 ib = kmalloc(struct_size(ib, data, nr_ptrs), GFP_DMA | GFP_KERNEL); in idal_buffer_alloc() 138 if (ib == NULL) in idal_buffer_alloc() 140 ib->size = size; in idal_buffer_alloc() 141 ib->page_order = page_order; in idal_buffer_alloc() 144 ib->data[i] = ib->data[i-1] + IDA_BLOCK_SIZE; in idal_buffer_alloc() 147 ib->data[i] = (void *) in idal_buffer_alloc() 149 if (ib in idal_buffer_alloc() 167 idal_buffer_free(struct idal_buffer *ib) idal_buffer_free() argument 182 __idal_buffer_is_needed(struct idal_buffer *ib) __idal_buffer_is_needed() argument 192 idal_buffer_set_cda(struct idal_buffer *ib, struct ccw1 *ccw) idal_buffer_set_cda() argument 208 idal_buffer_to_user(struct idal_buffer *ib, void __user *to, size_t count) idal_buffer_to_user() argument 228 idal_buffer_from_user(struct idal_buffer *ib, const void __user *from, size_t count) idal_buffer_from_user() argument [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_vce.c | 444 struct amdgpu_ib *ib; in amdgpu_vce_get_create_msg() local 454 ib = &job->ibs[0]; in amdgpu_vce_get_create_msg() 459 ib->length_dw = 0; in amdgpu_vce_get_create_msg() 460 ib->ptr[ib->length_dw++] = 0x0000000c; /* len */ in amdgpu_vce_get_create_msg() 461 ib->ptr[ib->length_dw++] = 0x00000001; /* session cmd */ in amdgpu_vce_get_create_msg() 462 ib->ptr[ib->length_dw++] = handle; in amdgpu_vce_get_create_msg() 465 ib in amdgpu_vce_get_create_msg() 524 struct amdgpu_ib *ib; amdgpu_vce_get_destroy_msg() local 721 struct amdgpu_ib *ib = &p->job->ibs[ib_idx]; amdgpu_vce_ring_parse_cs() local 957 struct amdgpu_ib *ib = &p->job->ibs[ib_idx]; amdgpu_vce_ring_parse_cs_vm() local 1046 amdgpu_vce_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) amdgpu_vce_ring_emit_ib() argument [all...] |
H A D | si_dma.c | 65 struct amdgpu_ib *ib, in si_dma_ring_emit_ib() 75 amdgpu_ring_write(ring, (ib->gpu_addr & 0xFFFFFFE0)); in si_dma_ring_emit_ib() 76 amdgpu_ring_write(ring, (ib->length_dw << 12) | (upper_32_bits(ib->gpu_addr) & 0xFF)); in si_dma_ring_emit_ib() 254 struct amdgpu_ib ib; in si_dma_ring_test_ib() local 268 memset(&ib, 0, sizeof(ib)); in si_dma_ring_test_ib() 270 AMDGPU_IB_POOL_DIRECT, &ib); in si_dma_ring_test_ib() 274 ib.ptr[0] = DMA_PACKET(DMA_PACKET_WRITE, 0, 0, 0, 1); in si_dma_ring_test_ib() 275 ib in si_dma_ring_test_ib() 63 si_dma_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) si_dma_ring_emit_ib() argument 314 si_dma_vm_copy_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t src, unsigned count) si_dma_vm_copy_pte() argument 339 si_dma_vm_write_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t value, unsigned count, uint32_t incr) si_dma_vm_write_pte() argument 367 si_dma_vm_set_pte_pde(struct amdgpu_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint64_t flags) si_dma_vm_set_pte_pde() argument 407 si_dma_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) si_dma_ring_pad_ib() argument 776 si_dma_emit_copy_buffer(struct amdgpu_ib *ib, uint64_t src_offset, uint64_t dst_offset, uint32_t byte_count, bool tmz) si_dma_emit_copy_buffer() argument 800 si_dma_emit_fill_buffer(struct amdgpu_ib *ib, uint32_t src_data, uint64_t dst_offset, uint32_t byte_count) si_dma_emit_fill_buffer() argument [all...] |
H A D | amdgpu_vcn.c | 439 struct amdgpu_ib *ib; in amdgpu_vcn_dec_send_msg() local 448 ib = &job->ibs[0]; in amdgpu_vcn_dec_send_msg() 450 ib->ptr[0] = PACKET0(adev->vcn.internal.data0, 0); in amdgpu_vcn_dec_send_msg() 451 ib->ptr[1] = addr; in amdgpu_vcn_dec_send_msg() 452 ib->ptr[2] = PACKET0(adev->vcn.internal.data1, 0); in amdgpu_vcn_dec_send_msg() 453 ib->ptr[3] = addr >> 32; in amdgpu_vcn_dec_send_msg() 454 ib->ptr[4] = PACKET0(adev->vcn.internal.cmd, 0); in amdgpu_vcn_dec_send_msg() 455 ib->ptr[5] = 0; in amdgpu_vcn_dec_send_msg() 457 ib->ptr[i] = PACKET0(adev->vcn.internal.nop, 0); in amdgpu_vcn_dec_send_msg() 458 ib in amdgpu_vcn_dec_send_msg() 606 struct amdgpu_ib *ib; amdgpu_vcn_enc_get_create_msg() local 660 struct amdgpu_ib *ib; amdgpu_vcn_enc_get_destroy_msg() local [all...] |
H A D | sdma_v2_4.c | 246 * @ib: IB object to schedule 252 struct amdgpu_ib *ib, in sdma_v2_4_ring_emit_ib() 263 amdgpu_ring_write(ring, lower_32_bits(ib->gpu_addr) & 0xffffffe0); in sdma_v2_4_ring_emit_ib() 264 amdgpu_ring_write(ring, upper_32_bits(ib->gpu_addr)); in sdma_v2_4_ring_emit_ib() 265 amdgpu_ring_write(ring, ib->length_dw); in sdma_v2_4_ring_emit_ib() 600 struct amdgpu_ib ib; in sdma_v2_4_ring_test_ib() local 614 memset(&ib, 0, sizeof(ib)); in sdma_v2_4_ring_test_ib() 616 AMDGPU_IB_POOL_DIRECT, &ib); in sdma_v2_4_ring_test_ib() 620 ib in sdma_v2_4_ring_test_ib() 250 sdma_v2_4_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) sdma_v2_4_ring_emit_ib() argument 666 sdma_v2_4_vm_copy_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t src, unsigned count) sdma_v2_4_vm_copy_pte() argument 693 sdma_v2_4_vm_write_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t value, unsigned count, uint32_t incr) sdma_v2_4_vm_write_pte() argument 723 sdma_v2_4_vm_set_pte_pde(struct amdgpu_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint64_t flags) sdma_v2_4_vm_set_pte_pde() argument 746 sdma_v2_4_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) sdma_v2_4_ring_pad_ib() argument 1200 sdma_v2_4_emit_copy_buffer(struct amdgpu_ib *ib, uint64_t src_offset, uint64_t dst_offset, uint32_t byte_count, bool tmz) sdma_v2_4_emit_copy_buffer() argument 1226 sdma_v2_4_emit_fill_buffer(struct amdgpu_ib *ib, uint32_t src_data, uint64_t dst_offset, uint32_t byte_count) sdma_v2_4_emit_fill_buffer() argument [all...] |
H A D | cik_sdma.c | 218 * @ib: IB object to schedule 224 struct amdgpu_ib *ib, in cik_sdma_ring_emit_ib() 234 amdgpu_ring_write(ring, ib->gpu_addr & 0xffffffe0); /* base must be 32 byte aligned */ in cik_sdma_ring_emit_ib() 235 amdgpu_ring_write(ring, upper_32_bits(ib->gpu_addr) & 0xffffffff); in cik_sdma_ring_emit_ib() 236 amdgpu_ring_write(ring, ib->length_dw); in cik_sdma_ring_emit_ib() 665 struct amdgpu_ib ib; in cik_sdma_ring_test_ib() local 679 memset(&ib, 0, sizeof(ib)); in cik_sdma_ring_test_ib() 681 AMDGPU_IB_POOL_DIRECT, &ib); in cik_sdma_ring_test_ib() 685 ib in cik_sdma_ring_test_ib() 222 cik_sdma_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) cik_sdma_ring_emit_ib() argument 727 cik_sdma_vm_copy_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t src, unsigned count) cik_sdma_vm_copy_pte() argument 754 cik_sdma_vm_write_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t value, unsigned count, uint32_t incr) cik_sdma_vm_write_pte() argument 784 cik_sdma_vm_set_pte_pde(struct amdgpu_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint64_t flags) cik_sdma_vm_set_pte_pde() argument 807 cik_sdma_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) cik_sdma_ring_pad_ib() argument 1310 cik_sdma_emit_copy_buffer(struct amdgpu_ib *ib, uint64_t src_offset, uint64_t dst_offset, uint32_t byte_count, bool tmz) cik_sdma_emit_copy_buffer() argument 1335 cik_sdma_emit_fill_buffer(struct amdgpu_ib *ib, uint32_t src_data, uint64_t dst_offset, uint32_t byte_count) cik_sdma_emit_fill_buffer() argument [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/amd/ |
H A D | 7990.c | 100 t, ib->brx_ring[t].rmd1_hadr, ib->brx_ring[t].rmd0, \ 101 ib->brx_ring[t].length, \ 102 ib->brx_ring[t].mblength, ib->brx_ring[t].rmd1_bits); \ 106 t, ib->btx_ring[t].tmd1_hadr, ib->btx_ring[t].tmd0, \ 107 ib->btx_ring[t].length, \ 108 ib->btx_ring[t].misc, ib 140 volatile struct lance_init_block *ib = lp->init_block; lance_init_ring() local 275 volatile struct lance_init_block *ib = lp->init_block; lance_rx() local 355 volatile struct lance_init_block *ib = lp->init_block; lance_tx() local 542 volatile struct lance_init_block *ib = lp->init_block; lance_start_xmit() local 601 volatile struct lance_init_block *ib = lp->init_block; lance_load_multicast() local 628 volatile struct lance_init_block *ib = lp->init_block; lance_set_multicast() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/amd/ |
H A D | 7990.c | 100 t, ib->brx_ring[t].rmd1_hadr, ib->brx_ring[t].rmd0, \ 101 ib->brx_ring[t].length, \ 102 ib->brx_ring[t].mblength, ib->brx_ring[t].rmd1_bits); \ 106 t, ib->btx_ring[t].tmd1_hadr, ib->btx_ring[t].tmd0, \ 107 ib->btx_ring[t].length, \ 108 ib->btx_ring[t].misc, ib 140 volatile struct lance_init_block *ib = lp->init_block; lance_init_ring() local 275 volatile struct lance_init_block *ib = lp->init_block; lance_rx() local 355 volatile struct lance_init_block *ib = lp->init_block; lance_tx() local 542 volatile struct lance_init_block *ib = lp->init_block; lance_start_xmit() local 601 volatile struct lance_init_block *ib = lp->init_block; lance_load_multicast() local 628 volatile struct lance_init_block *ib = lp->init_block; lance_set_multicast() local [all...] |