Home
last modified time | relevance | path

Searched refs:ib (Results 1 - 25 of 282) sorted by relevance

12345678910>>...12

/kernel/linux/linux-6.6/drivers/gpu/drm/radeon/
H A Dradeon_ib.c51 * @ib: IB object returned
59 struct radeon_ib *ib, struct radeon_vm *vm, in radeon_ib_get()
64 r = radeon_sa_bo_new(&rdev->ring_tmp_bo, &ib->sa_bo, size, 256); in radeon_ib_get()
70 radeon_sync_create(&ib->sync); in radeon_ib_get()
72 ib->ring = ring; in radeon_ib_get()
73 ib->fence = NULL; in radeon_ib_get()
74 ib->ptr = radeon_sa_bo_cpu_addr(ib->sa_bo); in radeon_ib_get()
75 ib->vm = vm; in radeon_ib_get()
77 /* ib poo in radeon_ib_get()
58 radeon_ib_get(struct radeon_device *rdev, int ring, struct radeon_ib *ib, struct radeon_vm *vm, unsigned size) radeon_ib_get() argument
97 radeon_ib_free(struct radeon_device *rdev, struct radeon_ib *ib) radeon_ib_free() argument
125 radeon_ib_schedule(struct radeon_device *rdev, struct radeon_ib *ib, struct radeon_ib *const_ib, bool hdp_flush) radeon_ib_schedule() argument
[all...]
H A Dsi_dma.c61 * @ib: indirect buffer to fill with commands
69 struct radeon_ib *ib, in si_dma_vm_copy_pages()
78 ib->ptr[ib->length_dw++] = DMA_PACKET(DMA_PACKET_COPY, in si_dma_vm_copy_pages()
80 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in si_dma_vm_copy_pages()
81 ib->ptr[ib->length_dw++] = lower_32_bits(src); in si_dma_vm_copy_pages()
82 ib->ptr[ib in si_dma_vm_copy_pages()
68 si_dma_vm_copy_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t src, unsigned count) si_dma_vm_copy_pages() argument
104 si_dma_vm_write_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint32_t flags) si_dma_vm_write_pages() argument
151 si_dma_vm_set_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint32_t flags) si_dma_vm_set_pages() argument
[all...]
H A Dradeon_vce.c349 struct radeon_ib ib; in radeon_vce_get_create_msg() local
353 r = radeon_ib_get(rdev, ring, &ib, NULL, ib_size_dw * 4); in radeon_vce_get_create_msg()
355 DRM_ERROR("radeon: failed to get ib (%d).\n", r); in radeon_vce_get_create_msg()
359 dummy = ib.gpu_addr + 1024; in radeon_vce_get_create_msg()
362 ib.length_dw = 0; in radeon_vce_get_create_msg()
363 ib.ptr[ib.length_dw++] = cpu_to_le32(0x0000000c); /* len */ in radeon_vce_get_create_msg()
364 ib.ptr[ib.length_dw++] = cpu_to_le32(0x00000001); /* session cmd */ in radeon_vce_get_create_msg()
365 ib in radeon_vce_get_create_msg()
416 struct radeon_ib ib; radeon_vce_get_destroy_msg() local
718 radeon_vce_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib) radeon_vce_ib_execute() argument
[all...]
H A Dni_dma.c117 * @ib: IB object to schedule
122 struct radeon_ib *ib) in cayman_dma_ring_ib_execute()
124 struct radeon_ring *ring = &rdev->ring[ib->ring]; in cayman_dma_ring_ib_execute()
125 unsigned vm_id = ib->vm ? ib->vm->ids[ib->ring].id : 0; in cayman_dma_ring_ib_execute()
144 radeon_ring_write(ring, (ib->gpu_addr & 0xFFFFFFE0)); in cayman_dma_ring_ib_execute()
145 radeon_ring_write(ring, (ib->length_dw << 12) | (upper_32_bits(ib->gpu_addr) & 0xFF)); in cayman_dma_ring_ib_execute()
307 * @ib
121 cayman_dma_ring_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib) cayman_dma_ring_ib_execute() argument
314 cayman_dma_vm_copy_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t src, unsigned count) cayman_dma_vm_copy_pages() argument
352 cayman_dma_vm_write_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint32_t flags) cayman_dma_vm_write_pages() argument
400 cayman_dma_vm_set_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint32_t flags) cayman_dma_vm_set_pages() argument
442 cayman_dma_vm_pad_ib(struct radeon_ib *ib) cayman_dma_vm_pad_ib() argument
[all...]
H A Dcik_sdma.c128 * @ib: IB object to schedule
133 struct radeon_ib *ib) in cik_sdma_ring_ib_execute()
135 struct radeon_ring *ring = &rdev->ring[ib->ring]; in cik_sdma_ring_ib_execute()
136 u32 extra_bits = (ib->vm ? ib->vm->ids[ib->ring].id : 0) & 0xf; in cik_sdma_ring_ib_execute()
154 radeon_ring_write(ring, ib->gpu_addr & 0xffffffe0); /* base must be 32 byte aligned */ in cik_sdma_ring_ib_execute()
155 radeon_ring_write(ring, upper_32_bits(ib->gpu_addr)); in cik_sdma_ring_ib_execute()
156 radeon_ring_write(ring, ib->length_dw); in cik_sdma_ring_ib_execute()
703 struct radeon_ib ib; in cik_sdma_ib_test() local
132 cik_sdma_ring_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib) cik_sdma_ring_ib_execute() argument
802 cik_sdma_vm_copy_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t src, unsigned count) cik_sdma_vm_copy_pages() argument
840 cik_sdma_vm_write_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint32_t flags) cik_sdma_vm_write_pages() argument
889 cik_sdma_vm_set_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint32_t flags) cik_sdma_vm_set_pages() argument
932 cik_sdma_vm_pad_ib(struct radeon_ib *ib) cik_sdma_vm_pad_ib() argument
[all...]
/kernel/linux/linux-6.6/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_vce.c442 struct amdgpu_ib *ib; in amdgpu_vce_get_create_msg() local
463 ib = &job->ibs[0]; in amdgpu_vce_get_create_msg()
468 ib->length_dw = 0; in amdgpu_vce_get_create_msg()
469 ib->ptr[ib->length_dw++] = 0x0000000c; /* len */ in amdgpu_vce_get_create_msg()
470 ib->ptr[ib->length_dw++] = 0x00000001; /* session cmd */ in amdgpu_vce_get_create_msg()
471 ib->ptr[ib->length_dw++] = handle; in amdgpu_vce_get_create_msg()
474 ib in amdgpu_vce_get_create_msg()
534 struct amdgpu_ib *ib; amdgpu_vce_get_destroy_msg() local
598 amdgpu_vce_validate_bo(struct amdgpu_cs_parser *p, struct amdgpu_ib *ib, int lo, int hi, unsigned int size, int32_t index) amdgpu_vce_validate_bo() argument
649 amdgpu_vce_cs_reloc(struct amdgpu_cs_parser *p, struct amdgpu_ib *ib, int lo, int hi, unsigned int size, uint32_t index) amdgpu_vce_cs_reloc() argument
735 amdgpu_vce_ring_parse_cs(struct amdgpu_cs_parser *p, struct amdgpu_job *job, struct amdgpu_ib *ib) amdgpu_vce_ring_parse_cs() argument
973 amdgpu_vce_ring_parse_cs_vm(struct amdgpu_cs_parser *p, struct amdgpu_job *job, struct amdgpu_ib *ib) amdgpu_vce_ring_parse_cs_vm() argument
1067 amdgpu_vce_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) amdgpu_vce_ring_emit_ib() argument
[all...]
H A Damdgpu_vcn.c534 struct amdgpu_ib *ib; in amdgpu_vcn_dec_send_msg() local
543 ib = &job->ibs[0]; in amdgpu_vcn_dec_send_msg()
544 ib->ptr[0] = PACKET0(adev->vcn.internal.data0, 0); in amdgpu_vcn_dec_send_msg()
545 ib->ptr[1] = addr; in amdgpu_vcn_dec_send_msg()
546 ib->ptr[2] = PACKET0(adev->vcn.internal.data1, 0); in amdgpu_vcn_dec_send_msg()
547 ib->ptr[3] = addr >> 32; in amdgpu_vcn_dec_send_msg()
548 ib->ptr[4] = PACKET0(adev->vcn.internal.cmd, 0); in amdgpu_vcn_dec_send_msg()
549 ib->ptr[5] = 0; in amdgpu_vcn_dec_send_msg()
551 ib->ptr[i] = PACKET0(adev->vcn.internal.nop, 0); in amdgpu_vcn_dec_send_msg()
552 ib in amdgpu_vcn_dec_send_msg()
575 amdgpu_vcn_dec_get_create_msg(struct amdgpu_ring *ring, uint32_t handle, struct amdgpu_ib *ib) amdgpu_vcn_dec_get_create_msg() argument
610 amdgpu_vcn_dec_get_destroy_msg(struct amdgpu_ring *ring, uint32_t handle, struct amdgpu_ib *ib) amdgpu_vcn_dec_get_destroy_msg() argument
640 struct amdgpu_ib ib; amdgpu_vcn_dec_ring_test_ib() local
669 amdgpu_vcn_unified_ring_ib_header(struct amdgpu_ib *ib, uint32_t ib_pack_in_dw, bool enc) amdgpu_vcn_unified_ring_ib_header() argument
708 struct amdgpu_ib *ib; amdgpu_vcn_dec_sw_send_msg() local
772 struct amdgpu_ib ib; amdgpu_vcn_dec_sw_ring_test_ib() local
838 struct amdgpu_ib *ib; amdgpu_vcn_enc_get_create_msg() local
905 struct amdgpu_ib *ib; amdgpu_vcn_enc_get_destroy_msg() local
970 struct amdgpu_ib ib; amdgpu_vcn_enc_ring_test_ib() local
[all...]
H A Dsi_dma.c64 struct amdgpu_ib *ib, in si_dma_ring_emit_ib()
74 amdgpu_ring_write(ring, (ib->gpu_addr & 0xFFFFFFE0)); in si_dma_ring_emit_ib()
75 amdgpu_ring_write(ring, (ib->length_dw << 12) | (upper_32_bits(ib->gpu_addr) & 0xFF)); in si_dma_ring_emit_ib()
251 struct amdgpu_ib ib; in si_dma_ring_test_ib() local
265 memset(&ib, 0, sizeof(ib)); in si_dma_ring_test_ib()
267 AMDGPU_IB_POOL_DIRECT, &ib); in si_dma_ring_test_ib()
271 ib.ptr[0] = DMA_PACKET(DMA_PACKET_WRITE, 0, 0, 0, 1); in si_dma_ring_test_ib()
272 ib in si_dma_ring_test_ib()
62 si_dma_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) si_dma_ring_emit_ib() argument
311 si_dma_vm_copy_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t src, unsigned count) si_dma_vm_copy_pte() argument
336 si_dma_vm_write_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t value, unsigned count, uint32_t incr) si_dma_vm_write_pte() argument
364 si_dma_vm_set_pte_pde(struct amdgpu_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint64_t flags) si_dma_vm_set_pte_pde() argument
405 si_dma_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) si_dma_ring_pad_ib() argument
775 si_dma_emit_copy_buffer(struct amdgpu_ib *ib, uint64_t src_offset, uint64_t dst_offset, uint32_t byte_count, bool tmz) si_dma_emit_copy_buffer() argument
799 si_dma_emit_fill_buffer(struct amdgpu_ib *ib, uint32_t src_data, uint64_t dst_offset, uint32_t byte_count) si_dma_emit_fill_buffer() argument
[all...]
H A Dsdma_v3_0.c415 * @ib: IB object to schedule
422 struct amdgpu_ib *ib, in sdma_v3_0_ring_emit_ib()
433 amdgpu_ring_write(ring, lower_32_bits(ib->gpu_addr) & 0xffffffe0); in sdma_v3_0_ring_emit_ib()
434 amdgpu_ring_write(ring, upper_32_bits(ib->gpu_addr)); in sdma_v3_0_ring_emit_ib()
435 amdgpu_ring_write(ring, ib->length_dw); in sdma_v3_0_ring_emit_ib()
863 struct amdgpu_ib ib; in sdma_v3_0_ring_test_ib() local
877 memset(&ib, 0, sizeof(ib)); in sdma_v3_0_ring_test_ib()
879 AMDGPU_IB_POOL_DIRECT, &ib); in sdma_v3_0_ring_test_ib()
883 ib in sdma_v3_0_ring_test_ib()
420 sdma_v3_0_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) sdma_v3_0_ring_emit_ib() argument
928 sdma_v3_0_vm_copy_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t src, unsigned count) sdma_v3_0_vm_copy_pte() argument
955 sdma_v3_0_vm_write_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t value, unsigned count, uint32_t incr) sdma_v3_0_vm_write_pte() argument
985 sdma_v3_0_vm_set_pte_pde(struct amdgpu_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint64_t flags) sdma_v3_0_vm_set_pte_pde() argument
1009 sdma_v3_0_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) sdma_v3_0_ring_pad_ib() argument
1631 sdma_v3_0_emit_copy_buffer(struct amdgpu_ib *ib, uint64_t src_offset, uint64_t dst_offset, uint32_t byte_count, bool tmz) sdma_v3_0_emit_copy_buffer() argument
1657 sdma_v3_0_emit_fill_buffer(struct amdgpu_ib *ib, uint32_t src_data, uint64_t dst_offset, uint32_t byte_count) sdma_v3_0_emit_fill_buffer() argument
[all...]
H A Dsdma_v2_4.c241 * @ib: IB object to schedule
248 struct amdgpu_ib *ib, in sdma_v2_4_ring_emit_ib()
259 amdgpu_ring_write(ring, lower_32_bits(ib->gpu_addr) & 0xffffffe0); in sdma_v2_4_ring_emit_ib()
260 amdgpu_ring_write(ring, upper_32_bits(ib->gpu_addr)); in sdma_v2_4_ring_emit_ib()
261 amdgpu_ring_write(ring, ib->length_dw); in sdma_v2_4_ring_emit_ib()
591 struct amdgpu_ib ib; in sdma_v2_4_ring_test_ib() local
605 memset(&ib, 0, sizeof(ib)); in sdma_v2_4_ring_test_ib()
607 AMDGPU_IB_POOL_DIRECT, &ib); in sdma_v2_4_ring_test_ib()
611 ib in sdma_v2_4_ring_test_ib()
246 sdma_v2_4_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) sdma_v2_4_ring_emit_ib() argument
657 sdma_v2_4_vm_copy_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t src, unsigned count) sdma_v2_4_vm_copy_pte() argument
684 sdma_v2_4_vm_write_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t value, unsigned count, uint32_t incr) sdma_v2_4_vm_write_pte() argument
714 sdma_v2_4_vm_set_pte_pde(struct amdgpu_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint64_t flags) sdma_v2_4_vm_set_pte_pde() argument
738 sdma_v2_4_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) sdma_v2_4_ring_pad_ib() argument
1193 sdma_v2_4_emit_copy_buffer(struct amdgpu_ib *ib, uint64_t src_offset, uint64_t dst_offset, uint32_t byte_count, bool tmz) sdma_v2_4_emit_copy_buffer() argument
1219 sdma_v2_4_emit_fill_buffer(struct amdgpu_ib *ib, uint32_t src_data, uint64_t dst_offset, uint32_t byte_count) sdma_v2_4_emit_fill_buffer() argument
[all...]
H A Dcik_sdma.c215 * @ib: IB object to schedule
222 struct amdgpu_ib *ib, in cik_sdma_ring_emit_ib()
232 amdgpu_ring_write(ring, ib->gpu_addr & 0xffffffe0); /* base must be 32 byte aligned */ in cik_sdma_ring_emit_ib()
233 amdgpu_ring_write(ring, upper_32_bits(ib->gpu_addr) & 0xffffffff); in cik_sdma_ring_emit_ib()
234 amdgpu_ring_write(ring, ib->length_dw); in cik_sdma_ring_emit_ib()
658 struct amdgpu_ib ib; in cik_sdma_ring_test_ib() local
672 memset(&ib, 0, sizeof(ib)); in cik_sdma_ring_test_ib()
674 AMDGPU_IB_POOL_DIRECT, &ib); in cik_sdma_ring_test_ib()
678 ib in cik_sdma_ring_test_ib()
220 cik_sdma_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) cik_sdma_ring_emit_ib() argument
720 cik_sdma_vm_copy_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t src, unsigned count) cik_sdma_vm_copy_pte() argument
747 cik_sdma_vm_write_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t value, unsigned count, uint32_t incr) cik_sdma_vm_write_pte() argument
777 cik_sdma_vm_set_pte_pde(struct amdgpu_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint64_t flags) cik_sdma_vm_set_pte_pde() argument
801 cik_sdma_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) cik_sdma_ring_pad_ib() argument
1305 cik_sdma_emit_copy_buffer(struct amdgpu_ib *ib, uint64_t src_offset, uint64_t dst_offset, uint32_t byte_count, bool tmz) cik_sdma_emit_copy_buffer() argument
1330 cik_sdma_emit_fill_buffer(struct amdgpu_ib *ib, uint32_t src_data, uint64_t dst_offset, uint32_t byte_count) cik_sdma_emit_fill_buffer() argument
[all...]
/kernel/linux/linux-5.10/drivers/gpu/drm/radeon/
H A Dsi_dma.c62 * @ib: indirect buffer to fill with commands
70 struct radeon_ib *ib, in si_dma_vm_copy_pages()
79 ib->ptr[ib->length_dw++] = DMA_PACKET(DMA_PACKET_COPY, in si_dma_vm_copy_pages()
81 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in si_dma_vm_copy_pages()
82 ib->ptr[ib->length_dw++] = lower_32_bits(src); in si_dma_vm_copy_pages()
83 ib->ptr[ib in si_dma_vm_copy_pages()
69 si_dma_vm_copy_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t src, unsigned count) si_dma_vm_copy_pages() argument
105 si_dma_vm_write_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint32_t flags) si_dma_vm_write_pages() argument
152 si_dma_vm_set_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint32_t flags) si_dma_vm_set_pages() argument
[all...]
H A Dradeon_ib.c51 * @ib: IB object returned
59 struct radeon_ib *ib, struct radeon_vm *vm, in radeon_ib_get()
64 r = radeon_sa_bo_new(rdev, &rdev->ring_tmp_bo, &ib->sa_bo, size, 256); in radeon_ib_get()
70 radeon_sync_create(&ib->sync); in radeon_ib_get()
72 ib->ring = ring; in radeon_ib_get()
73 ib->fence = NULL; in radeon_ib_get()
74 ib->ptr = radeon_sa_bo_cpu_addr(ib->sa_bo); in radeon_ib_get()
75 ib->vm = vm; in radeon_ib_get()
77 /* ib poo in radeon_ib_get()
58 radeon_ib_get(struct radeon_device *rdev, int ring, struct radeon_ib *ib, struct radeon_vm *vm, unsigned size) radeon_ib_get() argument
97 radeon_ib_free(struct radeon_device *rdev, struct radeon_ib *ib) radeon_ib_free() argument
125 radeon_ib_schedule(struct radeon_device *rdev, struct radeon_ib *ib, struct radeon_ib *const_ib, bool hdp_flush) radeon_ib_schedule() argument
[all...]
H A Dradeon_vce.c349 struct radeon_ib ib; in radeon_vce_get_create_msg() local
353 r = radeon_ib_get(rdev, ring, &ib, NULL, ib_size_dw * 4); in radeon_vce_get_create_msg()
355 DRM_ERROR("radeon: failed to get ib (%d).\n", r); in radeon_vce_get_create_msg()
359 dummy = ib.gpu_addr + 1024; in radeon_vce_get_create_msg()
362 ib.length_dw = 0; in radeon_vce_get_create_msg()
363 ib.ptr[ib.length_dw++] = cpu_to_le32(0x0000000c); /* len */ in radeon_vce_get_create_msg()
364 ib.ptr[ib.length_dw++] = cpu_to_le32(0x00000001); /* session cmd */ in radeon_vce_get_create_msg()
365 ib in radeon_vce_get_create_msg()
416 struct radeon_ib ib; radeon_vce_get_destroy_msg() local
718 radeon_vce_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib) radeon_vce_ib_execute() argument
[all...]
H A Dni_dma.c118 * @ib: IB object to schedule
123 struct radeon_ib *ib) in cayman_dma_ring_ib_execute()
125 struct radeon_ring *ring = &rdev->ring[ib->ring]; in cayman_dma_ring_ib_execute()
126 unsigned vm_id = ib->vm ? ib->vm->ids[ib->ring].id : 0; in cayman_dma_ring_ib_execute()
145 radeon_ring_write(ring, (ib->gpu_addr & 0xFFFFFFE0)); in cayman_dma_ring_ib_execute()
146 radeon_ring_write(ring, (ib->length_dw << 12) | (upper_32_bits(ib->gpu_addr) & 0xFF)); in cayman_dma_ring_ib_execute()
308 * @ib
122 cayman_dma_ring_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib) cayman_dma_ring_ib_execute() argument
315 cayman_dma_vm_copy_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t src, unsigned count) cayman_dma_vm_copy_pages() argument
353 cayman_dma_vm_write_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint32_t flags) cayman_dma_vm_write_pages() argument
401 cayman_dma_vm_set_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint32_t flags) cayman_dma_vm_set_pages() argument
443 cayman_dma_vm_pad_ib(struct radeon_ib *ib) cayman_dma_vm_pad_ib() argument
[all...]
H A Dcik_sdma.c129 * @ib: IB object to schedule
134 struct radeon_ib *ib) in cik_sdma_ring_ib_execute()
136 struct radeon_ring *ring = &rdev->ring[ib->ring]; in cik_sdma_ring_ib_execute()
137 u32 extra_bits = (ib->vm ? ib->vm->ids[ib->ring].id : 0) & 0xf; in cik_sdma_ring_ib_execute()
155 radeon_ring_write(ring, ib->gpu_addr & 0xffffffe0); /* base must be 32 byte aligned */ in cik_sdma_ring_ib_execute()
156 radeon_ring_write(ring, upper_32_bits(ib->gpu_addr)); in cik_sdma_ring_ib_execute()
157 radeon_ring_write(ring, ib->length_dw); in cik_sdma_ring_ib_execute()
704 struct radeon_ib ib; in cik_sdma_ib_test() local
133 cik_sdma_ring_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib) cik_sdma_ring_ib_execute() argument
803 cik_sdma_vm_copy_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t src, unsigned count) cik_sdma_vm_copy_pages() argument
841 cik_sdma_vm_write_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint32_t flags) cik_sdma_vm_write_pages() argument
890 cik_sdma_vm_set_pages(struct radeon_device *rdev, struct radeon_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint32_t flags) cik_sdma_vm_set_pages() argument
933 cik_sdma_vm_pad_ib(struct radeon_ib *ib) cik_sdma_vm_pad_ib() argument
[all...]
/kernel/linux/linux-5.10/arch/s390/include/asm/
H A Didals.h120 struct idal_buffer *ib; in idal_buffer_alloc() local
125 ib = kmalloc(struct_size(ib, data, nr_ptrs), GFP_DMA | GFP_KERNEL); in idal_buffer_alloc()
126 if (ib == NULL) in idal_buffer_alloc()
128 ib->size = size; in idal_buffer_alloc()
129 ib->page_order = page_order; in idal_buffer_alloc()
132 ib->data[i] = ib->data[i-1] + IDA_BLOCK_SIZE; in idal_buffer_alloc()
135 ib->data[i] = (void *) in idal_buffer_alloc()
137 if (ib in idal_buffer_alloc()
155 idal_buffer_free(struct idal_buffer *ib) idal_buffer_free() argument
170 __idal_buffer_is_needed(struct idal_buffer *ib) __idal_buffer_is_needed() argument
180 idal_buffer_set_cda(struct idal_buffer *ib, struct ccw1 *ccw) idal_buffer_set_cda() argument
196 idal_buffer_to_user(struct idal_buffer *ib, void __user *to, size_t count) idal_buffer_to_user() argument
216 idal_buffer_from_user(struct idal_buffer *ib, const void __user *from, size_t count) idal_buffer_from_user() argument
[all...]
/kernel/linux/linux-6.6/arch/s390/include/asm/
H A Didals.h132 struct idal_buffer *ib; in idal_buffer_alloc() local
137 ib = kmalloc(struct_size(ib, data, nr_ptrs), GFP_DMA | GFP_KERNEL); in idal_buffer_alloc()
138 if (ib == NULL) in idal_buffer_alloc()
140 ib->size = size; in idal_buffer_alloc()
141 ib->page_order = page_order; in idal_buffer_alloc()
144 ib->data[i] = ib->data[i-1] + IDA_BLOCK_SIZE; in idal_buffer_alloc()
147 ib->data[i] = (void *) in idal_buffer_alloc()
149 if (ib in idal_buffer_alloc()
167 idal_buffer_free(struct idal_buffer *ib) idal_buffer_free() argument
182 __idal_buffer_is_needed(struct idal_buffer *ib) __idal_buffer_is_needed() argument
192 idal_buffer_set_cda(struct idal_buffer *ib, struct ccw1 *ccw) idal_buffer_set_cda() argument
208 idal_buffer_to_user(struct idal_buffer *ib, void __user *to, size_t count) idal_buffer_to_user() argument
228 idal_buffer_from_user(struct idal_buffer *ib, const void __user *from, size_t count) idal_buffer_from_user() argument
[all...]
/kernel/linux/linux-5.10/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_vce.c444 struct amdgpu_ib *ib; in amdgpu_vce_get_create_msg() local
454 ib = &job->ibs[0]; in amdgpu_vce_get_create_msg()
459 ib->length_dw = 0; in amdgpu_vce_get_create_msg()
460 ib->ptr[ib->length_dw++] = 0x0000000c; /* len */ in amdgpu_vce_get_create_msg()
461 ib->ptr[ib->length_dw++] = 0x00000001; /* session cmd */ in amdgpu_vce_get_create_msg()
462 ib->ptr[ib->length_dw++] = handle; in amdgpu_vce_get_create_msg()
465 ib in amdgpu_vce_get_create_msg()
524 struct amdgpu_ib *ib; amdgpu_vce_get_destroy_msg() local
721 struct amdgpu_ib *ib = &p->job->ibs[ib_idx]; amdgpu_vce_ring_parse_cs() local
957 struct amdgpu_ib *ib = &p->job->ibs[ib_idx]; amdgpu_vce_ring_parse_cs_vm() local
1046 amdgpu_vce_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) amdgpu_vce_ring_emit_ib() argument
[all...]
H A Dsi_dma.c65 struct amdgpu_ib *ib, in si_dma_ring_emit_ib()
75 amdgpu_ring_write(ring, (ib->gpu_addr & 0xFFFFFFE0)); in si_dma_ring_emit_ib()
76 amdgpu_ring_write(ring, (ib->length_dw << 12) | (upper_32_bits(ib->gpu_addr) & 0xFF)); in si_dma_ring_emit_ib()
254 struct amdgpu_ib ib; in si_dma_ring_test_ib() local
268 memset(&ib, 0, sizeof(ib)); in si_dma_ring_test_ib()
270 AMDGPU_IB_POOL_DIRECT, &ib); in si_dma_ring_test_ib()
274 ib.ptr[0] = DMA_PACKET(DMA_PACKET_WRITE, 0, 0, 0, 1); in si_dma_ring_test_ib()
275 ib in si_dma_ring_test_ib()
63 si_dma_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) si_dma_ring_emit_ib() argument
314 si_dma_vm_copy_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t src, unsigned count) si_dma_vm_copy_pte() argument
339 si_dma_vm_write_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t value, unsigned count, uint32_t incr) si_dma_vm_write_pte() argument
367 si_dma_vm_set_pte_pde(struct amdgpu_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint64_t flags) si_dma_vm_set_pte_pde() argument
407 si_dma_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) si_dma_ring_pad_ib() argument
776 si_dma_emit_copy_buffer(struct amdgpu_ib *ib, uint64_t src_offset, uint64_t dst_offset, uint32_t byte_count, bool tmz) si_dma_emit_copy_buffer() argument
800 si_dma_emit_fill_buffer(struct amdgpu_ib *ib, uint32_t src_data, uint64_t dst_offset, uint32_t byte_count) si_dma_emit_fill_buffer() argument
[all...]
H A Damdgpu_vcn.c439 struct amdgpu_ib *ib; in amdgpu_vcn_dec_send_msg() local
448 ib = &job->ibs[0]; in amdgpu_vcn_dec_send_msg()
450 ib->ptr[0] = PACKET0(adev->vcn.internal.data0, 0); in amdgpu_vcn_dec_send_msg()
451 ib->ptr[1] = addr; in amdgpu_vcn_dec_send_msg()
452 ib->ptr[2] = PACKET0(adev->vcn.internal.data1, 0); in amdgpu_vcn_dec_send_msg()
453 ib->ptr[3] = addr >> 32; in amdgpu_vcn_dec_send_msg()
454 ib->ptr[4] = PACKET0(adev->vcn.internal.cmd, 0); in amdgpu_vcn_dec_send_msg()
455 ib->ptr[5] = 0; in amdgpu_vcn_dec_send_msg()
457 ib->ptr[i] = PACKET0(adev->vcn.internal.nop, 0); in amdgpu_vcn_dec_send_msg()
458 ib in amdgpu_vcn_dec_send_msg()
606 struct amdgpu_ib *ib; amdgpu_vcn_enc_get_create_msg() local
660 struct amdgpu_ib *ib; amdgpu_vcn_enc_get_destroy_msg() local
[all...]
H A Dsdma_v2_4.c246 * @ib: IB object to schedule
252 struct amdgpu_ib *ib, in sdma_v2_4_ring_emit_ib()
263 amdgpu_ring_write(ring, lower_32_bits(ib->gpu_addr) & 0xffffffe0); in sdma_v2_4_ring_emit_ib()
264 amdgpu_ring_write(ring, upper_32_bits(ib->gpu_addr)); in sdma_v2_4_ring_emit_ib()
265 amdgpu_ring_write(ring, ib->length_dw); in sdma_v2_4_ring_emit_ib()
600 struct amdgpu_ib ib; in sdma_v2_4_ring_test_ib() local
614 memset(&ib, 0, sizeof(ib)); in sdma_v2_4_ring_test_ib()
616 AMDGPU_IB_POOL_DIRECT, &ib); in sdma_v2_4_ring_test_ib()
620 ib in sdma_v2_4_ring_test_ib()
250 sdma_v2_4_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) sdma_v2_4_ring_emit_ib() argument
666 sdma_v2_4_vm_copy_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t src, unsigned count) sdma_v2_4_vm_copy_pte() argument
693 sdma_v2_4_vm_write_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t value, unsigned count, uint32_t incr) sdma_v2_4_vm_write_pte() argument
723 sdma_v2_4_vm_set_pte_pde(struct amdgpu_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint64_t flags) sdma_v2_4_vm_set_pte_pde() argument
746 sdma_v2_4_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) sdma_v2_4_ring_pad_ib() argument
1200 sdma_v2_4_emit_copy_buffer(struct amdgpu_ib *ib, uint64_t src_offset, uint64_t dst_offset, uint32_t byte_count, bool tmz) sdma_v2_4_emit_copy_buffer() argument
1226 sdma_v2_4_emit_fill_buffer(struct amdgpu_ib *ib, uint32_t src_data, uint64_t dst_offset, uint32_t byte_count) sdma_v2_4_emit_fill_buffer() argument
[all...]
H A Dcik_sdma.c218 * @ib: IB object to schedule
224 struct amdgpu_ib *ib, in cik_sdma_ring_emit_ib()
234 amdgpu_ring_write(ring, ib->gpu_addr & 0xffffffe0); /* base must be 32 byte aligned */ in cik_sdma_ring_emit_ib()
235 amdgpu_ring_write(ring, upper_32_bits(ib->gpu_addr) & 0xffffffff); in cik_sdma_ring_emit_ib()
236 amdgpu_ring_write(ring, ib->length_dw); in cik_sdma_ring_emit_ib()
665 struct amdgpu_ib ib; in cik_sdma_ring_test_ib() local
679 memset(&ib, 0, sizeof(ib)); in cik_sdma_ring_test_ib()
681 AMDGPU_IB_POOL_DIRECT, &ib); in cik_sdma_ring_test_ib()
685 ib in cik_sdma_ring_test_ib()
222 cik_sdma_ring_emit_ib(struct amdgpu_ring *ring, struct amdgpu_job *job, struct amdgpu_ib *ib, uint32_t flags) cik_sdma_ring_emit_ib() argument
727 cik_sdma_vm_copy_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t src, unsigned count) cik_sdma_vm_copy_pte() argument
754 cik_sdma_vm_write_pte(struct amdgpu_ib *ib, uint64_t pe, uint64_t value, unsigned count, uint32_t incr) cik_sdma_vm_write_pte() argument
784 cik_sdma_vm_set_pte_pde(struct amdgpu_ib *ib, uint64_t pe, uint64_t addr, unsigned count, uint32_t incr, uint64_t flags) cik_sdma_vm_set_pte_pde() argument
807 cik_sdma_ring_pad_ib(struct amdgpu_ring *ring, struct amdgpu_ib *ib) cik_sdma_ring_pad_ib() argument
1310 cik_sdma_emit_copy_buffer(struct amdgpu_ib *ib, uint64_t src_offset, uint64_t dst_offset, uint32_t byte_count, bool tmz) cik_sdma_emit_copy_buffer() argument
1335 cik_sdma_emit_fill_buffer(struct amdgpu_ib *ib, uint32_t src_data, uint64_t dst_offset, uint32_t byte_count) cik_sdma_emit_fill_buffer() argument
[all...]
/kernel/linux/linux-5.10/drivers/net/ethernet/amd/
H A D7990.c100 t, ib->brx_ring[t].rmd1_hadr, ib->brx_ring[t].rmd0, \
101 ib->brx_ring[t].length, \
102 ib->brx_ring[t].mblength, ib->brx_ring[t].rmd1_bits); \
106 t, ib->btx_ring[t].tmd1_hadr, ib->btx_ring[t].tmd0, \
107 ib->btx_ring[t].length, \
108 ib->btx_ring[t].misc, ib
140 volatile struct lance_init_block *ib = lp->init_block; lance_init_ring() local
275 volatile struct lance_init_block *ib = lp->init_block; lance_rx() local
355 volatile struct lance_init_block *ib = lp->init_block; lance_tx() local
542 volatile struct lance_init_block *ib = lp->init_block; lance_start_xmit() local
601 volatile struct lance_init_block *ib = lp->init_block; lance_load_multicast() local
628 volatile struct lance_init_block *ib = lp->init_block; lance_set_multicast() local
[all...]
/kernel/linux/linux-6.6/drivers/net/ethernet/amd/
H A D7990.c100 t, ib->brx_ring[t].rmd1_hadr, ib->brx_ring[t].rmd0, \
101 ib->brx_ring[t].length, \
102 ib->brx_ring[t].mblength, ib->brx_ring[t].rmd1_bits); \
106 t, ib->btx_ring[t].tmd1_hadr, ib->btx_ring[t].tmd0, \
107 ib->btx_ring[t].length, \
108 ib->btx_ring[t].misc, ib
140 volatile struct lance_init_block *ib = lp->init_block; lance_init_ring() local
275 volatile struct lance_init_block *ib = lp->init_block; lance_rx() local
355 volatile struct lance_init_block *ib = lp->init_block; lance_tx() local
542 volatile struct lance_init_block *ib = lp->init_block; lance_start_xmit() local
601 volatile struct lance_init_block *ib = lp->init_block; lance_load_multicast() local
628 volatile struct lance_init_block *ib = lp->init_block; lance_set_multicast() local
[all...]

Completed in 20 milliseconds

12345678910>>...12