Home
last modified time | relevance | path

Searched refs:dma (Results 1 - 25 of 36) sorted by relevance

12

/third_party/mesa3d/src/gallium/drivers/svga/
H A Dsvga_resource_buffer_upload.c128 assert(sbuf->handle || !sbuf->dma.pending); in svga_buffer_create_hw_storage()
141 assert(!sbuf->dma.pending); in svga_buffer_create_hw_storage()
238 sbuf->dma.flags.discard = TRUE; in svga_buffer_create_host_surface()
508 assert(sbuf->dma.updates == NULL); in svga_buffer_upload_gb_command()
535 sbuf->dma.updates = whole_update_cmd; in svga_buffer_upload_gb_command()
543 memcpy(whole_update_cmd, sbuf->dma.updates, sizeof(*whole_update_cmd)); in svga_buffer_upload_gb_command()
551 sbuf->dma.svga = svga; in svga_buffer_upload_gb_command()
557 sbuf->dma.flags.discard = FALSE; in svga_buffer_upload_gb_command()
619 sbuf->dma.boxes = (SVGA3dCopyBox *)&cmd[1]; in svga_buffer_upload_hb_command()
620 sbuf->dma in svga_buffer_upload_hb_command()
[all...]
H A Dsvga_resource_buffer.c167 if (sbuf->dma.pending) { in svga_buffer_transfer_map()
197 if (sbuf->dma.pending) { in svga_buffer_transfer_map()
213 sbuf->dma.flags.discard = TRUE; in svga_buffer_transfer_map()
223 sbuf->dma.flags.unsynchronized = TRUE; in svga_buffer_transfer_map()
233 if (sbuf->dma.pending) { in svga_buffer_transfer_map()
268 sbuf->dma.flags.unsynchronized = FALSE; in svga_buffer_transfer_map()
391 sbuf->dma.flags.discard = TRUE; in svga_buffer_transfer_unmap()
427 assert(!sbuf->dma.pending); in svga_resource_destroy()
H A Dsvga_resource_buffer.h142 * This information will be copied into dma.boxes, when emiting the
208 } dma; member
211 * Linked list head, used to gather all buffers with pending dma uploads on
212 * a context. It is only valid if the dma.pending is set above.
/third_party/mesa3d/src/gallium/drivers/r600/
H A Dr600_pipe_common.c218 struct radeon_cmdbuf *cs = &rctx->dma.cs; in r600_dma_emit_wait_idle()
231 uint64_t vram = (uint64_t)ctx->dma.cs.used_vram_kb * 1024; in r600_need_dma_space()
232 uint64_t gtt = (uint64_t)ctx->dma.cs.used_gart_kb * 1024; in r600_need_dma_space()
266 if (!ctx->ws->cs_check_space(&ctx->dma.cs, num_dw) || in r600_need_dma_space()
267 ctx->dma.cs.used_vram_kb + ctx->dma.cs.used_gart_kb > 64 * 1024 || in r600_need_dma_space()
268 !radeon_cs_memory_below_limit(ctx->screen, &ctx->dma.cs, vram, gtt)) { in r600_need_dma_space()
269 ctx->dma.flush(ctx, PIPE_FLUSH_ASYNC, NULL); in r600_need_dma_space()
270 assert((num_dw + ctx->dma.cs.current.cdw) <= ctx->dma in r600_need_dma_space()
[all...]
H A Dr600_buffer_common.c42 if (radeon_emitted(&ctx->dma.cs, 0) && in r600_rings_is_buffer_referenced()
43 ctx->ws->cs_is_buffer_referenced(&ctx->dma.cs, buf, usage)) { in r600_rings_is_buffer_referenced()
78 if (radeon_emitted(&ctx->dma.cs, 0) && in r600_buffer_map_sync_with_rings()
79 ctx->ws->cs_is_buffer_referenced(&ctx->dma.cs, in r600_buffer_map_sync_with_rings()
82 ctx->dma.flush(ctx, PIPE_FLUSH_ASYNC, NULL); in r600_buffer_map_sync_with_rings()
85 ctx->dma.flush(ctx, 0, NULL); in r600_buffer_map_sync_with_rings()
97 if (ctx->dma.cs.priv) in r600_buffer_map_sync_with_rings()
98 ctx->ws->cs_sync_flush(&ctx->dma.cs); in r600_buffer_map_sync_with_rings()
309 (dword_aligned && (rctx->dma.cs.priv || in r600_can_dma_copy_buffer()
H A Devergreen_hw_context.c38 struct radeon_cmdbuf *cs = &rctx->b.dma.cs; in evergreen_dma_copy_buffer()
67 radeon_add_to_buffer_list(&rctx->b, &rctx->b.dma, rsrc, RADEON_USAGE_READ); in evergreen_dma_copy_buffer()
68 radeon_add_to_buffer_list(&rctx->b, &rctx->b.dma, rdst, RADEON_USAGE_WRITE); in evergreen_dma_copy_buffer()
H A Dr600_hw_context.c37 if (radeon_emitted(&ctx->b.dma.cs, 0)) in r600_need_cs_space()
38 ctx->b.dma.flush(ctx, PIPE_FLUSH_ASYNC, NULL); in r600_need_cs_space()
586 struct radeon_cmdbuf *cs = &rctx->b.dma.cs; in r600_dma_copy_buffer()
604 radeon_add_to_buffer_list(&rctx->b, &rctx->b.dma, rsrc, RADEON_USAGE_READ); in r600_dma_copy_buffer()
605 radeon_add_to_buffer_list(&rctx->b, &rctx->b.dma, rdst, RADEON_USAGE_WRITE); in r600_dma_copy_buffer()
H A Devergreen_compute.c732 if (radeon_emitted(&rctx->b.dma.cs, 0)) { in compute_emit_cs()
733 rctx->b.dma.flush(rctx, PIPE_FLUSH_ASYNC, NULL); in compute_emit_cs()
/third_party/ltp/testcases/kernel/device-drivers/usb/tusb/
H A Dst_tusb.h39 /* allocate dma-consistent buffer for URB_DMA_NOMAPPING */
42 dma_addr_t *dma);
44 void *addr, dma_addr_t dma);
/third_party/mesa3d/src/imagination/vulkan/pds/
H A Dpvr_xgl_pds.c391 * Generates the PDS vertex primary program for the dma's listed in the input
529 for (uint32_t dma = 0; dma < input_program->dma_count; dma++) { in pvr_pds_generate_vertex_primary_program()
742 for (uint32_t dma = 0; dma < input_program->dma_count; dma++) { in pvr_pds_generate_vertex_primary_program()
743 uint32_t const_base = dma * PVR_PDS_DDMAD_NUM_CONSTS; in pvr_pds_generate_vertex_primary_program()
747 struct pvr_pds_vertex_dma *vertex_dma = &input_program->dma_list[dma]; in pvr_pds_generate_vertex_primary_program()
750 pvr_debug_pds_note("Vertex Attribute DMA %d (last=%d)", dma, last_DM in pvr_pds_generate_vertex_primary_program()
[all...]
H A Dpvr_pds.c1344 /* Set up the dma transfer control word. */ in pvr_pds_vertex_shader()
1371 /* If this is the last dma then also set the last flag. */ in pvr_pds_vertex_shader()
1495 /* If this is the last dma then also set the last flag. */ in pvr_pds_vertex_shader()
2882 for (uint32_t dma = 0; dma < program->num_texture_dma_kicks; dma++) { in pvr_pds_pixel_shader_uniform_texture_code()
2890 /* END */ dma == (program->num_texture_dma_kicks - 1), in pvr_pds_pixel_shader_uniform_texture_code()
2966 for (uint32_t dma = 0; dma < program->num_uniform_dma_kicks; dma in pvr_pds_pixel_shader_uniform_texture_code()
[all...]
/third_party/toybox/toys/net/
H A Difconfig.c285 ifre.ifr_map.mem_start || ifre.ifr_map.dma || ifre.ifr_map.base_addr)) in display_ifconfig()
293 if(ifre.ifr_map.dma) xprintf("DMA chan:%x ", ifre.ifr_map.dma); in display_ifconfig()
/third_party/toybox/porting/liteos_a/toys/net/
H A Difconfig.c284 ifre.ifr_map.mem_start || ifre.ifr_map.dma || ifre.ifr_map.base_addr)) in display_ifconfig()
292 if(ifre.ifr_map.dma) xprintf("DMA chan:%x ", ifre.ifr_map.dma); in display_ifconfig()
/third_party/musl/porting/liteos_a/kernel/include/net/
H A Dif.h69 unsigned char dma; member
/third_party/musl/porting/uniproton/kernel/include/net/
H A Dif.h69 unsigned char dma; member
/third_party/musl/porting/liteos_m/kernel/include/net/
H A Dif.h69 unsigned char dma; member
/third_party/musl/porting/liteos_m_iccarm/kernel/include/net/
H A Dif.h69 unsigned char dma; member
/third_party/musl/include/net/
H A Dif.h69 unsigned char dma; member
/third_party/libdrm/
H A Dxf86drm.c1885 drm_dma_t dma; in drmDMA() local
1888 dma.context = request->context; in drmDMA()
1889 dma.send_count = request->send_count; in drmDMA()
1890 dma.send_indices = request->send_list; in drmDMA()
1891 dma.send_sizes = request->send_sizes; in drmDMA()
1892 dma.flags = (enum drm_dma_flags)request->flags; in drmDMA()
1893 dma.request_count = request->request_count; in drmDMA()
1894 dma.request_size = request->request_size; in drmDMA()
1895 dma.request_indices = request->request_list; in drmDMA()
1896 dma in drmDMA()
[all...]
/third_party/NuttX/drivers/usbdev/gadget/fconfig/src/
H A Dusbd_generic_epfile.c88 req->dma = (DMA_ADDR_T)VMM_TO_UNCACHED_ADDR((unsigned long)req->buf); in generic_epfile_submit_sync()
179 req->dma = (DMA_ADDR_T)VMM_TO_UNCACHED_ADDR((unsigned long)req->buf); in generic_epfile_submit_request()
/third_party/libdrm/include/drm/
H A Dradeon_drm.h179 #define RADEON_CMD_DMA_DISCARD 4 /* discard current dma buf */
207 } dma; member
271 } dma; member
659 * - allows more than one dma buffer to be referenced per ioctl
/third_party/ffmpeg/libavutil/
H A Dhwcontext_drm.c29 #include <linux/dma-buf.h>
/third_party/NuttX/drivers/usbdev/gadget/
H A Dusbdev.h271 uintptr_t dma; member
H A Dcdcacm.c674 req->dma = (DMA_ADDR_T)VMM_TO_UNCACHED_ADDR((unsigned long)req->buf); in cdcacm_allocreq()
676 req->dma = NULL; in cdcacm_allocreq()
/third_party/mesa3d/src/vulkan/wsi/
H A Dwsi_common_drm.c35 #include <linux/dma-buf.h>

Completed in 30 milliseconds

12