/kernel/linux/linux-5.10/drivers/net/ethernet/intel/e1000/ |
H A D | e1000.h | 28 #include <linux/dma-mapping.h> 126 dma_addr_t dma; member 140 dma_addr_t dma; member 147 dma_addr_t dma; member 168 dma_addr_t dma; member
|
/kernel/linux/linux-5.10/drivers/net/vmxnet3/ |
H A D | vmxnet3_int.h | 45 #include <asm/dma.h> 417 #define VMXNET3_GET_ADDR_LO(dma) ((u32)(dma)) 418 #define VMXNET3_GET_ADDR_HI(dma) ((u32)(((u64)(dma)) >> 32))
|
/kernel/linux/linux-5.10/drivers/gpu/drm/shmobile/ |
H A D | shmob_drm_plane.c | 27 unsigned long dma[2]; member 48 splane->dma[0] = gem->paddr + fb->offsets[0] in shmob_drm_plane_compute_base() 54 splane->dma[1] = gem->paddr + fb->offsets[1] in shmob_drm_plane_compute_base() 142 lcdc_write(sdev, LDBnBSAYR(splane->index), splane->dma[0]); in __shmob_drm_plane_setup() 144 lcdc_write(sdev, LDBnBSACR(splane->index), splane->dma[1]); in __shmob_drm_plane_setup()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/intel/e1000/ |
H A D | e1000.h | 28 #include <linux/dma-mapping.h> 126 dma_addr_t dma; member 140 dma_addr_t dma; member 147 dma_addr_t dma; member 168 dma_addr_t dma; member
|
/kernel/linux/linux-6.6/drivers/ata/ |
H A D | pata_atiixp.c | 98 * Called by both the pio and dma setup functions to set the controller 155 int dma = adev->dma_mode; in atiixp_set_dmamode() local 165 dma -= XFER_UDMA_0; in atiixp_set_dmamode() 169 udma_mode_data |= dma << (4 * dn); in atiixp_set_dmamode() 175 dma -= XFER_MW_DMA_0; in atiixp_set_dmamode() 180 mwdma_timing_data |= (mwdma_timings[dma] << timing_shift); in atiixp_set_dmamode()
|
/kernel/linux/linux-6.6/drivers/crypto/qce/ |
H A D | core.c | 7 #include <linux/dma-mapping.h> 172 * Pipe pair number depends on the actual BAM dma pipe in qce_check_version() 173 * that is used for transfers. The BAM dma pipes are passed in qce_check_version() 176 * BAM dma pipes(rx, tx) CE pipe pair id in qce_check_version() 183 qce->pipe_pair_id = qce->dma.rxchan->chan_id >> 1; in qce_check_version() 244 ret = qce_dma_request(qce->dev, &qce->dma); in qce_crypto_probe() 267 qce_dma_release(&qce->dma); in qce_crypto_probe() 286 qce_dma_release(&qce->dma); in qce_crypto_remove()
|
H A D | skcipher.c | 7 #include <linux/dma-mapping.h> 34 struct qce_result_dump *result_buf = qce->dma.result_buf; in qce_skcipher_done() 44 error = qce_dma_terminate_all(&qce->dma); in qce_skcipher_done() 46 dev_dbg(qce->dev, "skcipher dma termination error (%d)\n", in qce_skcipher_done() 108 sg_init_one(&rctx->result_sg, qce->dma.result_buf, QCE_RESULT_BUF_SZ); in qce_skcipher_async_req_handle() 144 ret = qce_dma_prep_sgs(&qce->dma, rctx->src_sg, src_nents, in qce_skcipher_async_req_handle() 150 qce_dma_issue_pending(&qce->dma); in qce_skcipher_async_req_handle() 159 qce_dma_terminate_all(&qce->dma); in qce_skcipher_async_req_handle()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlx5/core/diag/ |
H A D | rsc_dump.c | 102 dma_addr_t dma; in mlx5_rsc_dump_trigger() local 105 dma = dma_map_page(ddev, page, 0, cmd->mem_size, DMA_FROM_DEVICE); in mlx5_rsc_dump_trigger() 106 if (unlikely(dma_mapping_error(ddev, dma))) in mlx5_rsc_dump_trigger() 111 MLX5_SET64(resource_dump, cmd->cmd, address, dma); in mlx5_rsc_dump_trigger() 123 dma_unmap_page(ddev, dma, cmd->mem_size, DMA_FROM_DEVICE); in mlx5_rsc_dump_trigger()
|
/kernel/linux/linux-6.6/drivers/media/pci/cx88/ |
H A D | cx88-vbi.c | 59 VBI_LINE_LENGTH, buf->risc.dma); in cx8800_start_vbi_dma() 76 /* start dma */ in cx8800_start_vbi_dma() 87 /* stop dma */ in cx8800_stop_vbi_dma() 107 dprintk(2, "restart_queue [%p/%d]: restart dma\n", in cx8800_restart_vbi_queue() 162 risc->dma); in buffer_finish() 175 buf->risc.cpu[1] = cpu_to_le32(buf->risc.dma + 8); in buffer_queue() 177 buf->risc.jmp[1] = cpu_to_le32(buf->risc.dma + 8); in buffer_queue() 188 prev->risc.jmp[1] = cpu_to_le32(buf->risc.dma); in buffer_queue()
|
/kernel/linux/linux-6.6/crypto/async_tx/ |
H A D | async_xor.c | 17 #include <linux/dma-mapping.h> 21 /* do_async_xor - dma map the pages and perform the xor with an engine */ 26 struct dma_device *dma = chan->device; in do_async_xor() local 41 xor_src_cnt = min(src_cnt, (int)dma->max_xor); in do_async_xor() 65 tx = dma->device_prep_dma_xor(chan, dma_dest, src_list, in do_async_xor() 75 tx = dma->device_prep_dma_xor(chan, dma_dest, in do_async_xor() 160 * async_xor_offs - attempt to xor a set of blocks with a dma engine. 173 * the calculation. The assumption with dma engines is that they only 227 WARN_ONCE(chan, "%s: no space for dma address conversion\n", in async_xor_offs() 252 * async_xor - attempt to xor a set of blocks with a dma engin [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/nvkm/falcon/ |
H A D | base.c | 43 const struct nvkm_falcon_func_dma *dma = nvkm_falcon_dma(falcon, &mem_type, &mem_base); in nvkm_falcon_dma_wr() local 50 if (WARN_ON(!dma->xfer)) in nvkm_falcon_dma_wr() 63 ret = dma->init(falcon, dma_addr, dmalen, mem_type, sec, &cmd); in nvkm_falcon_dma_wr() 71 dma->xfer(falcon, dst, src - dma_start, cmd); in nvkm_falcon_dma_wr() 89 if (dma->done(falcon)) in nvkm_falcon_dma_wr()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/renesas/shmobile/ |
H A D | shmob_drm_plane.c | 27 unsigned long dma[2]; member 48 splane->dma[0] = gem->dma_addr + fb->offsets[0] in shmob_drm_plane_compute_base() 54 splane->dma[1] = gem->dma_addr + fb->offsets[1] in shmob_drm_plane_compute_base() 146 lcdc_write(sdev, LDBnBSAYR(splane->index), splane->dma[0]); in __shmob_drm_plane_setup() 148 lcdc_write(sdev, LDBnBSACR(splane->index), splane->dma[1]); in __shmob_drm_plane_setup()
|
/kernel/linux/linux-6.6/drivers/scsi/arm/ |
H A D | arxescsi.c | 17 * 01-01-2000 SH 0.1.0 Added *real* pseudo dma writing 34 #include <asm/dma.h> 287 info->info.scsi.dma = NO_DMA; in arxescsi_probe() 297 info->info.dma.setup = arxescsi_dma_setup; in arxescsi_probe() 298 info->info.dma.pseudo = arxescsi_dma_pseudo; in arxescsi_probe() 299 info->info.dma.stop = arxescsi_dma_stop; in arxescsi_probe()
|
/kernel/linux/linux-6.6/sound/soc/fsl/ |
H A D | fsl_rpmsg.c | 224 rpmsg->dma = devm_clk_get_optional(&pdev->dev, "dma"); in fsl_rpmsg_probe() 225 if (IS_ERR(rpmsg->dma)) in fsl_rpmsg_probe() 226 return PTR_ERR(rpmsg->dma); in fsl_rpmsg_probe() 284 ret = clk_prepare_enable(rpmsg->dma); in fsl_rpmsg_runtime_resume() 286 dev_err(dev, "Failed to enable dma clock %d\n", ret); in fsl_rpmsg_runtime_resume() 302 clk_disable_unprepare(rpmsg->dma); in fsl_rpmsg_runtime_suspend()
|
/kernel/linux/linux-5.10/drivers/crypto/qce/ |
H A D | skcipher.c | 7 #include <linux/dma-mapping.h> 33 struct qce_result_dump *result_buf = qce->dma.result_buf; in qce_skcipher_done() 43 error = qce_dma_terminate_all(&qce->dma); in qce_skcipher_done() 45 dev_dbg(qce->dev, "skcipher dma termination error (%d)\n", in qce_skcipher_done() 107 sg_init_one(&rctx->result_sg, qce->dma.result_buf, QCE_RESULT_BUF_SZ); in qce_skcipher_async_req_handle() 143 ret = qce_dma_prep_sgs(&qce->dma, rctx->src_sg, src_nents, in qce_skcipher_async_req_handle() 149 qce_dma_issue_pending(&qce->dma); in qce_skcipher_async_req_handle() 158 qce_dma_terminate_all(&qce->dma); in qce_skcipher_async_req_handle()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/hisilicon/hns3/hns3pf/ |
H A D | hclge_cmd.c | 4 #include <linux/dma-mapping.h> 9 #include <linux/dma-direction.h> 101 dma_addr_t dma = ring->desc_dma_addr; in hclge_cmd_config_regs() local 108 lower_32_bits(dma)); in hclge_cmd_config_regs() 110 upper_32_bits(dma)); in hclge_cmd_config_regs() 119 lower_32_bits(dma)); in hclge_cmd_config_regs() 121 upper_32_bits(dma)); in hclge_cmd_config_regs()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/atheros/atlx/ |
H A D | atl1.h | 585 dma_addr_t dma; /* physical address*/ member 597 dma_addr_t dma; member 603 dma_addr_t dma; /* descriptor ring physical address */ member 615 dma_addr_t dma; /* descriptor ring physical address */ member 626 dma_addr_t dma; /* descriptor ring physical address */ member 636 dma_addr_t dma; member 642 dma_addr_t dma; member
|
/kernel/linux/linux-5.10/drivers/net/ethernet/stmicro/stmmac/ |
H A D | dwmac5.c | 247 u32 mtl, dma; in dwmac5_safety_feat_irq_status() local 254 dma = readl(ioaddr + DMA_SAFETY_INT_STATUS); in dwmac5_safety_feat_irq_status() 256 err = (mtl & MCSIS) || (dma & MCSIS); in dwmac5_safety_feat_irq_status() 263 err = (mtl & (MEUIS | MECIS)) || (dma & (MSUIS | MSCIS)); in dwmac5_safety_feat_irq_status() 264 corr = (mtl & MECIS) || (dma & MSCIS); in dwmac5_safety_feat_irq_status() 270 err = dma & (DEUIS | DECIS); in dwmac5_safety_feat_irq_status() 271 corr = dma & DECIS; in dwmac5_safety_feat_irq_status()
|
/kernel/linux/linux-5.10/drivers/pnp/ |
H A D | quirks.c | 73 struct pnp_dma *dma; in quirk_cmi8330_resources() local 89 dma = &option->u.dma; in quirk_cmi8330_resources() 90 if ((dma->flags & IORESOURCE_DMA_TYPE_MASK) == in quirk_cmi8330_resources() 92 dma->map != 0x0A) { in quirk_cmi8330_resources() 96 pnp_option_set(option), dma->map); in quirk_cmi8330_resources() 97 dma->map = 0x0A; in quirk_cmi8330_resources() 426 /* CMI 8330 interrupt and dma fix */
|
/kernel/linux/linux-5.10/drivers/scsi/lpfc/ |
H A D | lpfc_mem.c | 386 * @dma: the DMA-mapped address of the lpfc_mbuf_pool to be freed 397 __lpfc_mbuf_free(struct lpfc_hba * phba, void *virt, dma_addr_t dma) in __lpfc_mbuf_free() argument 403 pool->elements[pool->current_count].phys = dma; in __lpfc_mbuf_free() 406 dma_pool_free(phba->lpfc_mbuf_pool, virt, dma); in __lpfc_mbuf_free() 415 * @dma: the DMA-mapped address of the lpfc_mbuf_pool to be freed 425 lpfc_mbuf_free(struct lpfc_hba * phba, void *virt, dma_addr_t dma) in lpfc_mbuf_free() argument 430 __lpfc_mbuf_free(phba, virt, dma); in lpfc_mbuf_free() 463 * @dma: the DMA-mapped address of the lpfc_sg_dma_buf_pool to be freed 468 lpfc_nvmet_buf_free(struct lpfc_hba *phba, void *virt, dma_addr_t dma) in lpfc_nvmet_buf_free() argument 470 dma_pool_free(phba->lpfc_sg_dma_buf_pool, virt, dma); in lpfc_nvmet_buf_free() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/atheros/atlx/ |
H A D | atl1.h | 585 dma_addr_t dma; /* physical address*/ member 597 dma_addr_t dma; member 603 dma_addr_t dma; /* descriptor ring physical address */ member 615 dma_addr_t dma; /* descriptor ring physical address */ member 626 dma_addr_t dma; /* descriptor ring physical address */ member 636 dma_addr_t dma; member 642 dma_addr_t dma; member
|
/kernel/linux/linux-6.6/drivers/pnp/ |
H A D | quirks.c | 73 struct pnp_dma *dma; in quirk_cmi8330_resources() local 89 dma = &option->u.dma; in quirk_cmi8330_resources() 90 if ((dma->flags & IORESOURCE_DMA_TYPE_MASK) == in quirk_cmi8330_resources() 92 dma->map != 0x0A) { in quirk_cmi8330_resources() 96 pnp_option_set(option), dma->map); in quirk_cmi8330_resources() 97 dma->map = 0x0A; in quirk_cmi8330_resources() 417 /* CMI 8330 interrupt and dma fix */
|
/kernel/linux/linux-5.10/drivers/net/ethernet/intel/ixgbevf/ |
H A D | ixgbevf_main.c | 312 dma_unmap_addr(tx_buffer, dma), in ixgbevf_clean_tx_irq() 333 dma_unmap_addr(tx_buffer, dma), in ixgbevf_clean_tx_irq() 538 rx_buffer->dma, in ixgbevf_get_rx_buffer() 560 dma_unmap_page_attrs(rx_ring->dev, rx_buffer->dma, in ixgbevf_put_rx_buffer() 608 dma_addr_t dma; in ixgbevf_alloc_mapped_page() local 622 dma = dma_map_page_attrs(rx_ring->dev, page, 0, in ixgbevf_alloc_mapped_page() 629 if (dma_mapping_error(rx_ring->dev, dma)) { in ixgbevf_alloc_mapped_page() 636 bi->dma = dma; in ixgbevf_alloc_mapped_page() 670 dma_sync_single_range_for_device(rx_ring->dev, bi->dma, in ixgbevf_alloc_rx_buffers() 982 dma_addr_t dma; ixgbevf_xmit_xdp_ring() local 3976 dma_addr_t dma; ixgbevf_tx_map() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/intel/ixgbevf/ |
H A D | ixgbevf_main.c | 312 dma_unmap_addr(tx_buffer, dma), in ixgbevf_clean_tx_irq() 333 dma_unmap_addr(tx_buffer, dma), in ixgbevf_clean_tx_irq() 538 rx_buffer->dma, in ixgbevf_get_rx_buffer() 560 dma_unmap_page_attrs(rx_ring->dev, rx_buffer->dma, in ixgbevf_put_rx_buffer() 608 dma_addr_t dma; in ixgbevf_alloc_mapped_page() local 622 dma = dma_map_page_attrs(rx_ring->dev, page, 0, in ixgbevf_alloc_mapped_page() 629 if (dma_mapping_error(rx_ring->dev, dma)) { in ixgbevf_alloc_mapped_page() 636 bi->dma = dma; in ixgbevf_alloc_mapped_page() 670 dma_sync_single_range_for_device(rx_ring->dev, bi->dma, in ixgbevf_alloc_rx_buffers() 977 dma_addr_t dma; ixgbevf_xmit_xdp_ring() local 3972 dma_addr_t dma; ixgbevf_tx_map() local [all...] |
/kernel/linux/linux-5.10/arch/arm64/mm/ |
H A D | dma-mapping.c | 9 #include <linux/dma-map-ops.h> 10 #include <linux/dma-iommu.h>
|