/kernel/linux/linux-5.10/arch/s390/pci/ |
H A D | pci_dma.c | 13 #include <linux/dma-map-ops.h> 434 /* Map a segment into a contiguous dma address area */ 486 struct scatterlist *s = sg, *start = sg, *dma = sg; in s390_dma_map_sg() local 501 &dma->dma_address, dir)) in s390_dma_map_sg() 504 dma->dma_address += offset; in s390_dma_map_sg() 505 dma->dma_length = size - offset; in s390_dma_map_sg() 509 dma = sg_next(dma); in s390_dma_map_sg() 514 if (__s390_dma_map_sg(dev, start, size, &dma->dma_address, dir)) in s390_dma_map_sg() 517 dma in s390_dma_map_sg() [all...] |
/kernel/linux/linux-6.6/arch/s390/pci/ |
H A D | pci_dma.c | 13 #include <linux/dma-map-ops.h> 456 /* Map a segment into a contiguous dma address area */ 508 struct scatterlist *s = sg, *start = sg, *dma = sg; in s390_dma_map_sg() local 522 &dma->dma_address, dir); in s390_dma_map_sg() 526 dma->dma_address += offset; in s390_dma_map_sg() 527 dma->dma_length = size - offset; in s390_dma_map_sg() 531 dma = sg_next(dma); in s390_dma_map_sg() 536 ret = __s390_dma_map_sg(dev, start, size, &dma->dma_address, dir); in s390_dma_map_sg() 540 dma in s390_dma_map_sg() [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/intel/ixgbe/ |
H A D | ixgbe_xsk.c | 149 dma_addr_t dma; in ixgbe_alloc_rx_buffers_zc() local 167 dma = xsk_buff_xdp_get_dma(bi->xdp); in ixgbe_alloc_rx_buffers_zc() 172 rx_desc->read.pkt_addr = cpu_to_le64(dma); in ixgbe_alloc_rx_buffers_zc() 390 dma_addr_t dma; in ixgbe_xmit_zc() local 405 dma = xsk_buff_raw_get_dma(pool, desc.addr); in ixgbe_xmit_zc() 406 xsk_buff_raw_dma_sync_for_device(pool, dma, desc.len); in ixgbe_xmit_zc() 414 tx_desc->read.buffer_addr = cpu_to_le64(dma); in ixgbe_xmit_zc() 443 dma_unmap_addr(tx_bi, dma), in ixgbe_clean_xdp_tx_buffer()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_xsk.c | 250 dma_addr_t dma; in i40e_alloc_rx_buffers_zc() local 262 dma = xsk_buff_xdp_get_dma(xdp); in i40e_alloc_rx_buffers_zc() 263 rx_desc->read.pkt_addr = cpu_to_le64(dma); in i40e_alloc_rx_buffers_zc() 462 dma_addr_t dma; in i40e_xmit_zc() local 468 dma = xsk_buff_raw_get_dma(xdp_ring->xsk_pool, desc.addr); in i40e_xmit_zc() 469 xsk_buff_raw_dma_sync_for_device(xdp_ring->xsk_pool, dma, in i40e_xmit_zc() 476 tx_desc->buffer_addr = cpu_to_le64(dma); in i40e_xmit_zc() 514 dma_unmap_addr(tx_bi, dma), in i40e_clean_xdp_tx_buffer()
|
/kernel/linux/linux-5.10/drivers/media/pci/intel/ipu3/ |
H A D | ipu3-cio2.h | 132 #define CIO2_REG_INT_EN_IOS(dma) (1 << (((dma) >> 1) + 12)) 137 #define CIO2_INT_IOC(dma) (1 << ((dma) < 4 ? (dma) : ((dma) >> 1) + 2)) 140 #define CIO2_INT_IOS_IOLN(dma) (1 << (((dma) >> 1) + 12))
|
/kernel/linux/linux-6.6/drivers/net/ethernet/intel/ixgbe/ |
H A D | ixgbe_xsk.c | 156 dma_addr_t dma; in ixgbe_alloc_rx_buffers_zc() local 174 dma = xsk_buff_xdp_get_dma(bi->xdp); in ixgbe_alloc_rx_buffers_zc() 179 rx_desc->read.pkt_addr = cpu_to_le64(dma); in ixgbe_alloc_rx_buffers_zc() 403 dma_addr_t dma; in ixgbe_xmit_zc() local 418 dma = xsk_buff_raw_get_dma(pool, desc.addr); in ixgbe_xmit_zc() 419 xsk_buff_raw_dma_sync_for_device(pool, dma, desc.len); in ixgbe_xmit_zc() 427 tx_desc->read.buffer_addr = cpu_to_le64(dma); in ixgbe_xmit_zc() 456 dma_unmap_addr(tx_bi, dma), in ixgbe_clean_xdp_tx_buffer()
|
/third_party/mesa3d/src/gallium/drivers/svga/ |
H A D | svga_resource_buffer.c | 167 if (sbuf->dma.pending) { in svga_buffer_transfer_map() 197 if (sbuf->dma.pending) { in svga_buffer_transfer_map() 213 sbuf->dma.flags.discard = TRUE; in svga_buffer_transfer_map() 223 sbuf->dma.flags.unsynchronized = TRUE; in svga_buffer_transfer_map() 233 if (sbuf->dma.pending) { in svga_buffer_transfer_map() 268 sbuf->dma.flags.unsynchronized = FALSE; in svga_buffer_transfer_map() 391 sbuf->dma.flags.discard = TRUE; in svga_buffer_transfer_unmap() 427 assert(!sbuf->dma.pending); in svga_resource_destroy()
|
/kernel/linux/linux-6.6/drivers/media/platform/verisilicon/ |
H A D | rockchip_vpu981_hw_av1_dec.c | 230 av1_dec->db_data_col.dma); in rockchip_vpu981_av1_dec_tiles_free() 236 av1_dec->db_ctrl_col.dma); in rockchip_vpu981_av1_dec_tiles_free() 241 av1_dec->cdef_col.cpu, av1_dec->cdef_col.dma); in rockchip_vpu981_av1_dec_tiles_free() 246 av1_dec->sr_col.cpu, av1_dec->sr_col.dma); in rockchip_vpu981_av1_dec_tiles_free() 251 av1_dec->lr_col.cpu, av1_dec->lr_col.dma); in rockchip_vpu981_av1_dec_tiles_free() 274 &av1_dec->db_data_col.dma, in rockchip_vpu981_av1_dec_tiles_reallocate() 282 &av1_dec->db_ctrl_col.dma, in rockchip_vpu981_av1_dec_tiles_reallocate() 290 &av1_dec->cdef_col.dma, in rockchip_vpu981_av1_dec_tiles_reallocate() 298 &av1_dec->sr_col.dma, in rockchip_vpu981_av1_dec_tiles_reallocate() 306 &av1_dec->lr_col.dma, in rockchip_vpu981_av1_dec_tiles_reallocate() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/atheros/atl1c/ |
H A D | atl1c_main.c | 850 if (buffer_info->dma) { in atl1c_clean_buffer() 857 dma_unmap_single(&pdev->dev, buffer_info->dma, in atl1c_clean_buffer() 860 dma_unmap_page(&pdev->dev, buffer_info->dma, in atl1c_clean_buffer() 865 buffer_info->dma = 0; in atl1c_clean_buffer() 964 adapter->ring_header.desc, adapter->ring_header.dma); in atl1c_free_ring_resources() 1039 &ring_header->dma, GFP_KERNEL); in atl1c_setup_ring_resources() 1046 tpd_ring[0].dma = roundup(ring_header->dma, 8); in atl1c_setup_ring_resources() 1047 offset = tpd_ring[0].dma - ring_header->dma; in atl1c_setup_ring_resources() [all...] |
/kernel/linux/linux-5.10/drivers/ide/ |
H A D | pmac.c | 11 * Some code taken from drivers/ide/ide-dma.c: 65 * beeing done by the generic code about the kind of dma controller 66 * and format of the dma table. This will have to be fixed though. 1179 if (macio_request_resource(mdev, 1, "ide-pmac (dma)")) in pmac_ide_macio_attach() 1457 volatile struct dbdma_regs __iomem *dma = pmif->dma_regs; in pmac_ide_build_dmatable() local 1466 writel((RUN|PAUSE|FLUSH|WAKE|DEAD) << 16, &dma->control); in pmac_ide_build_dmatable() 1467 while (readl(&dma->status) & RUN) in pmac_ide_build_dmatable() 1516 writel(hwif->dmatable_dma, &dma->cmdptr); in pmac_ide_build_dmatable() 1558 volatile struct dbdma_regs __iomem *dma; in pmac_ide_dma_start() local 1560 dma in pmac_ide_dma_start() 1575 volatile struct dbdma_regs __iomem *dma = pmif->dma_regs; pmac_ide_dma_end() local 1599 volatile struct dbdma_regs __iomem *dma = pmif->dma_regs; pmac_ide_dma_test_irq() local 1653 volatile struct dbdma_regs __iomem *dma = pmif->dma_regs; pmac_ide_dma_lost_irq() local [all...] |
/kernel/linux/linux-5.10/drivers/usb/musb/ |
H A D | cppi_dma.c | 126 dma_addr_t dma; in cppi_pool_init() local 128 bd = dma_pool_alloc(cppi->pool, GFP_KERNEL, &dma); in cppi_pool_init() 129 bd->dma = dma; in cppi_pool_init() 149 dma_pool_free(cppi->pool, bd, bd->dma); in cppi_pool_free() 255 /* While dma channel is allocated, we only want the core irqs active 257 * Except for TX irqs, where dma done != fifo empty and reusable ... 261 * REVISIT same issue applies to pure PIO usage too, and non-cppi dma... 433 tag, bd->dma, in cppi_dump_rxbd() 520 * existing queue, processing completed "dma buffer [all...] |
/kernel/linux/linux-5.10/sound/ppc/ |
H A D | pmac.c | 17 #include <linux/dma-mapping.h> 133 while ((in_le32(&rec->dma->status) & RUN) && timeout-- > 0) in snd_pmac_wait_ack() 155 out_le32(&rec->dma->control, (RUN|WAKE|FLUSH|PAUSE) << 16); in snd_pmac_dma_stop() 164 out_le32(&rec->dma->cmdptr, cmd->addr); in snd_pmac_dma_set_command() 172 out_le32(&rec->dma->control, status | (status << 16)); in snd_pmac_dma_run() 262 (void)in_le32(&rec->dma->status); in snd_pmac_pcm_trigger() 296 #if 1 /* hmm.. how can we get the current dma pointer?? */ in snd_pmac_pcm_pointer() 391 (void)in_le32(&rec->dma->status); in snd_pmac_pcm_dead_xfer() 392 out_le32(&rec->dma->control, (RUN|PAUSE|FLUSH|WAKE) << 16); in snd_pmac_pcm_dead_xfer() 420 out_le32(&rec->dma in snd_pmac_pcm_dead_xfer() [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/savage/ |
H A D | savage_bci.c | 218 struct drm_device_dma *dma = dev->dma; in savage_freelist_init() local 222 DRM_DEBUG("count=%d\n", dma->buf_count); in savage_freelist_init() 232 for (i = 0; i < dma->buf_count; i++) { in savage_freelist_init() 233 buf = dma->buflist[i]; in savage_freelist_init() 689 DRM_ERROR("invalid dma memory type %d!\n", init->dma_type); in savage_do_init_bci() 761 if (dev->dma && dev->dma->buflist) { in savage_do_init_bci() 1016 struct drm_device_dma *dma = dev->dma; in savage_bci_buffers() local 1049 struct drm_device_dma *dma = dev->dma; savage_reclaim_buffers() local [all...] |
/kernel/linux/linux-6.6/sound/ppc/ |
H A D | pmac.c | 17 #include <linux/dma-mapping.h> 133 while ((in_le32(&rec->dma->status) & RUN) && timeout-- > 0) in snd_pmac_wait_ack() 155 out_le32(&rec->dma->control, (RUN|WAKE|FLUSH|PAUSE) << 16); in snd_pmac_dma_stop() 164 out_le32(&rec->dma->cmdptr, cmd->addr); in snd_pmac_dma_set_command() 172 out_le32(&rec->dma->control, status | (status << 16)); in snd_pmac_dma_run() 262 (void)in_le32(&rec->dma->status); in snd_pmac_pcm_trigger() 296 #if 1 /* hmm.. how can we get the current dma pointer?? */ in snd_pmac_pcm_pointer() 391 (void)in_le32(&rec->dma->status); in snd_pmac_pcm_dead_xfer() 392 out_le32(&rec->dma->control, (RUN|PAUSE|FLUSH|WAKE) << 16); in snd_pmac_pcm_dead_xfer() 420 out_le32(&rec->dma in snd_pmac_pcm_dead_xfer() [all...] |
/kernel/linux/linux-5.10/drivers/dma/ |
H A D | pxa_dma.c | 11 #include <linux/dma-mapping.h> 22 #include <linux/dma/pxa-dma.h> 25 #include "virt-dma.h" 377 * dma channel priorities in lookup_phy() 501 "%s(): no free dma channel\n", __func__); in pxad_launch_chan() 521 dma_addr_t dma = sw_desc->hw_desc[sw_desc->nb_desc - 2]->ddadr; in set_updater_desc() local 524 updater->dsadr = dma; in set_updater_desc() 525 updater->dtadr = dma + 8; in set_updater_desc() 723 dma_addr_t dma; in pxad_free_desc() local 742 dma_addr_t dma; pxad_alloc_desc() local 971 dma_addr_t dma; pxad_prep_slave_sg() local 1020 dma_addr_t dma; pxad_prep_dma_cyclic() local [all...] |
/kernel/linux/linux-5.10/drivers/crypto/ux500/hash/ |
H A D | hash_core.c | 18 #include <linux/dma-mapping.h> 118 dma_cap_zero(device_data->dma.mask); in hash_dma_setup_channel() 119 dma_cap_set(DMA_SLAVE, device_data->dma.mask); in hash_dma_setup_channel() 121 device_data->dma.cfg_mem2hash = platform_data->mem_to_engine; in hash_dma_setup_channel() 122 device_data->dma.chan_mem2hash = in hash_dma_setup_channel() 123 dma_request_channel(device_data->dma.mask, in hash_dma_setup_channel() 125 device_data->dma.cfg_mem2hash); in hash_dma_setup_channel() 127 dmaengine_slave_config(device_data->dma.chan_mem2hash, &conf); in hash_dma_setup_channel() 129 init_completion(&device_data->dma.complete); in hash_dma_setup_channel() 136 complete(&ctx->device->dma in hash_dma_callback() [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/ti/ |
H A D | netcp_core.c | 621 dma_addr_t dma; in netcp_empty_rx_queue() local 624 dma = knav_queue_pop(netcp->rx_queue, &dma_sz); in netcp_empty_rx_queue() 625 if (!dma) in netcp_empty_rx_queue() 628 desc = knav_pool_desc_unmap(netcp->rx_pool, dma, dma_sz); in netcp_empty_rx_queue() 793 dma_addr_t dma; in netcp_free_rx_buf() local 797 while ((dma = knav_queue_pop(netcp->rx_fdq[fdq], &dma_sz))) { in netcp_free_rx_buf() 798 desc = knav_pool_desc_unmap(netcp->rx_pool, dma, dma_sz); in netcp_free_rx_buf() 804 get_org_pkt_info(&dma, &buf_len, desc); in netcp_free_rx_buf() 810 if (unlikely(!dma)) { in netcp_free_rx_buf() 823 dma_unmap_single(netcp->dev, dma, buf_le in netcp_free_rx_buf() 858 dma_addr_t dma; netcp_allocate_rx_buf() local 1010 dma_addr_t dma; netcp_process_tx_compl_packets() local 1179 dma_addr_t dma; netcp_tx_submit_skb() local [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/freescale/enetc/ |
H A D | enetc_qos.c | 56 dma_addr_t dma; in enetc_setup_taprio() local 114 dma = dma_map_single(&priv->si->pdev->dev, gcl_data, in enetc_setup_taprio() 116 if (dma_mapping_error(&priv->si->pdev->dev, dma)) { in enetc_setup_taprio() 122 cbd.addr[0] = lower_32_bits(dma); in enetc_setup_taprio() 123 cbd.addr[1] = upper_32_bits(dma); in enetc_setup_taprio() 133 dma_unmap_single(&priv->si->pdev->dev, dma, data_size, DMA_TO_DEVICE); in enetc_setup_taprio() 477 dma_addr_t dma; in enetc_streamid_hw_set() local 498 dma = dma_map_single(&priv->si->pdev->dev, si_data, in enetc_streamid_hw_set() 500 if (dma_mapping_error(&priv->si->pdev->dev, dma)) { in enetc_streamid_hw_set() 506 cbd.addr[0] = lower_32_bits(dma); in enetc_streamid_hw_set() 639 dma_addr_t dma; enetc_streamcounter_hw_get() local 732 dma_addr_t dma; enetc_streamgate_hw_set() local [all...] |
/kernel/linux/linux-6.6/drivers/dma/ |
H A D | pxa_dma.c | 11 #include <linux/dma-mapping.h> 22 #include <linux/dma/pxa-dma.h> 25 #include "virt-dma.h" 377 * dma channel priorities in lookup_phy() 501 "%s(): no free dma channel\n", __func__); in pxad_launch_chan() 521 dma_addr_t dma = sw_desc->hw_desc[sw_desc->nb_desc - 2]->ddadr; in set_updater_desc() local 524 updater->dsadr = dma; in set_updater_desc() 525 updater->dtadr = dma + 8; in set_updater_desc() 722 dma_addr_t dma; in pxad_free_desc() local 741 dma_addr_t dma; pxad_alloc_desc() local 969 dma_addr_t dma; pxad_prep_slave_sg() local 1018 dma_addr_t dma; pxad_prep_dma_cyclic() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/ti/ |
H A D | netcp_core.c | 620 dma_addr_t dma; in netcp_empty_rx_queue() local 623 dma = knav_queue_pop(netcp->rx_queue, &dma_sz); in netcp_empty_rx_queue() 624 if (!dma) in netcp_empty_rx_queue() 627 desc = knav_pool_desc_unmap(netcp->rx_pool, dma, dma_sz); in netcp_empty_rx_queue() 792 dma_addr_t dma; in netcp_free_rx_buf() local 796 while ((dma = knav_queue_pop(netcp->rx_fdq[fdq], &dma_sz))) { in netcp_free_rx_buf() 797 desc = knav_pool_desc_unmap(netcp->rx_pool, dma, dma_sz); in netcp_free_rx_buf() 803 get_org_pkt_info(&dma, &buf_len, desc); in netcp_free_rx_buf() 809 if (unlikely(!dma)) { in netcp_free_rx_buf() 822 dma_unmap_single(netcp->dev, dma, buf_le in netcp_free_rx_buf() 857 dma_addr_t dma; netcp_allocate_rx_buf() local 1009 dma_addr_t dma; netcp_process_tx_compl_packets() local 1178 dma_addr_t dma; netcp_tx_submit_skb() local [all...] |
/kernel/linux/linux-5.10/include/linux/ |
H A D | dmaengine.h | 49 * automatically set as dma devices are registered. 73 * enum dma_transfer_direction - dma transfer mode and direction indicator 312 * @device: ptr to the dma device who supplies this channel, always !%NULL 384 * struct dma_slave_config - dma slave channel runtime config 418 * @slave_id: Slave requester id. Only valid for slave channels. The dma 419 * slave peripheral will have unique id as dma requester which need to be 482 * each type, the dma controller should set BIT(<TYPE>) and same 577 * ---dma generic offload fields--- 784 * @dev: struct device reference for dma mapping api 792 * each type, the dma controlle 1235 dma_set_maxpq(struct dma_device *dma, int maxpq, int has_pq_continue) dma_set_maxpq() argument 1254 dma_dev_has_pq_continue(struct dma_device *dma) dma_dev_has_pq_continue() argument 1259 dma_dev_to_maxpq(struct dma_device *dma) dma_dev_to_maxpq() argument 1277 dma_maxpq(struct dma_device *dma, enum dma_ctrl_flags flags) dma_maxpq() argument [all...] |
/kernel/linux/linux-6.6/drivers/accel/ivpu/ |
H A D | ivpu_mmu.c | 283 cdtab->base = dmam_alloc_coherent(vdev->drm.dev, size, &cdtab->dma, GFP_KERNEL); in ivpu_mmu_cdtab_alloc() 287 ivpu_dbg(vdev, MMU, "CDTAB alloc: dma=%pad size=%zu\n", &cdtab->dma, size); in ivpu_mmu_cdtab_alloc() 298 strtab->base = dmam_alloc_coherent(vdev->drm.dev, size, &strtab->dma, GFP_KERNEL); in ivpu_mmu_strtab_alloc() 304 strtab->dma_q |= strtab->dma & IVPU_MMU_STRTAB_BASE_ADDR_MASK; in ivpu_mmu_strtab_alloc() 306 ivpu_dbg(vdev, MMU, "STRTAB alloc: dma=%pad dma_q=%pad size=%zu\n", in ivpu_mmu_strtab_alloc() 307 &strtab->dma, &strtab->dma_q, size); in ivpu_mmu_strtab_alloc() 317 q->base = dmam_alloc_coherent(vdev->drm.dev, IVPU_MMU_CMDQ_SIZE, &q->dma, GFP_KERNEL); in ivpu_mmu_cmdq_alloc() 322 q->dma_q |= q->dma & IVPU_MMU_Q_BASE_ADDR_MASK; in ivpu_mmu_cmdq_alloc() 325 ivpu_dbg(vdev, MMU, "CMDQ alloc: dma in ivpu_mmu_cmdq_alloc() [all...] |
/kernel/linux/linux-6.6/include/linux/ |
H A D | dmaengine.h | 49 * automatically set as dma devices are registered. 73 * enum dma_transfer_direction - dma transfer mode and direction indicator 311 * @device: ptr to the dma device who supplies this channel, always !%NULL 359 * @chan_dma_dev: The channel is using custom/different dma-mapping 387 * struct dma_slave_config - dma slave channel runtime config 486 * each type, the dma controller should set BIT(<TYPE>) and same 581 * ---dma generic offload fields--- 791 * @dev: struct device reference for dma mapping api 800 * each type, the dma controller should set BIT(<TYPE>) and same 822 * @device_prep_slave_sg: prepares a slave dma operatio 1249 dma_set_maxpq(struct dma_device *dma, int maxpq, int has_pq_continue) dma_set_maxpq() argument 1268 dma_dev_has_pq_continue(struct dma_device *dma) dma_dev_has_pq_continue() argument 1273 dma_dev_to_maxpq(struct dma_device *dma) dma_dev_to_maxpq() argument 1291 dma_maxpq(struct dma_device *dma, enum dma_ctrl_flags flags) dma_maxpq() argument [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/intel/igbvf/ |
H A D | netdev.c | 189 buffer_info->dma = dma_map_single(&pdev->dev, skb->data, in igbvf_alloc_rx_buffers() 192 if (dma_mapping_error(&pdev->dev, buffer_info->dma)) { in igbvf_alloc_rx_buffers() 205 rx_desc->read.hdr_addr = cpu_to_le64(buffer_info->dma); in igbvf_alloc_rx_buffers() 207 rx_desc->read.pkt_addr = cpu_to_le64(buffer_info->dma); in igbvf_alloc_rx_buffers() 290 dma_unmap_single(&pdev->dev, buffer_info->dma, in igbvf_clean_rx_irq() 293 buffer_info->dma = 0; in igbvf_clean_rx_irq() 299 dma_unmap_single(&pdev->dev, buffer_info->dma, in igbvf_clean_rx_irq() 302 buffer_info->dma = 0; in igbvf_clean_rx_irq() 337 buffer_info->dma = next_buffer->dma; in igbvf_clean_rx_irq() [all...] |
/kernel/linux/linux-6.6/drivers/crypto/hisilicon/hpre/ |
H A D | hpre_crypto.c | 14 #include <linux/dma-mapping.h> 251 dev_err(dev, "dma map data err!\n"); in hpre_get_data_dma_addr() 1448 dma_addr_t dma; in hpre_ecdh_hw_data_clr_all() local 1450 dma = le64_to_cpu(sqe->in); in hpre_ecdh_hw_data_clr_all() 1451 if (unlikely(dma_mapping_error(dev, dma))) in hpre_ecdh_hw_data_clr_all() 1455 dma_free_coherent(dev, ctx->key_sz << 2, req->src, dma); in hpre_ecdh_hw_data_clr_all() 1457 dma = le64_to_cpu(sqe->out); in hpre_ecdh_hw_data_clr_all() 1458 if (unlikely(dma_mapping_error(dev, dma))) in hpre_ecdh_hw_data_clr_all() 1462 dma_free_coherent(dev, ctx->key_sz << 1, req->dst, dma); in hpre_ecdh_hw_data_clr_all() 1464 dma_unmap_single(dev, dma, ct in hpre_ecdh_hw_data_clr_all() 1537 dma_addr_t dma = 0; hpre_ecdh_src_data_init() local 1566 dma_addr_t dma; hpre_ecdh_dst_data_init() local 1764 dma_addr_t dma; hpre_curve25519_hw_data_clr_all() local 1861 dma_addr_t dma = 0; hpre_curve25519_src_init() local 1918 dma_addr_t dma; hpre_curve25519_dst_init() local [all...] |