Lines Matching defs:device
94 dev_dbg(iop_chan->device->common.dev,
112 dev_dbg(iop_chan->device->common.dev, "%s\n", __func__);
308 dev_dbg(iop_chan->device->common.dev,
352 dev_dbg(iop_chan->device->common.dev, "pending: %d\n",
399 dev_dbg(iop_chan->device->common.dev, "%s cookie: %d slot: %d\n",
414 * greater than 2x the number slots needed to satisfy a device->max_xor
426 dev_get_platdata(&iop_chan->device->pdev->dev);
441 hw_desc = (char *) iop_chan->device->dma_desc_pool_virt;
449 dma_desc = iop_chan->device->dma_desc_pool;
464 dev_dbg(iop_chan->device->common.dev,
471 iop_chan->device->common.cap_mask))
474 iop_chan->device->common.cap_mask))
490 dev_dbg(iop_chan->device->common.dev, "%s\n", __func__);
517 dev_dbg(iop_chan->device->common.dev, "%s len: %zu\n",
549 dev_dbg(iop_chan->device->common.dev,
583 dev_dbg(iop_chan->device->common.dev, "%s src_cnt: %d len: %zu\n",
620 dev_dbg(iop_chan->device->common.dev,
684 dev_dbg(iop_chan->device->common.dev, "%s src_cnt: %d len: %zu\n",
736 dev_dbg(iop_chan->device->common.dev, "%s slots_allocated %d\n",
772 dev_dbg(chan->device->common.dev, "%s\n", __func__);
785 dev_dbg(chan->device->common.dev, "%s\n", __func__);
799 dev_err(chan->device->common.dev,
831 static int iop_adma_memcpy_self_test(struct iop_adma_device *device)
842 dev_dbg(device->common.dev, "%s\n", __func__);
858 dma_chan = container_of(device->common.channels.next,
866 dest_dma = dma_map_single(dma_chan->device->dev, dest,
868 src_dma = dma_map_single(dma_chan->device->dev, src,
880 dev_err(dma_chan->device->dev,
887 dma_sync_single_for_cpu(&iop_chan->device->pdev->dev, dest_dma,
890 dev_err(dma_chan->device->dev,
906 iop_adma_xor_val_self_test(struct iop_adma_device *device)
923 dev_dbg(device->common.dev, "%s\n", __func__);
956 dma_chan = container_of(device->common.channels.next,
965 dest_dma = dma_map_page(dma_chan->device->dev, dest, 0,
968 dma_srcs[i] = dma_map_page(dma_chan->device->dev, xor_srcs[i],
980 dev_err(dma_chan->device->dev,
987 dma_sync_single_for_cpu(&iop_chan->device->pdev->dev, dest_dma,
992 dev_err(dma_chan->device->dev,
998 dma_sync_single_for_device(&iop_chan->device->pdev->dev, dest_dma,
1002 if (!dma_has_cap(DMA_XOR_VAL, dma_chan->device->cap_mask))
1013 dma_srcs[i] = dma_map_page(dma_chan->device->dev,
1026 dev_err(dma_chan->device->dev,
1033 dev_err(dma_chan->device->dev,
1042 dma_srcs[i] = dma_map_page(dma_chan->device->dev,
1055 dev_err(dma_chan->device->dev,
1062 dev_err(dma_chan->device->dev,
1080 iop_adma_pq_zero_sum_self_test(struct iop_adma_device *device)
1097 struct device *dev;
1099 dev_dbg(device->common.dev, "%s\n", __func__);
1118 dma_chan = container_of(device->common.channels.next,
1126 dev = dma_chan->device->dev;
1244 struct iop_adma_device *device = platform_get_drvdata(dev);
1249 dma_async_device_unregister(&device->common);
1252 device->dma_desc_pool_virt, device->dma_desc_pool);
1254 list_for_each_entry_safe(chan, _chan, &device->common.channels,
1260 kfree(device);
1346 iop_chan->device = adev;
1380 iop_chan->common.device = dma_dev;
1441 dev_dbg(iop_chan->device->common.dev, "%s\n", __func__);
1483 dev_err(iop_chan->device->common.dev,
1494 dev_dbg(iop_chan->device->common.dev, "%s\n", __func__);
1536 dev_err(iop_chan->device->common.dev,