Lines Matching refs:dma

18 #include <linux/dma-mapping.h>
30 /* Maps the buffer to dma */
36 struct dma_controller *dma = musb->dma_controller;
40 if (!is_dma_capable() || !musb_ep->dma)
47 if (dma->is_compatible)
48 compatible = dma->is_compatible(musb_ep->dma,
54 if (request->request.dma == DMA_ADDR_INVALID) {
69 request->request.dma = dma_addr;
73 request->request.dma,
82 /* Unmap the buffer from dma and maps it back to cpu */
88 if (!is_buffer_mapped(request) || !musb_ep->dma)
91 if (request->request.dma == DMA_ADDR_INVALID) {
98 request->request.dma,
103 request->request.dma = DMA_ADDR_INVALID;
106 request->request.dma,
143 if (!dma_mapping_error(&musb->g.dev, request->dma))
166 if (is_dma_capable() && ep->dma) {
187 value = c->channel_abort(ep->dma);
189 c->channel_release(ep->dma);
190 ep->dma = NULL;
205 * from the usb core ... sequenced a bit differently from mentor dma.
242 if (dma_channel_status(musb_ep->dma) == MUSB_DMA_STATUS_BUSY) {
243 musb_dbg(musb, "dma pending...");
277 musb_ep->dma->max_len);
279 use_dma = (request->dma != DMA_ADDR_INVALID && request_size);
285 musb_ep->dma->desired_mode = 0;
287 musb_ep->dma->desired_mode = 1;
290 musb_ep->dma, musb_ep->packet_sz,
291 musb_ep->dma->desired_mode,
292 request->dma + request->actual, request_size);
294 if (musb_ep->dma->desired_mode == 0) {
346 * OK since the transfer dma glue (between CPPI and
358 musb_ep->dma, musb_ep->packet_sz,
360 request->dma + request->actual,
363 c->channel_release(musb_ep->dma);
364 musb_ep->dma = NULL;
371 musb_ep->dma, musb_ep->packet_sz,
373 request->dma + request->actual,
380 * Unmap the dma buffer back to cpu if dma channel
395 musb_ep->end_point.name, use_dma ? "dma" : "pio",
414 struct dma_channel *dma;
423 dma = is_dma_capable() ? musb_ep->dma : NULL;
445 if (dma_channel_status(dma) == MUSB_DMA_STATUS_BUSY) {
450 musb_dbg(musb, "%s dma still busy?", musb_ep->end_point.name);
458 if (dma && (csr & MUSB_TXCSR_DMAENAB)) {
465 request->actual += musb_ep->dma->actual_len;
467 epnum, csr, musb_ep->dma->actual_len, request);
545 if (dma_channel_status(musb_ep->dma) == MUSB_DMA_STATUS_BUSY) {
558 struct dma_channel *channel = musb_ep->dma;
568 request->dma + request->actual,
608 channel = musb_ep->dma;
651 musb_ep->dma->desired_mode = 1;
661 musb_ep->dma->desired_mode = 0;
668 request->dma
684 channel = musb_ep->dma;
707 musb_ep->dma->desired_mode = 0;
709 musb_ep->dma->desired_mode = 1;
718 request->dma
735 struct dma_channel *channel = musb_ep->dma;
736 u32 dma_addr = request->dma + request->actual;
749 * Unmap the dma buffer back to cpu if dma channel
798 struct dma_channel *dma;
816 dma = is_dma_capable() ? musb_ep->dma : NULL;
819 csr, dma ? " (dma)" : "", request);
842 if (dma_channel_status(dma) == MUSB_DMA_STATUS_BUSY) {
849 if (dma && (csr & MUSB_RXCSR_DMAENAB)) {
856 request->actual += musb_ep->dma->actual_len;
861 if ((dma->desired_mode == 0 && !hw_ep->rx_double_buffered)
862 || (dma->actual_len
871 && (musb_ep->dma->actual_len
1057 musb_ep->dma = c->channel_alloc(c, hw_ep,
1060 musb_ep->dma = NULL;
1071 musb_ep->dma ? "dma, " : "",
1139 request->request.dma = DMA_ADDR_INVALID;
1165 dma_addr_t dma;
1299 /* ... else abort the dma transfer ... */
1300 else if (is_dma_capable() && musb_ep->dma) {
1305 status = c->channel_abort(musb_ep->dma);