Lines Matching defs:chan
44 return container_of(dchan, struct sf_pdma_chan, vchan.chan);
52 static struct sf_pdma_desc *sf_pdma_alloc_desc(struct sf_pdma_chan *chan)
60 desc->chan = chan;
74 static void sf_pdma_disclaim_chan(struct sf_pdma_chan *chan)
76 struct pdma_regs *regs = &chan->regs;
85 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan);
89 if (chan && (!len || !dest || !src)) {
90 dev_err(chan->pdma->dma_dev.dev,
95 desc = sf_pdma_alloc_desc(chan);
100 desc->async_tx = vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
102 spin_lock_irqsave(&chan->vchan.lock, iflags);
104 spin_unlock_irqrestore(&chan->vchan.lock, iflags);
112 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan);
114 memcpy(&chan->cfg, cfg, sizeof(*cfg));
121 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan);
122 struct pdma_regs *regs = &chan->regs;
130 static void sf_pdma_disable_request(struct sf_pdma_chan *chan)
132 struct pdma_regs *regs = &chan->regs;
139 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan);
143 spin_lock_irqsave(&chan->vchan.lock, flags);
144 sf_pdma_disable_request(chan);
145 kfree(chan->desc);
146 chan->desc = NULL;
147 vchan_get_all_descriptors(&chan->vchan, &head);
148 sf_pdma_disclaim_chan(chan);
149 spin_unlock_irqrestore(&chan->vchan.lock, flags);
150 vchan_dma_desc_free_list(&chan->vchan, &head);
153 static size_t sf_pdma_desc_residue(struct sf_pdma_chan *chan,
157 struct pdma_regs *regs = &chan->regs;
163 spin_lock_irqsave(&chan->vchan.lock, flags);
165 list_for_each_entry(vd, &chan->vchan.desc_submitted, node)
172 if (cookie == tx->chan->completed_cookie)
178 vd = vchan_find_desc(&chan->vchan, cookie);
187 spin_unlock_irqrestore(&chan->vchan.lock, flags);
196 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan);
202 dma_set_residue(txstate, sf_pdma_desc_residue(chan, cookie));
209 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan);
213 spin_lock_irqsave(&chan->vchan.lock, flags);
214 sf_pdma_disable_request(chan);
215 kfree(chan->desc);
216 chan->desc = NULL;
217 chan->xfer_err = false;
218 vchan_get_all_descriptors(&chan->vchan, &head);
219 spin_unlock_irqrestore(&chan->vchan.lock, flags);
220 vchan_dma_desc_free_list(&chan->vchan, &head);
225 static void sf_pdma_enable_request(struct sf_pdma_chan *chan)
227 struct pdma_regs *regs = &chan->regs;
238 static struct sf_pdma_desc *sf_pdma_get_first_pending_desc(struct sf_pdma_chan *chan)
240 struct virt_dma_chan *vchan = &chan->vchan;
251 static void sf_pdma_xfer_desc(struct sf_pdma_chan *chan)
253 struct sf_pdma_desc *desc = chan->desc;
254 struct pdma_regs *regs = &chan->regs;
257 dev_err(chan->pdma->dma_dev.dev, "NULL desc.\n");
266 chan->desc = desc;
267 chan->status = DMA_IN_PROGRESS;
268 sf_pdma_enable_request(chan);
273 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan);
276 spin_lock_irqsave(&chan->vchan.lock, flags);
278 if (!chan->desc && vchan_issue_pending(&chan->vchan)) {
280 chan->desc = sf_pdma_get_first_pending_desc(chan);
281 sf_pdma_xfer_desc(chan);
284 spin_unlock_irqrestore(&chan->vchan.lock, flags);
297 struct sf_pdma_chan *chan = from_tasklet(chan, t, done_tasklet);
300 spin_lock_irqsave(&chan->lock, flags);
301 if (chan->xfer_err) {
302 chan->retries = MAX_RETRY;
303 chan->status = DMA_COMPLETE;
304 chan->xfer_err = false;
306 spin_unlock_irqrestore(&chan->lock, flags);
308 spin_lock_irqsave(&chan->vchan.lock, flags);
309 list_del(&chan->desc->vdesc.node);
310 vchan_cookie_complete(&chan->desc->vdesc);
312 chan->desc = sf_pdma_get_first_pending_desc(chan);
313 if (chan->desc)
314 sf_pdma_xfer_desc(chan);
316 spin_unlock_irqrestore(&chan->vchan.lock, flags);
321 struct sf_pdma_chan *chan = from_tasklet(chan, t, err_tasklet);
322 struct sf_pdma_desc *desc = chan->desc;
325 spin_lock_irqsave(&chan->lock, flags);
326 if (chan->retries <= 0) {
328 spin_unlock_irqrestore(&chan->lock, flags);
332 chan->retries--;
333 chan->xfer_err = true;
334 chan->status = DMA_ERROR;
336 sf_pdma_enable_request(chan);
337 spin_unlock_irqrestore(&chan->lock, flags);
343 struct sf_pdma_chan *chan = dev_id;
344 struct pdma_regs *regs = &chan->regs;
347 spin_lock(&chan->vchan.lock);
352 tasklet_hi_schedule(&chan->done_tasklet);
355 struct sf_pdma_desc *desc = chan->desc;
361 sf_pdma_xfer_desc(chan);
364 spin_unlock(&chan->vchan.lock);
371 struct sf_pdma_chan *chan = dev_id;
372 struct pdma_regs *regs = &chan->regs;
374 spin_lock(&chan->lock);
376 spin_unlock(&chan->lock);
378 tasklet_schedule(&chan->err_tasklet);
401 struct sf_pdma_chan *chan;
404 chan = &pdma->chans[i];
411 dev_name(&pdev->dev), (void *)chan);
417 chan->txirq = irq;
424 dev_name(&pdev->dev), (void *)chan);
430 chan->errirq = irq;
451 struct sf_pdma_chan *chan;
456 chan = &pdma->chans[i];
458 chan->regs.ctrl =
460 chan->regs.xfer_type =
462 chan->regs.xfer_size =
464 chan->regs.dst_addr =
466 chan->regs.src_addr =
468 chan->regs.act_type =
470 chan->regs.residue =
472 chan->regs.cur_dst_addr =
474 chan->regs.cur_src_addr =
477 chan->pdma = pdma;
478 chan->pm_state = RUNNING;
479 chan->slave_id = i;
480 chan->xfer_err = false;
481 spin_lock_init(&chan->lock);
483 chan->vchan.desc_free = sf_pdma_free_desc;
484 vchan_init(&chan->vchan, &pdma->dma_dev);
486 writel(PDMA_CLEAR_CTRL, chan->regs.ctrl);
488 tasklet_setup(&chan->done_tasklet, sf_pdma_donebh_tasklet);
489 tasklet_setup(&chan->err_tasklet, sf_pdma_errbh_tasklet);
580 list_del(&ch->vchan.chan.device_node);