Lines Matching refs:vchan

44 	return container_of(dchan, struct sf_pdma_chan, vchan.chan);
100 desc->async_tx = vchan_tx_prep(&chan->vchan, &desc->vdesc, flags);
102 spin_lock_irqsave(&chan->vchan.lock, iflags);
104 spin_unlock_irqrestore(&chan->vchan.lock, iflags);
143 spin_lock_irqsave(&chan->vchan.lock, flags);
147 vchan_get_all_descriptors(&chan->vchan, &head);
149 spin_unlock_irqrestore(&chan->vchan.lock, flags);
150 vchan_dma_desc_free_list(&chan->vchan, &head);
163 spin_lock_irqsave(&chan->vchan.lock, flags);
165 list_for_each_entry(vd, &chan->vchan.desc_submitted, node)
178 vd = vchan_find_desc(&chan->vchan, cookie);
187 spin_unlock_irqrestore(&chan->vchan.lock, flags);
213 spin_lock_irqsave(&chan->vchan.lock, flags);
218 vchan_get_all_descriptors(&chan->vchan, &head);
219 spin_unlock_irqrestore(&chan->vchan.lock, flags);
220 vchan_dma_desc_free_list(&chan->vchan, &head);
240 struct virt_dma_chan *vchan = &chan->vchan;
243 if (list_empty(&vchan->desc_issued))
246 vdesc = list_first_entry(&vchan->desc_issued, struct virt_dma_desc, node);
276 spin_lock_irqsave(&chan->vchan.lock, flags);
278 if (!chan->desc && vchan_issue_pending(&chan->vchan)) {
284 spin_unlock_irqrestore(&chan->vchan.lock, flags);
308 spin_lock_irqsave(&chan->vchan.lock, flags);
316 spin_unlock_irqrestore(&chan->vchan.lock, flags);
347 spin_lock(&chan->vchan.lock);
364 spin_unlock(&chan->vchan.lock);
483 chan->vchan.desc_free = sf_pdma_free_desc;
484 vchan_init(&chan->vchan, &pdma->dma_dev);
580 list_del(&ch->vchan.chan.device_node);
581 tasklet_kill(&ch->vchan.task);