Lines Matching refs:txreq
27 * @txreq: sdma transmit request
35 struct sdma_txreq txreq;
112 * The size of the txreq ring is fixed at initialization.
142 sdma_txclean(priv->dd, &tx->txreq);
210 /* Finish storing txreq before incrementing head. */
226 static void hfi1_ipoib_sdma_complete(struct sdma_txreq *txreq, int status)
228 struct ipoib_txreq *tx = container_of(txreq, struct ipoib_txreq, txreq);
239 struct sdma_txreq *txreq = &tx->txreq;
245 ret = sdma_txadd_kvaddr(dd, txreq, skb->data, skb_headlen(skb));
254 txreq,
270 struct sdma_txreq *txreq = &tx->txreq;
276 ret = sdma_txinit(txreq, 0, pkt_bytes, hfi1_ipoib_sdma_complete);
282 txreq,
399 tx->txreq.num_desc = 0;
403 INIT_LIST_HEAD(&tx->txreq.list);
422 sdma_txclean(priv->dd, &tx->txreq);
470 &tx->txreq,
518 sdma_txclean(priv->dd, &tx->txreq);
560 list_add_tail(&tx->txreq.list, &txq->tx_list);
629 struct sdma_txreq *txreq,
639 if (sdma_progress(sde, seq, txreq)) {
644 if (list_empty(&txreq->list))
646 list_add_tail(&txreq->list, &txq->tx_list);
798 struct sdma_txreq *txreq;
802 list_for_each_entry_safe(txreq, txreq_tmp, &txq->tx_list, list) {
804 container_of(txreq, struct ipoib_txreq, txreq);
806 list_del(&txreq->list);
807 sdma_txclean(txq->priv->dd, &tx->txreq);