Lines Matching refs:entry

917 	/* Mark the last entry as wrapping the ring. */
950 int entry, free_count;
960 /* Calculate the next Tx descriptor entry. */
963 entry = ep->cur_tx % TX_RING_SIZE;
965 ep->tx_skbuff[entry] = skb;
966 ep->tx_ring[entry].bufaddr = dma_map_single(&ep->pci_dev->dev,
976 /* Leave room for an additional entry. */
980 ep->tx_ring[entry].buflength = ctrl_word | skb->len;
981 ep->tx_ring[entry].txstatus =
995 skb->len, entry, ctrl_word, er32(TxSTAT));
1033 int entry = dirty_tx % TX_RING_SIZE;
1034 int txstatus = ep->tx_ring[entry].txstatus;
1042 dev->stats.tx_bytes += ep->tx_skbuff[entry]->len;
1047 skb = ep->tx_skbuff[entry];
1049 ep->tx_ring[entry].bufaddr, skb->len,
1052 ep->tx_skbuff[entry] = NULL;
1144 int entry = ep->cur_rx % RX_RING_SIZE;
1149 netdev_dbg(dev, " In epic_rx(), entry %d %8.8x.\n", entry,
1150 ep->rx_ring[entry].rxstatus);
1155 /* If we own the next entry, it's a new packet. Send it up. */
1156 while ((ep->rx_ring[entry].rxstatus & DescOwn) == 0) {
1157 int status = ep->rx_ring[entry].rxstatus;
1192 ep->rx_ring[entry].bufaddr,
1195 skb_copy_to_linear_data(skb, ep->rx_skbuff[entry]->data, pkt_len);
1198 ep->rx_ring[entry].bufaddr,
1203 ep->rx_ring[entry].bufaddr,
1206 skb_put(skb = ep->rx_skbuff[entry], pkt_len);
1207 ep->rx_skbuff[entry] = NULL;
1215 entry = (++ep->cur_rx) % RX_RING_SIZE;
1220 entry = ep->dirty_rx % RX_RING_SIZE;
1221 if (ep->rx_skbuff[entry] == NULL) {
1223 skb = ep->rx_skbuff[entry] = netdev_alloc_skb(dev, ep->rx_buf_sz + 2);
1227 ep->rx_ring[entry].bufaddr = dma_map_single(&ep->pci_dev->dev,
1234 ep->rx_ring[entry].rxstatus = DescOwn;