Lines Matching refs:buffer

26 					 struct ef4_tx_buffer *buffer)
38 buffer->dma_addr = page_buf->dma_addr + offset;
39 buffer->unmap_len = 0;
44 struct ef4_tx_buffer *buffer, size_t len)
48 return ef4_tx_get_copy_buffer(tx_queue, buffer);
52 struct ef4_tx_buffer *buffer,
56 if (buffer->unmap_len) {
58 dma_addr_t unmap_addr = buffer->dma_addr - buffer->dma_offset;
59 if (buffer->flags & EF4_TX_BUF_MAP_SINGLE)
60 dma_unmap_single(dma_dev, unmap_addr, buffer->unmap_len,
63 dma_unmap_page(dma_dev, unmap_addr, buffer->unmap_len,
65 buffer->unmap_len = 0;
68 if (buffer->flags & EF4_TX_BUF_SKB) {
70 (*bytes_compl) += buffer->skb->len;
71 dev_consume_skb_any((struct sk_buff *)buffer->skb);
77 buffer->len = 0;
78 buffer->flags = 0;
152 struct ef4_tx_buffer *buffer;
158 buffer = ef4_tx_queue_get_insert_buffer(tx_queue);
160 copy_buffer = ef4_tx_get_copy_buffer(tx_queue, buffer);
168 buffer->len = min_len;
170 buffer->len = copy_len;
173 buffer->skb = skb;
174 buffer->flags = EF4_TX_BUF_SKB;
185 struct ef4_tx_buffer *buffer;
190 buffer = ef4_tx_queue_get_insert_buffer(tx_queue);
193 buffer->len = dma_len;
194 buffer->dma_addr = dma_addr;
195 buffer->flags = EF4_TX_BUF_CONT;
201 return buffer;
230 struct ef4_tx_buffer *buffer;
233 buffer = ef4_tx_map_chunk(tx_queue, dma_addr, len);
238 buffer->flags = EF4_TX_BUF_CONT | dma_flags;
239 buffer->unmap_len = unmap_len;
240 buffer->dma_offset = buffer->dma_addr - unmap_addr;
243 /* Store SKB details with the final buffer for
246 buffer->skb = skb;
247 buffer->flags = EF4_TX_BUF_SKB | dma_flags;
270 struct ef4_tx_buffer *buffer;
275 buffer = __ef4_tx_queue_get_insert_buffer(tx_queue);
276 ef4_dequeue_buffer(tx_queue, buffer, NULL, NULL);
281 * Add a socket buffer to a TX queue
283 * This maps all fragments of a socket buffer for DMA and adds them to
285 * the number of fragments in the socket buffer.
366 struct ef4_tx_buffer *buffer = &tx_queue->buffer[read_ptr];
368 if (!(buffer->flags & EF4_TX_BUF_OPTION) &&
369 unlikely(buffer->len == 0)) {
377 ef4_dequeue_buffer(tx_queue, buffer, pkts_compl, bytes_compl);
460 if (!tx_queue->buffer) {
554 tx_queue->buffer = kcalloc(entries, sizeof(*tx_queue->buffer),
556 if (!tx_queue->buffer)
577 kfree(tx_queue->buffer);
578 tx_queue->buffer = NULL;
608 struct ef4_tx_buffer *buffer;
613 if (!tx_queue->buffer)
619 buffer = &tx_queue->buffer[tx_queue->read_count & tx_queue->ptr_mask];
620 ef4_dequeue_buffer(tx_queue, buffer, &pkts_compl, &bytes_compl);
632 if (!tx_queue->buffer)
647 kfree(tx_queue->buffer);
648 tx_queue->buffer = NULL;