Lines Matching defs:queue

19  *    Fixes in packet dropping, queue length setting and queue wakeup.
20 * Increased default tx queue length.
131 * to serve as one transmit queue for tuntap device. The sock_fprog and
133 * netdevice not for a specific queue (at least I didn't see the requirement for
239 struct sk_buff_head *queue = &tfile->sk.sk_write_queue;
246 spin_lock(&queue->lock);
247 skb_queue_splice_tail_init(queue, &process_queue);
248 spin_unlock(&queue->lock);
256 spin_lock(&queue->lock);
257 skb_queue_splice(&process_queue, queue);
258 spin_unlock(&queue->lock);
499 /* TODO: keep queueing to old queue until it's empty? */
682 /* Drop read queue */
747 /* Drop read queue */
818 /* Setup XDP RX-queue info, for new tfile getting attached */
1055 /* Select queue was not called for the skbuff, so we extract the
1086 struct netdev_queue *queue;
1132 queue = netdev_get_tx_queue(dev, txq);
1133 queue->trans_start = jiffies;
1555 struct sk_buff_head *queue = &tfile->sk.sk_write_queue;
1560 if (!rx_batched || (!more && skb_queue_empty(queue))) {
1568 spin_lock(&queue->lock);
1569 if (!more || skb_queue_len(queue) == rx_batched) {
1571 skb_queue_splice_tail_init(queue, &process_queue);
1574 __skb_queue_tail(queue, skb);
1576 spin_unlock(&queue->lock);
2012 struct sk_buff_head *queue = &tfile->sk.sk_write_queue;
2015 spin_lock_bh(&queue->lock);
2016 __skb_queue_tail(queue, skb);
2017 queue_len = skb_queue_len(queue);
2018 spin_unlock(&queue->lock);
2353 /* We prefer our own queue length */
2788 /* One or more queue has already been attached, no need