/kernel/linux/linux-5.10/net/sched/ |
H A D | sch_generic.c | 72 const struct netdev_queue *txq = q->dev_queue; in __skb_dequeue_bad_txq() 225 const struct netdev_queue *txq = q->dev_queue; in dequeue_skb() 580 .dev_queue = &noop_netdev_queue, 845 struct Qdisc *qdisc_alloc(struct netdev_queue *dev_queue, in qdisc_alloc() argument 854 if (!dev_queue) { in qdisc_alloc() 860 dev = dev_queue->dev; in qdisc_alloc() 861 sch = kzalloc_node(size, GFP_KERNEL, netdev_queue_numa_node_read(dev_queue)); in qdisc_alloc() 900 sch->dev_queue = dev_queue; in qdisc_alloc() 912 struct Qdisc *qdisc_create_dflt(struct netdev_queue *dev_queue, in qdisc_create_dflt() argument 1039 dev_graft_qdisc(struct netdev_queue *dev_queue, struct Qdisc *qdisc) dev_graft_qdisc() argument 1060 shutdown_scheduler_queue(struct net_device *dev, struct netdev_queue *dev_queue, void *_qdisc_default) shutdown_scheduler_queue() argument 1075 attach_one_default_qdisc(struct net_device *dev, struct netdev_queue *dev_queue, void *_unused) attach_one_default_qdisc() argument 1137 transition_one_qdisc(struct net_device *dev, struct netdev_queue *dev_queue, void *_need_watchdog) transition_one_qdisc() argument 1190 dev_deactivate_queue(struct net_device *dev, struct netdev_queue *dev_queue, void *_qdisc_default) dev_deactivate_queue() argument 1204 dev_reset_queue(struct net_device *dev, struct netdev_queue *dev_queue, void *_unused) dev_reset_queue() argument 1235 struct netdev_queue *dev_queue; some_qdisc_is_busy() local 1314 qdisc_change_tx_queue_len(struct net_device *dev, struct netdev_queue *dev_queue) qdisc_change_tx_queue_len() argument 1356 dev_init_scheduler_queue(struct net_device *dev, struct netdev_queue *dev_queue, void *_qdisc) dev_init_scheduler_queue() argument [all...] |
H A D | sch_mq.c | 72 struct netdev_queue *dev_queue; in mq_init() local 89 dev_queue = netdev_get_tx_queue(dev, ntx); in mq_init() 90 qdisc = qdisc_create_dflt(dev_queue, get_default_qdisc_ops(dev, ntx), in mq_init() 115 old = dev_graft_qdisc(qdisc->dev_queue, qdisc); in mq_attach() 216 struct netdev_queue *dev_queue = mq_queue_get(sch, cl); in mq_graft() local 223 *old = dev_graft_qdisc(dev_queue, new); in mq_graft() 241 struct netdev_queue *dev_queue = mq_queue_get(sch, cl); in mq_leaf() local 243 return dev_queue->qdisc_sleeping; in mq_leaf() 258 struct netdev_queue *dev_queue = mq_queue_get(sch, cl); in mq_dump_class() local 262 tcm->tcm_info = dev_queue in mq_dump_class() 269 struct netdev_queue *dev_queue = mq_queue_get(sch, cl); mq_dump_class_stats() local [all...] |
H A D | sch_mqprio.c | 229 struct netdev_queue *dev_queue; in mqprio_init() local 269 dev_queue = netdev_get_tx_queue(dev, i); in mqprio_init() 270 qdisc = qdisc_create_dflt(dev_queue, in mqprio_init() 341 old = dev_graft_qdisc(qdisc->dev_queue, qdisc); in mqprio_attach() 388 struct netdev_queue *dev_queue = mqprio_queue_get(sch, cl); in mqprio_graft() local 390 if (!dev_queue) in mqprio_graft() 396 *old = dev_graft_qdisc(dev_queue, new); in mqprio_graft() 525 struct netdev_queue *dev_queue = mqprio_queue_get(sch, cl); in mqprio_leaf() local 527 if (!dev_queue) in mqprio_leaf() 530 return dev_queue in mqprio_leaf() 556 struct netdev_queue *dev_queue = mqprio_queue_get(sch, cl); mqprio_dump_class() local 628 struct netdev_queue *dev_queue = mqprio_queue_get(sch, cl); global() variable [all...] |
H A D | sch_api.c | 1047 struct netdev_queue *dev_queue; in qdisc_graft() local 1070 dev_queue = netdev_get_tx_queue(dev, i); in qdisc_graft() 1071 old = dev_graft_qdisc(dev_queue, new); in qdisc_graft() 1078 dev_queue = dev_ingress_queue(dev); in qdisc_graft() 1079 old = dev_graft_qdisc(dev_queue, new); in qdisc_graft() 1171 struct netdev_queue *dev_queue, in qdisc_create() 1217 sch = qdisc_alloc(dev_queue, ops, extack); in qdisc_create() 1707 struct netdev_queue *dev_queue; in tc_modify_qdisc() local 1710 dev_queue = p->ops->cl_ops->select_queue(p, tcm); in tc_modify_qdisc() 1712 dev_queue in tc_modify_qdisc() 1170 qdisc_create(struct net_device *dev, struct netdev_queue *dev_queue, struct Qdisc *p, u32 parent, u32 handle, struct nlattr **tca, int *errp, struct netlink_ext_ack *extack) qdisc_create() argument 1813 struct netdev_queue *dev_queue; tc_dump_qdisc() local 2257 struct netdev_queue *dev_queue; tc_dump_tclass() local [all...] |
H A D | sch_plug.c | 187 netif_schedule_queue(sch->dev_queue); in plug_change() 195 netif_schedule_queue(sch->dev_queue); in plug_change()
|
H A D | sch_taprio.c | 1712 struct netdev_queue *dev_queue; in taprio_init() local 1715 dev_queue = netdev_get_tx_queue(dev, i); in taprio_init() 1716 qdisc = qdisc_create_dflt(dev_queue, in taprio_init() 1751 struct netdev_queue *dev_queue = taprio_queue_get(sch, cl); in taprio_graft() local 1753 if (!dev_queue) in taprio_graft() 1909 struct netdev_queue *dev_queue = taprio_queue_get(sch, cl); in taprio_leaf() local 1911 if (!dev_queue) in taprio_leaf() 1914 return dev_queue->qdisc_sleeping; in taprio_leaf() 1929 struct netdev_queue *dev_queue = taprio_queue_get(sch, cl); in taprio_dump_class() local 1933 tcm->tcm_info = dev_queue in taprio_dump_class() 1943 struct netdev_queue *dev_queue = taprio_queue_get(sch, cl); global() variable [all...] |
H A D | sch_cbs.c | 411 q->qdisc = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in cbs_init() 422 q->queue = sch->dev_queue - netdev_get_tx_queue(dev, 0); in cbs_init() 497 new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in cbs_graft()
|
H A D | sch_dsmark.c | 74 new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in dsmark_graft() 390 p->q = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, sch->handle, in dsmark_init()
|
H A D | sch_prio.c | 198 queues[i] = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in prio_tune() 294 new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in prio_graft()
|
H A D | sch_drr.c | 111 cl->qdisc = qdisc_create_dflt(sch->dev_queue, in drr_change_class() 211 new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in drr_graft_class()
|
/kernel/linux/linux-6.6/net/sched/ |
H A D | sch_generic.c | 74 const struct netdev_queue *txq = q->dev_queue; in __skb_dequeue_bad_txq() 231 const struct netdev_queue *txq = q->dev_queue; in dequeue_skb() 660 .dev_queue = &noop_netdev_queue, 922 struct Qdisc *qdisc_alloc(struct netdev_queue *dev_queue, in qdisc_alloc() argument 931 if (!dev_queue) { in qdisc_alloc() 937 dev = dev_queue->dev; in qdisc_alloc() 938 sch = kzalloc_node(size, GFP_KERNEL, netdev_queue_numa_node_read(dev_queue)); in qdisc_alloc() 973 sch->dev_queue = dev_queue; in qdisc_alloc() 984 struct Qdisc *qdisc_create_dflt(struct netdev_queue *dev_queue, in qdisc_create_dflt() argument 1111 dev_graft_qdisc(struct netdev_queue *dev_queue, struct Qdisc *qdisc) dev_graft_qdisc() argument 1132 shutdown_scheduler_queue(struct net_device *dev, struct netdev_queue *dev_queue, void *_qdisc_default) shutdown_scheduler_queue() argument 1147 attach_one_default_qdisc(struct net_device *dev, struct netdev_queue *dev_queue, void *_unused) attach_one_default_qdisc() argument 1209 transition_one_qdisc(struct net_device *dev, struct netdev_queue *dev_queue, void *_need_watchdog) transition_one_qdisc() argument 1262 dev_deactivate_queue(struct net_device *dev, struct netdev_queue *dev_queue, void *_qdisc_default) dev_deactivate_queue() argument 1276 dev_reset_queue(struct net_device *dev, struct netdev_queue *dev_queue, void *_unused) dev_reset_queue() argument 1308 struct netdev_queue *dev_queue; some_qdisc_is_busy() local 1387 qdisc_change_tx_queue_len(struct net_device *dev, struct netdev_queue *dev_queue) qdisc_change_tx_queue_len() argument 1453 dev_init_scheduler_queue(struct net_device *dev, struct netdev_queue *dev_queue, void *_qdisc) dev_init_scheduler_queue() argument [all...] |
H A D | sch_mq.c | 72 struct netdev_queue *dev_queue; in mq_init() local 89 dev_queue = netdev_get_tx_queue(dev, ntx); in mq_init() 90 qdisc = qdisc_create_dflt(dev_queue, get_default_qdisc_ops(dev, ntx), in mq_init() 115 old = dev_graft_qdisc(qdisc->dev_queue, qdisc); in mq_attach() 178 struct netdev_queue *dev_queue = mq_queue_get(sch, cl); in mq_graft() local 185 *old = dev_graft_qdisc(dev_queue, new); in mq_graft() 203 struct netdev_queue *dev_queue = mq_queue_get(sch, cl); in mq_leaf() local 205 return rtnl_dereference(dev_queue->qdisc_sleeping); in mq_leaf() 220 struct netdev_queue *dev_queue = mq_queue_get(sch, cl); in mq_dump_class() local 224 tcm->tcm_info = rtnl_dereference(dev_queue in mq_dump_class() 231 struct netdev_queue *dev_queue = mq_queue_get(sch, cl); mq_dump_class_stats() local [all...] |
H A D | sch_htb.c | 1126 struct netdev_queue *dev_queue = netdev_get_tx_queue(dev, ntx); in htb_init() local 1129 qdisc = qdisc_create_dflt(dev_queue, &pfifo_qdisc_ops, in htb_init() 1169 old = dev_graft_qdisc(qdisc->dev_queue, qdisc); in htb_attach_offload() 1174 struct netdev_queue *dev_queue = netdev_get_tx_queue(dev, ntx); in htb_attach_offload() local 1175 struct Qdisc *old = dev_graft_qdisc(dev_queue, NULL); in htb_attach_offload() 1191 struct netdev_queue *dev_queue = netdev_get_tx_queue(dev, ntx); in htb_attach_software() local 1192 struct Qdisc *old = dev_graft_qdisc(dev_queue, sch); in htb_attach_software() 1377 return sch->dev_queue; in htb_select_queue() 1390 htb_graft_helper(struct netdev_queue *dev_queue, struct Qdisc *new_q) in htb_graft_helper() argument 1392 struct net_device *dev = dev_queue in htb_graft_helper() 1452 struct netdev_queue *dev_queue = sch->dev_queue; htb_graft() local 1534 htb_parent_to_leaf_offload(struct Qdisc *sch, struct netdev_queue *dev_queue, struct Qdisc *new_q) htb_parent_to_leaf_offload() argument 1552 struct netdev_queue *dev_queue; htb_destroy_class_offload() local 1728 struct netdev_queue *dev_queue = sch->dev_queue; htb_delete() local 1780 struct netdev_queue *dev_queue; htb_change_class() local [all...] |
H A D | sch_mqprio.c | 352 struct netdev_queue *dev_queue; in mqprio_init() local 399 dev_queue = netdev_get_tx_queue(dev, i); in mqprio_init() 400 qdisc = qdisc_create_dflt(dev_queue, in mqprio_init() 444 old = dev_graft_qdisc(qdisc->dev_queue, qdisc); in mqprio_attach() 469 struct netdev_queue *dev_queue = mqprio_queue_get(sch, cl); in mqprio_graft() local 471 if (!dev_queue) in mqprio_graft() 477 *old = dev_graft_qdisc(dev_queue, new); in mqprio_graft() 616 struct netdev_queue *dev_queue = mqprio_queue_get(sch, cl); in mqprio_leaf() local 618 if (!dev_queue) in mqprio_leaf() 621 return rtnl_dereference(dev_queue in mqprio_leaf() 647 struct netdev_queue *dev_queue = mqprio_queue_get(sch, cl); mqprio_dump_class() local 708 struct netdev_queue *dev_queue = mqprio_queue_get(sch, cl); global() variable [all...] |
H A D | sch_api.c | 1082 struct netdev_queue *dev_queue; in qdisc_graft() local 1089 dev_queue = dev_ingress_queue(dev); in qdisc_graft() 1090 if (!dev_queue) { in qdisc_graft() 1095 q = rtnl_dereference(dev_queue->qdisc_sleeping); in qdisc_graft() 1117 dev_queue = netdev_get_tx_queue(dev, i); in qdisc_graft() 1118 old = dev_graft_qdisc(dev_queue, new); in qdisc_graft() 1125 old = dev_graft_qdisc(dev_queue, NULL); in qdisc_graft() 1134 dev_graft_qdisc(dev_queue, new); in qdisc_graft() 1224 struct netdev_queue *dev_queue, in qdisc_create() 1270 sch = qdisc_alloc(dev_queue, op in qdisc_create() 1223 qdisc_create(struct net_device *dev, struct netdev_queue *dev_queue, u32 parent, u32 handle, struct nlattr **tca, int *errp, struct netlink_ext_ack *extack) qdisc_create() argument 1738 struct netdev_queue *dev_queue; tc_modify_qdisc() local 1844 struct netdev_queue *dev_queue; tc_dump_qdisc() local 2284 struct netdev_queue *dev_queue; tc_dump_tclass() local [all...] |
H A D | sch_plug.c | 184 netif_schedule_queue(sch->dev_queue); in plug_change() 192 netif_schedule_queue(sch->dev_queue); in plug_change()
|
H A D | sch_cbs.c | 412 q->qdisc = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in cbs_init() 423 q->queue = sch->dev_queue - netdev_get_tx_queue(dev, 0); in cbs_init() 498 new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in cbs_graft()
|
/kernel/linux/linux-6.6/include/linux/ |
H A D | netdevice.h | 3314 static __always_inline void netif_tx_start_queue(struct netdev_queue *dev_queue) in netif_tx_start_queue() argument 3316 clear_bit(__QUEUE_STATE_DRV_XOFF, &dev_queue->state); in netif_tx_start_queue() 3340 void netif_tx_wake_queue(struct netdev_queue *dev_queue); 3364 static __always_inline void netif_tx_stop_queue(struct netdev_queue *dev_queue) in netif_tx_stop_queue() argument 3367 set_bit(__QUEUE_STATE_DRV_XOFF, &dev_queue->state); in netif_tx_stop_queue() 3384 static inline bool netif_tx_queue_stopped(const struct netdev_queue *dev_queue) in netif_tx_queue_stopped() argument 3386 return test_bit(__QUEUE_STATE_DRV_XOFF, &dev_queue->state); in netif_tx_queue_stopped() 3400 static inline bool netif_xmit_stopped(const struct netdev_queue *dev_queue) in netif_xmit_stopped() argument 3402 return dev_queue->state & QUEUE_STATE_ANY_XOFF; in netif_xmit_stopped() 3406 netif_xmit_frozen_or_stopped(const struct netdev_queue *dev_queue) in netif_xmit_frozen_or_stopped() argument 3412 netif_xmit_frozen_or_drv_stopped(const struct netdev_queue *dev_queue) netif_xmit_frozen_or_drv_stopped() argument 3427 netdev_queue_set_dql_min_limit(struct netdev_queue *dev_queue, unsigned int min_limit) netdev_queue_set_dql_min_limit() argument 3442 netdev_txq_bql_enqueue_prefetchw(struct netdev_queue *dev_queue) netdev_txq_bql_enqueue_prefetchw() argument 3456 netdev_txq_bql_complete_prefetchw(struct netdev_queue *dev_queue) netdev_txq_bql_complete_prefetchw() argument 3473 netdev_tx_sent_queue(struct netdev_queue *dev_queue, unsigned int bytes) netdev_tx_sent_queue() argument 3503 __netdev_tx_sent_queue(struct netdev_queue *dev_queue, unsigned int bytes, bool xmit_more) __netdev_tx_sent_queue() argument 3549 netdev_tx_completed_queue(struct netdev_queue *dev_queue, unsigned int pkts, unsigned int bytes) netdev_tx_completed_queue() argument 3604 netdev_reset_queue(struct net_device *dev_queue) netdev_reset_queue() argument [all...] |
/kernel/linux/linux-6.6/include/net/ |
H A D | netdev_queues.h | 89 netdev_txq_completed_mb(struct netdev_queue *dev_queue, in netdev_txq_completed_mb() argument 93 netdev_tx_completed_queue(dev_queue, pkts, bytes); in netdev_txq_completed_mb()
|
H A D | sch_generic.h | 83 * q->dev_queue : It can test 104 struct netdev_queue *dev_queue; member 547 struct Qdisc *q = rcu_dereference_rtnl(qdisc->dev_queue->qdisc); in qdisc_root() 554 return rcu_dereference_bh(qdisc->dev_queue->qdisc); in qdisc_root_bh() 559 return rcu_dereference_rtnl(qdisc->dev_queue->qdisc_sleeping); in qdisc_root_sleeping() 572 return qdisc->dev_queue->dev; in qdisc_dev() 689 struct Qdisc *dev_graft_qdisc(struct netdev_queue *dev_queue, 723 struct Qdisc *qdisc_alloc(struct netdev_queue *dev_queue, 727 struct Qdisc *qdisc_create_dflt(struct netdev_queue *dev_queue,
|
/kernel/linux/linux-5.10/include/linux/ |
H A D | netdevice.h | 3324 static __always_inline void netif_tx_start_queue(struct netdev_queue *dev_queue) in netif_tx_start_queue() argument 3326 clear_bit(__QUEUE_STATE_DRV_XOFF, &dev_queue->state); in netif_tx_start_queue() 3350 void netif_tx_wake_queue(struct netdev_queue *dev_queue); 3374 static __always_inline void netif_tx_stop_queue(struct netdev_queue *dev_queue) in netif_tx_stop_queue() argument 3376 set_bit(__QUEUE_STATE_DRV_XOFF, &dev_queue->state); in netif_tx_stop_queue() 3393 static inline bool netif_tx_queue_stopped(const struct netdev_queue *dev_queue) in netif_tx_queue_stopped() argument 3395 return test_bit(__QUEUE_STATE_DRV_XOFF, &dev_queue->state); in netif_tx_queue_stopped() 3409 static inline bool netif_xmit_stopped(const struct netdev_queue *dev_queue) in netif_xmit_stopped() argument 3411 return dev_queue->state & QUEUE_STATE_ANY_XOFF; in netif_xmit_stopped() 3415 netif_xmit_frozen_or_stopped(const struct netdev_queue *dev_queue) in netif_xmit_frozen_or_stopped() argument 3421 netif_xmit_frozen_or_drv_stopped(const struct netdev_queue *dev_queue) netif_xmit_frozen_or_drv_stopped() argument 3433 netdev_txq_bql_enqueue_prefetchw(struct netdev_queue *dev_queue) netdev_txq_bql_enqueue_prefetchw() argument 3447 netdev_txq_bql_complete_prefetchw(struct netdev_queue *dev_queue) netdev_txq_bql_complete_prefetchw() argument 3454 netdev_tx_sent_queue(struct netdev_queue *dev_queue, unsigned int bytes) netdev_tx_sent_queue() argument 3484 __netdev_tx_sent_queue(struct netdev_queue *dev_queue, unsigned int bytes, bool xmit_more) __netdev_tx_sent_queue() argument 3520 netdev_tx_completed_queue(struct netdev_queue *dev_queue, unsigned int pkts, unsigned int bytes) netdev_tx_completed_queue() argument 3575 netdev_reset_queue(struct net_device *dev_queue) netdev_reset_queue() argument [all...] |
/kernel/linux/linux-5.10/include/net/ |
H A D | sch_generic.h | 69 * q->dev_queue : It can test 90 struct netdev_queue *dev_queue; member 544 struct Qdisc *q = rcu_dereference_rtnl(qdisc->dev_queue->qdisc); in qdisc_root() 551 return rcu_dereference_bh(qdisc->dev_queue->qdisc); in qdisc_root_bh() 556 return qdisc->dev_queue->qdisc_sleeping; in qdisc_root_sleeping() 596 return qdisc->dev_queue->dev; in qdisc_dev() 681 struct Qdisc *dev_graft_qdisc(struct netdev_queue *dev_queue, 711 struct Qdisc *qdisc_alloc(struct netdev_queue *dev_queue, 715 struct Qdisc *qdisc_create_dflt(struct netdev_queue *dev_queue,
|
H A D | pkt_sched.h | 144 return dev_net(q->dev_queue->dev); in qdisc_net()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/hisilicon/hns/ |
H A D | hns_enet.c | 312 struct netdev_queue *dev_queue; in hns_nic_net_xmit_hw() local 363 dev_queue = netdev_get_tx_queue(ndev, skb->queue_mapping); in hns_nic_net_xmit_hw() 364 netdev_tx_sent_queue(dev_queue, skb->len); in hns_nic_net_xmit_hw() 953 struct netdev_queue *dev_queue; in hns_nic_tx_poll_one() local 982 dev_queue = netdev_get_tx_queue(ndev, ring_data->queue_index); in hns_nic_tx_poll_one() 983 netdev_tx_completed_queue(dev_queue, pkts, bytes); in hns_nic_tx_poll_one() 994 if (netif_tx_queue_stopped(dev_queue) && in hns_nic_tx_poll_one() 996 netif_tx_wake_queue(dev_queue); in hns_nic_tx_poll_one() 1037 struct netdev_queue *dev_queue; in hns_nic_tx_clr_all_bufs() local 1047 dev_queue in hns_nic_tx_clr_all_bufs() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/hisilicon/hns/ |
H A D | hns_enet.c | 333 struct netdev_queue *dev_queue; in hns_nic_net_xmit_hw() local 386 dev_queue = netdev_get_tx_queue(ndev, skb->queue_mapping); in hns_nic_net_xmit_hw() 387 netdev_tx_sent_queue(dev_queue, skb->len); in hns_nic_net_xmit_hw() 976 struct netdev_queue *dev_queue; in hns_nic_tx_poll_one() local 1005 dev_queue = netdev_get_tx_queue(ndev, ring_data->queue_index); in hns_nic_tx_poll_one() 1006 netdev_tx_completed_queue(dev_queue, pkts, bytes); in hns_nic_tx_poll_one() 1017 if (netif_tx_queue_stopped(dev_queue) && in hns_nic_tx_poll_one() 1019 netif_tx_wake_queue(dev_queue); in hns_nic_tx_poll_one() 1060 struct netdev_queue *dev_queue; in hns_nic_tx_clr_all_bufs() local 1070 dev_queue in hns_nic_tx_clr_all_bufs() [all...] |