/kernel/linux/linux-5.10/drivers/net/wireless/broadcom/brcm80211/brcmfmac/ |
H A D | flowring.c | 142 if (flow->rings[i] == NULL) in brcmf_flowring_create() 160 flow->rings[i] = ring; in brcmf_flowring_create() 172 ring = flow->rings[flowid]; in brcmf_flowring_tid() 192 ring = flow->rings[flowid]; in brcmf_flowring_block() 201 if ((flow->rings[i]) && (i != flowid)) { in brcmf_flowring_block() 202 ring = flow->rings[i]; in brcmf_flowring_block() 212 flow->rings[flowid]->blocked = blocked; in brcmf_flowring_block() 236 ring = flow->rings[flowid]; in brcmf_flowring_delete() 247 flow->rings[flowid] = NULL; in brcmf_flowring_delete() 264 ring = flow->rings[flowi in brcmf_flowring_enqueue() [all...] |
/kernel/linux/linux-6.6/drivers/net/wireless/broadcom/brcm80211/brcmfmac/ |
H A D | flowring.c | 142 if (flow->rings[i] == NULL) in brcmf_flowring_create() 160 flow->rings[i] = ring; in brcmf_flowring_create() 172 ring = flow->rings[flowid]; in brcmf_flowring_tid() 192 ring = flow->rings[flowid]; in brcmf_flowring_block() 201 if ((flow->rings[i]) && (i != flowid)) { in brcmf_flowring_block() 202 ring = flow->rings[i]; in brcmf_flowring_block() 212 flow->rings[flowid]->blocked = blocked; in brcmf_flowring_block() 236 ring = flow->rings[flowid]; in brcmf_flowring_delete() 247 flow->rings[flowid] = NULL; in brcmf_flowring_delete() 264 ring = flow->rings[flowi in brcmf_flowring_enqueue() [all...] |
/kernel/linux/linux-5.10/net/9p/ |
H A D | trans_xen.c | 81 struct xen_9pfs_dataring *rings; member 158 ring = &priv->rings[num]; in p9_xen_request() 303 struct xen_9pfs_dataring *ring = &priv->rings[i]; in xen_9pfs_front_free() 307 if (!priv->rings[i].intf) in xen_9pfs_front_free() 309 if (priv->rings[i].irq > 0) in xen_9pfs_front_free() 310 unbind_from_irqhandler(priv->rings[i].irq, priv->dev); in xen_9pfs_front_free() 311 if (priv->rings[i].data.in) { in xen_9pfs_front_free() 313 j < (1 << priv->rings[i].intf->ring_order); in xen_9pfs_front_free() 317 ref = priv->rings[i].intf->ref[j]; in xen_9pfs_front_free() 320 free_pages_exact(priv->rings[ in xen_9pfs_front_free() [all...] |
/kernel/linux/linux-6.6/tools/net/ynl/samples/ |
H A D | ethtool.c | 16 struct ethtool_rings_get_list *rings; in main() local 42 rings = ethtool_rings_get_dump(ys, &rreq); in main() 43 if (!rings) in main() 47 ynl_dump_foreach(rings, dev) { in main() 55 ethtool_rings_get_list_free(rings); in main()
|
/kernel/linux/linux-6.6/net/9p/ |
H A D | trans_xen.c | 58 struct xen_9pfs_dataring *rings; member 135 ring = &priv->rings[num]; in p9_xen_request() 283 struct xen_9pfs_dataring *ring = &priv->rings[i]; in xen_9pfs_front_free() 287 if (!priv->rings[i].intf) in xen_9pfs_front_free() 289 if (priv->rings[i].irq > 0) in xen_9pfs_front_free() 290 unbind_from_irqhandler(priv->rings[i].irq, priv->dev); in xen_9pfs_front_free() 291 if (priv->rings[i].data.in) { in xen_9pfs_front_free() 293 j < (1 << priv->rings[i].intf->ring_order); in xen_9pfs_front_free() 297 ref = priv->rings[i].intf->ref[j]; in xen_9pfs_front_free() 300 free_pages_exact(priv->rings[ in xen_9pfs_front_free() [all...] |
/kernel/linux/linux-6.6/drivers/i3c/master/mipi-i3c-hci/ |
H A D | dma.c | 167 struct hci_rings_data *rings = hci->io_data; in hci_dma_cleanup() local 171 if (!rings) in hci_dma_cleanup() 174 for (i = 0; i < rings->total; i++) { in hci_dma_cleanup() 175 rh = &rings->headers[i]; in hci_dma_cleanup() 204 kfree(rings); in hci_dma_cleanup() 210 struct hci_rings_data *rings; in hci_dma_init() local 219 dev_info(&hci->master.dev, "%d DMA rings available\n", nr_rings); in hci_dma_init() 221 dev_err(&hci->master.dev, "number of rings should be <= 8\n"); in hci_dma_init() 226 rings = kzalloc(struct_size(rings, header in hci_dma_init() 357 struct hci_rings_data *rings = hci->io_data; hci_dma_queue_xfer() local 442 struct hci_rings_data *rings = hci->io_data; hci_dma_dequeue_xfer() local 733 struct hci_rings_data *rings = hci->io_data; hci_dma_irq_handler() local [all...] |
/kernel/linux/linux-5.10/tools/lib/bpf/ |
H A D | ringbuf.c | 36 struct ring *rings; member 82 tmp = libbpf_reallocarray(rb->rings, rb->ring_cnt + 1, sizeof(*rb->rings)); in ring_buffer__add() 85 rb->rings = tmp; in ring_buffer__add() 92 r = &rb->rings[rb->ring_cnt]; in ring_buffer__add() 155 ringbuf_unmap_ring(rb, &rb->rings[i]); in ring_buffer__free() 160 free(rb->rings); in ring_buffer__free() 263 struct ring *ring = &rb->rings[i]; in ring_buffer__consume() 290 struct ring *ring = &rb->rings[ring_id]; in ring_buffer__poll()
|
/kernel/linux/linux-5.10/tools/testing/selftests/net/ |
H A D | psock_fanout.c | 221 static int sock_fanout_read(int fds[], char *rings[], const int expect[]) in sock_fanout_read() argument 225 ret[0] = sock_fanout_read_ring(fds[0], rings[0]); in sock_fanout_read() 226 ret[1] = sock_fanout_read_ring(fds[1], rings[1]); in sock_fanout_read() 348 char *rings[2]; in test_datapath() local 367 rings[0] = sock_fanout_open_ring(fds[0]); in test_datapath() 368 rings[1] = sock_fanout_open_ring(fds[1]); in test_datapath() 371 sock_fanout_read(fds, rings, expect0); in test_datapath() 376 ret = sock_fanout_read(fds, rings, expect1); in test_datapath() 381 ret |= sock_fanout_read(fds, rings, expect2); in test_datapath() 383 if (munmap(rings[ in test_datapath() [all...] |
/kernel/linux/linux-6.6/tools/testing/selftests/net/ |
H A D | psock_fanout.c | 235 static int sock_fanout_read(int fds[], char *rings[], const int expect[]) in sock_fanout_read() argument 239 ret[0] = sock_fanout_read_ring(fds[0], rings[0]); in sock_fanout_read() 240 ret[1] = sock_fanout_read_ring(fds[1], rings[1]); in sock_fanout_read() 412 char *rings[2]; in test_datapath() local 431 rings[0] = sock_fanout_open_ring(fds[0]); in test_datapath() 432 rings[1] = sock_fanout_open_ring(fds[1]); in test_datapath() 435 sock_fanout_read(fds, rings, expect0); in test_datapath() 440 ret = sock_fanout_read(fds, rings, expect1); in test_datapath() 445 ret |= sock_fanout_read(fds, rings, expect2); in test_datapath() 447 if (munmap(rings[ in test_datapath() [all...] |
H A D | toeplitz.c | 104 static struct ring_state rings[RSS_MAX_CPUS]; variable 242 /* simple test: sleep once unconditionally and then process all rings */ 250 do {} while (recv_block(&rings[i])); in process_rings() 355 /* block packets until all rings are added to the fanout group: in create_ring() 404 rings[i].cpu = i; in setup_rings() 405 rings[i].fd = create_ring(&rings[i].mmap); in setup_rings() 408 /* accept packets once all rings in the fanout group are up */ in setup_rings() 410 set_filter(rings[i].fd); in setup_rings() 418 if (munmap(rings[ in cleanup_rings() [all...] |
/kernel/linux/linux-6.6/tools/lib/bpf/ |
H A D | ringbuf.c | 37 struct ring *rings; member 100 tmp = libbpf_reallocarray(rb->rings, rb->ring_cnt + 1, sizeof(*rb->rings)); in ring_buffer__add() 103 rb->rings = tmp; in ring_buffer__add() 110 r = &rb->rings[rb->ring_cnt]; in ring_buffer__add() 173 ringbuf_unmap_ring(rb, &rb->rings[i]); in ring_buffer__free() 178 free(rb->rings); in ring_buffer__free() 281 struct ring *ring = &rb->rings[i]; in ring_buffer__consume() 308 struct ring *ring = &rb->rings[ring_id]; in ring_buffer__poll()
|
/kernel/linux/linux-6.6/drivers/crypto/intel/qat/qat_common/ |
H A D | adf_transport.c | 267 ring = &bank->rings[ring_num]; in adf_create_ring() 338 adf_handle_response(&bank->rings[i]); in adf_ring_response_handler() 404 /* Allocate the rings in the bank */ in adf_init_bank() 406 bank->rings = kzalloc_node(size, GFP_KERNEL, in adf_init_bank() 408 if (!bank->rings) in adf_init_bank() 425 ring = &bank->rings[i]; in adf_init_bank() 436 "Invalid tx rings mask config\n"); in adf_init_bank() 439 tx_ring = &bank->rings[i - hw_data->tx_rx_gap]; in adf_init_bank() 456 ring = &bank->rings[i]; in adf_init_bank() 460 kfree(bank->rings); in adf_init_bank() [all...] |
/kernel/linux/linux-6.6/io_uring/ |
H A D | io_uring.h | 234 smp_store_release(&ctx->rings->cq.tail, ctx->cached_cq_tail); in io_commit_cqring() 263 struct io_rings *r = ctx->rings; in io_sqring_full() 270 struct io_rings *rings = ctx->rings; in io_sqring_entries() local 274 entries = smp_load_acquire(&rings->sq.tail) - ctx->cached_sq_head; in io_sqring_entries()
|
H A D | io_uring.c | 186 return ctx->cached_cq_tail - READ_ONCE(ctx->rings->cq.head); in __io_cqring_events() 191 return READ_ONCE(ctx->rings->cq.tail) - READ_ONCE(ctx->rings->cq.head); in __io_cqring_events_user() 354 struct io_rings *r = ctx->rings; in io_account_cq_overflow() 573 if (READ_ONCE(ctx->rings->cq_flags) & IORING_CQ_EVENTFD_DISABLED) in io_eventfd_signal() 648 /* IOPOLL rings only need to wake up if it's also SQPOLL */ in __io_cq_unlock_post() 708 atomic_andnot(IORING_SQ_CQ_OVERFLOW, &ctx->rings->sq_flags); in __io_cqring_overflow_flush() 802 atomic_or(IORING_SQ_CQ_OVERFLOW, &ctx->rings->sq_flags); in io_cqring_event_overflow() 831 struct io_rings *rings = ctx->rings; in io_cqe_cache_refill() local 2334 struct io_rings *rings = ctx->rings; io_commit_sqring() local 2542 struct io_rings *rings = ctx->rings; io_cqring_wait() local 2777 struct io_rings *rings; rings_size() local 3741 struct io_rings *rings; io_allocate_scq_urings() local [all...] |
/kernel/linux/linux-6.6/drivers/soc/ti/ |
H A D | k3-ringacc.c | 153 * @use_count: Use count for shared rings 194 * @rm_gp_range: general purpose rings range from tisci 198 * @rings: array of rings descriptors (struct @k3_ring) 200 * @req_lock: protect rings allocation 202 * @tisci_ring_ops: ti-sci rings ops 211 u32 num_rings; /* number of rings in Ringacc module */ 219 struct k3_ring *rings; member 221 struct mutex req_lock; /* protect rings allocation */ 358 !(ringacc->rings[i in k3_ringacc_request_ring() [all...] |
/kernel/linux/linux-5.10/drivers/block/xen-blkback/ |
H A D | xenbus.c | 84 if (!blkif->rings || !blkif->rings[0].irq || !blkif->vbd.bdev) in xen_update_blkif_status() 110 ring = &blkif->rings[i]; in xen_update_blkif_status() 124 ring = &blkif->rings[i]; in xen_update_blkif_status() 134 blkif->rings = kcalloc(blkif->nr_rings, sizeof(struct xen_blkif_ring), in xen_blkif_alloc_rings() 136 if (!blkif->rings) in xen_blkif_alloc_rings() 140 struct xen_blkif_ring *ring = &blkif->rings[r]; in xen_blkif_alloc_rings() 274 struct xen_blkif_ring *ring = &blkif->rings[r]; in xen_blkif_disconnect() 335 * blkif->rings was allocated in connect_ring, so we should free it in in xen_blkif_disconnect() 338 kfree(blkif->rings); in xen_blkif_disconnect() [all...] |
/kernel/linux/linux-5.10/drivers/soc/ti/ |
H A D | k3-ringacc.c | 142 * @use_count: Use count for shared rings 177 * @rm_gp_range: general purpose rings range from tisci 181 * @rings: array of rings descriptors (struct @k3_ring) 183 * @req_lock: protect rings allocation 185 * @tisci_ring_ops: ti-sci rings ops 193 u32 num_rings; /* number of rings in Ringacc module */ 201 struct k3_ring *rings; member 203 struct mutex req_lock; /* protect rings allocation */ 309 !(ringacc->rings[i in k3_ringacc_request_ring() [all...] |
/kernel/linux/linux-6.6/drivers/block/xen-blkback/ |
H A D | xenbus.c | 84 if (!blkif->rings || !blkif->rings[0].irq || !blkif->vbd.bdev) in xen_update_blkif_status() 110 ring = &blkif->rings[i]; in xen_update_blkif_status() 124 ring = &blkif->rings[i]; in xen_update_blkif_status() 134 blkif->rings = kcalloc(blkif->nr_rings, sizeof(struct xen_blkif_ring), in xen_blkif_alloc_rings() 136 if (!blkif->rings) in xen_blkif_alloc_rings() 140 struct xen_blkif_ring *ring = &blkif->rings[r]; in xen_blkif_alloc_rings() 274 struct xen_blkif_ring *ring = &blkif->rings[r]; in xen_blkif_disconnect() 335 * blkif->rings was allocated in connect_ring, so we should free it in in xen_blkif_disconnect() 338 kfree(blkif->rings); in xen_blkif_disconnect() [all...] |
/kernel/linux/linux-5.10/drivers/crypto/qat/qat_common/ |
H A D | adf_transport.c | 234 ring = &bank->rings[ring_num]; in adf_create_ring() 297 adf_handle_response(&bank->rings[i]); in adf_ring_response_handler() 370 ring = &bank->rings[i]; in adf_init_bank() 381 "Invalid tx rings mask config\n"); in adf_init_bank() 384 tx_ring = &bank->rings[i - hw_data->tx_rx_gap]; in adf_init_bank() 399 ring = &bank->rings[i]; in adf_init_bank() 407 * adf_init_etr_data() - Initialize transport rings for acceleration device 410 * Function is the initializes the communications channels (rings) to the 473 struct adf_etr_ring_data *ring = &bank->rings[i]; in cleanup_bank() 495 * adf_cleanup_etr_data() - Clear transport rings fo [all...] |
/kernel/linux/linux-5.10/include/linux/ |
H A D | ptr_ring.h | 619 static inline int ptr_ring_resize_multiple(struct ptr_ring **rings, in ptr_ring_resize_multiple() argument 639 spin_lock_irqsave(&(rings[i])->consumer_lock, flags); in ptr_ring_resize_multiple() 640 spin_lock(&(rings[i])->producer_lock); in ptr_ring_resize_multiple() 641 queues[i] = __ptr_ring_swap_queue(rings[i], queues[i], in ptr_ring_resize_multiple() 643 spin_unlock(&(rings[i])->producer_lock); in ptr_ring_resize_multiple() 644 spin_unlock_irqrestore(&(rings[i])->consumer_lock, flags); in ptr_ring_resize_multiple()
|
H A D | skb_array.h | 201 static inline int skb_array_resize_multiple(struct skb_array **rings, in skb_array_resize_multiple() argument 206 return ptr_ring_resize_multiple((struct ptr_ring **)rings, in skb_array_resize_multiple()
|
/kernel/linux/linux-6.6/include/linux/ |
H A D | ptr_ring.h | 619 static inline int ptr_ring_resize_multiple(struct ptr_ring **rings, in ptr_ring_resize_multiple() argument 639 spin_lock_irqsave(&(rings[i])->consumer_lock, flags); in ptr_ring_resize_multiple() 640 spin_lock(&(rings[i])->producer_lock); in ptr_ring_resize_multiple() 641 queues[i] = __ptr_ring_swap_queue(rings[i], queues[i], in ptr_ring_resize_multiple() 643 spin_unlock(&(rings[i])->producer_lock); in ptr_ring_resize_multiple() 644 spin_unlock_irqrestore(&(rings[i])->consumer_lock, flags); in ptr_ring_resize_multiple()
|
/kernel/linux/linux-5.10/drivers/crypto/inside-secure/ |
H A D | safexcel.c | 29 MODULE_PARM_DESC(max_rings, "Maximum number of rings to use."); 36 * Map all interfaces/rings to register index 0 in eip197_trc_cache_setupvirt() 51 for (i = 0; i < priv->config.rings; i++) { in eip197_trc_cache_setupvirt() 502 for (i = 0; i < priv->config.rings; i++) { in safexcel_hw_setup_cdesc_rings() 550 for (i = 0; i < priv->config.rings; i++) { in safexcel_hw_setup_rdesc_rings() 592 priv->config.pes, priv->config.rings); in safexcel_hw_init() 652 /* enable HIA input interface arbiter and rings */ in safexcel_hw_init() 654 GENMASK(priv->config.rings - 1, 0), in safexcel_hw_init() 712 for (i = 0; i < priv->config.rings; i++) { in safexcel_hw_init() 738 for (i = 0; i < priv->config.rings; in safexcel_hw_init() [all...] |
/kernel/linux/linux-6.6/drivers/crypto/inside-secure/ |
H A D | safexcel.c | 29 MODULE_PARM_DESC(max_rings, "Maximum number of rings to use."); 36 * Map all interfaces/rings to register index 0 in eip197_trc_cache_setupvirt() 51 for (i = 0; i < priv->config.rings; i++) { in eip197_trc_cache_setupvirt() 510 for (i = 0; i < priv->config.rings; i++) { in safexcel_hw_setup_cdesc_rings() 558 for (i = 0; i < priv->config.rings; i++) { in safexcel_hw_setup_rdesc_rings() 600 priv->config.pes, priv->config.rings); in safexcel_hw_init() 660 /* enable HIA input interface arbiter and rings */ in safexcel_hw_init() 662 GENMASK(priv->config.rings - 1, 0), in safexcel_hw_init() 720 for (i = 0; i < priv->config.rings; i++) { in safexcel_hw_init() 746 for (i = 0; i < priv->config.rings; in safexcel_hw_init() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/netronome/nfp/ |
H A D | Makefile | 24 nfd3/rings.o \ 27 nfdk/rings.o \
|