/kernel/linux/linux-6.6/drivers/vdpa/vdpa_sim/ |
H A D | vdpa_sim_net.c | 109 struct vdpasim_virtqueue *cvq = &vdpasim->vqs[2]; in vdpasim_handle_ctrl_mac() local 115 read = vringh_iov_pull_iotlb(&cvq->vring, &cvq->in_iov, in vdpasim_handle_ctrl_mac() 129 struct vdpasim_virtqueue *cvq = &vdpasim->vqs[2]; in vdpasim_handle_cvq() local 140 if (!cvq->ready) in vdpasim_handle_cvq() 144 err = vringh_getdesc_iotlb(&cvq->vring, &cvq->in_iov, in vdpasim_handle_cvq() 145 &cvq->out_iov, in vdpasim_handle_cvq() 146 &cvq->head, GFP_ATOMIC); in vdpasim_handle_cvq() 151 read = vringh_iov_pull_iotlb(&cvq in vdpasim_handle_cvq() [all...] |
/kernel/linux/linux-6.6/drivers/vdpa/mlx5/net/ |
H A D | mlx5_vnet.c | 1799 struct mlx5_control_vq *cvq = &mvdev->cvq; in handle_ctrl_mac() local 1808 read = vringh_iov_pull_iotlb(&cvq->vring, &cvq->riov, (void *)mac, ETH_ALEN); in handle_ctrl_mac() 1927 struct mlx5_control_vq *cvq = &mvdev->cvq; in handle_ctrl_mq() local 1946 read = vringh_iov_pull_iotlb(&cvq->vring, &cvq->riov, (void *)&mq, sizeof(mq)); in handle_ctrl_mq() 1975 struct mlx5_control_vq *cvq = &mvdev->cvq; in handle_ctrl_vlan() local 2017 struct mlx5_control_vq *cvq; mlx5_cvq_kick_handler() local 2141 struct mlx5_control_vq *cvq = &mvdev->cvq; mlx5_vdpa_set_vq_num() local 2163 struct mlx5_control_vq *cvq = container_of(vring, struct mlx5_control_vq, vring); mlx5_cvq_notify() local 2173 struct mlx5_control_vq *cvq = &mvdev->cvq; set_cvq_ready() local 2800 struct mlx5_control_vq *cvq = &mvdev->cvq; setup_cvq_vring() local 3098 struct mlx5_control_vq *cvq; mlx5_vdpa_get_vendor_vq_stats() local 3148 struct mlx5_control_vq *cvq; mlx5_vdpa_cvq_suspend() local [all...] |
/kernel/linux/linux-6.6/drivers/vdpa/mlx5/core/ |
H A D | resources.c | 232 mvdev->cvq.iotlb = vhost_iotlb_alloc(0, 0); in init_ctrl_vq() 233 if (!mvdev->cvq.iotlb) in init_ctrl_vq() 236 spin_lock_init(&mvdev->cvq.iommu_lock); in init_ctrl_vq() 237 vringh_set_iotlb(&mvdev->cvq.vring, mvdev->cvq.iotlb, &mvdev->cvq.iommu_lock); in init_ctrl_vq() 244 vhost_iotlb_free(mvdev->cvq.iotlb); in cleanup_ctrl_vq()
|
H A D | mlx5_vdpa.h | 97 struct mlx5_control_vq cvq; member
|
H A D | mr.c | 460 err = vhost_iotlb_add_range(mvdev->cvq.iotlb, start, last, start, VHOST_ACCESS_RW); in dup_iotlb() 466 err = vhost_iotlb_add_range(mvdev->cvq.iotlb, map->start, map->last, in dup_iotlb() 476 vhost_iotlb_del_range(mvdev->cvq.iotlb, 0, ULLONG_MAX); in prune_iotlb()
|
/kernel/linux/linux-5.10/drivers/net/ |
H A D | virtio_net.c | 183 struct virtqueue *cvq; member 274 * 0:rx0 1:tx0 2:rx1 3:tx1 ... 2N:rxN 2N+1:txN 2N+2:cvq 1779 virtqueue_add_sgs(vi->cvq, sgs, out_num, 1, vi, GFP_ATOMIC); in virtnet_send_command() 1781 if (unlikely(!virtqueue_kick(vi->cvq))) in virtnet_send_command() 1787 while (!virtqueue_get_buf(vi->cvq, &tmp) && in virtnet_send_command() 1788 !virtqueue_is_broken(vi->cvq)) in virtnet_send_command() 2877 vi->cvq = vqs[total_vqs - 1]; in virtnet_find_vqs()
|
/kernel/linux/linux-6.6/drivers/net/ |
H A D | virtio_net.c | 239 struct virtqueue *cvq; member 355 * 0:rx0 1:tx0 2:rx1 3:tx1 ... 2N:rxN 2N+1:txN 2N+2:cvq 2491 ret = virtqueue_add_sgs(vi->cvq, sgs, out_num, 1, vi, GFP_ATOMIC); in virtnet_send_command() 2498 if (unlikely(!virtqueue_kick(vi->cvq))) in virtnet_send_command() 2504 while (!virtqueue_get_buf(vi->cvq, &tmp) && in virtnet_send_command() 2505 !virtqueue_is_broken(vi->cvq)) in virtnet_send_command() 4154 vi->cvq = vqs[total_vqs - 1]; in virtnet_find_vqs()
|