Lines Matching refs:vqs

133 	struct vhost_net_virtqueue vqs[VHOST_NET_VQ_MAX];
274 kfree(n->vqs[i].ubuf_info);
275 n->vqs[i].ubuf_info = NULL;
288 n->vqs[i].ubuf_info =
290 sizeof(*n->vqs[i].ubuf_info),
292 if (!n->vqs[i].ubuf_info)
309 n->vqs[i].done_idx = 0;
310 n->vqs[i].upend_idx = 0;
311 n->vqs[i].ubufs = NULL;
312 n->vqs[i].vhost_hlen = 0;
313 n->vqs[i].sock_hlen = 0;
314 vhost_net_buf_init(&n->vqs[i].rxq);
426 struct vhost_poll *poll = n->poll + (nvq - n->vqs);
437 struct vhost_poll *poll = n->poll + (nvq - n->vqs);
574 struct vhost_net_virtqueue *rnvq = &net->vqs[VHOST_NET_VQ_RX];
599 struct vhost_net_virtqueue *nvq = &net->vqs[VHOST_NET_VQ_TX];
770 struct vhost_net_virtqueue *nvq = &net->vqs[VHOST_NET_VQ_TX];
859 struct vhost_net_virtqueue *nvq = &net->vqs[VHOST_NET_VQ_TX];
961 struct vhost_net_virtqueue *nvq = &net->vqs[VHOST_NET_VQ_TX];
1009 struct vhost_net_virtqueue *rnvq = &net->vqs[VHOST_NET_VQ_RX];
1010 struct vhost_net_virtqueue *tnvq = &net->vqs[VHOST_NET_VQ_TX];
1108 struct vhost_net_virtqueue *nvq = &net->vqs[VHOST_NET_VQ_RX];
1288 struct vhost_virtqueue **vqs;
1296 vqs = kmalloc_array(VHOST_NET_VQ_MAX, sizeof(*vqs), GFP_KERNEL);
1297 if (!vqs) {
1305 kfree(vqs);
1309 n->vqs[VHOST_NET_VQ_RX].rxq.queue = queue;
1313 kfree(vqs);
1318 n->vqs[VHOST_NET_VQ_TX].xdp = xdp;
1321 vqs[VHOST_NET_VQ_TX] = &n->vqs[VHOST_NET_VQ_TX].vq;
1322 vqs[VHOST_NET_VQ_RX] = &n->vqs[VHOST_NET_VQ_RX].vq;
1323 n->vqs[VHOST_NET_VQ_TX].vq.handle_kick = handle_tx_kick;
1324 n->vqs[VHOST_NET_VQ_RX].vq.handle_kick = handle_rx_kick;
1326 n->vqs[i].ubufs = NULL;
1327 n->vqs[i].ubuf_info = NULL;
1328 n->vqs[i].upend_idx = 0;
1329 n->vqs[i].done_idx = 0;
1330 n->vqs[i].batched_xdp = 0;
1331 n->vqs[i].vhost_hlen = 0;
1332 n->vqs[i].sock_hlen = 0;
1333 n->vqs[i].rx_ring = NULL;
1334 vhost_net_buf_init(&n->vqs[i].rxq);
1336 vhost_dev_init(dev, vqs, VHOST_NET_VQ_MAX,
1371 *tx_sock = vhost_net_stop_vq(n, &n->vqs[VHOST_NET_VQ_TX].vq);
1372 *rx_sock = vhost_net_stop_vq(n, &n->vqs[VHOST_NET_VQ_RX].vq);
1378 vhost_poll_flush(&n->vqs[index].vq.poll);
1385 if (n->vqs[VHOST_NET_VQ_TX].ubufs) {
1386 mutex_lock(&n->vqs[VHOST_NET_VQ_TX].vq.mutex);
1388 mutex_unlock(&n->vqs[VHOST_NET_VQ_TX].vq.mutex);
1390 vhost_net_ubuf_put_and_wait(n->vqs[VHOST_NET_VQ_TX].ubufs);
1391 mutex_lock(&n->vqs[VHOST_NET_VQ_TX].vq.mutex);
1393 atomic_set(&n->vqs[VHOST_NET_VQ_TX].ubufs->refcount, 1);
1394 mutex_unlock(&n->vqs[VHOST_NET_VQ_TX].vq.mutex);
1418 kfree(n->vqs[VHOST_NET_VQ_RX].rxq.queue);
1419 kfree(n->vqs[VHOST_NET_VQ_TX].xdp);
1420 kfree(n->dev.vqs);
1514 vq = &n->vqs[index].vq;
1515 nvq = &n->vqs[index];
1657 mutex_lock(&n->vqs[i].vq.mutex);
1658 n->vqs[i].vq.acked_features = features;
1659 n->vqs[i].vhost_hlen = vhost_hlen;
1660 n->vqs[i].sock_hlen = sock_hlen;
1661 mutex_unlock(&n->vqs[i].vq.mutex);