Lines Matching refs:vd_dev
119 struct virtio_vdpa_device *vd_dev = private;
121 virtio_config_changed(&vd_dev->vdev);
138 struct virtio_vdpa_device *vd_dev = to_virtio_vdpa_device(vdev);
202 spin_lock_irqsave(&vd_dev->lock, flags);
203 list_add(&info->node, &vd_dev->virtqueues);
204 spin_unlock_irqrestore(&vd_dev->lock, flags);
220 struct virtio_vdpa_device *vd_dev = to_virtio_vdpa_device(vq->vdev);
221 struct vdpa_device *vdpa = vd_dev->vdpa;
227 spin_lock_irqsave(&vd_dev->lock, flags);
229 spin_unlock_irqrestore(&vd_dev->lock, flags);
255 struct virtio_vdpa_device *vd_dev = to_virtio_vdpa_device(vdev);
277 cb.private = vd_dev;
307 struct virtio_vdpa_device *vd_dev = to_virtio_vdpa_device(vdev);
308 struct vdpa_device *vdpa = vd_dev->vdpa;
331 struct virtio_vdpa_device *vd_dev =
334 kfree(vd_dev);
340 struct virtio_vdpa_device *vd_dev, *reg_dev = NULL;
343 vd_dev = kzalloc(sizeof(*vd_dev), GFP_KERNEL);
344 if (!vd_dev)
347 vd_dev->vdev.dev.parent = vdpa_get_dma_dev(vdpa);
348 vd_dev->vdev.dev.release = virtio_vdpa_release_dev;
349 vd_dev->vdev.config = &virtio_vdpa_config_ops;
350 vd_dev->vdpa = vdpa;
351 INIT_LIST_HEAD(&vd_dev->virtqueues);
352 spin_lock_init(&vd_dev->lock);
354 vd_dev->vdev.id.device = ops->get_device_id(vdpa);
355 if (vd_dev->vdev.id.device == 0)
358 vd_dev->vdev.id.vendor = ops->get_vendor_id(vdpa);
359 ret = register_virtio_device(&vd_dev->vdev);
360 reg_dev = vd_dev;
364 vdpa_set_drvdata(vdpa, vd_dev);
370 put_device(&vd_dev->vdev.dev);
372 kfree(vd_dev);
378 struct virtio_vdpa_device *vd_dev = vdpa_get_drvdata(vdpa);
380 unregister_virtio_device(&vd_dev->vdev);