Lines Matching defs:device
90 struct nvme_rdma_device *device;
118 struct nvme_rdma_device *device;
224 * lifetime. It's safe, since any chage in the underlying RDMA device
259 struct nvme_rdma_device *dev = queue->device;
373 if (ndev->dev->node_guid == cm_id->device->node_guid &&
382 ndev->dev = cm_id->device;
429 dev = queue->device;
496 queue->device = nvme_rdma_find_get_device(queue->cm_id);
497 if (!queue->device) {
498 dev_err(queue->cm_id->device->dev.parent,
502 ibdev = queue->device->dev;
533 dev_err(queue->ctrl->ctrl.device,
544 dev_err(queue->ctrl->ctrl.device,
565 nvme_rdma_dev_put(queue->device);
595 dev_info(ctrl->ctrl.device,
609 dev_info(ctrl->ctrl.device,
616 dev_info(ctrl->ctrl.device,
691 dev_info(ctrl->ctrl.device,
728 dev_err(ctrl->ctrl.device,
734 dev_info(ctrl->ctrl.device,
773 nvme_rdma_free_qe(ctrl->device->dev, &ctrl->async_event_sqe,
790 ctrl->device = ctrl->queues[0].device;
791 ctrl->ctrl.numa_node = ibdev_to_node(ctrl->device->dev);
794 if (ctrl->device->dev->attrs.kernel_cap_flags &
798 ctrl->max_fr_pages = nvme_rdma_get_max_fr_pages(ctrl->device->dev,
803 * It's safe, since any chage in the underlying RDMA device will issue
806 error = nvme_rdma_alloc_qe(ctrl->device->dev, &ctrl->async_event_sqe,
855 nvme_rdma_free_qe(ctrl->device->dev, &ctrl->async_event_sqe,
996 dev_info(ctrl->ctrl.device, "Reconnecting in %d seconds...\n",
1016 dev_err(ctrl->ctrl.device, "icdoff is not supported!\n");
1022 dev_err(ctrl->ctrl.device,
1028 dev_warn(ctrl->ctrl.device,
1034 dev_warn(ctrl->ctrl.device,
1041 dev_warn(ctrl->ctrl.device,
1106 dev_info(ctrl->ctrl.device, "Successfully reconnected (%d attempts)\n",
1114 dev_info(ctrl->ctrl.device, "Failed reconnect attempt %d\n",
1149 dev_warn(ctrl->ctrl.device, "starting error recovery\n");
1170 dev_info(ctrl->ctrl.device,
1231 struct nvme_rdma_device *dev = queue->device;
1273 sge->lkey = queue->device->pd->local_dma_lkey;
1293 put_unaligned_le32(queue->device->pd->unsafe_global_rkey, sg->key);
1521 struct nvme_rdma_device *dev = queue->device;
1592 sge->lkey = queue->device->pd->local_dma_lkey;
1608 dev_err(queue->ctrl->ctrl.device,
1623 list.lkey = queue->device->pd->local_dma_lkey;
1634 dev_err(queue->ctrl->ctrl.device,
1659 struct ib_device *dev = queue->device->dev;
1690 dev_err(queue->ctrl->ctrl.device,
1704 dev_err(queue->ctrl->ctrl.device,
1714 dev_err(queue->ctrl->ctrl.device,
1731 struct ib_device *ibdev = queue->device->dev;
1742 dev_err(queue->ctrl->ctrl.device,
1794 dev_err(queue->ctrl->ctrl.device,
1798 dev_err(queue->ctrl->ctrl.device,
1818 dev_err(ctrl->device, "rdma_resolve_route failed (%d).\n",
1840 param.responder_resources = queue->device->dev->attrs.max_qp_rd_atom;
1868 dev_err(ctrl->ctrl.device,
1882 dev_dbg(queue->ctrl->ctrl.device, "%s (%d): status %d id %p\n",
1905 dev_dbg(queue->ctrl->ctrl.device,
1912 dev_dbg(queue->ctrl->ctrl.device,
1917 /* device removal is handled via the ib_client API */
1920 dev_err(queue->ctrl->ctrl.device,
1949 dev_warn(ctrl->ctrl.device, "I/O %d QID %d timeout\n",
1997 dev = queue->device->dev;
2026 dev_err(queue->ctrl->ctrl.device,
2101 struct ib_device *ibdev = queue->device->dev;
2219 static struct nvme_ctrl *nvme_rdma_create_ctrl(struct device *dev,
2293 dev_info(ctrl->ctrl.device, "new ctrl: NQN \"%s\", addr %pISpcs\n",
2344 /* Delete all controllers using this device */
2347 if (ctrl->device->dev != ib_device)