Lines Matching defs:dev

481 	dev_dbg(&qm->pdev->dev, "change qm state from %s to %s\n",
485 dev_warn(&qm->pdev->dev, "Can not change qm state from %s to %s\n",
527 dev_dbg(&qm->pdev->dev, "change qp state from %s to %s in QM %s\n",
531 dev_warn(&qm->pdev->dev,
658 dev_err(&qm->pdev->dev, "QM mailbox is busy to start!\n");
666 dev_err(&qm->pdev->dev, "QM mailbox operation timeout!\n");
673 dev_err(&qm->pdev->dev, "QM mailbox operation failed!\n");
691 dev_dbg(&qm->pdev->dev, "QM mailbox request to q%u: %u-%llx\n",
737 dev_dbg(&qm->pdev->dev, "QM doorbell request: qn=%u, cmd=%u, index=%u\n",
809 struct device *dev = &qm->pdev->dev;
817 dev_err(dev, "algs size %u is equal or larger than %d.\n",
822 algs = devm_kzalloc(dev, QM_DEV_ALG_MAX_LEN * sizeof(char), GFP_KERNEL);
850 struct device *dev = &qm->pdev->dev;
856 ret = pm_runtime_resume_and_get(dev);
858 dev_err(dev, "failed to get_sync(%d).\n", ret);
867 struct device *dev = &qm->pdev->dev;
872 pm_runtime_mark_last_busy(dev);
873 pm_runtime_put_autosuspend(dev);
994 dev_warn(&qm->pdev->dev, "Driver is down, message cannot be processed!\n");
1029 struct device *dev = &qm->pdev->dev;
1037 dev_err(dev, "reset function not ready\n");
1043 dev_err(dev, "failed to stop qm when reset function\n");
1049 dev_err(dev, "failed to start qm when reset function\n");
1070 dev_err(&qm->pdev->dev, "eq overflow, reset function\n");
1074 dev_err(&qm->pdev->dev, "cq overflow, stop qp(%u)\n",
1081 dev_err(&qm->pdev->dev, "unknown error type %u\n",
1113 struct device *dev = &qm->pdev->dev;
1130 dev_err(dev, "system page size is not support: %lu, default set to 4KB",
1290 dev_err(&qm->pdev->dev, "failed to calculate shaper parameter!\n");
1352 struct device *dev = &qm->pdev->dev;
1359 *dma_addr = dma_map_single(dev, ctx_addr, ctx_size, DMA_FROM_DEVICE);
1360 if (dma_mapping_error(dev, *dma_addr)) {
1361 dev_err(dev, "DMA mapping error!\n");
1372 struct device *dev = &qm->pdev->dev;
1374 dma_unmap_single(dev, *dma_addr, ctx_size, DMA_FROM_DEVICE);
1455 struct device *dev = &qm->pdev->dev;
1464 dev_err(dev, "%s [error status=0x%x] found\n",
1472 dev_err(dev, "qm %s doorbell timeout in function %u\n",
1481 dev_err(dev, "qm %s fifo overflow in function %u\n",
1484 dev_err(dev, "unknown error type\n");
1545 struct device *dev = &qm->pdev->dev;
1552 dev_err(dev, "failed to get msg from VF(%u)!\n", vf_id);
1559 dev_err(dev, "failed to stop VF(%u)!\n", vf_id);
1562 dev_err(dev, "failed to start VF(%u)!\n", vf_id);
1568 dev_err(dev, "unsupported cmd %u sent by VF(%u)!\n", cmd, vf_id);
1575 struct device *dev = &qm->pdev->dev;
1604 dev_err(dev, "VF(%u) not ping PF!\n", i);
1638 struct device *dev = &qm->pdev->dev;
1648 dev_err(dev, "failed to send command to vf(%u)!\n", fun_num);
1661 dev_err(dev, "failed to get response from VF(%u)!\n", fun_num);
1674 struct device *dev = &qm->pdev->dev;
1687 dev_err(dev, "failed to send command to VFs!\n");
1711 dev_err(dev, "failed to get response from VF(%u)!\n", i);
1728 dev_err(&qm->pdev->dev, "failed to send command to PF!\n");
1889 struct device *dev = &qm->pdev->dev;
1897 dev_info_ratelimited(dev, "All %u queues of QM are busy!\n",
1905 dev_info_ratelimited(dev, "All %u queues of QM are busy!\n",
1980 struct device *dev = &qm->pdev->dev;
2005 sqc_dma = dma_map_single(dev, sqc, sizeof(struct qm_sqc),
2007 if (dma_mapping_error(dev, sqc_dma)) {
2013 dma_unmap_single(dev, sqc_dma, sizeof(struct qm_sqc), DMA_TO_DEVICE);
2022 struct device *dev = &qm->pdev->dev;
2046 cqc_dma = dma_map_single(dev, cqc, sizeof(struct qm_cqc),
2048 if (dma_mapping_error(dev, cqc_dma)) {
2054 dma_unmap_single(dev, cqc_dma, sizeof(struct qm_cqc), DMA_TO_DEVICE);
2076 struct device *dev = &qm->pdev->dev;
2089 dev_dbg(dev, "queue %d started\n", qp_id);
2149 struct device *dev = &qm->pdev->dev;
2164 dev_err(dev, "Failed to stop qp(%u)!\n", qp->qp_id);
2170 dev_err(dev, "Failed to alloc ctx for sqc and cqc!\n");
2177 dev_err_ratelimited(dev, "Failed to dump sqc!\n");
2185 dev_err_ratelimited(dev, "Failed to dump cqc!\n");
2195 dev_err(dev, "Fail to empty queue %u!\n", qp->qp_id);
2210 struct device *dev = &qp->qm->pdev->dev;
2231 dev_err(dev, "Failed to drain out data for stopping!\n");
2238 dev_dbg(dev, "stop queue %u!", qp->qp_id);
2286 dev_info_ratelimited(&qp->qm->pdev->dev, "QP is stopped or resetting\n");
2314 dev_err(&qm->pdev->dev, "QM writeback sqc cache fail!\n");
2383 struct device *dev = &pdev->dev;
2416 ret = dma_mmap_coherent(dev, vma, qp->qdma.va,
2652 ret = strscpy(interface.name, dev_driver_string(&pdev->dev),
2657 uacce = uacce_alloc(&pdev->dev, &interface);
2734 struct pci_dev *dev;
2743 dev = qm->pdev;
2744 if (dev == pdev)
2746 if (pci_physfn(dev) == pdev) {
2747 vf_qm = pci_get_drvdata(dev);
2787 struct device *dev = &qm->pdev->dev;
2793 dma_free_coherent(dev, qdma->size, qdma->va, qdma->dma);
2804 struct device *dev = &qm->pdev->dev;
2815 qp->qdma.va = dma_alloc_coherent(dev, dma_size, &qp->qdma.dma,
2853 if (!acpi_device_power_manageable(ACPI_COMPANION(&pdev->dev)))
2854 dev_info(&pdev->dev, "_PS0 and _PR0 are not defined");
2919 struct device *dev = &qm->pdev->dev;
2924 dma_free_coherent(dev, qm->qdma.size,
2982 dev_err(&qm->pdev->dev, "Don't support vft read!\n");
3043 struct device *dev = &qm->pdev->dev;
3058 eqc_dma = dma_map_single(dev, eqc, sizeof(struct qm_eqc),
3060 if (dma_mapping_error(dev, eqc_dma)) {
3066 dma_unmap_single(dev, eqc_dma, sizeof(struct qm_eqc), DMA_TO_DEVICE);
3074 struct device *dev = &qm->pdev->dev;
3087 aeqc_dma = dma_map_single(dev, aeqc, sizeof(struct qm_aeqc),
3089 if (dma_mapping_error(dev, aeqc_dma)) {
3095 dma_unmap_single(dev, aeqc_dma, sizeof(struct qm_aeqc), DMA_TO_DEVICE);
3103 struct device *dev = &qm->pdev->dev;
3110 dev_err(dev, "Set eqc failed!\n");
3155 struct device *dev = &qm->pdev->dev;
3165 dev_dbg(dev, "qm start with %u queue pairs\n", qm->qp_num);
3168 dev_err(dev, "qp_num should not be 0\n");
3186 struct device *dev = &qm->pdev->dev;
3201 dev_err(dev, "Failed to start qp%d!\n", i);
3217 struct device *dev = &qm->pdev->dev;
3227 dev_err(dev, "Failed to stop qp%d!\n", i);
3268 struct device *dev = &qm->pdev->dev;
3284 dev_err(dev, "Failed to stop started qp!\n");
3294 dev_err(dev, "Failed to set vft!\n");
3312 dev_err(&qm->pdev->dev, "QM doesn't support hw error handling!\n");
3322 dev_err(&qm->pdev->dev, "Unexpected QM hw error uninit!\n");
3332 dev_err(&qm->pdev->dev, "QM doesn't support hw error report!\n");
3353 dev_err(&qm->pdev->dev, "Device doesn't support hw error init!\n");
3374 dev_err(&qm->pdev->dev, "Unexpected device hw error uninit!\n");
3414 struct device *dev;
3418 dev = &qm->pdev->dev;
3420 dev_node = dev_to_node(dev);
3562 struct device *dev = &qm->pdev->dev;
3574 dev_err(dev, "failed to calculate shaper parameter!\n");
3582 dev_err(dev, "type: %d, failed to set shaper vft!\n", i);
3642 struct device *dev = &qm->pdev->dev;
3649 dev_err(dev, "function(%u) failed to get qos by PF!\n", fun_num);
3656 dev_err(dev, "failed to send cmd to VF(%u)!\n", fun_num);
3702 pci_err(qm->pdev, "dev resetting, read alg qos failed!\n");
3732 const struct bus_type *bus_type = qm->pdev->dev.bus;
3736 struct device *dev;
3749 dev = bus_find_device_by_name(bus_type, NULL, tbuf_bdf);
3750 if (!dev) {
3755 pdev = container_of(dev, struct pci_dev, dev);
3788 pci_err(qm->pdev, "dev resetting, write alg qos failed!\n");
3965 dev_err(&qm->pdev->dev, "Device doesn't support get hw error status!\n");
4046 dev_err(&pdev->dev, "Fails to read QM reg!\n");
4055 dev_err(&pdev->dev, "Fails to read QM reg in the second time!\n");
4278 if (ACPI_HANDLE(&pdev->dev)) {
4282 s = acpi_evaluate_integer(ACPI_HANDLE(&pdev->dev),
4383 /* clear dev ecc 2bit error source if having */
4680 dev_err(&pdev->dev, "Fail to stop qm in shutdown!\n");
4700 dev_err(&qm->pdev->dev, "controller reset failed (%d)\n", ret);
4714 dev_err(&pdev->dev, "reset prepare not ready!\n");
4722 dev_err(&pdev->dev, "failed to stop QM, ret = %d.\n", ret);
4737 dev_warn(&pdev->dev, "PF responds timeout in reset prepare!\n");
4749 dev_err(&pdev->dev, "failed to start QM, ret = %d.\n", ret);
4756 dev_warn(&pdev->dev, "PF responds timeout in reset done!\n");
4763 struct device *dev = &qm->pdev->dev;
4774 dev_err(dev, "couldn't get reset done status from PF, timeout!\n");
4785 dev_err(dev, "failed to get msg from PF in reset done!\n");
4791 dev_err(dev, "the cmd(%u) is not reset done!\n", cmd);
4801 struct device *dev = &qm->pdev->dev;
4804 dev_info(dev, "device reset start...\n");
4816 dev_info(dev, "device reset done.\n");
4827 struct device *dev = &qm->pdev->dev;
4839 dev_err(dev, "failed to get msg from source!\n");
4858 dev_err(dev, "unsupported cmd %u sent by function(%u)!\n", cmd, fun_num);
4897 struct device *dev = &qm->pdev->dev;
4908 dev_info(dev, "HW V2 not both use uacce sva mode and hardware crypto algs.\n");
4980 dev_err(&qm->pdev->dev, "failed to request abnormal irq, ret = %d", ret);
5011 dev_err(&pdev->dev, "failed to request function communication irq, ret = %d", ret);
5043 dev_err(&pdev->dev, "failed to request eq irq, ret = %d", ret);
5074 dev_err(&pdev->dev, "failed to request eq irq, ret = %d", ret);
5120 struct device *dev = &qm->pdev->dev;
5142 dev_err(dev, "qp num(%u) is more than max qp num(%u)!\n",
5147 dev_info(dev, "Default qp num(%u) is too big, reset it to Function's max qp num(%u)!\n",
5162 qm_cap = devm_kzalloc(&pdev->dev, sizeof(*qm_cap) * size, GFP_KERNEL);
5216 struct device *dev = &pdev->dev;
5221 dev_err(dev, "Failed to request mem regions!\n");
5270 struct device *dev = &pdev->dev;
5276 dev_err(dev, "Failed to enable device mem!\n");
5284 ret = dma_set_mask_and_coherent(dev, DMA_BIT_MASK(64));
5292 dev_err(dev, "Failed to enable MSI vectors!\n");
5331 struct device *dev = &qm->pdev->dev;
5357 dev_dbg(dev, "allocate qp dma buf size=%zx)\n", qp_dma_size);
5369 struct device *dev = &qm->pdev->dev;
5395 qm->qdma.va = dma_alloc_coherent(dev, qm->qdma.size, &qm->qdma.dma,
5397 dev_dbg(dev, "allocate qm dma buf size=%zx)\n", qm->qdma.size);
5415 dma_free_coherent(dev, qm->qdma.size, qm->qdma.va, qm->qdma.dma);
5433 struct device *dev = &pdev->dev;
5452 dev_err(dev, "failed to reset device memory\n");
5460 dev_warn(dev, "fail to alloc uacce (%d)\n", ret);
5499 struct device *dev = &qm->pdev->dev;
5501 if (pm_runtime_suspended(dev)) {
5502 dev_info(dev, "can not read/write - device in suspended.\n");
5530 struct device *dev = &qm->pdev->dev;
5535 pm_runtime_set_autosuspend_delay(dev, QM_AUTOSUSPEND_DELAY);
5536 pm_runtime_use_autosuspend(dev);
5537 pm_runtime_put_noidle(dev);
5549 struct device *dev = &qm->pdev->dev;
5554 pm_runtime_get_noresume(dev);
5555 pm_runtime_dont_use_autosuspend(dev);
5628 * @dev: device to suspend.
5632 int hisi_qm_suspend(struct device *dev)
5634 struct pci_dev *pdev = to_pci_dev(dev);
5656 * @dev: device to resume.
5660 int hisi_qm_resume(struct device *dev)
5662 struct pci_dev *pdev = to_pci_dev(dev);