Lines Matching refs:ret
411 struct mlx5_ib_mr *ret;
450 ret = ERR_PTR(err);
455 ret = __xa_cmpxchg(&imr->implicit_children, idx, NULL, mr,
457 if (unlikely(ret)) {
458 if (xa_is_err(ret)) {
459 ret = ERR_PTR(xa_err(ret));
466 refcount_inc(&ret->mmkey.usecount);
478 return ret;
556 int page_shift, ret, np;
581 ret = mlx5r_umr_update_xlt(mr, start_idx, np, page_shift, xlt_flags);
584 if (ret < 0) {
585 if (ret != -EAGAIN)
601 return ret;
613 int ret;
632 ret = PTR_ERR(mtt);
646 ret = pagefault_real_mr(mtt, umem_odp, user_va, len,
651 if (ret < 0)
655 npages += ret;
658 ret = npages;
667 return ret;
687 return ret;
766 int ret;
768 ret = pagefault_real_mr(mr, to_ib_umem_odp(mr->umem), mr->umem->address,
771 return ret >= 0 ? 0 : ret;
776 int ret;
778 ret = pagefault_dmabuf_mr(mr, mr->umem->length, NULL,
781 return ret >= 0 ? 0 : ret;
819 int npages = 0, ret, i, outlen, cur_outlen = 0, depth = 0;
846 ret = 0;
854 ret = -EFAULT;
862 ret = pagefault_mr(mr, io_virt, bcnt, bytes_mapped, 0);
863 if (ret < 0)
866 mlx5_update_odp_stats(mr, faults, ret);
868 npages += ret;
869 ret = 0;
876 ret = -EFAULT;
887 ret = -ENOMEM;
896 ret = mlx5_core_query_mkey(dev->mdev, mmkey->key, out, outlen);
897 if (ret)
911 ret = -ENOMEM;
930 ret = -EFAULT;
959 return ret ? ret : npages;
987 int ret = 0, npages = 0;
1038 ret = pagefault_single_data_segment(dev, NULL, be32_to_cpu(key),
1042 if (ret < 0)
1044 npages += ret;
1047 return ret < 0 ? ret : npages;
1066 mlx5_ib_err(dev, "Unable to read the complete WQE. ds = 0x%x, ret = 0x%x\n",
1204 int ret = 0;
1228 ret = mlx5_ib_read_wqe_sq(qp, wqe_index, wqe, PAGE_SIZE,
1230 if (ret)
1232 ret = mlx5_ib_mr_initiator_pfault_handler(
1235 ret = mlx5_ib_read_wqe_rq(qp, wqe_index, wqe, PAGE_SIZE,
1237 if (ret)
1239 ret = mlx5_ib_mr_responder_pfault_handler_rq(
1244 ret = mlx5_ib_read_wqe_srq(srq, wqe_index, wqe, PAGE_SIZE,
1246 if (ret)
1248 ret = mlx5_ib_mr_responder_pfault_handler_srq(
1252 if (ret < 0 || wqe >= wqe_end)
1255 ret = pagefault_data_segments(dev, pfault, wqe, wqe_end, &bytes_mapped,
1257 if (ret == -EAGAIN)
1260 if (ret < 0 || total_wqe_bytes > bytes_mapped)
1264 ret = 0;
1268 if (ret)
1272 ret, wqe_index, pfault->token);
1297 int ret;
1322 ret = pagefault_single_data_segment(dev, NULL, rkey, address, length,
1324 if (ret == -EAGAIN) {
1327 } else if (ret < 0 || pages_in_range(address, length) > ret) {
1329 if (ret != -ENOENT)
1331 ret, pfault->token, pfault->type);
1348 ret = pagefault_single_data_segment(dev, NULL, rkey, address,
1351 if (ret < 0 && ret != -EAGAIN) {
1352 mlx5_ib_dbg(dev, "Prefetch failed. ret: %d, QP 0x%x, address: 0x%.16llx, length = 0x%.16x\n",
1353 ret, pfault->token, address, prefetch_len);
1707 int ret;
1713 ret = pagefault_mr(work->frags[i].mr, work->frags[i].io_virt,
1716 if (ret <= 0)
1718 mlx5_update_odp_stats(work->frags[i].mr, prefetch, ret);
1756 int ret = 0;
1765 ret = pagefault_mr(mr, sg_list[i].addr, sg_list[i].length,
1767 if (ret < 0) {
1769 return ret;
1771 mlx5_update_odp_stats(mr, prefetch, ret);