H A D | odp.c | 553 u64 user_va, size_t bcnt, u32 *bytes_mapped, in pagefault_real_mr() 591 if (bytes_mapped) { in pagefault_real_mr() 595 *bytes_mapped += min_t(u32, new_mappings, bcnt); in pagefault_real_mr() 606 size_t bcnt, u32 *bytes_mapped, u32 flags) in pagefault_implicit_mr() 647 bytes_mapped, flags); in pagefault_implicit_mr() 691 u32 *bytes_mapped, u32 flags) in pagefault_dmabuf_mr() 722 if (bytes_mapped) in pagefault_dmabuf_mr() 723 *bytes_mapped += bcnt; in pagefault_dmabuf_mr() 738 u32 *bytes_mapped, u32 flags) in pagefault_mr() 746 return pagefault_dmabuf_mr(mr, bcnt, bytes_mapped, flag in pagefault_mr() 552 pagefault_real_mr(struct mlx5_ib_mr *mr, struct ib_umem_odp *odp, u64 user_va, size_t bcnt, u32 *bytes_mapped, u32 flags) pagefault_real_mr() argument 604 pagefault_implicit_mr(struct mlx5_ib_mr *imr, struct ib_umem_odp *odp_imr, u64 user_va, size_t bcnt, u32 *bytes_mapped, u32 flags) pagefault_implicit_mr() argument 690 pagefault_dmabuf_mr(struct mlx5_ib_mr *mr, size_t bcnt, u32 *bytes_mapped, u32 flags) pagefault_dmabuf_mr() argument 737 pagefault_mr(struct mlx5_ib_mr *mr, u64 io_virt, size_t bcnt, u32 *bytes_mapped, u32 flags) pagefault_mr() argument 813 pagefault_single_data_segment(struct mlx5_ib_dev *dev, struct ib_pd *pd, u32 key, u64 io_virt, size_t bcnt, u32 *bytes_committed, u32 *bytes_mapped) pagefault_single_data_segment() argument 981 pagefault_data_segments(struct mlx5_ib_dev *dev, struct mlx5_pagefault *pfault, void *wqe, void *wqe_end, u32 *bytes_mapped, u32 *total_wqe_bytes, bool receive_queue) pagefault_data_segments() argument 1199 u32 bytes_mapped, total_wqe_bytes; mlx5_ib_mr_wqe_pfault_handler() local 1706 u32 bytes_mapped = 0; mlx5_ib_prefetch_mr_work() local 1755 u32 bytes_mapped = 0; mlx5_ib_prefetch_sg_list() local [all...] |