Lines Matching defs:umem_odp
230 struct ib_umem_odp *umem_odp =
244 mutex_lock(&umem_odp->umem_mutex);
247 * If npages is zero then umem_odp->private may not be setup yet. This
250 if (!umem_odp->npages)
252 mr = umem_odp->private;
254 start = max_t(u64, ib_umem_start(umem_odp), range->start);
255 end = min_t(u64, ib_umem_end(umem_odp), range->end);
263 for (addr = start; addr < end; addr += BIT(umem_odp->page_shift)) {
264 idx = (addr - ib_umem_start(umem_odp)) >> umem_odp->page_shift;
271 if (umem_odp->dma_list[idx] &
306 ib_umem_odp_unmap_dma_pages(umem_odp, start, end);
308 if (unlikely(!umem_odp->npages && mr->parent))
311 mutex_unlock(&umem_odp->umem_mutex);
485 struct ib_umem_odp *umem_odp;
492 umem_odp = ib_umem_odp_alloc_implicit(&dev->ib_dev, access_flags);
493 if (IS_ERR(umem_odp))
494 return ERR_CAST(umem_odp);
499 ib_umem_odp_release(umem_odp);
506 imr->umem = &umem_odp->umem;
622 struct ib_umem_odp *umem_odp;
642 umem_odp = to_ib_umem_odp(mtt->umem);
643 len = min_t(u64, user_va + bcnt, ib_umem_end(umem_odp)) -
646 ret = pagefault_real_mr(mtt, umem_odp, user_va, len,