Lines Matching refs:entry
296 struct gvt_dma *entry)
298 rb_erase(&entry->gfn_node, &vgpu->gfn_cache);
299 rb_erase(&entry->dma_addr_node, &vgpu->dma_addr_cache);
300 kfree(entry);
616 struct gvt_dma *entry;
622 entry = __gvt_cache_find_gfn(vgpu, iov_pfn);
623 if (!entry)
626 gvt_dma_unmap_page(vgpu, entry->gfn, entry->dma_addr,
627 entry->size);
628 __gvt_cache_remove_entry(vgpu, entry);
845 /* Only allow MMIO GGTT entry access */
867 /* Only support GGTT entry 8 bytes read */
942 /* Only support GGTT entry 8 bytes write */
1632 struct gvt_dma *entry;
1640 entry = __gvt_cache_find_gfn(vgpu, gfn);
1641 if (!entry) {
1649 } else if (entry->size != size) {
1651 gvt_dma_unmap_page(vgpu, gfn, entry->dma_addr, entry->size);
1652 __gvt_cache_remove_entry(vgpu, entry);
1662 kref_get(&entry->ref);
1663 *dma_addr = entry->dma_addr;
1678 struct gvt_dma *entry;
1685 entry = __gvt_cache_find_dma_addr(vgpu, dma_addr);
1686 if (entry)
1687 kref_get(&entry->ref);
1697 struct gvt_dma *entry = container_of(ref, typeof(*entry), ref);
1699 gvt_dma_unmap_page(entry->vgpu, entry->gfn, entry->dma_addr,
1700 entry->size);
1701 __gvt_cache_remove_entry(entry->vgpu, entry);
1707 struct gvt_dma *entry;
1713 entry = __gvt_cache_find_dma_addr(vgpu, dma_addr);
1714 if (entry)
1715 kref_put(&entry->ref, __gvt_dma_release);