Lines Matching refs:entry
341 struct gvt_dma *entry)
345 rb_erase(&entry->gfn_node, &vdev->gfn_cache);
346 rb_erase(&entry->dma_addr_node, &vdev->dma_addr_cache);
347 kfree(entry);
752 struct gvt_dma *entry;
760 entry = __gvt_cache_find_gfn(vgpu, iov_pfn);
761 if (!entry)
764 gvt_dma_unmap_page(vgpu, entry->gfn, entry->dma_addr,
765 entry->size);
766 __gvt_cache_remove_entry(vgpu, entry);
1058 /* Only allow MMIO GGTT entry access */
1079 /* Only support GGTT entry 8 bytes read */
1153 /* Only support GGTT entry 8 bytes write */
1954 struct gvt_dma *entry;
1965 entry = __gvt_cache_find_gfn(vgpu, gfn);
1966 if (!entry) {
1974 } else if (entry->size != size) {
1976 gvt_dma_unmap_page(vgpu, gfn, entry->dma_addr, entry->size);
1977 __gvt_cache_remove_entry(vgpu, entry);
1987 kref_get(&entry->ref);
1988 *dma_addr = entry->dma_addr;
2005 struct gvt_dma *entry;
2015 entry = __gvt_cache_find_dma_addr(info->vgpu, dma_addr);
2016 if (entry)
2017 kref_get(&entry->ref);
2027 struct gvt_dma *entry = container_of(ref, typeof(*entry), ref);
2029 gvt_dma_unmap_page(entry->vgpu, entry->gfn, entry->dma_addr,
2030 entry->size);
2031 __gvt_cache_remove_entry(entry->vgpu, entry);
2038 struct gvt_dma *entry;
2047 entry = __gvt_cache_find_dma_addr(vgpu, dma_addr);
2048 if (entry)
2049 kref_put(&entry->ref, __gvt_dma_release);