Lines Matching refs:ggtt

64 insert_mappable_node(struct i915_ggtt *ggtt, struct drm_mm_node *node, u32 size)
68 err = mutex_lock_interruptible(&ggtt->vm.mutex);
73 err = drm_mm_insert_node_in_range(&ggtt->vm.mm, node,
75 0, ggtt->mappable_end,
78 mutex_unlock(&ggtt->vm.mutex);
84 remove_mappable_node(struct i915_ggtt *ggtt, struct drm_mm_node *node)
86 mutex_lock(&ggtt->vm.mutex);
88 mutex_unlock(&ggtt->vm.mutex);
96 struct i915_ggtt *ggtt = to_gt(i915)->ggtt;
101 if (mutex_lock_interruptible(&ggtt->vm.mutex))
104 pinned = ggtt->vm.reserved;
105 list_for_each_entry(vma, &ggtt->vm.bound_list, vm_link)
109 mutex_unlock(&ggtt->vm.mutex);
111 args->aper_size = ggtt->vm.total;
309 struct i915_ggtt *ggtt = to_gt(i915)->ggtt;
337 ret = insert_mappable_node(ggtt, node, PAGE_SIZE);
347 ggtt->vm.clear_range(&ggtt->vm, node->start, node->size);
348 remove_mappable_node(ggtt, node);
370 struct i915_ggtt *ggtt = to_gt(i915)->ggtt;
374 ggtt->vm.clear_range(&ggtt->vm, node->start, node->size);
375 remove_mappable_node(ggtt, node);
386 struct i915_ggtt *ggtt = to_gt(i915)->ggtt;
422 ggtt->vm.insert_page(&ggtt->vm,
432 if (gtt_user_read(&ggtt->iomap, page_base, page_offset,
549 struct i915_ggtt *ggtt = to_gt(i915)->ggtt;
602 intel_gt_flush_ggtt_writes(ggtt->vm.gt);
603 ggtt->vm.insert_page(&ggtt->vm,
619 if (ggtt_write(&ggtt->iomap, page_base, page_offset,
630 intel_gt_flush_ggtt_writes(ggtt->vm.gt);
858 &to_gt(i915)->ggtt->userfault_list, userfault_link)
870 for (i = 0; i < to_gt(i915)->ggtt->num_fences; i++) {
871 struct i915_fence_reg *reg = &to_gt(i915)->ggtt->fence_regs[i];
912 struct i915_ggtt *ggtt = to_gt(i915)->ggtt;
928 if (obj->base.size > ggtt->mappable_end)
948 obj->base.size > ggtt->mappable_end / 2)
953 vma = i915_vma_instance(obj, &ggtt->vm, view);
973 (vma->fence_size > ggtt->mappable_end / 2 ||
994 mutex_lock(&ggtt->vm.mutex);
996 mutex_unlock(&ggtt->vm.mutex);
1244 i915_ggtt_resume(to_gt(dev_priv)->ggtt);