Lines Matching defs:device
57 * device virtual memory. Host mapping is conditional and is controlled by
60 * \param[in] device Logical device pointer.
61 * \param[in] heap Heap to allocate device virtual address from.
72 VkResult pvr_bo_alloc(struct pvr_device *device,
84 pvr_bo = vk_alloc(&device->vk.alloc,
89 return vk_error(device, VK_ERROR_OUT_OF_HOST_MEMORY);
91 result = device->ws->ops->buffer_create(device->ws,
101 void *map = device->ws->ops->buffer_map(pvr_bo->bo);
108 pvr_bo->vma = device->ws->ops->heap_alloc(heap, size, alignment);
114 addr = device->ws->ops->vma_map(pvr_bo->vma, pvr_bo->bo, 0, size);
125 device->ws->ops->heap_free(pvr_bo->vma);
129 device->ws->ops->buffer_unmap(pvr_bo->bo);
132 device->ws->ops->buffer_destroy(pvr_bo->bo);
135 vk_free(&device->vk.alloc, pvr_bo);
147 * \param[in] device Logical device pointer.
153 void *pvr_bo_cpu_map(struct pvr_device *device, struct pvr_bo *pvr_bo)
157 return device->ws->ops->buffer_map(pvr_bo->bo);
169 * \param[in] device Logical device pointer.
172 void pvr_bo_cpu_unmap(struct pvr_device *device, struct pvr_bo *pvr_bo)
175 device->ws->ops->buffer_unmap(pvr_bo->bo);
181 * \param[in] device Logical device pointer.
186 void pvr_bo_free(struct pvr_device *device, struct pvr_bo *pvr_bo)
191 device->ws->ops->vma_unmap(pvr_bo->vma);
192 device->ws->ops->heap_free(pvr_bo->vma);
195 device->ws->ops->buffer_unmap(pvr_bo->bo);
197 device->ws->ops->buffer_destroy(pvr_bo->bo);
199 vk_free(&device->vk.alloc, pvr_bo);