Lines Matching refs:buffer

1865     * render targets and one slot for the descriptor buffer. */
2278 * byte offset of the element in the buffer. The base address must
2289 * arbitrary byte alignment for linear and buffer surfaces. It's
2990 /* Finding a buffer for batch decoding */
3056 intel_aux_map_buffer_free(void *driver_ctx, struct intel_buffer *buffer)
3058 struct intel_aux_map_buffer *buf = (struct intel_aux_map_buffer*)buffer;
3897 /* Check if we need to support Android HW buffer export. If so,
3936 * from passing a buffer to a trusted client, lying about the size, and
3940 * this sort of attack but only if it can trust the buffer size.
4311 ANV_FROM_HANDLE(anv_buffer, buffer, pBindInfo->buffer);
4317 assert(mem->bo->size - pBindInfo->memoryOffset >= buffer->vk.size);
4318 buffer->address = (struct anv_address) {
4323 buffer->address = ANV_NULL_ADDRESS;
4489 ANV_FROM_HANDLE(anv_buffer, buffer, pInfo->buffer);
4492 buffer->vk.size,
4493 buffer->vk.usage,
4517 struct anv_buffer *buffer;
4520 * issue here is that we may align up the buffer size and we don't want
4522 * allocating a buffer larger than our GTT size.
4527 buffer = vk_buffer_create(&device->vk, pCreateInfo,
4528 pAllocator, sizeof(*buffer));
4529 if (buffer == NULL)
4532 buffer->address = ANV_NULL_ADDRESS;
4534 *pBuffer = anv_buffer_to_handle(buffer);
4545 ANV_FROM_HANDLE(anv_buffer, buffer, _buffer);
4547 if (!buffer)
4550 vk_buffer_destroy(&device->vk, pAllocator, &buffer->vk);
4557 ANV_FROM_HANDLE(anv_buffer, buffer, pInfo->buffer);
4559 assert(!anv_address_is_null(buffer->address));
4560 assert(anv_bo_is_pinned(buffer->address.bo));
4562 return anv_address_physical(buffer->address);