Lines Matching defs:address

363        * address support can still fail.  Just clamp the address space size to
874 * address is required for the vertex cache flush workaround.
938 * because it's just a matter of setting the sampler address in the sample
2154 * address. This means that we can have at must 1M surface states
2276 * specifies the base address of the first element of the surface,
2277 * computed in software by adding the surface base address to the
2278 * byte offset of the element in the buffer. The base address must
2974 uint64_t address)
2978 if (address >= bo_address && address < (bo_address + bo->size)) {
2992 decode_get_bo(void *v_batch, bool ppgtt, uint64_t address)
2999 if (get_bo_from_pool(&ret_bo, &device->dynamic_state_pool.block_pool, address))
3001 if (get_bo_from_pool(&ret_bo, &device->instruction_state_pool.block_pool, address))
3003 if (get_bo_from_pool(&ret_bo, &device->binding_table_pool.block_pool, address))
3005 if (get_bo_from_pool(&ret_bo, &device->surface_state_pool.block_pool, address))
3017 if (address >= bo_address && address < bo_address + (*bo)->bo->size) {
3254 /* keep the page with address zero out of the allocator */
3262 * base address + size can overflow 48 bits. For more information see
3328 * the base address 0 and start the pool memory at an offset. This way we
3348 * right off the bat, so they are close to the base address.
3370 * table its own base address separately from surface state base.
3753 uint64_t address, uint64_t size)
3755 const uint64_t addr_48b = intel_48b_address(address);
4318 buffer->address = (struct anv_address) {
4323 buffer->address = ANV_NULL_ADDRESS;
4519 /* Don't allow creating buffers bigger than our address space. The real
4532 buffer->address = ANV_NULL_ADDRESS;
4559 assert(!anv_address_is_null(buffer->address));
4560 assert(anv_bo_is_pinned(buffer->address.bo));
4562 return anv_address_physical(buffer->address);
4589 struct anv_address address,
4593 .address = anv_address_physical(address),
4595 address.bo && address.bo->is_external),