Lines Matching refs:base
58 return container_of(bo, struct vmw_buffer_object, base);
95 struct ttm_buffer_object *bo = &buf->base;
144 struct ttm_buffer_object *bo = &buf->base;
219 struct ttm_buffer_object *bo = &buf->base;
288 struct ttm_buffer_object *bo = &buf->base;
340 struct ttm_buffer_object *bo = &vbo->base;
344 dma_resv_assert_held(bo->base.resv);
391 struct ttm_buffer_object *bo = &vbo->base;
520 vmw_bo->base.priority = 3;
523 ret = ttm_bo_init(bdev, &vmw_bo->base, size,
532 * vmw_user_bo_release - TTM reference base object release callback for
535 * @p_base: The TTM base object pointer about to be unreferenced.
537 * Clears the TTM base object pointer and drops the reference the
538 * base object has on the underlying struct vmw_buffer_object.
543 struct ttm_base_object *base = *p_base;
547 if (unlikely(base == NULL))
550 vmw_user_bo = container_of(base, struct vmw_user_buffer_object,
551 prime.base);
552 ttm_bo_put(&vmw_user_bo->vbo.base);
560 * @base: Pointer to the TTM base object
566 static void vmw_user_bo_ref_obj_release(struct ttm_base_object *base,
571 user_bo = container_of(base, struct vmw_user_buffer_object, prime.base);
621 ttm_bo_get(&user_bo->vbo.base);
630 ttm_bo_put(&user_bo->vbo.base);
636 *p_base = &user_bo->prime.base;
639 *handle = user_bo->prime.base.handle;
664 if (likely(ttm_ref_object_exists(tfile, &vmw_user_bo->prime.base)))
692 struct ttm_buffer_object *bo = &user_bo->vbo.base;
700 (bo->base.resv, true, true,
721 ret = ttm_ref_object_add(tfile, &user_bo->prime.base,
852 rep->map_handle = drm_vma_node_offset_addr(&vbo->base.base.vma_node);
874 * handle to a TTM base object, optionally freeing the object.
895 * @p_base: Pointer to where a pointer to the TTM base object should be
899 * Both the output base object pointer and the vmw buffer object pointer
907 struct ttm_base_object *base;
909 base = ttm_base_object_lookup(tfile, handle);
910 if (unlikely(base == NULL)) {
916 if (unlikely(ttm_base_object_type(base) != ttm_buffer_type)) {
917 ttm_base_object_unref(&base);
923 vmw_user_bo = container_of(base, struct vmw_user_buffer_object,
924 prime.base);
925 ttm_bo_get(&vmw_user_bo->vbo.base);
927 *p_base = base;
929 ttm_base_object_unref(&base);
956 struct ttm_base_object *base;
958 base = ttm_base_object_noref_lookup(tfile, handle);
959 if (!base) {
965 if (unlikely(ttm_base_object_type(base) != ttm_buffer_type)) {
972 vmw_user_bo = container_of(base, struct vmw_user_buffer_object,
973 prime.base);
991 if (vbo->base.destroy != vmw_user_bo_destroy)
996 *handle = user_bo->prime.base.handle;
997 return ttm_ref_object_add(tfile, &user_bo->prime.base,
1024 dma_resv_add_excl_fence(bo->base.resv, &fence->base);
1025 dma_fence_put(&fence->base);
1027 dma_resv_add_excl_fence(bo->base.resv, &fence->base);
1094 *offset = drm_vma_node_offset_addr(&out_buf->base.base.vma_node);
1159 vbo = container_of(bo, struct vmw_buffer_object, base);