Lines Matching refs:base

51 	struct ttm_base_object base;
82 return container_of(fence->base.lock, struct vmw_fence_manager, lock);
126 container_of(f, struct vmw_fence_obj, base);
150 container_of(f, struct vmw_fence_obj, base);
156 if (seqno - fence->base.seqno < VMW_FENCE_WRAP)
163 struct dma_fence_cb base;
171 container_of(cb, struct vmwgfx_wait_cb, base);
181 container_of(f, struct vmw_fence_obj, base);
203 cb.base.func = vmwgfx_wait_cb;
205 list_add(&cb.base.node, &f->cb_list);
241 if (!list_empty(&cb.base.node))
242 list_del(&cb.base.node);
347 dma_fence_init(&fence->base, &vmw_fence_ops, &fman->lock,
420 fence->base.seqno);
449 if (dma_fence_is_signaled_locked(&fence->base))
454 goal_seqno - fence->base.seqno < VMW_FENCE_WRAP))
457 vmw_fence_goal_write(fman->dev_priv, fence->base.seqno);
473 if (seqno - fence->base.seqno < VMW_FENCE_WRAP) {
475 dma_fence_signal_locked(&fence->base);
514 if (test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &fence->base.flags))
519 return dma_fence_is_signaled(&fence->base);
525 long ret = dma_fence_wait_timeout(&fence->base, interruptible, timeout);
537 dma_fence_free(&fence->base);
570 ttm_base_object_kfree(ufence, base);
575 struct ttm_base_object *base = *p_base;
577 container_of(base, struct vmw_user_fence, base);
609 * The base object holds a reference which is freed in
614 ret = ttm_base_object_init(tfile, &ufence->base, false,
621 * Free the base object's reference
628 *p_handle = ufence->base.handle;
658 dma_fence_get(&fence->base);
666 dma_fence_signal(&fence->base);
674 dma_fence_put(&fence->base);
693 * @return: A struct vmw_user_fence base ttm object on success or
704 struct ttm_base_object *base = ttm_base_object_lookup(tfile, handle);
706 if (!base) {
712 if (base->refcount_release != vmw_user_fence_base_release) {
715 ttm_base_object_unref(&base);
719 return base;
729 struct ttm_base_object *base;
748 base = vmw_fence_obj_lookup(tfile, arg->handle);
749 if (IS_ERR(base))
750 return PTR_ERR(base);
752 fence = &(container_of(base, struct vmw_user_fence, base)->fence);
766 ttm_base_object_unref(&base);
782 struct ttm_base_object *base;
788 base = vmw_fence_obj_lookup(tfile, arg->handle);
789 if (IS_ERR(base))
790 return PTR_ERR(base);
792 fence = &(container_of(base, struct vmw_user_fence, base)->fence);
802 ttm_base_object_unref(&base);
892 if (dma_fence_is_signaled_locked(&fence->base)) {
971 struct drm_pending_event base;
993 event->event.base.type = DRM_VMW_EVENT_FENCE_SIGNALED;
994 event->event.base.length = sizeof(*event);
997 ret = drm_event_reserve_init(dev, file_priv, &event->base, &event->event.base);
1007 &event->base,
1013 &event->base,
1023 drm_event_cancel_free(dev, &event->base);
1049 struct ttm_base_object *base =
1052 if (IS_ERR(base))
1053 return PTR_ERR(base);
1055 fence = &(container_of(base, struct vmw_user_fence,
1056 base)->fence);
1060 ret = ttm_ref_object_add(vmw_fp->tfile, base,
1067 handle = base->handle;
1069 ttm_base_object_unref(&base);