drm/vmwgfx: switch driver from bo->resv to bo->base.resv
Signed-off-by: Gerd Hoffmann <kraxel@redhat.com> Acked-by: Thomas Hellstrom <thellstrom@vmware.com> Link: http://patchwork.freedesktop.org/patch/msgid/20190805140119.7337-13-kraxel@redhat.com
This commit is contained in:
Родитель
336ac942f1
Коммит
4922f55294
|
@ -459,9 +459,9 @@ int vmw_bo_cpu_blit(struct ttm_buffer_object *dst,
|
|||
|
||||
/* Buffer objects need to be either pinned or reserved: */
|
||||
if (!(dst->mem.placement & TTM_PL_FLAG_NO_EVICT))
|
||||
reservation_object_assert_held(dst->resv);
|
||||
reservation_object_assert_held(dst->base.resv);
|
||||
if (!(src->mem.placement & TTM_PL_FLAG_NO_EVICT))
|
||||
reservation_object_assert_held(src->resv);
|
||||
reservation_object_assert_held(src->base.resv);
|
||||
|
||||
if (dst->ttm->state == tt_unpopulated) {
|
||||
ret = dst->ttm->bdev->driver->ttm_tt_populate(dst->ttm, &ctx);
|
||||
|
|
|
@ -342,7 +342,7 @@ void vmw_bo_pin_reserved(struct vmw_buffer_object *vbo, bool pin)
|
|||
uint32_t old_mem_type = bo->mem.mem_type;
|
||||
int ret;
|
||||
|
||||
reservation_object_assert_held(bo->resv);
|
||||
reservation_object_assert_held(bo->base.resv);
|
||||
|
||||
if (pin) {
|
||||
if (vbo->pin_count++ > 0)
|
||||
|
@ -690,7 +690,7 @@ static int vmw_user_bo_synccpu_grab(struct vmw_user_buffer_object *user_bo,
|
|||
long lret;
|
||||
|
||||
lret = reservation_object_wait_timeout_rcu
|
||||
(bo->resv, true, true,
|
||||
(bo->base.resv, true, true,
|
||||
nonblock ? 0 : MAX_SCHEDULE_TIMEOUT);
|
||||
if (!lret)
|
||||
return -EBUSY;
|
||||
|
@ -1007,10 +1007,10 @@ void vmw_bo_fence_single(struct ttm_buffer_object *bo,
|
|||
|
||||
if (fence == NULL) {
|
||||
vmw_execbuf_fence_commands(NULL, dev_priv, &fence, NULL);
|
||||
reservation_object_add_excl_fence(bo->resv, &fence->base);
|
||||
reservation_object_add_excl_fence(bo->base.resv, &fence->base);
|
||||
dma_fence_put(&fence->base);
|
||||
} else
|
||||
reservation_object_add_excl_fence(bo->resv, &fence->base);
|
||||
reservation_object_add_excl_fence(bo->base.resv, &fence->base);
|
||||
}
|
||||
|
||||
|
||||
|
|
|
@ -169,7 +169,7 @@ static int vmw_cotable_unscrub(struct vmw_resource *res)
|
|||
} *cmd;
|
||||
|
||||
WARN_ON_ONCE(bo->mem.mem_type != VMW_PL_MOB);
|
||||
reservation_object_assert_held(bo->resv);
|
||||
reservation_object_assert_held(bo->base.resv);
|
||||
|
||||
cmd = VMW_FIFO_RESERVE(dev_priv, sizeof(*cmd));
|
||||
if (!cmd)
|
||||
|
@ -311,7 +311,7 @@ static int vmw_cotable_unbind(struct vmw_resource *res,
|
|||
return 0;
|
||||
|
||||
WARN_ON_ONCE(bo->mem.mem_type != VMW_PL_MOB);
|
||||
reservation_object_assert_held(bo->resv);
|
||||
reservation_object_assert_held(bo->base.resv);
|
||||
|
||||
mutex_lock(&dev_priv->binding_mutex);
|
||||
if (!vcotbl->scrubbed)
|
||||
|
|
|
@ -402,14 +402,14 @@ void vmw_resource_unreserve(struct vmw_resource *res,
|
|||
|
||||
if (switch_backup && new_backup != res->backup) {
|
||||
if (res->backup) {
|
||||
reservation_object_assert_held(res->backup->base.resv);
|
||||
reservation_object_assert_held(res->backup->base.base.resv);
|
||||
list_del_init(&res->mob_head);
|
||||
vmw_bo_unreference(&res->backup);
|
||||
}
|
||||
|
||||
if (new_backup) {
|
||||
res->backup = vmw_bo_reference(new_backup);
|
||||
reservation_object_assert_held(new_backup->base.resv);
|
||||
reservation_object_assert_held(new_backup->base.base.resv);
|
||||
list_add_tail(&res->mob_head, &new_backup->res_list);
|
||||
} else {
|
||||
res->backup = NULL;
|
||||
|
@ -691,7 +691,7 @@ void vmw_resource_unbind_list(struct vmw_buffer_object *vbo)
|
|||
.num_shared = 0
|
||||
};
|
||||
|
||||
reservation_object_assert_held(vbo->base.resv);
|
||||
reservation_object_assert_held(vbo->base.base.resv);
|
||||
list_for_each_entry_safe(res, next, &vbo->res_list, mob_head) {
|
||||
if (!res->func->unbind)
|
||||
continue;
|
||||
|
|
Загрузка…
Ссылка в новой задаче