drm/virtio: move virtio_gpu_object_{attach, detach} calls.
Drop the dummy ttm backend implementation, add a real one for TTM_PL_FLAG_TT objects. The bin/unbind callbacks will call virtio_gpu_object_{attach,detach}, to update the object state on the host side, instead of invoking those calls from the move_notify() callback. With that in place the move and move_notify callbacks are not needed any more, so drop them. Signed-off-by: Gerd Hoffmann <kraxel@redhat.com> Acked-by: Noralf Trønnes <noralf@tronnes.org> Link: http://patchwork.freedesktop.org/patch/msgid/20190318113332.10900-2-kraxel@redhat.com
This commit is contained in:
Родитель
6bbeae35b6
Коммит
42ca472603
|
@ -190,42 +190,45 @@ static void virtio_gpu_ttm_io_mem_free(struct ttm_bo_device *bdev,
|
|||
*/
|
||||
struct virtio_gpu_ttm_tt {
|
||||
struct ttm_dma_tt ttm;
|
||||
struct virtio_gpu_device *vgdev;
|
||||
u64 offset;
|
||||
struct virtio_gpu_object *obj;
|
||||
};
|
||||
|
||||
static int virtio_gpu_ttm_backend_bind(struct ttm_tt *ttm,
|
||||
struct ttm_mem_reg *bo_mem)
|
||||
static int virtio_gpu_ttm_tt_bind(struct ttm_tt *ttm,
|
||||
struct ttm_mem_reg *bo_mem)
|
||||
{
|
||||
struct virtio_gpu_ttm_tt *gtt = (void *)ttm;
|
||||
struct virtio_gpu_ttm_tt *gtt =
|
||||
container_of(ttm, struct virtio_gpu_ttm_tt, ttm.ttm);
|
||||
struct virtio_gpu_device *vgdev =
|
||||
virtio_gpu_get_vgdev(gtt->obj->tbo.bdev);
|
||||
|
||||
gtt->offset = (unsigned long)(bo_mem->start << PAGE_SHIFT);
|
||||
if (!ttm->num_pages)
|
||||
WARN(1, "nothing to bind %lu pages for mreg %p back %p!\n",
|
||||
ttm->num_pages, bo_mem, ttm);
|
||||
|
||||
/* Not implemented */
|
||||
virtio_gpu_object_attach(vgdev, gtt->obj, NULL);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int virtio_gpu_ttm_backend_unbind(struct ttm_tt *ttm)
|
||||
static int virtio_gpu_ttm_tt_unbind(struct ttm_tt *ttm)
|
||||
{
|
||||
/* Not implemented */
|
||||
struct virtio_gpu_ttm_tt *gtt =
|
||||
container_of(ttm, struct virtio_gpu_ttm_tt, ttm.ttm);
|
||||
struct virtio_gpu_device *vgdev =
|
||||
virtio_gpu_get_vgdev(gtt->obj->tbo.bdev);
|
||||
|
||||
virtio_gpu_object_detach(vgdev, gtt->obj);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void virtio_gpu_ttm_backend_destroy(struct ttm_tt *ttm)
|
||||
static void virtio_gpu_ttm_tt_destroy(struct ttm_tt *ttm)
|
||||
{
|
||||
struct virtio_gpu_ttm_tt *gtt = (void *)ttm;
|
||||
struct virtio_gpu_ttm_tt *gtt =
|
||||
container_of(ttm, struct virtio_gpu_ttm_tt, ttm.ttm);
|
||||
|
||||
ttm_dma_tt_fini(>t->ttm);
|
||||
kfree(gtt);
|
||||
}
|
||||
|
||||
static struct ttm_backend_func virtio_gpu_backend_func = {
|
||||
.bind = &virtio_gpu_ttm_backend_bind,
|
||||
.unbind = &virtio_gpu_ttm_backend_unbind,
|
||||
.destroy = &virtio_gpu_ttm_backend_destroy,
|
||||
static struct ttm_backend_func virtio_gpu_tt_func = {
|
||||
.bind = &virtio_gpu_ttm_tt_bind,
|
||||
.unbind = &virtio_gpu_ttm_tt_unbind,
|
||||
.destroy = &virtio_gpu_ttm_tt_destroy,
|
||||
};
|
||||
|
||||
static struct ttm_tt *virtio_gpu_ttm_tt_create(struct ttm_buffer_object *bo,
|
||||
|
@ -238,8 +241,8 @@ static struct ttm_tt *virtio_gpu_ttm_tt_create(struct ttm_buffer_object *bo,
|
|||
gtt = kzalloc(sizeof(struct virtio_gpu_ttm_tt), GFP_KERNEL);
|
||||
if (gtt == NULL)
|
||||
return NULL;
|
||||
gtt->ttm.ttm.func = &virtio_gpu_backend_func;
|
||||
gtt->vgdev = vgdev;
|
||||
gtt->ttm.ttm.func = &virtio_gpu_tt_func;
|
||||
gtt->obj = container_of(bo, struct virtio_gpu_object, tbo);
|
||||
if (ttm_dma_tt_init(>t->ttm, bo, page_flags)) {
|
||||
kfree(gtt);
|
||||
return NULL;
|
||||
|
@ -247,51 +250,6 @@ static struct ttm_tt *virtio_gpu_ttm_tt_create(struct ttm_buffer_object *bo,
|
|||
return >t->ttm.ttm;
|
||||
}
|
||||
|
||||
static void virtio_gpu_move_null(struct ttm_buffer_object *bo,
|
||||
struct ttm_mem_reg *new_mem)
|
||||
{
|
||||
struct ttm_mem_reg *old_mem = &bo->mem;
|
||||
|
||||
BUG_ON(old_mem->mm_node != NULL);
|
||||
*old_mem = *new_mem;
|
||||
new_mem->mm_node = NULL;
|
||||
}
|
||||
|
||||
static int virtio_gpu_bo_move(struct ttm_buffer_object *bo, bool evict,
|
||||
struct ttm_operation_ctx *ctx,
|
||||
struct ttm_mem_reg *new_mem)
|
||||
{
|
||||
int ret;
|
||||
|
||||
ret = ttm_bo_wait(bo, ctx->interruptible, ctx->no_wait_gpu);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
virtio_gpu_move_null(bo, new_mem);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void virtio_gpu_bo_move_notify(struct ttm_buffer_object *tbo,
|
||||
bool evict,
|
||||
struct ttm_mem_reg *new_mem)
|
||||
{
|
||||
struct virtio_gpu_object *bo;
|
||||
struct virtio_gpu_device *vgdev;
|
||||
|
||||
bo = container_of(tbo, struct virtio_gpu_object, tbo);
|
||||
vgdev = (struct virtio_gpu_device *)bo->gem_base.dev->dev_private;
|
||||
|
||||
if (!new_mem || (new_mem->placement & TTM_PL_FLAG_SYSTEM)) {
|
||||
if (bo->hw_res_handle)
|
||||
virtio_gpu_object_detach(vgdev, bo);
|
||||
|
||||
} else if (new_mem->placement & TTM_PL_FLAG_TT) {
|
||||
if (bo->hw_res_handle) {
|
||||
virtio_gpu_object_attach(vgdev, bo, NULL);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
static void virtio_gpu_bo_swap_notify(struct ttm_buffer_object *tbo)
|
||||
{
|
||||
struct virtio_gpu_object *bo;
|
||||
|
@ -308,11 +266,9 @@ static struct ttm_bo_driver virtio_gpu_bo_driver = {
|
|||
.init_mem_type = &virtio_gpu_init_mem_type,
|
||||
.eviction_valuable = ttm_bo_eviction_valuable,
|
||||
.evict_flags = &virtio_gpu_evict_flags,
|
||||
.move = &virtio_gpu_bo_move,
|
||||
.verify_access = &virtio_gpu_verify_access,
|
||||
.io_mem_reserve = &virtio_gpu_ttm_io_mem_reserve,
|
||||
.io_mem_free = &virtio_gpu_ttm_io_mem_free,
|
||||
.move_notify = &virtio_gpu_bo_move_notify,
|
||||
.swap_notify = &virtio_gpu_bo_swap_notify,
|
||||
};
|
||||
|
||||
|
|
Загрузка…
Ссылка в новой задаче