vmwgfx: Drop 3D Legacy Display Unit support
Since 3D requires HWv8 and screen objects is always available on those hosts we only need the screen objects path for surfaces. Signed-off-by: Jakob Bornecrantz <jakob@vmware.com> Signed-off-by: Thomas Hellstrom <thellstrom@vmware.com> Signed-off-by: Dave Airlie <airlied@redhat.com>
This commit is contained in:
Родитель
b7b70024d9
Коммит
01e81419ce
|
@ -457,9 +457,6 @@ static int vmw_driver_load(struct drm_device *dev, unsigned long chipset)
|
|||
if (unlikely(ret != 0))
|
||||
goto out_no_fifo;
|
||||
vmw_kms_save_vga(dev_priv);
|
||||
DRM_INFO("%s", vmw_fifo_have_3d(dev_priv) ?
|
||||
"Detected device 3D availability.\n" :
|
||||
"Detected no device 3D availability.\n");
|
||||
|
||||
/* Start kms and overlay systems, needs fifo. */
|
||||
ret = vmw_kms_init(dev_priv);
|
||||
|
@ -467,6 +464,11 @@ static int vmw_driver_load(struct drm_device *dev, unsigned long chipset)
|
|||
goto out_no_kms;
|
||||
vmw_overlay_init(dev_priv);
|
||||
|
||||
/* 3D Depends on Screen Objects being used. */
|
||||
DRM_INFO("%s", vmw_fifo_have_3d(dev_priv) ?
|
||||
"Detected device 3D availability.\n" :
|
||||
"Detected no device 3D availability.\n");
|
||||
|
||||
/* We might be done with the fifo now */
|
||||
if (dev_priv->enable_fb) {
|
||||
vmw_fb_init(dev_priv);
|
||||
|
@ -779,8 +781,6 @@ static void vmw_master_drop(struct drm_device *dev,
|
|||
|
||||
vmw_fp->locked_master = drm_master_get(file_priv->master);
|
||||
ret = ttm_vt_lock(&vmaster->lock, false, vmw_fp->tfile);
|
||||
vmw_kms_idle_workqueues(vmaster);
|
||||
|
||||
if (unlikely((ret != 0))) {
|
||||
DRM_ERROR("Unable to lock TTM at VT switch.\n");
|
||||
drm_master_put(&vmw_fp->locked_master);
|
||||
|
|
|
@ -48,6 +48,10 @@ bool vmw_fifo_have_3d(struct vmw_private *dev_priv)
|
|||
if (hwversion < SVGA3D_HWVERSION_WS8_B1)
|
||||
return false;
|
||||
|
||||
/* Non-Screen Object path does not support surfaces */
|
||||
if (!dev_priv->sou_priv)
|
||||
return false;
|
||||
|
||||
return true;
|
||||
}
|
||||
|
||||
|
|
|
@ -31,9 +31,6 @@
|
|||
/* Might need a hrtimer here? */
|
||||
#define VMWGFX_PRESENT_RATE ((HZ / 60 > 0) ? HZ / 60 : 1)
|
||||
|
||||
static int vmw_surface_dmabuf_pin(struct vmw_framebuffer *vfb);
|
||||
static int vmw_surface_dmabuf_unpin(struct vmw_framebuffer *vfb);
|
||||
|
||||
void vmw_display_unit_cleanup(struct vmw_display_unit *du)
|
||||
{
|
||||
if (du->cursor_surface)
|
||||
|
@ -330,41 +327,10 @@ struct vmw_framebuffer_surface {
|
|||
struct vmw_framebuffer base;
|
||||
struct vmw_surface *surface;
|
||||
struct vmw_dma_buffer *buffer;
|
||||
struct delayed_work d_work;
|
||||
struct mutex work_lock;
|
||||
bool present_fs;
|
||||
struct list_head head;
|
||||
struct drm_master *master;
|
||||
};
|
||||
|
||||
/**
|
||||
* vmw_kms_idle_workqueues - Flush workqueues on this master
|
||||
*
|
||||
* @vmaster - Pointer identifying the master, for the surfaces of which
|
||||
* we idle the dirty work queues.
|
||||
*
|
||||
* This function should be called with the ttm lock held in exclusive mode
|
||||
* to idle all dirty work queues before the fifo is taken down.
|
||||
*
|
||||
* The work task may actually requeue itself, but after the flush returns we're
|
||||
* sure that there's nothing to present, since the ttm lock is held in
|
||||
* exclusive mode, so the fifo will never get used.
|
||||
*/
|
||||
|
||||
void vmw_kms_idle_workqueues(struct vmw_master *vmaster)
|
||||
{
|
||||
struct vmw_framebuffer_surface *entry;
|
||||
|
||||
mutex_lock(&vmaster->fb_surf_mutex);
|
||||
list_for_each_entry(entry, &vmaster->fb_surf, head) {
|
||||
if (cancel_delayed_work_sync(&entry->d_work))
|
||||
(void) entry->d_work.work.func(&entry->d_work.work);
|
||||
|
||||
(void) cancel_delayed_work_sync(&entry->d_work);
|
||||
}
|
||||
mutex_unlock(&vmaster->fb_surf_mutex);
|
||||
}
|
||||
|
||||
void vmw_framebuffer_surface_destroy(struct drm_framebuffer *framebuffer)
|
||||
{
|
||||
struct vmw_framebuffer_surface *vfbs =
|
||||
|
@ -376,7 +342,6 @@ void vmw_framebuffer_surface_destroy(struct drm_framebuffer *framebuffer)
|
|||
list_del(&vfbs->head);
|
||||
mutex_unlock(&vmaster->fb_surf_mutex);
|
||||
|
||||
cancel_delayed_work_sync(&vfbs->d_work);
|
||||
drm_master_put(&vfbs->master);
|
||||
drm_framebuffer_cleanup(framebuffer);
|
||||
vmw_surface_unreference(&vfbs->surface);
|
||||
|
@ -384,101 +349,6 @@ void vmw_framebuffer_surface_destroy(struct drm_framebuffer *framebuffer)
|
|||
kfree(vfbs);
|
||||
}
|
||||
|
||||
static void vmw_framebuffer_present_fs_callback(struct work_struct *work)
|
||||
{
|
||||
struct delayed_work *d_work =
|
||||
container_of(work, struct delayed_work, work);
|
||||
struct vmw_framebuffer_surface *vfbs =
|
||||
container_of(d_work, struct vmw_framebuffer_surface, d_work);
|
||||
struct vmw_surface *surf = vfbs->surface;
|
||||
struct drm_framebuffer *framebuffer = &vfbs->base.base;
|
||||
struct vmw_private *dev_priv = vmw_priv(framebuffer->dev);
|
||||
|
||||
struct {
|
||||
SVGA3dCmdHeader header;
|
||||
SVGA3dCmdPresent body;
|
||||
SVGA3dCopyRect cr;
|
||||
} *cmd;
|
||||
|
||||
/**
|
||||
* Strictly we should take the ttm_lock in read mode before accessing
|
||||
* the fifo, to make sure the fifo is present and up. However,
|
||||
* instead we flush all workqueues under the ttm lock in exclusive mode
|
||||
* before taking down the fifo.
|
||||
*/
|
||||
mutex_lock(&vfbs->work_lock);
|
||||
if (!vfbs->present_fs)
|
||||
goto out_unlock;
|
||||
|
||||
cmd = vmw_fifo_reserve(dev_priv, sizeof(*cmd));
|
||||
if (unlikely(cmd == NULL))
|
||||
goto out_resched;
|
||||
|
||||
cmd->header.id = cpu_to_le32(SVGA_3D_CMD_PRESENT);
|
||||
cmd->header.size = cpu_to_le32(sizeof(cmd->body) + sizeof(cmd->cr));
|
||||
cmd->body.sid = cpu_to_le32(surf->res.id);
|
||||
cmd->cr.x = cpu_to_le32(0);
|
||||
cmd->cr.y = cpu_to_le32(0);
|
||||
cmd->cr.srcx = cmd->cr.x;
|
||||
cmd->cr.srcy = cmd->cr.y;
|
||||
cmd->cr.w = cpu_to_le32(framebuffer->width);
|
||||
cmd->cr.h = cpu_to_le32(framebuffer->height);
|
||||
vfbs->present_fs = false;
|
||||
vmw_fifo_commit(dev_priv, sizeof(*cmd));
|
||||
out_resched:
|
||||
/**
|
||||
* Will not re-add if already pending.
|
||||
*/
|
||||
schedule_delayed_work(&vfbs->d_work, VMWGFX_PRESENT_RATE);
|
||||
out_unlock:
|
||||
mutex_unlock(&vfbs->work_lock);
|
||||
}
|
||||
|
||||
static int do_surface_dirty_ldu(struct vmw_private *dev_priv,
|
||||
struct vmw_framebuffer *framebuffer,
|
||||
struct vmw_surface *surf,
|
||||
unsigned flags, unsigned color,
|
||||
struct drm_clip_rect *clips,
|
||||
unsigned num_clips, int inc)
|
||||
{
|
||||
SVGA3dCopyRect *cr;
|
||||
int i;
|
||||
|
||||
struct {
|
||||
SVGA3dCmdHeader header;
|
||||
SVGA3dCmdPresent body;
|
||||
SVGA3dCopyRect cr;
|
||||
} *cmd;
|
||||
|
||||
cmd = vmw_fifo_reserve(dev_priv, sizeof(*cmd) + (num_clips - 1) *
|
||||
sizeof(cmd->cr));
|
||||
if (unlikely(cmd == NULL)) {
|
||||
DRM_ERROR("Fifo reserve failed.\n");
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
||||
memset(cmd, 0, sizeof(*cmd));
|
||||
|
||||
cmd->header.id = cpu_to_le32(SVGA_3D_CMD_PRESENT);
|
||||
cmd->header.size = cpu_to_le32(sizeof(cmd->body) + num_clips *
|
||||
sizeof(cmd->cr));
|
||||
cmd->body.sid = cpu_to_le32(surf->res.id);
|
||||
|
||||
for (i = 0, cr = &cmd->cr; i < num_clips; i++, cr++, clips += inc) {
|
||||
cr->x = cpu_to_le16(clips->x1);
|
||||
cr->y = cpu_to_le16(clips->y1);
|
||||
cr->srcx = cr->x;
|
||||
cr->srcy = cr->y;
|
||||
cr->w = cpu_to_le16(clips->x2 - clips->x1);
|
||||
cr->h = cpu_to_le16(clips->y2 - clips->y1);
|
||||
}
|
||||
|
||||
vmw_fifo_commit(dev_priv, sizeof(*cmd) + (num_clips - 1) *
|
||||
sizeof(cmd->cr));
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int do_surface_dirty_sou(struct vmw_private *dev_priv,
|
||||
struct vmw_framebuffer *framebuffer,
|
||||
struct vmw_surface *surf,
|
||||
|
@ -551,28 +421,14 @@ int vmw_framebuffer_surface_dirty(struct drm_framebuffer *framebuffer,
|
|||
if (unlikely(vfbs->master != file_priv->master))
|
||||
return -EINVAL;
|
||||
|
||||
/* Require ScreenObject support for 3D */
|
||||
if (!dev_priv->sou_priv)
|
||||
return -EINVAL;
|
||||
|
||||
ret = ttm_read_lock(&vmaster->lock, true);
|
||||
if (unlikely(ret != 0))
|
||||
return ret;
|
||||
|
||||
/* Are we using screen objects? */
|
||||
if (!dev_priv->sou_priv) {
|
||||
int ret;
|
||||
|
||||
mutex_lock(&vfbs->work_lock);
|
||||
vfbs->present_fs = true;
|
||||
ret = schedule_delayed_work(&vfbs->d_work, VMWGFX_PRESENT_RATE);
|
||||
mutex_unlock(&vfbs->work_lock);
|
||||
if (ret) {
|
||||
/**
|
||||
* No work pending, Force immediate present.
|
||||
*/
|
||||
vmw_framebuffer_present_fs_callback(&vfbs->d_work.work);
|
||||
}
|
||||
ttm_read_unlock(&vmaster->lock);
|
||||
return 0;
|
||||
}
|
||||
|
||||
if (!num_clips) {
|
||||
num_clips = 1;
|
||||
clips = &norect;
|
||||
|
@ -584,14 +440,9 @@ int vmw_framebuffer_surface_dirty(struct drm_framebuffer *framebuffer,
|
|||
inc = 2; /* skip source rects */
|
||||
}
|
||||
|
||||
if (!dev_priv->sou_priv)
|
||||
ret = do_surface_dirty_ldu(dev_priv, &vfbs->base, surf,
|
||||
flags, color,
|
||||
clips, num_clips, inc);
|
||||
else
|
||||
ret = do_surface_dirty_sou(dev_priv, &vfbs->base, surf,
|
||||
flags, color,
|
||||
clips, num_clips, inc);
|
||||
ret = do_surface_dirty_sou(dev_priv, &vfbs->base, surf,
|
||||
flags, color,
|
||||
clips, num_clips, inc);
|
||||
|
||||
ttm_read_unlock(&vmaster->lock);
|
||||
return 0;
|
||||
|
@ -617,6 +468,10 @@ static int vmw_kms_new_framebuffer_surface(struct vmw_private *dev_priv,
|
|||
struct vmw_master *vmaster = vmw_master(file_priv->master);
|
||||
int ret;
|
||||
|
||||
/* 3D is only supported on HWv8 hosts which supports screen objects */
|
||||
if (!dev_priv->sou_priv)
|
||||
return -ENOSYS;
|
||||
|
||||
/*
|
||||
* Sanity checks.
|
||||
*/
|
||||
|
@ -679,19 +534,10 @@ static int vmw_kms_new_framebuffer_surface(struct vmw_private *dev_priv,
|
|||
vfbs->base.base.depth = mode_cmd->depth;
|
||||
vfbs->base.base.width = mode_cmd->width;
|
||||
vfbs->base.base.height = mode_cmd->height;
|
||||
/* Don't need to fill start of vram with empty
|
||||
* buffer if we have screen objects support.
|
||||
*/
|
||||
if (!dev_priv->sou_priv) {
|
||||
vfbs->base.pin = &vmw_surface_dmabuf_pin;
|
||||
vfbs->base.unpin = &vmw_surface_dmabuf_unpin;
|
||||
}
|
||||
vfbs->surface = surface;
|
||||
vfbs->master = drm_master_get(file_priv->master);
|
||||
mutex_init(&vfbs->work_lock);
|
||||
|
||||
mutex_lock(&vmaster->fb_surf_mutex);
|
||||
INIT_DELAYED_WORK(&vfbs->d_work, &vmw_framebuffer_present_fs_callback);
|
||||
list_add_tail(&vfbs->head, &vmaster->fb_surf);
|
||||
mutex_unlock(&vmaster->fb_surf_mutex);
|
||||
|
||||
|
@ -872,55 +718,6 @@ static struct drm_framebuffer_funcs vmw_framebuffer_dmabuf_funcs = {
|
|||
.create_handle = vmw_framebuffer_create_handle,
|
||||
};
|
||||
|
||||
/**
|
||||
* We need to reserve the start of vram because the host might
|
||||
* scribble to it at mode changes, so we need to reserve it.
|
||||
*/
|
||||
static int vmw_surface_dmabuf_pin(struct vmw_framebuffer *vfb)
|
||||
{
|
||||
struct vmw_private *dev_priv = vmw_priv(vfb->base.dev);
|
||||
struct vmw_framebuffer_surface *vfbs =
|
||||
vmw_framebuffer_to_vfbs(&vfb->base);
|
||||
unsigned long size = vfbs->base.base.pitch * vfbs->base.base.height;
|
||||
int ret;
|
||||
struct ttm_placement ne_placement = vmw_vram_ne_placement;
|
||||
|
||||
ne_placement.lpfn = (size + (PAGE_SIZE - 1)) / PAGE_SIZE;
|
||||
|
||||
vfbs->buffer = kzalloc(sizeof(*vfbs->buffer), GFP_KERNEL);
|
||||
if (unlikely(vfbs->buffer == NULL))
|
||||
return -ENOMEM;
|
||||
|
||||
vmw_overlay_pause_all(dev_priv);
|
||||
ret = vmw_dmabuf_init(dev_priv, vfbs->buffer, size,
|
||||
&vmw_vram_ne_placement,
|
||||
false, &vmw_dmabuf_bo_free);
|
||||
vmw_overlay_resume_all(dev_priv);
|
||||
if (unlikely(ret != 0))
|
||||
vfbs->buffer = NULL;
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
/**
|
||||
* See vmw_surface_dmabuf_pin.
|
||||
*/
|
||||
static int vmw_surface_dmabuf_unpin(struct vmw_framebuffer *vfb)
|
||||
{
|
||||
struct ttm_buffer_object *bo;
|
||||
struct vmw_framebuffer_surface *vfbs =
|
||||
vmw_framebuffer_to_vfbs(&vfb->base);
|
||||
|
||||
if (unlikely(vfbs->buffer == NULL))
|
||||
return 0;
|
||||
|
||||
bo = &vfbs->buffer->base;
|
||||
ttm_bo_unref(&bo);
|
||||
vfbs->buffer = NULL;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
/**
|
||||
* Pin the dmabuffer to the start of vram.
|
||||
*/
|
||||
|
|
Загрузка…
Ссылка в новой задаче