Merge branch 'drm-fixes' of git://people.freedesktop.org/~airlied/linux
Pull drm fixes from Dave Airlie: "One nouveau deadlock fix, one qxl irq handling fix, and a set of radeon pageflipping changes that fix regressions in pageflipping since -rc1 along with a leak and backlight fix. The pageflipping fixes are a bit bigger than I'd like, but there has been a few people focused on testing them" * 'drm-fixes' of git://people.freedesktop.org/~airlied/linux: drm/radeon: Make classic pageflip completion path less racy. drm/radeon: Add missing vblank_put in pageflip ioctl error path. drm/radeon: Remove redundant fence unref in pageflip path. drm/radeon: Complete page flip even if waiting on the BO fence fails drm/radeon: Move pinning the BO back to radeon_crtc_page_flip() drm/radeon: Prevent too early kms-pageflips triggered by vblank. drm/radeon: set default bl level to something reasonable drm/radeon: avoid leaking edid data drm/qxl: return IRQ_NONE if it was not our irq drm/nouveau/therm: fix a potential deadlock in the therm monitoring code
This commit is contained in:
Коммит
3e8e2756b6
|
@ -192,11 +192,11 @@ alarm_timer_callback(struct nouveau_alarm *alarm)
|
||||||
nouveau_therm_threshold_hyst_polling(therm, &sensor->thrs_shutdown,
|
nouveau_therm_threshold_hyst_polling(therm, &sensor->thrs_shutdown,
|
||||||
NOUVEAU_THERM_THRS_SHUTDOWN);
|
NOUVEAU_THERM_THRS_SHUTDOWN);
|
||||||
|
|
||||||
|
spin_unlock_irqrestore(&priv->sensor.alarm_program_lock, flags);
|
||||||
|
|
||||||
/* schedule the next poll in one second */
|
/* schedule the next poll in one second */
|
||||||
if (therm->temp_get(therm) >= 0 && list_empty(&alarm->head))
|
if (therm->temp_get(therm) >= 0 && list_empty(&alarm->head))
|
||||||
ptimer->alarm(ptimer, 1000 * 1000 * 1000, alarm);
|
ptimer->alarm(ptimer, 1000000000ULL, alarm);
|
||||||
|
|
||||||
spin_unlock_irqrestore(&priv->sensor.alarm_program_lock, flags);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
void
|
void
|
||||||
|
|
|
@ -33,6 +33,9 @@ irqreturn_t qxl_irq_handler(int irq, void *arg)
|
||||||
|
|
||||||
pending = xchg(&qdev->ram_header->int_pending, 0);
|
pending = xchg(&qdev->ram_header->int_pending, 0);
|
||||||
|
|
||||||
|
if (!pending)
|
||||||
|
return IRQ_NONE;
|
||||||
|
|
||||||
atomic_inc(&qdev->irq_received);
|
atomic_inc(&qdev->irq_received);
|
||||||
|
|
||||||
if (pending & QXL_INTERRUPT_DISPLAY) {
|
if (pending & QXL_INTERRUPT_DISPLAY) {
|
||||||
|
|
|
@ -1414,8 +1414,8 @@ static int dce4_crtc_do_set_base(struct drm_crtc *crtc,
|
||||||
tmp &= ~EVERGREEN_GRPH_SURFACE_UPDATE_H_RETRACE_EN;
|
tmp &= ~EVERGREEN_GRPH_SURFACE_UPDATE_H_RETRACE_EN;
|
||||||
WREG32(EVERGREEN_GRPH_FLIP_CONTROL + radeon_crtc->crtc_offset, tmp);
|
WREG32(EVERGREEN_GRPH_FLIP_CONTROL + radeon_crtc->crtc_offset, tmp);
|
||||||
|
|
||||||
/* set pageflip to happen anywhere in vblank interval */
|
/* set pageflip to happen only at start of vblank interval (front porch) */
|
||||||
WREG32(EVERGREEN_MASTER_UPDATE_MODE + radeon_crtc->crtc_offset, 0);
|
WREG32(EVERGREEN_MASTER_UPDATE_MODE + radeon_crtc->crtc_offset, 3);
|
||||||
|
|
||||||
if (!atomic && fb && fb != crtc->primary->fb) {
|
if (!atomic && fb && fb != crtc->primary->fb) {
|
||||||
radeon_fb = to_radeon_framebuffer(fb);
|
radeon_fb = to_radeon_framebuffer(fb);
|
||||||
|
@ -1614,8 +1614,8 @@ static int avivo_crtc_do_set_base(struct drm_crtc *crtc,
|
||||||
tmp &= ~AVIVO_D1GRPH_SURFACE_UPDATE_H_RETRACE_EN;
|
tmp &= ~AVIVO_D1GRPH_SURFACE_UPDATE_H_RETRACE_EN;
|
||||||
WREG32(AVIVO_D1GRPH_FLIP_CONTROL + radeon_crtc->crtc_offset, tmp);
|
WREG32(AVIVO_D1GRPH_FLIP_CONTROL + radeon_crtc->crtc_offset, tmp);
|
||||||
|
|
||||||
/* set pageflip to happen anywhere in vblank interval */
|
/* set pageflip to happen only at start of vblank interval (front porch) */
|
||||||
WREG32(AVIVO_D1MODE_MASTER_UPDATE_MODE + radeon_crtc->crtc_offset, 0);
|
WREG32(AVIVO_D1MODE_MASTER_UPDATE_MODE + radeon_crtc->crtc_offset, 3);
|
||||||
|
|
||||||
if (!atomic && fb && fb != crtc->primary->fb) {
|
if (!atomic && fb && fb != crtc->primary->fb) {
|
||||||
radeon_fb = to_radeon_framebuffer(fb);
|
radeon_fb = to_radeon_framebuffer(fb);
|
||||||
|
|
|
@ -183,7 +183,6 @@ void radeon_atom_backlight_init(struct radeon_encoder *radeon_encoder,
|
||||||
struct backlight_properties props;
|
struct backlight_properties props;
|
||||||
struct radeon_backlight_privdata *pdata;
|
struct radeon_backlight_privdata *pdata;
|
||||||
struct radeon_encoder_atom_dig *dig;
|
struct radeon_encoder_atom_dig *dig;
|
||||||
u8 backlight_level;
|
|
||||||
char bl_name[16];
|
char bl_name[16];
|
||||||
|
|
||||||
/* Mac laptops with multiple GPUs use the gmux driver for backlight
|
/* Mac laptops with multiple GPUs use the gmux driver for backlight
|
||||||
|
@ -222,12 +221,17 @@ void radeon_atom_backlight_init(struct radeon_encoder *radeon_encoder,
|
||||||
|
|
||||||
pdata->encoder = radeon_encoder;
|
pdata->encoder = radeon_encoder;
|
||||||
|
|
||||||
backlight_level = radeon_atom_get_backlight_level_from_reg(rdev);
|
|
||||||
|
|
||||||
dig = radeon_encoder->enc_priv;
|
dig = radeon_encoder->enc_priv;
|
||||||
dig->bl_dev = bd;
|
dig->bl_dev = bd;
|
||||||
|
|
||||||
bd->props.brightness = radeon_atom_backlight_get_brightness(bd);
|
bd->props.brightness = radeon_atom_backlight_get_brightness(bd);
|
||||||
|
/* Set a reasonable default here if the level is 0 otherwise
|
||||||
|
* fbdev will attempt to turn the backlight on after console
|
||||||
|
* unblanking and it will try and restore 0 which turns the backlight
|
||||||
|
* off again.
|
||||||
|
*/
|
||||||
|
if (bd->props.brightness == 0)
|
||||||
|
bd->props.brightness = RADEON_MAX_BL_LEVEL;
|
||||||
bd->props.power = FB_BLANK_UNBLANK;
|
bd->props.power = FB_BLANK_UNBLANK;
|
||||||
backlight_update_status(bd);
|
backlight_update_status(bd);
|
||||||
|
|
||||||
|
|
|
@ -2642,8 +2642,9 @@ void evergreen_mc_resume(struct radeon_device *rdev, struct evergreen_mc_save *s
|
||||||
for (i = 0; i < rdev->num_crtc; i++) {
|
for (i = 0; i < rdev->num_crtc; i++) {
|
||||||
if (save->crtc_enabled[i]) {
|
if (save->crtc_enabled[i]) {
|
||||||
tmp = RREG32(EVERGREEN_MASTER_UPDATE_MODE + crtc_offsets[i]);
|
tmp = RREG32(EVERGREEN_MASTER_UPDATE_MODE + crtc_offsets[i]);
|
||||||
if ((tmp & 0x3) != 0) {
|
if ((tmp & 0x7) != 3) {
|
||||||
tmp &= ~0x3;
|
tmp &= ~0x7;
|
||||||
|
tmp |= 0x3;
|
||||||
WREG32(EVERGREEN_MASTER_UPDATE_MODE + crtc_offsets[i], tmp);
|
WREG32(EVERGREEN_MASTER_UPDATE_MODE + crtc_offsets[i], tmp);
|
||||||
}
|
}
|
||||||
tmp = RREG32(EVERGREEN_GRPH_UPDATE + crtc_offsets[i]);
|
tmp = RREG32(EVERGREEN_GRPH_UPDATE + crtc_offsets[i]);
|
||||||
|
|
|
@ -239,7 +239,6 @@
|
||||||
# define EVERGREEN_CRTC_V_BLANK (1 << 0)
|
# define EVERGREEN_CRTC_V_BLANK (1 << 0)
|
||||||
#define EVERGREEN_CRTC_STATUS_POSITION 0x6e90
|
#define EVERGREEN_CRTC_STATUS_POSITION 0x6e90
|
||||||
#define EVERGREEN_CRTC_STATUS_HV_COUNT 0x6ea0
|
#define EVERGREEN_CRTC_STATUS_HV_COUNT 0x6ea0
|
||||||
#define EVERGREEN_MASTER_UPDATE_MODE 0x6ef8
|
|
||||||
#define EVERGREEN_CRTC_UPDATE_LOCK 0x6ed4
|
#define EVERGREEN_CRTC_UPDATE_LOCK 0x6ed4
|
||||||
#define EVERGREEN_MASTER_UPDATE_LOCK 0x6ef4
|
#define EVERGREEN_MASTER_UPDATE_LOCK 0x6ef4
|
||||||
#define EVERGREEN_MASTER_UPDATE_MODE 0x6ef8
|
#define EVERGREEN_MASTER_UPDATE_MODE 0x6ef8
|
||||||
|
|
|
@ -684,10 +684,9 @@ struct radeon_flip_work {
|
||||||
struct work_struct unpin_work;
|
struct work_struct unpin_work;
|
||||||
struct radeon_device *rdev;
|
struct radeon_device *rdev;
|
||||||
int crtc_id;
|
int crtc_id;
|
||||||
struct drm_framebuffer *fb;
|
uint64_t base;
|
||||||
struct drm_pending_vblank_event *event;
|
struct drm_pending_vblank_event *event;
|
||||||
struct radeon_bo *old_rbo;
|
struct radeon_bo *old_rbo;
|
||||||
struct radeon_bo *new_rbo;
|
|
||||||
struct radeon_fence *fence;
|
struct radeon_fence *fence;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
|
@ -366,7 +366,6 @@ void radeon_crtc_handle_flip(struct radeon_device *rdev, int crtc_id)
|
||||||
spin_unlock_irqrestore(&rdev->ddev->event_lock, flags);
|
spin_unlock_irqrestore(&rdev->ddev->event_lock, flags);
|
||||||
|
|
||||||
drm_vblank_put(rdev->ddev, radeon_crtc->crtc_id);
|
drm_vblank_put(rdev->ddev, radeon_crtc->crtc_id);
|
||||||
radeon_fence_unref(&work->fence);
|
|
||||||
radeon_irq_kms_pflip_irq_put(rdev, work->crtc_id);
|
radeon_irq_kms_pflip_irq_put(rdev, work->crtc_id);
|
||||||
queue_work(radeon_crtc->flip_queue, &work->unpin_work);
|
queue_work(radeon_crtc->flip_queue, &work->unpin_work);
|
||||||
}
|
}
|
||||||
|
@ -386,51 +385,108 @@ static void radeon_flip_work_func(struct work_struct *__work)
|
||||||
struct radeon_crtc *radeon_crtc = rdev->mode_info.crtcs[work->crtc_id];
|
struct radeon_crtc *radeon_crtc = rdev->mode_info.crtcs[work->crtc_id];
|
||||||
|
|
||||||
struct drm_crtc *crtc = &radeon_crtc->base;
|
struct drm_crtc *crtc = &radeon_crtc->base;
|
||||||
struct drm_framebuffer *fb = work->fb;
|
|
||||||
|
|
||||||
uint32_t tiling_flags, pitch_pixels;
|
|
||||||
uint64_t base;
|
|
||||||
|
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
int r;
|
int r;
|
||||||
|
|
||||||
down_read(&rdev->exclusive_lock);
|
down_read(&rdev->exclusive_lock);
|
||||||
while (work->fence) {
|
if (work->fence) {
|
||||||
r = radeon_fence_wait(work->fence, false);
|
r = radeon_fence_wait(work->fence, false);
|
||||||
if (r == -EDEADLK) {
|
if (r == -EDEADLK) {
|
||||||
up_read(&rdev->exclusive_lock);
|
up_read(&rdev->exclusive_lock);
|
||||||
r = radeon_gpu_reset(rdev);
|
r = radeon_gpu_reset(rdev);
|
||||||
down_read(&rdev->exclusive_lock);
|
down_read(&rdev->exclusive_lock);
|
||||||
}
|
}
|
||||||
|
if (r)
|
||||||
|
DRM_ERROR("failed to wait on page flip fence (%d)!\n", r);
|
||||||
|
|
||||||
|
/* We continue with the page flip even if we failed to wait on
|
||||||
|
* the fence, otherwise the DRM core and userspace will be
|
||||||
|
* confused about which BO the CRTC is scanning out
|
||||||
|
*/
|
||||||
|
|
||||||
if (r) {
|
|
||||||
DRM_ERROR("failed to wait on page flip fence (%d)!\n",
|
|
||||||
r);
|
|
||||||
goto cleanup;
|
|
||||||
} else
|
|
||||||
radeon_fence_unref(&work->fence);
|
radeon_fence_unref(&work->fence);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* pin the new buffer */
|
/* We borrow the event spin lock for protecting flip_status */
|
||||||
DRM_DEBUG_DRIVER("flip-ioctl() cur_fbo = %p, cur_bbo = %p\n",
|
spin_lock_irqsave(&crtc->dev->event_lock, flags);
|
||||||
work->old_rbo, work->new_rbo);
|
|
||||||
|
|
||||||
r = radeon_bo_reserve(work->new_rbo, false);
|
/* set the proper interrupt */
|
||||||
|
radeon_irq_kms_pflip_irq_get(rdev, radeon_crtc->crtc_id);
|
||||||
|
|
||||||
|
/* do the flip (mmio) */
|
||||||
|
radeon_page_flip(rdev, radeon_crtc->crtc_id, work->base);
|
||||||
|
|
||||||
|
radeon_crtc->flip_status = RADEON_FLIP_SUBMITTED;
|
||||||
|
spin_unlock_irqrestore(&crtc->dev->event_lock, flags);
|
||||||
|
up_read(&rdev->exclusive_lock);
|
||||||
|
}
|
||||||
|
|
||||||
|
static int radeon_crtc_page_flip(struct drm_crtc *crtc,
|
||||||
|
struct drm_framebuffer *fb,
|
||||||
|
struct drm_pending_vblank_event *event,
|
||||||
|
uint32_t page_flip_flags)
|
||||||
|
{
|
||||||
|
struct drm_device *dev = crtc->dev;
|
||||||
|
struct radeon_device *rdev = dev->dev_private;
|
||||||
|
struct radeon_crtc *radeon_crtc = to_radeon_crtc(crtc);
|
||||||
|
struct radeon_framebuffer *old_radeon_fb;
|
||||||
|
struct radeon_framebuffer *new_radeon_fb;
|
||||||
|
struct drm_gem_object *obj;
|
||||||
|
struct radeon_flip_work *work;
|
||||||
|
struct radeon_bo *new_rbo;
|
||||||
|
uint32_t tiling_flags, pitch_pixels;
|
||||||
|
uint64_t base;
|
||||||
|
unsigned long flags;
|
||||||
|
int r;
|
||||||
|
|
||||||
|
work = kzalloc(sizeof *work, GFP_KERNEL);
|
||||||
|
if (work == NULL)
|
||||||
|
return -ENOMEM;
|
||||||
|
|
||||||
|
INIT_WORK(&work->flip_work, radeon_flip_work_func);
|
||||||
|
INIT_WORK(&work->unpin_work, radeon_unpin_work_func);
|
||||||
|
|
||||||
|
work->rdev = rdev;
|
||||||
|
work->crtc_id = radeon_crtc->crtc_id;
|
||||||
|
work->event = event;
|
||||||
|
|
||||||
|
/* schedule unpin of the old buffer */
|
||||||
|
old_radeon_fb = to_radeon_framebuffer(crtc->primary->fb);
|
||||||
|
obj = old_radeon_fb->obj;
|
||||||
|
|
||||||
|
/* take a reference to the old object */
|
||||||
|
drm_gem_object_reference(obj);
|
||||||
|
work->old_rbo = gem_to_radeon_bo(obj);
|
||||||
|
|
||||||
|
new_radeon_fb = to_radeon_framebuffer(fb);
|
||||||
|
obj = new_radeon_fb->obj;
|
||||||
|
new_rbo = gem_to_radeon_bo(obj);
|
||||||
|
|
||||||
|
spin_lock(&new_rbo->tbo.bdev->fence_lock);
|
||||||
|
if (new_rbo->tbo.sync_obj)
|
||||||
|
work->fence = radeon_fence_ref(new_rbo->tbo.sync_obj);
|
||||||
|
spin_unlock(&new_rbo->tbo.bdev->fence_lock);
|
||||||
|
|
||||||
|
/* pin the new buffer */
|
||||||
|
DRM_DEBUG_DRIVER("flip-ioctl() cur_rbo = %p, new_rbo = %p\n",
|
||||||
|
work->old_rbo, new_rbo);
|
||||||
|
|
||||||
|
r = radeon_bo_reserve(new_rbo, false);
|
||||||
if (unlikely(r != 0)) {
|
if (unlikely(r != 0)) {
|
||||||
DRM_ERROR("failed to reserve new rbo buffer before flip\n");
|
DRM_ERROR("failed to reserve new rbo buffer before flip\n");
|
||||||
goto cleanup;
|
goto cleanup;
|
||||||
}
|
}
|
||||||
/* Only 27 bit offset for legacy CRTC */
|
/* Only 27 bit offset for legacy CRTC */
|
||||||
r = radeon_bo_pin_restricted(work->new_rbo, RADEON_GEM_DOMAIN_VRAM,
|
r = radeon_bo_pin_restricted(new_rbo, RADEON_GEM_DOMAIN_VRAM,
|
||||||
ASIC_IS_AVIVO(rdev) ? 0 : 1 << 27, &base);
|
ASIC_IS_AVIVO(rdev) ? 0 : 1 << 27, &base);
|
||||||
if (unlikely(r != 0)) {
|
if (unlikely(r != 0)) {
|
||||||
radeon_bo_unreserve(work->new_rbo);
|
radeon_bo_unreserve(new_rbo);
|
||||||
r = -EINVAL;
|
r = -EINVAL;
|
||||||
DRM_ERROR("failed to pin new rbo buffer before flip\n");
|
DRM_ERROR("failed to pin new rbo buffer before flip\n");
|
||||||
goto cleanup;
|
goto cleanup;
|
||||||
}
|
}
|
||||||
radeon_bo_get_tiling_flags(work->new_rbo, &tiling_flags, NULL);
|
radeon_bo_get_tiling_flags(new_rbo, &tiling_flags, NULL);
|
||||||
radeon_bo_unreserve(work->new_rbo);
|
radeon_bo_unreserve(new_rbo);
|
||||||
|
|
||||||
if (!ASIC_IS_AVIVO(rdev)) {
|
if (!ASIC_IS_AVIVO(rdev)) {
|
||||||
/* crtc offset is from display base addr not FB location */
|
/* crtc offset is from display base addr not FB location */
|
||||||
|
@ -467,6 +523,7 @@ static void radeon_flip_work_func(struct work_struct *__work)
|
||||||
}
|
}
|
||||||
base &= ~7;
|
base &= ~7;
|
||||||
}
|
}
|
||||||
|
work->base = base;
|
||||||
|
|
||||||
r = drm_vblank_get(crtc->dev, radeon_crtc->crtc_id);
|
r = drm_vblank_get(crtc->dev, radeon_crtc->crtc_id);
|
||||||
if (r) {
|
if (r) {
|
||||||
|
@ -477,88 +534,11 @@ static void radeon_flip_work_func(struct work_struct *__work)
|
||||||
/* We borrow the event spin lock for protecting flip_work */
|
/* We borrow the event spin lock for protecting flip_work */
|
||||||
spin_lock_irqsave(&crtc->dev->event_lock, flags);
|
spin_lock_irqsave(&crtc->dev->event_lock, flags);
|
||||||
|
|
||||||
/* set the proper interrupt */
|
|
||||||
radeon_irq_kms_pflip_irq_get(rdev, radeon_crtc->crtc_id);
|
|
||||||
|
|
||||||
/* do the flip (mmio) */
|
|
||||||
radeon_page_flip(rdev, radeon_crtc->crtc_id, base);
|
|
||||||
|
|
||||||
radeon_crtc->flip_status = RADEON_FLIP_SUBMITTED;
|
|
||||||
spin_unlock_irqrestore(&crtc->dev->event_lock, flags);
|
|
||||||
up_read(&rdev->exclusive_lock);
|
|
||||||
|
|
||||||
return;
|
|
||||||
|
|
||||||
pflip_cleanup:
|
|
||||||
if (unlikely(radeon_bo_reserve(work->new_rbo, false) != 0)) {
|
|
||||||
DRM_ERROR("failed to reserve new rbo in error path\n");
|
|
||||||
goto cleanup;
|
|
||||||
}
|
|
||||||
if (unlikely(radeon_bo_unpin(work->new_rbo) != 0)) {
|
|
||||||
DRM_ERROR("failed to unpin new rbo in error path\n");
|
|
||||||
}
|
|
||||||
radeon_bo_unreserve(work->new_rbo);
|
|
||||||
|
|
||||||
cleanup:
|
|
||||||
drm_gem_object_unreference_unlocked(&work->old_rbo->gem_base);
|
|
||||||
radeon_fence_unref(&work->fence);
|
|
||||||
kfree(work);
|
|
||||||
up_read(&rdev->exclusive_lock);
|
|
||||||
}
|
|
||||||
|
|
||||||
static int radeon_crtc_page_flip(struct drm_crtc *crtc,
|
|
||||||
struct drm_framebuffer *fb,
|
|
||||||
struct drm_pending_vblank_event *event,
|
|
||||||
uint32_t page_flip_flags)
|
|
||||||
{
|
|
||||||
struct drm_device *dev = crtc->dev;
|
|
||||||
struct radeon_device *rdev = dev->dev_private;
|
|
||||||
struct radeon_crtc *radeon_crtc = to_radeon_crtc(crtc);
|
|
||||||
struct radeon_framebuffer *old_radeon_fb;
|
|
||||||
struct radeon_framebuffer *new_radeon_fb;
|
|
||||||
struct drm_gem_object *obj;
|
|
||||||
struct radeon_flip_work *work;
|
|
||||||
unsigned long flags;
|
|
||||||
|
|
||||||
work = kzalloc(sizeof *work, GFP_KERNEL);
|
|
||||||
if (work == NULL)
|
|
||||||
return -ENOMEM;
|
|
||||||
|
|
||||||
INIT_WORK(&work->flip_work, radeon_flip_work_func);
|
|
||||||
INIT_WORK(&work->unpin_work, radeon_unpin_work_func);
|
|
||||||
|
|
||||||
work->rdev = rdev;
|
|
||||||
work->crtc_id = radeon_crtc->crtc_id;
|
|
||||||
work->fb = fb;
|
|
||||||
work->event = event;
|
|
||||||
|
|
||||||
/* schedule unpin of the old buffer */
|
|
||||||
old_radeon_fb = to_radeon_framebuffer(crtc->primary->fb);
|
|
||||||
obj = old_radeon_fb->obj;
|
|
||||||
|
|
||||||
/* take a reference to the old object */
|
|
||||||
drm_gem_object_reference(obj);
|
|
||||||
work->old_rbo = gem_to_radeon_bo(obj);
|
|
||||||
|
|
||||||
new_radeon_fb = to_radeon_framebuffer(fb);
|
|
||||||
obj = new_radeon_fb->obj;
|
|
||||||
work->new_rbo = gem_to_radeon_bo(obj);
|
|
||||||
|
|
||||||
spin_lock(&work->new_rbo->tbo.bdev->fence_lock);
|
|
||||||
if (work->new_rbo->tbo.sync_obj)
|
|
||||||
work->fence = radeon_fence_ref(work->new_rbo->tbo.sync_obj);
|
|
||||||
spin_unlock(&work->new_rbo->tbo.bdev->fence_lock);
|
|
||||||
|
|
||||||
/* We borrow the event spin lock for protecting flip_work */
|
|
||||||
spin_lock_irqsave(&crtc->dev->event_lock, flags);
|
|
||||||
|
|
||||||
if (radeon_crtc->flip_status != RADEON_FLIP_NONE) {
|
if (radeon_crtc->flip_status != RADEON_FLIP_NONE) {
|
||||||
DRM_DEBUG_DRIVER("flip queue: crtc already busy\n");
|
DRM_DEBUG_DRIVER("flip queue: crtc already busy\n");
|
||||||
spin_unlock_irqrestore(&crtc->dev->event_lock, flags);
|
spin_unlock_irqrestore(&crtc->dev->event_lock, flags);
|
||||||
drm_gem_object_unreference_unlocked(&work->old_rbo->gem_base);
|
r = -EBUSY;
|
||||||
radeon_fence_unref(&work->fence);
|
goto vblank_cleanup;
|
||||||
kfree(work);
|
|
||||||
return -EBUSY;
|
|
||||||
}
|
}
|
||||||
radeon_crtc->flip_status = RADEON_FLIP_PENDING;
|
radeon_crtc->flip_status = RADEON_FLIP_PENDING;
|
||||||
radeon_crtc->flip_work = work;
|
radeon_crtc->flip_work = work;
|
||||||
|
@ -569,8 +549,27 @@ static int radeon_crtc_page_flip(struct drm_crtc *crtc,
|
||||||
spin_unlock_irqrestore(&crtc->dev->event_lock, flags);
|
spin_unlock_irqrestore(&crtc->dev->event_lock, flags);
|
||||||
|
|
||||||
queue_work(radeon_crtc->flip_queue, &work->flip_work);
|
queue_work(radeon_crtc->flip_queue, &work->flip_work);
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
|
vblank_cleanup:
|
||||||
|
drm_vblank_put(crtc->dev, radeon_crtc->crtc_id);
|
||||||
|
|
||||||
|
pflip_cleanup:
|
||||||
|
if (unlikely(radeon_bo_reserve(new_rbo, false) != 0)) {
|
||||||
|
DRM_ERROR("failed to reserve new rbo in error path\n");
|
||||||
|
goto cleanup;
|
||||||
|
}
|
||||||
|
if (unlikely(radeon_bo_unpin(new_rbo) != 0)) {
|
||||||
|
DRM_ERROR("failed to unpin new rbo in error path\n");
|
||||||
|
}
|
||||||
|
radeon_bo_unreserve(new_rbo);
|
||||||
|
|
||||||
|
cleanup:
|
||||||
|
drm_gem_object_unreference_unlocked(&work->old_rbo->gem_base);
|
||||||
|
radeon_fence_unref(&work->fence);
|
||||||
|
kfree(work);
|
||||||
|
|
||||||
|
return r;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int
|
static int
|
||||||
|
@ -830,6 +829,10 @@ int radeon_ddc_get_modes(struct radeon_connector *radeon_connector)
|
||||||
struct radeon_device *rdev = dev->dev_private;
|
struct radeon_device *rdev = dev->dev_private;
|
||||||
int ret = 0;
|
int ret = 0;
|
||||||
|
|
||||||
|
/* don't leak the edid if we already fetched it in detect() */
|
||||||
|
if (radeon_connector->edid)
|
||||||
|
goto got_edid;
|
||||||
|
|
||||||
/* on hw with routers, select right port */
|
/* on hw with routers, select right port */
|
||||||
if (radeon_connector->router.ddc_valid)
|
if (radeon_connector->router.ddc_valid)
|
||||||
radeon_router_select_ddc_port(radeon_connector);
|
radeon_router_select_ddc_port(radeon_connector);
|
||||||
|
@ -868,6 +871,7 @@ int radeon_ddc_get_modes(struct radeon_connector *radeon_connector)
|
||||||
radeon_connector->edid = radeon_bios_get_hardcoded_edid(rdev);
|
radeon_connector->edid = radeon_bios_get_hardcoded_edid(rdev);
|
||||||
}
|
}
|
||||||
if (radeon_connector->edid) {
|
if (radeon_connector->edid) {
|
||||||
|
got_edid:
|
||||||
drm_mode_connector_update_edid_property(&radeon_connector->base, radeon_connector->edid);
|
drm_mode_connector_update_edid_property(&radeon_connector->base, radeon_connector->edid);
|
||||||
ret = drm_add_edid_modes(&radeon_connector->base, radeon_connector->edid);
|
ret = drm_add_edid_modes(&radeon_connector->base, radeon_connector->edid);
|
||||||
drm_edid_to_eld(&radeon_connector->base, radeon_connector->edid);
|
drm_edid_to_eld(&radeon_connector->base, radeon_connector->edid);
|
||||||
|
|
|
@ -406,8 +406,9 @@ void rv515_mc_resume(struct radeon_device *rdev, struct rv515_mc_save *save)
|
||||||
for (i = 0; i < rdev->num_crtc; i++) {
|
for (i = 0; i < rdev->num_crtc; i++) {
|
||||||
if (save->crtc_enabled[i]) {
|
if (save->crtc_enabled[i]) {
|
||||||
tmp = RREG32(AVIVO_D1MODE_MASTER_UPDATE_MODE + crtc_offsets[i]);
|
tmp = RREG32(AVIVO_D1MODE_MASTER_UPDATE_MODE + crtc_offsets[i]);
|
||||||
if ((tmp & 0x3) != 0) {
|
if ((tmp & 0x7) != 3) {
|
||||||
tmp &= ~0x3;
|
tmp &= ~0x7;
|
||||||
|
tmp |= 0x3;
|
||||||
WREG32(AVIVO_D1MODE_MASTER_UPDATE_MODE + crtc_offsets[i], tmp);
|
WREG32(AVIVO_D1MODE_MASTER_UPDATE_MODE + crtc_offsets[i], tmp);
|
||||||
}
|
}
|
||||||
tmp = RREG32(AVIVO_D1GRPH_UPDATE + crtc_offsets[i]);
|
tmp = RREG32(AVIVO_D1GRPH_UPDATE + crtc_offsets[i]);
|
||||||
|
|
Загрузка…
Ссылка в новой задаче