2010-01-13 01:54:34 +03:00
|
|
|
/*
|
|
|
|
* Copyright 2010 Advanced Micro Devices, Inc.
|
|
|
|
*
|
|
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
|
|
* to deal in the Software without restriction, including without limitation
|
|
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
|
|
*
|
|
|
|
* The above copyright notice and this permission notice shall be included in
|
|
|
|
* all copies or substantial portions of the Software.
|
|
|
|
*
|
|
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
|
|
* THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
|
|
|
|
* OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
|
|
|
|
* ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
|
|
|
|
* OTHER DEALINGS IN THE SOFTWARE.
|
|
|
|
*
|
|
|
|
* Authors: Alex Deucher
|
|
|
|
*/
|
|
|
|
#include <linux/firmware.h>
|
include cleanup: Update gfp.h and slab.h includes to prepare for breaking implicit slab.h inclusion from percpu.h
percpu.h is included by sched.h and module.h and thus ends up being
included when building most .c files. percpu.h includes slab.h which
in turn includes gfp.h making everything defined by the two files
universally available and complicating inclusion dependencies.
percpu.h -> slab.h dependency is about to be removed. Prepare for
this change by updating users of gfp and slab facilities include those
headers directly instead of assuming availability. As this conversion
needs to touch large number of source files, the following script is
used as the basis of conversion.
http://userweb.kernel.org/~tj/misc/slabh-sweep.py
The script does the followings.
* Scan files for gfp and slab usages and update includes such that
only the necessary includes are there. ie. if only gfp is used,
gfp.h, if slab is used, slab.h.
* When the script inserts a new include, it looks at the include
blocks and try to put the new include such that its order conforms
to its surrounding. It's put in the include block which contains
core kernel includes, in the same order that the rest are ordered -
alphabetical, Christmas tree, rev-Xmas-tree or at the end if there
doesn't seem to be any matching order.
* If the script can't find a place to put a new include (mostly
because the file doesn't have fitting include block), it prints out
an error message indicating which .h file needs to be added to the
file.
The conversion was done in the following steps.
1. The initial automatic conversion of all .c files updated slightly
over 4000 files, deleting around 700 includes and adding ~480 gfp.h
and ~3000 slab.h inclusions. The script emitted errors for ~400
files.
2. Each error was manually checked. Some didn't need the inclusion,
some needed manual addition while adding it to implementation .h or
embedding .c file was more appropriate for others. This step added
inclusions to around 150 files.
3. The script was run again and the output was compared to the edits
from #2 to make sure no file was left behind.
4. Several build tests were done and a couple of problems were fixed.
e.g. lib/decompress_*.c used malloc/free() wrappers around slab
APIs requiring slab.h to be added manually.
5. The script was run on all .h files but without automatically
editing them as sprinkling gfp.h and slab.h inclusions around .h
files could easily lead to inclusion dependency hell. Most gfp.h
inclusion directives were ignored as stuff from gfp.h was usually
wildly available and often used in preprocessor macros. Each
slab.h inclusion directive was examined and added manually as
necessary.
6. percpu.h was updated not to include slab.h.
7. Build test were done on the following configurations and failures
were fixed. CONFIG_GCOV_KERNEL was turned off for all tests (as my
distributed build env didn't work with gcov compiles) and a few
more options had to be turned off depending on archs to make things
build (like ipr on powerpc/64 which failed due to missing writeq).
* x86 and x86_64 UP and SMP allmodconfig and a custom test config.
* powerpc and powerpc64 SMP allmodconfig
* sparc and sparc64 SMP allmodconfig
* ia64 SMP allmodconfig
* s390 SMP allmodconfig
* alpha SMP allmodconfig
* um on x86_64 SMP allmodconfig
8. percpu.h modifications were reverted so that it could be applied as
a separate patch and serve as bisection point.
Given the fact that I had only a couple of failures from tests on step
6, I'm fairly confident about the coverage of this conversion patch.
If there is a breakage, it's likely to be something in one of the arch
headers which should be easily discoverable easily on most builds of
the specific arch.
Signed-off-by: Tejun Heo <tj@kernel.org>
Guess-its-ok-by: Christoph Lameter <cl@linux-foundation.org>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Lee Schermerhorn <Lee.Schermerhorn@hp.com>
2010-03-24 11:04:11 +03:00
|
|
|
#include <linux/slab.h>
|
2012-10-02 21:01:07 +04:00
|
|
|
#include <drm/drmP.h>
|
2010-01-13 01:54:34 +03:00
|
|
|
#include "radeon.h"
|
2010-03-12 00:19:17 +03:00
|
|
|
#include "radeon_asic.h"
|
2014-12-23 01:26:51 +03:00
|
|
|
#include "radeon_audio.h"
|
2012-10-02 21:01:07 +04:00
|
|
|
#include <drm/radeon_drm.h>
|
2010-03-24 20:20:41 +03:00
|
|
|
#include "evergreend.h"
|
2010-01-13 01:54:34 +03:00
|
|
|
#include "atom.h"
|
|
|
|
#include "avivod.h"
|
|
|
|
#include "evergreen_reg.h"
|
2010-10-21 21:31:38 +04:00
|
|
|
#include "evergreen_blit_shaders.h"
|
2013-01-12 00:33:13 +04:00
|
|
|
#include "radeon_ucode.h"
|
2010-03-24 20:36:43 +03:00
|
|
|
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
#define DC_HPDx_CONTROL(x) (DC_HPD1_CONTROL + (x * 0xc))
|
|
|
|
#define DC_HPDx_INT_CONTROL(x) (DC_HPD1_INT_CONTROL + (x * 0xc))
|
|
|
|
#define DC_HPDx_INT_STATUS_REG(x) (DC_HPD1_INT_STATUS + (x * 0xc))
|
|
|
|
|
radeon: Deinline indirect register accessor functions
This patch deinlines indirect register accessor functions.
These functions perform two mmio accesses, framed by spin lock/unlock.
Spin lock/unlock by itself takes more than 50 cycles in ideal case
(if lock is exclusively cached on current CPU).
With this .config: http://busybox.net/~vda/kernel_config,
after uninlining these functions have sizes and callsite counts
as follows:
r600_uvd_ctx_rreg: 111 bytes, 4 callsites
r600_uvd_ctx_wreg: 113 bytes, 5 callsites
eg_pif_phy0_rreg: 106 bytes, 13 callsites
eg_pif_phy0_wreg: 108 bytes, 13 callsites
eg_pif_phy1_rreg: 107 bytes, 13 callsites
eg_pif_phy1_wreg: 108 bytes, 13 callsites
rv370_pcie_rreg: 111 bytes, 21 callsites
rv370_pcie_wreg: 113 bytes, 24 callsites
r600_rcu_rreg: 111 bytes, 16 callsites
r600_rcu_wreg: 113 bytes, 25 callsites
cik_didt_rreg: 106 bytes, 10 callsites
cik_didt_wreg: 107 bytes, 10 callsites
tn_smc_rreg: 106 bytes, 126 callsites
tn_smc_wreg: 107 bytes, 116 callsites
eg_cg_rreg: 107 bytes, 20 callsites
eg_cg_wreg: 108 bytes, 52 callsites
Functions r100_mm_rreg() and r100_mm_rreg() have a fast path and
a locked (slow) path. This patch deinlines only slow path.
r100_mm_rreg_slow: 78 bytes, 2083 callsites
r100_mm_wreg_slow: 81 bytes, 3570 callsites
Reduction in code size is more than 65,000 bytes:
text data bss dec hex filename
85740176 22294680 20627456 128662312 7ab3b28 vmlinux.before
85674192 22294776 20627456 128598664 7aa4288 vmlinux
Signed-off-by: Denys Vlasenko <dvlasenk@redhat.com>
Cc: Christian König <christian.koenig@amd.com>
Cc: Alex Deucher <alexander.deucher@amd.com>
Cc: linux-kernel@vger.kernel.org
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2015-05-20 14:02:37 +03:00
|
|
|
/*
|
|
|
|
* Indirect registers accessor
|
|
|
|
*/
|
|
|
|
u32 eg_cg_rreg(struct radeon_device *rdev, u32 reg)
|
|
|
|
{
|
|
|
|
unsigned long flags;
|
|
|
|
u32 r;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&rdev->cg_idx_lock, flags);
|
|
|
|
WREG32(EVERGREEN_CG_IND_ADDR, ((reg) & 0xffff));
|
|
|
|
r = RREG32(EVERGREEN_CG_IND_DATA);
|
|
|
|
spin_unlock_irqrestore(&rdev->cg_idx_lock, flags);
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
|
|
|
|
void eg_cg_wreg(struct radeon_device *rdev, u32 reg, u32 v)
|
|
|
|
{
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&rdev->cg_idx_lock, flags);
|
|
|
|
WREG32(EVERGREEN_CG_IND_ADDR, ((reg) & 0xffff));
|
|
|
|
WREG32(EVERGREEN_CG_IND_DATA, (v));
|
|
|
|
spin_unlock_irqrestore(&rdev->cg_idx_lock, flags);
|
|
|
|
}
|
|
|
|
|
|
|
|
u32 eg_pif_phy0_rreg(struct radeon_device *rdev, u32 reg)
|
|
|
|
{
|
|
|
|
unsigned long flags;
|
|
|
|
u32 r;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&rdev->pif_idx_lock, flags);
|
|
|
|
WREG32(EVERGREEN_PIF_PHY0_INDEX, ((reg) & 0xffff));
|
|
|
|
r = RREG32(EVERGREEN_PIF_PHY0_DATA);
|
|
|
|
spin_unlock_irqrestore(&rdev->pif_idx_lock, flags);
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
|
|
|
|
void eg_pif_phy0_wreg(struct radeon_device *rdev, u32 reg, u32 v)
|
|
|
|
{
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&rdev->pif_idx_lock, flags);
|
|
|
|
WREG32(EVERGREEN_PIF_PHY0_INDEX, ((reg) & 0xffff));
|
|
|
|
WREG32(EVERGREEN_PIF_PHY0_DATA, (v));
|
|
|
|
spin_unlock_irqrestore(&rdev->pif_idx_lock, flags);
|
|
|
|
}
|
|
|
|
|
|
|
|
u32 eg_pif_phy1_rreg(struct radeon_device *rdev, u32 reg)
|
|
|
|
{
|
|
|
|
unsigned long flags;
|
|
|
|
u32 r;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&rdev->pif_idx_lock, flags);
|
|
|
|
WREG32(EVERGREEN_PIF_PHY1_INDEX, ((reg) & 0xffff));
|
|
|
|
r = RREG32(EVERGREEN_PIF_PHY1_DATA);
|
|
|
|
spin_unlock_irqrestore(&rdev->pif_idx_lock, flags);
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
|
|
|
|
void eg_pif_phy1_wreg(struct radeon_device *rdev, u32 reg, u32 v)
|
|
|
|
{
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&rdev->pif_idx_lock, flags);
|
|
|
|
WREG32(EVERGREEN_PIF_PHY1_INDEX, ((reg) & 0xffff));
|
|
|
|
WREG32(EVERGREEN_PIF_PHY1_DATA, (v));
|
|
|
|
spin_unlock_irqrestore(&rdev->pif_idx_lock, flags);
|
|
|
|
}
|
|
|
|
|
2012-08-16 01:13:53 +04:00
|
|
|
static const u32 crtc_offsets[6] =
|
|
|
|
{
|
|
|
|
EVERGREEN_CRTC0_REGISTER_OFFSET,
|
|
|
|
EVERGREEN_CRTC1_REGISTER_OFFSET,
|
|
|
|
EVERGREEN_CRTC2_REGISTER_OFFSET,
|
|
|
|
EVERGREEN_CRTC3_REGISTER_OFFSET,
|
|
|
|
EVERGREEN_CRTC4_REGISTER_OFFSET,
|
|
|
|
EVERGREEN_CRTC5_REGISTER_OFFSET
|
|
|
|
};
|
|
|
|
|
2013-04-12 21:52:52 +04:00
|
|
|
#include "clearstate_evergreen.h"
|
|
|
|
|
2013-04-18 01:53:50 +04:00
|
|
|
static const u32 sumo_rlc_save_restore_register_list[] =
|
2013-04-12 21:52:52 +04:00
|
|
|
{
|
|
|
|
0x98fc,
|
|
|
|
0x9830,
|
|
|
|
0x9834,
|
|
|
|
0x9838,
|
|
|
|
0x9870,
|
|
|
|
0x9874,
|
|
|
|
0x8a14,
|
|
|
|
0x8b24,
|
|
|
|
0x8bcc,
|
|
|
|
0x8b10,
|
|
|
|
0x8d00,
|
|
|
|
0x8d04,
|
|
|
|
0x8c00,
|
|
|
|
0x8c04,
|
|
|
|
0x8c08,
|
|
|
|
0x8c0c,
|
|
|
|
0x8d8c,
|
|
|
|
0x8c20,
|
|
|
|
0x8c24,
|
|
|
|
0x8c28,
|
|
|
|
0x8c18,
|
|
|
|
0x8c1c,
|
|
|
|
0x8cf0,
|
|
|
|
0x8e2c,
|
|
|
|
0x8e38,
|
|
|
|
0x8c30,
|
|
|
|
0x9508,
|
|
|
|
0x9688,
|
|
|
|
0x9608,
|
|
|
|
0x960c,
|
|
|
|
0x9610,
|
|
|
|
0x9614,
|
|
|
|
0x88c4,
|
|
|
|
0x88d4,
|
|
|
|
0xa008,
|
|
|
|
0x900c,
|
|
|
|
0x9100,
|
|
|
|
0x913c,
|
|
|
|
0x98f8,
|
|
|
|
0x98f4,
|
|
|
|
0x9b7c,
|
|
|
|
0x3f8c,
|
|
|
|
0x8950,
|
|
|
|
0x8954,
|
|
|
|
0x8a18,
|
|
|
|
0x8b28,
|
|
|
|
0x9144,
|
|
|
|
0x9148,
|
|
|
|
0x914c,
|
|
|
|
0x3f90,
|
|
|
|
0x3f94,
|
|
|
|
0x915c,
|
|
|
|
0x9160,
|
|
|
|
0x9178,
|
|
|
|
0x917c,
|
|
|
|
0x9180,
|
|
|
|
0x918c,
|
|
|
|
0x9190,
|
|
|
|
0x9194,
|
|
|
|
0x9198,
|
|
|
|
0x919c,
|
|
|
|
0x91a8,
|
|
|
|
0x91ac,
|
|
|
|
0x91b0,
|
|
|
|
0x91b4,
|
|
|
|
0x91b8,
|
|
|
|
0x91c4,
|
|
|
|
0x91c8,
|
|
|
|
0x91cc,
|
|
|
|
0x91d0,
|
|
|
|
0x91d4,
|
|
|
|
0x91e0,
|
|
|
|
0x91e4,
|
|
|
|
0x91ec,
|
|
|
|
0x91f0,
|
|
|
|
0x91f4,
|
|
|
|
0x9200,
|
|
|
|
0x9204,
|
|
|
|
0x929c,
|
|
|
|
0x9150,
|
|
|
|
0x802c,
|
|
|
|
};
|
|
|
|
|
2010-01-13 01:54:34 +03:00
|
|
|
static void evergreen_gpu_init(struct radeon_device *rdev);
|
|
|
|
void evergreen_fini(struct radeon_device *rdev);
|
2011-09-20 18:22:58 +04:00
|
|
|
void evergreen_pcie_gen2_enable(struct radeon_device *rdev);
|
2013-02-15 20:02:50 +04:00
|
|
|
void evergreen_program_aspm(struct radeon_device *rdev);
|
2011-11-18 05:13:28 +04:00
|
|
|
extern void cayman_cp_int_cntl_setup(struct radeon_device *rdev,
|
|
|
|
int ring, u32 cp_int_cntl);
|
2013-06-14 02:26:25 +04:00
|
|
|
extern void cayman_vm_decode_fault(struct radeon_device *rdev,
|
|
|
|
u32 status, u32 addr);
|
2013-07-23 17:41:05 +04:00
|
|
|
void cik_init_cp_pg_table(struct radeon_device *rdev);
|
2010-01-13 01:54:34 +03:00
|
|
|
|
2013-08-13 20:48:06 +04:00
|
|
|
extern u32 si_get_csb_size(struct radeon_device *rdev);
|
|
|
|
extern void si_get_csb_buffer(struct radeon_device *rdev, volatile u32 *buffer);
|
2013-08-22 19:57:46 +04:00
|
|
|
extern u32 cik_get_csb_size(struct radeon_device *rdev);
|
|
|
|
extern void cik_get_csb_buffer(struct radeon_device *rdev, volatile u32 *buffer);
|
2013-11-02 00:25:10 +04:00
|
|
|
extern void rv770_set_clk_bypass_mode(struct radeon_device *rdev);
|
2010-01-13 01:54:34 +03:00
|
|
|
|
2013-02-28 23:40:09 +04:00
|
|
|
static const u32 evergreen_golden_registers[] =
|
|
|
|
{
|
|
|
|
0x3f90, 0xffff0000, 0xff000000,
|
|
|
|
0x9148, 0xffff0000, 0xff000000,
|
|
|
|
0x3f94, 0xffff0000, 0xff000000,
|
|
|
|
0x914c, 0xffff0000, 0xff000000,
|
|
|
|
0x9b7c, 0xffffffff, 0x00000000,
|
|
|
|
0x8a14, 0xffffffff, 0x00000007,
|
|
|
|
0x8b10, 0xffffffff, 0x00000000,
|
|
|
|
0x960c, 0xffffffff, 0x54763210,
|
|
|
|
0x88c4, 0xffffffff, 0x000000c2,
|
|
|
|
0x88d4, 0xffffffff, 0x00000010,
|
|
|
|
0x8974, 0xffffffff, 0x00000000,
|
|
|
|
0xc78, 0x00000080, 0x00000080,
|
|
|
|
0x5eb4, 0xffffffff, 0x00000002,
|
|
|
|
0x5e78, 0xffffffff, 0x001000f0,
|
|
|
|
0x6104, 0x01000300, 0x00000000,
|
|
|
|
0x5bc0, 0x00300000, 0x00000000,
|
|
|
|
0x7030, 0xffffffff, 0x00000011,
|
|
|
|
0x7c30, 0xffffffff, 0x00000011,
|
|
|
|
0x10830, 0xffffffff, 0x00000011,
|
|
|
|
0x11430, 0xffffffff, 0x00000011,
|
|
|
|
0x12030, 0xffffffff, 0x00000011,
|
|
|
|
0x12c30, 0xffffffff, 0x00000011,
|
|
|
|
0xd02c, 0xffffffff, 0x08421000,
|
|
|
|
0x240c, 0xffffffff, 0x00000380,
|
|
|
|
0x8b24, 0xffffffff, 0x00ff0fff,
|
|
|
|
0x28a4c, 0x06000000, 0x06000000,
|
|
|
|
0x10c, 0x00000001, 0x00000001,
|
|
|
|
0x8d00, 0xffffffff, 0x100e4848,
|
|
|
|
0x8d04, 0xffffffff, 0x00164745,
|
|
|
|
0x8c00, 0xffffffff, 0xe4000003,
|
|
|
|
0x8c04, 0xffffffff, 0x40600060,
|
|
|
|
0x8c08, 0xffffffff, 0x001c001c,
|
|
|
|
0x8cf0, 0xffffffff, 0x08e00620,
|
|
|
|
0x8c20, 0xffffffff, 0x00800080,
|
|
|
|
0x8c24, 0xffffffff, 0x00800080,
|
|
|
|
0x8c18, 0xffffffff, 0x20202078,
|
|
|
|
0x8c1c, 0xffffffff, 0x00001010,
|
|
|
|
0x28350, 0xffffffff, 0x00000000,
|
|
|
|
0xa008, 0xffffffff, 0x00010000,
|
2014-07-08 01:59:37 +04:00
|
|
|
0x5c4, 0xffffffff, 0x00000001,
|
2013-02-28 23:40:09 +04:00
|
|
|
0x9508, 0xffffffff, 0x00000002,
|
|
|
|
0x913c, 0x0000000f, 0x0000000a
|
|
|
|
};
|
|
|
|
|
|
|
|
static const u32 evergreen_golden_registers2[] =
|
|
|
|
{
|
|
|
|
0x2f4c, 0xffffffff, 0x00000000,
|
|
|
|
0x54f4, 0xffffffff, 0x00000000,
|
|
|
|
0x54f0, 0xffffffff, 0x00000000,
|
|
|
|
0x5498, 0xffffffff, 0x00000000,
|
|
|
|
0x549c, 0xffffffff, 0x00000000,
|
|
|
|
0x5494, 0xffffffff, 0x00000000,
|
|
|
|
0x53cc, 0xffffffff, 0x00000000,
|
|
|
|
0x53c8, 0xffffffff, 0x00000000,
|
|
|
|
0x53c4, 0xffffffff, 0x00000000,
|
|
|
|
0x53c0, 0xffffffff, 0x00000000,
|
|
|
|
0x53bc, 0xffffffff, 0x00000000,
|
|
|
|
0x53b8, 0xffffffff, 0x00000000,
|
|
|
|
0x53b4, 0xffffffff, 0x00000000,
|
|
|
|
0x53b0, 0xffffffff, 0x00000000
|
|
|
|
};
|
|
|
|
|
|
|
|
static const u32 cypress_mgcg_init[] =
|
|
|
|
{
|
|
|
|
0x802c, 0xffffffff, 0xc0000000,
|
|
|
|
0x5448, 0xffffffff, 0x00000100,
|
|
|
|
0x55e4, 0xffffffff, 0x00000100,
|
|
|
|
0x160c, 0xffffffff, 0x00000100,
|
|
|
|
0x5644, 0xffffffff, 0x00000100,
|
|
|
|
0xc164, 0xffffffff, 0x00000100,
|
|
|
|
0x8a18, 0xffffffff, 0x00000100,
|
|
|
|
0x897c, 0xffffffff, 0x06000100,
|
|
|
|
0x8b28, 0xffffffff, 0x00000100,
|
|
|
|
0x9144, 0xffffffff, 0x00000100,
|
|
|
|
0x9a60, 0xffffffff, 0x00000100,
|
|
|
|
0x9868, 0xffffffff, 0x00000100,
|
|
|
|
0x8d58, 0xffffffff, 0x00000100,
|
|
|
|
0x9510, 0xffffffff, 0x00000100,
|
|
|
|
0x949c, 0xffffffff, 0x00000100,
|
|
|
|
0x9654, 0xffffffff, 0x00000100,
|
|
|
|
0x9030, 0xffffffff, 0x00000100,
|
|
|
|
0x9034, 0xffffffff, 0x00000100,
|
|
|
|
0x9038, 0xffffffff, 0x00000100,
|
|
|
|
0x903c, 0xffffffff, 0x00000100,
|
|
|
|
0x9040, 0xffffffff, 0x00000100,
|
|
|
|
0xa200, 0xffffffff, 0x00000100,
|
|
|
|
0xa204, 0xffffffff, 0x00000100,
|
|
|
|
0xa208, 0xffffffff, 0x00000100,
|
|
|
|
0xa20c, 0xffffffff, 0x00000100,
|
|
|
|
0x971c, 0xffffffff, 0x00000100,
|
|
|
|
0x977c, 0xffffffff, 0x00000100,
|
|
|
|
0x3f80, 0xffffffff, 0x00000100,
|
|
|
|
0xa210, 0xffffffff, 0x00000100,
|
|
|
|
0xa214, 0xffffffff, 0x00000100,
|
|
|
|
0x4d8, 0xffffffff, 0x00000100,
|
|
|
|
0x9784, 0xffffffff, 0x00000100,
|
|
|
|
0x9698, 0xffffffff, 0x00000100,
|
|
|
|
0x4d4, 0xffffffff, 0x00000200,
|
|
|
|
0x30cc, 0xffffffff, 0x00000100,
|
|
|
|
0xd0c0, 0xffffffff, 0xff000100,
|
|
|
|
0x802c, 0xffffffff, 0x40000000,
|
|
|
|
0x915c, 0xffffffff, 0x00010000,
|
|
|
|
0x9160, 0xffffffff, 0x00030002,
|
|
|
|
0x9178, 0xffffffff, 0x00070000,
|
|
|
|
0x917c, 0xffffffff, 0x00030002,
|
|
|
|
0x9180, 0xffffffff, 0x00050004,
|
|
|
|
0x918c, 0xffffffff, 0x00010006,
|
|
|
|
0x9190, 0xffffffff, 0x00090008,
|
|
|
|
0x9194, 0xffffffff, 0x00070000,
|
|
|
|
0x9198, 0xffffffff, 0x00030002,
|
|
|
|
0x919c, 0xffffffff, 0x00050004,
|
|
|
|
0x91a8, 0xffffffff, 0x00010006,
|
|
|
|
0x91ac, 0xffffffff, 0x00090008,
|
|
|
|
0x91b0, 0xffffffff, 0x00070000,
|
|
|
|
0x91b4, 0xffffffff, 0x00030002,
|
|
|
|
0x91b8, 0xffffffff, 0x00050004,
|
|
|
|
0x91c4, 0xffffffff, 0x00010006,
|
|
|
|
0x91c8, 0xffffffff, 0x00090008,
|
|
|
|
0x91cc, 0xffffffff, 0x00070000,
|
|
|
|
0x91d0, 0xffffffff, 0x00030002,
|
|
|
|
0x91d4, 0xffffffff, 0x00050004,
|
|
|
|
0x91e0, 0xffffffff, 0x00010006,
|
|
|
|
0x91e4, 0xffffffff, 0x00090008,
|
|
|
|
0x91e8, 0xffffffff, 0x00000000,
|
|
|
|
0x91ec, 0xffffffff, 0x00070000,
|
|
|
|
0x91f0, 0xffffffff, 0x00030002,
|
|
|
|
0x91f4, 0xffffffff, 0x00050004,
|
|
|
|
0x9200, 0xffffffff, 0x00010006,
|
|
|
|
0x9204, 0xffffffff, 0x00090008,
|
|
|
|
0x9208, 0xffffffff, 0x00070000,
|
|
|
|
0x920c, 0xffffffff, 0x00030002,
|
|
|
|
0x9210, 0xffffffff, 0x00050004,
|
|
|
|
0x921c, 0xffffffff, 0x00010006,
|
|
|
|
0x9220, 0xffffffff, 0x00090008,
|
|
|
|
0x9224, 0xffffffff, 0x00070000,
|
|
|
|
0x9228, 0xffffffff, 0x00030002,
|
|
|
|
0x922c, 0xffffffff, 0x00050004,
|
|
|
|
0x9238, 0xffffffff, 0x00010006,
|
|
|
|
0x923c, 0xffffffff, 0x00090008,
|
|
|
|
0x9240, 0xffffffff, 0x00070000,
|
|
|
|
0x9244, 0xffffffff, 0x00030002,
|
|
|
|
0x9248, 0xffffffff, 0x00050004,
|
|
|
|
0x9254, 0xffffffff, 0x00010006,
|
|
|
|
0x9258, 0xffffffff, 0x00090008,
|
|
|
|
0x925c, 0xffffffff, 0x00070000,
|
|
|
|
0x9260, 0xffffffff, 0x00030002,
|
|
|
|
0x9264, 0xffffffff, 0x00050004,
|
|
|
|
0x9270, 0xffffffff, 0x00010006,
|
|
|
|
0x9274, 0xffffffff, 0x00090008,
|
|
|
|
0x9278, 0xffffffff, 0x00070000,
|
|
|
|
0x927c, 0xffffffff, 0x00030002,
|
|
|
|
0x9280, 0xffffffff, 0x00050004,
|
|
|
|
0x928c, 0xffffffff, 0x00010006,
|
|
|
|
0x9290, 0xffffffff, 0x00090008,
|
|
|
|
0x9294, 0xffffffff, 0x00000000,
|
|
|
|
0x929c, 0xffffffff, 0x00000001,
|
|
|
|
0x802c, 0xffffffff, 0x40010000,
|
|
|
|
0x915c, 0xffffffff, 0x00010000,
|
|
|
|
0x9160, 0xffffffff, 0x00030002,
|
|
|
|
0x9178, 0xffffffff, 0x00070000,
|
|
|
|
0x917c, 0xffffffff, 0x00030002,
|
|
|
|
0x9180, 0xffffffff, 0x00050004,
|
|
|
|
0x918c, 0xffffffff, 0x00010006,
|
|
|
|
0x9190, 0xffffffff, 0x00090008,
|
|
|
|
0x9194, 0xffffffff, 0x00070000,
|
|
|
|
0x9198, 0xffffffff, 0x00030002,
|
|
|
|
0x919c, 0xffffffff, 0x00050004,
|
|
|
|
0x91a8, 0xffffffff, 0x00010006,
|
|
|
|
0x91ac, 0xffffffff, 0x00090008,
|
|
|
|
0x91b0, 0xffffffff, 0x00070000,
|
|
|
|
0x91b4, 0xffffffff, 0x00030002,
|
|
|
|
0x91b8, 0xffffffff, 0x00050004,
|
|
|
|
0x91c4, 0xffffffff, 0x00010006,
|
|
|
|
0x91c8, 0xffffffff, 0x00090008,
|
|
|
|
0x91cc, 0xffffffff, 0x00070000,
|
|
|
|
0x91d0, 0xffffffff, 0x00030002,
|
|
|
|
0x91d4, 0xffffffff, 0x00050004,
|
|
|
|
0x91e0, 0xffffffff, 0x00010006,
|
|
|
|
0x91e4, 0xffffffff, 0x00090008,
|
|
|
|
0x91e8, 0xffffffff, 0x00000000,
|
|
|
|
0x91ec, 0xffffffff, 0x00070000,
|
|
|
|
0x91f0, 0xffffffff, 0x00030002,
|
|
|
|
0x91f4, 0xffffffff, 0x00050004,
|
|
|
|
0x9200, 0xffffffff, 0x00010006,
|
|
|
|
0x9204, 0xffffffff, 0x00090008,
|
|
|
|
0x9208, 0xffffffff, 0x00070000,
|
|
|
|
0x920c, 0xffffffff, 0x00030002,
|
|
|
|
0x9210, 0xffffffff, 0x00050004,
|
|
|
|
0x921c, 0xffffffff, 0x00010006,
|
|
|
|
0x9220, 0xffffffff, 0x00090008,
|
|
|
|
0x9224, 0xffffffff, 0x00070000,
|
|
|
|
0x9228, 0xffffffff, 0x00030002,
|
|
|
|
0x922c, 0xffffffff, 0x00050004,
|
|
|
|
0x9238, 0xffffffff, 0x00010006,
|
|
|
|
0x923c, 0xffffffff, 0x00090008,
|
|
|
|
0x9240, 0xffffffff, 0x00070000,
|
|
|
|
0x9244, 0xffffffff, 0x00030002,
|
|
|
|
0x9248, 0xffffffff, 0x00050004,
|
|
|
|
0x9254, 0xffffffff, 0x00010006,
|
|
|
|
0x9258, 0xffffffff, 0x00090008,
|
|
|
|
0x925c, 0xffffffff, 0x00070000,
|
|
|
|
0x9260, 0xffffffff, 0x00030002,
|
|
|
|
0x9264, 0xffffffff, 0x00050004,
|
|
|
|
0x9270, 0xffffffff, 0x00010006,
|
|
|
|
0x9274, 0xffffffff, 0x00090008,
|
|
|
|
0x9278, 0xffffffff, 0x00070000,
|
|
|
|
0x927c, 0xffffffff, 0x00030002,
|
|
|
|
0x9280, 0xffffffff, 0x00050004,
|
|
|
|
0x928c, 0xffffffff, 0x00010006,
|
|
|
|
0x9290, 0xffffffff, 0x00090008,
|
|
|
|
0x9294, 0xffffffff, 0x00000000,
|
|
|
|
0x929c, 0xffffffff, 0x00000001,
|
|
|
|
0x802c, 0xffffffff, 0xc0000000
|
|
|
|
};
|
|
|
|
|
|
|
|
static const u32 redwood_mgcg_init[] =
|
|
|
|
{
|
|
|
|
0x802c, 0xffffffff, 0xc0000000,
|
|
|
|
0x5448, 0xffffffff, 0x00000100,
|
|
|
|
0x55e4, 0xffffffff, 0x00000100,
|
|
|
|
0x160c, 0xffffffff, 0x00000100,
|
|
|
|
0x5644, 0xffffffff, 0x00000100,
|
|
|
|
0xc164, 0xffffffff, 0x00000100,
|
|
|
|
0x8a18, 0xffffffff, 0x00000100,
|
|
|
|
0x897c, 0xffffffff, 0x06000100,
|
|
|
|
0x8b28, 0xffffffff, 0x00000100,
|
|
|
|
0x9144, 0xffffffff, 0x00000100,
|
|
|
|
0x9a60, 0xffffffff, 0x00000100,
|
|
|
|
0x9868, 0xffffffff, 0x00000100,
|
|
|
|
0x8d58, 0xffffffff, 0x00000100,
|
|
|
|
0x9510, 0xffffffff, 0x00000100,
|
|
|
|
0x949c, 0xffffffff, 0x00000100,
|
|
|
|
0x9654, 0xffffffff, 0x00000100,
|
|
|
|
0x9030, 0xffffffff, 0x00000100,
|
|
|
|
0x9034, 0xffffffff, 0x00000100,
|
|
|
|
0x9038, 0xffffffff, 0x00000100,
|
|
|
|
0x903c, 0xffffffff, 0x00000100,
|
|
|
|
0x9040, 0xffffffff, 0x00000100,
|
|
|
|
0xa200, 0xffffffff, 0x00000100,
|
|
|
|
0xa204, 0xffffffff, 0x00000100,
|
|
|
|
0xa208, 0xffffffff, 0x00000100,
|
|
|
|
0xa20c, 0xffffffff, 0x00000100,
|
|
|
|
0x971c, 0xffffffff, 0x00000100,
|
|
|
|
0x977c, 0xffffffff, 0x00000100,
|
|
|
|
0x3f80, 0xffffffff, 0x00000100,
|
|
|
|
0xa210, 0xffffffff, 0x00000100,
|
|
|
|
0xa214, 0xffffffff, 0x00000100,
|
|
|
|
0x4d8, 0xffffffff, 0x00000100,
|
|
|
|
0x9784, 0xffffffff, 0x00000100,
|
|
|
|
0x9698, 0xffffffff, 0x00000100,
|
|
|
|
0x4d4, 0xffffffff, 0x00000200,
|
|
|
|
0x30cc, 0xffffffff, 0x00000100,
|
|
|
|
0xd0c0, 0xffffffff, 0xff000100,
|
|
|
|
0x802c, 0xffffffff, 0x40000000,
|
|
|
|
0x915c, 0xffffffff, 0x00010000,
|
|
|
|
0x9160, 0xffffffff, 0x00030002,
|
|
|
|
0x9178, 0xffffffff, 0x00070000,
|
|
|
|
0x917c, 0xffffffff, 0x00030002,
|
|
|
|
0x9180, 0xffffffff, 0x00050004,
|
|
|
|
0x918c, 0xffffffff, 0x00010006,
|
|
|
|
0x9190, 0xffffffff, 0x00090008,
|
|
|
|
0x9194, 0xffffffff, 0x00070000,
|
|
|
|
0x9198, 0xffffffff, 0x00030002,
|
|
|
|
0x919c, 0xffffffff, 0x00050004,
|
|
|
|
0x91a8, 0xffffffff, 0x00010006,
|
|
|
|
0x91ac, 0xffffffff, 0x00090008,
|
|
|
|
0x91b0, 0xffffffff, 0x00070000,
|
|
|
|
0x91b4, 0xffffffff, 0x00030002,
|
|
|
|
0x91b8, 0xffffffff, 0x00050004,
|
|
|
|
0x91c4, 0xffffffff, 0x00010006,
|
|
|
|
0x91c8, 0xffffffff, 0x00090008,
|
|
|
|
0x91cc, 0xffffffff, 0x00070000,
|
|
|
|
0x91d0, 0xffffffff, 0x00030002,
|
|
|
|
0x91d4, 0xffffffff, 0x00050004,
|
|
|
|
0x91e0, 0xffffffff, 0x00010006,
|
|
|
|
0x91e4, 0xffffffff, 0x00090008,
|
|
|
|
0x91e8, 0xffffffff, 0x00000000,
|
|
|
|
0x91ec, 0xffffffff, 0x00070000,
|
|
|
|
0x91f0, 0xffffffff, 0x00030002,
|
|
|
|
0x91f4, 0xffffffff, 0x00050004,
|
|
|
|
0x9200, 0xffffffff, 0x00010006,
|
|
|
|
0x9204, 0xffffffff, 0x00090008,
|
|
|
|
0x9294, 0xffffffff, 0x00000000,
|
|
|
|
0x929c, 0xffffffff, 0x00000001,
|
|
|
|
0x802c, 0xffffffff, 0xc0000000
|
|
|
|
};
|
|
|
|
|
|
|
|
static const u32 cedar_golden_registers[] =
|
|
|
|
{
|
|
|
|
0x3f90, 0xffff0000, 0xff000000,
|
|
|
|
0x9148, 0xffff0000, 0xff000000,
|
|
|
|
0x3f94, 0xffff0000, 0xff000000,
|
|
|
|
0x914c, 0xffff0000, 0xff000000,
|
|
|
|
0x9b7c, 0xffffffff, 0x00000000,
|
|
|
|
0x8a14, 0xffffffff, 0x00000007,
|
|
|
|
0x8b10, 0xffffffff, 0x00000000,
|
|
|
|
0x960c, 0xffffffff, 0x54763210,
|
|
|
|
0x88c4, 0xffffffff, 0x000000c2,
|
|
|
|
0x88d4, 0xffffffff, 0x00000000,
|
|
|
|
0x8974, 0xffffffff, 0x00000000,
|
|
|
|
0xc78, 0x00000080, 0x00000080,
|
|
|
|
0x5eb4, 0xffffffff, 0x00000002,
|
|
|
|
0x5e78, 0xffffffff, 0x001000f0,
|
|
|
|
0x6104, 0x01000300, 0x00000000,
|
|
|
|
0x5bc0, 0x00300000, 0x00000000,
|
|
|
|
0x7030, 0xffffffff, 0x00000011,
|
|
|
|
0x7c30, 0xffffffff, 0x00000011,
|
|
|
|
0x10830, 0xffffffff, 0x00000011,
|
|
|
|
0x11430, 0xffffffff, 0x00000011,
|
|
|
|
0xd02c, 0xffffffff, 0x08421000,
|
|
|
|
0x240c, 0xffffffff, 0x00000380,
|
|
|
|
0x8b24, 0xffffffff, 0x00ff0fff,
|
|
|
|
0x28a4c, 0x06000000, 0x06000000,
|
|
|
|
0x10c, 0x00000001, 0x00000001,
|
|
|
|
0x8d00, 0xffffffff, 0x100e4848,
|
|
|
|
0x8d04, 0xffffffff, 0x00164745,
|
|
|
|
0x8c00, 0xffffffff, 0xe4000003,
|
|
|
|
0x8c04, 0xffffffff, 0x40600060,
|
|
|
|
0x8c08, 0xffffffff, 0x001c001c,
|
|
|
|
0x8cf0, 0xffffffff, 0x08e00410,
|
|
|
|
0x8c20, 0xffffffff, 0x00800080,
|
|
|
|
0x8c24, 0xffffffff, 0x00800080,
|
|
|
|
0x8c18, 0xffffffff, 0x20202078,
|
|
|
|
0x8c1c, 0xffffffff, 0x00001010,
|
|
|
|
0x28350, 0xffffffff, 0x00000000,
|
|
|
|
0xa008, 0xffffffff, 0x00010000,
|
2014-07-08 01:59:37 +04:00
|
|
|
0x5c4, 0xffffffff, 0x00000001,
|
2013-02-28 23:40:09 +04:00
|
|
|
0x9508, 0xffffffff, 0x00000002
|
|
|
|
};
|
|
|
|
|
|
|
|
static const u32 cedar_mgcg_init[] =
|
|
|
|
{
|
|
|
|
0x802c, 0xffffffff, 0xc0000000,
|
|
|
|
0x5448, 0xffffffff, 0x00000100,
|
|
|
|
0x55e4, 0xffffffff, 0x00000100,
|
|
|
|
0x160c, 0xffffffff, 0x00000100,
|
|
|
|
0x5644, 0xffffffff, 0x00000100,
|
|
|
|
0xc164, 0xffffffff, 0x00000100,
|
|
|
|
0x8a18, 0xffffffff, 0x00000100,
|
|
|
|
0x897c, 0xffffffff, 0x06000100,
|
|
|
|
0x8b28, 0xffffffff, 0x00000100,
|
|
|
|
0x9144, 0xffffffff, 0x00000100,
|
|
|
|
0x9a60, 0xffffffff, 0x00000100,
|
|
|
|
0x9868, 0xffffffff, 0x00000100,
|
|
|
|
0x8d58, 0xffffffff, 0x00000100,
|
|
|
|
0x9510, 0xffffffff, 0x00000100,
|
|
|
|
0x949c, 0xffffffff, 0x00000100,
|
|
|
|
0x9654, 0xffffffff, 0x00000100,
|
|
|
|
0x9030, 0xffffffff, 0x00000100,
|
|
|
|
0x9034, 0xffffffff, 0x00000100,
|
|
|
|
0x9038, 0xffffffff, 0x00000100,
|
|
|
|
0x903c, 0xffffffff, 0x00000100,
|
|
|
|
0x9040, 0xffffffff, 0x00000100,
|
|
|
|
0xa200, 0xffffffff, 0x00000100,
|
|
|
|
0xa204, 0xffffffff, 0x00000100,
|
|
|
|
0xa208, 0xffffffff, 0x00000100,
|
|
|
|
0xa20c, 0xffffffff, 0x00000100,
|
|
|
|
0x971c, 0xffffffff, 0x00000100,
|
|
|
|
0x977c, 0xffffffff, 0x00000100,
|
|
|
|
0x3f80, 0xffffffff, 0x00000100,
|
|
|
|
0xa210, 0xffffffff, 0x00000100,
|
|
|
|
0xa214, 0xffffffff, 0x00000100,
|
|
|
|
0x4d8, 0xffffffff, 0x00000100,
|
|
|
|
0x9784, 0xffffffff, 0x00000100,
|
|
|
|
0x9698, 0xffffffff, 0x00000100,
|
|
|
|
0x4d4, 0xffffffff, 0x00000200,
|
|
|
|
0x30cc, 0xffffffff, 0x00000100,
|
|
|
|
0xd0c0, 0xffffffff, 0xff000100,
|
|
|
|
0x802c, 0xffffffff, 0x40000000,
|
|
|
|
0x915c, 0xffffffff, 0x00010000,
|
|
|
|
0x9178, 0xffffffff, 0x00050000,
|
|
|
|
0x917c, 0xffffffff, 0x00030002,
|
|
|
|
0x918c, 0xffffffff, 0x00010004,
|
|
|
|
0x9190, 0xffffffff, 0x00070006,
|
|
|
|
0x9194, 0xffffffff, 0x00050000,
|
|
|
|
0x9198, 0xffffffff, 0x00030002,
|
|
|
|
0x91a8, 0xffffffff, 0x00010004,
|
|
|
|
0x91ac, 0xffffffff, 0x00070006,
|
|
|
|
0x91e8, 0xffffffff, 0x00000000,
|
|
|
|
0x9294, 0xffffffff, 0x00000000,
|
|
|
|
0x929c, 0xffffffff, 0x00000001,
|
|
|
|
0x802c, 0xffffffff, 0xc0000000
|
|
|
|
};
|
|
|
|
|
|
|
|
static const u32 juniper_mgcg_init[] =
|
|
|
|
{
|
|
|
|
0x802c, 0xffffffff, 0xc0000000,
|
|
|
|
0x5448, 0xffffffff, 0x00000100,
|
|
|
|
0x55e4, 0xffffffff, 0x00000100,
|
|
|
|
0x160c, 0xffffffff, 0x00000100,
|
|
|
|
0x5644, 0xffffffff, 0x00000100,
|
|
|
|
0xc164, 0xffffffff, 0x00000100,
|
|
|
|
0x8a18, 0xffffffff, 0x00000100,
|
|
|
|
0x897c, 0xffffffff, 0x06000100,
|
|
|
|
0x8b28, 0xffffffff, 0x00000100,
|
|
|
|
0x9144, 0xffffffff, 0x00000100,
|
|
|
|
0x9a60, 0xffffffff, 0x00000100,
|
|
|
|
0x9868, 0xffffffff, 0x00000100,
|
|
|
|
0x8d58, 0xffffffff, 0x00000100,
|
|
|
|
0x9510, 0xffffffff, 0x00000100,
|
|
|
|
0x949c, 0xffffffff, 0x00000100,
|
|
|
|
0x9654, 0xffffffff, 0x00000100,
|
|
|
|
0x9030, 0xffffffff, 0x00000100,
|
|
|
|
0x9034, 0xffffffff, 0x00000100,
|
|
|
|
0x9038, 0xffffffff, 0x00000100,
|
|
|
|
0x903c, 0xffffffff, 0x00000100,
|
|
|
|
0x9040, 0xffffffff, 0x00000100,
|
|
|
|
0xa200, 0xffffffff, 0x00000100,
|
|
|
|
0xa204, 0xffffffff, 0x00000100,
|
|
|
|
0xa208, 0xffffffff, 0x00000100,
|
|
|
|
0xa20c, 0xffffffff, 0x00000100,
|
|
|
|
0x971c, 0xffffffff, 0x00000100,
|
|
|
|
0xd0c0, 0xffffffff, 0xff000100,
|
|
|
|
0x802c, 0xffffffff, 0x40000000,
|
|
|
|
0x915c, 0xffffffff, 0x00010000,
|
|
|
|
0x9160, 0xffffffff, 0x00030002,
|
|
|
|
0x9178, 0xffffffff, 0x00070000,
|
|
|
|
0x917c, 0xffffffff, 0x00030002,
|
|
|
|
0x9180, 0xffffffff, 0x00050004,
|
|
|
|
0x918c, 0xffffffff, 0x00010006,
|
|
|
|
0x9190, 0xffffffff, 0x00090008,
|
|
|
|
0x9194, 0xffffffff, 0x00070000,
|
|
|
|
0x9198, 0xffffffff, 0x00030002,
|
|
|
|
0x919c, 0xffffffff, 0x00050004,
|
|
|
|
0x91a8, 0xffffffff, 0x00010006,
|
|
|
|
0x91ac, 0xffffffff, 0x00090008,
|
|
|
|
0x91b0, 0xffffffff, 0x00070000,
|
|
|
|
0x91b4, 0xffffffff, 0x00030002,
|
|
|
|
0x91b8, 0xffffffff, 0x00050004,
|
|
|
|
0x91c4, 0xffffffff, 0x00010006,
|
|
|
|
0x91c8, 0xffffffff, 0x00090008,
|
|
|
|
0x91cc, 0xffffffff, 0x00070000,
|
|
|
|
0x91d0, 0xffffffff, 0x00030002,
|
|
|
|
0x91d4, 0xffffffff, 0x00050004,
|
|
|
|
0x91e0, 0xffffffff, 0x00010006,
|
|
|
|
0x91e4, 0xffffffff, 0x00090008,
|
|
|
|
0x91e8, 0xffffffff, 0x00000000,
|
|
|
|
0x91ec, 0xffffffff, 0x00070000,
|
|
|
|
0x91f0, 0xffffffff, 0x00030002,
|
|
|
|
0x91f4, 0xffffffff, 0x00050004,
|
|
|
|
0x9200, 0xffffffff, 0x00010006,
|
|
|
|
0x9204, 0xffffffff, 0x00090008,
|
|
|
|
0x9208, 0xffffffff, 0x00070000,
|
|
|
|
0x920c, 0xffffffff, 0x00030002,
|
|
|
|
0x9210, 0xffffffff, 0x00050004,
|
|
|
|
0x921c, 0xffffffff, 0x00010006,
|
|
|
|
0x9220, 0xffffffff, 0x00090008,
|
|
|
|
0x9224, 0xffffffff, 0x00070000,
|
|
|
|
0x9228, 0xffffffff, 0x00030002,
|
|
|
|
0x922c, 0xffffffff, 0x00050004,
|
|
|
|
0x9238, 0xffffffff, 0x00010006,
|
|
|
|
0x923c, 0xffffffff, 0x00090008,
|
|
|
|
0x9240, 0xffffffff, 0x00070000,
|
|
|
|
0x9244, 0xffffffff, 0x00030002,
|
|
|
|
0x9248, 0xffffffff, 0x00050004,
|
|
|
|
0x9254, 0xffffffff, 0x00010006,
|
|
|
|
0x9258, 0xffffffff, 0x00090008,
|
|
|
|
0x925c, 0xffffffff, 0x00070000,
|
|
|
|
0x9260, 0xffffffff, 0x00030002,
|
|
|
|
0x9264, 0xffffffff, 0x00050004,
|
|
|
|
0x9270, 0xffffffff, 0x00010006,
|
|
|
|
0x9274, 0xffffffff, 0x00090008,
|
|
|
|
0x9278, 0xffffffff, 0x00070000,
|
|
|
|
0x927c, 0xffffffff, 0x00030002,
|
|
|
|
0x9280, 0xffffffff, 0x00050004,
|
|
|
|
0x928c, 0xffffffff, 0x00010006,
|
|
|
|
0x9290, 0xffffffff, 0x00090008,
|
|
|
|
0x9294, 0xffffffff, 0x00000000,
|
|
|
|
0x929c, 0xffffffff, 0x00000001,
|
|
|
|
0x802c, 0xffffffff, 0xc0000000,
|
|
|
|
0x977c, 0xffffffff, 0x00000100,
|
|
|
|
0x3f80, 0xffffffff, 0x00000100,
|
|
|
|
0xa210, 0xffffffff, 0x00000100,
|
|
|
|
0xa214, 0xffffffff, 0x00000100,
|
|
|
|
0x4d8, 0xffffffff, 0x00000100,
|
|
|
|
0x9784, 0xffffffff, 0x00000100,
|
|
|
|
0x9698, 0xffffffff, 0x00000100,
|
|
|
|
0x4d4, 0xffffffff, 0x00000200,
|
|
|
|
0x30cc, 0xffffffff, 0x00000100,
|
|
|
|
0x802c, 0xffffffff, 0xc0000000
|
|
|
|
};
|
|
|
|
|
|
|
|
static const u32 supersumo_golden_registers[] =
|
|
|
|
{
|
|
|
|
0x5eb4, 0xffffffff, 0x00000002,
|
2014-07-08 01:59:37 +04:00
|
|
|
0x5c4, 0xffffffff, 0x00000001,
|
2013-02-28 23:40:09 +04:00
|
|
|
0x7030, 0xffffffff, 0x00000011,
|
|
|
|
0x7c30, 0xffffffff, 0x00000011,
|
|
|
|
0x6104, 0x01000300, 0x00000000,
|
|
|
|
0x5bc0, 0x00300000, 0x00000000,
|
|
|
|
0x8c04, 0xffffffff, 0x40600060,
|
|
|
|
0x8c08, 0xffffffff, 0x001c001c,
|
|
|
|
0x8c20, 0xffffffff, 0x00800080,
|
|
|
|
0x8c24, 0xffffffff, 0x00800080,
|
|
|
|
0x8c18, 0xffffffff, 0x20202078,
|
|
|
|
0x8c1c, 0xffffffff, 0x00001010,
|
|
|
|
0x918c, 0xffffffff, 0x00010006,
|
|
|
|
0x91a8, 0xffffffff, 0x00010006,
|
|
|
|
0x91c4, 0xffffffff, 0x00010006,
|
|
|
|
0x91e0, 0xffffffff, 0x00010006,
|
|
|
|
0x9200, 0xffffffff, 0x00010006,
|
|
|
|
0x9150, 0xffffffff, 0x6e944040,
|
|
|
|
0x917c, 0xffffffff, 0x00030002,
|
|
|
|
0x9180, 0xffffffff, 0x00050004,
|
|
|
|
0x9198, 0xffffffff, 0x00030002,
|
|
|
|
0x919c, 0xffffffff, 0x00050004,
|
|
|
|
0x91b4, 0xffffffff, 0x00030002,
|
|
|
|
0x91b8, 0xffffffff, 0x00050004,
|
|
|
|
0x91d0, 0xffffffff, 0x00030002,
|
|
|
|
0x91d4, 0xffffffff, 0x00050004,
|
|
|
|
0x91f0, 0xffffffff, 0x00030002,
|
|
|
|
0x91f4, 0xffffffff, 0x00050004,
|
|
|
|
0x915c, 0xffffffff, 0x00010000,
|
|
|
|
0x9160, 0xffffffff, 0x00030002,
|
|
|
|
0x3f90, 0xffff0000, 0xff000000,
|
|
|
|
0x9178, 0xffffffff, 0x00070000,
|
|
|
|
0x9194, 0xffffffff, 0x00070000,
|
|
|
|
0x91b0, 0xffffffff, 0x00070000,
|
|
|
|
0x91cc, 0xffffffff, 0x00070000,
|
|
|
|
0x91ec, 0xffffffff, 0x00070000,
|
|
|
|
0x9148, 0xffff0000, 0xff000000,
|
|
|
|
0x9190, 0xffffffff, 0x00090008,
|
|
|
|
0x91ac, 0xffffffff, 0x00090008,
|
|
|
|
0x91c8, 0xffffffff, 0x00090008,
|
|
|
|
0x91e4, 0xffffffff, 0x00090008,
|
|
|
|
0x9204, 0xffffffff, 0x00090008,
|
|
|
|
0x3f94, 0xffff0000, 0xff000000,
|
|
|
|
0x914c, 0xffff0000, 0xff000000,
|
|
|
|
0x929c, 0xffffffff, 0x00000001,
|
|
|
|
0x8a18, 0xffffffff, 0x00000100,
|
|
|
|
0x8b28, 0xffffffff, 0x00000100,
|
|
|
|
0x9144, 0xffffffff, 0x00000100,
|
|
|
|
0x5644, 0xffffffff, 0x00000100,
|
|
|
|
0x9b7c, 0xffffffff, 0x00000000,
|
|
|
|
0x8030, 0xffffffff, 0x0000100a,
|
|
|
|
0x8a14, 0xffffffff, 0x00000007,
|
|
|
|
0x8b24, 0xffffffff, 0x00ff0fff,
|
|
|
|
0x8b10, 0xffffffff, 0x00000000,
|
|
|
|
0x28a4c, 0x06000000, 0x06000000,
|
|
|
|
0x4d8, 0xffffffff, 0x00000100,
|
|
|
|
0x913c, 0xffff000f, 0x0100000a,
|
|
|
|
0x960c, 0xffffffff, 0x54763210,
|
|
|
|
0x88c4, 0xffffffff, 0x000000c2,
|
|
|
|
0x88d4, 0xffffffff, 0x00000010,
|
|
|
|
0x8974, 0xffffffff, 0x00000000,
|
|
|
|
0xc78, 0x00000080, 0x00000080,
|
|
|
|
0x5e78, 0xffffffff, 0x001000f0,
|
|
|
|
0xd02c, 0xffffffff, 0x08421000,
|
|
|
|
0xa008, 0xffffffff, 0x00010000,
|
|
|
|
0x8d00, 0xffffffff, 0x100e4848,
|
|
|
|
0x8d04, 0xffffffff, 0x00164745,
|
|
|
|
0x8c00, 0xffffffff, 0xe4000003,
|
|
|
|
0x8cf0, 0x1fffffff, 0x08e00620,
|
|
|
|
0x28350, 0xffffffff, 0x00000000,
|
|
|
|
0x9508, 0xffffffff, 0x00000002
|
|
|
|
};
|
|
|
|
|
|
|
|
static const u32 sumo_golden_registers[] =
|
|
|
|
{
|
|
|
|
0x900c, 0x00ffffff, 0x0017071f,
|
|
|
|
0x8c18, 0xffffffff, 0x10101060,
|
|
|
|
0x8c1c, 0xffffffff, 0x00001010,
|
|
|
|
0x8c30, 0x0000000f, 0x00000005,
|
|
|
|
0x9688, 0x0000000f, 0x00000007
|
|
|
|
};
|
|
|
|
|
|
|
|
static const u32 wrestler_golden_registers[] =
|
|
|
|
{
|
|
|
|
0x5eb4, 0xffffffff, 0x00000002,
|
2014-07-08 01:59:37 +04:00
|
|
|
0x5c4, 0xffffffff, 0x00000001,
|
2013-02-28 23:40:09 +04:00
|
|
|
0x7030, 0xffffffff, 0x00000011,
|
|
|
|
0x7c30, 0xffffffff, 0x00000011,
|
|
|
|
0x6104, 0x01000300, 0x00000000,
|
|
|
|
0x5bc0, 0x00300000, 0x00000000,
|
|
|
|
0x918c, 0xffffffff, 0x00010006,
|
|
|
|
0x91a8, 0xffffffff, 0x00010006,
|
|
|
|
0x9150, 0xffffffff, 0x6e944040,
|
|
|
|
0x917c, 0xffffffff, 0x00030002,
|
|
|
|
0x9198, 0xffffffff, 0x00030002,
|
|
|
|
0x915c, 0xffffffff, 0x00010000,
|
|
|
|
0x3f90, 0xffff0000, 0xff000000,
|
|
|
|
0x9178, 0xffffffff, 0x00070000,
|
|
|
|
0x9194, 0xffffffff, 0x00070000,
|
|
|
|
0x9148, 0xffff0000, 0xff000000,
|
|
|
|
0x9190, 0xffffffff, 0x00090008,
|
|
|
|
0x91ac, 0xffffffff, 0x00090008,
|
|
|
|
0x3f94, 0xffff0000, 0xff000000,
|
|
|
|
0x914c, 0xffff0000, 0xff000000,
|
|
|
|
0x929c, 0xffffffff, 0x00000001,
|
|
|
|
0x8a18, 0xffffffff, 0x00000100,
|
|
|
|
0x8b28, 0xffffffff, 0x00000100,
|
|
|
|
0x9144, 0xffffffff, 0x00000100,
|
|
|
|
0x9b7c, 0xffffffff, 0x00000000,
|
|
|
|
0x8030, 0xffffffff, 0x0000100a,
|
|
|
|
0x8a14, 0xffffffff, 0x00000001,
|
|
|
|
0x8b24, 0xffffffff, 0x00ff0fff,
|
|
|
|
0x8b10, 0xffffffff, 0x00000000,
|
|
|
|
0x28a4c, 0x06000000, 0x06000000,
|
|
|
|
0x4d8, 0xffffffff, 0x00000100,
|
|
|
|
0x913c, 0xffff000f, 0x0100000a,
|
|
|
|
0x960c, 0xffffffff, 0x54763210,
|
|
|
|
0x88c4, 0xffffffff, 0x000000c2,
|
|
|
|
0x88d4, 0xffffffff, 0x00000010,
|
|
|
|
0x8974, 0xffffffff, 0x00000000,
|
|
|
|
0xc78, 0x00000080, 0x00000080,
|
|
|
|
0x5e78, 0xffffffff, 0x001000f0,
|
|
|
|
0xd02c, 0xffffffff, 0x08421000,
|
|
|
|
0xa008, 0xffffffff, 0x00010000,
|
|
|
|
0x8d00, 0xffffffff, 0x100e4848,
|
|
|
|
0x8d04, 0xffffffff, 0x00164745,
|
|
|
|
0x8c00, 0xffffffff, 0xe4000003,
|
|
|
|
0x8cf0, 0x1fffffff, 0x08e00410,
|
|
|
|
0x28350, 0xffffffff, 0x00000000,
|
|
|
|
0x9508, 0xffffffff, 0x00000002,
|
|
|
|
0x900c, 0xffffffff, 0x0017071f,
|
|
|
|
0x8c18, 0xffffffff, 0x10101060,
|
|
|
|
0x8c1c, 0xffffffff, 0x00001010
|
|
|
|
};
|
|
|
|
|
|
|
|
static const u32 barts_golden_registers[] =
|
|
|
|
{
|
|
|
|
0x5eb4, 0xffffffff, 0x00000002,
|
|
|
|
0x5e78, 0x8f311ff1, 0x001000f0,
|
|
|
|
0x3f90, 0xffff0000, 0xff000000,
|
|
|
|
0x9148, 0xffff0000, 0xff000000,
|
|
|
|
0x3f94, 0xffff0000, 0xff000000,
|
|
|
|
0x914c, 0xffff0000, 0xff000000,
|
|
|
|
0xc78, 0x00000080, 0x00000080,
|
|
|
|
0xbd4, 0x70073777, 0x00010001,
|
|
|
|
0xd02c, 0xbfffff1f, 0x08421000,
|
|
|
|
0xd0b8, 0x03773777, 0x02011003,
|
|
|
|
0x5bc0, 0x00200000, 0x50100000,
|
|
|
|
0x98f8, 0x33773777, 0x02011003,
|
|
|
|
0x98fc, 0xffffffff, 0x76543210,
|
|
|
|
0x7030, 0x31000311, 0x00000011,
|
|
|
|
0x2f48, 0x00000007, 0x02011003,
|
|
|
|
0x6b28, 0x00000010, 0x00000012,
|
|
|
|
0x7728, 0x00000010, 0x00000012,
|
|
|
|
0x10328, 0x00000010, 0x00000012,
|
|
|
|
0x10f28, 0x00000010, 0x00000012,
|
|
|
|
0x11b28, 0x00000010, 0x00000012,
|
|
|
|
0x12728, 0x00000010, 0x00000012,
|
|
|
|
0x240c, 0x000007ff, 0x00000380,
|
|
|
|
0x8a14, 0xf000001f, 0x00000007,
|
|
|
|
0x8b24, 0x3fff3fff, 0x00ff0fff,
|
|
|
|
0x8b10, 0x0000ff0f, 0x00000000,
|
|
|
|
0x28a4c, 0x07ffffff, 0x06000000,
|
|
|
|
0x10c, 0x00000001, 0x00010003,
|
|
|
|
0xa02c, 0xffffffff, 0x0000009b,
|
|
|
|
0x913c, 0x0000000f, 0x0100000a,
|
|
|
|
0x8d00, 0xffff7f7f, 0x100e4848,
|
|
|
|
0x8d04, 0x00ffffff, 0x00164745,
|
|
|
|
0x8c00, 0xfffc0003, 0xe4000003,
|
|
|
|
0x8c04, 0xf8ff00ff, 0x40600060,
|
|
|
|
0x8c08, 0x00ff00ff, 0x001c001c,
|
|
|
|
0x8cf0, 0x1fff1fff, 0x08e00620,
|
|
|
|
0x8c20, 0x0fff0fff, 0x00800080,
|
|
|
|
0x8c24, 0x0fff0fff, 0x00800080,
|
|
|
|
0x8c18, 0xffffffff, 0x20202078,
|
|
|
|
0x8c1c, 0x0000ffff, 0x00001010,
|
|
|
|
0x28350, 0x00000f01, 0x00000000,
|
|
|
|
0x9508, 0x3700001f, 0x00000002,
|
|
|
|
0x960c, 0xffffffff, 0x54763210,
|
|
|
|
0x88c4, 0x001f3ae3, 0x000000c2,
|
|
|
|
0x88d4, 0x0000001f, 0x00000010,
|
|
|
|
0x8974, 0xffffffff, 0x00000000
|
|
|
|
};
|
|
|
|
|
|
|
|
static const u32 turks_golden_registers[] =
|
|
|
|
{
|
|
|
|
0x5eb4, 0xffffffff, 0x00000002,
|
|
|
|
0x5e78, 0x8f311ff1, 0x001000f0,
|
|
|
|
0x8c8, 0x00003000, 0x00001070,
|
|
|
|
0x8cc, 0x000fffff, 0x00040035,
|
|
|
|
0x3f90, 0xffff0000, 0xfff00000,
|
|
|
|
0x9148, 0xffff0000, 0xfff00000,
|
|
|
|
0x3f94, 0xffff0000, 0xfff00000,
|
|
|
|
0x914c, 0xffff0000, 0xfff00000,
|
|
|
|
0xc78, 0x00000080, 0x00000080,
|
|
|
|
0xbd4, 0x00073007, 0x00010002,
|
|
|
|
0xd02c, 0xbfffff1f, 0x08421000,
|
|
|
|
0xd0b8, 0x03773777, 0x02010002,
|
|
|
|
0x5bc0, 0x00200000, 0x50100000,
|
|
|
|
0x98f8, 0x33773777, 0x00010002,
|
|
|
|
0x98fc, 0xffffffff, 0x33221100,
|
|
|
|
0x7030, 0x31000311, 0x00000011,
|
|
|
|
0x2f48, 0x33773777, 0x00010002,
|
|
|
|
0x6b28, 0x00000010, 0x00000012,
|
|
|
|
0x7728, 0x00000010, 0x00000012,
|
|
|
|
0x10328, 0x00000010, 0x00000012,
|
|
|
|
0x10f28, 0x00000010, 0x00000012,
|
|
|
|
0x11b28, 0x00000010, 0x00000012,
|
|
|
|
0x12728, 0x00000010, 0x00000012,
|
|
|
|
0x240c, 0x000007ff, 0x00000380,
|
|
|
|
0x8a14, 0xf000001f, 0x00000007,
|
|
|
|
0x8b24, 0x3fff3fff, 0x00ff0fff,
|
|
|
|
0x8b10, 0x0000ff0f, 0x00000000,
|
|
|
|
0x28a4c, 0x07ffffff, 0x06000000,
|
|
|
|
0x10c, 0x00000001, 0x00010003,
|
|
|
|
0xa02c, 0xffffffff, 0x0000009b,
|
|
|
|
0x913c, 0x0000000f, 0x0100000a,
|
|
|
|
0x8d00, 0xffff7f7f, 0x100e4848,
|
|
|
|
0x8d04, 0x00ffffff, 0x00164745,
|
|
|
|
0x8c00, 0xfffc0003, 0xe4000003,
|
|
|
|
0x8c04, 0xf8ff00ff, 0x40600060,
|
|
|
|
0x8c08, 0x00ff00ff, 0x001c001c,
|
|
|
|
0x8cf0, 0x1fff1fff, 0x08e00410,
|
|
|
|
0x8c20, 0x0fff0fff, 0x00800080,
|
|
|
|
0x8c24, 0x0fff0fff, 0x00800080,
|
|
|
|
0x8c18, 0xffffffff, 0x20202078,
|
|
|
|
0x8c1c, 0x0000ffff, 0x00001010,
|
|
|
|
0x28350, 0x00000f01, 0x00000000,
|
|
|
|
0x9508, 0x3700001f, 0x00000002,
|
|
|
|
0x960c, 0xffffffff, 0x54763210,
|
|
|
|
0x88c4, 0x001f3ae3, 0x000000c2,
|
|
|
|
0x88d4, 0x0000001f, 0x00000010,
|
|
|
|
0x8974, 0xffffffff, 0x00000000
|
|
|
|
};
|
|
|
|
|
|
|
|
static const u32 caicos_golden_registers[] =
|
|
|
|
{
|
|
|
|
0x5eb4, 0xffffffff, 0x00000002,
|
|
|
|
0x5e78, 0x8f311ff1, 0x001000f0,
|
|
|
|
0x8c8, 0x00003420, 0x00001450,
|
|
|
|
0x8cc, 0x000fffff, 0x00040035,
|
|
|
|
0x3f90, 0xffff0000, 0xfffc0000,
|
|
|
|
0x9148, 0xffff0000, 0xfffc0000,
|
|
|
|
0x3f94, 0xffff0000, 0xfffc0000,
|
|
|
|
0x914c, 0xffff0000, 0xfffc0000,
|
|
|
|
0xc78, 0x00000080, 0x00000080,
|
|
|
|
0xbd4, 0x00073007, 0x00010001,
|
|
|
|
0xd02c, 0xbfffff1f, 0x08421000,
|
|
|
|
0xd0b8, 0x03773777, 0x02010001,
|
|
|
|
0x5bc0, 0x00200000, 0x50100000,
|
|
|
|
0x98f8, 0x33773777, 0x02010001,
|
|
|
|
0x98fc, 0xffffffff, 0x33221100,
|
|
|
|
0x7030, 0x31000311, 0x00000011,
|
|
|
|
0x2f48, 0x33773777, 0x02010001,
|
|
|
|
0x6b28, 0x00000010, 0x00000012,
|
|
|
|
0x7728, 0x00000010, 0x00000012,
|
|
|
|
0x10328, 0x00000010, 0x00000012,
|
|
|
|
0x10f28, 0x00000010, 0x00000012,
|
|
|
|
0x11b28, 0x00000010, 0x00000012,
|
|
|
|
0x12728, 0x00000010, 0x00000012,
|
|
|
|
0x240c, 0x000007ff, 0x00000380,
|
|
|
|
0x8a14, 0xf000001f, 0x00000001,
|
|
|
|
0x8b24, 0x3fff3fff, 0x00ff0fff,
|
|
|
|
0x8b10, 0x0000ff0f, 0x00000000,
|
|
|
|
0x28a4c, 0x07ffffff, 0x06000000,
|
|
|
|
0x10c, 0x00000001, 0x00010003,
|
|
|
|
0xa02c, 0xffffffff, 0x0000009b,
|
|
|
|
0x913c, 0x0000000f, 0x0100000a,
|
|
|
|
0x8d00, 0xffff7f7f, 0x100e4848,
|
|
|
|
0x8d04, 0x00ffffff, 0x00164745,
|
|
|
|
0x8c00, 0xfffc0003, 0xe4000003,
|
|
|
|
0x8c04, 0xf8ff00ff, 0x40600060,
|
|
|
|
0x8c08, 0x00ff00ff, 0x001c001c,
|
|
|
|
0x8cf0, 0x1fff1fff, 0x08e00410,
|
|
|
|
0x8c20, 0x0fff0fff, 0x00800080,
|
|
|
|
0x8c24, 0x0fff0fff, 0x00800080,
|
|
|
|
0x8c18, 0xffffffff, 0x20202078,
|
|
|
|
0x8c1c, 0x0000ffff, 0x00001010,
|
|
|
|
0x28350, 0x00000f01, 0x00000000,
|
|
|
|
0x9508, 0x3700001f, 0x00000002,
|
|
|
|
0x960c, 0xffffffff, 0x54763210,
|
|
|
|
0x88c4, 0x001f3ae3, 0x000000c2,
|
|
|
|
0x88d4, 0x0000001f, 0x00000010,
|
|
|
|
0x8974, 0xffffffff, 0x00000000
|
|
|
|
};
|
|
|
|
|
|
|
|
static void evergreen_init_golden_registers(struct radeon_device *rdev)
|
|
|
|
{
|
|
|
|
switch (rdev->family) {
|
|
|
|
case CHIP_CYPRESS:
|
|
|
|
case CHIP_HEMLOCK:
|
|
|
|
radeon_program_register_sequence(rdev,
|
|
|
|
evergreen_golden_registers,
|
|
|
|
(const u32)ARRAY_SIZE(evergreen_golden_registers));
|
|
|
|
radeon_program_register_sequence(rdev,
|
|
|
|
evergreen_golden_registers2,
|
|
|
|
(const u32)ARRAY_SIZE(evergreen_golden_registers2));
|
|
|
|
radeon_program_register_sequence(rdev,
|
|
|
|
cypress_mgcg_init,
|
|
|
|
(const u32)ARRAY_SIZE(cypress_mgcg_init));
|
|
|
|
break;
|
|
|
|
case CHIP_JUNIPER:
|
|
|
|
radeon_program_register_sequence(rdev,
|
|
|
|
evergreen_golden_registers,
|
|
|
|
(const u32)ARRAY_SIZE(evergreen_golden_registers));
|
|
|
|
radeon_program_register_sequence(rdev,
|
|
|
|
evergreen_golden_registers2,
|
|
|
|
(const u32)ARRAY_SIZE(evergreen_golden_registers2));
|
|
|
|
radeon_program_register_sequence(rdev,
|
|
|
|
juniper_mgcg_init,
|
|
|
|
(const u32)ARRAY_SIZE(juniper_mgcg_init));
|
|
|
|
break;
|
|
|
|
case CHIP_REDWOOD:
|
|
|
|
radeon_program_register_sequence(rdev,
|
|
|
|
evergreen_golden_registers,
|
|
|
|
(const u32)ARRAY_SIZE(evergreen_golden_registers));
|
|
|
|
radeon_program_register_sequence(rdev,
|
|
|
|
evergreen_golden_registers2,
|
|
|
|
(const u32)ARRAY_SIZE(evergreen_golden_registers2));
|
|
|
|
radeon_program_register_sequence(rdev,
|
|
|
|
redwood_mgcg_init,
|
|
|
|
(const u32)ARRAY_SIZE(redwood_mgcg_init));
|
|
|
|
break;
|
|
|
|
case CHIP_CEDAR:
|
|
|
|
radeon_program_register_sequence(rdev,
|
|
|
|
cedar_golden_registers,
|
|
|
|
(const u32)ARRAY_SIZE(cedar_golden_registers));
|
|
|
|
radeon_program_register_sequence(rdev,
|
|
|
|
evergreen_golden_registers2,
|
|
|
|
(const u32)ARRAY_SIZE(evergreen_golden_registers2));
|
|
|
|
radeon_program_register_sequence(rdev,
|
|
|
|
cedar_mgcg_init,
|
|
|
|
(const u32)ARRAY_SIZE(cedar_mgcg_init));
|
|
|
|
break;
|
|
|
|
case CHIP_PALM:
|
|
|
|
radeon_program_register_sequence(rdev,
|
|
|
|
wrestler_golden_registers,
|
|
|
|
(const u32)ARRAY_SIZE(wrestler_golden_registers));
|
|
|
|
break;
|
|
|
|
case CHIP_SUMO:
|
|
|
|
radeon_program_register_sequence(rdev,
|
|
|
|
supersumo_golden_registers,
|
|
|
|
(const u32)ARRAY_SIZE(supersumo_golden_registers));
|
|
|
|
break;
|
|
|
|
case CHIP_SUMO2:
|
|
|
|
radeon_program_register_sequence(rdev,
|
|
|
|
supersumo_golden_registers,
|
|
|
|
(const u32)ARRAY_SIZE(supersumo_golden_registers));
|
|
|
|
radeon_program_register_sequence(rdev,
|
|
|
|
sumo_golden_registers,
|
|
|
|
(const u32)ARRAY_SIZE(sumo_golden_registers));
|
|
|
|
break;
|
|
|
|
case CHIP_BARTS:
|
|
|
|
radeon_program_register_sequence(rdev,
|
|
|
|
barts_golden_registers,
|
|
|
|
(const u32)ARRAY_SIZE(barts_golden_registers));
|
|
|
|
break;
|
|
|
|
case CHIP_TURKS:
|
|
|
|
radeon_program_register_sequence(rdev,
|
|
|
|
turks_golden_registers,
|
|
|
|
(const u32)ARRAY_SIZE(turks_golden_registers));
|
|
|
|
break;
|
|
|
|
case CHIP_CAICOS:
|
|
|
|
radeon_program_register_sequence(rdev,
|
|
|
|
caicos_golden_registers,
|
|
|
|
(const u32)ARRAY_SIZE(caicos_golden_registers));
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2014-10-01 17:43:38 +04:00
|
|
|
/**
|
|
|
|
* evergreen_get_allowed_info_register - fetch the register for the info ioctl
|
|
|
|
*
|
|
|
|
* @rdev: radeon_device pointer
|
|
|
|
* @reg: register offset in bytes
|
|
|
|
* @val: register value
|
|
|
|
*
|
|
|
|
* Returns 0 for success or -EINVAL for an invalid register
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
int evergreen_get_allowed_info_register(struct radeon_device *rdev,
|
|
|
|
u32 reg, u32 *val)
|
|
|
|
{
|
|
|
|
switch (reg) {
|
|
|
|
case GRBM_STATUS:
|
|
|
|
case GRBM_STATUS_SE0:
|
|
|
|
case GRBM_STATUS_SE1:
|
|
|
|
case SRBM_STATUS:
|
|
|
|
case SRBM_STATUS2:
|
|
|
|
case DMA_STATUS_REG:
|
|
|
|
case UVD_STATUS:
|
|
|
|
*val = RREG32(reg);
|
|
|
|
return 0;
|
|
|
|
default:
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2011-12-17 02:03:42 +04:00
|
|
|
void evergreen_tiling_fields(unsigned tiling_flags, unsigned *bankw,
|
|
|
|
unsigned *bankh, unsigned *mtaspect,
|
|
|
|
unsigned *tile_split)
|
|
|
|
{
|
|
|
|
*bankw = (tiling_flags >> RADEON_TILING_EG_BANKW_SHIFT) & RADEON_TILING_EG_BANKW_MASK;
|
|
|
|
*bankh = (tiling_flags >> RADEON_TILING_EG_BANKH_SHIFT) & RADEON_TILING_EG_BANKH_MASK;
|
|
|
|
*mtaspect = (tiling_flags >> RADEON_TILING_EG_MACRO_TILE_ASPECT_SHIFT) & RADEON_TILING_EG_MACRO_TILE_ASPECT_MASK;
|
|
|
|
*tile_split = (tiling_flags >> RADEON_TILING_EG_TILE_SPLIT_SHIFT) & RADEON_TILING_EG_TILE_SPLIT_MASK;
|
|
|
|
switch (*bankw) {
|
|
|
|
default:
|
|
|
|
case 1: *bankw = EVERGREEN_ADDR_SURF_BANK_WIDTH_1; break;
|
|
|
|
case 2: *bankw = EVERGREEN_ADDR_SURF_BANK_WIDTH_2; break;
|
|
|
|
case 4: *bankw = EVERGREEN_ADDR_SURF_BANK_WIDTH_4; break;
|
|
|
|
case 8: *bankw = EVERGREEN_ADDR_SURF_BANK_WIDTH_8; break;
|
|
|
|
}
|
|
|
|
switch (*bankh) {
|
|
|
|
default:
|
|
|
|
case 1: *bankh = EVERGREEN_ADDR_SURF_BANK_HEIGHT_1; break;
|
|
|
|
case 2: *bankh = EVERGREEN_ADDR_SURF_BANK_HEIGHT_2; break;
|
|
|
|
case 4: *bankh = EVERGREEN_ADDR_SURF_BANK_HEIGHT_4; break;
|
|
|
|
case 8: *bankh = EVERGREEN_ADDR_SURF_BANK_HEIGHT_8; break;
|
|
|
|
}
|
|
|
|
switch (*mtaspect) {
|
|
|
|
default:
|
|
|
|
case 1: *mtaspect = EVERGREEN_ADDR_SURF_MACRO_TILE_ASPECT_1; break;
|
|
|
|
case 2: *mtaspect = EVERGREEN_ADDR_SURF_MACRO_TILE_ASPECT_2; break;
|
|
|
|
case 4: *mtaspect = EVERGREEN_ADDR_SURF_MACRO_TILE_ASPECT_4; break;
|
|
|
|
case 8: *mtaspect = EVERGREEN_ADDR_SURF_MACRO_TILE_ASPECT_8; break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2013-04-08 14:41:32 +04:00
|
|
|
static int sumo_set_uvd_clock(struct radeon_device *rdev, u32 clock,
|
|
|
|
u32 cntl_reg, u32 status_reg)
|
|
|
|
{
|
|
|
|
int r, i;
|
|
|
|
struct atom_clock_dividers dividers;
|
|
|
|
|
2016-03-16 14:56:45 +03:00
|
|
|
r = radeon_atom_get_clock_dividers(rdev, COMPUTE_ENGINE_PLL_PARAM,
|
2013-04-08 14:41:32 +04:00
|
|
|
clock, false, ÷rs);
|
|
|
|
if (r)
|
|
|
|
return r;
|
|
|
|
|
|
|
|
WREG32_P(cntl_reg, dividers.post_div, ~(DCLK_DIR_CNTL_EN|DCLK_DIVIDER_MASK));
|
|
|
|
|
|
|
|
for (i = 0; i < 100; i++) {
|
|
|
|
if (RREG32(status_reg) & DCLK_STATUS)
|
|
|
|
break;
|
|
|
|
mdelay(10);
|
|
|
|
}
|
|
|
|
if (i == 100)
|
|
|
|
return -ETIMEDOUT;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
int sumo_set_uvd_clocks(struct radeon_device *rdev, u32 vclk, u32 dclk)
|
|
|
|
{
|
|
|
|
int r = 0;
|
|
|
|
u32 cg_scratch = RREG32(CG_SCRATCH1);
|
|
|
|
|
|
|
|
r = sumo_set_uvd_clock(rdev, vclk, CG_VCLK_CNTL, CG_VCLK_STATUS);
|
|
|
|
if (r)
|
|
|
|
goto done;
|
|
|
|
cg_scratch &= 0xffff0000;
|
|
|
|
cg_scratch |= vclk / 100; /* Mhz */
|
|
|
|
|
|
|
|
r = sumo_set_uvd_clock(rdev, dclk, CG_DCLK_CNTL, CG_DCLK_STATUS);
|
|
|
|
if (r)
|
|
|
|
goto done;
|
|
|
|
cg_scratch &= 0x0000ffff;
|
|
|
|
cg_scratch |= (dclk / 100) << 16; /* Mhz */
|
|
|
|
|
|
|
|
done:
|
|
|
|
WREG32(CG_SCRATCH1, cg_scratch);
|
|
|
|
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
|
2013-04-08 14:41:33 +04:00
|
|
|
int evergreen_set_uvd_clocks(struct radeon_device *rdev, u32 vclk, u32 dclk)
|
|
|
|
{
|
|
|
|
/* start off with something large */
|
2013-04-29 13:55:02 +04:00
|
|
|
unsigned fb_div = 0, vclk_div = 0, dclk_div = 0;
|
2013-04-08 14:41:33 +04:00
|
|
|
int r;
|
|
|
|
|
2013-04-18 17:25:58 +04:00
|
|
|
/* bypass vclk and dclk with bclk */
|
|
|
|
WREG32_P(CG_UPLL_FUNC_CNTL_2,
|
|
|
|
VCLK_SRC_SEL(1) | DCLK_SRC_SEL(1),
|
|
|
|
~(VCLK_SRC_SEL_MASK | DCLK_SRC_SEL_MASK));
|
|
|
|
|
|
|
|
/* put PLL in bypass mode */
|
|
|
|
WREG32_P(CG_UPLL_FUNC_CNTL, UPLL_BYPASS_EN_MASK, ~UPLL_BYPASS_EN_MASK);
|
|
|
|
|
|
|
|
if (!vclk || !dclk) {
|
|
|
|
/* keep the Bypass mode, put PLL to sleep */
|
|
|
|
WREG32_P(CG_UPLL_FUNC_CNTL, UPLL_SLEEP_MASK, ~UPLL_SLEEP_MASK);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2013-04-29 13:55:02 +04:00
|
|
|
r = radeon_uvd_calc_upll_dividers(rdev, vclk, dclk, 125000, 250000,
|
|
|
|
16384, 0x03FFFFFF, 0, 128, 5,
|
|
|
|
&fb_div, &vclk_div, &dclk_div);
|
|
|
|
if (r)
|
|
|
|
return r;
|
2013-04-08 14:41:33 +04:00
|
|
|
|
|
|
|
/* set VCO_MODE to 1 */
|
|
|
|
WREG32_P(CG_UPLL_FUNC_CNTL, UPLL_VCO_MODE_MASK, ~UPLL_VCO_MODE_MASK);
|
|
|
|
|
|
|
|
/* toggle UPLL_SLEEP to 1 then back to 0 */
|
|
|
|
WREG32_P(CG_UPLL_FUNC_CNTL, UPLL_SLEEP_MASK, ~UPLL_SLEEP_MASK);
|
|
|
|
WREG32_P(CG_UPLL_FUNC_CNTL, 0, ~UPLL_SLEEP_MASK);
|
|
|
|
|
|
|
|
/* deassert UPLL_RESET */
|
|
|
|
WREG32_P(CG_UPLL_FUNC_CNTL, 0, ~UPLL_RESET_MASK);
|
|
|
|
|
|
|
|
mdelay(1);
|
|
|
|
|
2013-04-29 13:55:02 +04:00
|
|
|
r = radeon_uvd_send_upll_ctlreq(rdev, CG_UPLL_FUNC_CNTL);
|
2013-04-08 14:41:33 +04:00
|
|
|
if (r)
|
|
|
|
return r;
|
|
|
|
|
|
|
|
/* assert UPLL_RESET again */
|
|
|
|
WREG32_P(CG_UPLL_FUNC_CNTL, UPLL_RESET_MASK, ~UPLL_RESET_MASK);
|
|
|
|
|
|
|
|
/* disable spread spectrum. */
|
|
|
|
WREG32_P(CG_UPLL_SPREAD_SPECTRUM, 0, ~SSEN_MASK);
|
|
|
|
|
|
|
|
/* set feedback divider */
|
2013-04-29 13:55:02 +04:00
|
|
|
WREG32_P(CG_UPLL_FUNC_CNTL_3, UPLL_FB_DIV(fb_div), ~UPLL_FB_DIV_MASK);
|
2013-04-08 14:41:33 +04:00
|
|
|
|
|
|
|
/* set ref divider to 0 */
|
|
|
|
WREG32_P(CG_UPLL_FUNC_CNTL, 0, ~UPLL_REF_DIV_MASK);
|
|
|
|
|
2013-04-29 13:55:02 +04:00
|
|
|
if (fb_div < 307200)
|
2013-04-08 14:41:33 +04:00
|
|
|
WREG32_P(CG_UPLL_FUNC_CNTL_4, 0, ~UPLL_SPARE_ISPARE9);
|
|
|
|
else
|
|
|
|
WREG32_P(CG_UPLL_FUNC_CNTL_4, UPLL_SPARE_ISPARE9, ~UPLL_SPARE_ISPARE9);
|
|
|
|
|
|
|
|
/* set PDIV_A and PDIV_B */
|
|
|
|
WREG32_P(CG_UPLL_FUNC_CNTL_2,
|
2013-04-29 13:55:02 +04:00
|
|
|
UPLL_PDIV_A(vclk_div) | UPLL_PDIV_B(dclk_div),
|
2013-04-08 14:41:33 +04:00
|
|
|
~(UPLL_PDIV_A_MASK | UPLL_PDIV_B_MASK));
|
|
|
|
|
|
|
|
/* give the PLL some time to settle */
|
|
|
|
mdelay(15);
|
|
|
|
|
|
|
|
/* deassert PLL_RESET */
|
|
|
|
WREG32_P(CG_UPLL_FUNC_CNTL, 0, ~UPLL_RESET_MASK);
|
|
|
|
|
|
|
|
mdelay(15);
|
|
|
|
|
|
|
|
/* switch from bypass mode to normal mode */
|
|
|
|
WREG32_P(CG_UPLL_FUNC_CNTL, 0, ~UPLL_BYPASS_EN_MASK);
|
|
|
|
|
2013-04-29 13:55:02 +04:00
|
|
|
r = radeon_uvd_send_upll_ctlreq(rdev, CG_UPLL_FUNC_CNTL);
|
2013-04-08 14:41:33 +04:00
|
|
|
if (r)
|
|
|
|
return r;
|
|
|
|
|
|
|
|
/* switch VCLK and DCLK selection */
|
|
|
|
WREG32_P(CG_UPLL_FUNC_CNTL_2,
|
|
|
|
VCLK_SRC_SEL(2) | DCLK_SRC_SEL(2),
|
|
|
|
~(VCLK_SRC_SEL_MASK | DCLK_SRC_SEL_MASK));
|
|
|
|
|
|
|
|
mdelay(100);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2011-09-01 21:46:15 +04:00
|
|
|
void evergreen_fix_pci_max_read_req_size(struct radeon_device *rdev)
|
|
|
|
{
|
2013-09-09 17:13:08 +04:00
|
|
|
int readrq;
|
|
|
|
u16 v;
|
2011-09-01 21:46:15 +04:00
|
|
|
|
2013-09-09 17:13:08 +04:00
|
|
|
readrq = pcie_get_readrq(rdev->pdev);
|
|
|
|
v = ffs(readrq) - 8;
|
2011-09-01 21:46:15 +04:00
|
|
|
/* if bios or OS sets MAX_READ_REQUEST_SIZE to an invalid value, fix it
|
|
|
|
* to avoid hangs or perfomance issues
|
|
|
|
*/
|
2013-09-09 17:13:08 +04:00
|
|
|
if ((v == 0) || (v == 6) || (v == 7))
|
|
|
|
pcie_set_readrq(rdev->pdev, 512);
|
2011-09-01 21:46:15 +04:00
|
|
|
}
|
|
|
|
|
2013-09-23 20:22:11 +04:00
|
|
|
void dce4_program_fmt(struct drm_encoder *encoder)
|
|
|
|
{
|
|
|
|
struct drm_device *dev = encoder->dev;
|
|
|
|
struct radeon_device *rdev = dev->dev_private;
|
|
|
|
struct radeon_encoder *radeon_encoder = to_radeon_encoder(encoder);
|
|
|
|
struct radeon_crtc *radeon_crtc = to_radeon_crtc(encoder->crtc);
|
|
|
|
struct drm_connector *connector = radeon_get_connector_for_encoder(encoder);
|
|
|
|
int bpc = 0;
|
|
|
|
u32 tmp = 0;
|
2013-09-25 01:26:26 +04:00
|
|
|
enum radeon_connector_dither dither = RADEON_FMT_DITHER_DISABLE;
|
2013-09-23 20:22:11 +04:00
|
|
|
|
2013-09-25 01:26:26 +04:00
|
|
|
if (connector) {
|
|
|
|
struct radeon_connector *radeon_connector = to_radeon_connector(connector);
|
2013-09-23 20:22:11 +04:00
|
|
|
bpc = radeon_get_monitor_bpc(connector);
|
2013-09-25 01:26:26 +04:00
|
|
|
dither = radeon_connector->dither;
|
|
|
|
}
|
2013-09-23 20:22:11 +04:00
|
|
|
|
|
|
|
/* LVDS/eDP FMT is set up by atom */
|
|
|
|
if (radeon_encoder->devices & ATOM_DEVICE_LCD_SUPPORT)
|
|
|
|
return;
|
|
|
|
|
|
|
|
/* not needed for analog */
|
|
|
|
if ((radeon_encoder->encoder_id == ENCODER_OBJECT_ID_INTERNAL_KLDSCP_DAC1) ||
|
|
|
|
(radeon_encoder->encoder_id == ENCODER_OBJECT_ID_INTERNAL_KLDSCP_DAC2))
|
|
|
|
return;
|
|
|
|
|
|
|
|
if (bpc == 0)
|
|
|
|
return;
|
|
|
|
|
|
|
|
switch (bpc) {
|
|
|
|
case 6:
|
2013-09-25 01:26:26 +04:00
|
|
|
if (dither == RADEON_FMT_DITHER_ENABLE)
|
2013-09-23 20:22:11 +04:00
|
|
|
/* XXX sort out optimal dither settings */
|
|
|
|
tmp |= (FMT_FRAME_RANDOM_ENABLE | FMT_HIGHPASS_RANDOM_ENABLE |
|
|
|
|
FMT_SPATIAL_DITHER_EN);
|
|
|
|
else
|
|
|
|
tmp |= FMT_TRUNCATE_EN;
|
|
|
|
break;
|
|
|
|
case 8:
|
2013-09-25 01:26:26 +04:00
|
|
|
if (dither == RADEON_FMT_DITHER_ENABLE)
|
2013-09-23 20:22:11 +04:00
|
|
|
/* XXX sort out optimal dither settings */
|
|
|
|
tmp |= (FMT_FRAME_RANDOM_ENABLE | FMT_HIGHPASS_RANDOM_ENABLE |
|
|
|
|
FMT_RGB_RANDOM_ENABLE |
|
|
|
|
FMT_SPATIAL_DITHER_EN | FMT_SPATIAL_DITHER_DEPTH);
|
|
|
|
else
|
|
|
|
tmp |= (FMT_TRUNCATE_EN | FMT_TRUNCATE_DEPTH);
|
|
|
|
break;
|
|
|
|
case 10:
|
|
|
|
default:
|
|
|
|
/* not needed */
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
WREG32(FMT_BIT_DEPTH_CONTROL + radeon_crtc->crtc_offset, tmp);
|
|
|
|
}
|
|
|
|
|
2013-04-10 02:49:59 +04:00
|
|
|
static bool dce4_is_in_vblank(struct radeon_device *rdev, int crtc)
|
|
|
|
{
|
|
|
|
if (RREG32(EVERGREEN_CRTC_STATUS + crtc_offsets[crtc]) & EVERGREEN_CRTC_V_BLANK)
|
|
|
|
return true;
|
|
|
|
else
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
static bool dce4_is_counter_moving(struct radeon_device *rdev, int crtc)
|
|
|
|
{
|
|
|
|
u32 pos1, pos2;
|
|
|
|
|
|
|
|
pos1 = RREG32(EVERGREEN_CRTC_STATUS_POSITION + crtc_offsets[crtc]);
|
|
|
|
pos2 = RREG32(EVERGREEN_CRTC_STATUS_POSITION + crtc_offsets[crtc]);
|
|
|
|
|
|
|
|
if (pos1 != pos2)
|
|
|
|
return true;
|
|
|
|
else
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2012-07-17 22:02:42 +04:00
|
|
|
/**
|
|
|
|
* dce4_wait_for_vblank - vblank wait asic callback.
|
|
|
|
*
|
|
|
|
* @rdev: radeon_device pointer
|
|
|
|
* @crtc: crtc to wait for vblank on
|
|
|
|
*
|
|
|
|
* Wait for vblank on the requested crtc (evergreen+).
|
|
|
|
*/
|
2012-02-24 02:53:37 +04:00
|
|
|
void dce4_wait_for_vblank(struct radeon_device *rdev, int crtc)
|
|
|
|
{
|
2013-04-10 02:49:59 +04:00
|
|
|
unsigned i = 0;
|
2012-02-24 02:53:37 +04:00
|
|
|
|
2012-08-16 01:13:53 +04:00
|
|
|
if (crtc >= rdev->num_crtc)
|
|
|
|
return;
|
|
|
|
|
2013-04-10 02:49:59 +04:00
|
|
|
if (!(RREG32(EVERGREEN_CRTC_CONTROL + crtc_offsets[crtc]) & EVERGREEN_CRTC_MASTER_EN))
|
|
|
|
return;
|
|
|
|
|
|
|
|
/* depending on when we hit vblank, we may be close to active; if so,
|
|
|
|
* wait for another frame.
|
|
|
|
*/
|
|
|
|
while (dce4_is_in_vblank(rdev, crtc)) {
|
|
|
|
if (i++ % 100 == 0) {
|
|
|
|
if (!dce4_is_counter_moving(rdev, crtc))
|
2012-02-24 02:53:37 +04:00
|
|
|
break;
|
|
|
|
}
|
2013-04-10 02:49:59 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
while (!dce4_is_in_vblank(rdev, crtc)) {
|
|
|
|
if (i++ % 100 == 0) {
|
|
|
|
if (!dce4_is_counter_moving(rdev, crtc))
|
2012-02-24 02:53:37 +04:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2012-07-17 22:02:42 +04:00
|
|
|
/**
|
|
|
|
* evergreen_page_flip - pageflip callback.
|
|
|
|
*
|
|
|
|
* @rdev: radeon_device pointer
|
|
|
|
* @crtc_id: crtc to cleanup pageflip on
|
|
|
|
* @crtc_base: new address of the crtc (GPU MC address)
|
|
|
|
*
|
2015-10-23 19:56:51 +03:00
|
|
|
* Triggers the actual pageflip by updating the primary
|
|
|
|
* surface base address (evergreen+).
|
2012-07-17 22:02:42 +04:00
|
|
|
*/
|
2016-04-01 12:51:34 +03:00
|
|
|
void evergreen_page_flip(struct radeon_device *rdev, int crtc_id, u64 crtc_base,
|
|
|
|
bool async)
|
2010-11-21 18:59:01 +03:00
|
|
|
{
|
|
|
|
struct radeon_crtc *radeon_crtc = rdev->mode_info.crtcs[crtc_id];
|
|
|
|
|
|
|
|
/* update the scanout addresses */
|
2016-04-01 12:51:34 +03:00
|
|
|
WREG32(EVERGREEN_GRPH_FLIP_CONTROL + radeon_crtc->crtc_offset,
|
|
|
|
async ? EVERGREEN_GRPH_SURFACE_UPDATE_H_RETRACE_EN : 0);
|
2010-11-21 18:59:01 +03:00
|
|
|
WREG32(EVERGREEN_GRPH_PRIMARY_SURFACE_ADDRESS_HIGH + radeon_crtc->crtc_offset,
|
|
|
|
upper_32_bits(crtc_base));
|
|
|
|
WREG32(EVERGREEN_GRPH_PRIMARY_SURFACE_ADDRESS + radeon_crtc->crtc_offset,
|
|
|
|
(u32)crtc_base);
|
2015-10-23 19:56:51 +03:00
|
|
|
/* post the write */
|
|
|
|
RREG32(EVERGREEN_GRPH_PRIMARY_SURFACE_ADDRESS + radeon_crtc->crtc_offset);
|
2014-05-27 18:49:20 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* evergreen_page_flip_pending - check if page flip is still pending
|
|
|
|
*
|
|
|
|
* @rdev: radeon_device pointer
|
|
|
|
* @crtc_id: crtc to check
|
|
|
|
*
|
|
|
|
* Returns the current update pending status.
|
|
|
|
*/
|
|
|
|
bool evergreen_page_flip_pending(struct radeon_device *rdev, int crtc_id)
|
|
|
|
{
|
|
|
|
struct radeon_crtc *radeon_crtc = rdev->mode_info.crtcs[crtc_id];
|
2010-11-21 18:59:01 +03:00
|
|
|
|
|
|
|
/* Return current update_pending status: */
|
2014-05-27 18:49:20 +04:00
|
|
|
return !!(RREG32(EVERGREEN_GRPH_UPDATE + radeon_crtc->crtc_offset) &
|
|
|
|
EVERGREEN_GRPH_SURFACE_UPDATE_PENDING);
|
2010-11-21 18:59:01 +03:00
|
|
|
}
|
|
|
|
|
2010-07-02 20:58:16 +04:00
|
|
|
/* get temperature in millidegrees */
|
2011-02-02 00:12:34 +03:00
|
|
|
int evergreen_get_temp(struct radeon_device *rdev)
|
2010-07-02 20:58:16 +04:00
|
|
|
{
|
2011-06-14 23:15:53 +04:00
|
|
|
u32 temp, toffset;
|
|
|
|
int actual_temp = 0;
|
2011-05-26 02:45:37 +04:00
|
|
|
|
|
|
|
if (rdev->family == CHIP_JUNIPER) {
|
|
|
|
toffset = (RREG32(CG_THERMAL_CTRL) & TOFFSET_MASK) >>
|
|
|
|
TOFFSET_SHIFT;
|
|
|
|
temp = (RREG32(CG_TS0_STATUS) & TS0_ADC_DOUT_MASK) >>
|
|
|
|
TS0_ADC_DOUT_SHIFT;
|
|
|
|
|
|
|
|
if (toffset & 0x100)
|
|
|
|
actual_temp = temp / 2 - (0x200 - toffset);
|
|
|
|
else
|
|
|
|
actual_temp = temp / 2 + toffset;
|
|
|
|
|
|
|
|
actual_temp = actual_temp * 1000;
|
|
|
|
|
|
|
|
} else {
|
|
|
|
temp = (RREG32(CG_MULT_THERMAL_STATUS) & ASIC_T_MASK) >>
|
|
|
|
ASIC_T_SHIFT;
|
|
|
|
|
|
|
|
if (temp & 0x400)
|
|
|
|
actual_temp = -256;
|
|
|
|
else if (temp & 0x200)
|
|
|
|
actual_temp = 255;
|
|
|
|
else if (temp & 0x100) {
|
|
|
|
actual_temp = temp & 0x1ff;
|
|
|
|
actual_temp |= ~0x1ff;
|
|
|
|
} else
|
|
|
|
actual_temp = temp & 0xff;
|
|
|
|
|
|
|
|
actual_temp = (actual_temp * 1000) / 2;
|
|
|
|
}
|
2010-07-02 20:58:16 +04:00
|
|
|
|
2011-05-26 02:45:37 +04:00
|
|
|
return actual_temp;
|
2010-07-02 20:58:16 +04:00
|
|
|
}
|
|
|
|
|
2011-02-02 00:12:34 +03:00
|
|
|
int sumo_get_temp(struct radeon_device *rdev)
|
2010-11-23 01:56:32 +03:00
|
|
|
{
|
|
|
|
u32 temp = RREG32(CG_THERMAL_STATUS) & 0xff;
|
2011-02-02 00:12:34 +03:00
|
|
|
int actual_temp = temp - 49;
|
2010-11-23 01:56:32 +03:00
|
|
|
|
|
|
|
return actual_temp * 1000;
|
|
|
|
}
|
|
|
|
|
2012-07-17 22:02:42 +04:00
|
|
|
/**
|
|
|
|
* sumo_pm_init_profile - Initialize power profiles callback.
|
|
|
|
*
|
|
|
|
* @rdev: radeon_device pointer
|
|
|
|
*
|
|
|
|
* Initialize the power states used in profile mode
|
|
|
|
* (sumo, trinity, SI).
|
|
|
|
* Used for profile mode only.
|
|
|
|
*/
|
2011-11-04 18:09:41 +04:00
|
|
|
void sumo_pm_init_profile(struct radeon_device *rdev)
|
|
|
|
{
|
|
|
|
int idx;
|
|
|
|
|
|
|
|
/* default */
|
|
|
|
rdev->pm.profiles[PM_PROFILE_DEFAULT_IDX].dpms_off_ps_idx = rdev->pm.default_power_state_index;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_DEFAULT_IDX].dpms_on_ps_idx = rdev->pm.default_power_state_index;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_DEFAULT_IDX].dpms_off_cm_idx = 0;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_DEFAULT_IDX].dpms_on_cm_idx = 0;
|
|
|
|
|
|
|
|
/* low,mid sh/mh */
|
|
|
|
if (rdev->flags & RADEON_IS_MOBILITY)
|
|
|
|
idx = radeon_pm_get_type_index(rdev, POWER_STATE_TYPE_BATTERY, 0);
|
|
|
|
else
|
|
|
|
idx = radeon_pm_get_type_index(rdev, POWER_STATE_TYPE_PERFORMANCE, 0);
|
|
|
|
|
|
|
|
rdev->pm.profiles[PM_PROFILE_LOW_SH_IDX].dpms_off_ps_idx = idx;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_LOW_SH_IDX].dpms_on_ps_idx = idx;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_LOW_SH_IDX].dpms_off_cm_idx = 0;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_LOW_SH_IDX].dpms_on_cm_idx = 0;
|
|
|
|
|
|
|
|
rdev->pm.profiles[PM_PROFILE_LOW_MH_IDX].dpms_off_ps_idx = idx;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_LOW_MH_IDX].dpms_on_ps_idx = idx;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_LOW_MH_IDX].dpms_off_cm_idx = 0;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_LOW_MH_IDX].dpms_on_cm_idx = 0;
|
|
|
|
|
|
|
|
rdev->pm.profiles[PM_PROFILE_MID_SH_IDX].dpms_off_ps_idx = idx;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_MID_SH_IDX].dpms_on_ps_idx = idx;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_MID_SH_IDX].dpms_off_cm_idx = 0;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_MID_SH_IDX].dpms_on_cm_idx = 0;
|
|
|
|
|
|
|
|
rdev->pm.profiles[PM_PROFILE_MID_MH_IDX].dpms_off_ps_idx = idx;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_MID_MH_IDX].dpms_on_ps_idx = idx;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_MID_MH_IDX].dpms_off_cm_idx = 0;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_MID_MH_IDX].dpms_on_cm_idx = 0;
|
|
|
|
|
|
|
|
/* high sh/mh */
|
|
|
|
idx = radeon_pm_get_type_index(rdev, POWER_STATE_TYPE_PERFORMANCE, 0);
|
|
|
|
rdev->pm.profiles[PM_PROFILE_HIGH_SH_IDX].dpms_off_ps_idx = idx;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_HIGH_SH_IDX].dpms_on_ps_idx = idx;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_HIGH_SH_IDX].dpms_off_cm_idx = 0;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_HIGH_SH_IDX].dpms_on_cm_idx =
|
|
|
|
rdev->pm.power_state[idx].num_clock_modes - 1;
|
|
|
|
|
|
|
|
rdev->pm.profiles[PM_PROFILE_HIGH_MH_IDX].dpms_off_ps_idx = idx;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_HIGH_MH_IDX].dpms_on_ps_idx = idx;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_HIGH_MH_IDX].dpms_off_cm_idx = 0;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_HIGH_MH_IDX].dpms_on_cm_idx =
|
|
|
|
rdev->pm.power_state[idx].num_clock_modes - 1;
|
|
|
|
}
|
|
|
|
|
2012-10-02 03:25:11 +04:00
|
|
|
/**
|
|
|
|
* btc_pm_init_profile - Initialize power profiles callback.
|
|
|
|
*
|
|
|
|
* @rdev: radeon_device pointer
|
|
|
|
*
|
|
|
|
* Initialize the power states used in profile mode
|
|
|
|
* (BTC, cayman).
|
|
|
|
* Used for profile mode only.
|
|
|
|
*/
|
|
|
|
void btc_pm_init_profile(struct radeon_device *rdev)
|
|
|
|
{
|
|
|
|
int idx;
|
|
|
|
|
|
|
|
/* default */
|
|
|
|
rdev->pm.profiles[PM_PROFILE_DEFAULT_IDX].dpms_off_ps_idx = rdev->pm.default_power_state_index;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_DEFAULT_IDX].dpms_on_ps_idx = rdev->pm.default_power_state_index;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_DEFAULT_IDX].dpms_off_cm_idx = 0;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_DEFAULT_IDX].dpms_on_cm_idx = 2;
|
|
|
|
/* starting with BTC, there is one state that is used for both
|
|
|
|
* MH and SH. Difference is that we always use the high clock index for
|
|
|
|
* mclk.
|
|
|
|
*/
|
|
|
|
if (rdev->flags & RADEON_IS_MOBILITY)
|
|
|
|
idx = radeon_pm_get_type_index(rdev, POWER_STATE_TYPE_BATTERY, 0);
|
|
|
|
else
|
|
|
|
idx = radeon_pm_get_type_index(rdev, POWER_STATE_TYPE_PERFORMANCE, 0);
|
|
|
|
/* low sh */
|
|
|
|
rdev->pm.profiles[PM_PROFILE_LOW_SH_IDX].dpms_off_ps_idx = idx;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_LOW_SH_IDX].dpms_on_ps_idx = idx;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_LOW_SH_IDX].dpms_off_cm_idx = 0;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_LOW_SH_IDX].dpms_on_cm_idx = 0;
|
|
|
|
/* mid sh */
|
|
|
|
rdev->pm.profiles[PM_PROFILE_MID_SH_IDX].dpms_off_ps_idx = idx;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_MID_SH_IDX].dpms_on_ps_idx = idx;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_MID_SH_IDX].dpms_off_cm_idx = 0;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_MID_SH_IDX].dpms_on_cm_idx = 1;
|
|
|
|
/* high sh */
|
|
|
|
rdev->pm.profiles[PM_PROFILE_HIGH_SH_IDX].dpms_off_ps_idx = idx;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_HIGH_SH_IDX].dpms_on_ps_idx = idx;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_HIGH_SH_IDX].dpms_off_cm_idx = 0;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_HIGH_SH_IDX].dpms_on_cm_idx = 2;
|
|
|
|
/* low mh */
|
|
|
|
rdev->pm.profiles[PM_PROFILE_LOW_MH_IDX].dpms_off_ps_idx = idx;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_LOW_MH_IDX].dpms_on_ps_idx = idx;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_LOW_MH_IDX].dpms_off_cm_idx = 0;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_LOW_MH_IDX].dpms_on_cm_idx = 0;
|
|
|
|
/* mid mh */
|
|
|
|
rdev->pm.profiles[PM_PROFILE_MID_MH_IDX].dpms_off_ps_idx = idx;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_MID_MH_IDX].dpms_on_ps_idx = idx;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_MID_MH_IDX].dpms_off_cm_idx = 0;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_MID_MH_IDX].dpms_on_cm_idx = 1;
|
|
|
|
/* high mh */
|
|
|
|
rdev->pm.profiles[PM_PROFILE_HIGH_MH_IDX].dpms_off_ps_idx = idx;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_HIGH_MH_IDX].dpms_on_ps_idx = idx;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_HIGH_MH_IDX].dpms_off_cm_idx = 0;
|
|
|
|
rdev->pm.profiles[PM_PROFILE_HIGH_MH_IDX].dpms_on_cm_idx = 2;
|
|
|
|
}
|
|
|
|
|
2012-07-17 22:02:42 +04:00
|
|
|
/**
|
|
|
|
* evergreen_pm_misc - set additional pm hw parameters callback.
|
|
|
|
*
|
|
|
|
* @rdev: radeon_device pointer
|
|
|
|
*
|
|
|
|
* Set non-clock parameters associated with a power state
|
|
|
|
* (voltage, etc.) (evergreen+).
|
|
|
|
*/
|
2010-04-24 01:57:27 +04:00
|
|
|
void evergreen_pm_misc(struct radeon_device *rdev)
|
|
|
|
{
|
2010-06-08 02:20:25 +04:00
|
|
|
int req_ps_idx = rdev->pm.requested_power_state_index;
|
|
|
|
int req_cm_idx = rdev->pm.requested_clock_mode_index;
|
|
|
|
struct radeon_power_state *ps = &rdev->pm.power_state[req_ps_idx];
|
|
|
|
struct radeon_voltage *voltage = &ps->clock_info[req_cm_idx].voltage;
|
2010-04-24 01:57:27 +04:00
|
|
|
|
2011-04-12 22:49:24 +04:00
|
|
|
if (voltage->type == VOLTAGE_SW) {
|
2013-07-05 21:14:30 +04:00
|
|
|
/* 0xff0x are flags rather then an actual voltage */
|
|
|
|
if ((voltage->voltage & 0xff00) == 0xff00)
|
2011-06-20 21:00:31 +04:00
|
|
|
return;
|
2011-04-12 22:49:24 +04:00
|
|
|
if (voltage->voltage && (voltage->voltage != rdev->pm.current_vddc)) {
|
2011-04-12 22:49:23 +04:00
|
|
|
radeon_atom_set_voltage(rdev, voltage->voltage, SET_VOLTAGE_TYPE_ASIC_VDDC);
|
2010-06-08 02:15:18 +04:00
|
|
|
rdev->pm.current_vddc = voltage->voltage;
|
2011-04-12 22:49:24 +04:00
|
|
|
DRM_DEBUG("Setting: vddc: %d\n", voltage->voltage);
|
|
|
|
}
|
2013-02-11 17:44:48 +04:00
|
|
|
|
|
|
|
/* starting with BTC, there is one state that is used for both
|
|
|
|
* MH and SH. Difference is that we always use the high clock index for
|
|
|
|
* mclk and vddci.
|
|
|
|
*/
|
|
|
|
if ((rdev->pm.pm_method == PM_METHOD_PROFILE) &&
|
|
|
|
(rdev->family >= CHIP_BARTS) &&
|
|
|
|
rdev->pm.active_crtc_count &&
|
|
|
|
((rdev->pm.profile_index == PM_PROFILE_MID_MH_IDX) ||
|
|
|
|
(rdev->pm.profile_index == PM_PROFILE_LOW_MH_IDX)))
|
|
|
|
voltage = &rdev->pm.power_state[req_ps_idx].
|
|
|
|
clock_info[rdev->pm.profiles[PM_PROFILE_HIGH_MH_IDX].dpms_on_cm_idx].voltage;
|
|
|
|
|
2013-07-05 21:14:30 +04:00
|
|
|
/* 0xff0x are flags rather then an actual voltage */
|
|
|
|
if ((voltage->vddci & 0xff00) == 0xff00)
|
2011-06-20 21:00:31 +04:00
|
|
|
return;
|
2011-04-12 22:49:24 +04:00
|
|
|
if (voltage->vddci && (voltage->vddci != rdev->pm.current_vddci)) {
|
|
|
|
radeon_atom_set_voltage(rdev, voltage->vddci, SET_VOLTAGE_TYPE_ASIC_VDDCI);
|
|
|
|
rdev->pm.current_vddci = voltage->vddci;
|
|
|
|
DRM_DEBUG("Setting: vddci: %d\n", voltage->vddci);
|
2010-06-08 02:15:18 +04:00
|
|
|
}
|
|
|
|
}
|
2010-04-24 01:57:27 +04:00
|
|
|
}
|
|
|
|
|
2012-07-17 22:02:42 +04:00
|
|
|
/**
|
|
|
|
* evergreen_pm_prepare - pre-power state change callback.
|
|
|
|
*
|
|
|
|
* @rdev: radeon_device pointer
|
|
|
|
*
|
|
|
|
* Prepare for a power state change (evergreen+).
|
|
|
|
*/
|
2010-04-24 01:57:27 +04:00
|
|
|
void evergreen_pm_prepare(struct radeon_device *rdev)
|
|
|
|
{
|
|
|
|
struct drm_device *ddev = rdev->ddev;
|
|
|
|
struct drm_crtc *crtc;
|
|
|
|
struct radeon_crtc *radeon_crtc;
|
|
|
|
u32 tmp;
|
|
|
|
|
|
|
|
/* disable any active CRTCs */
|
|
|
|
list_for_each_entry(crtc, &ddev->mode_config.crtc_list, head) {
|
|
|
|
radeon_crtc = to_radeon_crtc(crtc);
|
|
|
|
if (radeon_crtc->enabled) {
|
|
|
|
tmp = RREG32(EVERGREEN_CRTC_CONTROL + radeon_crtc->crtc_offset);
|
|
|
|
tmp |= EVERGREEN_CRTC_DISP_READ_REQUEST_DISABLE;
|
|
|
|
WREG32(EVERGREEN_CRTC_CONTROL + radeon_crtc->crtc_offset, tmp);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2012-07-17 22:02:42 +04:00
|
|
|
/**
|
|
|
|
* evergreen_pm_finish - post-power state change callback.
|
|
|
|
*
|
|
|
|
* @rdev: radeon_device pointer
|
|
|
|
*
|
|
|
|
* Clean up after a power state change (evergreen+).
|
|
|
|
*/
|
2010-04-24 01:57:27 +04:00
|
|
|
void evergreen_pm_finish(struct radeon_device *rdev)
|
|
|
|
{
|
|
|
|
struct drm_device *ddev = rdev->ddev;
|
|
|
|
struct drm_crtc *crtc;
|
|
|
|
struct radeon_crtc *radeon_crtc;
|
|
|
|
u32 tmp;
|
|
|
|
|
|
|
|
/* enable any active CRTCs */
|
|
|
|
list_for_each_entry(crtc, &ddev->mode_config.crtc_list, head) {
|
|
|
|
radeon_crtc = to_radeon_crtc(crtc);
|
|
|
|
if (radeon_crtc->enabled) {
|
|
|
|
tmp = RREG32(EVERGREEN_CRTC_CONTROL + radeon_crtc->crtc_offset);
|
|
|
|
tmp &= ~EVERGREEN_CRTC_DISP_READ_REQUEST_DISABLE;
|
|
|
|
WREG32(EVERGREEN_CRTC_CONTROL + radeon_crtc->crtc_offset, tmp);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2012-07-17 22:02:42 +04:00
|
|
|
/**
|
|
|
|
* evergreen_hpd_sense - hpd sense callback.
|
|
|
|
*
|
|
|
|
* @rdev: radeon_device pointer
|
|
|
|
* @hpd: hpd (hotplug detect) pin
|
|
|
|
*
|
|
|
|
* Checks if a digital monitor is connected (evergreen+).
|
|
|
|
* Returns true if connected, false if not connected.
|
|
|
|
*/
|
2010-01-13 01:54:34 +03:00
|
|
|
bool evergreen_hpd_sense(struct radeon_device *rdev, enum radeon_hpd_id hpd)
|
|
|
|
{
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
if (hpd == RADEON_HPD_NONE)
|
|
|
|
return false;
|
2010-02-26 21:57:45 +03:00
|
|
|
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
return !!(RREG32(DC_HPDx_INT_STATUS_REG(hpd)) & DC_HPDx_SENSE);
|
2010-01-13 01:54:34 +03:00
|
|
|
}
|
|
|
|
|
2012-07-17 22:02:42 +04:00
|
|
|
/**
|
|
|
|
* evergreen_hpd_set_polarity - hpd set polarity callback.
|
|
|
|
*
|
|
|
|
* @rdev: radeon_device pointer
|
|
|
|
* @hpd: hpd (hotplug detect) pin
|
|
|
|
*
|
|
|
|
* Set the polarity of the hpd pin (evergreen+).
|
|
|
|
*/
|
2010-01-13 01:54:34 +03:00
|
|
|
void evergreen_hpd_set_polarity(struct radeon_device *rdev,
|
|
|
|
enum radeon_hpd_id hpd)
|
|
|
|
{
|
2010-02-26 21:57:45 +03:00
|
|
|
bool connected = evergreen_hpd_sense(rdev, hpd);
|
|
|
|
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
if (hpd == RADEON_HPD_NONE)
|
|
|
|
return;
|
|
|
|
|
|
|
|
if (connected)
|
|
|
|
WREG32_AND(DC_HPDx_INT_CONTROL(hpd), ~DC_HPDx_INT_POLARITY);
|
|
|
|
else
|
|
|
|
WREG32_OR(DC_HPDx_INT_CONTROL(hpd), DC_HPDx_INT_POLARITY);
|
2010-01-13 01:54:34 +03:00
|
|
|
}
|
|
|
|
|
2012-07-17 22:02:42 +04:00
|
|
|
/**
|
|
|
|
* evergreen_hpd_init - hpd setup callback.
|
|
|
|
*
|
|
|
|
* @rdev: radeon_device pointer
|
|
|
|
*
|
|
|
|
* Setup the hpd pins used by the card (evergreen+).
|
|
|
|
* Enable the pin, set the polarity, and enable the hpd interrupts.
|
|
|
|
*/
|
2010-01-13 01:54:34 +03:00
|
|
|
void evergreen_hpd_init(struct radeon_device *rdev)
|
|
|
|
{
|
2010-02-26 21:57:45 +03:00
|
|
|
struct drm_device *dev = rdev->ddev;
|
|
|
|
struct drm_connector *connector;
|
2012-05-17 03:33:30 +04:00
|
|
|
unsigned enabled = 0;
|
2010-02-26 21:57:45 +03:00
|
|
|
u32 tmp = DC_HPDx_CONNECTION_TIMER(0x9c4) |
|
|
|
|
DC_HPDx_RX_INT_TIMER(0xfa) | DC_HPDx_EN;
|
2010-01-13 01:54:34 +03:00
|
|
|
|
2010-02-26 21:57:45 +03:00
|
|
|
list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
enum radeon_hpd_id hpd =
|
|
|
|
to_radeon_connector(connector)->hpd.hpd;
|
2013-04-11 20:45:34 +04:00
|
|
|
|
|
|
|
if (connector->connector_type == DRM_MODE_CONNECTOR_eDP ||
|
|
|
|
connector->connector_type == DRM_MODE_CONNECTOR_LVDS) {
|
|
|
|
/* don't try to enable hpd on eDP or LVDS avoid breaking the
|
|
|
|
* aux dp channel on imac and help (but not completely fix)
|
|
|
|
* https://bugzilla.redhat.com/show_bug.cgi?id=726143
|
|
|
|
* also avoid interrupt storms during dpms.
|
|
|
|
*/
|
|
|
|
continue;
|
|
|
|
}
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
|
|
|
|
if (hpd == RADEON_HPD_NONE)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
WREG32(DC_HPDx_CONTROL(hpd), tmp);
|
|
|
|
enabled |= 1 << hpd;
|
|
|
|
|
|
|
|
radeon_hpd_set_polarity(rdev, hpd);
|
2010-02-26 21:57:45 +03:00
|
|
|
}
|
2012-05-17 03:33:30 +04:00
|
|
|
radeon_irq_kms_enable_hpd(rdev, enabled);
|
2010-01-13 01:54:34 +03:00
|
|
|
}
|
|
|
|
|
2012-07-17 22:02:42 +04:00
|
|
|
/**
|
|
|
|
* evergreen_hpd_fini - hpd tear down callback.
|
|
|
|
*
|
|
|
|
* @rdev: radeon_device pointer
|
|
|
|
*
|
|
|
|
* Tear down the hpd pins used by the card (evergreen+).
|
|
|
|
* Disable the hpd interrupts.
|
|
|
|
*/
|
2010-02-26 21:57:45 +03:00
|
|
|
void evergreen_hpd_fini(struct radeon_device *rdev)
|
2010-01-13 01:54:34 +03:00
|
|
|
{
|
2010-02-26 21:57:45 +03:00
|
|
|
struct drm_device *dev = rdev->ddev;
|
|
|
|
struct drm_connector *connector;
|
2012-05-17 03:33:30 +04:00
|
|
|
unsigned disabled = 0;
|
2010-02-26 21:57:45 +03:00
|
|
|
|
|
|
|
list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
enum radeon_hpd_id hpd =
|
|
|
|
to_radeon_connector(connector)->hpd.hpd;
|
|
|
|
|
|
|
|
if (hpd == RADEON_HPD_NONE)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
WREG32(DC_HPDx_CONTROL(hpd), 0);
|
|
|
|
disabled |= 1 << hpd;
|
2010-02-26 21:57:45 +03:00
|
|
|
}
|
2012-05-17 03:33:30 +04:00
|
|
|
radeon_irq_kms_disable_hpd(rdev, disabled);
|
2010-01-13 01:54:34 +03:00
|
|
|
}
|
|
|
|
|
2010-10-22 10:51:05 +04:00
|
|
|
/* watermark setup */
|
|
|
|
|
|
|
|
static u32 evergreen_line_buffer_adjust(struct radeon_device *rdev,
|
|
|
|
struct radeon_crtc *radeon_crtc,
|
|
|
|
struct drm_display_mode *mode,
|
|
|
|
struct drm_display_mode *other_mode)
|
|
|
|
{
|
2013-08-19 19:06:50 +04:00
|
|
|
u32 tmp, buffer_alloc, i;
|
|
|
|
u32 pipe_offset = radeon_crtc->crtc_id * 0x20;
|
2010-10-22 10:51:05 +04:00
|
|
|
/*
|
|
|
|
* Line Buffer Setup
|
|
|
|
* There are 3 line buffers, each one shared by 2 display controllers.
|
|
|
|
* DC_LB_MEMORY_SPLIT controls how that line buffer is shared between
|
|
|
|
* the display controllers. The paritioning is done via one of four
|
|
|
|
* preset allocations specified in bits 2:0:
|
|
|
|
* first display controller
|
|
|
|
* 0 - first half of lb (3840 * 2)
|
|
|
|
* 1 - first 3/4 of lb (5760 * 2)
|
2011-04-15 03:07:34 +04:00
|
|
|
* 2 - whole lb (7680 * 2), other crtc must be disabled
|
2010-10-22 10:51:05 +04:00
|
|
|
* 3 - first 1/4 of lb (1920 * 2)
|
|
|
|
* second display controller
|
|
|
|
* 4 - second half of lb (3840 * 2)
|
|
|
|
* 5 - second 3/4 of lb (5760 * 2)
|
2011-04-15 03:07:34 +04:00
|
|
|
* 6 - whole lb (7680 * 2), other crtc must be disabled
|
2010-10-22 10:51:05 +04:00
|
|
|
* 7 - last 1/4 of lb (1920 * 2)
|
|
|
|
*/
|
2011-04-15 03:07:34 +04:00
|
|
|
/* this can get tricky if we have two large displays on a paired group
|
|
|
|
* of crtcs. Ideally for multiple large displays we'd assign them to
|
|
|
|
* non-linked crtcs for maximum line buffer allocation.
|
|
|
|
*/
|
|
|
|
if (radeon_crtc->base.enabled && mode) {
|
2013-08-19 19:06:50 +04:00
|
|
|
if (other_mode) {
|
2010-10-22 10:51:05 +04:00
|
|
|
tmp = 0; /* 1/2 */
|
2013-08-19 19:06:50 +04:00
|
|
|
buffer_alloc = 1;
|
|
|
|
} else {
|
2011-04-15 03:07:34 +04:00
|
|
|
tmp = 2; /* whole */
|
2013-08-19 19:06:50 +04:00
|
|
|
buffer_alloc = 2;
|
|
|
|
}
|
|
|
|
} else {
|
2011-04-15 03:07:34 +04:00
|
|
|
tmp = 0;
|
2013-08-19 19:06:50 +04:00
|
|
|
buffer_alloc = 0;
|
|
|
|
}
|
2010-10-22 10:51:05 +04:00
|
|
|
|
|
|
|
/* second controller of the pair uses second half of the lb */
|
|
|
|
if (radeon_crtc->crtc_id % 2)
|
|
|
|
tmp += 4;
|
|
|
|
WREG32(DC_LB_MEMORY_SPLIT + radeon_crtc->crtc_offset, tmp);
|
|
|
|
|
2013-08-19 19:06:50 +04:00
|
|
|
if (ASIC_IS_DCE41(rdev) || ASIC_IS_DCE5(rdev)) {
|
|
|
|
WREG32(PIPE0_DMIF_BUFFER_CONTROL + pipe_offset,
|
|
|
|
DMIF_BUFFERS_ALLOCATED(buffer_alloc));
|
|
|
|
for (i = 0; i < rdev->usec_timeout; i++) {
|
|
|
|
if (RREG32(PIPE0_DMIF_BUFFER_CONTROL + pipe_offset) &
|
|
|
|
DMIF_BUFFERS_ALLOCATED_COMPLETED)
|
|
|
|
break;
|
|
|
|
udelay(1);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2011-04-15 03:07:34 +04:00
|
|
|
if (radeon_crtc->base.enabled && mode) {
|
|
|
|
switch (tmp) {
|
|
|
|
case 0:
|
|
|
|
case 4:
|
|
|
|
default:
|
|
|
|
if (ASIC_IS_DCE5(rdev))
|
|
|
|
return 4096 * 2;
|
|
|
|
else
|
|
|
|
return 3840 * 2;
|
|
|
|
case 1:
|
|
|
|
case 5:
|
|
|
|
if (ASIC_IS_DCE5(rdev))
|
|
|
|
return 6144 * 2;
|
|
|
|
else
|
|
|
|
return 5760 * 2;
|
|
|
|
case 2:
|
|
|
|
case 6:
|
|
|
|
if (ASIC_IS_DCE5(rdev))
|
|
|
|
return 8192 * 2;
|
|
|
|
else
|
|
|
|
return 7680 * 2;
|
|
|
|
case 3:
|
|
|
|
case 7:
|
|
|
|
if (ASIC_IS_DCE5(rdev))
|
|
|
|
return 2048 * 2;
|
|
|
|
else
|
|
|
|
return 1920 * 2;
|
|
|
|
}
|
2010-10-22 10:51:05 +04:00
|
|
|
}
|
2011-04-15 03:07:34 +04:00
|
|
|
|
|
|
|
/* controller not enabled, so no lb used */
|
|
|
|
return 0;
|
2010-10-22 10:51:05 +04:00
|
|
|
}
|
|
|
|
|
2012-03-21 01:18:30 +04:00
|
|
|
u32 evergreen_get_number_of_dram_channels(struct radeon_device *rdev)
|
2010-10-22 10:51:05 +04:00
|
|
|
{
|
|
|
|
u32 tmp = RREG32(MC_SHARED_CHMAP);
|
|
|
|
|
|
|
|
switch ((tmp & NOOFCHAN_MASK) >> NOOFCHAN_SHIFT) {
|
|
|
|
case 0:
|
|
|
|
default:
|
|
|
|
return 1;
|
|
|
|
case 1:
|
|
|
|
return 2;
|
|
|
|
case 2:
|
|
|
|
return 4;
|
|
|
|
case 3:
|
|
|
|
return 8;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
struct evergreen_wm_params {
|
|
|
|
u32 dram_channels; /* number of dram channels */
|
|
|
|
u32 yclk; /* bandwidth per dram data pin in kHz */
|
|
|
|
u32 sclk; /* engine clock in kHz */
|
|
|
|
u32 disp_clk; /* display clock in kHz */
|
|
|
|
u32 src_width; /* viewport width */
|
|
|
|
u32 active_time; /* active display time in ns */
|
|
|
|
u32 blank_time; /* blank time in ns */
|
|
|
|
bool interlaced; /* mode is interlaced */
|
|
|
|
fixed20_12 vsc; /* vertical scale ratio */
|
|
|
|
u32 num_heads; /* number of active crtcs */
|
|
|
|
u32 bytes_per_pixel; /* bytes per pixel display + overlay */
|
|
|
|
u32 lb_size; /* line buffer allocated to pipe */
|
|
|
|
u32 vtaps; /* vertical scaler taps */
|
|
|
|
};
|
|
|
|
|
|
|
|
static u32 evergreen_dram_bandwidth(struct evergreen_wm_params *wm)
|
|
|
|
{
|
|
|
|
/* Calculate DRAM Bandwidth and the part allocated to display. */
|
|
|
|
fixed20_12 dram_efficiency; /* 0.7 */
|
|
|
|
fixed20_12 yclk, dram_channels, bandwidth;
|
|
|
|
fixed20_12 a;
|
|
|
|
|
|
|
|
a.full = dfixed_const(1000);
|
|
|
|
yclk.full = dfixed_const(wm->yclk);
|
|
|
|
yclk.full = dfixed_div(yclk, a);
|
|
|
|
dram_channels.full = dfixed_const(wm->dram_channels * 4);
|
|
|
|
a.full = dfixed_const(10);
|
|
|
|
dram_efficiency.full = dfixed_const(7);
|
|
|
|
dram_efficiency.full = dfixed_div(dram_efficiency, a);
|
|
|
|
bandwidth.full = dfixed_mul(dram_channels, yclk);
|
|
|
|
bandwidth.full = dfixed_mul(bandwidth, dram_efficiency);
|
|
|
|
|
|
|
|
return dfixed_trunc(bandwidth);
|
|
|
|
}
|
|
|
|
|
|
|
|
static u32 evergreen_dram_bandwidth_for_display(struct evergreen_wm_params *wm)
|
|
|
|
{
|
|
|
|
/* Calculate DRAM Bandwidth and the part allocated to display. */
|
|
|
|
fixed20_12 disp_dram_allocation; /* 0.3 to 0.7 */
|
|
|
|
fixed20_12 yclk, dram_channels, bandwidth;
|
|
|
|
fixed20_12 a;
|
|
|
|
|
|
|
|
a.full = dfixed_const(1000);
|
|
|
|
yclk.full = dfixed_const(wm->yclk);
|
|
|
|
yclk.full = dfixed_div(yclk, a);
|
|
|
|
dram_channels.full = dfixed_const(wm->dram_channels * 4);
|
|
|
|
a.full = dfixed_const(10);
|
|
|
|
disp_dram_allocation.full = dfixed_const(3); /* XXX worse case value 0.3 */
|
|
|
|
disp_dram_allocation.full = dfixed_div(disp_dram_allocation, a);
|
|
|
|
bandwidth.full = dfixed_mul(dram_channels, yclk);
|
|
|
|
bandwidth.full = dfixed_mul(bandwidth, disp_dram_allocation);
|
|
|
|
|
|
|
|
return dfixed_trunc(bandwidth);
|
|
|
|
}
|
|
|
|
|
|
|
|
static u32 evergreen_data_return_bandwidth(struct evergreen_wm_params *wm)
|
|
|
|
{
|
|
|
|
/* Calculate the display Data return Bandwidth */
|
|
|
|
fixed20_12 return_efficiency; /* 0.8 */
|
|
|
|
fixed20_12 sclk, bandwidth;
|
|
|
|
fixed20_12 a;
|
|
|
|
|
|
|
|
a.full = dfixed_const(1000);
|
|
|
|
sclk.full = dfixed_const(wm->sclk);
|
|
|
|
sclk.full = dfixed_div(sclk, a);
|
|
|
|
a.full = dfixed_const(10);
|
|
|
|
return_efficiency.full = dfixed_const(8);
|
|
|
|
return_efficiency.full = dfixed_div(return_efficiency, a);
|
|
|
|
a.full = dfixed_const(32);
|
|
|
|
bandwidth.full = dfixed_mul(a, sclk);
|
|
|
|
bandwidth.full = dfixed_mul(bandwidth, return_efficiency);
|
|
|
|
|
|
|
|
return dfixed_trunc(bandwidth);
|
|
|
|
}
|
|
|
|
|
|
|
|
static u32 evergreen_dmif_request_bandwidth(struct evergreen_wm_params *wm)
|
|
|
|
{
|
|
|
|
/* Calculate the DMIF Request Bandwidth */
|
|
|
|
fixed20_12 disp_clk_request_efficiency; /* 0.8 */
|
|
|
|
fixed20_12 disp_clk, bandwidth;
|
|
|
|
fixed20_12 a;
|
|
|
|
|
|
|
|
a.full = dfixed_const(1000);
|
|
|
|
disp_clk.full = dfixed_const(wm->disp_clk);
|
|
|
|
disp_clk.full = dfixed_div(disp_clk, a);
|
|
|
|
a.full = dfixed_const(10);
|
|
|
|
disp_clk_request_efficiency.full = dfixed_const(8);
|
|
|
|
disp_clk_request_efficiency.full = dfixed_div(disp_clk_request_efficiency, a);
|
|
|
|
a.full = dfixed_const(32);
|
|
|
|
bandwidth.full = dfixed_mul(a, disp_clk);
|
|
|
|
bandwidth.full = dfixed_mul(bandwidth, disp_clk_request_efficiency);
|
|
|
|
|
|
|
|
return dfixed_trunc(bandwidth);
|
|
|
|
}
|
|
|
|
|
|
|
|
static u32 evergreen_available_bandwidth(struct evergreen_wm_params *wm)
|
|
|
|
{
|
|
|
|
/* Calculate the Available bandwidth. Display can use this temporarily but not in average. */
|
|
|
|
u32 dram_bandwidth = evergreen_dram_bandwidth(wm);
|
|
|
|
u32 data_return_bandwidth = evergreen_data_return_bandwidth(wm);
|
|
|
|
u32 dmif_req_bandwidth = evergreen_dmif_request_bandwidth(wm);
|
|
|
|
|
|
|
|
return min(dram_bandwidth, min(data_return_bandwidth, dmif_req_bandwidth));
|
|
|
|
}
|
|
|
|
|
|
|
|
static u32 evergreen_average_bandwidth(struct evergreen_wm_params *wm)
|
|
|
|
{
|
|
|
|
/* Calculate the display mode Average Bandwidth
|
|
|
|
* DisplayMode should contain the source and destination dimensions,
|
|
|
|
* timing, etc.
|
|
|
|
*/
|
|
|
|
fixed20_12 bpp;
|
|
|
|
fixed20_12 line_time;
|
|
|
|
fixed20_12 src_width;
|
|
|
|
fixed20_12 bandwidth;
|
|
|
|
fixed20_12 a;
|
|
|
|
|
|
|
|
a.full = dfixed_const(1000);
|
|
|
|
line_time.full = dfixed_const(wm->active_time + wm->blank_time);
|
|
|
|
line_time.full = dfixed_div(line_time, a);
|
|
|
|
bpp.full = dfixed_const(wm->bytes_per_pixel);
|
|
|
|
src_width.full = dfixed_const(wm->src_width);
|
|
|
|
bandwidth.full = dfixed_mul(src_width, bpp);
|
|
|
|
bandwidth.full = dfixed_mul(bandwidth, wm->vsc);
|
|
|
|
bandwidth.full = dfixed_div(bandwidth, line_time);
|
|
|
|
|
|
|
|
return dfixed_trunc(bandwidth);
|
|
|
|
}
|
|
|
|
|
|
|
|
static u32 evergreen_latency_watermark(struct evergreen_wm_params *wm)
|
|
|
|
{
|
|
|
|
/* First calcualte the latency in ns */
|
|
|
|
u32 mc_latency = 2000; /* 2000 ns. */
|
|
|
|
u32 available_bandwidth = evergreen_available_bandwidth(wm);
|
|
|
|
u32 worst_chunk_return_time = (512 * 8 * 1000) / available_bandwidth;
|
|
|
|
u32 cursor_line_pair_return_time = (128 * 4 * 1000) / available_bandwidth;
|
|
|
|
u32 dc_latency = 40000000 / wm->disp_clk; /* dc pipe latency */
|
|
|
|
u32 other_heads_data_return_time = ((wm->num_heads + 1) * worst_chunk_return_time) +
|
|
|
|
(wm->num_heads * cursor_line_pair_return_time);
|
|
|
|
u32 latency = mc_latency + other_heads_data_return_time + dc_latency;
|
|
|
|
u32 max_src_lines_per_dst_line, lb_fill_bw, line_fill_time;
|
|
|
|
fixed20_12 a, b, c;
|
|
|
|
|
|
|
|
if (wm->num_heads == 0)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
a.full = dfixed_const(2);
|
|
|
|
b.full = dfixed_const(1);
|
|
|
|
if ((wm->vsc.full > a.full) ||
|
|
|
|
((wm->vsc.full > b.full) && (wm->vtaps >= 3)) ||
|
|
|
|
(wm->vtaps >= 5) ||
|
|
|
|
((wm->vsc.full >= a.full) && wm->interlaced))
|
|
|
|
max_src_lines_per_dst_line = 4;
|
|
|
|
else
|
|
|
|
max_src_lines_per_dst_line = 2;
|
|
|
|
|
|
|
|
a.full = dfixed_const(available_bandwidth);
|
|
|
|
b.full = dfixed_const(wm->num_heads);
|
|
|
|
a.full = dfixed_div(a, b);
|
|
|
|
|
2017-04-24 02:33:08 +03:00
|
|
|
lb_fill_bw = min(dfixed_trunc(a), wm->disp_clk * wm->bytes_per_pixel / 1000);
|
2010-10-22 10:51:05 +04:00
|
|
|
|
|
|
|
a.full = dfixed_const(max_src_lines_per_dst_line * wm->src_width * wm->bytes_per_pixel);
|
|
|
|
b.full = dfixed_const(1000);
|
|
|
|
c.full = dfixed_const(lb_fill_bw);
|
|
|
|
b.full = dfixed_div(c, b);
|
|
|
|
a.full = dfixed_div(a, b);
|
|
|
|
line_fill_time = dfixed_trunc(a);
|
|
|
|
|
|
|
|
if (line_fill_time < wm->active_time)
|
|
|
|
return latency;
|
|
|
|
else
|
|
|
|
return latency + (line_fill_time - wm->active_time);
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
static bool evergreen_average_bandwidth_vs_dram_bandwidth_for_display(struct evergreen_wm_params *wm)
|
|
|
|
{
|
|
|
|
if (evergreen_average_bandwidth(wm) <=
|
|
|
|
(evergreen_dram_bandwidth_for_display(wm) / wm->num_heads))
|
|
|
|
return true;
|
|
|
|
else
|
|
|
|
return false;
|
|
|
|
};
|
|
|
|
|
|
|
|
static bool evergreen_average_bandwidth_vs_available_bandwidth(struct evergreen_wm_params *wm)
|
|
|
|
{
|
|
|
|
if (evergreen_average_bandwidth(wm) <=
|
|
|
|
(evergreen_available_bandwidth(wm) / wm->num_heads))
|
|
|
|
return true;
|
|
|
|
else
|
|
|
|
return false;
|
|
|
|
};
|
|
|
|
|
|
|
|
static bool evergreen_check_latency_hiding(struct evergreen_wm_params *wm)
|
|
|
|
{
|
|
|
|
u32 lb_partitions = wm->lb_size / wm->src_width;
|
|
|
|
u32 line_time = wm->active_time + wm->blank_time;
|
|
|
|
u32 latency_tolerant_lines;
|
|
|
|
u32 latency_hiding;
|
|
|
|
fixed20_12 a;
|
|
|
|
|
|
|
|
a.full = dfixed_const(1);
|
|
|
|
if (wm->vsc.full > a.full)
|
|
|
|
latency_tolerant_lines = 1;
|
|
|
|
else {
|
|
|
|
if (lb_partitions <= (wm->vtaps + 1))
|
|
|
|
latency_tolerant_lines = 1;
|
|
|
|
else
|
|
|
|
latency_tolerant_lines = 2;
|
|
|
|
}
|
|
|
|
|
|
|
|
latency_hiding = (latency_tolerant_lines * line_time + wm->blank_time);
|
|
|
|
|
|
|
|
if (evergreen_latency_watermark(wm) <= latency_hiding)
|
|
|
|
return true;
|
|
|
|
else
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void evergreen_program_watermarks(struct radeon_device *rdev,
|
|
|
|
struct radeon_crtc *radeon_crtc,
|
|
|
|
u32 lb_size, u32 num_heads)
|
|
|
|
{
|
|
|
|
struct drm_display_mode *mode = &radeon_crtc->base.mode;
|
2012-03-14 00:25:11 +04:00
|
|
|
struct evergreen_wm_params wm_low, wm_high;
|
|
|
|
u32 dram_channels;
|
2017-04-24 02:33:09 +03:00
|
|
|
u32 active_time;
|
2010-10-22 10:51:05 +04:00
|
|
|
u32 line_time = 0;
|
|
|
|
u32 latency_watermark_a = 0, latency_watermark_b = 0;
|
|
|
|
u32 priority_a_mark = 0, priority_b_mark = 0;
|
|
|
|
u32 priority_a_cnt = PRIORITY_OFF;
|
|
|
|
u32 priority_b_cnt = PRIORITY_OFF;
|
|
|
|
u32 pipe_offset = radeon_crtc->crtc_id * 16;
|
|
|
|
u32 tmp, arb_control3;
|
|
|
|
fixed20_12 a, b, c;
|
|
|
|
|
|
|
|
if (radeon_crtc->base.enabled && num_heads && mode) {
|
2017-06-13 08:17:11 +03:00
|
|
|
active_time = (u32) div_u64((u64)mode->crtc_hdisplay * 1000000,
|
|
|
|
(u32)mode->clock);
|
|
|
|
line_time = (u32) div_u64((u64)mode->crtc_htotal * 1000000,
|
|
|
|
(u32)mode->clock);
|
|
|
|
line_time = min(line_time, (u32)65535);
|
2010-10-22 10:51:05 +04:00
|
|
|
priority_a_cnt = 0;
|
|
|
|
priority_b_cnt = 0;
|
2012-03-14 00:25:11 +04:00
|
|
|
dram_channels = evergreen_get_number_of_dram_channels(rdev);
|
|
|
|
|
|
|
|
/* watermark for high clocks */
|
|
|
|
if ((rdev->pm.pm_method == PM_METHOD_DPM) && rdev->pm.dpm_enabled) {
|
|
|
|
wm_high.yclk =
|
|
|
|
radeon_dpm_get_mclk(rdev, false) * 10;
|
|
|
|
wm_high.sclk =
|
|
|
|
radeon_dpm_get_sclk(rdev, false) * 10;
|
|
|
|
} else {
|
|
|
|
wm_high.yclk = rdev->pm.current_mclk * 10;
|
|
|
|
wm_high.sclk = rdev->pm.current_sclk * 10;
|
|
|
|
}
|
2010-10-22 10:51:05 +04:00
|
|
|
|
2012-03-14 00:25:11 +04:00
|
|
|
wm_high.disp_clk = mode->clock;
|
|
|
|
wm_high.src_width = mode->crtc_hdisplay;
|
2017-04-24 02:33:09 +03:00
|
|
|
wm_high.active_time = active_time;
|
2012-03-14 00:25:11 +04:00
|
|
|
wm_high.blank_time = line_time - wm_high.active_time;
|
|
|
|
wm_high.interlaced = false;
|
2010-10-22 10:51:05 +04:00
|
|
|
if (mode->flags & DRM_MODE_FLAG_INTERLACE)
|
2012-03-14 00:25:11 +04:00
|
|
|
wm_high.interlaced = true;
|
|
|
|
wm_high.vsc = radeon_crtc->vsc;
|
|
|
|
wm_high.vtaps = 1;
|
2010-10-22 10:51:05 +04:00
|
|
|
if (radeon_crtc->rmx_type != RMX_OFF)
|
2012-03-14 00:25:11 +04:00
|
|
|
wm_high.vtaps = 2;
|
|
|
|
wm_high.bytes_per_pixel = 4; /* XXX: get this from fb config */
|
|
|
|
wm_high.lb_size = lb_size;
|
|
|
|
wm_high.dram_channels = dram_channels;
|
|
|
|
wm_high.num_heads = num_heads;
|
|
|
|
|
|
|
|
/* watermark for low clocks */
|
|
|
|
if ((rdev->pm.pm_method == PM_METHOD_DPM) && rdev->pm.dpm_enabled) {
|
|
|
|
wm_low.yclk =
|
|
|
|
radeon_dpm_get_mclk(rdev, true) * 10;
|
|
|
|
wm_low.sclk =
|
|
|
|
radeon_dpm_get_sclk(rdev, true) * 10;
|
|
|
|
} else {
|
|
|
|
wm_low.yclk = rdev->pm.current_mclk * 10;
|
|
|
|
wm_low.sclk = rdev->pm.current_sclk * 10;
|
|
|
|
}
|
|
|
|
|
|
|
|
wm_low.disp_clk = mode->clock;
|
|
|
|
wm_low.src_width = mode->crtc_hdisplay;
|
2017-04-24 02:33:09 +03:00
|
|
|
wm_low.active_time = active_time;
|
2012-03-14 00:25:11 +04:00
|
|
|
wm_low.blank_time = line_time - wm_low.active_time;
|
|
|
|
wm_low.interlaced = false;
|
|
|
|
if (mode->flags & DRM_MODE_FLAG_INTERLACE)
|
|
|
|
wm_low.interlaced = true;
|
|
|
|
wm_low.vsc = radeon_crtc->vsc;
|
|
|
|
wm_low.vtaps = 1;
|
|
|
|
if (radeon_crtc->rmx_type != RMX_OFF)
|
|
|
|
wm_low.vtaps = 2;
|
|
|
|
wm_low.bytes_per_pixel = 4; /* XXX: get this from fb config */
|
|
|
|
wm_low.lb_size = lb_size;
|
|
|
|
wm_low.dram_channels = dram_channels;
|
|
|
|
wm_low.num_heads = num_heads;
|
2010-10-22 10:51:05 +04:00
|
|
|
|
|
|
|
/* set for high clocks */
|
2012-03-14 00:25:11 +04:00
|
|
|
latency_watermark_a = min(evergreen_latency_watermark(&wm_high), (u32)65535);
|
2010-10-22 10:51:05 +04:00
|
|
|
/* set for low clocks */
|
2012-03-14 00:25:11 +04:00
|
|
|
latency_watermark_b = min(evergreen_latency_watermark(&wm_low), (u32)65535);
|
2010-10-22 10:51:05 +04:00
|
|
|
|
|
|
|
/* possibly force display priority to high */
|
|
|
|
/* should really do this at mode validation time... */
|
2012-03-14 00:25:11 +04:00
|
|
|
if (!evergreen_average_bandwidth_vs_dram_bandwidth_for_display(&wm_high) ||
|
|
|
|
!evergreen_average_bandwidth_vs_available_bandwidth(&wm_high) ||
|
|
|
|
!evergreen_check_latency_hiding(&wm_high) ||
|
2010-10-22 10:51:05 +04:00
|
|
|
(rdev->disp_priority == 2)) {
|
2012-03-14 00:25:11 +04:00
|
|
|
DRM_DEBUG_KMS("force priority a to high\n");
|
2010-10-22 10:51:05 +04:00
|
|
|
priority_a_cnt |= PRIORITY_ALWAYS_ON;
|
2012-03-14 00:25:11 +04:00
|
|
|
}
|
|
|
|
if (!evergreen_average_bandwidth_vs_dram_bandwidth_for_display(&wm_low) ||
|
|
|
|
!evergreen_average_bandwidth_vs_available_bandwidth(&wm_low) ||
|
|
|
|
!evergreen_check_latency_hiding(&wm_low) ||
|
|
|
|
(rdev->disp_priority == 2)) {
|
|
|
|
DRM_DEBUG_KMS("force priority b to high\n");
|
2010-10-22 10:51:05 +04:00
|
|
|
priority_b_cnt |= PRIORITY_ALWAYS_ON;
|
|
|
|
}
|
|
|
|
|
|
|
|
a.full = dfixed_const(1000);
|
|
|
|
b.full = dfixed_const(mode->clock);
|
|
|
|
b.full = dfixed_div(b, a);
|
|
|
|
c.full = dfixed_const(latency_watermark_a);
|
|
|
|
c.full = dfixed_mul(c, b);
|
|
|
|
c.full = dfixed_mul(c, radeon_crtc->hsc);
|
|
|
|
c.full = dfixed_div(c, a);
|
|
|
|
a.full = dfixed_const(16);
|
|
|
|
c.full = dfixed_div(c, a);
|
|
|
|
priority_a_mark = dfixed_trunc(c);
|
|
|
|
priority_a_cnt |= priority_a_mark & PRIORITY_MARK_MASK;
|
|
|
|
|
|
|
|
a.full = dfixed_const(1000);
|
|
|
|
b.full = dfixed_const(mode->clock);
|
|
|
|
b.full = dfixed_div(b, a);
|
|
|
|
c.full = dfixed_const(latency_watermark_b);
|
|
|
|
c.full = dfixed_mul(c, b);
|
|
|
|
c.full = dfixed_mul(c, radeon_crtc->hsc);
|
|
|
|
c.full = dfixed_div(c, a);
|
|
|
|
a.full = dfixed_const(16);
|
|
|
|
c.full = dfixed_div(c, a);
|
|
|
|
priority_b_mark = dfixed_trunc(c);
|
|
|
|
priority_b_cnt |= priority_b_mark & PRIORITY_MARK_MASK;
|
drm/radeon: Fixup hw vblank counter/ts for new drm_update_vblank_count() (v2)
commit 4dfd6486 "drm: Use vblank timestamps to guesstimate how many
vblanks were missed" introduced in Linux 4.4-rc1 makes the drm core
more fragile to drivers which don't update hw vblank counters and
vblank timestamps in sync with firing of the vblank irq and
essentially at leading edge of vblank.
This exposed a problem with radeon-kms/amdgpu-kms which do not
satisfy above requirements:
The vblank irq fires a few scanlines before start of vblank, but
programmed pageflips complete at start of vblank and
vblank timestamps update at start of vblank, whereas the
hw vblank counter increments only later, at start of vsync.
This leads to problems like off by one errors for vblank counter
updates, vblank counters apparently going backwards or vblank
timestamps apparently having time going backwards. The net result
is stuttering of graphics in games, or little hangs, as well as
total failure of timing sensitive applications.
See bug #93147 for an example of the regression on Linux 4.4-rc:
https://bugs.freedesktop.org/show_bug.cgi?id=93147
This patch tries to align all above events better from the
viewpoint of the drm core / of external callers to fix the problem:
1. The apparent start of vblank is shifted a few scanlines earlier,
so the vblank irq now always happens after start of this extended
vblank interval and thereby drm_update_vblank_count() always samples
the updated vblank count and timestamp of the new vblank interval.
To achieve this, the reporting of scanout positions by
radeon_get_crtc_scanoutpos() now operates as if the vblank starts
radeon_crtc->lb_vblank_lead_lines before the real start of the hw
vblank interval. This means that the vblank timestamps which are based
on these scanout positions will now update at this earlier start of
vblank.
2. The driver->get_vblank_counter() function will bump the returned
vblank count as read from the hw by +1 if the query happens after
the shifted earlier start of the vblank, but before the real hw increment
at start of vsync, so the counter appears to increment at start of vblank
in sync with the timestamp update.
3. Calls from vblank irq-context and regular non-irq calls are now
treated identical, always simulating the shifted vblank start, to
avoid inconsistent results for queries happening from vblank irq vs.
happening from drm_vblank_enable() or vblank_disable_fn().
4. The radeon_flip_work_func will delay mmio programming a pageflip until
the start of the real vblank iff it happens to execute inside the shifted
earlier start of the vblank, so pageflips now also appear to execute at
start of the shifted vblank, in sync with vblank counter and timestamp
updates. This to avoid some races between updates of vblank count and
timestamps that are used for swap scheduling and pageflip execution which
could cause pageflips to execute before the scheduled target vblank.
The lb_vblank_lead_lines "fudge" value is calculated as the size of
the display controllers line buffer in scanlines for the given video
mode: Vblank irq's are triggered by the line buffer logic when the line
buffer refill for a video frame ends, ie. when the line buffer source read
position enters the hw vblank. This means that a vblank irq could fire at
most as many scanlines before the current reported scanout position of the
crtc timing generator as the number of scanlines the line buffer can
maximally hold for a given video mode.
This patch has been successfully tested on a RV730 card with DCE-3 display
engine and on a evergreen card with DCE-4 display engine, in single-display
and dual-display configuration, with different video modes.
A similar patch is needed for amdgpu-kms to fix the same problem.
Limitations:
- Line buffer sizes in pixels are hard-coded on < DCE-4 to a value
i just guessed to be high enough to work ok, lacking info on the true
sizes atm.
Fixes: fdo#93147
Signed-off-by: Mario Kleiner <mario.kleiner.de@gmail.com>
Cc: Alex Deucher <alexander.deucher@amd.com>
Cc: Michel Dänzer <michel.daenzer@amd.com>
Cc: Harry Wentland <Harry.Wentland@amd.com>
Cc: Ville Syrjälä <ville.syrjala@linux.intel.com>
(v1) Tested-by: Dave Witbrodt <dawitbro@sbcglobal.net>
(v2) Refine radeon_flip_work_func() for better efficiency:
In radeon_flip_work_func, replace the busy waiting udelay(5)
with event lock held by a more performance and energy efficient
usleep_range() until at least predicted true start of hw vblank,
with some slack for scheduler happiness. Release the event lock
during waits to not delay other outputs in doing their stuff, as
the waiting can last up to 200 usecs in some cases.
Retested on DCE-3 and DCE-4 to verify it still works nicely.
(v2) Signed-off-by: Mario Kleiner <mario.kleiner.de@gmail.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2015-11-25 22:14:31 +03:00
|
|
|
|
|
|
|
/* Save number of lines the linebuffer leads before the scanout */
|
|
|
|
radeon_crtc->lb_vblank_lead_lines = DIV_ROUND_UP(lb_size, mode->crtc_hdisplay);
|
2010-10-22 10:51:05 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
/* select wm A */
|
|
|
|
arb_control3 = RREG32(PIPE0_ARBITRATION_CONTROL3 + pipe_offset);
|
|
|
|
tmp = arb_control3;
|
|
|
|
tmp &= ~LATENCY_WATERMARK_MASK(3);
|
|
|
|
tmp |= LATENCY_WATERMARK_MASK(1);
|
|
|
|
WREG32(PIPE0_ARBITRATION_CONTROL3 + pipe_offset, tmp);
|
|
|
|
WREG32(PIPE0_LATENCY_CONTROL + pipe_offset,
|
|
|
|
(LATENCY_LOW_WATERMARK(latency_watermark_a) |
|
|
|
|
LATENCY_HIGH_WATERMARK(line_time)));
|
|
|
|
/* select wm B */
|
|
|
|
tmp = RREG32(PIPE0_ARBITRATION_CONTROL3 + pipe_offset);
|
|
|
|
tmp &= ~LATENCY_WATERMARK_MASK(3);
|
|
|
|
tmp |= LATENCY_WATERMARK_MASK(2);
|
|
|
|
WREG32(PIPE0_ARBITRATION_CONTROL3 + pipe_offset, tmp);
|
|
|
|
WREG32(PIPE0_LATENCY_CONTROL + pipe_offset,
|
|
|
|
(LATENCY_LOW_WATERMARK(latency_watermark_b) |
|
|
|
|
LATENCY_HIGH_WATERMARK(line_time)));
|
|
|
|
/* restore original selection */
|
|
|
|
WREG32(PIPE0_ARBITRATION_CONTROL3 + pipe_offset, arb_control3);
|
|
|
|
|
|
|
|
/* write the priority marks */
|
|
|
|
WREG32(PRIORITY_A_CNT + radeon_crtc->crtc_offset, priority_a_cnt);
|
|
|
|
WREG32(PRIORITY_B_CNT + radeon_crtc->crtc_offset, priority_b_cnt);
|
|
|
|
|
2013-03-21 18:38:49 +04:00
|
|
|
/* save values for DPM */
|
|
|
|
radeon_crtc->line_time = line_time;
|
|
|
|
radeon_crtc->wm_high = latency_watermark_a;
|
|
|
|
radeon_crtc->wm_low = latency_watermark_b;
|
2010-10-22 10:51:05 +04:00
|
|
|
}
|
|
|
|
|
2012-07-17 22:02:42 +04:00
|
|
|
/**
|
|
|
|
* evergreen_bandwidth_update - update display watermarks callback.
|
|
|
|
*
|
|
|
|
* @rdev: radeon_device pointer
|
|
|
|
*
|
|
|
|
* Update the display watermarks based on the requested mode(s)
|
|
|
|
* (evergreen+).
|
|
|
|
*/
|
2010-02-26 21:57:45 +03:00
|
|
|
void evergreen_bandwidth_update(struct radeon_device *rdev)
|
2010-01-13 01:54:34 +03:00
|
|
|
{
|
2010-10-22 10:51:05 +04:00
|
|
|
struct drm_display_mode *mode0 = NULL;
|
|
|
|
struct drm_display_mode *mode1 = NULL;
|
|
|
|
u32 num_heads = 0, lb_size;
|
|
|
|
int i;
|
|
|
|
|
2014-11-03 17:57:46 +03:00
|
|
|
if (!rdev->mode_info.mode_config_initialized)
|
|
|
|
return;
|
|
|
|
|
2010-10-22 10:51:05 +04:00
|
|
|
radeon_update_display_priority(rdev);
|
|
|
|
|
|
|
|
for (i = 0; i < rdev->num_crtc; i++) {
|
|
|
|
if (rdev->mode_info.crtcs[i]->base.enabled)
|
|
|
|
num_heads++;
|
|
|
|
}
|
|
|
|
for (i = 0; i < rdev->num_crtc; i += 2) {
|
|
|
|
mode0 = &rdev->mode_info.crtcs[i]->base.mode;
|
|
|
|
mode1 = &rdev->mode_info.crtcs[i+1]->base.mode;
|
|
|
|
lb_size = evergreen_line_buffer_adjust(rdev, rdev->mode_info.crtcs[i], mode0, mode1);
|
|
|
|
evergreen_program_watermarks(rdev, rdev->mode_info.crtcs[i], lb_size, num_heads);
|
|
|
|
lb_size = evergreen_line_buffer_adjust(rdev, rdev->mode_info.crtcs[i+1], mode1, mode0);
|
|
|
|
evergreen_program_watermarks(rdev, rdev->mode_info.crtcs[i+1], lb_size, num_heads);
|
|
|
|
}
|
2010-01-13 01:54:34 +03:00
|
|
|
}
|
|
|
|
|
2012-07-17 22:02:42 +04:00
|
|
|
/**
|
|
|
|
* evergreen_mc_wait_for_idle - wait for MC idle callback.
|
|
|
|
*
|
|
|
|
* @rdev: radeon_device pointer
|
|
|
|
*
|
|
|
|
* Wait for the MC (memory controller) to be idle.
|
|
|
|
* (evergreen+).
|
|
|
|
* Returns 0 if the MC is idle, -1 if not.
|
|
|
|
*/
|
2011-03-03 04:07:33 +03:00
|
|
|
int evergreen_mc_wait_for_idle(struct radeon_device *rdev)
|
2010-01-13 01:54:34 +03:00
|
|
|
{
|
|
|
|
unsigned i;
|
|
|
|
u32 tmp;
|
|
|
|
|
|
|
|
for (i = 0; i < rdev->usec_timeout; i++) {
|
|
|
|
/* read MC_STATUS */
|
|
|
|
tmp = RREG32(SRBM_STATUS) & 0x1F00;
|
|
|
|
if (!tmp)
|
|
|
|
return 0;
|
|
|
|
udelay(1);
|
|
|
|
}
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* GART
|
|
|
|
*/
|
2010-03-24 20:20:41 +03:00
|
|
|
void evergreen_pcie_gart_tlb_flush(struct radeon_device *rdev)
|
|
|
|
{
|
|
|
|
unsigned i;
|
|
|
|
u32 tmp;
|
|
|
|
|
2010-12-15 19:01:56 +03:00
|
|
|
WREG32(HDP_MEM_COHERENCY_FLUSH_CNTL, 0x1);
|
|
|
|
|
2010-03-24 20:20:41 +03:00
|
|
|
WREG32(VM_CONTEXT0_REQUEST_RESPONSE, REQUEST_TYPE(1));
|
|
|
|
for (i = 0; i < rdev->usec_timeout; i++) {
|
|
|
|
/* read MC_STATUS */
|
|
|
|
tmp = RREG32(VM_CONTEXT0_REQUEST_RESPONSE);
|
|
|
|
tmp = (tmp & RESPONSE_TYPE_MASK) >> RESPONSE_TYPE_SHIFT;
|
|
|
|
if (tmp == 2) {
|
2017-02-28 15:55:52 +03:00
|
|
|
pr_warn("[drm] r600 flush TLB failed\n");
|
2010-03-24 20:20:41 +03:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
if (tmp) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
udelay(1);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2012-08-31 21:43:50 +04:00
|
|
|
static int evergreen_pcie_gart_enable(struct radeon_device *rdev)
|
2010-01-13 01:54:34 +03:00
|
|
|
{
|
|
|
|
u32 tmp;
|
2010-03-24 20:20:41 +03:00
|
|
|
int r;
|
2010-01-13 01:54:34 +03:00
|
|
|
|
2011-11-03 19:16:49 +04:00
|
|
|
if (rdev->gart.robj == NULL) {
|
2010-01-13 01:54:34 +03:00
|
|
|
dev_err(rdev->dev, "No VRAM object for PCIE GART.\n");
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
r = radeon_gart_table_vram_pin(rdev);
|
|
|
|
if (r)
|
|
|
|
return r;
|
|
|
|
/* Setup L2 cache */
|
|
|
|
WREG32(VM_L2_CNTL, ENABLE_L2_CACHE | ENABLE_L2_FRAGMENT_PROCESSING |
|
|
|
|
ENABLE_L2_PTE_CACHE_LRU_UPDATE_BY_WRITE |
|
|
|
|
EFFECTIVE_L2_QUEUE_SIZE(7));
|
|
|
|
WREG32(VM_L2_CNTL2, 0);
|
|
|
|
WREG32(VM_L2_CNTL3, BANK_SELECT(0) | CACHE_UPDATE_MODE(2));
|
|
|
|
/* Setup TLB control */
|
|
|
|
tmp = ENABLE_L1_TLB | ENABLE_L1_FRAGMENT_PROCESSING |
|
|
|
|
SYSTEM_ACCESS_MODE_NOT_IN_SYS |
|
|
|
|
SYSTEM_APERTURE_UNMAPPED_ACCESS_PASS_THRU |
|
|
|
|
EFFECTIVE_L1_TLB_SIZE(5) | EFFECTIVE_L1_QUEUE_SIZE(5);
|
2011-05-04 03:28:02 +04:00
|
|
|
if (rdev->flags & RADEON_IS_IGP) {
|
|
|
|
WREG32(FUS_MC_VM_MD_L1_TLB0_CNTL, tmp);
|
|
|
|
WREG32(FUS_MC_VM_MD_L1_TLB1_CNTL, tmp);
|
|
|
|
WREG32(FUS_MC_VM_MD_L1_TLB2_CNTL, tmp);
|
|
|
|
} else {
|
|
|
|
WREG32(MC_VM_MD_L1_TLB0_CNTL, tmp);
|
|
|
|
WREG32(MC_VM_MD_L1_TLB1_CNTL, tmp);
|
|
|
|
WREG32(MC_VM_MD_L1_TLB2_CNTL, tmp);
|
2012-06-01 02:54:43 +04:00
|
|
|
if ((rdev->family == CHIP_JUNIPER) ||
|
|
|
|
(rdev->family == CHIP_CYPRESS) ||
|
|
|
|
(rdev->family == CHIP_HEMLOCK) ||
|
|
|
|
(rdev->family == CHIP_BARTS))
|
|
|
|
WREG32(MC_VM_MD_L1_TLB3_CNTL, tmp);
|
2011-05-04 03:28:02 +04:00
|
|
|
}
|
2010-01-13 01:54:34 +03:00
|
|
|
WREG32(MC_VM_MB_L1_TLB0_CNTL, tmp);
|
|
|
|
WREG32(MC_VM_MB_L1_TLB1_CNTL, tmp);
|
|
|
|
WREG32(MC_VM_MB_L1_TLB2_CNTL, tmp);
|
|
|
|
WREG32(MC_VM_MB_L1_TLB3_CNTL, tmp);
|
|
|
|
WREG32(VM_CONTEXT0_PAGE_TABLE_START_ADDR, rdev->mc.gtt_start >> 12);
|
2015-05-28 16:51:59 +03:00
|
|
|
WREG32(VM_CONTEXT0_PAGE_TABLE_END_ADDR, rdev->mc.gtt_end >> 12);
|
2010-01-13 01:54:34 +03:00
|
|
|
WREG32(VM_CONTEXT0_PAGE_TABLE_BASE_ADDR, rdev->gart.table_addr >> 12);
|
|
|
|
WREG32(VM_CONTEXT0_CNTL, ENABLE_CONTEXT | PAGE_TABLE_DEPTH(0) |
|
|
|
|
RANGE_PROTECTION_FAULT_ENABLE_DEFAULT);
|
|
|
|
WREG32(VM_CONTEXT0_PROTECTION_FAULT_DEFAULT_ADDR,
|
|
|
|
(u32)(rdev->dummy_page.addr >> 12));
|
2010-03-24 20:20:41 +03:00
|
|
|
WREG32(VM_CONTEXT1_CNTL, 0);
|
2010-01-13 01:54:34 +03:00
|
|
|
|
2010-03-24 20:20:41 +03:00
|
|
|
evergreen_pcie_gart_tlb_flush(rdev);
|
2011-09-01 01:54:07 +04:00
|
|
|
DRM_INFO("PCIE GART of %uM enabled (table at 0x%016llX).\n",
|
|
|
|
(unsigned)(rdev->mc.gtt_size >> 20),
|
|
|
|
(unsigned long long)rdev->gart.table_addr);
|
2010-01-13 01:54:34 +03:00
|
|
|
rdev->gart.ready = true;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2012-08-31 21:43:50 +04:00
|
|
|
static void evergreen_pcie_gart_disable(struct radeon_device *rdev)
|
2010-01-13 01:54:34 +03:00
|
|
|
{
|
|
|
|
u32 tmp;
|
|
|
|
|
|
|
|
/* Disable all tables */
|
2010-03-24 20:20:41 +03:00
|
|
|
WREG32(VM_CONTEXT0_CNTL, 0);
|
|
|
|
WREG32(VM_CONTEXT1_CNTL, 0);
|
2010-01-13 01:54:34 +03:00
|
|
|
|
|
|
|
/* Setup L2 cache */
|
|
|
|
WREG32(VM_L2_CNTL, ENABLE_L2_FRAGMENT_PROCESSING |
|
|
|
|
EFFECTIVE_L2_QUEUE_SIZE(7));
|
|
|
|
WREG32(VM_L2_CNTL2, 0);
|
|
|
|
WREG32(VM_L2_CNTL3, BANK_SELECT(0) | CACHE_UPDATE_MODE(2));
|
|
|
|
/* Setup TLB control */
|
|
|
|
tmp = EFFECTIVE_L1_TLB_SIZE(5) | EFFECTIVE_L1_QUEUE_SIZE(5);
|
|
|
|
WREG32(MC_VM_MD_L1_TLB0_CNTL, tmp);
|
|
|
|
WREG32(MC_VM_MD_L1_TLB1_CNTL, tmp);
|
|
|
|
WREG32(MC_VM_MD_L1_TLB2_CNTL, tmp);
|
|
|
|
WREG32(MC_VM_MB_L1_TLB0_CNTL, tmp);
|
|
|
|
WREG32(MC_VM_MB_L1_TLB1_CNTL, tmp);
|
|
|
|
WREG32(MC_VM_MB_L1_TLB2_CNTL, tmp);
|
|
|
|
WREG32(MC_VM_MB_L1_TLB3_CNTL, tmp);
|
2011-11-03 19:16:49 +04:00
|
|
|
radeon_gart_table_vram_unpin(rdev);
|
2010-01-13 01:54:34 +03:00
|
|
|
}
|
|
|
|
|
2012-08-31 21:43:50 +04:00
|
|
|
static void evergreen_pcie_gart_fini(struct radeon_device *rdev)
|
2010-01-13 01:54:34 +03:00
|
|
|
{
|
|
|
|
evergreen_pcie_gart_disable(rdev);
|
|
|
|
radeon_gart_table_vram_free(rdev);
|
|
|
|
radeon_gart_fini(rdev);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2012-08-31 21:43:50 +04:00
|
|
|
static void evergreen_agp_enable(struct radeon_device *rdev)
|
2010-01-13 01:54:34 +03:00
|
|
|
{
|
|
|
|
u32 tmp;
|
|
|
|
|
|
|
|
/* Setup L2 cache */
|
|
|
|
WREG32(VM_L2_CNTL, ENABLE_L2_CACHE | ENABLE_L2_FRAGMENT_PROCESSING |
|
|
|
|
ENABLE_L2_PTE_CACHE_LRU_UPDATE_BY_WRITE |
|
|
|
|
EFFECTIVE_L2_QUEUE_SIZE(7));
|
|
|
|
WREG32(VM_L2_CNTL2, 0);
|
|
|
|
WREG32(VM_L2_CNTL3, BANK_SELECT(0) | CACHE_UPDATE_MODE(2));
|
|
|
|
/* Setup TLB control */
|
|
|
|
tmp = ENABLE_L1_TLB | ENABLE_L1_FRAGMENT_PROCESSING |
|
|
|
|
SYSTEM_ACCESS_MODE_NOT_IN_SYS |
|
|
|
|
SYSTEM_APERTURE_UNMAPPED_ACCESS_PASS_THRU |
|
|
|
|
EFFECTIVE_L1_TLB_SIZE(5) | EFFECTIVE_L1_QUEUE_SIZE(5);
|
|
|
|
WREG32(MC_VM_MD_L1_TLB0_CNTL, tmp);
|
|
|
|
WREG32(MC_VM_MD_L1_TLB1_CNTL, tmp);
|
|
|
|
WREG32(MC_VM_MD_L1_TLB2_CNTL, tmp);
|
|
|
|
WREG32(MC_VM_MB_L1_TLB0_CNTL, tmp);
|
|
|
|
WREG32(MC_VM_MB_L1_TLB1_CNTL, tmp);
|
|
|
|
WREG32(MC_VM_MB_L1_TLB2_CNTL, tmp);
|
|
|
|
WREG32(MC_VM_MB_L1_TLB3_CNTL, tmp);
|
2010-03-24 20:20:41 +03:00
|
|
|
WREG32(VM_CONTEXT0_CNTL, 0);
|
|
|
|
WREG32(VM_CONTEXT1_CNTL, 0);
|
2010-01-13 01:54:34 +03:00
|
|
|
}
|
|
|
|
|
2016-04-14 20:34:03 +03:00
|
|
|
static const unsigned ni_dig_offsets[] =
|
|
|
|
{
|
|
|
|
NI_DIG0_REGISTER_OFFSET,
|
|
|
|
NI_DIG1_REGISTER_OFFSET,
|
|
|
|
NI_DIG2_REGISTER_OFFSET,
|
|
|
|
NI_DIG3_REGISTER_OFFSET,
|
|
|
|
NI_DIG4_REGISTER_OFFSET,
|
|
|
|
NI_DIG5_REGISTER_OFFSET
|
|
|
|
};
|
|
|
|
|
|
|
|
static const unsigned ni_tx_offsets[] =
|
|
|
|
{
|
|
|
|
NI_DCIO_UNIPHY0_UNIPHY_TX_CONTROL1,
|
|
|
|
NI_DCIO_UNIPHY1_UNIPHY_TX_CONTROL1,
|
|
|
|
NI_DCIO_UNIPHY2_UNIPHY_TX_CONTROL1,
|
|
|
|
NI_DCIO_UNIPHY3_UNIPHY_TX_CONTROL1,
|
|
|
|
NI_DCIO_UNIPHY4_UNIPHY_TX_CONTROL1,
|
|
|
|
NI_DCIO_UNIPHY5_UNIPHY_TX_CONTROL1
|
|
|
|
};
|
|
|
|
|
|
|
|
static const unsigned evergreen_dp_offsets[] =
|
|
|
|
{
|
|
|
|
EVERGREEN_DP0_REGISTER_OFFSET,
|
|
|
|
EVERGREEN_DP1_REGISTER_OFFSET,
|
|
|
|
EVERGREEN_DP2_REGISTER_OFFSET,
|
|
|
|
EVERGREEN_DP3_REGISTER_OFFSET,
|
|
|
|
EVERGREEN_DP4_REGISTER_OFFSET,
|
|
|
|
EVERGREEN_DP5_REGISTER_OFFSET
|
|
|
|
};
|
|
|
|
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
static const unsigned evergreen_disp_int_status[] =
|
|
|
|
{
|
|
|
|
DISP_INTERRUPT_STATUS,
|
|
|
|
DISP_INTERRUPT_STATUS_CONTINUE,
|
|
|
|
DISP_INTERRUPT_STATUS_CONTINUE2,
|
|
|
|
DISP_INTERRUPT_STATUS_CONTINUE3,
|
|
|
|
DISP_INTERRUPT_STATUS_CONTINUE4,
|
|
|
|
DISP_INTERRUPT_STATUS_CONTINUE5
|
|
|
|
};
|
2016-04-14 20:34:03 +03:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Assumption is that EVERGREEN_CRTC_MASTER_EN enable for requested crtc
|
|
|
|
* We go from crtc to connector and it is not relible since it
|
|
|
|
* should be an opposite direction .If crtc is enable then
|
|
|
|
* find the dig_fe which selects this crtc and insure that it enable.
|
|
|
|
* if such dig_fe is found then find dig_be which selects found dig_be and
|
|
|
|
* insure that it enable and in DP_SST mode.
|
|
|
|
* if UNIPHY_PLL_CONTROL1.enable then we should disconnect timing
|
|
|
|
* from dp symbols clocks .
|
|
|
|
*/
|
|
|
|
static bool evergreen_is_dp_sst_stream_enabled(struct radeon_device *rdev,
|
|
|
|
unsigned crtc_id, unsigned *ret_dig_fe)
|
|
|
|
{
|
|
|
|
unsigned i;
|
|
|
|
unsigned dig_fe;
|
|
|
|
unsigned dig_be;
|
|
|
|
unsigned dig_en_be;
|
|
|
|
unsigned uniphy_pll;
|
|
|
|
unsigned digs_fe_selected;
|
|
|
|
unsigned dig_be_mode;
|
|
|
|
unsigned dig_fe_mask;
|
|
|
|
bool is_enabled = false;
|
|
|
|
bool found_crtc = false;
|
|
|
|
|
|
|
|
/* loop through all running dig_fe to find selected crtc */
|
|
|
|
for (i = 0; i < ARRAY_SIZE(ni_dig_offsets); i++) {
|
|
|
|
dig_fe = RREG32(NI_DIG_FE_CNTL + ni_dig_offsets[i]);
|
|
|
|
if (dig_fe & NI_DIG_FE_CNTL_SYMCLK_FE_ON &&
|
|
|
|
crtc_id == NI_DIG_FE_CNTL_SOURCE_SELECT(dig_fe)) {
|
|
|
|
/* found running pipe */
|
|
|
|
found_crtc = true;
|
|
|
|
dig_fe_mask = 1 << i;
|
|
|
|
dig_fe = i;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (found_crtc) {
|
|
|
|
/* loop through all running dig_be to find selected dig_fe */
|
|
|
|
for (i = 0; i < ARRAY_SIZE(ni_dig_offsets); i++) {
|
|
|
|
dig_be = RREG32(NI_DIG_BE_CNTL + ni_dig_offsets[i]);
|
|
|
|
/* if dig_fe_selected by dig_be? */
|
|
|
|
digs_fe_selected = NI_DIG_BE_CNTL_FE_SOURCE_SELECT(dig_be);
|
|
|
|
dig_be_mode = NI_DIG_FE_CNTL_MODE(dig_be);
|
|
|
|
if (dig_fe_mask & digs_fe_selected &&
|
|
|
|
/* if dig_be in sst mode? */
|
|
|
|
dig_be_mode == NI_DIG_BE_DPSST) {
|
|
|
|
dig_en_be = RREG32(NI_DIG_BE_EN_CNTL +
|
|
|
|
ni_dig_offsets[i]);
|
|
|
|
uniphy_pll = RREG32(NI_DCIO_UNIPHY0_PLL_CONTROL1 +
|
|
|
|
ni_tx_offsets[i]);
|
|
|
|
/* dig_be enable and tx is running */
|
|
|
|
if (dig_en_be & NI_DIG_BE_EN_CNTL_ENABLE &&
|
|
|
|
dig_en_be & NI_DIG_BE_EN_CNTL_SYMBCLK_ON &&
|
|
|
|
uniphy_pll & NI_DCIO_UNIPHY0_PLL_CONTROL1_ENABLE) {
|
|
|
|
is_enabled = true;
|
|
|
|
*ret_dig_fe = dig_fe;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return is_enabled;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Blank dig when in dp sst mode
|
|
|
|
* Dig ignores crtc timing
|
|
|
|
*/
|
|
|
|
static void evergreen_blank_dp_output(struct radeon_device *rdev,
|
|
|
|
unsigned dig_fe)
|
|
|
|
{
|
|
|
|
unsigned stream_ctrl;
|
|
|
|
unsigned fifo_ctrl;
|
|
|
|
unsigned counter = 0;
|
|
|
|
|
|
|
|
if (dig_fe >= ARRAY_SIZE(evergreen_dp_offsets)) {
|
|
|
|
DRM_ERROR("invalid dig_fe %d\n", dig_fe);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
stream_ctrl = RREG32(EVERGREEN_DP_VID_STREAM_CNTL +
|
|
|
|
evergreen_dp_offsets[dig_fe]);
|
|
|
|
if (!(stream_ctrl & EVERGREEN_DP_VID_STREAM_CNTL_ENABLE)) {
|
|
|
|
DRM_ERROR("dig %d , should be enable\n", dig_fe);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
stream_ctrl &=~EVERGREEN_DP_VID_STREAM_CNTL_ENABLE;
|
|
|
|
WREG32(EVERGREEN_DP_VID_STREAM_CNTL +
|
|
|
|
evergreen_dp_offsets[dig_fe], stream_ctrl);
|
|
|
|
|
|
|
|
stream_ctrl = RREG32(EVERGREEN_DP_VID_STREAM_CNTL +
|
|
|
|
evergreen_dp_offsets[dig_fe]);
|
|
|
|
while (counter < 32 && stream_ctrl & EVERGREEN_DP_VID_STREAM_STATUS) {
|
|
|
|
msleep(1);
|
|
|
|
counter++;
|
|
|
|
stream_ctrl = RREG32(EVERGREEN_DP_VID_STREAM_CNTL +
|
|
|
|
evergreen_dp_offsets[dig_fe]);
|
|
|
|
}
|
|
|
|
if (counter >= 32 )
|
|
|
|
DRM_ERROR("counter exceeds %d\n", counter);
|
|
|
|
|
|
|
|
fifo_ctrl = RREG32(EVERGREEN_DP_STEER_FIFO + evergreen_dp_offsets[dig_fe]);
|
|
|
|
fifo_ctrl |= EVERGREEN_DP_STEER_FIFO_RESET;
|
|
|
|
WREG32(EVERGREEN_DP_STEER_FIFO + evergreen_dp_offsets[dig_fe], fifo_ctrl);
|
|
|
|
|
|
|
|
}
|
|
|
|
|
2011-03-03 04:07:33 +03:00
|
|
|
void evergreen_mc_stop(struct radeon_device *rdev, struct evergreen_mc_save *save)
|
2010-01-13 01:54:34 +03:00
|
|
|
{
|
2012-08-16 01:18:42 +04:00
|
|
|
u32 crtc_enabled, tmp, frame_count, blackout;
|
|
|
|
int i, j;
|
2016-04-14 20:34:03 +03:00
|
|
|
unsigned dig_fe;
|
2012-08-16 01:18:42 +04:00
|
|
|
|
2012-08-30 22:34:30 +04:00
|
|
|
if (!ASIC_IS_NODCE(rdev)) {
|
|
|
|
save->vga_render_control = RREG32(VGA_RENDER_CONTROL);
|
|
|
|
save->vga_hdp_control = RREG32(VGA_HDP_CONTROL);
|
2010-01-13 01:54:34 +03:00
|
|
|
|
2012-08-30 22:34:30 +04:00
|
|
|
/* disable VGA render */
|
|
|
|
WREG32(VGA_RENDER_CONTROL, 0);
|
|
|
|
}
|
2012-08-16 01:18:42 +04:00
|
|
|
/* blank the display controllers */
|
|
|
|
for (i = 0; i < rdev->num_crtc; i++) {
|
|
|
|
crtc_enabled = RREG32(EVERGREEN_CRTC_CONTROL + crtc_offsets[i]) & EVERGREEN_CRTC_MASTER_EN;
|
|
|
|
if (crtc_enabled) {
|
|
|
|
save->crtc_enabled[i] = true;
|
|
|
|
if (ASIC_IS_DCE6(rdev)) {
|
|
|
|
tmp = RREG32(EVERGREEN_CRTC_BLANK_CONTROL + crtc_offsets[i]);
|
|
|
|
if (!(tmp & EVERGREEN_CRTC_BLANK_DATA_EN)) {
|
|
|
|
radeon_wait_for_vblank(rdev, i);
|
2013-04-11 03:08:14 +04:00
|
|
|
WREG32(EVERGREEN_CRTC_UPDATE_LOCK + crtc_offsets[i], 1);
|
2012-08-16 01:18:42 +04:00
|
|
|
tmp |= EVERGREEN_CRTC_BLANK_DATA_EN;
|
|
|
|
WREG32(EVERGREEN_CRTC_BLANK_CONTROL + crtc_offsets[i], tmp);
|
2014-11-06 01:14:32 +03:00
|
|
|
WREG32(EVERGREEN_CRTC_UPDATE_LOCK + crtc_offsets[i], 0);
|
2012-08-16 01:18:42 +04:00
|
|
|
}
|
|
|
|
} else {
|
|
|
|
tmp = RREG32(EVERGREEN_CRTC_CONTROL + crtc_offsets[i]);
|
|
|
|
if (!(tmp & EVERGREEN_CRTC_DISP_READ_REQUEST_DISABLE)) {
|
|
|
|
radeon_wait_for_vblank(rdev, i);
|
2013-04-11 03:08:14 +04:00
|
|
|
WREG32(EVERGREEN_CRTC_UPDATE_LOCK + crtc_offsets[i], 1);
|
2012-08-16 01:18:42 +04:00
|
|
|
tmp |= EVERGREEN_CRTC_DISP_READ_REQUEST_DISABLE;
|
|
|
|
WREG32(EVERGREEN_CRTC_CONTROL + crtc_offsets[i], tmp);
|
2013-04-11 03:08:14 +04:00
|
|
|
WREG32(EVERGREEN_CRTC_UPDATE_LOCK + crtc_offsets[i], 0);
|
2012-08-16 01:18:42 +04:00
|
|
|
}
|
|
|
|
}
|
|
|
|
/* wait for the next frame */
|
|
|
|
frame_count = radeon_get_vblank_counter(rdev, i);
|
|
|
|
for (j = 0; j < rdev->usec_timeout; j++) {
|
|
|
|
if (radeon_get_vblank_counter(rdev, i) != frame_count)
|
|
|
|
break;
|
|
|
|
udelay(1);
|
|
|
|
}
|
2016-04-14 20:34:03 +03:00
|
|
|
/*we should disable dig if it drives dp sst*/
|
|
|
|
/*but we are in radeon_device_init and the topology is unknown*/
|
|
|
|
/*and it is available after radeon_modeset_init*/
|
|
|
|
/*the following method radeon_atom_encoder_dpms_dig*/
|
|
|
|
/*does the job if we initialize it properly*/
|
|
|
|
/*for now we do it this manually*/
|
|
|
|
/**/
|
|
|
|
if (ASIC_IS_DCE5(rdev) &&
|
|
|
|
evergreen_is_dp_sst_stream_enabled(rdev, i ,&dig_fe))
|
|
|
|
evergreen_blank_dp_output(rdev, dig_fe);
|
|
|
|
/*we could remove 6 lines below*/
|
2013-04-11 03:08:14 +04:00
|
|
|
/* XXX this is a hack to avoid strange behavior with EFI on certain systems */
|
|
|
|
WREG32(EVERGREEN_CRTC_UPDATE_LOCK + crtc_offsets[i], 1);
|
|
|
|
tmp = RREG32(EVERGREEN_CRTC_CONTROL + crtc_offsets[i]);
|
|
|
|
tmp &= ~EVERGREEN_CRTC_MASTER_EN;
|
|
|
|
WREG32(EVERGREEN_CRTC_CONTROL + crtc_offsets[i], tmp);
|
|
|
|
WREG32(EVERGREEN_CRTC_UPDATE_LOCK + crtc_offsets[i], 0);
|
|
|
|
save->crtc_enabled[i] = false;
|
|
|
|
/* ***** */
|
2012-11-19 18:11:27 +04:00
|
|
|
} else {
|
|
|
|
save->crtc_enabled[i] = false;
|
2012-08-16 01:18:42 +04:00
|
|
|
}
|
2010-11-23 01:56:28 +03:00
|
|
|
}
|
2010-01-13 01:54:34 +03:00
|
|
|
|
2012-08-16 01:18:42 +04:00
|
|
|
radeon_mc_wait_for_idle(rdev);
|
|
|
|
|
|
|
|
blackout = RREG32(MC_SHARED_BLACKOUT_CNTL);
|
|
|
|
if ((blackout & BLACKOUT_MODE_MASK) != 1) {
|
|
|
|
/* Block CPU access */
|
|
|
|
WREG32(BIF_FB_EN, 0);
|
|
|
|
/* blackout the MC */
|
|
|
|
blackout &= ~BLACKOUT_MODE_MASK;
|
|
|
|
WREG32(MC_SHARED_BLACKOUT_CNTL, blackout | 1);
|
2011-07-08 19:44:56 +04:00
|
|
|
}
|
2013-01-31 18:00:52 +04:00
|
|
|
/* wait for the MC to settle */
|
|
|
|
udelay(100);
|
2013-04-10 17:58:42 +04:00
|
|
|
|
|
|
|
/* lock double buffered regs */
|
|
|
|
for (i = 0; i < rdev->num_crtc; i++) {
|
|
|
|
if (save->crtc_enabled[i]) {
|
|
|
|
tmp = RREG32(EVERGREEN_GRPH_UPDATE + crtc_offsets[i]);
|
|
|
|
if (!(tmp & EVERGREEN_GRPH_UPDATE_LOCK)) {
|
|
|
|
tmp |= EVERGREEN_GRPH_UPDATE_LOCK;
|
|
|
|
WREG32(EVERGREEN_GRPH_UPDATE + crtc_offsets[i], tmp);
|
|
|
|
}
|
|
|
|
tmp = RREG32(EVERGREEN_MASTER_UPDATE_LOCK + crtc_offsets[i]);
|
|
|
|
if (!(tmp & 1)) {
|
|
|
|
tmp |= 1;
|
|
|
|
WREG32(EVERGREEN_MASTER_UPDATE_LOCK + crtc_offsets[i], tmp);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2010-01-13 01:54:34 +03:00
|
|
|
}
|
|
|
|
|
2011-03-03 04:07:33 +03:00
|
|
|
void evergreen_mc_resume(struct radeon_device *rdev, struct evergreen_mc_save *save)
|
2010-01-13 01:54:34 +03:00
|
|
|
{
|
2012-08-16 01:18:42 +04:00
|
|
|
u32 tmp, frame_count;
|
|
|
|
int i, j;
|
2010-11-23 01:56:28 +03:00
|
|
|
|
2012-08-16 01:18:42 +04:00
|
|
|
/* update crtc base addresses */
|
|
|
|
for (i = 0; i < rdev->num_crtc; i++) {
|
|
|
|
WREG32(EVERGREEN_GRPH_PRIMARY_SURFACE_ADDRESS_HIGH + crtc_offsets[i],
|
2010-11-23 01:56:28 +03:00
|
|
|
upper_32_bits(rdev->mc.vram_start));
|
2012-08-16 01:18:42 +04:00
|
|
|
WREG32(EVERGREEN_GRPH_SECONDARY_SURFACE_ADDRESS_HIGH + crtc_offsets[i],
|
2010-11-23 01:56:28 +03:00
|
|
|
upper_32_bits(rdev->mc.vram_start));
|
2012-08-16 01:18:42 +04:00
|
|
|
WREG32(EVERGREEN_GRPH_PRIMARY_SURFACE_ADDRESS + crtc_offsets[i],
|
2010-11-23 01:56:28 +03:00
|
|
|
(u32)rdev->mc.vram_start);
|
2012-08-16 01:18:42 +04:00
|
|
|
WREG32(EVERGREEN_GRPH_SECONDARY_SURFACE_ADDRESS + crtc_offsets[i],
|
2010-11-23 01:56:28 +03:00
|
|
|
(u32)rdev->mc.vram_start);
|
|
|
|
}
|
2012-08-30 22:34:30 +04:00
|
|
|
|
|
|
|
if (!ASIC_IS_NODCE(rdev)) {
|
|
|
|
WREG32(EVERGREEN_VGA_MEMORY_BASE_ADDRESS_HIGH, upper_32_bits(rdev->mc.vram_start));
|
|
|
|
WREG32(EVERGREEN_VGA_MEMORY_BASE_ADDRESS, (u32)rdev->mc.vram_start);
|
|
|
|
}
|
2012-08-16 01:18:42 +04:00
|
|
|
|
2013-04-10 17:58:42 +04:00
|
|
|
/* unlock regs and wait for update */
|
|
|
|
for (i = 0; i < rdev->num_crtc; i++) {
|
|
|
|
if (save->crtc_enabled[i]) {
|
|
|
|
tmp = RREG32(EVERGREEN_MASTER_UPDATE_MODE + crtc_offsets[i]);
|
2016-08-04 06:39:40 +03:00
|
|
|
if ((tmp & 0x7) != 0) {
|
drm/radeon: Prevent too early kms-pageflips triggered by vblank.
Since 3.16-rc1 we have this new failure:
When the userspace XOrg ddx schedules vblank events to
trigger deferred kms-pageflips, e.g., via the OML_sync_control
extension call glXSwapBuffersMscOML(), or if a glXSwapBuffers()
is called immediately after completion of a previous swapbuffers
call, e.g., in a tight rendering loop with minimal rendering,
it happens frequently that the pageflip ioctl() is executed
within the same vblank in which a previous kms-pageflip completed,
or - for deferred swaps - always one vblank earlier than requested
by the client app.
This causes premature pageflips and detection of failure by
the ddx, e.g., XOrg log warnings like...
"(WW) RADEON(1): radeon_dri2_flip_event_handler: Pageflip
completion event has impossible msc 201025 < target_msc 201026"
... and error/invalid return values of glXWaitForSbcOML() and
Intel_swap_events extension.
Reason is the new way in which kms-pageflips are programmed
since 3.16.
This commit changes the time window in which the hw can
execute pending programmed pageflips. Before, a pending flip
would get executed anywhere within the vblank interval. Now
a pending flip only gets executed at the leading edge of
vblank (start of front porch), making sure that a invocation
of the pageflip ioctl() within a given vblank interval will
only lead to pageflip completion in the following vblank.
Tested to death on a DCE-4 card.
Signed-off-by: Mario Kleiner <mario.kleiner.de@gmail.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2014-07-03 05:45:02 +04:00
|
|
|
tmp &= ~0x7;
|
2013-04-10 17:58:42 +04:00
|
|
|
WREG32(EVERGREEN_MASTER_UPDATE_MODE + crtc_offsets[i], tmp);
|
|
|
|
}
|
|
|
|
tmp = RREG32(EVERGREEN_GRPH_UPDATE + crtc_offsets[i]);
|
|
|
|
if (tmp & EVERGREEN_GRPH_UPDATE_LOCK) {
|
|
|
|
tmp &= ~EVERGREEN_GRPH_UPDATE_LOCK;
|
|
|
|
WREG32(EVERGREEN_GRPH_UPDATE + crtc_offsets[i], tmp);
|
|
|
|
}
|
|
|
|
tmp = RREG32(EVERGREEN_MASTER_UPDATE_LOCK + crtc_offsets[i]);
|
|
|
|
if (tmp & 1) {
|
|
|
|
tmp &= ~1;
|
|
|
|
WREG32(EVERGREEN_MASTER_UPDATE_LOCK + crtc_offsets[i], tmp);
|
|
|
|
}
|
|
|
|
for (j = 0; j < rdev->usec_timeout; j++) {
|
|
|
|
tmp = RREG32(EVERGREEN_GRPH_UPDATE + crtc_offsets[i]);
|
|
|
|
if ((tmp & EVERGREEN_GRPH_SURFACE_UPDATE_PENDING) == 0)
|
|
|
|
break;
|
|
|
|
udelay(1);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2012-08-16 01:18:42 +04:00
|
|
|
/* unblackout the MC */
|
|
|
|
tmp = RREG32(MC_SHARED_BLACKOUT_CNTL);
|
|
|
|
tmp &= ~BLACKOUT_MODE_MASK;
|
|
|
|
WREG32(MC_SHARED_BLACKOUT_CNTL, tmp);
|
|
|
|
/* allow CPU access */
|
|
|
|
WREG32(BIF_FB_EN, FB_READ_EN | FB_WRITE_EN);
|
|
|
|
|
|
|
|
for (i = 0; i < rdev->num_crtc; i++) {
|
2012-11-05 20:34:58 +04:00
|
|
|
if (save->crtc_enabled[i]) {
|
2012-08-16 01:18:42 +04:00
|
|
|
if (ASIC_IS_DCE6(rdev)) {
|
|
|
|
tmp = RREG32(EVERGREEN_CRTC_BLANK_CONTROL + crtc_offsets[i]);
|
2014-07-31 01:57:28 +04:00
|
|
|
tmp &= ~EVERGREEN_CRTC_BLANK_DATA_EN;
|
2013-01-26 20:10:58 +04:00
|
|
|
WREG32(EVERGREEN_CRTC_UPDATE_LOCK + crtc_offsets[i], 1);
|
2012-08-16 01:18:42 +04:00
|
|
|
WREG32(EVERGREEN_CRTC_BLANK_CONTROL + crtc_offsets[i], tmp);
|
2013-01-26 20:10:58 +04:00
|
|
|
WREG32(EVERGREEN_CRTC_UPDATE_LOCK + crtc_offsets[i], 0);
|
2012-08-16 01:18:42 +04:00
|
|
|
} else {
|
|
|
|
tmp = RREG32(EVERGREEN_CRTC_CONTROL + crtc_offsets[i]);
|
|
|
|
tmp &= ~EVERGREEN_CRTC_DISP_READ_REQUEST_DISABLE;
|
2013-01-26 20:10:58 +04:00
|
|
|
WREG32(EVERGREEN_CRTC_UPDATE_LOCK + crtc_offsets[i], 1);
|
2012-08-16 01:18:42 +04:00
|
|
|
WREG32(EVERGREEN_CRTC_CONTROL + crtc_offsets[i], tmp);
|
2013-01-26 20:10:58 +04:00
|
|
|
WREG32(EVERGREEN_CRTC_UPDATE_LOCK + crtc_offsets[i], 0);
|
2012-08-16 01:18:42 +04:00
|
|
|
}
|
|
|
|
/* wait for the next frame */
|
|
|
|
frame_count = radeon_get_vblank_counter(rdev, i);
|
|
|
|
for (j = 0; j < rdev->usec_timeout; j++) {
|
|
|
|
if (radeon_get_vblank_counter(rdev, i) != frame_count)
|
|
|
|
break;
|
|
|
|
udelay(1);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2012-08-30 22:34:30 +04:00
|
|
|
if (!ASIC_IS_NODCE(rdev)) {
|
|
|
|
/* Unlock vga access */
|
|
|
|
WREG32(VGA_HDP_CONTROL, save->vga_hdp_control);
|
|
|
|
mdelay(1);
|
|
|
|
WREG32(VGA_RENDER_CONTROL, save->vga_render_control);
|
|
|
|
}
|
2010-01-13 01:54:34 +03:00
|
|
|
}
|
|
|
|
|
2011-03-03 04:07:34 +03:00
|
|
|
void evergreen_mc_program(struct radeon_device *rdev)
|
2010-01-13 01:54:34 +03:00
|
|
|
{
|
|
|
|
struct evergreen_mc_save save;
|
|
|
|
u32 tmp;
|
|
|
|
int i, j;
|
|
|
|
|
|
|
|
/* Initialize HDP */
|
|
|
|
for (i = 0, j = 0; i < 32; i++, j += 0x18) {
|
|
|
|
WREG32((0x2c14 + j), 0x00000000);
|
|
|
|
WREG32((0x2c18 + j), 0x00000000);
|
|
|
|
WREG32((0x2c1c + j), 0x00000000);
|
|
|
|
WREG32((0x2c20 + j), 0x00000000);
|
|
|
|
WREG32((0x2c24 + j), 0x00000000);
|
|
|
|
}
|
|
|
|
WREG32(HDP_REG_COHERENCY_FLUSH_CNTL, 0);
|
|
|
|
|
|
|
|
evergreen_mc_stop(rdev, &save);
|
|
|
|
if (evergreen_mc_wait_for_idle(rdev)) {
|
|
|
|
dev_warn(rdev->dev, "Wait for MC idle timedout !\n");
|
|
|
|
}
|
|
|
|
/* Lockout access through VGA aperture*/
|
|
|
|
WREG32(VGA_HDP_CONTROL, VGA_MEMORY_DISABLE);
|
|
|
|
/* Update configuration */
|
|
|
|
if (rdev->flags & RADEON_IS_AGP) {
|
|
|
|
if (rdev->mc.vram_start < rdev->mc.gtt_start) {
|
|
|
|
/* VRAM before AGP */
|
|
|
|
WREG32(MC_VM_SYSTEM_APERTURE_LOW_ADDR,
|
|
|
|
rdev->mc.vram_start >> 12);
|
|
|
|
WREG32(MC_VM_SYSTEM_APERTURE_HIGH_ADDR,
|
|
|
|
rdev->mc.gtt_end >> 12);
|
|
|
|
} else {
|
|
|
|
/* VRAM after AGP */
|
|
|
|
WREG32(MC_VM_SYSTEM_APERTURE_LOW_ADDR,
|
|
|
|
rdev->mc.gtt_start >> 12);
|
|
|
|
WREG32(MC_VM_SYSTEM_APERTURE_HIGH_ADDR,
|
|
|
|
rdev->mc.vram_end >> 12);
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
WREG32(MC_VM_SYSTEM_APERTURE_LOW_ADDR,
|
|
|
|
rdev->mc.vram_start >> 12);
|
|
|
|
WREG32(MC_VM_SYSTEM_APERTURE_HIGH_ADDR,
|
|
|
|
rdev->mc.vram_end >> 12);
|
|
|
|
}
|
2011-11-10 17:59:39 +04:00
|
|
|
WREG32(MC_VM_SYSTEM_APERTURE_DEFAULT_ADDR, rdev->vram_scratch.gpu_addr >> 12);
|
2012-03-21 01:18:37 +04:00
|
|
|
/* llano/ontario only */
|
|
|
|
if ((rdev->family == CHIP_PALM) ||
|
|
|
|
(rdev->family == CHIP_SUMO) ||
|
|
|
|
(rdev->family == CHIP_SUMO2)) {
|
2010-12-15 19:04:10 +03:00
|
|
|
tmp = RREG32(MC_FUS_VM_FB_OFFSET) & 0x000FFFFF;
|
|
|
|
tmp |= ((rdev->mc.vram_end >> 20) & 0xF) << 24;
|
|
|
|
tmp |= ((rdev->mc.vram_start >> 20) & 0xF) << 20;
|
|
|
|
WREG32(MC_FUS_VM_FB_OFFSET, tmp);
|
|
|
|
}
|
2010-01-13 01:54:34 +03:00
|
|
|
tmp = ((rdev->mc.vram_end >> 24) & 0xFFFF) << 16;
|
|
|
|
tmp |= ((rdev->mc.vram_start >> 24) & 0xFFFF);
|
|
|
|
WREG32(MC_VM_FB_LOCATION, tmp);
|
|
|
|
WREG32(HDP_NONSURFACE_BASE, (rdev->mc.vram_start >> 8));
|
2011-01-07 03:12:37 +03:00
|
|
|
WREG32(HDP_NONSURFACE_INFO, (2 << 7) | (1 << 30));
|
2010-06-03 21:34:48 +04:00
|
|
|
WREG32(HDP_NONSURFACE_SIZE, 0x3FFFFFFF);
|
2010-01-13 01:54:34 +03:00
|
|
|
if (rdev->flags & RADEON_IS_AGP) {
|
|
|
|
WREG32(MC_VM_AGP_TOP, rdev->mc.gtt_end >> 16);
|
|
|
|
WREG32(MC_VM_AGP_BOT, rdev->mc.gtt_start >> 16);
|
|
|
|
WREG32(MC_VM_AGP_BASE, rdev->mc.agp_base >> 22);
|
|
|
|
} else {
|
|
|
|
WREG32(MC_VM_AGP_BASE, 0);
|
|
|
|
WREG32(MC_VM_AGP_TOP, 0x0FFFFFFF);
|
|
|
|
WREG32(MC_VM_AGP_BOT, 0x0FFFFFFF);
|
|
|
|
}
|
|
|
|
if (evergreen_mc_wait_for_idle(rdev)) {
|
|
|
|
dev_warn(rdev->dev, "Wait for MC idle timedout !\n");
|
|
|
|
}
|
|
|
|
evergreen_mc_resume(rdev, &save);
|
|
|
|
/* we need to own VRAM, so turn off the VGA renderer here
|
|
|
|
* to stop it overwriting our objects */
|
|
|
|
rv515_vga_render_disable(rdev);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* CP.
|
|
|
|
*/
|
2011-02-02 20:37:40 +03:00
|
|
|
void evergreen_ring_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib)
|
|
|
|
{
|
2012-05-08 16:24:01 +04:00
|
|
|
struct radeon_ring *ring = &rdev->ring[ib->ring];
|
2012-07-17 22:02:31 +04:00
|
|
|
u32 next_rptr;
|
2011-09-23 17:11:23 +04:00
|
|
|
|
2011-02-02 20:37:40 +03:00
|
|
|
/* set to DX10/11 mode */
|
2011-10-23 14:56:27 +04:00
|
|
|
radeon_ring_write(ring, PACKET3(PACKET3_MODE_CONTROL, 0));
|
|
|
|
radeon_ring_write(ring, 1);
|
2012-07-06 18:22:55 +04:00
|
|
|
|
|
|
|
if (ring->rptr_save_reg) {
|
2012-07-17 22:02:31 +04:00
|
|
|
next_rptr = ring->wptr + 3 + 4;
|
2012-07-06 18:22:55 +04:00
|
|
|
radeon_ring_write(ring, PACKET3(PACKET3_SET_CONFIG_REG, 1));
|
|
|
|
radeon_ring_write(ring, ((ring->rptr_save_reg -
|
|
|
|
PACKET3_SET_CONFIG_REG_START) >> 2));
|
|
|
|
radeon_ring_write(ring, next_rptr);
|
2012-07-17 22:02:31 +04:00
|
|
|
} else if (rdev->wb.enabled) {
|
|
|
|
next_rptr = ring->wptr + 5 + 4;
|
|
|
|
radeon_ring_write(ring, PACKET3(PACKET3_MEM_WRITE, 3));
|
|
|
|
radeon_ring_write(ring, ring->next_rptr_gpu_addr & 0xfffffffc);
|
|
|
|
radeon_ring_write(ring, (upper_32_bits(ring->next_rptr_gpu_addr) & 0xff) | (1 << 18));
|
|
|
|
radeon_ring_write(ring, next_rptr);
|
|
|
|
radeon_ring_write(ring, 0);
|
2012-07-06 18:22:55 +04:00
|
|
|
}
|
|
|
|
|
2011-10-23 14:56:27 +04:00
|
|
|
radeon_ring_write(ring, PACKET3(PACKET3_INDIRECT_BUFFER, 2));
|
|
|
|
radeon_ring_write(ring,
|
2011-02-14 03:06:33 +03:00
|
|
|
#ifdef __BIG_ENDIAN
|
|
|
|
(2 << 0) |
|
|
|
|
#endif
|
|
|
|
(ib->gpu_addr & 0xFFFFFFFC));
|
2011-10-23 14:56:27 +04:00
|
|
|
radeon_ring_write(ring, upper_32_bits(ib->gpu_addr) & 0xFF);
|
|
|
|
radeon_ring_write(ring, ib->length_dw);
|
2011-02-02 20:37:40 +03:00
|
|
|
}
|
|
|
|
|
2010-01-13 01:54:34 +03:00
|
|
|
|
|
|
|
static int evergreen_cp_load_microcode(struct radeon_device *rdev)
|
|
|
|
{
|
2010-03-24 20:36:43 +03:00
|
|
|
const __be32 *fw_data;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
if (!rdev->me_fw || !rdev->pfp_fw)
|
|
|
|
return -EINVAL;
|
2010-01-13 01:54:34 +03:00
|
|
|
|
2010-03-24 20:36:43 +03:00
|
|
|
r700_cp_stop(rdev);
|
2011-02-14 03:06:33 +03:00
|
|
|
WREG32(CP_RB_CNTL,
|
|
|
|
#ifdef __BIG_ENDIAN
|
|
|
|
BUF_SWAP_32BIT |
|
|
|
|
#endif
|
|
|
|
RB_NO_UPDATE | RB_BLKSZ(15) | RB_BUFSZ(3));
|
2010-03-24 20:36:43 +03:00
|
|
|
|
|
|
|
fw_data = (const __be32 *)rdev->pfp_fw->data;
|
|
|
|
WREG32(CP_PFP_UCODE_ADDR, 0);
|
|
|
|
for (i = 0; i < EVERGREEN_PFP_UCODE_SIZE; i++)
|
|
|
|
WREG32(CP_PFP_UCODE_DATA, be32_to_cpup(fw_data++));
|
|
|
|
WREG32(CP_PFP_UCODE_ADDR, 0);
|
|
|
|
|
|
|
|
fw_data = (const __be32 *)rdev->me_fw->data;
|
|
|
|
WREG32(CP_ME_RAM_WADDR, 0);
|
|
|
|
for (i = 0; i < EVERGREEN_PM4_UCODE_SIZE; i++)
|
|
|
|
WREG32(CP_ME_RAM_DATA, be32_to_cpup(fw_data++));
|
|
|
|
|
|
|
|
WREG32(CP_PFP_UCODE_ADDR, 0);
|
|
|
|
WREG32(CP_ME_RAM_WADDR, 0);
|
|
|
|
WREG32(CP_ME_RAM_RADDR, 0);
|
2010-01-13 01:54:34 +03:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2010-09-03 05:32:32 +04:00
|
|
|
static int evergreen_cp_start(struct radeon_device *rdev)
|
|
|
|
{
|
2011-10-23 14:56:27 +04:00
|
|
|
struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX];
|
2010-10-21 21:31:38 +04:00
|
|
|
int r, i;
|
2010-09-03 05:32:32 +04:00
|
|
|
uint32_t cp_me;
|
|
|
|
|
2011-10-23 14:56:27 +04:00
|
|
|
r = radeon_ring_lock(rdev, ring, 7);
|
2010-09-03 05:32:32 +04:00
|
|
|
if (r) {
|
|
|
|
DRM_ERROR("radeon: cp failed to lock ring (%d).\n", r);
|
|
|
|
return r;
|
|
|
|
}
|
2011-10-23 14:56:27 +04:00
|
|
|
radeon_ring_write(ring, PACKET3(PACKET3_ME_INITIALIZE, 5));
|
|
|
|
radeon_ring_write(ring, 0x1);
|
|
|
|
radeon_ring_write(ring, 0x0);
|
|
|
|
radeon_ring_write(ring, rdev->config.evergreen.max_hw_contexts - 1);
|
|
|
|
radeon_ring_write(ring, PACKET3_ME_INITIALIZE_DEVICE_ID(1));
|
|
|
|
radeon_ring_write(ring, 0);
|
|
|
|
radeon_ring_write(ring, 0);
|
2014-08-18 12:34:55 +04:00
|
|
|
radeon_ring_unlock_commit(rdev, ring, false);
|
2010-09-03 05:32:32 +04:00
|
|
|
|
|
|
|
cp_me = 0xff;
|
|
|
|
WREG32(CP_ME_CNTL, cp_me);
|
|
|
|
|
2011-10-23 14:56:27 +04:00
|
|
|
r = radeon_ring_lock(rdev, ring, evergreen_default_size + 19);
|
2010-09-03 05:32:32 +04:00
|
|
|
if (r) {
|
|
|
|
DRM_ERROR("radeon: cp failed to lock ring (%d).\n", r);
|
|
|
|
return r;
|
|
|
|
}
|
2010-10-21 21:31:38 +04:00
|
|
|
|
|
|
|
/* setup clear context state */
|
2011-10-23 14:56:27 +04:00
|
|
|
radeon_ring_write(ring, PACKET3(PACKET3_PREAMBLE_CNTL, 0));
|
|
|
|
radeon_ring_write(ring, PACKET3_PREAMBLE_BEGIN_CLEAR_STATE);
|
2010-10-21 21:31:38 +04:00
|
|
|
|
|
|
|
for (i = 0; i < evergreen_default_size; i++)
|
2011-10-23 14:56:27 +04:00
|
|
|
radeon_ring_write(ring, evergreen_default_state[i]);
|
2010-10-21 21:31:38 +04:00
|
|
|
|
2011-10-23 14:56:27 +04:00
|
|
|
radeon_ring_write(ring, PACKET3(PACKET3_PREAMBLE_CNTL, 0));
|
|
|
|
radeon_ring_write(ring, PACKET3_PREAMBLE_END_CLEAR_STATE);
|
2010-10-21 21:31:38 +04:00
|
|
|
|
|
|
|
/* set clear context state */
|
2011-10-23 14:56:27 +04:00
|
|
|
radeon_ring_write(ring, PACKET3(PACKET3_CLEAR_STATE, 0));
|
|
|
|
radeon_ring_write(ring, 0);
|
2010-10-21 21:31:38 +04:00
|
|
|
|
|
|
|
/* SQ_VTX_BASE_VTX_LOC */
|
2011-10-23 14:56:27 +04:00
|
|
|
radeon_ring_write(ring, 0xc0026f00);
|
|
|
|
radeon_ring_write(ring, 0x00000000);
|
|
|
|
radeon_ring_write(ring, 0x00000000);
|
|
|
|
radeon_ring_write(ring, 0x00000000);
|
2010-10-21 21:31:38 +04:00
|
|
|
|
|
|
|
/* Clear consts */
|
2011-10-23 14:56:27 +04:00
|
|
|
radeon_ring_write(ring, 0xc0036f00);
|
|
|
|
radeon_ring_write(ring, 0x00000bc4);
|
|
|
|
radeon_ring_write(ring, 0xffffffff);
|
|
|
|
radeon_ring_write(ring, 0xffffffff);
|
|
|
|
radeon_ring_write(ring, 0xffffffff);
|
2010-10-21 21:31:38 +04:00
|
|
|
|
2011-10-23 14:56:27 +04:00
|
|
|
radeon_ring_write(ring, 0xc0026900);
|
|
|
|
radeon_ring_write(ring, 0x00000316);
|
|
|
|
radeon_ring_write(ring, 0x0000000e); /* VGT_VERTEX_REUSE_BLOCK_CNTL */
|
|
|
|
radeon_ring_write(ring, 0x00000010); /* */
|
2011-02-02 20:37:41 +03:00
|
|
|
|
2014-08-18 12:34:55 +04:00
|
|
|
radeon_ring_unlock_commit(rdev, ring, false);
|
2010-09-03 05:32:32 +04:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2012-08-31 21:43:50 +04:00
|
|
|
static int evergreen_cp_resume(struct radeon_device *rdev)
|
2010-03-24 20:36:43 +03:00
|
|
|
{
|
2011-10-23 14:56:27 +04:00
|
|
|
struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX];
|
2010-03-24 20:36:43 +03:00
|
|
|
u32 tmp;
|
|
|
|
u32 rb_bufsz;
|
|
|
|
int r;
|
|
|
|
|
|
|
|
/* Reset cp; if cp is reset, then PA, SH, VGT also need to be reset */
|
|
|
|
WREG32(GRBM_SOFT_RESET, (SOFT_RESET_CP |
|
|
|
|
SOFT_RESET_PA |
|
|
|
|
SOFT_RESET_SH |
|
|
|
|
SOFT_RESET_VGT |
|
2011-08-25 00:00:17 +04:00
|
|
|
SOFT_RESET_SPI |
|
2010-03-24 20:36:43 +03:00
|
|
|
SOFT_RESET_SX));
|
|
|
|
RREG32(GRBM_SOFT_RESET);
|
|
|
|
mdelay(15);
|
|
|
|
WREG32(GRBM_SOFT_RESET, 0);
|
|
|
|
RREG32(GRBM_SOFT_RESET);
|
|
|
|
|
|
|
|
/* Set ring buffer size */
|
2013-07-10 16:11:59 +04:00
|
|
|
rb_bufsz = order_base_2(ring->ring_size / 8);
|
|
|
|
tmp = (order_base_2(RADEON_GPU_PAGE_SIZE/8) << 8) | rb_bufsz;
|
2010-03-24 20:36:43 +03:00
|
|
|
#ifdef __BIG_ENDIAN
|
|
|
|
tmp |= BUF_SWAP_32BIT;
|
2010-03-24 20:33:47 +03:00
|
|
|
#endif
|
2010-03-24 20:36:43 +03:00
|
|
|
WREG32(CP_RB_CNTL, tmp);
|
2011-09-15 21:02:22 +04:00
|
|
|
WREG32(CP_SEM_WAIT_TIMER, 0x0);
|
2012-01-20 23:47:43 +04:00
|
|
|
WREG32(CP_SEM_INCOMPLETE_TIMER_CNTL, 0x0);
|
2010-03-24 20:36:43 +03:00
|
|
|
|
|
|
|
/* Set the write pointer delay */
|
|
|
|
WREG32(CP_RB_WPTR_DELAY, 0);
|
|
|
|
|
|
|
|
/* Initialize the ring buffer's read and write pointers */
|
|
|
|
WREG32(CP_RB_CNTL, tmp | RB_RPTR_WR_ENA);
|
|
|
|
WREG32(CP_RB_RPTR_WR, 0);
|
2011-10-23 14:56:27 +04:00
|
|
|
ring->wptr = 0;
|
|
|
|
WREG32(CP_RB_WPTR, ring->wptr);
|
2010-08-28 02:25:25 +04:00
|
|
|
|
2012-09-20 05:48:00 +04:00
|
|
|
/* set the wb address whether it's enabled or not */
|
2011-02-14 03:06:33 +03:00
|
|
|
WREG32(CP_RB_RPTR_ADDR,
|
|
|
|
((rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFFFFFFFC));
|
2010-08-28 02:25:25 +04:00
|
|
|
WREG32(CP_RB_RPTR_ADDR_HI, upper_32_bits(rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFF);
|
|
|
|
WREG32(SCRATCH_ADDR, ((rdev->wb.gpu_addr + RADEON_WB_SCRATCH_OFFSET) >> 8) & 0xFFFFFFFF);
|
|
|
|
|
|
|
|
if (rdev->wb.enabled)
|
|
|
|
WREG32(SCRATCH_UMSK, 0xff);
|
|
|
|
else {
|
|
|
|
tmp |= RB_NO_UPDATE;
|
|
|
|
WREG32(SCRATCH_UMSK, 0);
|
|
|
|
}
|
|
|
|
|
2010-03-24 20:36:43 +03:00
|
|
|
mdelay(1);
|
|
|
|
WREG32(CP_RB_CNTL, tmp);
|
|
|
|
|
2011-10-23 14:56:27 +04:00
|
|
|
WREG32(CP_RB_BASE, ring->gpu_addr >> 8);
|
2010-03-24 20:36:43 +03:00
|
|
|
WREG32(CP_DEBUG, (1 << 27) | (1 << 28));
|
|
|
|
|
2010-09-03 05:32:32 +04:00
|
|
|
evergreen_cp_start(rdev);
|
2011-10-23 14:56:27 +04:00
|
|
|
ring->ready = true;
|
2012-02-24 02:53:45 +04:00
|
|
|
r = radeon_ring_test(rdev, RADEON_RING_TYPE_GFX_INDEX, ring);
|
2010-03-24 20:36:43 +03:00
|
|
|
if (r) {
|
2011-10-23 14:56:27 +04:00
|
|
|
ring->ready = false;
|
2010-03-24 20:36:43 +03:00
|
|
|
return r;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
2010-01-13 01:54:34 +03:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Core functions
|
|
|
|
*/
|
|
|
|
static void evergreen_gpu_init(struct radeon_device *rdev)
|
|
|
|
{
|
2012-06-01 03:00:25 +04:00
|
|
|
u32 gb_addr_config;
|
2010-03-24 20:33:47 +03:00
|
|
|
u32 mc_shared_chmap, mc_arb_ramcfg;
|
|
|
|
u32 sx_debug_1;
|
|
|
|
u32 smx_dc_ctl0;
|
|
|
|
u32 sq_config;
|
|
|
|
u32 sq_lds_resource_mgmt;
|
|
|
|
u32 sq_gpr_resource_mgmt_1;
|
|
|
|
u32 sq_gpr_resource_mgmt_2;
|
|
|
|
u32 sq_gpr_resource_mgmt_3;
|
|
|
|
u32 sq_thread_resource_mgmt;
|
|
|
|
u32 sq_thread_resource_mgmt_2;
|
|
|
|
u32 sq_stack_resource_mgmt_1;
|
|
|
|
u32 sq_stack_resource_mgmt_2;
|
|
|
|
u32 sq_stack_resource_mgmt_3;
|
|
|
|
u32 vgt_cache_invalidation;
|
2011-05-19 19:07:57 +04:00
|
|
|
u32 hdp_host_path_cntl, tmp;
|
2012-06-01 03:00:25 +04:00
|
|
|
u32 disabled_rb_mask;
|
2014-10-29 18:47:07 +03:00
|
|
|
int i, j, ps_thread_count;
|
2010-03-24 20:33:47 +03:00
|
|
|
|
|
|
|
switch (rdev->family) {
|
|
|
|
case CHIP_CYPRESS:
|
|
|
|
case CHIP_HEMLOCK:
|
|
|
|
rdev->config.evergreen.num_ses = 2;
|
|
|
|
rdev->config.evergreen.max_pipes = 4;
|
|
|
|
rdev->config.evergreen.max_tile_pipes = 8;
|
|
|
|
rdev->config.evergreen.max_simds = 10;
|
|
|
|
rdev->config.evergreen.max_backends = 4 * rdev->config.evergreen.num_ses;
|
|
|
|
rdev->config.evergreen.max_gprs = 256;
|
|
|
|
rdev->config.evergreen.max_threads = 248;
|
|
|
|
rdev->config.evergreen.max_gs_threads = 32;
|
|
|
|
rdev->config.evergreen.max_stack_entries = 512;
|
|
|
|
rdev->config.evergreen.sx_num_of_sets = 4;
|
|
|
|
rdev->config.evergreen.sx_max_export_size = 256;
|
|
|
|
rdev->config.evergreen.sx_max_export_pos_size = 64;
|
|
|
|
rdev->config.evergreen.sx_max_export_smx_size = 192;
|
|
|
|
rdev->config.evergreen.max_hw_contexts = 8;
|
|
|
|
rdev->config.evergreen.sq_num_cf_insts = 2;
|
|
|
|
|
|
|
|
rdev->config.evergreen.sc_prim_fifo_size = 0x100;
|
|
|
|
rdev->config.evergreen.sc_hiz_tile_fifo_size = 0x30;
|
|
|
|
rdev->config.evergreen.sc_earlyz_tile_fifo_size = 0x130;
|
2012-06-01 03:00:25 +04:00
|
|
|
gb_addr_config = CYPRESS_GB_ADDR_CONFIG_GOLDEN;
|
2010-03-24 20:33:47 +03:00
|
|
|
break;
|
|
|
|
case CHIP_JUNIPER:
|
|
|
|
rdev->config.evergreen.num_ses = 1;
|
|
|
|
rdev->config.evergreen.max_pipes = 4;
|
|
|
|
rdev->config.evergreen.max_tile_pipes = 4;
|
|
|
|
rdev->config.evergreen.max_simds = 10;
|
|
|
|
rdev->config.evergreen.max_backends = 4 * rdev->config.evergreen.num_ses;
|
|
|
|
rdev->config.evergreen.max_gprs = 256;
|
|
|
|
rdev->config.evergreen.max_threads = 248;
|
|
|
|
rdev->config.evergreen.max_gs_threads = 32;
|
|
|
|
rdev->config.evergreen.max_stack_entries = 512;
|
|
|
|
rdev->config.evergreen.sx_num_of_sets = 4;
|
|
|
|
rdev->config.evergreen.sx_max_export_size = 256;
|
|
|
|
rdev->config.evergreen.sx_max_export_pos_size = 64;
|
|
|
|
rdev->config.evergreen.sx_max_export_smx_size = 192;
|
|
|
|
rdev->config.evergreen.max_hw_contexts = 8;
|
|
|
|
rdev->config.evergreen.sq_num_cf_insts = 2;
|
|
|
|
|
|
|
|
rdev->config.evergreen.sc_prim_fifo_size = 0x100;
|
|
|
|
rdev->config.evergreen.sc_hiz_tile_fifo_size = 0x30;
|
|
|
|
rdev->config.evergreen.sc_earlyz_tile_fifo_size = 0x130;
|
2012-06-01 03:00:25 +04:00
|
|
|
gb_addr_config = JUNIPER_GB_ADDR_CONFIG_GOLDEN;
|
2010-03-24 20:33:47 +03:00
|
|
|
break;
|
|
|
|
case CHIP_REDWOOD:
|
|
|
|
rdev->config.evergreen.num_ses = 1;
|
|
|
|
rdev->config.evergreen.max_pipes = 4;
|
|
|
|
rdev->config.evergreen.max_tile_pipes = 4;
|
|
|
|
rdev->config.evergreen.max_simds = 5;
|
|
|
|
rdev->config.evergreen.max_backends = 2 * rdev->config.evergreen.num_ses;
|
|
|
|
rdev->config.evergreen.max_gprs = 256;
|
|
|
|
rdev->config.evergreen.max_threads = 248;
|
|
|
|
rdev->config.evergreen.max_gs_threads = 32;
|
|
|
|
rdev->config.evergreen.max_stack_entries = 256;
|
|
|
|
rdev->config.evergreen.sx_num_of_sets = 4;
|
|
|
|
rdev->config.evergreen.sx_max_export_size = 256;
|
|
|
|
rdev->config.evergreen.sx_max_export_pos_size = 64;
|
|
|
|
rdev->config.evergreen.sx_max_export_smx_size = 192;
|
|
|
|
rdev->config.evergreen.max_hw_contexts = 8;
|
|
|
|
rdev->config.evergreen.sq_num_cf_insts = 2;
|
|
|
|
|
|
|
|
rdev->config.evergreen.sc_prim_fifo_size = 0x100;
|
|
|
|
rdev->config.evergreen.sc_hiz_tile_fifo_size = 0x30;
|
|
|
|
rdev->config.evergreen.sc_earlyz_tile_fifo_size = 0x130;
|
2012-06-01 03:00:25 +04:00
|
|
|
gb_addr_config = REDWOOD_GB_ADDR_CONFIG_GOLDEN;
|
2010-03-24 20:33:47 +03:00
|
|
|
break;
|
|
|
|
case CHIP_CEDAR:
|
|
|
|
default:
|
|
|
|
rdev->config.evergreen.num_ses = 1;
|
|
|
|
rdev->config.evergreen.max_pipes = 2;
|
|
|
|
rdev->config.evergreen.max_tile_pipes = 2;
|
|
|
|
rdev->config.evergreen.max_simds = 2;
|
|
|
|
rdev->config.evergreen.max_backends = 1 * rdev->config.evergreen.num_ses;
|
|
|
|
rdev->config.evergreen.max_gprs = 256;
|
|
|
|
rdev->config.evergreen.max_threads = 192;
|
|
|
|
rdev->config.evergreen.max_gs_threads = 16;
|
|
|
|
rdev->config.evergreen.max_stack_entries = 256;
|
|
|
|
rdev->config.evergreen.sx_num_of_sets = 4;
|
|
|
|
rdev->config.evergreen.sx_max_export_size = 128;
|
|
|
|
rdev->config.evergreen.sx_max_export_pos_size = 32;
|
|
|
|
rdev->config.evergreen.sx_max_export_smx_size = 96;
|
|
|
|
rdev->config.evergreen.max_hw_contexts = 4;
|
|
|
|
rdev->config.evergreen.sq_num_cf_insts = 1;
|
|
|
|
|
2010-11-23 01:56:29 +03:00
|
|
|
rdev->config.evergreen.sc_prim_fifo_size = 0x40;
|
|
|
|
rdev->config.evergreen.sc_hiz_tile_fifo_size = 0x30;
|
|
|
|
rdev->config.evergreen.sc_earlyz_tile_fifo_size = 0x130;
|
2012-06-01 03:00:25 +04:00
|
|
|
gb_addr_config = CEDAR_GB_ADDR_CONFIG_GOLDEN;
|
2010-11-23 01:56:29 +03:00
|
|
|
break;
|
|
|
|
case CHIP_PALM:
|
|
|
|
rdev->config.evergreen.num_ses = 1;
|
|
|
|
rdev->config.evergreen.max_pipes = 2;
|
|
|
|
rdev->config.evergreen.max_tile_pipes = 2;
|
|
|
|
rdev->config.evergreen.max_simds = 2;
|
|
|
|
rdev->config.evergreen.max_backends = 1 * rdev->config.evergreen.num_ses;
|
|
|
|
rdev->config.evergreen.max_gprs = 256;
|
|
|
|
rdev->config.evergreen.max_threads = 192;
|
|
|
|
rdev->config.evergreen.max_gs_threads = 16;
|
|
|
|
rdev->config.evergreen.max_stack_entries = 256;
|
|
|
|
rdev->config.evergreen.sx_num_of_sets = 4;
|
|
|
|
rdev->config.evergreen.sx_max_export_size = 128;
|
|
|
|
rdev->config.evergreen.sx_max_export_pos_size = 32;
|
|
|
|
rdev->config.evergreen.sx_max_export_smx_size = 96;
|
|
|
|
rdev->config.evergreen.max_hw_contexts = 4;
|
|
|
|
rdev->config.evergreen.sq_num_cf_insts = 1;
|
|
|
|
|
2011-05-31 23:42:48 +04:00
|
|
|
rdev->config.evergreen.sc_prim_fifo_size = 0x40;
|
|
|
|
rdev->config.evergreen.sc_hiz_tile_fifo_size = 0x30;
|
|
|
|
rdev->config.evergreen.sc_earlyz_tile_fifo_size = 0x130;
|
2012-06-01 03:00:25 +04:00
|
|
|
gb_addr_config = CEDAR_GB_ADDR_CONFIG_GOLDEN;
|
2011-05-31 23:42:48 +04:00
|
|
|
break;
|
|
|
|
case CHIP_SUMO:
|
|
|
|
rdev->config.evergreen.num_ses = 1;
|
|
|
|
rdev->config.evergreen.max_pipes = 4;
|
2012-12-11 20:56:52 +04:00
|
|
|
rdev->config.evergreen.max_tile_pipes = 4;
|
2011-05-31 23:42:48 +04:00
|
|
|
if (rdev->pdev->device == 0x9648)
|
|
|
|
rdev->config.evergreen.max_simds = 3;
|
|
|
|
else if ((rdev->pdev->device == 0x9647) ||
|
|
|
|
(rdev->pdev->device == 0x964a))
|
|
|
|
rdev->config.evergreen.max_simds = 4;
|
|
|
|
else
|
|
|
|
rdev->config.evergreen.max_simds = 5;
|
|
|
|
rdev->config.evergreen.max_backends = 2 * rdev->config.evergreen.num_ses;
|
|
|
|
rdev->config.evergreen.max_gprs = 256;
|
|
|
|
rdev->config.evergreen.max_threads = 248;
|
|
|
|
rdev->config.evergreen.max_gs_threads = 32;
|
|
|
|
rdev->config.evergreen.max_stack_entries = 256;
|
|
|
|
rdev->config.evergreen.sx_num_of_sets = 4;
|
|
|
|
rdev->config.evergreen.sx_max_export_size = 256;
|
|
|
|
rdev->config.evergreen.sx_max_export_pos_size = 64;
|
|
|
|
rdev->config.evergreen.sx_max_export_smx_size = 192;
|
|
|
|
rdev->config.evergreen.max_hw_contexts = 8;
|
|
|
|
rdev->config.evergreen.sq_num_cf_insts = 2;
|
|
|
|
|
|
|
|
rdev->config.evergreen.sc_prim_fifo_size = 0x40;
|
|
|
|
rdev->config.evergreen.sc_hiz_tile_fifo_size = 0x30;
|
|
|
|
rdev->config.evergreen.sc_earlyz_tile_fifo_size = 0x130;
|
2012-12-11 20:56:52 +04:00
|
|
|
gb_addr_config = SUMO_GB_ADDR_CONFIG_GOLDEN;
|
2011-05-31 23:42:48 +04:00
|
|
|
break;
|
|
|
|
case CHIP_SUMO2:
|
|
|
|
rdev->config.evergreen.num_ses = 1;
|
|
|
|
rdev->config.evergreen.max_pipes = 4;
|
|
|
|
rdev->config.evergreen.max_tile_pipes = 4;
|
|
|
|
rdev->config.evergreen.max_simds = 2;
|
|
|
|
rdev->config.evergreen.max_backends = 1 * rdev->config.evergreen.num_ses;
|
|
|
|
rdev->config.evergreen.max_gprs = 256;
|
|
|
|
rdev->config.evergreen.max_threads = 248;
|
|
|
|
rdev->config.evergreen.max_gs_threads = 32;
|
|
|
|
rdev->config.evergreen.max_stack_entries = 512;
|
|
|
|
rdev->config.evergreen.sx_num_of_sets = 4;
|
|
|
|
rdev->config.evergreen.sx_max_export_size = 256;
|
|
|
|
rdev->config.evergreen.sx_max_export_pos_size = 64;
|
|
|
|
rdev->config.evergreen.sx_max_export_smx_size = 192;
|
2013-10-02 03:54:33 +04:00
|
|
|
rdev->config.evergreen.max_hw_contexts = 4;
|
2011-05-31 23:42:48 +04:00
|
|
|
rdev->config.evergreen.sq_num_cf_insts = 2;
|
|
|
|
|
2011-01-07 05:19:24 +03:00
|
|
|
rdev->config.evergreen.sc_prim_fifo_size = 0x40;
|
|
|
|
rdev->config.evergreen.sc_hiz_tile_fifo_size = 0x30;
|
|
|
|
rdev->config.evergreen.sc_earlyz_tile_fifo_size = 0x130;
|
2012-12-11 20:56:52 +04:00
|
|
|
gb_addr_config = SUMO2_GB_ADDR_CONFIG_GOLDEN;
|
2011-01-07 05:19:24 +03:00
|
|
|
break;
|
|
|
|
case CHIP_BARTS:
|
|
|
|
rdev->config.evergreen.num_ses = 2;
|
|
|
|
rdev->config.evergreen.max_pipes = 4;
|
|
|
|
rdev->config.evergreen.max_tile_pipes = 8;
|
|
|
|
rdev->config.evergreen.max_simds = 7;
|
|
|
|
rdev->config.evergreen.max_backends = 4 * rdev->config.evergreen.num_ses;
|
|
|
|
rdev->config.evergreen.max_gprs = 256;
|
|
|
|
rdev->config.evergreen.max_threads = 248;
|
|
|
|
rdev->config.evergreen.max_gs_threads = 32;
|
|
|
|
rdev->config.evergreen.max_stack_entries = 512;
|
|
|
|
rdev->config.evergreen.sx_num_of_sets = 4;
|
|
|
|
rdev->config.evergreen.sx_max_export_size = 256;
|
|
|
|
rdev->config.evergreen.sx_max_export_pos_size = 64;
|
|
|
|
rdev->config.evergreen.sx_max_export_smx_size = 192;
|
|
|
|
rdev->config.evergreen.max_hw_contexts = 8;
|
|
|
|
rdev->config.evergreen.sq_num_cf_insts = 2;
|
|
|
|
|
|
|
|
rdev->config.evergreen.sc_prim_fifo_size = 0x100;
|
|
|
|
rdev->config.evergreen.sc_hiz_tile_fifo_size = 0x30;
|
|
|
|
rdev->config.evergreen.sc_earlyz_tile_fifo_size = 0x130;
|
2012-06-01 03:00:25 +04:00
|
|
|
gb_addr_config = BARTS_GB_ADDR_CONFIG_GOLDEN;
|
2011-01-07 05:19:24 +03:00
|
|
|
break;
|
|
|
|
case CHIP_TURKS:
|
|
|
|
rdev->config.evergreen.num_ses = 1;
|
|
|
|
rdev->config.evergreen.max_pipes = 4;
|
|
|
|
rdev->config.evergreen.max_tile_pipes = 4;
|
|
|
|
rdev->config.evergreen.max_simds = 6;
|
|
|
|
rdev->config.evergreen.max_backends = 2 * rdev->config.evergreen.num_ses;
|
|
|
|
rdev->config.evergreen.max_gprs = 256;
|
|
|
|
rdev->config.evergreen.max_threads = 248;
|
|
|
|
rdev->config.evergreen.max_gs_threads = 32;
|
|
|
|
rdev->config.evergreen.max_stack_entries = 256;
|
|
|
|
rdev->config.evergreen.sx_num_of_sets = 4;
|
|
|
|
rdev->config.evergreen.sx_max_export_size = 256;
|
|
|
|
rdev->config.evergreen.sx_max_export_pos_size = 64;
|
|
|
|
rdev->config.evergreen.sx_max_export_smx_size = 192;
|
|
|
|
rdev->config.evergreen.max_hw_contexts = 8;
|
|
|
|
rdev->config.evergreen.sq_num_cf_insts = 2;
|
|
|
|
|
|
|
|
rdev->config.evergreen.sc_prim_fifo_size = 0x100;
|
|
|
|
rdev->config.evergreen.sc_hiz_tile_fifo_size = 0x30;
|
|
|
|
rdev->config.evergreen.sc_earlyz_tile_fifo_size = 0x130;
|
2012-06-01 03:00:25 +04:00
|
|
|
gb_addr_config = TURKS_GB_ADDR_CONFIG_GOLDEN;
|
2011-01-07 05:19:24 +03:00
|
|
|
break;
|
|
|
|
case CHIP_CAICOS:
|
|
|
|
rdev->config.evergreen.num_ses = 1;
|
2012-12-11 20:56:52 +04:00
|
|
|
rdev->config.evergreen.max_pipes = 2;
|
2011-01-07 05:19:24 +03:00
|
|
|
rdev->config.evergreen.max_tile_pipes = 2;
|
|
|
|
rdev->config.evergreen.max_simds = 2;
|
|
|
|
rdev->config.evergreen.max_backends = 1 * rdev->config.evergreen.num_ses;
|
|
|
|
rdev->config.evergreen.max_gprs = 256;
|
|
|
|
rdev->config.evergreen.max_threads = 192;
|
|
|
|
rdev->config.evergreen.max_gs_threads = 16;
|
|
|
|
rdev->config.evergreen.max_stack_entries = 256;
|
|
|
|
rdev->config.evergreen.sx_num_of_sets = 4;
|
|
|
|
rdev->config.evergreen.sx_max_export_size = 128;
|
|
|
|
rdev->config.evergreen.sx_max_export_pos_size = 32;
|
|
|
|
rdev->config.evergreen.sx_max_export_smx_size = 96;
|
|
|
|
rdev->config.evergreen.max_hw_contexts = 4;
|
|
|
|
rdev->config.evergreen.sq_num_cf_insts = 1;
|
|
|
|
|
2010-03-24 20:33:47 +03:00
|
|
|
rdev->config.evergreen.sc_prim_fifo_size = 0x40;
|
|
|
|
rdev->config.evergreen.sc_hiz_tile_fifo_size = 0x30;
|
|
|
|
rdev->config.evergreen.sc_earlyz_tile_fifo_size = 0x130;
|
2012-06-01 03:00:25 +04:00
|
|
|
gb_addr_config = CAICOS_GB_ADDR_CONFIG_GOLDEN;
|
2010-03-24 20:33:47 +03:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Initialize HDP */
|
|
|
|
for (i = 0, j = 0; i < 32; i++, j += 0x18) {
|
|
|
|
WREG32((0x2c14 + j), 0x00000000);
|
|
|
|
WREG32((0x2c18 + j), 0x00000000);
|
|
|
|
WREG32((0x2c1c + j), 0x00000000);
|
|
|
|
WREG32((0x2c20 + j), 0x00000000);
|
|
|
|
WREG32((0x2c24 + j), 0x00000000);
|
|
|
|
}
|
|
|
|
|
|
|
|
WREG32(GRBM_CNTL, GRBM_READ_TIMEOUT(0xff));
|
2015-02-18 15:19:28 +03:00
|
|
|
WREG32(SRBM_INT_CNTL, 0x1);
|
|
|
|
WREG32(SRBM_INT_ACK, 0x1);
|
2010-03-24 20:33:47 +03:00
|
|
|
|
2011-09-01 21:46:15 +04:00
|
|
|
evergreen_fix_pci_max_read_req_size(rdev);
|
|
|
|
|
2010-03-24 20:33:47 +03:00
|
|
|
mc_shared_chmap = RREG32(MC_SHARED_CHMAP);
|
2012-03-21 01:18:37 +04:00
|
|
|
if ((rdev->family == CHIP_PALM) ||
|
|
|
|
(rdev->family == CHIP_SUMO) ||
|
|
|
|
(rdev->family == CHIP_SUMO2))
|
2011-05-11 11:15:24 +04:00
|
|
|
mc_arb_ramcfg = RREG32(FUS_MC_ARB_RAMCFG);
|
|
|
|
else
|
|
|
|
mc_arb_ramcfg = RREG32(MC_ARB_RAMCFG);
|
2010-03-24 20:33:47 +03:00
|
|
|
|
2010-11-17 20:11:03 +03:00
|
|
|
/* setup tiling info dword. gb_addr_config is not adequate since it does
|
|
|
|
* not have bank info, so create a custom tiling dword.
|
|
|
|
* bits 3:0 num_pipes
|
|
|
|
* bits 7:4 num_banks
|
|
|
|
* bits 11:8 group_size
|
|
|
|
* bits 15:12 row_size
|
|
|
|
*/
|
|
|
|
rdev->config.evergreen.tile_config = 0;
|
|
|
|
switch (rdev->config.evergreen.max_tile_pipes) {
|
|
|
|
case 1:
|
|
|
|
default:
|
|
|
|
rdev->config.evergreen.tile_config |= (0 << 0);
|
|
|
|
break;
|
|
|
|
case 2:
|
|
|
|
rdev->config.evergreen.tile_config |= (1 << 0);
|
|
|
|
break;
|
|
|
|
case 4:
|
|
|
|
rdev->config.evergreen.tile_config |= (2 << 0);
|
|
|
|
break;
|
|
|
|
case 8:
|
|
|
|
rdev->config.evergreen.tile_config |= (3 << 0);
|
|
|
|
break;
|
|
|
|
}
|
2011-06-23 08:49:29 +04:00
|
|
|
/* num banks is 8 on all fusion asics. 0 = 4, 1 = 8, 2 = 16 */
|
2011-05-20 20:35:22 +04:00
|
|
|
if (rdev->flags & RADEON_IS_IGP)
|
2011-06-23 08:49:29 +04:00
|
|
|
rdev->config.evergreen.tile_config |= 1 << 4;
|
2012-06-01 02:53:36 +04:00
|
|
|
else {
|
2012-07-31 19:01:10 +04:00
|
|
|
switch ((mc_arb_ramcfg & NOOFBANK_MASK) >> NOOFBANK_SHIFT) {
|
|
|
|
case 0: /* four banks */
|
2012-06-01 02:53:36 +04:00
|
|
|
rdev->config.evergreen.tile_config |= 0 << 4;
|
2012-07-31 19:01:10 +04:00
|
|
|
break;
|
|
|
|
case 1: /* eight banks */
|
|
|
|
rdev->config.evergreen.tile_config |= 1 << 4;
|
|
|
|
break;
|
|
|
|
case 2: /* sixteen banks */
|
|
|
|
default:
|
|
|
|
rdev->config.evergreen.tile_config |= 2 << 4;
|
|
|
|
break;
|
|
|
|
}
|
2012-06-01 02:53:36 +04:00
|
|
|
}
|
2012-06-01 03:00:25 +04:00
|
|
|
rdev->config.evergreen.tile_config |= 0 << 8;
|
2010-11-17 20:11:03 +03:00
|
|
|
rdev->config.evergreen.tile_config |=
|
|
|
|
((gb_addr_config & 0x30000000) >> 28) << 12;
|
|
|
|
|
2012-06-01 03:00:25 +04:00
|
|
|
if ((rdev->family >= CHIP_CEDAR) && (rdev->family <= CHIP_HEMLOCK)) {
|
|
|
|
u32 efuse_straps_4;
|
|
|
|
u32 efuse_straps_3;
|
2010-03-24 20:33:47 +03:00
|
|
|
|
2013-04-12 19:27:20 +04:00
|
|
|
efuse_straps_4 = RREG32_RCU(0x204);
|
|
|
|
efuse_straps_3 = RREG32_RCU(0x203);
|
2012-06-01 03:00:25 +04:00
|
|
|
tmp = (((efuse_straps_4 & 0xf) << 4) |
|
|
|
|
((efuse_straps_3 & 0xf0000000) >> 28));
|
|
|
|
} else {
|
|
|
|
tmp = 0;
|
|
|
|
for (i = (rdev->config.evergreen.num_ses - 1); i >= 0; i--) {
|
|
|
|
u32 rb_disable_bitmap;
|
|
|
|
|
|
|
|
WREG32(GRBM_GFX_INDEX, INSTANCE_BROADCAST_WRITES | SE_INDEX(i));
|
|
|
|
WREG32(RLC_GFX_INDEX, INSTANCE_BROADCAST_WRITES | SE_INDEX(i));
|
|
|
|
rb_disable_bitmap = (RREG32(CC_RB_BACKEND_DISABLE) & 0x00ff0000) >> 16;
|
|
|
|
tmp <<= 4;
|
|
|
|
tmp |= rb_disable_bitmap;
|
2010-03-24 20:33:47 +03:00
|
|
|
}
|
2012-06-01 03:00:25 +04:00
|
|
|
}
|
|
|
|
/* enabled rb are just the one not disabled :) */
|
|
|
|
disabled_rb_mask = tmp;
|
2013-04-09 18:13:22 +04:00
|
|
|
tmp = 0;
|
|
|
|
for (i = 0; i < rdev->config.evergreen.max_backends; i++)
|
|
|
|
tmp |= (1 << i);
|
|
|
|
/* if all the backends are disabled, fix it up here */
|
|
|
|
if ((disabled_rb_mask & tmp) == tmp) {
|
|
|
|
for (i = 0; i < rdev->config.evergreen.max_backends; i++)
|
|
|
|
disabled_rb_mask &= ~(1 << i);
|
|
|
|
}
|
2010-03-24 20:33:47 +03:00
|
|
|
|
2014-06-03 00:13:21 +04:00
|
|
|
for (i = 0; i < rdev->config.evergreen.num_ses; i++) {
|
|
|
|
u32 simd_disable_bitmap;
|
|
|
|
|
|
|
|
WREG32(GRBM_GFX_INDEX, INSTANCE_BROADCAST_WRITES | SE_INDEX(i));
|
|
|
|
WREG32(RLC_GFX_INDEX, INSTANCE_BROADCAST_WRITES | SE_INDEX(i));
|
|
|
|
simd_disable_bitmap = (RREG32(CC_GC_SHADER_PIPE_CONFIG) & 0xffff0000) >> 16;
|
|
|
|
simd_disable_bitmap |= 0xffffffff << rdev->config.evergreen.max_simds;
|
|
|
|
tmp <<= 16;
|
|
|
|
tmp |= simd_disable_bitmap;
|
|
|
|
}
|
|
|
|
rdev->config.evergreen.active_simds = hweight32(~tmp);
|
|
|
|
|
2012-06-01 03:00:25 +04:00
|
|
|
WREG32(GRBM_GFX_INDEX, INSTANCE_BROADCAST_WRITES | SE_BROADCAST_WRITES);
|
|
|
|
WREG32(RLC_GFX_INDEX, INSTANCE_BROADCAST_WRITES | SE_BROADCAST_WRITES);
|
2010-03-24 20:33:47 +03:00
|
|
|
|
2012-06-01 03:00:25 +04:00
|
|
|
WREG32(GB_ADDR_CONFIG, gb_addr_config);
|
|
|
|
WREG32(DMIF_ADDR_CONFIG, gb_addr_config);
|
|
|
|
WREG32(HDP_ADDR_CONFIG, gb_addr_config);
|
2012-12-05 00:25:59 +04:00
|
|
|
WREG32(DMA_TILING_CONFIG, gb_addr_config);
|
2013-04-08 14:41:37 +04:00
|
|
|
WREG32(UVD_UDEC_ADDR_CONFIG, gb_addr_config);
|
|
|
|
WREG32(UVD_UDEC_DB_ADDR_CONFIG, gb_addr_config);
|
|
|
|
WREG32(UVD_UDEC_DBW_ADDR_CONFIG, gb_addr_config);
|
2010-03-24 20:33:47 +03:00
|
|
|
|
2013-01-30 22:57:40 +04:00
|
|
|
if ((rdev->config.evergreen.max_backends == 1) &&
|
|
|
|
(rdev->flags & RADEON_IS_IGP)) {
|
|
|
|
if ((disabled_rb_mask & 3) == 1) {
|
|
|
|
/* RB0 disabled, RB1 enabled */
|
|
|
|
tmp = 0x11111111;
|
|
|
|
} else {
|
|
|
|
/* RB1 disabled, RB0 enabled */
|
|
|
|
tmp = 0x00000000;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
tmp = gb_addr_config & NUM_PIPES_MASK;
|
|
|
|
tmp = r6xx_remap_render_backend(rdev, tmp, rdev->config.evergreen.max_backends,
|
|
|
|
EVERGREEN_MAX_BACKENDS, disabled_rb_mask);
|
|
|
|
}
|
2012-06-01 03:00:25 +04:00
|
|
|
WREG32(GB_BACKEND_MAP, tmp);
|
2010-03-24 20:33:47 +03:00
|
|
|
|
|
|
|
WREG32(CGTS_SYS_TCC_DISABLE, 0);
|
|
|
|
WREG32(CGTS_TCC_DISABLE, 0);
|
|
|
|
WREG32(CGTS_USER_SYS_TCC_DISABLE, 0);
|
|
|
|
WREG32(CGTS_USER_TCC_DISABLE, 0);
|
|
|
|
|
|
|
|
/* set HW defaults for 3D engine */
|
|
|
|
WREG32(CP_QUEUE_THRESHOLDS, (ROQ_IB1_START(0x16) |
|
|
|
|
ROQ_IB2_START(0x2b)));
|
|
|
|
|
|
|
|
WREG32(CP_MEQ_THRESHOLDS, STQ_SPLIT(0x30));
|
|
|
|
|
|
|
|
WREG32(TA_CNTL_AUX, (DISABLE_CUBE_ANISO |
|
|
|
|
SYNC_GRADIENT |
|
|
|
|
SYNC_WALKER |
|
|
|
|
SYNC_ALIGNER));
|
|
|
|
|
|
|
|
sx_debug_1 = RREG32(SX_DEBUG_1);
|
|
|
|
sx_debug_1 |= ENABLE_NEW_SMX_ADDRESS;
|
|
|
|
WREG32(SX_DEBUG_1, sx_debug_1);
|
|
|
|
|
|
|
|
|
|
|
|
smx_dc_ctl0 = RREG32(SMX_DC_CTL0);
|
|
|
|
smx_dc_ctl0 &= ~NUMBER_OF_SETS(0x1ff);
|
|
|
|
smx_dc_ctl0 |= NUMBER_OF_SETS(rdev->config.evergreen.sx_num_of_sets);
|
|
|
|
WREG32(SMX_DC_CTL0, smx_dc_ctl0);
|
|
|
|
|
2012-06-15 00:06:36 +04:00
|
|
|
if (rdev->family <= CHIP_SUMO2)
|
|
|
|
WREG32(SMX_SAR_CTL0, 0x00010000);
|
|
|
|
|
2010-03-24 20:33:47 +03:00
|
|
|
WREG32(SX_EXPORT_BUFFER_SIZES, (COLOR_BUFFER_SIZE((rdev->config.evergreen.sx_max_export_size / 4) - 1) |
|
|
|
|
POSITION_BUFFER_SIZE((rdev->config.evergreen.sx_max_export_pos_size / 4) - 1) |
|
|
|
|
SMX_BUFFER_SIZE((rdev->config.evergreen.sx_max_export_smx_size / 4) - 1)));
|
|
|
|
|
|
|
|
WREG32(PA_SC_FIFO_SIZE, (SC_PRIM_FIFO_SIZE(rdev->config.evergreen.sc_prim_fifo_size) |
|
|
|
|
SC_HIZ_TILE_FIFO_SIZE(rdev->config.evergreen.sc_hiz_tile_fifo_size) |
|
|
|
|
SC_EARLYZ_TILE_FIFO_SIZE(rdev->config.evergreen.sc_earlyz_tile_fifo_size)));
|
|
|
|
|
|
|
|
WREG32(VGT_NUM_INSTANCES, 1);
|
|
|
|
WREG32(SPI_CONFIG_CNTL, 0);
|
|
|
|
WREG32(SPI_CONFIG_CNTL_1, VTX_DONE_DELAY(4));
|
|
|
|
WREG32(CP_PERFMON_CNTL, 0);
|
|
|
|
|
|
|
|
WREG32(SQ_MS_FIFO_SIZES, (CACHE_FIFO_SIZE(16 * rdev->config.evergreen.sq_num_cf_insts) |
|
|
|
|
FETCH_FIFO_HIWATER(0x4) |
|
|
|
|
DONE_FIFO_HIWATER(0xe0) |
|
|
|
|
ALU_UPDATE_FIFO_HIWATER(0x8)));
|
|
|
|
|
|
|
|
sq_config = RREG32(SQ_CONFIG);
|
|
|
|
sq_config &= ~(PS_PRIO(3) |
|
|
|
|
VS_PRIO(3) |
|
|
|
|
GS_PRIO(3) |
|
|
|
|
ES_PRIO(3));
|
|
|
|
sq_config |= (VC_ENABLE |
|
|
|
|
EXPORT_SRC_C |
|
|
|
|
PS_PRIO(0) |
|
|
|
|
VS_PRIO(1) |
|
|
|
|
GS_PRIO(2) |
|
|
|
|
ES_PRIO(3));
|
|
|
|
|
2010-11-23 01:56:29 +03:00
|
|
|
switch (rdev->family) {
|
|
|
|
case CHIP_CEDAR:
|
|
|
|
case CHIP_PALM:
|
2011-05-31 23:42:48 +04:00
|
|
|
case CHIP_SUMO:
|
|
|
|
case CHIP_SUMO2:
|
2011-01-07 05:19:24 +03:00
|
|
|
case CHIP_CAICOS:
|
2010-03-24 20:33:47 +03:00
|
|
|
/* no vertex cache */
|
|
|
|
sq_config &= ~VC_ENABLE;
|
2010-11-23 01:56:29 +03:00
|
|
|
break;
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
2010-03-24 20:33:47 +03:00
|
|
|
|
|
|
|
sq_lds_resource_mgmt = RREG32(SQ_LDS_RESOURCE_MGMT);
|
|
|
|
|
|
|
|
sq_gpr_resource_mgmt_1 = NUM_PS_GPRS((rdev->config.evergreen.max_gprs - (4 * 2))* 12 / 32);
|
|
|
|
sq_gpr_resource_mgmt_1 |= NUM_VS_GPRS((rdev->config.evergreen.max_gprs - (4 * 2)) * 6 / 32);
|
|
|
|
sq_gpr_resource_mgmt_1 |= NUM_CLAUSE_TEMP_GPRS(4);
|
|
|
|
sq_gpr_resource_mgmt_2 = NUM_GS_GPRS((rdev->config.evergreen.max_gprs - (4 * 2)) * 4 / 32);
|
|
|
|
sq_gpr_resource_mgmt_2 |= NUM_ES_GPRS((rdev->config.evergreen.max_gprs - (4 * 2)) * 4 / 32);
|
|
|
|
sq_gpr_resource_mgmt_3 = NUM_HS_GPRS((rdev->config.evergreen.max_gprs - (4 * 2)) * 3 / 32);
|
|
|
|
sq_gpr_resource_mgmt_3 |= NUM_LS_GPRS((rdev->config.evergreen.max_gprs - (4 * 2)) * 3 / 32);
|
|
|
|
|
2010-11-23 01:56:29 +03:00
|
|
|
switch (rdev->family) {
|
|
|
|
case CHIP_CEDAR:
|
|
|
|
case CHIP_PALM:
|
2011-05-31 23:42:48 +04:00
|
|
|
case CHIP_SUMO:
|
|
|
|
case CHIP_SUMO2:
|
2010-03-24 20:33:47 +03:00
|
|
|
ps_thread_count = 96;
|
2010-11-23 01:56:29 +03:00
|
|
|
break;
|
|
|
|
default:
|
2010-03-24 20:33:47 +03:00
|
|
|
ps_thread_count = 128;
|
2010-11-23 01:56:29 +03:00
|
|
|
break;
|
|
|
|
}
|
2010-03-24 20:33:47 +03:00
|
|
|
|
|
|
|
sq_thread_resource_mgmt = NUM_PS_THREADS(ps_thread_count);
|
2010-06-16 20:24:07 +04:00
|
|
|
sq_thread_resource_mgmt |= NUM_VS_THREADS((((rdev->config.evergreen.max_threads - ps_thread_count) / 6) / 8) * 8);
|
|
|
|
sq_thread_resource_mgmt |= NUM_GS_THREADS((((rdev->config.evergreen.max_threads - ps_thread_count) / 6) / 8) * 8);
|
|
|
|
sq_thread_resource_mgmt |= NUM_ES_THREADS((((rdev->config.evergreen.max_threads - ps_thread_count) / 6) / 8) * 8);
|
|
|
|
sq_thread_resource_mgmt_2 = NUM_HS_THREADS((((rdev->config.evergreen.max_threads - ps_thread_count) / 6) / 8) * 8);
|
|
|
|
sq_thread_resource_mgmt_2 |= NUM_LS_THREADS((((rdev->config.evergreen.max_threads - ps_thread_count) / 6) / 8) * 8);
|
2010-03-24 20:33:47 +03:00
|
|
|
|
|
|
|
sq_stack_resource_mgmt_1 = NUM_PS_STACK_ENTRIES((rdev->config.evergreen.max_stack_entries * 1) / 6);
|
|
|
|
sq_stack_resource_mgmt_1 |= NUM_VS_STACK_ENTRIES((rdev->config.evergreen.max_stack_entries * 1) / 6);
|
|
|
|
sq_stack_resource_mgmt_2 = NUM_GS_STACK_ENTRIES((rdev->config.evergreen.max_stack_entries * 1) / 6);
|
|
|
|
sq_stack_resource_mgmt_2 |= NUM_ES_STACK_ENTRIES((rdev->config.evergreen.max_stack_entries * 1) / 6);
|
|
|
|
sq_stack_resource_mgmt_3 = NUM_HS_STACK_ENTRIES((rdev->config.evergreen.max_stack_entries * 1) / 6);
|
|
|
|
sq_stack_resource_mgmt_3 |= NUM_LS_STACK_ENTRIES((rdev->config.evergreen.max_stack_entries * 1) / 6);
|
|
|
|
|
|
|
|
WREG32(SQ_CONFIG, sq_config);
|
|
|
|
WREG32(SQ_GPR_RESOURCE_MGMT_1, sq_gpr_resource_mgmt_1);
|
|
|
|
WREG32(SQ_GPR_RESOURCE_MGMT_2, sq_gpr_resource_mgmt_2);
|
|
|
|
WREG32(SQ_GPR_RESOURCE_MGMT_3, sq_gpr_resource_mgmt_3);
|
|
|
|
WREG32(SQ_THREAD_RESOURCE_MGMT, sq_thread_resource_mgmt);
|
|
|
|
WREG32(SQ_THREAD_RESOURCE_MGMT_2, sq_thread_resource_mgmt_2);
|
|
|
|
WREG32(SQ_STACK_RESOURCE_MGMT_1, sq_stack_resource_mgmt_1);
|
|
|
|
WREG32(SQ_STACK_RESOURCE_MGMT_2, sq_stack_resource_mgmt_2);
|
|
|
|
WREG32(SQ_STACK_RESOURCE_MGMT_3, sq_stack_resource_mgmt_3);
|
|
|
|
WREG32(SQ_DYN_GPR_CNTL_PS_FLUSH_REQ, 0);
|
|
|
|
WREG32(SQ_LDS_RESOURCE_MGMT, sq_lds_resource_mgmt);
|
|
|
|
|
|
|
|
WREG32(PA_SC_FORCE_EOV_MAX_CNTS, (FORCE_EOV_MAX_CLK_CNT(4095) |
|
|
|
|
FORCE_EOV_MAX_REZ_CNT(255)));
|
|
|
|
|
2010-11-23 01:56:29 +03:00
|
|
|
switch (rdev->family) {
|
|
|
|
case CHIP_CEDAR:
|
|
|
|
case CHIP_PALM:
|
2011-05-31 23:42:48 +04:00
|
|
|
case CHIP_SUMO:
|
|
|
|
case CHIP_SUMO2:
|
2011-01-07 05:19:24 +03:00
|
|
|
case CHIP_CAICOS:
|
2010-03-24 20:33:47 +03:00
|
|
|
vgt_cache_invalidation = CACHE_INVALIDATION(TC_ONLY);
|
2010-11-23 01:56:29 +03:00
|
|
|
break;
|
|
|
|
default:
|
2010-03-24 20:33:47 +03:00
|
|
|
vgt_cache_invalidation = CACHE_INVALIDATION(VC_AND_TC);
|
2010-11-23 01:56:29 +03:00
|
|
|
break;
|
|
|
|
}
|
2010-03-24 20:33:47 +03:00
|
|
|
vgt_cache_invalidation |= AUTO_INVLD_EN(ES_AND_GS_AUTO);
|
|
|
|
WREG32(VGT_CACHE_INVALIDATION, vgt_cache_invalidation);
|
|
|
|
|
|
|
|
WREG32(VGT_GS_VERTEX_REUSE, 16);
|
2011-02-02 20:37:40 +03:00
|
|
|
WREG32(PA_SU_LINE_STIPPLE_VALUE, 0);
|
2010-03-24 20:33:47 +03:00
|
|
|
WREG32(PA_SC_LINE_STIPPLE_STATE, 0);
|
|
|
|
|
2010-06-30 01:03:35 +04:00
|
|
|
WREG32(VGT_VERTEX_REUSE_BLOCK_CNTL, 14);
|
|
|
|
WREG32(VGT_OUT_DEALLOC_CNTL, 16);
|
|
|
|
|
2010-03-24 20:33:47 +03:00
|
|
|
WREG32(CB_PERF_CTR0_SEL_0, 0);
|
|
|
|
WREG32(CB_PERF_CTR0_SEL_1, 0);
|
|
|
|
WREG32(CB_PERF_CTR1_SEL_0, 0);
|
|
|
|
WREG32(CB_PERF_CTR1_SEL_1, 0);
|
|
|
|
WREG32(CB_PERF_CTR2_SEL_0, 0);
|
|
|
|
WREG32(CB_PERF_CTR2_SEL_1, 0);
|
|
|
|
WREG32(CB_PERF_CTR3_SEL_0, 0);
|
|
|
|
WREG32(CB_PERF_CTR3_SEL_1, 0);
|
|
|
|
|
2010-06-30 01:03:35 +04:00
|
|
|
/* clear render buffer base addresses */
|
|
|
|
WREG32(CB_COLOR0_BASE, 0);
|
|
|
|
WREG32(CB_COLOR1_BASE, 0);
|
|
|
|
WREG32(CB_COLOR2_BASE, 0);
|
|
|
|
WREG32(CB_COLOR3_BASE, 0);
|
|
|
|
WREG32(CB_COLOR4_BASE, 0);
|
|
|
|
WREG32(CB_COLOR5_BASE, 0);
|
|
|
|
WREG32(CB_COLOR6_BASE, 0);
|
|
|
|
WREG32(CB_COLOR7_BASE, 0);
|
|
|
|
WREG32(CB_COLOR8_BASE, 0);
|
|
|
|
WREG32(CB_COLOR9_BASE, 0);
|
|
|
|
WREG32(CB_COLOR10_BASE, 0);
|
|
|
|
WREG32(CB_COLOR11_BASE, 0);
|
|
|
|
|
|
|
|
/* set the shader const cache sizes to 0 */
|
|
|
|
for (i = SQ_ALU_CONST_BUFFER_SIZE_PS_0; i < 0x28200; i += 4)
|
|
|
|
WREG32(i, 0);
|
|
|
|
for (i = SQ_ALU_CONST_BUFFER_SIZE_HS_0; i < 0x29000; i += 4)
|
|
|
|
WREG32(i, 0);
|
|
|
|
|
2011-05-19 19:07:57 +04:00
|
|
|
tmp = RREG32(HDP_MISC_CNTL);
|
|
|
|
tmp |= HDP_FLUSH_INVALIDATE_CACHE;
|
|
|
|
WREG32(HDP_MISC_CNTL, tmp);
|
|
|
|
|
2010-03-24 20:33:47 +03:00
|
|
|
hdp_host_path_cntl = RREG32(HDP_HOST_PATH_CNTL);
|
|
|
|
WREG32(HDP_HOST_PATH_CNTL, hdp_host_path_cntl);
|
|
|
|
|
|
|
|
WREG32(PA_CL_ENHANCE, CLIP_VTX_REORDER_ENA | NUM_CLIP_SEQ(3));
|
|
|
|
|
|
|
|
udelay(50);
|
|
|
|
|
2010-01-13 01:54:34 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
int evergreen_mc_init(struct radeon_device *rdev)
|
|
|
|
{
|
|
|
|
u32 tmp;
|
|
|
|
int chansize, numchan;
|
|
|
|
|
|
|
|
/* Get VRAM informations */
|
|
|
|
rdev->mc.vram_is_ddr = true;
|
2012-03-21 01:18:37 +04:00
|
|
|
if ((rdev->family == CHIP_PALM) ||
|
|
|
|
(rdev->family == CHIP_SUMO) ||
|
|
|
|
(rdev->family == CHIP_SUMO2))
|
2011-07-01 21:18:28 +04:00
|
|
|
tmp = RREG32(FUS_MC_ARB_RAMCFG);
|
|
|
|
else
|
|
|
|
tmp = RREG32(MC_ARB_RAMCFG);
|
2010-01-13 01:54:34 +03:00
|
|
|
if (tmp & CHANSIZE_OVERRIDE) {
|
|
|
|
chansize = 16;
|
|
|
|
} else if (tmp & CHANSIZE_MASK) {
|
|
|
|
chansize = 64;
|
|
|
|
} else {
|
|
|
|
chansize = 32;
|
|
|
|
}
|
|
|
|
tmp = RREG32(MC_SHARED_CHMAP);
|
|
|
|
switch ((tmp & NOOFCHAN_MASK) >> NOOFCHAN_SHIFT) {
|
|
|
|
case 0:
|
|
|
|
default:
|
|
|
|
numchan = 1;
|
|
|
|
break;
|
|
|
|
case 1:
|
|
|
|
numchan = 2;
|
|
|
|
break;
|
|
|
|
case 2:
|
|
|
|
numchan = 4;
|
|
|
|
break;
|
|
|
|
case 3:
|
|
|
|
numchan = 8;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
rdev->mc.vram_width = numchan * chansize;
|
|
|
|
/* Could aper size report 0 ? */
|
2010-05-27 23:40:24 +04:00
|
|
|
rdev->mc.aper_base = pci_resource_start(rdev->pdev, 0);
|
|
|
|
rdev->mc.aper_size = pci_resource_len(rdev->pdev, 0);
|
2010-01-13 01:54:34 +03:00
|
|
|
/* Setup GPU memory space */
|
2012-03-21 01:18:37 +04:00
|
|
|
if ((rdev->family == CHIP_PALM) ||
|
|
|
|
(rdev->family == CHIP_SUMO) ||
|
|
|
|
(rdev->family == CHIP_SUMO2)) {
|
2010-11-23 01:56:27 +03:00
|
|
|
/* size in bytes on fusion */
|
|
|
|
rdev->mc.mc_vram_size = RREG32(CONFIG_MEMSIZE);
|
|
|
|
rdev->mc.real_vram_size = RREG32(CONFIG_MEMSIZE);
|
|
|
|
} else {
|
2012-03-21 01:18:37 +04:00
|
|
|
/* size in MB on evergreen/cayman/tn */
|
2013-05-18 23:19:23 +04:00
|
|
|
rdev->mc.mc_vram_size = RREG32(CONFIG_MEMSIZE) * 1024ULL * 1024ULL;
|
|
|
|
rdev->mc.real_vram_size = RREG32(CONFIG_MEMSIZE) * 1024ULL * 1024ULL;
|
2010-11-23 01:56:27 +03:00
|
|
|
}
|
2010-02-19 17:33:54 +03:00
|
|
|
rdev->mc.visible_vram_size = rdev->mc.aper_size;
|
2010-11-23 01:56:26 +03:00
|
|
|
r700_vram_gtt_location(rdev, &rdev->mc);
|
2010-03-17 03:54:38 +03:00
|
|
|
radeon_update_bandwidth_info(rdev);
|
|
|
|
|
2010-01-13 01:54:34 +03:00
|
|
|
return 0;
|
|
|
|
}
|
drm/radeon/kms: simplify memory controller setup V2
Get rid of _location and use _start/_end also simplify the
computation of vram_start|end & gtt_start|end. For R1XX-R2XX
we place VRAM at the same address of PCI aperture, those GPU
shouldn't have much memory and seems to behave better when
setup that way. For R3XX and newer we place VRAM at 0. For
R6XX-R7XX AGP we place VRAM before or after AGP aperture this
might limit to limit the VRAM size but it's very unlikely.
For IGP we don't change the VRAM placement.
Tested on (compiz,quake3,suspend/resume):
PCI/PCIE:RV280,R420,RV515,RV570,RV610,RV710
AGP:RV100,RV280,R420,RV350,RV620(RPB*),RV730
IGP:RS480(RPB*),RS690,RS780(RPB*),RS880
RPB: resume previously broken
V2 correct commit message to reflect more accurately the bug
and move VRAM placement to 0 for most of the GPU to avoid
limiting VRAM.
Signed-off-by: Jerome Glisse <jglisse@redhat.com>
Signed-off-by: Dave Airlie <airlied@redhat.com>
2010-02-18 00:54:29 +03:00
|
|
|
|
2013-01-18 23:51:38 +04:00
|
|
|
void evergreen_print_gpu_status_regs(struct radeon_device *rdev)
|
2010-01-13 01:54:34 +03:00
|
|
|
{
|
2013-01-03 02:30:35 +04:00
|
|
|
dev_info(rdev->dev, " GRBM_STATUS = 0x%08X\n",
|
2010-03-24 20:26:36 +03:00
|
|
|
RREG32(GRBM_STATUS));
|
2013-01-03 02:30:35 +04:00
|
|
|
dev_info(rdev->dev, " GRBM_STATUS_SE0 = 0x%08X\n",
|
2010-03-24 20:26:36 +03:00
|
|
|
RREG32(GRBM_STATUS_SE0));
|
2013-01-03 02:30:35 +04:00
|
|
|
dev_info(rdev->dev, " GRBM_STATUS_SE1 = 0x%08X\n",
|
2010-03-24 20:26:36 +03:00
|
|
|
RREG32(GRBM_STATUS_SE1));
|
2013-01-03 02:30:35 +04:00
|
|
|
dev_info(rdev->dev, " SRBM_STATUS = 0x%08X\n",
|
2010-03-24 20:26:36 +03:00
|
|
|
RREG32(SRBM_STATUS));
|
2013-01-19 03:55:54 +04:00
|
|
|
dev_info(rdev->dev, " SRBM_STATUS2 = 0x%08X\n",
|
|
|
|
RREG32(SRBM_STATUS2));
|
2012-06-27 20:25:01 +04:00
|
|
|
dev_info(rdev->dev, " R_008674_CP_STALLED_STAT1 = 0x%08X\n",
|
|
|
|
RREG32(CP_STALLED_STAT1));
|
|
|
|
dev_info(rdev->dev, " R_008678_CP_STALLED_STAT2 = 0x%08X\n",
|
|
|
|
RREG32(CP_STALLED_STAT2));
|
|
|
|
dev_info(rdev->dev, " R_00867C_CP_BUSY_STAT = 0x%08X\n",
|
|
|
|
RREG32(CP_BUSY_STAT));
|
|
|
|
dev_info(rdev->dev, " R_008680_CP_STAT = 0x%08X\n",
|
|
|
|
RREG32(CP_STAT));
|
2013-01-03 00:12:15 +04:00
|
|
|
dev_info(rdev->dev, " R_00D034_DMA_STATUS_REG = 0x%08X\n",
|
|
|
|
RREG32(DMA_STATUS_REG));
|
2013-01-19 04:17:22 +04:00
|
|
|
if (rdev->family >= CHIP_CAYMAN) {
|
|
|
|
dev_info(rdev->dev, " R_00D834_DMA_STATUS_REG = 0x%08X\n",
|
|
|
|
RREG32(DMA_STATUS_REG + 0x800));
|
|
|
|
}
|
2013-01-03 21:40:13 +04:00
|
|
|
}
|
|
|
|
|
2013-01-19 04:17:22 +04:00
|
|
|
bool evergreen_is_display_hung(struct radeon_device *rdev)
|
2013-01-03 21:40:13 +04:00
|
|
|
{
|
2013-01-19 03:55:54 +04:00
|
|
|
u32 crtc_hung = 0;
|
|
|
|
u32 crtc_status[6];
|
|
|
|
u32 i, j, tmp;
|
|
|
|
|
|
|
|
for (i = 0; i < rdev->num_crtc; i++) {
|
|
|
|
if (RREG32(EVERGREEN_CRTC_CONTROL + crtc_offsets[i]) & EVERGREEN_CRTC_MASTER_EN) {
|
|
|
|
crtc_status[i] = RREG32(EVERGREEN_CRTC_STATUS_HV_COUNT + crtc_offsets[i]);
|
|
|
|
crtc_hung |= (1 << i);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
for (j = 0; j < 10; j++) {
|
|
|
|
for (i = 0; i < rdev->num_crtc; i++) {
|
|
|
|
if (crtc_hung & (1 << i)) {
|
|
|
|
tmp = RREG32(EVERGREEN_CRTC_STATUS_HV_COUNT + crtc_offsets[i]);
|
|
|
|
if (tmp != crtc_status[i])
|
|
|
|
crtc_hung &= ~(1 << i);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (crtc_hung == 0)
|
|
|
|
return false;
|
|
|
|
udelay(100);
|
|
|
|
}
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2013-08-13 13:56:54 +04:00
|
|
|
u32 evergreen_gpu_check_soft_reset(struct radeon_device *rdev)
|
2013-01-19 03:55:54 +04:00
|
|
|
{
|
|
|
|
u32 reset_mask = 0;
|
2013-01-18 23:28:41 +04:00
|
|
|
u32 tmp;
|
2013-01-03 21:40:13 +04:00
|
|
|
|
2013-01-19 03:55:54 +04:00
|
|
|
/* GRBM_STATUS */
|
|
|
|
tmp = RREG32(GRBM_STATUS);
|
|
|
|
if (tmp & (PA_BUSY | SC_BUSY |
|
|
|
|
SH_BUSY | SX_BUSY |
|
|
|
|
TA_BUSY | VGT_BUSY |
|
|
|
|
DB_BUSY | CB_BUSY |
|
|
|
|
SPI_BUSY | VGT_BUSY_NO_DMA))
|
|
|
|
reset_mask |= RADEON_RESET_GFX;
|
|
|
|
|
|
|
|
if (tmp & (CF_RQ_PENDING | PF_RQ_PENDING |
|
|
|
|
CP_BUSY | CP_COHERENCY_BUSY))
|
|
|
|
reset_mask |= RADEON_RESET_CP;
|
|
|
|
|
|
|
|
if (tmp & GRBM_EE_BUSY)
|
|
|
|
reset_mask |= RADEON_RESET_GRBM | RADEON_RESET_GFX | RADEON_RESET_CP;
|
2013-01-14 20:04:39 +04:00
|
|
|
|
2013-01-19 03:55:54 +04:00
|
|
|
/* DMA_STATUS_REG */
|
|
|
|
tmp = RREG32(DMA_STATUS_REG);
|
|
|
|
if (!(tmp & DMA_IDLE))
|
|
|
|
reset_mask |= RADEON_RESET_DMA;
|
|
|
|
|
|
|
|
/* SRBM_STATUS2 */
|
|
|
|
tmp = RREG32(SRBM_STATUS2);
|
|
|
|
if (tmp & DMA_BUSY)
|
|
|
|
reset_mask |= RADEON_RESET_DMA;
|
|
|
|
|
|
|
|
/* SRBM_STATUS */
|
|
|
|
tmp = RREG32(SRBM_STATUS);
|
|
|
|
if (tmp & (RLC_RQ_PENDING | RLC_BUSY))
|
|
|
|
reset_mask |= RADEON_RESET_RLC;
|
|
|
|
|
|
|
|
if (tmp & IH_BUSY)
|
|
|
|
reset_mask |= RADEON_RESET_IH;
|
|
|
|
|
|
|
|
if (tmp & SEM_BUSY)
|
|
|
|
reset_mask |= RADEON_RESET_SEM;
|
|
|
|
|
|
|
|
if (tmp & GRBM_RQ_PENDING)
|
|
|
|
reset_mask |= RADEON_RESET_GRBM;
|
|
|
|
|
|
|
|
if (tmp & VMC_BUSY)
|
|
|
|
reset_mask |= RADEON_RESET_VMC;
|
|
|
|
|
|
|
|
if (tmp & (MCB_BUSY | MCB_NON_DISPLAY_BUSY |
|
|
|
|
MCC_BUSY | MCD_BUSY))
|
|
|
|
reset_mask |= RADEON_RESET_MC;
|
|
|
|
|
|
|
|
if (evergreen_is_display_hung(rdev))
|
|
|
|
reset_mask |= RADEON_RESET_DISPLAY;
|
|
|
|
|
|
|
|
/* VM_L2_STATUS */
|
|
|
|
tmp = RREG32(VM_L2_STATUS);
|
|
|
|
if (tmp & L2_BUSY)
|
|
|
|
reset_mask |= RADEON_RESET_VMC;
|
|
|
|
|
2013-02-28 19:03:08 +04:00
|
|
|
/* Skip MC reset as it's mostly likely not hung, just busy */
|
|
|
|
if (reset_mask & RADEON_RESET_MC) {
|
|
|
|
DRM_DEBUG("MC busy: 0x%08X, clearing.\n", reset_mask);
|
|
|
|
reset_mask &= ~RADEON_RESET_MC;
|
|
|
|
}
|
|
|
|
|
2013-01-19 03:55:54 +04:00
|
|
|
return reset_mask;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void evergreen_gpu_soft_reset(struct radeon_device *rdev, u32 reset_mask)
|
|
|
|
{
|
|
|
|
struct evergreen_mc_save save;
|
|
|
|
u32 grbm_soft_reset = 0, srbm_soft_reset = 0;
|
|
|
|
u32 tmp;
|
2013-01-14 20:04:39 +04:00
|
|
|
|
2013-01-03 21:40:13 +04:00
|
|
|
if (reset_mask == 0)
|
2013-01-19 03:55:54 +04:00
|
|
|
return;
|
2013-01-03 21:40:13 +04:00
|
|
|
|
|
|
|
dev_info(rdev->dev, "GPU softreset: 0x%08X\n", reset_mask);
|
|
|
|
|
2013-01-18 23:28:41 +04:00
|
|
|
evergreen_print_gpu_status_regs(rdev);
|
|
|
|
|
|
|
|
/* Disable CP parsing/prefetching */
|
|
|
|
WREG32(CP_ME_CNTL, CP_ME_HALT | CP_PFP_HALT);
|
|
|
|
|
|
|
|
if (reset_mask & RADEON_RESET_DMA) {
|
|
|
|
/* Disable DMA */
|
|
|
|
tmp = RREG32(DMA_RB_CNTL);
|
|
|
|
tmp &= ~DMA_RB_ENABLE;
|
|
|
|
WREG32(DMA_RB_CNTL, tmp);
|
|
|
|
}
|
|
|
|
|
2013-01-24 03:57:56 +04:00
|
|
|
udelay(50);
|
|
|
|
|
|
|
|
evergreen_mc_stop(rdev, &save);
|
|
|
|
if (evergreen_mc_wait_for_idle(rdev)) {
|
|
|
|
dev_warn(rdev->dev, "Wait for MC idle timedout !\n");
|
|
|
|
}
|
|
|
|
|
2013-01-18 23:28:41 +04:00
|
|
|
if (reset_mask & (RADEON_RESET_GFX | RADEON_RESET_COMPUTE)) {
|
|
|
|
grbm_soft_reset |= SOFT_RESET_DB |
|
|
|
|
SOFT_RESET_CB |
|
|
|
|
SOFT_RESET_PA |
|
|
|
|
SOFT_RESET_SC |
|
|
|
|
SOFT_RESET_SPI |
|
|
|
|
SOFT_RESET_SX |
|
|
|
|
SOFT_RESET_SH |
|
|
|
|
SOFT_RESET_TC |
|
|
|
|
SOFT_RESET_TA |
|
|
|
|
SOFT_RESET_VC |
|
|
|
|
SOFT_RESET_VGT;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (reset_mask & RADEON_RESET_CP) {
|
|
|
|
grbm_soft_reset |= SOFT_RESET_CP |
|
|
|
|
SOFT_RESET_VGT;
|
|
|
|
|
|
|
|
srbm_soft_reset |= SOFT_RESET_GRBM;
|
|
|
|
}
|
2013-01-03 21:40:13 +04:00
|
|
|
|
|
|
|
if (reset_mask & RADEON_RESET_DMA)
|
2013-01-18 23:28:41 +04:00
|
|
|
srbm_soft_reset |= SOFT_RESET_DMA;
|
|
|
|
|
2013-01-19 03:55:54 +04:00
|
|
|
if (reset_mask & RADEON_RESET_DISPLAY)
|
|
|
|
srbm_soft_reset |= SOFT_RESET_DC;
|
|
|
|
|
|
|
|
if (reset_mask & RADEON_RESET_RLC)
|
|
|
|
srbm_soft_reset |= SOFT_RESET_RLC;
|
|
|
|
|
|
|
|
if (reset_mask & RADEON_RESET_SEM)
|
|
|
|
srbm_soft_reset |= SOFT_RESET_SEM;
|
|
|
|
|
|
|
|
if (reset_mask & RADEON_RESET_IH)
|
|
|
|
srbm_soft_reset |= SOFT_RESET_IH;
|
|
|
|
|
|
|
|
if (reset_mask & RADEON_RESET_GRBM)
|
|
|
|
srbm_soft_reset |= SOFT_RESET_GRBM;
|
|
|
|
|
|
|
|
if (reset_mask & RADEON_RESET_VMC)
|
|
|
|
srbm_soft_reset |= SOFT_RESET_VMC;
|
|
|
|
|
2013-01-25 00:00:17 +04:00
|
|
|
if (!(rdev->flags & RADEON_IS_IGP)) {
|
|
|
|
if (reset_mask & RADEON_RESET_MC)
|
|
|
|
srbm_soft_reset |= SOFT_RESET_MC;
|
|
|
|
}
|
2013-01-19 03:55:54 +04:00
|
|
|
|
2013-01-18 23:28:41 +04:00
|
|
|
if (grbm_soft_reset) {
|
|
|
|
tmp = RREG32(GRBM_SOFT_RESET);
|
|
|
|
tmp |= grbm_soft_reset;
|
|
|
|
dev_info(rdev->dev, "GRBM_SOFT_RESET=0x%08X\n", tmp);
|
|
|
|
WREG32(GRBM_SOFT_RESET, tmp);
|
|
|
|
tmp = RREG32(GRBM_SOFT_RESET);
|
|
|
|
|
|
|
|
udelay(50);
|
|
|
|
|
|
|
|
tmp &= ~grbm_soft_reset;
|
|
|
|
WREG32(GRBM_SOFT_RESET, tmp);
|
|
|
|
tmp = RREG32(GRBM_SOFT_RESET);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (srbm_soft_reset) {
|
|
|
|
tmp = RREG32(SRBM_SOFT_RESET);
|
|
|
|
tmp |= srbm_soft_reset;
|
|
|
|
dev_info(rdev->dev, "SRBM_SOFT_RESET=0x%08X\n", tmp);
|
|
|
|
WREG32(SRBM_SOFT_RESET, tmp);
|
|
|
|
tmp = RREG32(SRBM_SOFT_RESET);
|
|
|
|
|
|
|
|
udelay(50);
|
|
|
|
|
|
|
|
tmp &= ~srbm_soft_reset;
|
|
|
|
WREG32(SRBM_SOFT_RESET, tmp);
|
|
|
|
tmp = RREG32(SRBM_SOFT_RESET);
|
|
|
|
}
|
2013-01-03 21:40:13 +04:00
|
|
|
|
|
|
|
/* Wait a little for things to settle down */
|
|
|
|
udelay(50);
|
|
|
|
|
2010-03-24 20:26:36 +03:00
|
|
|
evergreen_mc_resume(rdev, &save);
|
2013-01-18 23:28:41 +04:00
|
|
|
udelay(50);
|
|
|
|
|
|
|
|
evergreen_print_gpu_status_regs(rdev);
|
2010-01-13 01:54:34 +03:00
|
|
|
}
|
|
|
|
|
2013-11-02 00:25:10 +04:00
|
|
|
void evergreen_gpu_pci_config_reset(struct radeon_device *rdev)
|
|
|
|
{
|
|
|
|
struct evergreen_mc_save save;
|
|
|
|
u32 tmp, i;
|
|
|
|
|
|
|
|
dev_info(rdev->dev, "GPU pci config reset\n");
|
|
|
|
|
|
|
|
/* disable dpm? */
|
|
|
|
|
|
|
|
/* Disable CP parsing/prefetching */
|
|
|
|
WREG32(CP_ME_CNTL, CP_ME_HALT | CP_PFP_HALT);
|
|
|
|
udelay(50);
|
|
|
|
/* Disable DMA */
|
|
|
|
tmp = RREG32(DMA_RB_CNTL);
|
|
|
|
tmp &= ~DMA_RB_ENABLE;
|
|
|
|
WREG32(DMA_RB_CNTL, tmp);
|
|
|
|
/* XXX other engines? */
|
|
|
|
|
|
|
|
/* halt the rlc */
|
|
|
|
r600_rlc_stop(rdev);
|
|
|
|
|
|
|
|
udelay(50);
|
|
|
|
|
|
|
|
/* set mclk/sclk to bypass */
|
|
|
|
rv770_set_clk_bypass_mode(rdev);
|
|
|
|
/* disable BM */
|
|
|
|
pci_clear_master(rdev->pdev);
|
|
|
|
/* disable mem access */
|
|
|
|
evergreen_mc_stop(rdev, &save);
|
|
|
|
if (evergreen_mc_wait_for_idle(rdev)) {
|
|
|
|
dev_warn(rdev->dev, "Wait for MC idle timed out !\n");
|
|
|
|
}
|
|
|
|
/* reset */
|
|
|
|
radeon_pci_config_reset(rdev);
|
|
|
|
/* wait for asic to come out of reset */
|
|
|
|
for (i = 0; i < rdev->usec_timeout; i++) {
|
|
|
|
if (RREG32(CONFIG_MEMSIZE) != 0xffffffff)
|
|
|
|
break;
|
|
|
|
udelay(1);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2016-03-18 18:58:38 +03:00
|
|
|
int evergreen_asic_reset(struct radeon_device *rdev, bool hard)
|
2010-01-13 01:54:34 +03:00
|
|
|
{
|
2013-01-19 03:55:54 +04:00
|
|
|
u32 reset_mask;
|
|
|
|
|
2016-03-18 18:58:38 +03:00
|
|
|
if (hard) {
|
|
|
|
evergreen_gpu_pci_config_reset(rdev);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2013-01-19 03:55:54 +04:00
|
|
|
reset_mask = evergreen_gpu_check_soft_reset(rdev);
|
|
|
|
|
|
|
|
if (reset_mask)
|
|
|
|
r600_set_bios_scratch_engine_hung(rdev, true);
|
|
|
|
|
2013-11-02 00:25:10 +04:00
|
|
|
/* try soft reset */
|
2013-01-19 03:55:54 +04:00
|
|
|
evergreen_gpu_soft_reset(rdev, reset_mask);
|
|
|
|
|
|
|
|
reset_mask = evergreen_gpu_check_soft_reset(rdev);
|
|
|
|
|
2013-11-02 00:25:10 +04:00
|
|
|
/* try pci config reset */
|
|
|
|
if (reset_mask && radeon_hard_reset)
|
|
|
|
evergreen_gpu_pci_config_reset(rdev);
|
|
|
|
|
|
|
|
reset_mask = evergreen_gpu_check_soft_reset(rdev);
|
|
|
|
|
2013-01-19 03:55:54 +04:00
|
|
|
if (!reset_mask)
|
|
|
|
r600_set_bios_scratch_engine_hung(rdev, false);
|
|
|
|
|
|
|
|
return 0;
|
2010-03-24 20:26:36 +03:00
|
|
|
}
|
|
|
|
|
2013-01-24 20:37:19 +04:00
|
|
|
/**
|
|
|
|
* evergreen_gfx_is_lockup - Check if the GFX engine is locked up
|
|
|
|
*
|
|
|
|
* @rdev: radeon_device pointer
|
|
|
|
* @ring: radeon_ring structure holding ring information
|
|
|
|
*
|
|
|
|
* Check if the GFX engine is locked up.
|
|
|
|
* Returns true if the engine appears to be locked up, false if not.
|
|
|
|
*/
|
|
|
|
bool evergreen_gfx_is_lockup(struct radeon_device *rdev, struct radeon_ring *ring)
|
|
|
|
{
|
|
|
|
u32 reset_mask = evergreen_gpu_check_soft_reset(rdev);
|
|
|
|
|
|
|
|
if (!(reset_mask & (RADEON_RESET_GFX |
|
|
|
|
RADEON_RESET_COMPUTE |
|
|
|
|
RADEON_RESET_CP))) {
|
2014-02-18 17:52:33 +04:00
|
|
|
radeon_ring_lockup_update(rdev, ring);
|
2013-01-24 20:37:19 +04:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
return radeon_ring_test_lockup(rdev, ring);
|
|
|
|
}
|
|
|
|
|
2013-04-12 21:52:52 +04:00
|
|
|
/*
|
|
|
|
* RLC
|
|
|
|
*/
|
|
|
|
#define RLC_SAVE_RESTORE_LIST_END_MARKER 0x00000000
|
|
|
|
#define RLC_CLEAR_STATE_END_MARKER 0x00000001
|
|
|
|
|
|
|
|
void sumo_rlc_fini(struct radeon_device *rdev)
|
|
|
|
{
|
|
|
|
int r;
|
|
|
|
|
|
|
|
/* save restore block */
|
|
|
|
if (rdev->rlc.save_restore_obj) {
|
|
|
|
r = radeon_bo_reserve(rdev->rlc.save_restore_obj, false);
|
|
|
|
if (unlikely(r != 0))
|
|
|
|
dev_warn(rdev->dev, "(%d) reserve RLC sr bo failed\n", r);
|
|
|
|
radeon_bo_unpin(rdev->rlc.save_restore_obj);
|
|
|
|
radeon_bo_unreserve(rdev->rlc.save_restore_obj);
|
|
|
|
|
|
|
|
radeon_bo_unref(&rdev->rlc.save_restore_obj);
|
|
|
|
rdev->rlc.save_restore_obj = NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* clear state block */
|
|
|
|
if (rdev->rlc.clear_state_obj) {
|
|
|
|
r = radeon_bo_reserve(rdev->rlc.clear_state_obj, false);
|
|
|
|
if (unlikely(r != 0))
|
|
|
|
dev_warn(rdev->dev, "(%d) reserve RLC c bo failed\n", r);
|
|
|
|
radeon_bo_unpin(rdev->rlc.clear_state_obj);
|
|
|
|
radeon_bo_unreserve(rdev->rlc.clear_state_obj);
|
|
|
|
|
|
|
|
radeon_bo_unref(&rdev->rlc.clear_state_obj);
|
|
|
|
rdev->rlc.clear_state_obj = NULL;
|
|
|
|
}
|
2013-07-23 17:41:05 +04:00
|
|
|
|
|
|
|
/* clear state block */
|
|
|
|
if (rdev->rlc.cp_table_obj) {
|
|
|
|
r = radeon_bo_reserve(rdev->rlc.cp_table_obj, false);
|
|
|
|
if (unlikely(r != 0))
|
|
|
|
dev_warn(rdev->dev, "(%d) reserve RLC cp table bo failed\n", r);
|
|
|
|
radeon_bo_unpin(rdev->rlc.cp_table_obj);
|
|
|
|
radeon_bo_unreserve(rdev->rlc.cp_table_obj);
|
|
|
|
|
|
|
|
radeon_bo_unref(&rdev->rlc.cp_table_obj);
|
|
|
|
rdev->rlc.cp_table_obj = NULL;
|
|
|
|
}
|
2013-04-12 21:52:52 +04:00
|
|
|
}
|
|
|
|
|
2013-07-23 17:41:05 +04:00
|
|
|
#define CP_ME_TABLE_SIZE 96
|
|
|
|
|
2013-04-12 21:52:52 +04:00
|
|
|
int sumo_rlc_init(struct radeon_device *rdev)
|
|
|
|
{
|
2013-04-18 01:53:50 +04:00
|
|
|
const u32 *src_ptr;
|
2013-04-12 21:52:52 +04:00
|
|
|
volatile u32 *dst_ptr;
|
|
|
|
u32 dws, data, i, j, k, reg_num;
|
2013-08-13 20:48:06 +04:00
|
|
|
u32 reg_list_num, reg_list_hdr_blk_index, reg_list_blk_index = 0;
|
2013-04-12 21:52:52 +04:00
|
|
|
u64 reg_list_mc_addr;
|
2013-04-18 01:53:50 +04:00
|
|
|
const struct cs_section_def *cs_data;
|
2013-04-12 21:52:52 +04:00
|
|
|
int r;
|
|
|
|
|
|
|
|
src_ptr = rdev->rlc.reg_list;
|
|
|
|
dws = rdev->rlc.reg_list_size;
|
2013-08-22 19:57:46 +04:00
|
|
|
if (rdev->family >= CHIP_BONAIRE) {
|
|
|
|
dws += (5 * 16) + 48 + 48 + 64;
|
|
|
|
}
|
2013-04-12 21:52:52 +04:00
|
|
|
cs_data = rdev->rlc.cs_data;
|
|
|
|
|
2013-04-18 01:22:05 +04:00
|
|
|
if (src_ptr) {
|
|
|
|
/* save restore block */
|
|
|
|
if (rdev->rlc.save_restore_obj == NULL) {
|
|
|
|
r = radeon_bo_create(rdev, dws * 4, PAGE_SIZE, true,
|
2014-07-17 14:01:08 +04:00
|
|
|
RADEON_GEM_DOMAIN_VRAM, 0, NULL,
|
2014-09-18 16:11:56 +04:00
|
|
|
NULL, &rdev->rlc.save_restore_obj);
|
2013-04-18 01:22:05 +04:00
|
|
|
if (r) {
|
|
|
|
dev_warn(rdev->dev, "(%d) create RLC sr bo failed\n", r);
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
r = radeon_bo_reserve(rdev->rlc.save_restore_obj, false);
|
|
|
|
if (unlikely(r != 0)) {
|
|
|
|
sumo_rlc_fini(rdev);
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
r = radeon_bo_pin(rdev->rlc.save_restore_obj, RADEON_GEM_DOMAIN_VRAM,
|
|
|
|
&rdev->rlc.save_restore_gpu_addr);
|
2013-04-12 21:52:52 +04:00
|
|
|
if (r) {
|
2013-04-18 01:22:05 +04:00
|
|
|
radeon_bo_unreserve(rdev->rlc.save_restore_obj);
|
|
|
|
dev_warn(rdev->dev, "(%d) pin RLC sr bo failed\n", r);
|
|
|
|
sumo_rlc_fini(rdev);
|
2013-04-12 21:52:52 +04:00
|
|
|
return r;
|
|
|
|
}
|
|
|
|
|
2013-04-18 01:22:05 +04:00
|
|
|
r = radeon_bo_kmap(rdev->rlc.save_restore_obj, (void **)&rdev->rlc.sr_ptr);
|
|
|
|
if (r) {
|
|
|
|
dev_warn(rdev->dev, "(%d) map RLC sr bo failed\n", r);
|
|
|
|
sumo_rlc_fini(rdev);
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
/* write the sr buffer */
|
|
|
|
dst_ptr = rdev->rlc.sr_ptr;
|
2013-04-18 01:53:50 +04:00
|
|
|
if (rdev->family >= CHIP_TAHITI) {
|
|
|
|
/* SI */
|
2013-08-13 20:48:06 +04:00
|
|
|
for (i = 0; i < rdev->rlc.reg_list_size; i++)
|
2013-10-24 02:27:10 +04:00
|
|
|
dst_ptr[i] = cpu_to_le32(src_ptr[i]);
|
2013-04-18 01:53:50 +04:00
|
|
|
} else {
|
|
|
|
/* ON/LN/TN */
|
|
|
|
/* format:
|
|
|
|
* dw0: (reg2 << 16) | reg1
|
|
|
|
* dw1: reg1 save space
|
|
|
|
* dw2: reg2 save space
|
|
|
|
*/
|
|
|
|
for (i = 0; i < dws; i++) {
|
|
|
|
data = src_ptr[i] >> 2;
|
|
|
|
i++;
|
|
|
|
if (i < dws)
|
|
|
|
data |= (src_ptr[i] >> 2) << 16;
|
|
|
|
j = (((i - 1) * 3) / 2);
|
2013-10-24 02:27:10 +04:00
|
|
|
dst_ptr[j] = cpu_to_le32(data);
|
2013-04-18 01:53:50 +04:00
|
|
|
}
|
|
|
|
j = ((i * 3) / 2);
|
2013-10-24 02:27:10 +04:00
|
|
|
dst_ptr[j] = cpu_to_le32(RLC_SAVE_RESTORE_LIST_END_MARKER);
|
2013-04-18 01:22:05 +04:00
|
|
|
}
|
|
|
|
radeon_bo_kunmap(rdev->rlc.save_restore_obj);
|
2013-04-12 21:52:52 +04:00
|
|
|
radeon_bo_unreserve(rdev->rlc.save_restore_obj);
|
|
|
|
}
|
|
|
|
|
2013-04-18 01:22:05 +04:00
|
|
|
if (cs_data) {
|
|
|
|
/* clear state block */
|
2013-08-22 19:57:46 +04:00
|
|
|
if (rdev->family >= CHIP_BONAIRE) {
|
|
|
|
rdev->rlc.clear_state_size = dws = cik_get_csb_size(rdev);
|
|
|
|
} else if (rdev->family >= CHIP_TAHITI) {
|
2013-08-13 20:48:06 +04:00
|
|
|
rdev->rlc.clear_state_size = si_get_csb_size(rdev);
|
|
|
|
dws = rdev->rlc.clear_state_size + (256 / 4);
|
|
|
|
} else {
|
|
|
|
reg_list_num = 0;
|
|
|
|
dws = 0;
|
|
|
|
for (i = 0; cs_data[i].section != NULL; i++) {
|
|
|
|
for (j = 0; cs_data[i].section[j].extent != NULL; j++) {
|
|
|
|
reg_list_num++;
|
|
|
|
dws += cs_data[i].section[j].reg_count;
|
|
|
|
}
|
2013-04-18 01:22:05 +04:00
|
|
|
}
|
2013-08-13 20:48:06 +04:00
|
|
|
reg_list_blk_index = (3 * reg_list_num + 2);
|
|
|
|
dws += reg_list_blk_index;
|
|
|
|
rdev->rlc.clear_state_size = dws;
|
2013-04-12 21:52:52 +04:00
|
|
|
}
|
|
|
|
|
2013-04-18 01:22:05 +04:00
|
|
|
if (rdev->rlc.clear_state_obj == NULL) {
|
2013-08-13 20:48:06 +04:00
|
|
|
r = radeon_bo_create(rdev, dws * 4, PAGE_SIZE, true,
|
2014-07-17 14:01:08 +04:00
|
|
|
RADEON_GEM_DOMAIN_VRAM, 0, NULL,
|
2014-09-18 16:11:56 +04:00
|
|
|
NULL, &rdev->rlc.clear_state_obj);
|
2013-04-18 01:22:05 +04:00
|
|
|
if (r) {
|
|
|
|
dev_warn(rdev->dev, "(%d) create RLC c bo failed\n", r);
|
|
|
|
sumo_rlc_fini(rdev);
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
r = radeon_bo_reserve(rdev->rlc.clear_state_obj, false);
|
|
|
|
if (unlikely(r != 0)) {
|
|
|
|
sumo_rlc_fini(rdev);
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
r = radeon_bo_pin(rdev->rlc.clear_state_obj, RADEON_GEM_DOMAIN_VRAM,
|
|
|
|
&rdev->rlc.clear_state_gpu_addr);
|
2013-04-12 21:52:52 +04:00
|
|
|
if (r) {
|
2013-04-18 01:22:05 +04:00
|
|
|
radeon_bo_unreserve(rdev->rlc.clear_state_obj);
|
|
|
|
dev_warn(rdev->dev, "(%d) pin RLC c bo failed\n", r);
|
2013-04-12 21:52:52 +04:00
|
|
|
sumo_rlc_fini(rdev);
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
|
2013-04-18 01:22:05 +04:00
|
|
|
r = radeon_bo_kmap(rdev->rlc.clear_state_obj, (void **)&rdev->rlc.cs_ptr);
|
|
|
|
if (r) {
|
|
|
|
dev_warn(rdev->dev, "(%d) map RLC c bo failed\n", r);
|
|
|
|
sumo_rlc_fini(rdev);
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
/* set up the cs buffer */
|
|
|
|
dst_ptr = rdev->rlc.cs_ptr;
|
2013-08-22 19:57:46 +04:00
|
|
|
if (rdev->family >= CHIP_BONAIRE) {
|
|
|
|
cik_get_csb_buffer(rdev, dst_ptr);
|
|
|
|
} else if (rdev->family >= CHIP_TAHITI) {
|
2013-08-13 20:48:06 +04:00
|
|
|
reg_list_mc_addr = rdev->rlc.clear_state_gpu_addr + 256;
|
2013-10-24 02:27:10 +04:00
|
|
|
dst_ptr[0] = cpu_to_le32(upper_32_bits(reg_list_mc_addr));
|
|
|
|
dst_ptr[1] = cpu_to_le32(lower_32_bits(reg_list_mc_addr));
|
|
|
|
dst_ptr[2] = cpu_to_le32(rdev->rlc.clear_state_size);
|
2013-08-13 20:48:06 +04:00
|
|
|
si_get_csb_buffer(rdev, &dst_ptr[(256/4)]);
|
|
|
|
} else {
|
|
|
|
reg_list_hdr_blk_index = 0;
|
|
|
|
reg_list_mc_addr = rdev->rlc.clear_state_gpu_addr + (reg_list_blk_index * 4);
|
|
|
|
data = upper_32_bits(reg_list_mc_addr);
|
2013-10-24 02:27:10 +04:00
|
|
|
dst_ptr[reg_list_hdr_blk_index] = cpu_to_le32(data);
|
2013-04-12 21:52:52 +04:00
|
|
|
reg_list_hdr_blk_index++;
|
2013-08-13 20:48:06 +04:00
|
|
|
for (i = 0; cs_data[i].section != NULL; i++) {
|
|
|
|
for (j = 0; cs_data[i].section[j].extent != NULL; j++) {
|
|
|
|
reg_num = cs_data[i].section[j].reg_count;
|
|
|
|
data = reg_list_mc_addr & 0xffffffff;
|
2013-10-24 02:27:10 +04:00
|
|
|
dst_ptr[reg_list_hdr_blk_index] = cpu_to_le32(data);
|
2013-08-13 20:48:06 +04:00
|
|
|
reg_list_hdr_blk_index++;
|
|
|
|
|
|
|
|
data = (cs_data[i].section[j].reg_index * 4) & 0xffffffff;
|
2013-10-24 02:27:10 +04:00
|
|
|
dst_ptr[reg_list_hdr_blk_index] = cpu_to_le32(data);
|
2013-08-13 20:48:06 +04:00
|
|
|
reg_list_hdr_blk_index++;
|
|
|
|
|
|
|
|
data = 0x08000000 | (reg_num * 4);
|
2013-10-24 02:27:10 +04:00
|
|
|
dst_ptr[reg_list_hdr_blk_index] = cpu_to_le32(data);
|
2013-08-13 20:48:06 +04:00
|
|
|
reg_list_hdr_blk_index++;
|
|
|
|
|
|
|
|
for (k = 0; k < reg_num; k++) {
|
|
|
|
data = cs_data[i].section[j].extent[k];
|
2013-10-24 02:27:10 +04:00
|
|
|
dst_ptr[reg_list_blk_index + k] = cpu_to_le32(data);
|
2013-08-13 20:48:06 +04:00
|
|
|
}
|
|
|
|
reg_list_mc_addr += reg_num * 4;
|
|
|
|
reg_list_blk_index += reg_num;
|
2013-04-18 01:22:05 +04:00
|
|
|
}
|
2013-04-12 21:52:52 +04:00
|
|
|
}
|
2013-10-24 02:27:10 +04:00
|
|
|
dst_ptr[reg_list_hdr_blk_index] = cpu_to_le32(RLC_CLEAR_STATE_END_MARKER);
|
2013-04-12 21:52:52 +04:00
|
|
|
}
|
2013-04-18 01:22:05 +04:00
|
|
|
radeon_bo_kunmap(rdev->rlc.clear_state_obj);
|
|
|
|
radeon_bo_unreserve(rdev->rlc.clear_state_obj);
|
2013-04-12 21:52:52 +04:00
|
|
|
}
|
|
|
|
|
2013-07-23 17:41:05 +04:00
|
|
|
if (rdev->rlc.cp_table_size) {
|
|
|
|
if (rdev->rlc.cp_table_obj == NULL) {
|
2014-07-17 14:01:08 +04:00
|
|
|
r = radeon_bo_create(rdev, rdev->rlc.cp_table_size,
|
|
|
|
PAGE_SIZE, true,
|
|
|
|
RADEON_GEM_DOMAIN_VRAM, 0, NULL,
|
2014-09-18 16:11:56 +04:00
|
|
|
NULL, &rdev->rlc.cp_table_obj);
|
2013-07-23 17:41:05 +04:00
|
|
|
if (r) {
|
|
|
|
dev_warn(rdev->dev, "(%d) create RLC cp table bo failed\n", r);
|
|
|
|
sumo_rlc_fini(rdev);
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
r = radeon_bo_reserve(rdev->rlc.cp_table_obj, false);
|
|
|
|
if (unlikely(r != 0)) {
|
|
|
|
dev_warn(rdev->dev, "(%d) reserve RLC cp table bo failed\n", r);
|
|
|
|
sumo_rlc_fini(rdev);
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
r = radeon_bo_pin(rdev->rlc.cp_table_obj, RADEON_GEM_DOMAIN_VRAM,
|
|
|
|
&rdev->rlc.cp_table_gpu_addr);
|
|
|
|
if (r) {
|
|
|
|
radeon_bo_unreserve(rdev->rlc.cp_table_obj);
|
|
|
|
dev_warn(rdev->dev, "(%d) pin RLC cp_table bo failed\n", r);
|
|
|
|
sumo_rlc_fini(rdev);
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
r = radeon_bo_kmap(rdev->rlc.cp_table_obj, (void **)&rdev->rlc.cp_table_ptr);
|
|
|
|
if (r) {
|
|
|
|
dev_warn(rdev->dev, "(%d) map RLC cp table bo failed\n", r);
|
|
|
|
sumo_rlc_fini(rdev);
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
|
|
|
|
cik_init_cp_pg_table(rdev);
|
|
|
|
|
|
|
|
radeon_bo_kunmap(rdev->rlc.cp_table_obj);
|
|
|
|
radeon_bo_unreserve(rdev->rlc.cp_table_obj);
|
|
|
|
|
|
|
|
}
|
2013-04-12 21:52:52 +04:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void evergreen_rlc_start(struct radeon_device *rdev)
|
|
|
|
{
|
2013-02-16 01:26:33 +04:00
|
|
|
u32 mask = RLC_ENABLE;
|
|
|
|
|
|
|
|
if (rdev->flags & RADEON_IS_IGP) {
|
|
|
|
mask |= GFX_POWER_GATING_ENABLE | GFX_POWER_GATING_SRC;
|
|
|
|
}
|
|
|
|
|
|
|
|
WREG32(RLC_CNTL, mask);
|
2013-04-12 21:52:52 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
int evergreen_rlc_resume(struct radeon_device *rdev)
|
|
|
|
{
|
|
|
|
u32 i;
|
|
|
|
const __be32 *fw_data;
|
|
|
|
|
|
|
|
if (!rdev->rlc_fw)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
r600_rlc_stop(rdev);
|
|
|
|
|
|
|
|
WREG32(RLC_HB_CNTL, 0);
|
|
|
|
|
|
|
|
if (rdev->flags & RADEON_IS_IGP) {
|
2013-02-16 01:26:33 +04:00
|
|
|
if (rdev->family == CHIP_ARUBA) {
|
|
|
|
u32 always_on_bitmap =
|
|
|
|
3 | (3 << (16 * rdev->config.cayman.max_shader_engines));
|
|
|
|
/* find out the number of active simds */
|
|
|
|
u32 tmp = (RREG32(CC_GC_SHADER_PIPE_CONFIG) & 0xffff0000) >> 16;
|
|
|
|
tmp |= 0xffffffff << rdev->config.cayman.max_simds_per_se;
|
|
|
|
tmp = hweight32(~tmp);
|
|
|
|
if (tmp == rdev->config.cayman.max_simds_per_se) {
|
|
|
|
WREG32(TN_RLC_LB_ALWAYS_ACTIVE_SIMD_MASK, always_on_bitmap);
|
|
|
|
WREG32(TN_RLC_LB_PARAMS, 0x00601004);
|
|
|
|
WREG32(TN_RLC_LB_INIT_SIMD_MASK, 0xffffffff);
|
|
|
|
WREG32(TN_RLC_LB_CNTR_INIT, 0x00000000);
|
|
|
|
WREG32(TN_RLC_LB_CNTR_MAX, 0x00002000);
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
WREG32(RLC_HB_WPTR_LSB_ADDR, 0);
|
|
|
|
WREG32(RLC_HB_WPTR_MSB_ADDR, 0);
|
|
|
|
}
|
2013-04-12 21:52:52 +04:00
|
|
|
WREG32(TN_RLC_SAVE_AND_RESTORE_BASE, rdev->rlc.save_restore_gpu_addr >> 8);
|
|
|
|
WREG32(TN_RLC_CLEAR_STATE_RESTORE_BASE, rdev->rlc.clear_state_gpu_addr >> 8);
|
|
|
|
} else {
|
|
|
|
WREG32(RLC_HB_BASE, 0);
|
|
|
|
WREG32(RLC_HB_RPTR, 0);
|
|
|
|
WREG32(RLC_HB_WPTR, 0);
|
2013-02-16 01:26:33 +04:00
|
|
|
WREG32(RLC_HB_WPTR_LSB_ADDR, 0);
|
|
|
|
WREG32(RLC_HB_WPTR_MSB_ADDR, 0);
|
2013-04-12 21:52:52 +04:00
|
|
|
}
|
|
|
|
WREG32(RLC_MC_CNTL, 0);
|
|
|
|
WREG32(RLC_UCODE_CNTL, 0);
|
|
|
|
|
|
|
|
fw_data = (const __be32 *)rdev->rlc_fw->data;
|
|
|
|
if (rdev->family >= CHIP_ARUBA) {
|
|
|
|
for (i = 0; i < ARUBA_RLC_UCODE_SIZE; i++) {
|
|
|
|
WREG32(RLC_UCODE_ADDR, i);
|
|
|
|
WREG32(RLC_UCODE_DATA, be32_to_cpup(fw_data++));
|
|
|
|
}
|
|
|
|
} else if (rdev->family >= CHIP_CAYMAN) {
|
|
|
|
for (i = 0; i < CAYMAN_RLC_UCODE_SIZE; i++) {
|
|
|
|
WREG32(RLC_UCODE_ADDR, i);
|
|
|
|
WREG32(RLC_UCODE_DATA, be32_to_cpup(fw_data++));
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
for (i = 0; i < EVERGREEN_RLC_UCODE_SIZE; i++) {
|
|
|
|
WREG32(RLC_UCODE_ADDR, i);
|
|
|
|
WREG32(RLC_UCODE_DATA, be32_to_cpup(fw_data++));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
WREG32(RLC_UCODE_ADDR, 0);
|
|
|
|
|
|
|
|
evergreen_rlc_start(rdev);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2010-03-24 20:55:51 +03:00
|
|
|
/* Interrupts */
|
|
|
|
|
|
|
|
u32 evergreen_get_vblank_counter(struct radeon_device *rdev, int crtc)
|
|
|
|
{
|
2012-08-16 01:10:32 +04:00
|
|
|
if (crtc >= rdev->num_crtc)
|
2010-03-24 20:55:51 +03:00
|
|
|
return 0;
|
2012-08-16 01:10:32 +04:00
|
|
|
else
|
|
|
|
return RREG32(CRTC_STATUS_FRAME_COUNT + crtc_offsets[crtc]);
|
2010-03-24 20:55:51 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
void evergreen_disable_interrupt_state(struct radeon_device *rdev)
|
|
|
|
{
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
int i;
|
2010-03-24 20:55:51 +03:00
|
|
|
u32 tmp;
|
|
|
|
|
2011-11-18 05:13:28 +04:00
|
|
|
if (rdev->family >= CHIP_CAYMAN) {
|
|
|
|
cayman_cp_int_cntl_setup(rdev, 0,
|
|
|
|
CNTX_BUSY_INT_ENABLE | CNTX_EMPTY_INT_ENABLE);
|
|
|
|
cayman_cp_int_cntl_setup(rdev, 1, 0);
|
|
|
|
cayman_cp_int_cntl_setup(rdev, 2, 0);
|
2012-12-05 00:27:33 +04:00
|
|
|
tmp = RREG32(CAYMAN_DMA1_CNTL) & ~TRAP_ENABLE;
|
|
|
|
WREG32(CAYMAN_DMA1_CNTL, tmp);
|
2011-11-18 05:13:28 +04:00
|
|
|
} else
|
|
|
|
WREG32(CP_INT_CNTL, CNTX_BUSY_INT_ENABLE | CNTX_EMPTY_INT_ENABLE);
|
2012-12-05 00:25:59 +04:00
|
|
|
tmp = RREG32(DMA_CNTL) & ~TRAP_ENABLE;
|
|
|
|
WREG32(DMA_CNTL, tmp);
|
2010-03-24 20:55:51 +03:00
|
|
|
WREG32(GRBM_INT_CNTL, 0);
|
2015-02-18 15:19:28 +03:00
|
|
|
WREG32(SRBM_INT_CNTL, 0);
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
for (i = 0; i < rdev->num_crtc; i++)
|
|
|
|
WREG32(INT_MASK + crtc_offsets[i], 0);
|
2017-05-20 02:48:39 +03:00
|
|
|
for (i = 0; i < rdev->num_crtc; i++)
|
|
|
|
WREG32(GRPH_INT_CONTROL + crtc_offsets[i], 0);
|
2010-03-24 20:55:51 +03:00
|
|
|
|
2014-01-27 20:54:44 +04:00
|
|
|
/* only one DAC on DCE5 */
|
|
|
|
if (!ASIC_IS_DCE5(rdev))
|
2012-03-21 01:18:37 +04:00
|
|
|
WREG32(DACA_AUTODETECT_INT_CONTROL, 0);
|
2010-03-24 20:55:51 +03:00
|
|
|
WREG32(DACB_AUTODETECT_INT_CONTROL, 0);
|
|
|
|
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
for (i = 0; i < 6; i++)
|
|
|
|
WREG32_AND(DC_HPDx_INT_CONTROL(i), DC_HPDx_INT_POLARITY);
|
2010-03-24 20:55:51 +03:00
|
|
|
}
|
|
|
|
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
/* Note that the order we write back regs here is important */
|
2010-03-24 20:55:51 +03:00
|
|
|
int evergreen_irq_set(struct radeon_device *rdev)
|
|
|
|
{
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
int i;
|
2010-03-24 20:55:51 +03:00
|
|
|
u32 cp_int_cntl = CNTX_BUSY_INT_ENABLE | CNTX_EMPTY_INT_ENABLE;
|
2011-11-18 05:13:28 +04:00
|
|
|
u32 cp_int_cntl1 = 0, cp_int_cntl2 = 0;
|
2010-04-22 20:52:11 +04:00
|
|
|
u32 grbm_int_cntl = 0;
|
2012-12-05 00:27:33 +04:00
|
|
|
u32 dma_cntl, dma_cntl1 = 0;
|
2013-06-26 08:33:35 +04:00
|
|
|
u32 thermal_int = 0;
|
2010-03-24 20:55:51 +03:00
|
|
|
|
|
|
|
if (!rdev->irq.installed) {
|
2010-10-31 01:08:30 +04:00
|
|
|
WARN(1, "Can't enable IRQ/MSI because no handler is installed\n");
|
2010-03-24 20:55:51 +03:00
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
/* don't enable anything if the ih is disabled */
|
|
|
|
if (!rdev->ih.enabled) {
|
|
|
|
r600_disable_interrupts(rdev);
|
|
|
|
/* force the active interrupt state to all disabled */
|
|
|
|
evergreen_disable_interrupt_state(rdev);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2013-04-13 00:40:41 +04:00
|
|
|
if (rdev->family == CHIP_ARUBA)
|
|
|
|
thermal_int = RREG32(TN_CG_THERMAL_INT_CTRL) &
|
|
|
|
~(THERM_INT_MASK_HIGH | THERM_INT_MASK_LOW);
|
|
|
|
else
|
|
|
|
thermal_int = RREG32(CG_THERMAL_INT) &
|
|
|
|
~(THERM_INT_MASK_HIGH | THERM_INT_MASK_LOW);
|
2010-03-24 20:55:51 +03:00
|
|
|
|
2012-12-05 00:25:59 +04:00
|
|
|
dma_cntl = RREG32(DMA_CNTL) & ~TRAP_ENABLE;
|
|
|
|
|
2011-11-18 05:13:28 +04:00
|
|
|
if (rdev->family >= CHIP_CAYMAN) {
|
|
|
|
/* enable CP interrupts on all rings */
|
2012-05-17 21:52:00 +04:00
|
|
|
if (atomic_read(&rdev->irq.ring_int[RADEON_RING_TYPE_GFX_INDEX])) {
|
2011-11-18 05:13:28 +04:00
|
|
|
DRM_DEBUG("evergreen_irq_set: sw int gfx\n");
|
|
|
|
cp_int_cntl |= TIME_STAMP_INT_ENABLE;
|
|
|
|
}
|
2012-05-17 21:52:00 +04:00
|
|
|
if (atomic_read(&rdev->irq.ring_int[CAYMAN_RING_TYPE_CP1_INDEX])) {
|
2011-11-18 05:13:28 +04:00
|
|
|
DRM_DEBUG("evergreen_irq_set: sw int cp1\n");
|
|
|
|
cp_int_cntl1 |= TIME_STAMP_INT_ENABLE;
|
|
|
|
}
|
2012-05-17 21:52:00 +04:00
|
|
|
if (atomic_read(&rdev->irq.ring_int[CAYMAN_RING_TYPE_CP2_INDEX])) {
|
2011-11-18 05:13:28 +04:00
|
|
|
DRM_DEBUG("evergreen_irq_set: sw int cp2\n");
|
|
|
|
cp_int_cntl2 |= TIME_STAMP_INT_ENABLE;
|
|
|
|
}
|
|
|
|
} else {
|
2012-05-17 21:52:00 +04:00
|
|
|
if (atomic_read(&rdev->irq.ring_int[RADEON_RING_TYPE_GFX_INDEX])) {
|
2011-11-18 05:13:28 +04:00
|
|
|
DRM_DEBUG("evergreen_irq_set: sw int gfx\n");
|
|
|
|
cp_int_cntl |= RB_INT_ENABLE;
|
|
|
|
cp_int_cntl |= TIME_STAMP_INT_ENABLE;
|
|
|
|
}
|
2010-03-24 20:55:51 +03:00
|
|
|
}
|
2011-11-18 05:13:28 +04:00
|
|
|
|
2012-12-05 00:25:59 +04:00
|
|
|
if (atomic_read(&rdev->irq.ring_int[R600_RING_TYPE_DMA_INDEX])) {
|
|
|
|
DRM_DEBUG("r600_irq_set: sw int dma\n");
|
|
|
|
dma_cntl |= TRAP_ENABLE;
|
|
|
|
}
|
|
|
|
|
2012-12-05 00:27:33 +04:00
|
|
|
if (rdev->family >= CHIP_CAYMAN) {
|
|
|
|
dma_cntl1 = RREG32(CAYMAN_DMA1_CNTL) & ~TRAP_ENABLE;
|
|
|
|
if (atomic_read(&rdev->irq.ring_int[CAYMAN_RING_TYPE_DMA1_INDEX])) {
|
|
|
|
DRM_DEBUG("r600_irq_set: sw int dma1\n");
|
|
|
|
dma_cntl1 |= TRAP_ENABLE;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2013-06-26 08:33:35 +04:00
|
|
|
if (rdev->irq.dpm_thermal) {
|
|
|
|
DRM_DEBUG("dpm thermal\n");
|
|
|
|
thermal_int |= THERM_INT_MASK_HIGH | THERM_INT_MASK_LOW;
|
|
|
|
}
|
|
|
|
|
2011-11-18 05:13:28 +04:00
|
|
|
if (rdev->family >= CHIP_CAYMAN) {
|
|
|
|
cayman_cp_int_cntl_setup(rdev, 0, cp_int_cntl);
|
|
|
|
cayman_cp_int_cntl_setup(rdev, 1, cp_int_cntl1);
|
|
|
|
cayman_cp_int_cntl_setup(rdev, 2, cp_int_cntl2);
|
|
|
|
} else
|
|
|
|
WREG32(CP_INT_CNTL, cp_int_cntl);
|
2012-12-05 00:25:59 +04:00
|
|
|
|
|
|
|
WREG32(DMA_CNTL, dma_cntl);
|
|
|
|
|
2012-12-05 00:27:33 +04:00
|
|
|
if (rdev->family >= CHIP_CAYMAN)
|
|
|
|
WREG32(CAYMAN_DMA1_CNTL, dma_cntl1);
|
|
|
|
|
2010-04-22 20:52:11 +04:00
|
|
|
WREG32(GRBM_INT_CNTL, grbm_int_cntl);
|
2010-03-24 20:55:51 +03:00
|
|
|
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
for (i = 0; i < rdev->num_crtc; i++) {
|
|
|
|
radeon_irq_kms_set_irq_n_enabled(
|
|
|
|
rdev, INT_MASK + crtc_offsets[i],
|
|
|
|
VBLANK_INT_MASK,
|
|
|
|
rdev->irq.crtc_vblank_int[i] ||
|
|
|
|
atomic_read(&rdev->irq.pflip[i]), "vblank", i);
|
2010-11-23 01:56:28 +03:00
|
|
|
}
|
2010-03-24 20:55:51 +03:00
|
|
|
|
2017-05-20 02:48:39 +03:00
|
|
|
for (i = 0; i < rdev->num_crtc; i++)
|
|
|
|
WREG32(GRPH_INT_CONTROL + crtc_offsets[i], GRPH_PFLIP_INT_MASK);
|
2010-11-21 18:59:01 +03:00
|
|
|
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
for (i = 0; i < 6; i++) {
|
|
|
|
radeon_irq_kms_set_irq_n_enabled(
|
|
|
|
rdev, DC_HPDx_INT_CONTROL(i),
|
|
|
|
DC_HPDx_INT_EN | DC_HPDx_RX_INT_EN,
|
|
|
|
rdev->irq.hpd[i], "HPD", i);
|
|
|
|
}
|
|
|
|
|
2013-04-13 00:40:41 +04:00
|
|
|
if (rdev->family == CHIP_ARUBA)
|
|
|
|
WREG32(TN_CG_THERMAL_INT_CTRL, thermal_int);
|
|
|
|
else
|
|
|
|
WREG32(CG_THERMAL_INT, thermal_int);
|
2010-03-24 20:55:51 +03:00
|
|
|
|
2017-05-20 02:48:38 +03:00
|
|
|
for (i = 0; i < 6; i++) {
|
|
|
|
radeon_irq_kms_set_irq_n_enabled(
|
|
|
|
rdev, AFMT_AUDIO_PACKET_CONTROL + crtc_offsets[i],
|
|
|
|
AFMT_AZ_FORMAT_WTRIG_MASK,
|
|
|
|
rdev->irq.afmt[i], "HDMI", i);
|
|
|
|
}
|
2012-03-30 16:59:57 +04:00
|
|
|
|
2015-03-03 04:42:53 +03:00
|
|
|
/* posting read */
|
|
|
|
RREG32(SRBM_STATUS);
|
|
|
|
|
2010-01-13 01:54:34 +03:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
/* Note that the order we write back regs here is important */
|
2011-10-14 03:08:46 +04:00
|
|
|
static void evergreen_irq_ack(struct radeon_device *rdev)
|
2010-03-24 20:55:51 +03:00
|
|
|
{
|
2017-05-20 02:48:39 +03:00
|
|
|
int i, j;
|
|
|
|
u32 *grph_int = rdev->irq.stat_regs.evergreen.grph_int;
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
u32 *disp_int = rdev->irq.stat_regs.evergreen.disp_int;
|
2017-05-20 02:48:38 +03:00
|
|
|
u32 *afmt_status = rdev->irq.stat_regs.evergreen.afmt_status;
|
2010-03-24 20:55:51 +03:00
|
|
|
|
2017-05-20 02:48:38 +03:00
|
|
|
for (i = 0; i < 6; i++) {
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
disp_int[i] = RREG32(evergreen_disp_int_status[i]);
|
2017-05-20 02:48:38 +03:00
|
|
|
afmt_status[i] = RREG32(AFMT_STATUS + crtc_offsets[i]);
|
2017-05-20 02:48:39 +03:00
|
|
|
if (i < rdev->num_crtc)
|
|
|
|
grph_int[i] = RREG32(GRPH_INT_STATUS + crtc_offsets[i]);
|
2017-05-20 02:48:38 +03:00
|
|
|
}
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
|
2017-05-20 02:48:39 +03:00
|
|
|
/* We write back each interrupt register in pairs of two */
|
|
|
|
for (i = 0; i < rdev->num_crtc; i += 2) {
|
|
|
|
for (j = i; j < (i + 2); j++) {
|
|
|
|
if (grph_int[j] & GRPH_PFLIP_INT_OCCURRED)
|
|
|
|
WREG32(GRPH_INT_STATUS + crtc_offsets[j],
|
|
|
|
GRPH_PFLIP_INT_CLEAR);
|
|
|
|
}
|
2011-07-08 19:44:56 +04:00
|
|
|
|
2017-05-20 02:48:39 +03:00
|
|
|
for (j = i; j < (i + 2); j++) {
|
|
|
|
if (disp_int[j] & LB_D1_VBLANK_INTERRUPT)
|
|
|
|
WREG32(VBLANK_STATUS + crtc_offsets[j],
|
|
|
|
VBLANK_ACK);
|
|
|
|
if (disp_int[j] & LB_D1_VLINE_INTERRUPT)
|
|
|
|
WREG32(VLINE_STATUS + crtc_offsets[j],
|
|
|
|
VLINE_ACK);
|
|
|
|
}
|
2011-07-08 19:44:56 +04:00
|
|
|
}
|
2010-03-24 20:55:51 +03:00
|
|
|
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
for (i = 0; i < 6; i++) {
|
|
|
|
if (disp_int[i] & DC_HPD1_INTERRUPT)
|
|
|
|
WREG32_OR(DC_HPDx_INT_CONTROL(i), DC_HPDx_INT_ACK);
|
2010-03-24 20:55:51 +03:00
|
|
|
}
|
2015-02-24 02:23:57 +03:00
|
|
|
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
for (i = 0; i < 6; i++) {
|
|
|
|
if (disp_int[i] & DC_HPD1_RX_INTERRUPT)
|
|
|
|
WREG32_OR(DC_HPDx_INT_CONTROL(i), DC_HPDx_RX_INT_ACK);
|
2015-02-24 02:23:57 +03:00
|
|
|
}
|
|
|
|
|
2017-05-20 02:48:38 +03:00
|
|
|
for (i = 0; i < 6; i++) {
|
|
|
|
if (afmt_status[i] & AFMT_AZ_FORMAT_WTRIG)
|
|
|
|
WREG32_OR(AFMT_AUDIO_PACKET_CONTROL + crtc_offsets[i],
|
|
|
|
AFMT_AZ_FORMAT_WTRIG_ACK);
|
2012-03-30 16:59:57 +04:00
|
|
|
}
|
2010-03-24 20:55:51 +03:00
|
|
|
}
|
|
|
|
|
2012-08-31 21:43:50 +04:00
|
|
|
static void evergreen_irq_disable(struct radeon_device *rdev)
|
2010-03-24 20:55:51 +03:00
|
|
|
{
|
|
|
|
r600_disable_interrupts(rdev);
|
|
|
|
/* Wait and acknowledge irq */
|
|
|
|
mdelay(1);
|
2010-11-21 18:59:01 +03:00
|
|
|
evergreen_irq_ack(rdev);
|
2010-03-24 20:55:51 +03:00
|
|
|
evergreen_disable_interrupt_state(rdev);
|
|
|
|
}
|
|
|
|
|
2011-03-03 04:07:34 +03:00
|
|
|
void evergreen_irq_suspend(struct radeon_device *rdev)
|
2010-03-24 20:55:51 +03:00
|
|
|
{
|
|
|
|
evergreen_irq_disable(rdev);
|
|
|
|
r600_rlc_stop(rdev);
|
|
|
|
}
|
|
|
|
|
2011-10-14 03:08:46 +04:00
|
|
|
static u32 evergreen_get_ih_wptr(struct radeon_device *rdev)
|
2010-03-24 20:55:51 +03:00
|
|
|
{
|
|
|
|
u32 wptr, tmp;
|
|
|
|
|
2010-08-28 02:25:25 +04:00
|
|
|
if (rdev->wb.enabled)
|
2011-04-19 19:07:13 +04:00
|
|
|
wptr = le32_to_cpu(rdev->wb.wb[R600_WB_IH_WPTR_OFFSET/4]);
|
2010-08-28 02:25:25 +04:00
|
|
|
else
|
|
|
|
wptr = RREG32(IH_RB_WPTR);
|
2010-03-24 20:55:51 +03:00
|
|
|
|
|
|
|
if (wptr & RB_OVERFLOW) {
|
2014-09-19 07:07:11 +04:00
|
|
|
wptr &= ~RB_OVERFLOW;
|
2010-03-24 20:55:51 +03:00
|
|
|
/* When a ring buffer overflow happen start parsing interrupt
|
|
|
|
* from the last not overwritten vector (wptr + 16). Hopefully
|
|
|
|
* this should allow us to catchup.
|
|
|
|
*/
|
2014-09-19 07:22:07 +04:00
|
|
|
dev_warn(rdev->dev, "IH ring buffer overflow (0x%08X, 0x%08X, 0x%08X)\n",
|
|
|
|
wptr, rdev->ih.rptr, (wptr + 16) & rdev->ih.ptr_mask);
|
2010-03-24 20:55:51 +03:00
|
|
|
rdev->ih.rptr = (wptr + 16) & rdev->ih.ptr_mask;
|
|
|
|
tmp = RREG32(IH_RB_CNTL);
|
|
|
|
tmp |= IH_WPTR_OVERFLOW_CLEAR;
|
|
|
|
WREG32(IH_RB_CNTL, tmp);
|
|
|
|
}
|
|
|
|
return (wptr & rdev->ih.ptr_mask);
|
|
|
|
}
|
|
|
|
|
|
|
|
int evergreen_irq_process(struct radeon_device *rdev)
|
|
|
|
{
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
u32 *disp_int = rdev->irq.stat_regs.evergreen.disp_int;
|
2017-05-20 02:48:38 +03:00
|
|
|
u32 *afmt_status = rdev->irq.stat_regs.evergreen.afmt_status;
|
|
|
|
u32 crtc_idx, hpd_idx, afmt_idx;
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
u32 mask;
|
2011-06-18 07:59:51 +04:00
|
|
|
u32 wptr;
|
|
|
|
u32 rptr;
|
2010-03-24 20:55:51 +03:00
|
|
|
u32 src_id, src_data;
|
|
|
|
u32 ring_index;
|
|
|
|
bool queue_hotplug = false;
|
2012-03-30 16:59:57 +04:00
|
|
|
bool queue_hdmi = false;
|
2015-02-24 02:23:57 +03:00
|
|
|
bool queue_dp = false;
|
2013-06-26 08:33:35 +04:00
|
|
|
bool queue_thermal = false;
|
2013-06-14 02:26:25 +04:00
|
|
|
u32 status, addr;
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
const char *event_name;
|
2010-03-24 20:55:51 +03:00
|
|
|
|
2011-06-18 07:59:51 +04:00
|
|
|
if (!rdev->ih.enabled || rdev->shutdown)
|
2010-03-24 20:55:51 +03:00
|
|
|
return IRQ_NONE;
|
|
|
|
|
2011-06-18 07:59:51 +04:00
|
|
|
wptr = evergreen_get_ih_wptr(rdev);
|
2012-05-16 23:45:24 +04:00
|
|
|
|
|
|
|
restart_ih:
|
|
|
|
/* is somebody else already processing irqs? */
|
|
|
|
if (atomic_xchg(&rdev->ih.lock, 1))
|
|
|
|
return IRQ_NONE;
|
|
|
|
|
2011-06-18 07:59:51 +04:00
|
|
|
rptr = rdev->ih.rptr;
|
drm/radeon: Handle irqs only based on irq ring, not irq status regs.
Trying to resolve issues with missed vblanks and impossible
values inside delivered kms pageflip completion events showed
that radeon's irq handling sometimes doesn't handle valid irqs,
but silently skips them. This was observed for vblank interrupts.
Although those irqs have corresponding events queued in the gpu's
irq ring at time of interrupt, and therefore the corresponding
handling code gets triggered by these events, the handling code
sometimes silently skipped processing the irq. The reason for those
skips is that the handling code double-checks for each irq event if
the corresponding irq status bits in the irq status registers
are set. Sometimes those bits are not set at time of check
for valid irqs, maybe due to some hardware race on some setups?
The problem only seems to happen on some machine + card combos
sometimes, e.g., never happened during my testing of different PC
cards of the DCE-2/3/4 generation a year ago, but happens consistently
now on two different Apple Mac cards (RV730, DCE-3, Apple iMac and
Evergreen JUNIPER, DCE-4 in a Apple MacPro). It also doesn't happen
at each interrupt but only occassionally every couple of
hundred or thousand vblank interrupts.
This results in XOrg warning messages like
"[ 7084.472] (WW) RADEON(0): radeon_dri2_flip_event_handler:
Pageflip completion event has impossible msc 420120 < target_msc 420121"
as well as skipped frames and problems for applications that
use kms pageflip events or vblank events, e.g., users of DRI2 and
DRI3/Present, Waylands Weston compositor, etc. See also
https://bugs.freedesktop.org/show_bug.cgi?id=85203
After some talking to Alex and Michel, we decided to fix this
by turning the double-check for asserted irq status bits into a
warning. Whenever a irq event is queued in the IH ring, always
execute the corresponding interrupt handler. Still check the irq
status bits, but only to log a DRM_DEBUG message on a mismatch.
This fixed the problems reliably on both previously failing
cards, RV-730 dual-head tested on both crtcs (pipes D1 and D2)
and a triple-output Juniper HD-5770 card tested on all three
available crtcs (D1/D2/D3). The r600 and evergreen irq handling
is therefore tested, but the cik an si handling is only compile
tested due to lack of hw.
Reviewed-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Mario Kleiner <mario.kleiner.de@gmail.com>
CC: Michel Dänzer <michel.daenzer@amd.com>
CC: Alex Deucher <alexander.deucher@amd.com>
CC: <stable@vger.kernel.org> # v3.16+
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2015-07-03 07:03:06 +03:00
|
|
|
DRM_DEBUG("evergreen_irq_process start: rptr %d, wptr %d\n", rptr, wptr);
|
2010-03-24 20:55:51 +03:00
|
|
|
|
2011-07-13 10:28:19 +04:00
|
|
|
/* Order reading of wptr vs. reading of IH ring data */
|
|
|
|
rmb();
|
|
|
|
|
2010-03-24 20:55:51 +03:00
|
|
|
/* display interrupts */
|
2010-11-21 18:59:01 +03:00
|
|
|
evergreen_irq_ack(rdev);
|
2010-03-24 20:55:51 +03:00
|
|
|
|
|
|
|
while (rptr != wptr) {
|
|
|
|
/* wptr/rptr are in bytes! */
|
|
|
|
ring_index = rptr / 4;
|
2011-02-14 03:06:33 +03:00
|
|
|
src_id = le32_to_cpu(rdev->ih.ring[ring_index]) & 0xff;
|
|
|
|
src_data = le32_to_cpu(rdev->ih.ring[ring_index + 1]) & 0xfffffff;
|
2010-03-24 20:55:51 +03:00
|
|
|
|
|
|
|
switch (src_id) {
|
|
|
|
case 1: /* D1 vblank/vline */
|
|
|
|
case 2: /* D2 vblank/vline */
|
|
|
|
case 3: /* D3 vblank/vline */
|
|
|
|
case 4: /* D4 vblank/vline */
|
|
|
|
case 5: /* D5 vblank/vline */
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
case 6: /* D6 vblank/vline */
|
|
|
|
crtc_idx = src_id - 1;
|
|
|
|
|
|
|
|
if (src_data == 0) { /* vblank */
|
|
|
|
mask = LB_D1_VBLANK_INTERRUPT;
|
|
|
|
event_name = "vblank";
|
drm/radeon: Handle irqs only based on irq ring, not irq status regs.
Trying to resolve issues with missed vblanks and impossible
values inside delivered kms pageflip completion events showed
that radeon's irq handling sometimes doesn't handle valid irqs,
but silently skips them. This was observed for vblank interrupts.
Although those irqs have corresponding events queued in the gpu's
irq ring at time of interrupt, and therefore the corresponding
handling code gets triggered by these events, the handling code
sometimes silently skipped processing the irq. The reason for those
skips is that the handling code double-checks for each irq event if
the corresponding irq status bits in the irq status registers
are set. Sometimes those bits are not set at time of check
for valid irqs, maybe due to some hardware race on some setups?
The problem only seems to happen on some machine + card combos
sometimes, e.g., never happened during my testing of different PC
cards of the DCE-2/3/4 generation a year ago, but happens consistently
now on two different Apple Mac cards (RV730, DCE-3, Apple iMac and
Evergreen JUNIPER, DCE-4 in a Apple MacPro). It also doesn't happen
at each interrupt but only occassionally every couple of
hundred or thousand vblank interrupts.
This results in XOrg warning messages like
"[ 7084.472] (WW) RADEON(0): radeon_dri2_flip_event_handler:
Pageflip completion event has impossible msc 420120 < target_msc 420121"
as well as skipped frames and problems for applications that
use kms pageflip events or vblank events, e.g., users of DRI2 and
DRI3/Present, Waylands Weston compositor, etc. See also
https://bugs.freedesktop.org/show_bug.cgi?id=85203
After some talking to Alex and Michel, we decided to fix this
by turning the double-check for asserted irq status bits into a
warning. Whenever a irq event is queued in the IH ring, always
execute the corresponding interrupt handler. Still check the irq
status bits, but only to log a DRM_DEBUG message on a mismatch.
This fixed the problems reliably on both previously failing
cards, RV-730 dual-head tested on both crtcs (pipes D1 and D2)
and a triple-output Juniper HD-5770 card tested on all three
available crtcs (D1/D2/D3). The r600 and evergreen irq handling
is therefore tested, but the cik an si handling is only compile
tested due to lack of hw.
Reviewed-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Mario Kleiner <mario.kleiner.de@gmail.com>
CC: Michel Dänzer <michel.daenzer@amd.com>
CC: Alex Deucher <alexander.deucher@amd.com>
CC: <stable@vger.kernel.org> # v3.16+
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2015-07-03 07:03:06 +03:00
|
|
|
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
if (rdev->irq.crtc_vblank_int[crtc_idx]) {
|
|
|
|
drm_handle_vblank(rdev->ddev, crtc_idx);
|
drm/radeon: Handle irqs only based on irq ring, not irq status regs.
Trying to resolve issues with missed vblanks and impossible
values inside delivered kms pageflip completion events showed
that radeon's irq handling sometimes doesn't handle valid irqs,
but silently skips them. This was observed for vblank interrupts.
Although those irqs have corresponding events queued in the gpu's
irq ring at time of interrupt, and therefore the corresponding
handling code gets triggered by these events, the handling code
sometimes silently skipped processing the irq. The reason for those
skips is that the handling code double-checks for each irq event if
the corresponding irq status bits in the irq status registers
are set. Sometimes those bits are not set at time of check
for valid irqs, maybe due to some hardware race on some setups?
The problem only seems to happen on some machine + card combos
sometimes, e.g., never happened during my testing of different PC
cards of the DCE-2/3/4 generation a year ago, but happens consistently
now on two different Apple Mac cards (RV730, DCE-3, Apple iMac and
Evergreen JUNIPER, DCE-4 in a Apple MacPro). It also doesn't happen
at each interrupt but only occassionally every couple of
hundred or thousand vblank interrupts.
This results in XOrg warning messages like
"[ 7084.472] (WW) RADEON(0): radeon_dri2_flip_event_handler:
Pageflip completion event has impossible msc 420120 < target_msc 420121"
as well as skipped frames and problems for applications that
use kms pageflip events or vblank events, e.g., users of DRI2 and
DRI3/Present, Waylands Weston compositor, etc. See also
https://bugs.freedesktop.org/show_bug.cgi?id=85203
After some talking to Alex and Michel, we decided to fix this
by turning the double-check for asserted irq status bits into a
warning. Whenever a irq event is queued in the IH ring, always
execute the corresponding interrupt handler. Still check the irq
status bits, but only to log a DRM_DEBUG message on a mismatch.
This fixed the problems reliably on both previously failing
cards, RV-730 dual-head tested on both crtcs (pipes D1 and D2)
and a triple-output Juniper HD-5770 card tested on all three
available crtcs (D1/D2/D3). The r600 and evergreen irq handling
is therefore tested, but the cik an si handling is only compile
tested due to lack of hw.
Reviewed-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Mario Kleiner <mario.kleiner.de@gmail.com>
CC: Michel Dänzer <michel.daenzer@amd.com>
CC: Alex Deucher <alexander.deucher@amd.com>
CC: <stable@vger.kernel.org> # v3.16+
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2015-07-03 07:03:06 +03:00
|
|
|
rdev->pm.vblank_sync = true;
|
|
|
|
wake_up(&rdev->irq.vblank_queue);
|
2010-03-24 20:55:51 +03:00
|
|
|
}
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
if (atomic_read(&rdev->irq.pflip[crtc_idx])) {
|
|
|
|
radeon_crtc_handle_vblank(rdev,
|
|
|
|
crtc_idx);
|
|
|
|
}
|
drm/radeon: Handle irqs only based on irq ring, not irq status regs.
Trying to resolve issues with missed vblanks and impossible
values inside delivered kms pageflip completion events showed
that radeon's irq handling sometimes doesn't handle valid irqs,
but silently skips them. This was observed for vblank interrupts.
Although those irqs have corresponding events queued in the gpu's
irq ring at time of interrupt, and therefore the corresponding
handling code gets triggered by these events, the handling code
sometimes silently skipped processing the irq. The reason for those
skips is that the handling code double-checks for each irq event if
the corresponding irq status bits in the irq status registers
are set. Sometimes those bits are not set at time of check
for valid irqs, maybe due to some hardware race on some setups?
The problem only seems to happen on some machine + card combos
sometimes, e.g., never happened during my testing of different PC
cards of the DCE-2/3/4 generation a year ago, but happens consistently
now on two different Apple Mac cards (RV730, DCE-3, Apple iMac and
Evergreen JUNIPER, DCE-4 in a Apple MacPro). It also doesn't happen
at each interrupt but only occassionally every couple of
hundred or thousand vblank interrupts.
This results in XOrg warning messages like
"[ 7084.472] (WW) RADEON(0): radeon_dri2_flip_event_handler:
Pageflip completion event has impossible msc 420120 < target_msc 420121"
as well as skipped frames and problems for applications that
use kms pageflip events or vblank events, e.g., users of DRI2 and
DRI3/Present, Waylands Weston compositor, etc. See also
https://bugs.freedesktop.org/show_bug.cgi?id=85203
After some talking to Alex and Michel, we decided to fix this
by turning the double-check for asserted irq status bits into a
warning. Whenever a irq event is queued in the IH ring, always
execute the corresponding interrupt handler. Still check the irq
status bits, but only to log a DRM_DEBUG message on a mismatch.
This fixed the problems reliably on both previously failing
cards, RV-730 dual-head tested on both crtcs (pipes D1 and D2)
and a triple-output Juniper HD-5770 card tested on all three
available crtcs (D1/D2/D3). The r600 and evergreen irq handling
is therefore tested, but the cik an si handling is only compile
tested due to lack of hw.
Reviewed-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Mario Kleiner <mario.kleiner.de@gmail.com>
CC: Michel Dänzer <michel.daenzer@amd.com>
CC: Alex Deucher <alexander.deucher@amd.com>
CC: <stable@vger.kernel.org> # v3.16+
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2015-07-03 07:03:06 +03:00
|
|
|
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
} else if (src_data == 1) { /* vline */
|
|
|
|
mask = LB_D1_VLINE_INTERRUPT;
|
|
|
|
event_name = "vline";
|
|
|
|
} else {
|
|
|
|
DRM_DEBUG("Unhandled interrupt: %d %d\n",
|
|
|
|
src_id, src_data);
|
2010-03-24 20:55:51 +03:00
|
|
|
break;
|
|
|
|
}
|
drm/radeon: Handle irqs only based on irq ring, not irq status regs.
Trying to resolve issues with missed vblanks and impossible
values inside delivered kms pageflip completion events showed
that radeon's irq handling sometimes doesn't handle valid irqs,
but silently skips them. This was observed for vblank interrupts.
Although those irqs have corresponding events queued in the gpu's
irq ring at time of interrupt, and therefore the corresponding
handling code gets triggered by these events, the handling code
sometimes silently skipped processing the irq. The reason for those
skips is that the handling code double-checks for each irq event if
the corresponding irq status bits in the irq status registers
are set. Sometimes those bits are not set at time of check
for valid irqs, maybe due to some hardware race on some setups?
The problem only seems to happen on some machine + card combos
sometimes, e.g., never happened during my testing of different PC
cards of the DCE-2/3/4 generation a year ago, but happens consistently
now on two different Apple Mac cards (RV730, DCE-3, Apple iMac and
Evergreen JUNIPER, DCE-4 in a Apple MacPro). It also doesn't happen
at each interrupt but only occassionally every couple of
hundred or thousand vblank interrupts.
This results in XOrg warning messages like
"[ 7084.472] (WW) RADEON(0): radeon_dri2_flip_event_handler:
Pageflip completion event has impossible msc 420120 < target_msc 420121"
as well as skipped frames and problems for applications that
use kms pageflip events or vblank events, e.g., users of DRI2 and
DRI3/Present, Waylands Weston compositor, etc. See also
https://bugs.freedesktop.org/show_bug.cgi?id=85203
After some talking to Alex and Michel, we decided to fix this
by turning the double-check for asserted irq status bits into a
warning. Whenever a irq event is queued in the IH ring, always
execute the corresponding interrupt handler. Still check the irq
status bits, but only to log a DRM_DEBUG message on a mismatch.
This fixed the problems reliably on both previously failing
cards, RV-730 dual-head tested on both crtcs (pipes D1 and D2)
and a triple-output Juniper HD-5770 card tested on all three
available crtcs (D1/D2/D3). The r600 and evergreen irq handling
is therefore tested, but the cik an si handling is only compile
tested due to lack of hw.
Reviewed-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Mario Kleiner <mario.kleiner.de@gmail.com>
CC: Michel Dänzer <michel.daenzer@amd.com>
CC: Alex Deucher <alexander.deucher@amd.com>
CC: <stable@vger.kernel.org> # v3.16+
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2015-07-03 07:03:06 +03:00
|
|
|
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
if (!(disp_int[crtc_idx] & mask)) {
|
|
|
|
DRM_DEBUG("IH: D%d %s - IH event w/o asserted irq bit?\n",
|
|
|
|
crtc_idx + 1, event_name);
|
|
|
|
}
|
drm/radeon: Handle irqs only based on irq ring, not irq status regs.
Trying to resolve issues with missed vblanks and impossible
values inside delivered kms pageflip completion events showed
that radeon's irq handling sometimes doesn't handle valid irqs,
but silently skips them. This was observed for vblank interrupts.
Although those irqs have corresponding events queued in the gpu's
irq ring at time of interrupt, and therefore the corresponding
handling code gets triggered by these events, the handling code
sometimes silently skipped processing the irq. The reason for those
skips is that the handling code double-checks for each irq event if
the corresponding irq status bits in the irq status registers
are set. Sometimes those bits are not set at time of check
for valid irqs, maybe due to some hardware race on some setups?
The problem only seems to happen on some machine + card combos
sometimes, e.g., never happened during my testing of different PC
cards of the DCE-2/3/4 generation a year ago, but happens consistently
now on two different Apple Mac cards (RV730, DCE-3, Apple iMac and
Evergreen JUNIPER, DCE-4 in a Apple MacPro). It also doesn't happen
at each interrupt but only occassionally every couple of
hundred or thousand vblank interrupts.
This results in XOrg warning messages like
"[ 7084.472] (WW) RADEON(0): radeon_dri2_flip_event_handler:
Pageflip completion event has impossible msc 420120 < target_msc 420121"
as well as skipped frames and problems for applications that
use kms pageflip events or vblank events, e.g., users of DRI2 and
DRI3/Present, Waylands Weston compositor, etc. See also
https://bugs.freedesktop.org/show_bug.cgi?id=85203
After some talking to Alex and Michel, we decided to fix this
by turning the double-check for asserted irq status bits into a
warning. Whenever a irq event is queued in the IH ring, always
execute the corresponding interrupt handler. Still check the irq
status bits, but only to log a DRM_DEBUG message on a mismatch.
This fixed the problems reliably on both previously failing
cards, RV-730 dual-head tested on both crtcs (pipes D1 and D2)
and a triple-output Juniper HD-5770 card tested on all three
available crtcs (D1/D2/D3). The r600 and evergreen irq handling
is therefore tested, but the cik an si handling is only compile
tested due to lack of hw.
Reviewed-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Mario Kleiner <mario.kleiner.de@gmail.com>
CC: Michel Dänzer <michel.daenzer@amd.com>
CC: Alex Deucher <alexander.deucher@amd.com>
CC: <stable@vger.kernel.org> # v3.16+
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2015-07-03 07:03:06 +03:00
|
|
|
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
disp_int[crtc_idx] &= ~mask;
|
|
|
|
DRM_DEBUG("IH: D%d %s\n", crtc_idx + 1, event_name);
|
drm/radeon: Handle irqs only based on irq ring, not irq status regs.
Trying to resolve issues with missed vblanks and impossible
values inside delivered kms pageflip completion events showed
that radeon's irq handling sometimes doesn't handle valid irqs,
but silently skips them. This was observed for vblank interrupts.
Although those irqs have corresponding events queued in the gpu's
irq ring at time of interrupt, and therefore the corresponding
handling code gets triggered by these events, the handling code
sometimes silently skipped processing the irq. The reason for those
skips is that the handling code double-checks for each irq event if
the corresponding irq status bits in the irq status registers
are set. Sometimes those bits are not set at time of check
for valid irqs, maybe due to some hardware race on some setups?
The problem only seems to happen on some machine + card combos
sometimes, e.g., never happened during my testing of different PC
cards of the DCE-2/3/4 generation a year ago, but happens consistently
now on two different Apple Mac cards (RV730, DCE-3, Apple iMac and
Evergreen JUNIPER, DCE-4 in a Apple MacPro). It also doesn't happen
at each interrupt but only occassionally every couple of
hundred or thousand vblank interrupts.
This results in XOrg warning messages like
"[ 7084.472] (WW) RADEON(0): radeon_dri2_flip_event_handler:
Pageflip completion event has impossible msc 420120 < target_msc 420121"
as well as skipped frames and problems for applications that
use kms pageflip events or vblank events, e.g., users of DRI2 and
DRI3/Present, Waylands Weston compositor, etc. See also
https://bugs.freedesktop.org/show_bug.cgi?id=85203
After some talking to Alex and Michel, we decided to fix this
by turning the double-check for asserted irq status bits into a
warning. Whenever a irq event is queued in the IH ring, always
execute the corresponding interrupt handler. Still check the irq
status bits, but only to log a DRM_DEBUG message on a mismatch.
This fixed the problems reliably on both previously failing
cards, RV-730 dual-head tested on both crtcs (pipes D1 and D2)
and a triple-output Juniper HD-5770 card tested on all three
available crtcs (D1/D2/D3). The r600 and evergreen irq handling
is therefore tested, but the cik an si handling is only compile
tested due to lack of hw.
Reviewed-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Mario Kleiner <mario.kleiner.de@gmail.com>
CC: Michel Dänzer <michel.daenzer@amd.com>
CC: Alex Deucher <alexander.deucher@amd.com>
CC: <stable@vger.kernel.org> # v3.16+
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2015-07-03 07:03:06 +03:00
|
|
|
|
2010-03-24 20:55:51 +03:00
|
|
|
break;
|
2014-04-23 22:46:06 +04:00
|
|
|
case 8: /* D1 page flip */
|
|
|
|
case 10: /* D2 page flip */
|
|
|
|
case 12: /* D3 page flip */
|
|
|
|
case 14: /* D4 page flip */
|
|
|
|
case 16: /* D5 page flip */
|
|
|
|
case 18: /* D6 page flip */
|
|
|
|
DRM_DEBUG("IH: D%d flip\n", ((src_id - 8) >> 1) + 1);
|
2014-07-29 08:21:44 +04:00
|
|
|
if (radeon_use_pflipirq > 0)
|
|
|
|
radeon_crtc_handle_flip(rdev, (src_id - 8) >> 1);
|
2014-04-23 22:46:06 +04:00
|
|
|
break;
|
2010-03-24 20:55:51 +03:00
|
|
|
case 42: /* HPD hotplug */
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
if (src_data <= 5) {
|
|
|
|
hpd_idx = src_data;
|
|
|
|
mask = DC_HPD1_INTERRUPT;
|
drm/radeon: Handle irqs only based on irq ring, not irq status regs.
Trying to resolve issues with missed vblanks and impossible
values inside delivered kms pageflip completion events showed
that radeon's irq handling sometimes doesn't handle valid irqs,
but silently skips them. This was observed for vblank interrupts.
Although those irqs have corresponding events queued in the gpu's
irq ring at time of interrupt, and therefore the corresponding
handling code gets triggered by these events, the handling code
sometimes silently skipped processing the irq. The reason for those
skips is that the handling code double-checks for each irq event if
the corresponding irq status bits in the irq status registers
are set. Sometimes those bits are not set at time of check
for valid irqs, maybe due to some hardware race on some setups?
The problem only seems to happen on some machine + card combos
sometimes, e.g., never happened during my testing of different PC
cards of the DCE-2/3/4 generation a year ago, but happens consistently
now on two different Apple Mac cards (RV730, DCE-3, Apple iMac and
Evergreen JUNIPER, DCE-4 in a Apple MacPro). It also doesn't happen
at each interrupt but only occassionally every couple of
hundred or thousand vblank interrupts.
This results in XOrg warning messages like
"[ 7084.472] (WW) RADEON(0): radeon_dri2_flip_event_handler:
Pageflip completion event has impossible msc 420120 < target_msc 420121"
as well as skipped frames and problems for applications that
use kms pageflip events or vblank events, e.g., users of DRI2 and
DRI3/Present, Waylands Weston compositor, etc. See also
https://bugs.freedesktop.org/show_bug.cgi?id=85203
After some talking to Alex and Michel, we decided to fix this
by turning the double-check for asserted irq status bits into a
warning. Whenever a irq event is queued in the IH ring, always
execute the corresponding interrupt handler. Still check the irq
status bits, but only to log a DRM_DEBUG message on a mismatch.
This fixed the problems reliably on both previously failing
cards, RV-730 dual-head tested on both crtcs (pipes D1 and D2)
and a triple-output Juniper HD-5770 card tested on all three
available crtcs (D1/D2/D3). The r600 and evergreen irq handling
is therefore tested, but the cik an si handling is only compile
tested due to lack of hw.
Reviewed-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Mario Kleiner <mario.kleiner.de@gmail.com>
CC: Michel Dänzer <michel.daenzer@amd.com>
CC: Alex Deucher <alexander.deucher@amd.com>
CC: <stable@vger.kernel.org> # v3.16+
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2015-07-03 07:03:06 +03:00
|
|
|
queue_hotplug = true;
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
event_name = "HPD";
|
drm/radeon: Handle irqs only based on irq ring, not irq status regs.
Trying to resolve issues with missed vblanks and impossible
values inside delivered kms pageflip completion events showed
that radeon's irq handling sometimes doesn't handle valid irqs,
but silently skips them. This was observed for vblank interrupts.
Although those irqs have corresponding events queued in the gpu's
irq ring at time of interrupt, and therefore the corresponding
handling code gets triggered by these events, the handling code
sometimes silently skipped processing the irq. The reason for those
skips is that the handling code double-checks for each irq event if
the corresponding irq status bits in the irq status registers
are set. Sometimes those bits are not set at time of check
for valid irqs, maybe due to some hardware race on some setups?
The problem only seems to happen on some machine + card combos
sometimes, e.g., never happened during my testing of different PC
cards of the DCE-2/3/4 generation a year ago, but happens consistently
now on two different Apple Mac cards (RV730, DCE-3, Apple iMac and
Evergreen JUNIPER, DCE-4 in a Apple MacPro). It also doesn't happen
at each interrupt but only occassionally every couple of
hundred or thousand vblank interrupts.
This results in XOrg warning messages like
"[ 7084.472] (WW) RADEON(0): radeon_dri2_flip_event_handler:
Pageflip completion event has impossible msc 420120 < target_msc 420121"
as well as skipped frames and problems for applications that
use kms pageflip events or vblank events, e.g., users of DRI2 and
DRI3/Present, Waylands Weston compositor, etc. See also
https://bugs.freedesktop.org/show_bug.cgi?id=85203
After some talking to Alex and Michel, we decided to fix this
by turning the double-check for asserted irq status bits into a
warning. Whenever a irq event is queued in the IH ring, always
execute the corresponding interrupt handler. Still check the irq
status bits, but only to log a DRM_DEBUG message on a mismatch.
This fixed the problems reliably on both previously failing
cards, RV-730 dual-head tested on both crtcs (pipes D1 and D2)
and a triple-output Juniper HD-5770 card tested on all three
available crtcs (D1/D2/D3). The r600 and evergreen irq handling
is therefore tested, but the cik an si handling is only compile
tested due to lack of hw.
Reviewed-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Mario Kleiner <mario.kleiner.de@gmail.com>
CC: Michel Dänzer <michel.daenzer@amd.com>
CC: Alex Deucher <alexander.deucher@amd.com>
CC: <stable@vger.kernel.org> # v3.16+
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2015-07-03 07:03:06 +03:00
|
|
|
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
} else if (src_data <= 11) {
|
|
|
|
hpd_idx = src_data - 6;
|
|
|
|
mask = DC_HPD1_RX_INTERRUPT;
|
drm/radeon: Handle irqs only based on irq ring, not irq status regs.
Trying to resolve issues with missed vblanks and impossible
values inside delivered kms pageflip completion events showed
that radeon's irq handling sometimes doesn't handle valid irqs,
but silently skips them. This was observed for vblank interrupts.
Although those irqs have corresponding events queued in the gpu's
irq ring at time of interrupt, and therefore the corresponding
handling code gets triggered by these events, the handling code
sometimes silently skipped processing the irq. The reason for those
skips is that the handling code double-checks for each irq event if
the corresponding irq status bits in the irq status registers
are set. Sometimes those bits are not set at time of check
for valid irqs, maybe due to some hardware race on some setups?
The problem only seems to happen on some machine + card combos
sometimes, e.g., never happened during my testing of different PC
cards of the DCE-2/3/4 generation a year ago, but happens consistently
now on two different Apple Mac cards (RV730, DCE-3, Apple iMac and
Evergreen JUNIPER, DCE-4 in a Apple MacPro). It also doesn't happen
at each interrupt but only occassionally every couple of
hundred or thousand vblank interrupts.
This results in XOrg warning messages like
"[ 7084.472] (WW) RADEON(0): radeon_dri2_flip_event_handler:
Pageflip completion event has impossible msc 420120 < target_msc 420121"
as well as skipped frames and problems for applications that
use kms pageflip events or vblank events, e.g., users of DRI2 and
DRI3/Present, Waylands Weston compositor, etc. See also
https://bugs.freedesktop.org/show_bug.cgi?id=85203
After some talking to Alex and Michel, we decided to fix this
by turning the double-check for asserted irq status bits into a
warning. Whenever a irq event is queued in the IH ring, always
execute the corresponding interrupt handler. Still check the irq
status bits, but only to log a DRM_DEBUG message on a mismatch.
This fixed the problems reliably on both previously failing
cards, RV-730 dual-head tested on both crtcs (pipes D1 and D2)
and a triple-output Juniper HD-5770 card tested on all three
available crtcs (D1/D2/D3). The r600 and evergreen irq handling
is therefore tested, but the cik an si handling is only compile
tested due to lack of hw.
Reviewed-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Mario Kleiner <mario.kleiner.de@gmail.com>
CC: Michel Dänzer <michel.daenzer@amd.com>
CC: Alex Deucher <alexander.deucher@amd.com>
CC: <stable@vger.kernel.org> # v3.16+
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2015-07-03 07:03:06 +03:00
|
|
|
queue_dp = true;
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
event_name = "HPD_RX";
|
drm/radeon: Handle irqs only based on irq ring, not irq status regs.
Trying to resolve issues with missed vblanks and impossible
values inside delivered kms pageflip completion events showed
that radeon's irq handling sometimes doesn't handle valid irqs,
but silently skips them. This was observed for vblank interrupts.
Although those irqs have corresponding events queued in the gpu's
irq ring at time of interrupt, and therefore the corresponding
handling code gets triggered by these events, the handling code
sometimes silently skipped processing the irq. The reason for those
skips is that the handling code double-checks for each irq event if
the corresponding irq status bits in the irq status registers
are set. Sometimes those bits are not set at time of check
for valid irqs, maybe due to some hardware race on some setups?
The problem only seems to happen on some machine + card combos
sometimes, e.g., never happened during my testing of different PC
cards of the DCE-2/3/4 generation a year ago, but happens consistently
now on two different Apple Mac cards (RV730, DCE-3, Apple iMac and
Evergreen JUNIPER, DCE-4 in a Apple MacPro). It also doesn't happen
at each interrupt but only occassionally every couple of
hundred or thousand vblank interrupts.
This results in XOrg warning messages like
"[ 7084.472] (WW) RADEON(0): radeon_dri2_flip_event_handler:
Pageflip completion event has impossible msc 420120 < target_msc 420121"
as well as skipped frames and problems for applications that
use kms pageflip events or vblank events, e.g., users of DRI2 and
DRI3/Present, Waylands Weston compositor, etc. See also
https://bugs.freedesktop.org/show_bug.cgi?id=85203
After some talking to Alex and Michel, we decided to fix this
by turning the double-check for asserted irq status bits into a
warning. Whenever a irq event is queued in the IH ring, always
execute the corresponding interrupt handler. Still check the irq
status bits, but only to log a DRM_DEBUG message on a mismatch.
This fixed the problems reliably on both previously failing
cards, RV-730 dual-head tested on both crtcs (pipes D1 and D2)
and a triple-output Juniper HD-5770 card tested on all three
available crtcs (D1/D2/D3). The r600 and evergreen irq handling
is therefore tested, but the cik an si handling is only compile
tested due to lack of hw.
Reviewed-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Mario Kleiner <mario.kleiner.de@gmail.com>
CC: Michel Dänzer <michel.daenzer@amd.com>
CC: Alex Deucher <alexander.deucher@amd.com>
CC: <stable@vger.kernel.org> # v3.16+
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2015-07-03 07:03:06 +03:00
|
|
|
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
} else {
|
|
|
|
DRM_DEBUG("Unhandled interrupt: %d %d\n",
|
|
|
|
src_id, src_data);
|
2015-02-24 02:23:57 +03:00
|
|
|
break;
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
}
|
drm/radeon: Handle irqs only based on irq ring, not irq status regs.
Trying to resolve issues with missed vblanks and impossible
values inside delivered kms pageflip completion events showed
that radeon's irq handling sometimes doesn't handle valid irqs,
but silently skips them. This was observed for vblank interrupts.
Although those irqs have corresponding events queued in the gpu's
irq ring at time of interrupt, and therefore the corresponding
handling code gets triggered by these events, the handling code
sometimes silently skipped processing the irq. The reason for those
skips is that the handling code double-checks for each irq event if
the corresponding irq status bits in the irq status registers
are set. Sometimes those bits are not set at time of check
for valid irqs, maybe due to some hardware race on some setups?
The problem only seems to happen on some machine + card combos
sometimes, e.g., never happened during my testing of different PC
cards of the DCE-2/3/4 generation a year ago, but happens consistently
now on two different Apple Mac cards (RV730, DCE-3, Apple iMac and
Evergreen JUNIPER, DCE-4 in a Apple MacPro). It also doesn't happen
at each interrupt but only occassionally every couple of
hundred or thousand vblank interrupts.
This results in XOrg warning messages like
"[ 7084.472] (WW) RADEON(0): radeon_dri2_flip_event_handler:
Pageflip completion event has impossible msc 420120 < target_msc 420121"
as well as skipped frames and problems for applications that
use kms pageflip events or vblank events, e.g., users of DRI2 and
DRI3/Present, Waylands Weston compositor, etc. See also
https://bugs.freedesktop.org/show_bug.cgi?id=85203
After some talking to Alex and Michel, we decided to fix this
by turning the double-check for asserted irq status bits into a
warning. Whenever a irq event is queued in the IH ring, always
execute the corresponding interrupt handler. Still check the irq
status bits, but only to log a DRM_DEBUG message on a mismatch.
This fixed the problems reliably on both previously failing
cards, RV-730 dual-head tested on both crtcs (pipes D1 and D2)
and a triple-output Juniper HD-5770 card tested on all three
available crtcs (D1/D2/D3). The r600 and evergreen irq handling
is therefore tested, but the cik an si handling is only compile
tested due to lack of hw.
Reviewed-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Mario Kleiner <mario.kleiner.de@gmail.com>
CC: Michel Dänzer <michel.daenzer@amd.com>
CC: Alex Deucher <alexander.deucher@amd.com>
CC: <stable@vger.kernel.org> # v3.16+
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2015-07-03 07:03:06 +03:00
|
|
|
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
if (!(disp_int[hpd_idx] & mask))
|
|
|
|
DRM_DEBUG("IH: IH event w/o asserted irq bit?\n");
|
drm/radeon: Handle irqs only based on irq ring, not irq status regs.
Trying to resolve issues with missed vblanks and impossible
values inside delivered kms pageflip completion events showed
that radeon's irq handling sometimes doesn't handle valid irqs,
but silently skips them. This was observed for vblank interrupts.
Although those irqs have corresponding events queued in the gpu's
irq ring at time of interrupt, and therefore the corresponding
handling code gets triggered by these events, the handling code
sometimes silently skipped processing the irq. The reason for those
skips is that the handling code double-checks for each irq event if
the corresponding irq status bits in the irq status registers
are set. Sometimes those bits are not set at time of check
for valid irqs, maybe due to some hardware race on some setups?
The problem only seems to happen on some machine + card combos
sometimes, e.g., never happened during my testing of different PC
cards of the DCE-2/3/4 generation a year ago, but happens consistently
now on two different Apple Mac cards (RV730, DCE-3, Apple iMac and
Evergreen JUNIPER, DCE-4 in a Apple MacPro). It also doesn't happen
at each interrupt but only occassionally every couple of
hundred or thousand vblank interrupts.
This results in XOrg warning messages like
"[ 7084.472] (WW) RADEON(0): radeon_dri2_flip_event_handler:
Pageflip completion event has impossible msc 420120 < target_msc 420121"
as well as skipped frames and problems for applications that
use kms pageflip events or vblank events, e.g., users of DRI2 and
DRI3/Present, Waylands Weston compositor, etc. See also
https://bugs.freedesktop.org/show_bug.cgi?id=85203
After some talking to Alex and Michel, we decided to fix this
by turning the double-check for asserted irq status bits into a
warning. Whenever a irq event is queued in the IH ring, always
execute the corresponding interrupt handler. Still check the irq
status bits, but only to log a DRM_DEBUG message on a mismatch.
This fixed the problems reliably on both previously failing
cards, RV-730 dual-head tested on both crtcs (pipes D1 and D2)
and a triple-output Juniper HD-5770 card tested on all three
available crtcs (D1/D2/D3). The r600 and evergreen irq handling
is therefore tested, but the cik an si handling is only compile
tested due to lack of hw.
Reviewed-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Mario Kleiner <mario.kleiner.de@gmail.com>
CC: Michel Dänzer <michel.daenzer@amd.com>
CC: Alex Deucher <alexander.deucher@amd.com>
CC: <stable@vger.kernel.org> # v3.16+
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2015-07-03 07:03:06 +03:00
|
|
|
|
drm/radeon: Cleanup display interrupt handling for evergreen, si
The current code here is really, really bad. A huge amount of it looks
to be copy pasted, it has some weird hatred of arrays and code sharing,
switch cases everywhere for things that really don't need them, and it
makes the file seem immensely more complex then it actually is. This is
a pain for maintanence, and is vulnerable to more weird irq handling
bugs.
So, let's start cleaning this up a bit. Modify all of the IRQ handlers
for evergreen/si so that they just use for loops. As well, we add a
helper function radeon_irq_kms_set_irq_n_enabled(), whose purpose is
just to update the state of registers that enable/disable interrupts
while printing any changes to the set of enabled interrupts to the
kernel log.
Note in this commit, since vblank/vline irq acking is intertwined with
page flip irq acking, we can't cut out all of the copy paste in
evergreen/si_irq_ack() just yet.
Changes since v1:
- Preserve order we write back all registers
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Lyude <lyude@redhat.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2017-05-20 02:48:37 +03:00
|
|
|
disp_int[hpd_idx] &= ~mask;
|
|
|
|
DRM_DEBUG("IH: %s%d\n", event_name, hpd_idx + 1);
|
drm/radeon: Handle irqs only based on irq ring, not irq status regs.
Trying to resolve issues with missed vblanks and impossible
values inside delivered kms pageflip completion events showed
that radeon's irq handling sometimes doesn't handle valid irqs,
but silently skips them. This was observed for vblank interrupts.
Although those irqs have corresponding events queued in the gpu's
irq ring at time of interrupt, and therefore the corresponding
handling code gets triggered by these events, the handling code
sometimes silently skipped processing the irq. The reason for those
skips is that the handling code double-checks for each irq event if
the corresponding irq status bits in the irq status registers
are set. Sometimes those bits are not set at time of check
for valid irqs, maybe due to some hardware race on some setups?
The problem only seems to happen on some machine + card combos
sometimes, e.g., never happened during my testing of different PC
cards of the DCE-2/3/4 generation a year ago, but happens consistently
now on two different Apple Mac cards (RV730, DCE-3, Apple iMac and
Evergreen JUNIPER, DCE-4 in a Apple MacPro). It also doesn't happen
at each interrupt but only occassionally every couple of
hundred or thousand vblank interrupts.
This results in XOrg warning messages like
"[ 7084.472] (WW) RADEON(0): radeon_dri2_flip_event_handler:
Pageflip completion event has impossible msc 420120 < target_msc 420121"
as well as skipped frames and problems for applications that
use kms pageflip events or vblank events, e.g., users of DRI2 and
DRI3/Present, Waylands Weston compositor, etc. See also
https://bugs.freedesktop.org/show_bug.cgi?id=85203
After some talking to Alex and Michel, we decided to fix this
by turning the double-check for asserted irq status bits into a
warning. Whenever a irq event is queued in the IH ring, always
execute the corresponding interrupt handler. Still check the irq
status bits, but only to log a DRM_DEBUG message on a mismatch.
This fixed the problems reliably on both previously failing
cards, RV-730 dual-head tested on both crtcs (pipes D1 and D2)
and a triple-output Juniper HD-5770 card tested on all three
available crtcs (D1/D2/D3). The r600 and evergreen irq handling
is therefore tested, but the cik an si handling is only compile
tested due to lack of hw.
Reviewed-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Mario Kleiner <mario.kleiner.de@gmail.com>
CC: Michel Dänzer <michel.daenzer@amd.com>
CC: Alex Deucher <alexander.deucher@amd.com>
CC: <stable@vger.kernel.org> # v3.16+
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2015-07-03 07:03:06 +03:00
|
|
|
|
2010-03-24 20:55:51 +03:00
|
|
|
break;
|
2012-03-30 16:59:57 +04:00
|
|
|
case 44: /* hdmi */
|
2017-05-20 02:48:38 +03:00
|
|
|
afmt_idx = src_data;
|
|
|
|
if (!(afmt_status[afmt_idx] & AFMT_AZ_FORMAT_WTRIG))
|
|
|
|
DRM_DEBUG("IH: IH event w/o asserted irq bit?\n");
|
drm/radeon: Handle irqs only based on irq ring, not irq status regs.
Trying to resolve issues with missed vblanks and impossible
values inside delivered kms pageflip completion events showed
that radeon's irq handling sometimes doesn't handle valid irqs,
but silently skips them. This was observed for vblank interrupts.
Although those irqs have corresponding events queued in the gpu's
irq ring at time of interrupt, and therefore the corresponding
handling code gets triggered by these events, the handling code
sometimes silently skipped processing the irq. The reason for those
skips is that the handling code double-checks for each irq event if
the corresponding irq status bits in the irq status registers
are set. Sometimes those bits are not set at time of check
for valid irqs, maybe due to some hardware race on some setups?
The problem only seems to happen on some machine + card combos
sometimes, e.g., never happened during my testing of different PC
cards of the DCE-2/3/4 generation a year ago, but happens consistently
now on two different Apple Mac cards (RV730, DCE-3, Apple iMac and
Evergreen JUNIPER, DCE-4 in a Apple MacPro). It also doesn't happen
at each interrupt but only occassionally every couple of
hundred or thousand vblank interrupts.
This results in XOrg warning messages like
"[ 7084.472] (WW) RADEON(0): radeon_dri2_flip_event_handler:
Pageflip completion event has impossible msc 420120 < target_msc 420121"
as well as skipped frames and problems for applications that
use kms pageflip events or vblank events, e.g., users of DRI2 and
DRI3/Present, Waylands Weston compositor, etc. See also
https://bugs.freedesktop.org/show_bug.cgi?id=85203
After some talking to Alex and Michel, we decided to fix this
by turning the double-check for asserted irq status bits into a
warning. Whenever a irq event is queued in the IH ring, always
execute the corresponding interrupt handler. Still check the irq
status bits, but only to log a DRM_DEBUG message on a mismatch.
This fixed the problems reliably on both previously failing
cards, RV-730 dual-head tested on both crtcs (pipes D1 and D2)
and a triple-output Juniper HD-5770 card tested on all three
available crtcs (D1/D2/D3). The r600 and evergreen irq handling
is therefore tested, but the cik an si handling is only compile
tested due to lack of hw.
Reviewed-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Mario Kleiner <mario.kleiner.de@gmail.com>
CC: Michel Dänzer <michel.daenzer@amd.com>
CC: Alex Deucher <alexander.deucher@amd.com>
CC: <stable@vger.kernel.org> # v3.16+
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
2015-07-03 07:03:06 +03:00
|
|
|
|
2017-05-20 02:48:38 +03:00
|
|
|
if (afmt_idx > 5) {
|
|
|
|
DRM_ERROR("Unhandled interrupt: %d %d\n",
|
|
|
|
src_id, src_data);
|
2012-03-30 16:59:57 +04:00
|
|
|
break;
|
|
|
|
}
|
2017-05-20 02:48:38 +03:00
|
|
|
afmt_status[afmt_idx] &= ~AFMT_AZ_FORMAT_WTRIG;
|
|
|
|
queue_hdmi = true;
|
|
|
|
DRM_DEBUG("IH: HDMI%d\n", afmt_idx + 1);
|
|
|
|
break;
|
2015-02-18 15:19:28 +03:00
|
|
|
case 96:
|
|
|
|
DRM_ERROR("SRBM_READ_ERROR: 0x%x\n", RREG32(SRBM_READ_ERROR));
|
|
|
|
WREG32(SRBM_INT_ACK, 0x1);
|
|
|
|
break;
|
2013-04-08 14:41:29 +04:00
|
|
|
case 124: /* UVD */
|
|
|
|
DRM_DEBUG("IH: UVD int: 0x%08x\n", src_data);
|
|
|
|
radeon_fence_process(rdev, R600_RING_TYPE_UVD_INDEX);
|
2012-03-30 16:59:57 +04:00
|
|
|
break;
|
2012-09-18 23:30:44 +04:00
|
|
|
case 146:
|
|
|
|
case 147:
|
2013-06-14 02:26:25 +04:00
|
|
|
addr = RREG32(VM_CONTEXT1_PROTECTION_FAULT_ADDR);
|
|
|
|
status = RREG32(VM_CONTEXT1_PROTECTION_FAULT_STATUS);
|
2014-07-07 13:16:29 +04:00
|
|
|
/* reset addr and status */
|
|
|
|
WREG32_P(VM_CONTEXT1_CNTL2, 1, ~1);
|
|
|
|
if (addr == 0x0 && status == 0x0)
|
|
|
|
break;
|
2012-09-18 23:30:44 +04:00
|
|
|
dev_err(rdev->dev, "GPU fault detected: %d 0x%08x\n", src_id, src_data);
|
|
|
|
dev_err(rdev->dev, " VM_CONTEXT1_PROTECTION_FAULT_ADDR 0x%08X\n",
|
2013-06-14 02:26:25 +04:00
|
|
|
addr);
|
2012-09-18 23:30:44 +04:00
|
|
|
dev_err(rdev->dev, " VM_CONTEXT1_PROTECTION_FAULT_STATUS 0x%08X\n",
|
2013-06-14 02:26:25 +04:00
|
|
|
status);
|
|
|
|
cayman_vm_decode_fault(rdev, status, addr);
|
2012-09-18 23:30:44 +04:00
|
|
|
break;
|
2010-03-24 20:55:51 +03:00
|
|
|
case 176: /* CP_INT in ring buffer */
|
|
|
|
case 177: /* CP_INT in IB1 */
|
|
|
|
case 178: /* CP_INT in IB2 */
|
|
|
|
DRM_DEBUG("IH: CP int: 0x%08x\n", src_data);
|
2011-08-25 21:39:48 +04:00
|
|
|
radeon_fence_process(rdev, RADEON_RING_TYPE_GFX_INDEX);
|
2010-03-24 20:55:51 +03:00
|
|
|
break;
|
|
|
|
case 181: /* CP EOP event */
|
|
|
|
DRM_DEBUG("IH: CP EOP\n");
|
2011-11-18 05:13:28 +04:00
|
|
|
if (rdev->family >= CHIP_CAYMAN) {
|
|
|
|
switch (src_data) {
|
|
|
|
case 0:
|
|
|
|
radeon_fence_process(rdev, RADEON_RING_TYPE_GFX_INDEX);
|
|
|
|
break;
|
|
|
|
case 1:
|
|
|
|
radeon_fence_process(rdev, CAYMAN_RING_TYPE_CP1_INDEX);
|
|
|
|
break;
|
|
|
|
case 2:
|
|
|
|
radeon_fence_process(rdev, CAYMAN_RING_TYPE_CP2_INDEX);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
} else
|
|
|
|
radeon_fence_process(rdev, RADEON_RING_TYPE_GFX_INDEX);
|
2010-03-24 20:55:51 +03:00
|
|
|
break;
|
2012-12-05 00:25:59 +04:00
|
|
|
case 224: /* DMA trap event */
|
|
|
|
DRM_DEBUG("IH: DMA trap\n");
|
|
|
|
radeon_fence_process(rdev, R600_RING_TYPE_DMA_INDEX);
|
|
|
|
break;
|
2013-06-26 08:33:35 +04:00
|
|
|
case 230: /* thermal low to high */
|
|
|
|
DRM_DEBUG("IH: thermal low to high\n");
|
|
|
|
rdev->pm.dpm.thermal.high_to_low = false;
|
|
|
|
queue_thermal = true;
|
|
|
|
break;
|
|
|
|
case 231: /* thermal high to low */
|
|
|
|
DRM_DEBUG("IH: thermal high to low\n");
|
|
|
|
rdev->pm.dpm.thermal.high_to_low = true;
|
|
|
|
queue_thermal = true;
|
|
|
|
break;
|
2010-04-22 20:52:11 +04:00
|
|
|
case 233: /* GUI IDLE */
|
2011-06-07 22:54:48 +04:00
|
|
|
DRM_DEBUG("IH: GUI idle\n");
|
2010-04-22 20:52:11 +04:00
|
|
|
break;
|
2012-12-05 00:27:33 +04:00
|
|
|
case 244: /* DMA trap event */
|
|
|
|
if (rdev->family >= CHIP_CAYMAN) {
|
|
|
|
DRM_DEBUG("IH: DMA1 trap\n");
|
|
|
|
radeon_fence_process(rdev, CAYMAN_RING_TYPE_DMA1_INDEX);
|
|
|
|
}
|
|
|
|
break;
|
2010-03-24 20:55:51 +03:00
|
|
|
default:
|
|
|
|
DRM_DEBUG("Unhandled interrupt: %d %d\n", src_id, src_data);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* wptr/rptr are in bytes! */
|
|
|
|
rptr += 16;
|
|
|
|
rptr &= rdev->ih.ptr_mask;
|
2014-09-19 07:22:10 +04:00
|
|
|
WREG32(IH_RB_RPTR, rptr);
|
2010-03-24 20:55:51 +03:00
|
|
|
}
|
2015-02-24 02:23:57 +03:00
|
|
|
if (queue_dp)
|
|
|
|
schedule_work(&rdev->dp_work);
|
2010-03-24 20:55:51 +03:00
|
|
|
if (queue_hotplug)
|
2015-12-04 02:26:07 +03:00
|
|
|
schedule_delayed_work(&rdev->hotplug_work, 0);
|
2012-03-30 16:59:57 +04:00
|
|
|
if (queue_hdmi)
|
|
|
|
schedule_work(&rdev->audio_work);
|
2013-06-26 08:33:35 +04:00
|
|
|
if (queue_thermal && rdev->pm.dpm_enabled)
|
|
|
|
schedule_work(&rdev->pm.dpm.thermal.work);
|
2010-03-24 20:55:51 +03:00
|
|
|
rdev->ih.rptr = rptr;
|
2012-05-16 23:45:24 +04:00
|
|
|
atomic_set(&rdev->ih.lock, 0);
|
|
|
|
|
|
|
|
/* make sure wptr hasn't changed while processing */
|
|
|
|
wptr = evergreen_get_ih_wptr(rdev);
|
|
|
|
if (wptr != rptr)
|
|
|
|
goto restart_ih;
|
|
|
|
|
2010-03-24 20:55:51 +03:00
|
|
|
return IRQ_HANDLED;
|
|
|
|
}
|
|
|
|
|
2016-03-18 18:58:28 +03:00
|
|
|
static void evergreen_uvd_init(struct radeon_device *rdev)
|
|
|
|
{
|
|
|
|
int r;
|
|
|
|
|
|
|
|
if (!rdev->has_uvd)
|
|
|
|
return;
|
|
|
|
|
|
|
|
r = radeon_uvd_init(rdev);
|
|
|
|
if (r) {
|
|
|
|
dev_err(rdev->dev, "failed UVD (%d) init.\n", r);
|
|
|
|
/*
|
|
|
|
* At this point rdev->uvd.vcpu_bo is NULL which trickles down
|
|
|
|
* to early fails uvd_v2_2_resume() and thus nothing happens
|
|
|
|
* there. So it is pointless to try to go through that code
|
|
|
|
* hence why we disable uvd here.
|
|
|
|
*/
|
|
|
|
rdev->has_uvd = 0;
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_obj = NULL;
|
|
|
|
r600_ring_init(rdev, &rdev->ring[R600_RING_TYPE_UVD_INDEX], 4096);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void evergreen_uvd_start(struct radeon_device *rdev)
|
|
|
|
{
|
|
|
|
int r;
|
|
|
|
|
|
|
|
if (!rdev->has_uvd)
|
|
|
|
return;
|
|
|
|
|
|
|
|
r = uvd_v2_2_resume(rdev);
|
|
|
|
if (r) {
|
|
|
|
dev_err(rdev->dev, "failed UVD resume (%d).\n", r);
|
|
|
|
goto error;
|
|
|
|
}
|
|
|
|
r = radeon_fence_driver_start_ring(rdev, R600_RING_TYPE_UVD_INDEX);
|
|
|
|
if (r) {
|
|
|
|
dev_err(rdev->dev, "failed initializing UVD fences (%d).\n", r);
|
|
|
|
goto error;
|
|
|
|
}
|
|
|
|
return;
|
|
|
|
|
|
|
|
error:
|
|
|
|
rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_size = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void evergreen_uvd_resume(struct radeon_device *rdev)
|
|
|
|
{
|
|
|
|
struct radeon_ring *ring;
|
|
|
|
int r;
|
|
|
|
|
|
|
|
if (!rdev->has_uvd || !rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_size)
|
|
|
|
return;
|
|
|
|
|
|
|
|
ring = &rdev->ring[R600_RING_TYPE_UVD_INDEX];
|
2016-08-23 17:07:28 +03:00
|
|
|
r = radeon_ring_init(rdev, ring, ring->ring_size, 0, PACKET0(UVD_NO_OP, 0));
|
2016-03-18 18:58:28 +03:00
|
|
|
if (r) {
|
|
|
|
dev_err(rdev->dev, "failed initializing UVD ring (%d).\n", r);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
r = uvd_v1_0_init(rdev);
|
|
|
|
if (r) {
|
|
|
|
dev_err(rdev->dev, "failed initializing UVD (%d).\n", r);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2010-01-13 01:54:34 +03:00
|
|
|
static int evergreen_startup(struct radeon_device *rdev)
|
|
|
|
{
|
2013-04-08 14:41:29 +04:00
|
|
|
struct radeon_ring *ring;
|
2010-01-13 01:54:34 +03:00
|
|
|
int r;
|
|
|
|
|
2011-01-07 02:49:35 +03:00
|
|
|
/* enable pcie gen2 link */
|
2011-09-20 18:22:57 +04:00
|
|
|
evergreen_pcie_gen2_enable(rdev);
|
2013-02-15 20:02:50 +04:00
|
|
|
/* enable aspm */
|
|
|
|
evergreen_program_aspm(rdev);
|
2011-01-07 02:49:35 +03:00
|
|
|
|
2013-08-30 16:58:20 +04:00
|
|
|
/* scratch needs to be initialized before MC */
|
|
|
|
r = r600_vram_scratch_init(rdev);
|
|
|
|
if (r)
|
|
|
|
return r;
|
|
|
|
|
2013-08-04 20:13:17 +04:00
|
|
|
evergreen_mc_program(rdev);
|
|
|
|
|
2013-12-18 23:07:14 +04:00
|
|
|
if (ASIC_IS_DCE5(rdev) && !rdev->pm.dpm_enabled) {
|
2011-03-03 04:07:34 +03:00
|
|
|
r = ni_mc_load_microcode(rdev);
|
2010-01-13 01:54:34 +03:00
|
|
|
if (r) {
|
2011-01-07 05:19:31 +03:00
|
|
|
DRM_ERROR("Failed to load MC firmware!\n");
|
2010-01-13 01:54:34 +03:00
|
|
|
return r;
|
|
|
|
}
|
|
|
|
}
|
2010-03-24 20:36:43 +03:00
|
|
|
|
2010-01-13 01:54:34 +03:00
|
|
|
if (rdev->flags & RADEON_IS_AGP) {
|
2010-03-24 20:20:41 +03:00
|
|
|
evergreen_agp_enable(rdev);
|
2010-01-13 01:54:34 +03:00
|
|
|
} else {
|
|
|
|
r = evergreen_pcie_gart_enable(rdev);
|
|
|
|
if (r)
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
evergreen_gpu_init(rdev);
|
|
|
|
|
2013-04-12 21:52:52 +04:00
|
|
|
/* allocate rlc buffers */
|
|
|
|
if (rdev->flags & RADEON_IS_IGP) {
|
|
|
|
rdev->rlc.reg_list = sumo_rlc_save_restore_register_list;
|
2013-04-18 01:53:50 +04:00
|
|
|
rdev->rlc.reg_list_size =
|
|
|
|
(u32)ARRAY_SIZE(sumo_rlc_save_restore_register_list);
|
2013-04-12 21:52:52 +04:00
|
|
|
rdev->rlc.cs_data = evergreen_cs_data;
|
|
|
|
r = sumo_rlc_init(rdev);
|
|
|
|
if (r) {
|
|
|
|
DRM_ERROR("Failed to init rlc BOs!\n");
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2010-08-28 02:25:25 +04:00
|
|
|
/* allocate wb buffer */
|
|
|
|
r = radeon_wb_init(rdev);
|
|
|
|
if (r)
|
|
|
|
return r;
|
|
|
|
|
2011-11-21 00:45:34 +04:00
|
|
|
r = radeon_fence_driver_start_ring(rdev, RADEON_RING_TYPE_GFX_INDEX);
|
|
|
|
if (r) {
|
|
|
|
dev_err(rdev->dev, "failed initializing CP fences (%d).\n", r);
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
|
2012-12-05 00:25:59 +04:00
|
|
|
r = radeon_fence_driver_start_ring(rdev, R600_RING_TYPE_DMA_INDEX);
|
|
|
|
if (r) {
|
|
|
|
dev_err(rdev->dev, "failed initializing DMA fences (%d).\n", r);
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
|
2016-03-18 18:58:28 +03:00
|
|
|
evergreen_uvd_start(rdev);
|
2013-04-08 14:41:29 +04:00
|
|
|
|
2010-01-13 01:54:34 +03:00
|
|
|
/* Enable IRQ */
|
radeon: Fix system hang issue when using KMS with older cards
The current radeon driver initialization routines, when using KMS, are written
so that the IRQ installation routine is called before initializing the WB buffer
and the CP rings. With some ASICs, though, the IRQ routine tries to access the
GFX_INDEX ring causing a call to RREG32 with the value of -1 in
radeon_fence_read. This, in turn causes the system to completely hang with some
cards, requiring a hard reset.
A call stack that can cause such a hang looks like this (using rv515 ASIC for the
example here):
* rv515_init (rv515.c)
* radeon_irq_kms_init (radeon_irq_kms.c)
* drm_irq_install (drm_irq.c)
* radeon_driver_irq_preinstall_kms (radeon_irq_kms.c)
* rs600_irq_process (rs600.c)
* radeon_fence_process - due to SW interrupt (radeon_fence.c)
* radeon_fence_read (radeon_fence.c)
* hang due to RREG32(-1)
The patch moves the IRQ installation to the card startup routine, after the ring
has been initialized, but before the IRQ has been set. This fixes the issue, but
requires a check to see if the IRQ is already installed, as is the case in the
system resume codepath.
I have tested the patch on three machines using the rv515, the rv770 and the
evergreen ASIC. They worked without issues.
This seems to be a known issue and has been reported on several bug tracking
sites by various distributions (see links below). Most of reports recommend
booting the system with KMS disabled and then enabling KMS by reloading the
radeon module. For some reason, this was indeed a usable workaround, however,
UMS is now deprecated and disabled by default.
Bug reports:
https://bugzilla.redhat.com/show_bug.cgi?id=845745
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/561789
https://bbs.archlinux.org/viewtopic.php?id=156964
Signed-off-by: Adis Hamzić <adis@hamzadis.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
Cc: stable@vger.kernel.org
2013-06-02 18:47:54 +04:00
|
|
|
if (!rdev->irq.installed) {
|
|
|
|
r = radeon_irq_kms_init(rdev);
|
|
|
|
if (r)
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
|
2010-01-13 01:54:34 +03:00
|
|
|
r = r600_irq_init(rdev);
|
|
|
|
if (r) {
|
|
|
|
DRM_ERROR("radeon: IH init failed (%d).\n", r);
|
|
|
|
radeon_irq_kms_fini(rdev);
|
|
|
|
return r;
|
|
|
|
}
|
2010-03-24 20:55:51 +03:00
|
|
|
evergreen_irq_set(rdev);
|
2010-01-13 01:54:34 +03:00
|
|
|
|
2013-04-08 14:41:29 +04:00
|
|
|
ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX];
|
2011-10-23 14:56:27 +04:00
|
|
|
r = radeon_ring_init(rdev, ring, ring->ring_size, RADEON_WB_CP_RPTR_OFFSET,
|
2013-08-13 13:56:52 +04:00
|
|
|
RADEON_CP_PACKET2);
|
2010-01-13 01:54:34 +03:00
|
|
|
if (r)
|
|
|
|
return r;
|
2012-12-05 00:25:59 +04:00
|
|
|
|
|
|
|
ring = &rdev->ring[R600_RING_TYPE_DMA_INDEX];
|
|
|
|
r = radeon_ring_init(rdev, ring, ring->ring_size, R600_WB_DMA_RPTR_OFFSET,
|
2013-08-13 13:56:52 +04:00
|
|
|
DMA_PACKET(DMA_PACKET_NOP, 0, 0));
|
2012-12-05 00:25:59 +04:00
|
|
|
if (r)
|
|
|
|
return r;
|
|
|
|
|
2010-01-13 01:54:34 +03:00
|
|
|
r = evergreen_cp_load_microcode(rdev);
|
|
|
|
if (r)
|
|
|
|
return r;
|
2010-03-24 20:36:43 +03:00
|
|
|
r = evergreen_cp_resume(rdev);
|
2012-12-05 00:25:59 +04:00
|
|
|
if (r)
|
|
|
|
return r;
|
|
|
|
r = r600_dma_resume(rdev);
|
2010-01-13 01:54:34 +03:00
|
|
|
if (r)
|
|
|
|
return r;
|
2010-03-24 20:36:43 +03:00
|
|
|
|
2016-03-18 18:58:28 +03:00
|
|
|
evergreen_uvd_resume(rdev);
|
2013-04-08 14:41:29 +04:00
|
|
|
|
2012-07-05 13:55:34 +04:00
|
|
|
r = radeon_ib_pool_init(rdev);
|
|
|
|
if (r) {
|
|
|
|
dev_err(rdev->dev, "IB initialization failed (%d).\n", r);
|
drm/radeon: introduce a sub allocator and convert ib pool to it v4
Somewhat specializaed sub-allocator designed to perform sub-allocation
for command buffer not only for current cs ioctl but for future command
submission ioctl as well. Patch also convert current ib pool to use
the sub allocator. Idea is that ib poll buffer can be share with other
command buffer submission not having 64K granularity.
v2 Harmonize pool handling and add suspend/resume callback to pin/unpin
sa bo (tested on rv280, rv370, r420, rv515, rv610, rv710, redwood, cayman,
rs480, rs690, rs880)
v3 Simplify allocator
v4 Fix radeon_ib_get error path to properly free fence
Signed-off-by: Jerome Glisse <jglisse@redhat.com>
Signed-off-by: Dave Airlie <airlied@redhat.com>
2011-11-15 20:48:34 +04:00
|
|
|
return r;
|
2012-07-05 13:55:34 +04:00
|
|
|
}
|
drm/radeon: introduce a sub allocator and convert ib pool to it v4
Somewhat specializaed sub-allocator designed to perform sub-allocation
for command buffer not only for current cs ioctl but for future command
submission ioctl as well. Patch also convert current ib pool to use
the sub allocator. Idea is that ib poll buffer can be share with other
command buffer submission not having 64K granularity.
v2 Harmonize pool handling and add suspend/resume callback to pin/unpin
sa bo (tested on rv280, rv370, r420, rv515, rv610, rv710, redwood, cayman,
rs480, rs690, rs880)
v3 Simplify allocator
v4 Fix radeon_ib_get error path to properly free fence
Signed-off-by: Jerome Glisse <jglisse@redhat.com>
Signed-off-by: Dave Airlie <airlied@redhat.com>
2011-11-15 20:48:34 +04:00
|
|
|
|
2014-12-23 01:26:51 +03:00
|
|
|
r = radeon_audio_init(rdev);
|
2011-12-08 02:32:24 +04:00
|
|
|
if (r) {
|
|
|
|
DRM_ERROR("radeon: audio init failed\n");
|
drm/radeon: introduce a sub allocator and convert ib pool to it v4
Somewhat specializaed sub-allocator designed to perform sub-allocation
for command buffer not only for current cs ioctl but for future command
submission ioctl as well. Patch also convert current ib pool to use
the sub allocator. Idea is that ib poll buffer can be share with other
command buffer submission not having 64K granularity.
v2 Harmonize pool handling and add suspend/resume callback to pin/unpin
sa bo (tested on rv280, rv370, r420, rv515, rv610, rv710, redwood, cayman,
rs480, rs690, rs880)
v3 Simplify allocator
v4 Fix radeon_ib_get error path to properly free fence
Signed-off-by: Jerome Glisse <jglisse@redhat.com>
Signed-off-by: Dave Airlie <airlied@redhat.com>
2011-11-15 20:48:34 +04:00
|
|
|
return r;
|
|
|
|
}
|
|
|
|
|
2010-01-13 01:54:34 +03:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
int evergreen_resume(struct radeon_device *rdev)
|
|
|
|
{
|
|
|
|
int r;
|
|
|
|
|
2010-12-20 20:35:04 +03:00
|
|
|
/* reset the asic, the gfx blocks are often in a bad state
|
|
|
|
* after the driver is unloaded or after a resume
|
|
|
|
*/
|
|
|
|
if (radeon_asic_reset(rdev))
|
|
|
|
dev_warn(rdev->dev, "GPU reset failed !\n");
|
2010-01-13 01:54:34 +03:00
|
|
|
/* Do not reset GPU before posting, on rv770 hw unlike on r500 hw,
|
|
|
|
* posting will perform necessary task to bring back GPU into good
|
|
|
|
* shape.
|
|
|
|
*/
|
|
|
|
/* post card */
|
|
|
|
atom_asic_init(rdev->mode_info.atom_context);
|
|
|
|
|
2013-02-28 23:40:09 +04:00
|
|
|
/* init golden registers */
|
|
|
|
evergreen_init_golden_registers(rdev);
|
|
|
|
|
2014-02-25 21:01:28 +04:00
|
|
|
if (rdev->pm.pm_method == PM_METHOD_DPM)
|
|
|
|
radeon_pm_resume(rdev);
|
2013-12-18 23:07:14 +04:00
|
|
|
|
drm/radeon: introduce a sub allocator and convert ib pool to it v4
Somewhat specializaed sub-allocator designed to perform sub-allocation
for command buffer not only for current cs ioctl but for future command
submission ioctl as well. Patch also convert current ib pool to use
the sub allocator. Idea is that ib poll buffer can be share with other
command buffer submission not having 64K granularity.
v2 Harmonize pool handling and add suspend/resume callback to pin/unpin
sa bo (tested on rv280, rv370, r420, rv515, rv610, rv710, redwood, cayman,
rs480, rs690, rs880)
v3 Simplify allocator
v4 Fix radeon_ib_get error path to properly free fence
Signed-off-by: Jerome Glisse <jglisse@redhat.com>
Signed-off-by: Dave Airlie <airlied@redhat.com>
2011-11-15 20:48:34 +04:00
|
|
|
rdev->accel_working = true;
|
2010-01-13 01:54:34 +03:00
|
|
|
r = evergreen_startup(rdev);
|
|
|
|
if (r) {
|
2011-03-03 04:07:34 +03:00
|
|
|
DRM_ERROR("evergreen startup failed on resume\n");
|
2012-02-21 02:57:20 +04:00
|
|
|
rdev->accel_working = false;
|
2010-01-13 01:54:34 +03:00
|
|
|
return r;
|
|
|
|
}
|
2010-03-24 20:36:43 +03:00
|
|
|
|
2010-01-13 01:54:34 +03:00
|
|
|
return r;
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
int evergreen_suspend(struct radeon_device *rdev)
|
|
|
|
{
|
2013-12-18 23:07:14 +04:00
|
|
|
radeon_pm_suspend(rdev);
|
2014-12-04 01:07:01 +03:00
|
|
|
radeon_audio_fini(rdev);
|
2016-03-18 18:58:28 +03:00
|
|
|
if (rdev->has_uvd) {
|
|
|
|
uvd_v1_0_fini(rdev);
|
|
|
|
radeon_uvd_suspend(rdev);
|
|
|
|
}
|
2010-01-13 01:54:34 +03:00
|
|
|
r700_cp_stop(rdev);
|
2012-12-05 00:25:59 +04:00
|
|
|
r600_dma_stop(rdev);
|
2010-03-24 20:55:51 +03:00
|
|
|
evergreen_irq_suspend(rdev);
|
2010-08-28 02:25:25 +04:00
|
|
|
radeon_wb_disable(rdev);
|
2010-01-13 01:54:34 +03:00
|
|
|
evergreen_pcie_gart_disable(rdev);
|
2010-09-09 19:33:36 +04:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2010-01-13 01:54:34 +03:00
|
|
|
/* Plan is to move initialization in that function and use
|
|
|
|
* helper function so that radeon_device_init pretty much
|
|
|
|
* do nothing more than calling asic specific function. This
|
|
|
|
* should also allow to remove a bunch of callback function
|
|
|
|
* like vram_info.
|
|
|
|
*/
|
|
|
|
int evergreen_init(struct radeon_device *rdev)
|
|
|
|
{
|
|
|
|
int r;
|
|
|
|
|
|
|
|
/* Read BIOS */
|
|
|
|
if (!radeon_get_bios(rdev)) {
|
|
|
|
if (ASIC_IS_AVIVO(rdev))
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
/* Must be an ATOMBIOS */
|
|
|
|
if (!rdev->is_atom_bios) {
|
2011-03-03 04:07:34 +03:00
|
|
|
dev_err(rdev->dev, "Expecting atombios for evergreen GPU\n");
|
2010-01-13 01:54:34 +03:00
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
r = radeon_atombios_init(rdev);
|
|
|
|
if (r)
|
|
|
|
return r;
|
2010-12-20 20:35:04 +03:00
|
|
|
/* reset the asic, the gfx blocks are often in a bad state
|
|
|
|
* after the driver is unloaded or after a resume
|
|
|
|
*/
|
|
|
|
if (radeon_asic_reset(rdev))
|
|
|
|
dev_warn(rdev->dev, "GPU reset failed !\n");
|
2010-01-13 01:54:34 +03:00
|
|
|
/* Post card if necessary */
|
2011-01-12 02:08:59 +03:00
|
|
|
if (!radeon_card_posted(rdev)) {
|
2010-01-13 01:54:34 +03:00
|
|
|
if (!rdev->bios) {
|
|
|
|
dev_err(rdev->dev, "Card not posted and no BIOS - ignoring\n");
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
DRM_INFO("GPU not posted. posting now...\n");
|
|
|
|
atom_asic_init(rdev->mode_info.atom_context);
|
|
|
|
}
|
2013-02-28 23:40:09 +04:00
|
|
|
/* init golden registers */
|
|
|
|
evergreen_init_golden_registers(rdev);
|
2010-01-13 01:54:34 +03:00
|
|
|
/* Initialize scratch registers */
|
|
|
|
r600_scratch_init(rdev);
|
|
|
|
/* Initialize surface registers */
|
|
|
|
radeon_surface_init(rdev);
|
|
|
|
/* Initialize clocks */
|
|
|
|
radeon_get_clock_info(rdev->ddev);
|
|
|
|
/* Fence driver */
|
|
|
|
r = radeon_fence_driver_init(rdev);
|
|
|
|
if (r)
|
|
|
|
return r;
|
drm/radeon/kms: simplify memory controller setup V2
Get rid of _location and use _start/_end also simplify the
computation of vram_start|end & gtt_start|end. For R1XX-R2XX
we place VRAM at the same address of PCI aperture, those GPU
shouldn't have much memory and seems to behave better when
setup that way. For R3XX and newer we place VRAM at 0. For
R6XX-R7XX AGP we place VRAM before or after AGP aperture this
might limit to limit the VRAM size but it's very unlikely.
For IGP we don't change the VRAM placement.
Tested on (compiz,quake3,suspend/resume):
PCI/PCIE:RV280,R420,RV515,RV570,RV610,RV710
AGP:RV100,RV280,R420,RV350,RV620(RPB*),RV730
IGP:RS480(RPB*),RS690,RS780(RPB*),RS880
RPB: resume previously broken
V2 correct commit message to reflect more accurately the bug
and move VRAM placement to 0 for most of the GPU to avoid
limiting VRAM.
Signed-off-by: Jerome Glisse <jglisse@redhat.com>
Signed-off-by: Dave Airlie <airlied@redhat.com>
2010-02-18 00:54:29 +03:00
|
|
|
/* initialize AGP */
|
|
|
|
if (rdev->flags & RADEON_IS_AGP) {
|
|
|
|
r = radeon_agp_init(rdev);
|
|
|
|
if (r)
|
|
|
|
radeon_agp_disable(rdev);
|
|
|
|
}
|
|
|
|
/* initialize memory controller */
|
2010-01-13 01:54:34 +03:00
|
|
|
r = evergreen_mc_init(rdev);
|
|
|
|
if (r)
|
|
|
|
return r;
|
|
|
|
/* Memory manager */
|
|
|
|
r = radeon_bo_init(rdev);
|
|
|
|
if (r)
|
|
|
|
return r;
|
2010-03-24 20:55:51 +03:00
|
|
|
|
2013-12-19 04:11:27 +04:00
|
|
|
if (ASIC_IS_DCE5(rdev)) {
|
|
|
|
if (!rdev->me_fw || !rdev->pfp_fw || !rdev->rlc_fw || !rdev->mc_fw) {
|
|
|
|
r = ni_init_microcode(rdev);
|
|
|
|
if (r) {
|
|
|
|
DRM_ERROR("Failed to load firmware!\n");
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
if (!rdev->me_fw || !rdev->pfp_fw || !rdev->rlc_fw) {
|
|
|
|
r = r600_init_microcode(rdev);
|
|
|
|
if (r) {
|
|
|
|
DRM_ERROR("Failed to load firmware!\n");
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2013-12-18 23:07:14 +04:00
|
|
|
/* Initialize power management */
|
|
|
|
radeon_pm_init(rdev);
|
|
|
|
|
2011-10-23 14:56:27 +04:00
|
|
|
rdev->ring[RADEON_RING_TYPE_GFX_INDEX].ring_obj = NULL;
|
|
|
|
r600_ring_init(rdev, &rdev->ring[RADEON_RING_TYPE_GFX_INDEX], 1024 * 1024);
|
2010-01-13 01:54:34 +03:00
|
|
|
|
2012-12-05 00:25:59 +04:00
|
|
|
rdev->ring[R600_RING_TYPE_DMA_INDEX].ring_obj = NULL;
|
|
|
|
r600_ring_init(rdev, &rdev->ring[R600_RING_TYPE_DMA_INDEX], 64 * 1024);
|
|
|
|
|
2016-03-18 18:58:28 +03:00
|
|
|
evergreen_uvd_init(rdev);
|
2013-04-08 14:41:29 +04:00
|
|
|
|
2010-01-13 01:54:34 +03:00
|
|
|
rdev->ih.ring_obj = NULL;
|
|
|
|
r600_ih_ring_init(rdev, 64 * 1024);
|
|
|
|
|
|
|
|
r = r600_pcie_gart_init(rdev);
|
|
|
|
if (r)
|
|
|
|
return r;
|
2010-03-24 20:20:41 +03:00
|
|
|
|
2010-06-04 03:00:03 +04:00
|
|
|
rdev->accel_working = true;
|
2010-01-13 01:54:34 +03:00
|
|
|
r = evergreen_startup(rdev);
|
|
|
|
if (r) {
|
2010-03-24 20:36:43 +03:00
|
|
|
dev_err(rdev->dev, "disabling GPU acceleration\n");
|
|
|
|
r700_cp_fini(rdev);
|
2012-12-05 00:25:59 +04:00
|
|
|
r600_dma_fini(rdev);
|
2010-03-24 20:36:43 +03:00
|
|
|
r600_irq_fini(rdev);
|
2013-04-12 21:52:52 +04:00
|
|
|
if (rdev->flags & RADEON_IS_IGP)
|
|
|
|
sumo_rlc_fini(rdev);
|
2010-08-28 02:25:25 +04:00
|
|
|
radeon_wb_fini(rdev);
|
2012-07-05 13:55:34 +04:00
|
|
|
radeon_ib_pool_fini(rdev);
|
2010-03-24 20:36:43 +03:00
|
|
|
radeon_irq_kms_fini(rdev);
|
2010-03-24 20:20:41 +03:00
|
|
|
evergreen_pcie_gart_fini(rdev);
|
2010-01-13 01:54:34 +03:00
|
|
|
rdev->accel_working = false;
|
|
|
|
}
|
2011-12-21 20:58:17 +04:00
|
|
|
|
|
|
|
/* Don't start up if the MC ucode is missing on BTC parts.
|
|
|
|
* The default clocks and voltages before the MC ucode
|
|
|
|
* is loaded are not suffient for advanced operations.
|
|
|
|
*/
|
|
|
|
if (ASIC_IS_DCE5(rdev)) {
|
|
|
|
if (!rdev->mc_fw && !(rdev->flags & RADEON_IS_IGP)) {
|
|
|
|
DRM_ERROR("radeon: MC ucode required for NI+.\n");
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2010-01-13 01:54:34 +03:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
void evergreen_fini(struct radeon_device *rdev)
|
|
|
|
{
|
2013-12-18 23:07:14 +04:00
|
|
|
radeon_pm_fini(rdev);
|
2014-12-04 01:07:01 +03:00
|
|
|
radeon_audio_fini(rdev);
|
2010-03-24 20:55:51 +03:00
|
|
|
r700_cp_fini(rdev);
|
2012-12-05 00:25:59 +04:00
|
|
|
r600_dma_fini(rdev);
|
2010-01-13 01:54:34 +03:00
|
|
|
r600_irq_fini(rdev);
|
2013-04-12 21:52:52 +04:00
|
|
|
if (rdev->flags & RADEON_IS_IGP)
|
|
|
|
sumo_rlc_fini(rdev);
|
2010-08-28 02:25:25 +04:00
|
|
|
radeon_wb_fini(rdev);
|
2012-07-05 13:55:34 +04:00
|
|
|
radeon_ib_pool_fini(rdev);
|
2010-01-13 01:54:34 +03:00
|
|
|
radeon_irq_kms_fini(rdev);
|
2013-08-13 13:56:53 +04:00
|
|
|
uvd_v1_0_fini(rdev);
|
2013-04-08 14:41:29 +04:00
|
|
|
radeon_uvd_fini(rdev);
|
2014-02-27 04:22:47 +04:00
|
|
|
evergreen_pcie_gart_fini(rdev);
|
2011-10-28 18:30:02 +04:00
|
|
|
r600_vram_scratch_fini(rdev);
|
2010-01-13 01:54:34 +03:00
|
|
|
radeon_gem_fini(rdev);
|
|
|
|
radeon_fence_driver_fini(rdev);
|
|
|
|
radeon_agp_fini(rdev);
|
|
|
|
radeon_bo_fini(rdev);
|
|
|
|
radeon_atombios_fini(rdev);
|
|
|
|
kfree(rdev->bios);
|
|
|
|
rdev->bios = NULL;
|
|
|
|
}
|
2011-01-07 02:49:35 +03:00
|
|
|
|
2011-09-20 18:22:58 +04:00
|
|
|
void evergreen_pcie_gen2_enable(struct radeon_device *rdev)
|
2011-01-07 02:49:35 +03:00
|
|
|
{
|
2013-05-04 02:43:13 +04:00
|
|
|
u32 link_width_cntl, speed_cntl;
|
2011-01-07 02:49:35 +03:00
|
|
|
|
2011-01-13 04:05:11 +03:00
|
|
|
if (radeon_pcie_gen2 == 0)
|
|
|
|
return;
|
|
|
|
|
2011-01-07 02:49:35 +03:00
|
|
|
if (rdev->flags & RADEON_IS_IGP)
|
|
|
|
return;
|
|
|
|
|
|
|
|
if (!(rdev->flags & RADEON_IS_PCIE))
|
|
|
|
return;
|
|
|
|
|
|
|
|
/* x2 cards have a special sequence */
|
|
|
|
if (ASIC_IS_X2(rdev))
|
|
|
|
return;
|
|
|
|
|
2013-05-04 02:43:13 +04:00
|
|
|
if ((rdev->pdev->bus->max_bus_speed != PCIE_SPEED_5_0GT) &&
|
|
|
|
(rdev->pdev->bus->max_bus_speed != PCIE_SPEED_8_0GT))
|
2012-06-27 11:35:54 +04:00
|
|
|
return;
|
|
|
|
|
2012-10-26 00:06:59 +04:00
|
|
|
speed_cntl = RREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL);
|
2012-10-09 01:46:27 +04:00
|
|
|
if (speed_cntl & LC_CURRENT_DATA_RATE) {
|
|
|
|
DRM_INFO("PCIE gen 2 link speeds already enabled\n");
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2012-06-27 11:35:54 +04:00
|
|
|
DRM_INFO("enabling PCIE gen 2 link speeds, disable with radeon.pcie_gen2=0\n");
|
|
|
|
|
2011-01-07 02:49:35 +03:00
|
|
|
if ((speed_cntl & LC_OTHER_SIDE_EVER_SENT_GEN2) ||
|
|
|
|
(speed_cntl & LC_OTHER_SIDE_SUPPORTS_GEN2)) {
|
|
|
|
|
2012-10-26 00:06:59 +04:00
|
|
|
link_width_cntl = RREG32_PCIE_PORT(PCIE_LC_LINK_WIDTH_CNTL);
|
2011-01-07 02:49:35 +03:00
|
|
|
link_width_cntl &= ~LC_UPCONFIGURE_DIS;
|
2012-10-26 00:06:59 +04:00
|
|
|
WREG32_PCIE_PORT(PCIE_LC_LINK_WIDTH_CNTL, link_width_cntl);
|
2011-01-07 02:49:35 +03:00
|
|
|
|
2012-10-26 00:06:59 +04:00
|
|
|
speed_cntl = RREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL);
|
2011-01-07 02:49:35 +03:00
|
|
|
speed_cntl &= ~LC_TARGET_LINK_SPEED_OVERRIDE_EN;
|
2012-10-26 00:06:59 +04:00
|
|
|
WREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL, speed_cntl);
|
2011-01-07 02:49:35 +03:00
|
|
|
|
2012-10-26 00:06:59 +04:00
|
|
|
speed_cntl = RREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL);
|
2011-01-07 02:49:35 +03:00
|
|
|
speed_cntl |= LC_CLR_FAILED_SPD_CHANGE_CNT;
|
2012-10-26 00:06:59 +04:00
|
|
|
WREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL, speed_cntl);
|
2011-01-07 02:49:35 +03:00
|
|
|
|
2012-10-26 00:06:59 +04:00
|
|
|
speed_cntl = RREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL);
|
2011-01-07 02:49:35 +03:00
|
|
|
speed_cntl &= ~LC_CLR_FAILED_SPD_CHANGE_CNT;
|
2012-10-26 00:06:59 +04:00
|
|
|
WREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL, speed_cntl);
|
2011-01-07 02:49:35 +03:00
|
|
|
|
2012-10-26 00:06:59 +04:00
|
|
|
speed_cntl = RREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL);
|
2011-01-07 02:49:35 +03:00
|
|
|
speed_cntl |= LC_GEN2_EN_STRAP;
|
2012-10-26 00:06:59 +04:00
|
|
|
WREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL, speed_cntl);
|
2011-01-07 02:49:35 +03:00
|
|
|
|
|
|
|
} else {
|
2012-10-26 00:06:59 +04:00
|
|
|
link_width_cntl = RREG32_PCIE_PORT(PCIE_LC_LINK_WIDTH_CNTL);
|
2011-01-07 02:49:35 +03:00
|
|
|
/* XXX: only disable it if gen1 bridge vendor == 0x111d or 0x1106 */
|
|
|
|
if (1)
|
|
|
|
link_width_cntl |= LC_UPCONFIGURE_DIS;
|
|
|
|
else
|
|
|
|
link_width_cntl &= ~LC_UPCONFIGURE_DIS;
|
2012-10-26 00:06:59 +04:00
|
|
|
WREG32_PCIE_PORT(PCIE_LC_LINK_WIDTH_CNTL, link_width_cntl);
|
2011-01-07 02:49:35 +03:00
|
|
|
}
|
|
|
|
}
|
2013-02-15 20:02:50 +04:00
|
|
|
|
|
|
|
void evergreen_program_aspm(struct radeon_device *rdev)
|
|
|
|
{
|
|
|
|
u32 data, orig;
|
|
|
|
u32 pcie_lc_cntl, pcie_lc_cntl_old;
|
|
|
|
bool disable_l0s, disable_l1 = false, disable_plloff_in_l1 = false;
|
|
|
|
/* fusion_platform = true
|
|
|
|
* if the system is a fusion system
|
|
|
|
* (APU or DGPU in a fusion system).
|
|
|
|
* todo: check if the system is a fusion platform.
|
|
|
|
*/
|
|
|
|
bool fusion_platform = false;
|
|
|
|
|
2013-07-16 23:58:50 +04:00
|
|
|
if (radeon_aspm == 0)
|
|
|
|
return;
|
|
|
|
|
2013-02-15 20:02:50 +04:00
|
|
|
if (!(rdev->flags & RADEON_IS_PCIE))
|
|
|
|
return;
|
|
|
|
|
|
|
|
switch (rdev->family) {
|
|
|
|
case CHIP_CYPRESS:
|
|
|
|
case CHIP_HEMLOCK:
|
|
|
|
case CHIP_JUNIPER:
|
|
|
|
case CHIP_REDWOOD:
|
|
|
|
case CHIP_CEDAR:
|
|
|
|
case CHIP_SUMO:
|
|
|
|
case CHIP_SUMO2:
|
|
|
|
case CHIP_PALM:
|
|
|
|
case CHIP_ARUBA:
|
|
|
|
disable_l0s = true;
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
disable_l0s = false;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (rdev->flags & RADEON_IS_IGP)
|
|
|
|
fusion_platform = true; /* XXX also dGPUs in a fusion system */
|
|
|
|
|
|
|
|
data = orig = RREG32_PIF_PHY0(PB0_PIF_PAIRING);
|
|
|
|
if (fusion_platform)
|
|
|
|
data &= ~MULTI_PIF;
|
|
|
|
else
|
|
|
|
data |= MULTI_PIF;
|
|
|
|
if (data != orig)
|
|
|
|
WREG32_PIF_PHY0(PB0_PIF_PAIRING, data);
|
|
|
|
|
|
|
|
data = orig = RREG32_PIF_PHY1(PB1_PIF_PAIRING);
|
|
|
|
if (fusion_platform)
|
|
|
|
data &= ~MULTI_PIF;
|
|
|
|
else
|
|
|
|
data |= MULTI_PIF;
|
|
|
|
if (data != orig)
|
|
|
|
WREG32_PIF_PHY1(PB1_PIF_PAIRING, data);
|
|
|
|
|
|
|
|
pcie_lc_cntl = pcie_lc_cntl_old = RREG32_PCIE_PORT(PCIE_LC_CNTL);
|
|
|
|
pcie_lc_cntl &= ~(LC_L0S_INACTIVITY_MASK | LC_L1_INACTIVITY_MASK);
|
|
|
|
if (!disable_l0s) {
|
|
|
|
if (rdev->family >= CHIP_BARTS)
|
|
|
|
pcie_lc_cntl |= LC_L0S_INACTIVITY(7);
|
|
|
|
else
|
|
|
|
pcie_lc_cntl |= LC_L0S_INACTIVITY(3);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!disable_l1) {
|
|
|
|
if (rdev->family >= CHIP_BARTS)
|
|
|
|
pcie_lc_cntl |= LC_L1_INACTIVITY(7);
|
|
|
|
else
|
|
|
|
pcie_lc_cntl |= LC_L1_INACTIVITY(8);
|
|
|
|
|
|
|
|
if (!disable_plloff_in_l1) {
|
|
|
|
data = orig = RREG32_PIF_PHY0(PB0_PIF_PWRDOWN_0);
|
|
|
|
data &= ~(PLL_POWER_STATE_IN_OFF_0_MASK | PLL_POWER_STATE_IN_TXS2_0_MASK);
|
|
|
|
data |= PLL_POWER_STATE_IN_OFF_0(7) | PLL_POWER_STATE_IN_TXS2_0(7);
|
|
|
|
if (data != orig)
|
|
|
|
WREG32_PIF_PHY0(PB0_PIF_PWRDOWN_0, data);
|
|
|
|
|
|
|
|
data = orig = RREG32_PIF_PHY0(PB0_PIF_PWRDOWN_1);
|
|
|
|
data &= ~(PLL_POWER_STATE_IN_OFF_1_MASK | PLL_POWER_STATE_IN_TXS2_1_MASK);
|
|
|
|
data |= PLL_POWER_STATE_IN_OFF_1(7) | PLL_POWER_STATE_IN_TXS2_1(7);
|
|
|
|
if (data != orig)
|
|
|
|
WREG32_PIF_PHY0(PB0_PIF_PWRDOWN_1, data);
|
|
|
|
|
|
|
|
data = orig = RREG32_PIF_PHY1(PB1_PIF_PWRDOWN_0);
|
|
|
|
data &= ~(PLL_POWER_STATE_IN_OFF_0_MASK | PLL_POWER_STATE_IN_TXS2_0_MASK);
|
|
|
|
data |= PLL_POWER_STATE_IN_OFF_0(7) | PLL_POWER_STATE_IN_TXS2_0(7);
|
|
|
|
if (data != orig)
|
|
|
|
WREG32_PIF_PHY1(PB1_PIF_PWRDOWN_0, data);
|
|
|
|
|
|
|
|
data = orig = RREG32_PIF_PHY1(PB1_PIF_PWRDOWN_1);
|
|
|
|
data &= ~(PLL_POWER_STATE_IN_OFF_1_MASK | PLL_POWER_STATE_IN_TXS2_1_MASK);
|
|
|
|
data |= PLL_POWER_STATE_IN_OFF_1(7) | PLL_POWER_STATE_IN_TXS2_1(7);
|
|
|
|
if (data != orig)
|
|
|
|
WREG32_PIF_PHY1(PB1_PIF_PWRDOWN_1, data);
|
|
|
|
|
|
|
|
if (rdev->family >= CHIP_BARTS) {
|
|
|
|
data = orig = RREG32_PIF_PHY0(PB0_PIF_PWRDOWN_0);
|
|
|
|
data &= ~PLL_RAMP_UP_TIME_0_MASK;
|
|
|
|
data |= PLL_RAMP_UP_TIME_0(4);
|
|
|
|
if (data != orig)
|
|
|
|
WREG32_PIF_PHY0(PB0_PIF_PWRDOWN_0, data);
|
|
|
|
|
|
|
|
data = orig = RREG32_PIF_PHY0(PB0_PIF_PWRDOWN_1);
|
|
|
|
data &= ~PLL_RAMP_UP_TIME_1_MASK;
|
|
|
|
data |= PLL_RAMP_UP_TIME_1(4);
|
|
|
|
if (data != orig)
|
|
|
|
WREG32_PIF_PHY0(PB0_PIF_PWRDOWN_1, data);
|
|
|
|
|
|
|
|
data = orig = RREG32_PIF_PHY1(PB1_PIF_PWRDOWN_0);
|
|
|
|
data &= ~PLL_RAMP_UP_TIME_0_MASK;
|
|
|
|
data |= PLL_RAMP_UP_TIME_0(4);
|
|
|
|
if (data != orig)
|
|
|
|
WREG32_PIF_PHY1(PB1_PIF_PWRDOWN_0, data);
|
|
|
|
|
|
|
|
data = orig = RREG32_PIF_PHY1(PB1_PIF_PWRDOWN_1);
|
|
|
|
data &= ~PLL_RAMP_UP_TIME_1_MASK;
|
|
|
|
data |= PLL_RAMP_UP_TIME_1(4);
|
|
|
|
if (data != orig)
|
|
|
|
WREG32_PIF_PHY1(PB1_PIF_PWRDOWN_1, data);
|
|
|
|
}
|
|
|
|
|
|
|
|
data = orig = RREG32_PCIE_PORT(PCIE_LC_LINK_WIDTH_CNTL);
|
|
|
|
data &= ~LC_DYN_LANES_PWR_STATE_MASK;
|
|
|
|
data |= LC_DYN_LANES_PWR_STATE(3);
|
|
|
|
if (data != orig)
|
|
|
|
WREG32_PCIE_PORT(PCIE_LC_LINK_WIDTH_CNTL, data);
|
|
|
|
|
|
|
|
if (rdev->family >= CHIP_BARTS) {
|
|
|
|
data = orig = RREG32_PIF_PHY0(PB0_PIF_CNTL);
|
|
|
|
data &= ~LS2_EXIT_TIME_MASK;
|
|
|
|
data |= LS2_EXIT_TIME(1);
|
|
|
|
if (data != orig)
|
|
|
|
WREG32_PIF_PHY0(PB0_PIF_CNTL, data);
|
|
|
|
|
|
|
|
data = orig = RREG32_PIF_PHY1(PB1_PIF_CNTL);
|
|
|
|
data &= ~LS2_EXIT_TIME_MASK;
|
|
|
|
data |= LS2_EXIT_TIME(1);
|
|
|
|
if (data != orig)
|
|
|
|
WREG32_PIF_PHY1(PB1_PIF_CNTL, data);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* evergreen parts only */
|
|
|
|
if (rdev->family < CHIP_BARTS)
|
|
|
|
pcie_lc_cntl |= LC_PMI_TO_L1_DIS;
|
|
|
|
|
|
|
|
if (pcie_lc_cntl != pcie_lc_cntl_old)
|
|
|
|
WREG32_PCIE_PORT(PCIE_LC_CNTL, pcie_lc_cntl);
|
|
|
|
}
|