powerpc/e500: SPE register saving: take arbitrary struct offset
Previously, these macros hardcoded THREAD_EVR0 as the base of the save area, relative to the base register passed. This base offset is now passed as a separate macro parameter, allowing reuse with other SPE save areas, such as used by KVM. Acked-by: Kumar Gala <galak@kernel.crashing.org> Signed-off-by: Scott Wood <scottwood@freescale.com> Signed-off-by: Alexander Graf <agraf@suse.de>
This commit is contained in:
Родитель
685659ee70
Коммит
c51584d52e
|
@ -150,18 +150,22 @@ END_FW_FTR_SECTION_IFSET(FW_FEATURE_SPLPAR)
|
||||||
#define REST_16VSRSU(n,b,base) REST_8VSRSU(n,b,base); REST_8VSRSU(n+8,b,base)
|
#define REST_16VSRSU(n,b,base) REST_8VSRSU(n,b,base); REST_8VSRSU(n+8,b,base)
|
||||||
#define REST_32VSRSU(n,b,base) REST_16VSRSU(n,b,base); REST_16VSRSU(n+16,b,base)
|
#define REST_32VSRSU(n,b,base) REST_16VSRSU(n,b,base); REST_16VSRSU(n+16,b,base)
|
||||||
|
|
||||||
#define SAVE_EVR(n,s,base) evmergehi s,s,n; stw s,THREAD_EVR0+4*(n)(base)
|
/*
|
||||||
#define SAVE_2EVRS(n,s,base) SAVE_EVR(n,s,base); SAVE_EVR(n+1,s,base)
|
* b = base register for addressing, o = base offset from register of 1st EVR
|
||||||
#define SAVE_4EVRS(n,s,base) SAVE_2EVRS(n,s,base); SAVE_2EVRS(n+2,s,base)
|
* n = first EVR, s = scratch
|
||||||
#define SAVE_8EVRS(n,s,base) SAVE_4EVRS(n,s,base); SAVE_4EVRS(n+4,s,base)
|
*/
|
||||||
#define SAVE_16EVRS(n,s,base) SAVE_8EVRS(n,s,base); SAVE_8EVRS(n+8,s,base)
|
#define SAVE_EVR(n,s,b,o) evmergehi s,s,n; stw s,o+4*(n)(b)
|
||||||
#define SAVE_32EVRS(n,s,base) SAVE_16EVRS(n,s,base); SAVE_16EVRS(n+16,s,base)
|
#define SAVE_2EVRS(n,s,b,o) SAVE_EVR(n,s,b,o); SAVE_EVR(n+1,s,b,o)
|
||||||
#define REST_EVR(n,s,base) lwz s,THREAD_EVR0+4*(n)(base); evmergelo n,s,n
|
#define SAVE_4EVRS(n,s,b,o) SAVE_2EVRS(n,s,b,o); SAVE_2EVRS(n+2,s,b,o)
|
||||||
#define REST_2EVRS(n,s,base) REST_EVR(n,s,base); REST_EVR(n+1,s,base)
|
#define SAVE_8EVRS(n,s,b,o) SAVE_4EVRS(n,s,b,o); SAVE_4EVRS(n+4,s,b,o)
|
||||||
#define REST_4EVRS(n,s,base) REST_2EVRS(n,s,base); REST_2EVRS(n+2,s,base)
|
#define SAVE_16EVRS(n,s,b,o) SAVE_8EVRS(n,s,b,o); SAVE_8EVRS(n+8,s,b,o)
|
||||||
#define REST_8EVRS(n,s,base) REST_4EVRS(n,s,base); REST_4EVRS(n+4,s,base)
|
#define SAVE_32EVRS(n,s,b,o) SAVE_16EVRS(n,s,b,o); SAVE_16EVRS(n+16,s,b,o)
|
||||||
#define REST_16EVRS(n,s,base) REST_8EVRS(n,s,base); REST_8EVRS(n+8,s,base)
|
#define REST_EVR(n,s,b,o) lwz s,o+4*(n)(b); evmergelo n,s,n
|
||||||
#define REST_32EVRS(n,s,base) REST_16EVRS(n,s,base); REST_16EVRS(n+16,s,base)
|
#define REST_2EVRS(n,s,b,o) REST_EVR(n,s,b,o); REST_EVR(n+1,s,b,o)
|
||||||
|
#define REST_4EVRS(n,s,b,o) REST_2EVRS(n,s,b,o); REST_2EVRS(n+2,s,b,o)
|
||||||
|
#define REST_8EVRS(n,s,b,o) REST_4EVRS(n,s,b,o); REST_4EVRS(n+4,s,b,o)
|
||||||
|
#define REST_16EVRS(n,s,b,o) REST_8EVRS(n,s,b,o); REST_8EVRS(n+8,s,b,o)
|
||||||
|
#define REST_32EVRS(n,s,b,o) REST_16EVRS(n,s,b,o); REST_16EVRS(n+16,s,b,o)
|
||||||
|
|
||||||
/* Macros to adjust thread priority for hardware multithreading */
|
/* Macros to adjust thread priority for hardware multithreading */
|
||||||
#define HMT_VERY_LOW or 31,31,31 # very low priority
|
#define HMT_VERY_LOW or 31,31,31 # very low priority
|
||||||
|
|
|
@ -656,7 +656,7 @@ load_up_spe:
|
||||||
cmpi 0,r4,0
|
cmpi 0,r4,0
|
||||||
beq 1f
|
beq 1f
|
||||||
addi r4,r4,THREAD /* want THREAD of last_task_used_spe */
|
addi r4,r4,THREAD /* want THREAD of last_task_used_spe */
|
||||||
SAVE_32EVRS(0,r10,r4)
|
SAVE_32EVRS(0,r10,r4,THREAD_EVR0)
|
||||||
evxor evr10, evr10, evr10 /* clear out evr10 */
|
evxor evr10, evr10, evr10 /* clear out evr10 */
|
||||||
evmwumiaa evr10, evr10, evr10 /* evr10 <- ACC = 0 * 0 + ACC */
|
evmwumiaa evr10, evr10, evr10 /* evr10 <- ACC = 0 * 0 + ACC */
|
||||||
li r5,THREAD_ACC
|
li r5,THREAD_ACC
|
||||||
|
@ -676,7 +676,7 @@ load_up_spe:
|
||||||
stw r4,THREAD_USED_SPE(r5)
|
stw r4,THREAD_USED_SPE(r5)
|
||||||
evlddx evr4,r10,r5
|
evlddx evr4,r10,r5
|
||||||
evmra evr4,evr4
|
evmra evr4,evr4
|
||||||
REST_32EVRS(0,r10,r5)
|
REST_32EVRS(0,r10,r5,THREAD_EVR0)
|
||||||
#ifndef CONFIG_SMP
|
#ifndef CONFIG_SMP
|
||||||
subi r4,r5,THREAD
|
subi r4,r5,THREAD
|
||||||
stw r4,last_task_used_spe@l(r3)
|
stw r4,last_task_used_spe@l(r3)
|
||||||
|
@ -787,7 +787,7 @@ _GLOBAL(giveup_spe)
|
||||||
addi r3,r3,THREAD /* want THREAD of task */
|
addi r3,r3,THREAD /* want THREAD of task */
|
||||||
lwz r5,PT_REGS(r3)
|
lwz r5,PT_REGS(r3)
|
||||||
cmpi 0,r5,0
|
cmpi 0,r5,0
|
||||||
SAVE_32EVRS(0, r4, r3)
|
SAVE_32EVRS(0, r4, r3, THREAD_EVR0)
|
||||||
evxor evr6, evr6, evr6 /* clear out evr6 */
|
evxor evr6, evr6, evr6 /* clear out evr6 */
|
||||||
evmwumiaa evr6, evr6, evr6 /* evr6 <- ACC = 0 * 0 + ACC */
|
evmwumiaa evr6, evr6, evr6 /* evr6 <- ACC = 0 * 0 + ACC */
|
||||||
li r4,THREAD_ACC
|
li r4,THREAD_ACC
|
||||||
|
|
Загрузка…
Ссылка в новой задаче