303 lines
7.0 KiB
ArmAsm
303 lines
7.0 KiB
ArmAsm
/*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License, version 2, as
|
|
* published by the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program; if not, write to the Free Software
|
|
* Foundation, 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
|
|
*
|
|
* Copyright SUSE Linux Products GmbH 2009
|
|
*
|
|
* Authors: Alexander Graf <agraf@suse.de>
|
|
*/
|
|
|
|
#define SHADOW_SLB_ESID(num) (SLBSHADOW_SAVEAREA + (num * 0x10))
|
|
#define SHADOW_SLB_VSID(num) (SLBSHADOW_SAVEAREA + (num * 0x10) + 0x8)
|
|
#define UNBOLT_SLB_ENTRY(num) \
|
|
ld r9, SHADOW_SLB_ESID(num)(r12); \
|
|
/* Invalid? Skip. */; \
|
|
rldicl. r0, r9, 37, 63; \
|
|
beq slb_entry_skip_ ## num; \
|
|
xoris r9, r9, SLB_ESID_V@h; \
|
|
std r9, SHADOW_SLB_ESID(num)(r12); \
|
|
slb_entry_skip_ ## num:
|
|
|
|
#define REBOLT_SLB_ENTRY(num) \
|
|
ld r10, SHADOW_SLB_ESID(num)(r11); \
|
|
cmpdi r10, 0; \
|
|
beq slb_exit_skip_ ## num; \
|
|
oris r10, r10, SLB_ESID_V@h; \
|
|
ld r9, SHADOW_SLB_VSID(num)(r11); \
|
|
slbmte r9, r10; \
|
|
std r10, SHADOW_SLB_ESID(num)(r11); \
|
|
slb_exit_skip_ ## num:
|
|
|
|
/******************************************************************************
|
|
* *
|
|
* Entry code *
|
|
* *
|
|
*****************************************************************************/
|
|
|
|
.global kvmppc_handler_trampoline_enter
|
|
kvmppc_handler_trampoline_enter:
|
|
|
|
/* Required state:
|
|
*
|
|
* MSR = ~IR|DR
|
|
* R13 = PACA
|
|
* R1 = host R1
|
|
* R2 = host R2
|
|
* R9 = guest IP
|
|
* R10 = guest MSR
|
|
* all other GPRS = free
|
|
* PACA[KVM_CR] = guest CR
|
|
* PACA[KVM_XER] = guest XER
|
|
*/
|
|
|
|
mtsrr0 r9
|
|
mtsrr1 r10
|
|
|
|
/* Activate guest mode, so faults get handled by KVM */
|
|
li r11, KVM_GUEST_MODE_GUEST
|
|
stb r11, PACA_KVM_IN_GUEST(r13)
|
|
|
|
/* Remove LPAR shadow entries */
|
|
|
|
#if SLB_NUM_BOLTED == 3
|
|
|
|
ld r12, PACA_SLBSHADOWPTR(r13)
|
|
|
|
/* Save off the first entry so we can slbie it later */
|
|
ld r10, SHADOW_SLB_ESID(0)(r12)
|
|
ld r11, SHADOW_SLB_VSID(0)(r12)
|
|
|
|
/* Remove bolted entries */
|
|
UNBOLT_SLB_ENTRY(0)
|
|
UNBOLT_SLB_ENTRY(1)
|
|
UNBOLT_SLB_ENTRY(2)
|
|
|
|
#else
|
|
#error unknown number of bolted entries
|
|
#endif
|
|
|
|
/* Flush SLB */
|
|
|
|
slbia
|
|
|
|
/* r0 = esid & ESID_MASK */
|
|
rldicr r10, r10, 0, 35
|
|
/* r0 |= CLASS_BIT(VSID) */
|
|
rldic r12, r11, 56 - 36, 36
|
|
or r10, r10, r12
|
|
slbie r10
|
|
|
|
isync
|
|
|
|
/* Fill SLB with our shadow */
|
|
|
|
lbz r12, PACA_KVM_SLB_MAX(r13)
|
|
mulli r12, r12, 16
|
|
addi r12, r12, PACA_KVM_SLB
|
|
add r12, r12, r13
|
|
|
|
/* for (r11 = kvm_slb; r11 < kvm_slb + kvm_slb_size; r11+=slb_entry) */
|
|
li r11, PACA_KVM_SLB
|
|
add r11, r11, r13
|
|
|
|
slb_loop_enter:
|
|
|
|
ld r10, 0(r11)
|
|
|
|
rldicl. r0, r10, 37, 63
|
|
beq slb_loop_enter_skip
|
|
|
|
ld r9, 8(r11)
|
|
slbmte r9, r10
|
|
|
|
slb_loop_enter_skip:
|
|
addi r11, r11, 16
|
|
cmpd cr0, r11, r12
|
|
blt slb_loop_enter
|
|
|
|
slb_do_enter:
|
|
|
|
/* Enter guest */
|
|
|
|
ld r0, (PACA_KVM_R0)(r13)
|
|
ld r1, (PACA_KVM_R1)(r13)
|
|
ld r2, (PACA_KVM_R2)(r13)
|
|
ld r3, (PACA_KVM_R3)(r13)
|
|
ld r4, (PACA_KVM_R4)(r13)
|
|
ld r5, (PACA_KVM_R5)(r13)
|
|
ld r6, (PACA_KVM_R6)(r13)
|
|
ld r7, (PACA_KVM_R7)(r13)
|
|
ld r8, (PACA_KVM_R8)(r13)
|
|
ld r9, (PACA_KVM_R9)(r13)
|
|
ld r10, (PACA_KVM_R10)(r13)
|
|
ld r12, (PACA_KVM_R12)(r13)
|
|
|
|
lwz r11, (PACA_KVM_CR)(r13)
|
|
mtcr r11
|
|
|
|
ld r11, (PACA_KVM_XER)(r13)
|
|
mtxer r11
|
|
|
|
ld r11, (PACA_KVM_R11)(r13)
|
|
ld r13, (PACA_KVM_R13)(r13)
|
|
|
|
RFI
|
|
kvmppc_handler_trampoline_enter_end:
|
|
|
|
|
|
|
|
/******************************************************************************
|
|
* *
|
|
* Exit code *
|
|
* *
|
|
*****************************************************************************/
|
|
|
|
.global kvmppc_handler_trampoline_exit
|
|
kvmppc_handler_trampoline_exit:
|
|
|
|
/* Register usage at this point:
|
|
*
|
|
* SPRG_SCRATCH0 = guest R13
|
|
* R12 = exit handler id
|
|
* R13 = PACA
|
|
* PACA.KVM.SCRATCH0 = guest R12
|
|
* PACA.KVM.SCRATCH1 = guest CR
|
|
*
|
|
*/
|
|
|
|
/* Save registers */
|
|
|
|
std r0, PACA_KVM_R0(r13)
|
|
std r1, PACA_KVM_R1(r13)
|
|
std r2, PACA_KVM_R2(r13)
|
|
std r3, PACA_KVM_R3(r13)
|
|
std r4, PACA_KVM_R4(r13)
|
|
std r5, PACA_KVM_R5(r13)
|
|
std r6, PACA_KVM_R6(r13)
|
|
std r7, PACA_KVM_R7(r13)
|
|
std r8, PACA_KVM_R8(r13)
|
|
std r9, PACA_KVM_R9(r13)
|
|
std r10, PACA_KVM_R10(r13)
|
|
std r11, PACA_KVM_R11(r13)
|
|
|
|
/* Restore R1/R2 so we can handle faults */
|
|
ld r1, PACA_KVM_HOST_R1(r13)
|
|
ld r2, PACA_KVM_HOST_R2(r13)
|
|
|
|
/* Save guest PC and MSR in GPRs */
|
|
mfsrr0 r3
|
|
mfsrr1 r4
|
|
|
|
/* Get scratch'ed off registers */
|
|
mfspr r9, SPRN_SPRG_SCRATCH0
|
|
std r9, PACA_KVM_R13(r13)
|
|
|
|
ld r8, PACA_KVM_SCRATCH0(r13)
|
|
std r8, PACA_KVM_R12(r13)
|
|
|
|
lwz r7, PACA_KVM_SCRATCH1(r13)
|
|
stw r7, PACA_KVM_CR(r13)
|
|
|
|
/* Save more register state */
|
|
|
|
mfxer r6
|
|
stw r6, PACA_KVM_XER(r13)
|
|
|
|
mfdar r5
|
|
mfdsisr r6
|
|
|
|
/*
|
|
* In order for us to easily get the last instruction,
|
|
* we got the #vmexit at, we exploit the fact that the
|
|
* virtual layout is still the same here, so we can just
|
|
* ld from the guest's PC address
|
|
*/
|
|
|
|
/* We only load the last instruction when it's safe */
|
|
cmpwi r12, BOOK3S_INTERRUPT_DATA_STORAGE
|
|
beq ld_last_inst
|
|
cmpwi r12, BOOK3S_INTERRUPT_PROGRAM
|
|
beq ld_last_inst
|
|
|
|
b no_ld_last_inst
|
|
|
|
ld_last_inst:
|
|
/* Save off the guest instruction we're at */
|
|
|
|
/* Set guest mode to 'jump over instruction' so if lwz faults
|
|
* we'll just continue at the next IP. */
|
|
li r9, KVM_GUEST_MODE_SKIP
|
|
stb r9, PACA_KVM_IN_GUEST(r13)
|
|
|
|
/* 1) enable paging for data */
|
|
mfmsr r9
|
|
ori r11, r9, MSR_DR /* Enable paging for data */
|
|
mtmsr r11
|
|
/* 2) fetch the instruction */
|
|
li r0, KVM_INST_FETCH_FAILED /* In case lwz faults */
|
|
lwz r0, 0(r3)
|
|
/* 3) disable paging again */
|
|
mtmsr r9
|
|
|
|
no_ld_last_inst:
|
|
|
|
/* Unset guest mode */
|
|
li r9, KVM_GUEST_MODE_NONE
|
|
stb r9, PACA_KVM_IN_GUEST(r13)
|
|
|
|
/* Restore bolted entries from the shadow and fix it along the way */
|
|
|
|
/* We don't store anything in entry 0, so we don't need to take care of it */
|
|
slbia
|
|
isync
|
|
|
|
#if SLB_NUM_BOLTED == 3
|
|
|
|
ld r11, PACA_SLBSHADOWPTR(r13)
|
|
|
|
REBOLT_SLB_ENTRY(0)
|
|
REBOLT_SLB_ENTRY(1)
|
|
REBOLT_SLB_ENTRY(2)
|
|
|
|
#else
|
|
#error unknown number of bolted entries
|
|
#endif
|
|
|
|
slb_do_exit:
|
|
|
|
/* Register usage at this point:
|
|
*
|
|
* R0 = guest last inst
|
|
* R1 = host R1
|
|
* R2 = host R2
|
|
* R3 = guest PC
|
|
* R4 = guest MSR
|
|
* R5 = guest DAR
|
|
* R6 = guest DSISR
|
|
* R12 = exit handler id
|
|
* R13 = PACA
|
|
* PACA.KVM.* = guest *
|
|
*
|
|
*/
|
|
|
|
/* RFI into the highmem handler */
|
|
mfmsr r7
|
|
ori r7, r7, MSR_IR|MSR_DR|MSR_RI /* Enable paging */
|
|
mtsrr1 r7
|
|
ld r8, PACA_KVM_VMHANDLER(r13) /* Highmem handler address */
|
|
mtsrr0 r8
|
|
|
|
RFI
|
|
kvmppc_handler_trampoline_exit_end:
|
|
|