/* * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License, version 2, as * published by the Free Software Foundation. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, write to the Free Software * Foundation, 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. * * Copyright SUSE Linux Products GmbH 2009 * * Authors: Alexander Graf */ #define SHADOW_SLB_ESID(num) (SLBSHADOW_SAVEAREA + (num * 0x10)) #define SHADOW_SLB_VSID(num) (SLBSHADOW_SAVEAREA + (num * 0x10) + 0x8) #define UNBOLT_SLB_ENTRY(num) \ ld r9, SHADOW_SLB_ESID(num)(r12); \ /* Invalid? Skip. */; \ rldicl. r0, r9, 37, 63; \ beq slb_entry_skip_ ## num; \ xoris r9, r9, SLB_ESID_V@h; \ std r9, SHADOW_SLB_ESID(num)(r12); \ slb_entry_skip_ ## num: #define REBOLT_SLB_ENTRY(num) \ ld r10, SHADOW_SLB_ESID(num)(r11); \ cmpdi r10, 0; \ beq slb_exit_skip_ ## num; \ oris r10, r10, SLB_ESID_V@h; \ ld r9, SHADOW_SLB_VSID(num)(r11); \ slbmte r9, r10; \ std r10, SHADOW_SLB_ESID(num)(r11); \ slb_exit_skip_ ## num: /****************************************************************************** * * * Entry code * * * *****************************************************************************/ .global kvmppc_handler_trampoline_enter kvmppc_handler_trampoline_enter: /* Required state: * * MSR = ~IR|DR * R13 = PACA * R1 = host R1 * R2 = host R2 * R9 = guest IP * R10 = guest MSR * all other GPRS = free * PACA[KVM_CR] = guest CR * PACA[KVM_XER] = guest XER */ mtsrr0 r9 mtsrr1 r10 /* Activate guest mode, so faults get handled by KVM */ li r11, KVM_GUEST_MODE_GUEST stb r11, PACA_KVM_IN_GUEST(r13) /* Remove LPAR shadow entries */ #if SLB_NUM_BOLTED == 3 ld r12, PACA_SLBSHADOWPTR(r13) /* Save off the first entry so we can slbie it later */ ld r10, SHADOW_SLB_ESID(0)(r12) ld r11, SHADOW_SLB_VSID(0)(r12) /* Remove bolted entries */ UNBOLT_SLB_ENTRY(0) UNBOLT_SLB_ENTRY(1) UNBOLT_SLB_ENTRY(2) #else #error unknown number of bolted entries #endif /* Flush SLB */ slbia /* r0 = esid & ESID_MASK */ rldicr r10, r10, 0, 35 /* r0 |= CLASS_BIT(VSID) */ rldic r12, r11, 56 - 36, 36 or r10, r10, r12 slbie r10 isync /* Fill SLB with our shadow */ lbz r12, PACA_KVM_SLB_MAX(r13) mulli r12, r12, 16 addi r12, r12, PACA_KVM_SLB add r12, r12, r13 /* for (r11 = kvm_slb; r11 < kvm_slb + kvm_slb_size; r11+=slb_entry) */ li r11, PACA_KVM_SLB add r11, r11, r13 slb_loop_enter: ld r10, 0(r11) rldicl. r0, r10, 37, 63 beq slb_loop_enter_skip ld r9, 8(r11) slbmte r9, r10 slb_loop_enter_skip: addi r11, r11, 16 cmpd cr0, r11, r12 blt slb_loop_enter slb_do_enter: /* Enter guest */ ld r0, (PACA_KVM_R0)(r13) ld r1, (PACA_KVM_R1)(r13) ld r2, (PACA_KVM_R2)(r13) ld r3, (PACA_KVM_R3)(r13) ld r4, (PACA_KVM_R4)(r13) ld r5, (PACA_KVM_R5)(r13) ld r6, (PACA_KVM_R6)(r13) ld r7, (PACA_KVM_R7)(r13) ld r8, (PACA_KVM_R8)(r13) ld r9, (PACA_KVM_R9)(r13) ld r10, (PACA_KVM_R10)(r13) ld r12, (PACA_KVM_R12)(r13) lwz r11, (PACA_KVM_CR)(r13) mtcr r11 ld r11, (PACA_KVM_XER)(r13) mtxer r11 ld r11, (PACA_KVM_R11)(r13) ld r13, (PACA_KVM_R13)(r13) RFI kvmppc_handler_trampoline_enter_end: /****************************************************************************** * * * Exit code * * * *****************************************************************************/ .global kvmppc_handler_trampoline_exit kvmppc_handler_trampoline_exit: /* Register usage at this point: * * SPRG_SCRATCH0 = guest R13 * R12 = exit handler id * R13 = PACA * PACA.KVM.SCRATCH0 = guest R12 * PACA.KVM.SCRATCH1 = guest CR * */ /* Save registers */ std r0, PACA_KVM_R0(r13) std r1, PACA_KVM_R1(r13) std r2, PACA_KVM_R2(r13) std r3, PACA_KVM_R3(r13) std r4, PACA_KVM_R4(r13) std r5, PACA_KVM_R5(r13) std r6, PACA_KVM_R6(r13) std r7, PACA_KVM_R7(r13) std r8, PACA_KVM_R8(r13) std r9, PACA_KVM_R9(r13) std r10, PACA_KVM_R10(r13) std r11, PACA_KVM_R11(r13) /* Restore R1/R2 so we can handle faults */ ld r1, PACA_KVM_HOST_R1(r13) ld r2, PACA_KVM_HOST_R2(r13) /* Save guest PC and MSR in GPRs */ mfsrr0 r3 mfsrr1 r4 /* Get scratch'ed off registers */ mfspr r9, SPRN_SPRG_SCRATCH0 std r9, PACA_KVM_R13(r13) ld r8, PACA_KVM_SCRATCH0(r13) std r8, PACA_KVM_R12(r13) lwz r7, PACA_KVM_SCRATCH1(r13) stw r7, PACA_KVM_CR(r13) /* Save more register state */ mfxer r6 stw r6, PACA_KVM_XER(r13) mfdar r5 mfdsisr r6 /* * In order for us to easily get the last instruction, * we got the #vmexit at, we exploit the fact that the * virtual layout is still the same here, so we can just * ld from the guest's PC address */ /* We only load the last instruction when it's safe */ cmpwi r12, BOOK3S_INTERRUPT_DATA_STORAGE beq ld_last_inst cmpwi r12, BOOK3S_INTERRUPT_PROGRAM beq ld_last_inst b no_ld_last_inst ld_last_inst: /* Save off the guest instruction we're at */ /* Set guest mode to 'jump over instruction' so if lwz faults * we'll just continue at the next IP. */ li r9, KVM_GUEST_MODE_SKIP stb r9, PACA_KVM_IN_GUEST(r13) /* 1) enable paging for data */ mfmsr r9 ori r11, r9, MSR_DR /* Enable paging for data */ mtmsr r11 /* 2) fetch the instruction */ li r0, KVM_INST_FETCH_FAILED /* In case lwz faults */ lwz r0, 0(r3) /* 3) disable paging again */ mtmsr r9 no_ld_last_inst: /* Unset guest mode */ li r9, KVM_GUEST_MODE_NONE stb r9, PACA_KVM_IN_GUEST(r13) /* Restore bolted entries from the shadow and fix it along the way */ /* We don't store anything in entry 0, so we don't need to take care of it */ slbia isync #if SLB_NUM_BOLTED == 3 ld r11, PACA_SLBSHADOWPTR(r13) REBOLT_SLB_ENTRY(0) REBOLT_SLB_ENTRY(1) REBOLT_SLB_ENTRY(2) #else #error unknown number of bolted entries #endif slb_do_exit: /* Register usage at this point: * * R0 = guest last inst * R1 = host R1 * R2 = host R2 * R3 = guest PC * R4 = guest MSR * R5 = guest DAR * R6 = guest DSISR * R12 = exit handler id * R13 = PACA * PACA.KVM.* = guest * * */ /* RFI into the highmem handler */ mfmsr r7 ori r7, r7, MSR_IR|MSR_DR|MSR_RI /* Enable paging */ mtsrr1 r7 ld r8, PACA_KVM_VMHANDLER(r13) /* Highmem handler address */ mtsrr0 r8 RFI kvmppc_handler_trampoline_exit_end: