diff options
Diffstat (limited to 'src/VBox/VMM/VMMR3/VMMSwitcher.cpp')
-rw-r--r-- | src/VBox/VMM/VMMR3/VMMSwitcher.cpp | 1188 |
1 files changed, 1188 insertions, 0 deletions
diff --git a/src/VBox/VMM/VMMR3/VMMSwitcher.cpp b/src/VBox/VMM/VMMR3/VMMSwitcher.cpp new file mode 100644 index 00000000..ac15c7ca --- /dev/null +++ b/src/VBox/VMM/VMMR3/VMMSwitcher.cpp @@ -0,0 +1,1188 @@ +/* $Id: VMMSwitcher.cpp $ */ +/** @file + * VMM - The Virtual Machine Monitor, World Switcher(s). + */ + +/* + * Copyright (C) 2006-2019 Oracle Corporation + * + * This file is part of VirtualBox Open Source Edition (OSE), as + * available from http://www.virtualbox.org. This file is free software; + * you can redistribute it and/or modify it under the terms of the GNU + * General Public License (GPL) as published by the Free Software + * Foundation, in version 2 as it comes in the "COPYING" file of the + * VirtualBox OSE distribution. VirtualBox OSE is distributed in the + * hope that it will be useful, but WITHOUT ANY WARRANTY of any kind. + */ + + +/********************************************************************************************************************************* +* Header Files * +*********************************************************************************************************************************/ +#define LOG_GROUP LOG_GROUP_VMM +#include <VBox/vmm/vmm.h> +#include <VBox/vmm/pgm.h> +#include <VBox/vmm/hm.h> +#include <VBox/vmm/selm.h> +#include <VBox/vmm/mm.h> +#include <VBox/sup.h> +#include "VMMInternal.h" +#include "VMMSwitcher.h" +#include <VBox/vmm/vm.h> +#include <VBox/dis.h> + +#include <VBox/err.h> +#include <VBox/param.h> +#include <iprt/assert.h> +#include <iprt/alloc.h> +#include <iprt/asm.h> +#include <iprt/asm-amd64-x86.h> +#include <iprt/string.h> +#include <iprt/ctype.h> + + +/********************************************************************************************************************************* +* Global Variables * +*********************************************************************************************************************************/ +#if defined(VBOX_WITH_RAW_MODE) || HC_ARCH_BITS != 64 + +/** Array of switcher definitions. + * The type and index shall match! + */ +static PVMMSWITCHERDEF g_apRawModeSwitchers[VMMSWITCHER_MAX] = +{ + NULL, /* invalid entry */ +# ifdef VBOX_WITH_RAW_MODE +# ifndef RT_ARCH_AMD64 + &vmmR3Switcher32BitTo32Bit_Def, + &vmmR3Switcher32BitToPAE_Def, + NULL, //&vmmR3Switcher32BitToAMD64_Def, + &vmmR3SwitcherPAETo32Bit_Def, + &vmmR3SwitcherPAEToPAE_Def, + NULL, //&vmmR3SwitcherPAEToAMD64_Def, + NULL, //&vmmR3SwitcherPAETo32Bit_Def, + NULL, //&vmmR3SwitcherAMD64ToPAE_Def, + NULL, //&vmmR3SwitcherAMD64ToAMD64_Def, +# else /* RT_ARCH_AMD64 */ + NULL, //&vmmR3Switcher32BitTo32Bit_Def, + NULL, //&vmmR3Switcher32BitToPAE_Def, + NULL, //&vmmR3Switcher32BitToAMD64_Def, + NULL, //&vmmR3SwitcherPAETo32Bit_Def, + NULL, //&vmmR3SwitcherPAEToPAE_Def, + NULL, //&vmmR3SwitcherPAEToAMD64_Def, + &vmmR3SwitcherAMD64To32Bit_Def, + &vmmR3SwitcherAMD64ToPAE_Def, + NULL, //&vmmR3SwitcherAMD64ToAMD64_Def, +# endif /* RT_ARCH_AMD64 */ +# else /* !VBOX_WITH_RAW_MODE */ + NULL, + NULL, + NULL, + NULL, + NULL, + NULL, + NULL, + NULL, + NULL, +# endif /* !VBOX_WITH_RAW_MODE */ +# ifndef RT_ARCH_AMD64 + &vmmR3SwitcherX86Stub_Def, + NULL, +# else + NULL, + &vmmR3SwitcherAMD64Stub_Def, +# endif +}; + +/** Array of switcher definitions. + * The type and index shall match! + */ +static PVMMSWITCHERDEF g_apHmSwitchers[VMMSWITCHER_MAX] = +{ + NULL, /* invalid entry */ +# if HC_ARCH_BITS == 32 + NULL, //&vmmR3Switcher32BitTo32Bit_Def, + NULL, //&vmmR3Switcher32BitToPAE_Def, + &vmmR3Switcher32BitToAMD64_Def, + NULL, //&vmmR3SwitcherPAETo32Bit_Def, + NULL, //&vmmR3SwitcherPAEToPAE_Def, + &vmmR3SwitcherPAEToAMD64_Def, + NULL, //&vmmR3SwitcherPAETo32Bit_Def, + NULL, //&vmmR3SwitcherAMD64ToPAE_Def, + NULL, //&vmmR3SwitcherAMD64ToAMD64_Def, +# else /* !VBOX_WITH_RAW_MODE */ + NULL, + NULL, + NULL, + NULL, + NULL, + NULL, + NULL, + NULL, + NULL, +# endif /* !VBOX_WITH_RAW_MODE */ +# ifndef RT_ARCH_AMD64 + &vmmR3SwitcherX86Stub_Def, + NULL, +# else + NULL, + &vmmR3SwitcherAMD64Stub_Def, +# endif +}; + +#endif /* VBOX_WITH_RAW_MODE || HC_ARCH_BITS != 64 */ + + +# ifdef VBOX_WITH_64ON32_IDT +/** + * Initializes the 64-bit IDT for 64-bit guest on 32-bit host switchers. + * + * This is only used as a debugging aid when we cannot find out why something + * goes haywire in the intermediate context. + * + * @param pVM The cross context VM structure. + * @param pSwitcher The switcher descriptor. + * @param pbDst Where the switcher code was just copied. + * @param HCPhysDst The host physical address corresponding to @a pbDst. + */ +static void vmmR3Switcher32On64IdtInit(PVM pVM, PVMMSWITCHERDEF pSwitcher, uint8_t *pbDst, RTHCPHYS HCPhysDst) +{ + AssertRelease(pSwitcher->offGCCode > 0 && pSwitcher->offGCCode < pSwitcher->cbCode); + AssertRelease(pSwitcher->cbCode < _64K); + RTSEL uCs64 = SELMGetHyperCS64(pVM); + + PX86DESC64GATE paIdt = (PX86DESC64GATE)(pbDst + pSwitcher->offGCCode); + for (uint32_t i = 0 ; i < 256; i++) + { + AssertRelease(((uint64_t *)&paIdt[i])[0] < pSwitcher->cbCode); + AssertRelease(((uint64_t *)&paIdt[i])[1] == 0); + uint64_t uHandler = HCPhysDst + paIdt[i].u16OffsetLow; + paIdt[i].u16OffsetLow = (uint16_t)uHandler; + paIdt[i].u16Sel = uCs64; + paIdt[i].u3IST = 0; + paIdt[i].u5Reserved = 0; + paIdt[i].u4Type = AMD64_SEL_TYPE_SYS_INT_GATE; + paIdt[i].u1DescType = 0 /* system */; + paIdt[i].u2Dpl = 3; + paIdt[i].u1Present = 1; + paIdt[i].u16OffsetHigh = (uint16_t)(uHandler >> 16); + paIdt[i].u32Reserved = (uint32_t)(uHandler >> 32); + } + + for (VMCPUID iCpu = 0; iCpu < pVM->cCpus; iCpu++) + { + uint64_t uIdtr = HCPhysDst + pSwitcher->offGCCode; AssertRelease(uIdtr < UINT32_MAX); + CPUMSetHyperIDTR(&pVM->aCpus[iCpu], uIdtr, 16*256 + iCpu); + } +} + + +/** + * Relocates the 64-bit IDT for 64-bit guest on 32-bit host switchers. + * + * @param pVM The cross context VM structure. + * @param pSwitcher The switcher descriptor. + * @param pbDst Where the switcher code was just copied. + * @param HCPhysDst The host physical address corresponding to @a pbDst. + */ +static void vmmR3Switcher32On64IdtRelocate(PVM pVM, PVMMSWITCHERDEF pSwitcher, uint8_t *pbDst, RTHCPHYS HCPhysDst) +{ + AssertRelease(pSwitcher->offGCCode > 0 && pSwitcher->offGCCode < pSwitcher->cbCode && pSwitcher->cbCode < _64K); + + /* The intermediate context doesn't move, but the CS may. */ + RTSEL uCs64 = SELMGetHyperCS64(pVM); + PX86DESC64GATE paIdt = (PX86DESC64GATE)(pbDst + pSwitcher->offGCCode); + for (uint32_t i = 0 ; i < 256; i++) + paIdt[i].u16Sel = uCs64; + + /* Just in case... */ + for (VMCPUID iCpu = 0; iCpu < pVM->cCpus; iCpu++) + { + uint64_t uIdtr = HCPhysDst + pSwitcher->offGCCode; AssertRelease(uIdtr < UINT32_MAX); + CPUMSetHyperIDTR(&pVM->aCpus[iCpu], uIdtr, 16*256 + iCpu); + } +} +# endif /* VBOX_WITH_64ON32_IDT */ + + +/** + * VMMR3Init worker that initiates the switcher code (aka core code). + * + * This is core per VM code which might need fixups and/or for ease of use are + * put on linear contiguous backing. + * + * @returns VBox status code. + * @param pVM The cross context VM structure. + */ +int vmmR3SwitcherInit(PVM pVM) +{ +#if !defined(VBOX_WITH_RAW_MODE) && (HC_ARCH_BITS == 64) + RT_NOREF(pVM); + return VINF_SUCCESS; +#else + + /* + * Calc the size. + */ + const PVMMSWITCHERDEF *papSwitchers = VM_IS_RAW_MODE_ENABLED(pVM) ? g_apRawModeSwitchers : g_apHmSwitchers; + unsigned cbCoreCode = 0; + for (unsigned iSwitcher = 0; iSwitcher < VMMSWITCHER_MAX; iSwitcher++) + { + pVM->vmm.s.aoffSwitchers[iSwitcher] = cbCoreCode; + PVMMSWITCHERDEF pSwitcher = papSwitchers[iSwitcher]; + if (pSwitcher) + { + AssertRelease((unsigned)pSwitcher->enmType == iSwitcher); + cbCoreCode += RT_ALIGN_32(pSwitcher->cbCode + 1, 32); + } + } + + /* + * Allocate contiguous pages for switchers and deal with + * conflicts in the intermediate mapping of the code. + */ + pVM->vmm.s.cbCoreCode = RT_ALIGN_32(cbCoreCode, PAGE_SIZE); + pVM->vmm.s.pvCoreCodeR3 = SUPR3ContAlloc(pVM->vmm.s.cbCoreCode >> PAGE_SHIFT, &pVM->vmm.s.pvCoreCodeR0, &pVM->vmm.s.HCPhysCoreCode); + int rc = VERR_NO_MEMORY; + if (pVM->vmm.s.pvCoreCodeR3) + { + rc = PGMR3MapIntermediate(pVM, pVM->vmm.s.pvCoreCodeR0, pVM->vmm.s.HCPhysCoreCode, cbCoreCode); + if (rc == VERR_PGM_INTERMEDIATE_PAGING_CONFLICT) + { + /* try more allocations - Solaris, Linux. */ + const unsigned cTries = 8234; + struct VMMInitBadTry + { + RTR0PTR pvR0; + void *pvR3; + RTHCPHYS HCPhys; + RTUINT cb; + } *paBadTries = (struct VMMInitBadTry *)RTMemTmpAlloc(sizeof(*paBadTries) * cTries); + AssertReturn(paBadTries, VERR_NO_TMP_MEMORY); + unsigned i = 0; + do + { + paBadTries[i].pvR3 = pVM->vmm.s.pvCoreCodeR3; + paBadTries[i].pvR0 = pVM->vmm.s.pvCoreCodeR0; + paBadTries[i].HCPhys = pVM->vmm.s.HCPhysCoreCode; + i++; + pVM->vmm.s.pvCoreCodeR0 = NIL_RTR0PTR; + pVM->vmm.s.HCPhysCoreCode = NIL_RTHCPHYS; + pVM->vmm.s.pvCoreCodeR3 = SUPR3ContAlloc(pVM->vmm.s.cbCoreCode >> PAGE_SHIFT, &pVM->vmm.s.pvCoreCodeR0, &pVM->vmm.s.HCPhysCoreCode); + if (!pVM->vmm.s.pvCoreCodeR3) + break; + rc = PGMR3MapIntermediate(pVM, pVM->vmm.s.pvCoreCodeR0, pVM->vmm.s.HCPhysCoreCode, cbCoreCode); + } while ( rc == VERR_PGM_INTERMEDIATE_PAGING_CONFLICT + && i < cTries - 1); + + /* cleanup */ + if (RT_FAILURE(rc)) + { + paBadTries[i].pvR3 = pVM->vmm.s.pvCoreCodeR3; + paBadTries[i].pvR0 = pVM->vmm.s.pvCoreCodeR0; + paBadTries[i].HCPhys = pVM->vmm.s.HCPhysCoreCode; + paBadTries[i].cb = pVM->vmm.s.cbCoreCode; + i++; + LogRel(("VMM: Failed to allocated and map core code: rc=%Rrc\n", rc)); + } + while (i-- > 0) + { + LogRel(("VMM: Core code alloc attempt #%d: pvR3=%p pvR0=%RKv HCPhys=%RHp\n", + i, paBadTries[i].pvR3, paBadTries[i].pvR0, paBadTries[i].HCPhys)); + SUPR3ContFree(paBadTries[i].pvR3, paBadTries[i].cb >> PAGE_SHIFT); + } + RTMemTmpFree(paBadTries); + } + } + if (RT_SUCCESS(rc)) + { + /* + * Copy the code. + */ + for (unsigned iSwitcher = 0; iSwitcher < VMMSWITCHER_MAX; iSwitcher++) + { + PVMMSWITCHERDEF pSwitcher = papSwitchers[iSwitcher]; + if (pSwitcher) + { + uint8_t *pbDst = (uint8_t *)pVM->vmm.s.pvCoreCodeR3 + pVM->vmm.s.aoffSwitchers[iSwitcher]; + memcpy(pbDst, pSwitcher->pvCode, pSwitcher->cbCode); +# ifdef VBOX_WITH_64ON32_IDT + if ( pSwitcher->enmType == VMMSWITCHER_32_TO_AMD64 + || pSwitcher->enmType == VMMSWITCHER_PAE_TO_AMD64) + vmmR3Switcher32On64IdtInit(pVM, pSwitcher, pbDst, + pVM->vmm.s.HCPhysCoreCode + pVM->vmm.s.aoffSwitchers[iSwitcher]); +# endif + } + } + + /* + * Map the code into the GC address space. + */ + RTGCPTR GCPtr; + rc = MMR3HyperMapHCPhys(pVM, pVM->vmm.s.pvCoreCodeR3, pVM->vmm.s.pvCoreCodeR0, pVM->vmm.s.HCPhysCoreCode, + cbCoreCode, "Core Code", &GCPtr); + if (RT_SUCCESS(rc)) + { + pVM->vmm.s.pvCoreCodeRC = GCPtr; + MMR3HyperReserve(pVM, PAGE_SIZE, "fence", NULL); + LogRel(("VMM: CoreCode: R3=%RHv R0=%RKv RC=%RRv Phys=%RHp cb=%#x\n", + pVM->vmm.s.pvCoreCodeR3, pVM->vmm.s.pvCoreCodeR0, pVM->vmm.s.pvCoreCodeRC, pVM->vmm.s.HCPhysCoreCode, pVM->vmm.s.cbCoreCode)); + + /* + * Finally, PGM probably has selected a switcher already but we need + * to get the routine addresses, so we'll reselect it. + * This may legally fail so, we're ignoring the rc. + * Note! See HMIsEnabled hack in selector function. + */ + VMMR3SelectSwitcher(pVM, pVM->vmm.s.enmSwitcher); + return rc; + } + + /* shit */ + AssertMsgFailed(("PGMR3Map(,%RRv, %RHp, %#x, 0) failed with rc=%Rrc\n", pVM->vmm.s.pvCoreCodeRC, pVM->vmm.s.HCPhysCoreCode, cbCoreCode, rc)); + SUPR3ContFree(pVM->vmm.s.pvCoreCodeR3, pVM->vmm.s.cbCoreCode >> PAGE_SHIFT); + } + else + VMSetError(pVM, rc, RT_SRC_POS, + N_("Failed to allocate %d bytes of contiguous memory for the world switcher code"), + cbCoreCode); + + pVM->vmm.s.pvCoreCodeR3 = NULL; + pVM->vmm.s.pvCoreCodeR0 = NIL_RTR0PTR; + pVM->vmm.s.pvCoreCodeRC = 0; + return rc; +#endif +} + +/** + * Relocate the switchers, called by VMMR#Relocate. + * + * @param pVM The cross context VM structure. + * @param offDelta The relocation delta. + */ +void vmmR3SwitcherRelocate(PVM pVM, RTGCINTPTR offDelta) +{ +#if defined(VBOX_WITH_RAW_MODE) || (HC_ARCH_BITS != 64) + /* + * Relocate all the switchers. + */ + const PVMMSWITCHERDEF *papSwitchers = VM_IS_RAW_MODE_ENABLED(pVM) ? g_apRawModeSwitchers : g_apHmSwitchers; + for (unsigned iSwitcher = 0; iSwitcher < VMMSWITCHER_MAX; iSwitcher++) + { + PVMMSWITCHERDEF pSwitcher = papSwitchers[iSwitcher]; + if (pSwitcher && pSwitcher->pfnRelocate) + { + unsigned off = pVM->vmm.s.aoffSwitchers[iSwitcher]; + pSwitcher->pfnRelocate(pVM, + pSwitcher, + pVM->vmm.s.pvCoreCodeR0 + off, + (uint8_t *)pVM->vmm.s.pvCoreCodeR3 + off, + pVM->vmm.s.pvCoreCodeRC + off, + pVM->vmm.s.HCPhysCoreCode + off); +# ifdef VBOX_WITH_64ON32_IDT + if ( pSwitcher->enmType == VMMSWITCHER_32_TO_AMD64 + || pSwitcher->enmType == VMMSWITCHER_PAE_TO_AMD64) + vmmR3Switcher32On64IdtRelocate(pVM, pSwitcher, + (uint8_t *)pVM->vmm.s.pvCoreCodeR3 + off, + pVM->vmm.s.HCPhysCoreCode + off); +# endif + } + } + + /* + * Recalc the RC address for the current switcher. + */ + PVMMSWITCHERDEF pSwitcher = papSwitchers[pVM->vmm.s.enmSwitcher]; + if (pSwitcher) + { + RTRCPTR RCPtr = pVM->vmm.s.pvCoreCodeRC + pVM->vmm.s.aoffSwitchers[pVM->vmm.s.enmSwitcher]; + pVM->vmm.s.pfnRCToHost = RCPtr + pSwitcher->offRCToHost; + pVM->vmm.s.pfnCallTrampolineRC = RCPtr + pSwitcher->offRCCallTrampoline; + pVM->pfnVMMRCToHostAsm = RCPtr + pSwitcher->offRCToHostAsm; + pVM->pfnVMMRCToHostAsmNoReturn = RCPtr + pSwitcher->offRCToHostAsmNoReturn; + } + else + AssertRelease(!VM_IS_RAW_MODE_ENABLED(pVM)); + +#else + NOREF(pVM); +#endif + NOREF(offDelta); +} + + +#if defined(VBOX_WITH_RAW_MODE) || (HC_ARCH_BITS != 64) + +/** + * Generic switcher code relocator. + * + * @param pVM The cross context VM structure. + * @param pSwitcher The switcher definition. + * @param pu8CodeR3 Pointer to the core code block for the switcher, ring-3 mapping. + * @param R0PtrCode Pointer to the core code block for the switcher, ring-0 mapping. + * @param GCPtrCode The guest context address corresponding to pu8Code. + * @param u32IDCode The identity mapped (ID) address corresponding to pu8Code. + * @param SelCS The hypervisor CS selector. + * @param SelDS The hypervisor DS selector. + * @param SelTSS The hypervisor TSS selector. + * @param GCPtrGDT The GC address of the hypervisor GDT. + * @param SelCS64 The 64-bit mode hypervisor CS selector. + */ +static void vmmR3SwitcherGenericRelocate(PVM pVM, PVMMSWITCHERDEF pSwitcher, + RTR0PTR R0PtrCode, uint8_t *pu8CodeR3, RTGCPTR GCPtrCode, uint32_t u32IDCode, + RTSEL SelCS, RTSEL SelDS, RTSEL SelTSS, RTGCPTR GCPtrGDT, RTSEL SelCS64) +{ + union + { + const uint8_t *pu8; + const uint16_t *pu16; + const uint32_t *pu32; + const uint64_t *pu64; + const void *pv; + uintptr_t u; + } u; + u.pv = pSwitcher->pvFixups; + + /* + * Process fixups. + */ + uint8_t u8; + while ((u8 = *u.pu8++) != FIX_THE_END) + { + /* + * Get the source (where to write the fixup). + */ + uint32_t offSrc = *u.pu32++; + Assert(offSrc < pSwitcher->cbCode); + union + { + uint8_t *pu8; + uint16_t *pu16; + uint32_t *pu32; + uint64_t *pu64; + uintptr_t u; + } uSrc; + uSrc.pu8 = pu8CodeR3 + offSrc; + + /* The fixup target and method depends on the type. */ + switch (u8) + { + /* + * 32-bit relative, source in HC and target in GC. + */ + case FIX_HC_2_GC_NEAR_REL: + { + Assert(offSrc - pSwitcher->offHCCode0 < pSwitcher->cbHCCode0 || offSrc - pSwitcher->offHCCode1 < pSwitcher->cbHCCode1); + uint32_t offTrg = *u.pu32++; + Assert(offTrg - pSwitcher->offGCCode < pSwitcher->cbGCCode); + *uSrc.pu32 = (uint32_t)((GCPtrCode + offTrg) - (uSrc.u + 4)); + break; + } + + /* + * 32-bit relative, source in HC and target in ID. + */ + case FIX_HC_2_ID_NEAR_REL: + { + Assert(offSrc - pSwitcher->offHCCode0 < pSwitcher->cbHCCode0 || offSrc - pSwitcher->offHCCode1 < pSwitcher->cbHCCode1); + uint32_t offTrg = *u.pu32++; + Assert(offTrg - pSwitcher->offIDCode0 < pSwitcher->cbIDCode0 || offTrg - pSwitcher->offIDCode1 < pSwitcher->cbIDCode1); + *uSrc.pu32 = (uint32_t)((u32IDCode + offTrg) - (R0PtrCode + offSrc + 4)); + break; + } + + /* + * 32-bit relative, source in GC and target in HC. + */ + case FIX_GC_2_HC_NEAR_REL: + { + Assert(offSrc - pSwitcher->offGCCode < pSwitcher->cbGCCode); + uint32_t offTrg = *u.pu32++; + Assert(offTrg - pSwitcher->offHCCode0 < pSwitcher->cbHCCode0 || offTrg - pSwitcher->offHCCode1 < pSwitcher->cbHCCode1); + *uSrc.pu32 = (uint32_t)((R0PtrCode + offTrg) - (GCPtrCode + offSrc + 4)); + break; + } + + /* + * 32-bit relative, source in GC and target in ID. + */ + case FIX_GC_2_ID_NEAR_REL: + { + AssertMsg(offSrc - pSwitcher->offGCCode < pSwitcher->cbGCCode, ("%x - %x < %x\n", offSrc, pSwitcher->offGCCode, pSwitcher->cbGCCode)); + uint32_t offTrg = *u.pu32++; + Assert(offTrg - pSwitcher->offIDCode0 < pSwitcher->cbIDCode0 || offTrg - pSwitcher->offIDCode1 < pSwitcher->cbIDCode1); + *uSrc.pu32 = (uint32_t)((u32IDCode + offTrg) - (GCPtrCode + offSrc + 4)); + break; + } + + /* + * 32-bit relative, source in ID and target in HC. + */ + case FIX_ID_2_HC_NEAR_REL: + { + Assert(offSrc - pSwitcher->offIDCode0 < pSwitcher->cbIDCode0 || offSrc - pSwitcher->offIDCode1 < pSwitcher->cbIDCode1); + uint32_t offTrg = *u.pu32++; + Assert(offTrg - pSwitcher->offHCCode0 < pSwitcher->cbHCCode0 || offTrg - pSwitcher->offHCCode1 < pSwitcher->cbHCCode1); + *uSrc.pu32 = (uint32_t)((R0PtrCode + offTrg) - (u32IDCode + offSrc + 4)); + break; + } + + /* + * 32-bit relative, source in ID and target in HC. + */ + case FIX_ID_2_GC_NEAR_REL: + { + Assert(offSrc - pSwitcher->offIDCode0 < pSwitcher->cbIDCode0 || offSrc - pSwitcher->offIDCode1 < pSwitcher->cbIDCode1); + uint32_t offTrg = *u.pu32++; + Assert(offTrg - pSwitcher->offGCCode < pSwitcher->cbGCCode); + *uSrc.pu32 = (uint32_t)((GCPtrCode + offTrg) - (u32IDCode + offSrc + 4)); + break; + } + + /* + * 16:32 far jump, target in GC. + */ + case FIX_GC_FAR32: + { + uint32_t offTrg = *u.pu32++; + Assert(offTrg - pSwitcher->offGCCode < pSwitcher->cbGCCode); + *uSrc.pu32++ = (uint32_t)(GCPtrCode + offTrg); + *uSrc.pu16++ = SelCS; + break; + } + + /* + * Make 32-bit GC pointer given CPUM offset. + */ + case FIX_GC_CPUM_OFF: + { + uint32_t offCPUM = *u.pu32++; + Assert(offCPUM < sizeof(pVM->cpum)); + *uSrc.pu32 = (uint32_t)(VM_RC_ADDR(pVM, &pVM->cpum) + offCPUM); + break; + } + + /* + * Make 32-bit GC pointer given CPUMCPU offset. + */ + case FIX_GC_CPUMCPU_OFF: + { + uint32_t offCPUM = *u.pu32++; + Assert(offCPUM < sizeof(pVM->aCpus[0].cpum)); + *uSrc.pu32 = (uint32_t)(VM_RC_ADDR(pVM, &pVM->aCpus[0].cpum) + offCPUM); + break; + } + + /* + * Make 32-bit GC pointer given VM offset. + */ + case FIX_GC_VM_OFF: + { + uint32_t offVM = *u.pu32++; + Assert(offVM < sizeof(VM)); + *uSrc.pu32 = (uint32_t)(VM_RC_ADDR(pVM, pVM) + offVM); + break; + } + + /* + * Make 32-bit HC pointer given CPUM offset. + */ + case FIX_HC_CPUM_OFF: + { + uint32_t offCPUM = *u.pu32++; + Assert(offCPUM < sizeof(pVM->cpum)); + *uSrc.pu32 = (uint32_t)pVM->pVMR0 + RT_UOFFSETOF(VM, cpum) + offCPUM; + break; + } + + /* + * Make 32-bit R0 pointer given VM offset. + */ + case FIX_HC_VM_OFF: + { + uint32_t offVM = *u.pu32++; + Assert(offVM < sizeof(VM)); + *uSrc.pu32 = (uint32_t)pVM->pVMR0 + offVM; + break; + } + + /* + * Store the 32-Bit CR3 (32-bit) for the intermediate memory context. + */ + case FIX_INTER_32BIT_CR3: + { + + *uSrc.pu32 = PGMGetInter32BitCR3(pVM); + break; + } + + /* + * Store the PAE CR3 (32-bit) for the intermediate memory context. + */ + case FIX_INTER_PAE_CR3: + { + + *uSrc.pu32 = PGMGetInterPaeCR3(pVM); + break; + } + + /* + * Store the AMD64 CR3 (32-bit) for the intermediate memory context. + */ + case FIX_INTER_AMD64_CR3: + { + + *uSrc.pu32 = PGMGetInterAmd64CR3(pVM); + break; + } + + /* + * Store Hypervisor CS (16-bit). + */ + case FIX_HYPER_CS: + { + *uSrc.pu16 = SelCS; + break; + } + + /* + * Store Hypervisor DS (16-bit). + */ + case FIX_HYPER_DS: + { + *uSrc.pu16 = SelDS; + break; + } + + /* + * Store Hypervisor TSS (16-bit). + */ + case FIX_HYPER_TSS: + { + *uSrc.pu16 = SelTSS; + break; + } + + /* + * Store the 32-bit GC address of the 2nd dword of the TSS descriptor (in the GDT). + */ + case FIX_GC_TSS_GDTE_DW2: + { + RTGCPTR GCPtr = GCPtrGDT + (SelTSS & ~7) + 4; + *uSrc.pu32 = (uint32_t)GCPtr; + break; + } + + /* + * Store the EFER or mask for the 32->64 bit switcher. + */ + case FIX_EFER_OR_MASK: + { + uint32_t u32OrMask = MSR_K6_EFER_LME | MSR_K6_EFER_SCE; + /* + * We don't care if cpuid 0x8000001 isn't supported as that implies + * long mode isn't supported either, so this switched would never be used. + */ + if (!!(ASMCpuId_EDX(0x80000001) & X86_CPUID_EXT_FEATURE_EDX_NX)) + u32OrMask |= MSR_K6_EFER_NXE; + + *uSrc.pu32 = u32OrMask; + break; + } + +#if 0 /* Reusable for XSAVE. */ + /* + * Insert relative jump to specified target it FXSAVE/FXRSTOR isn't supported by the cpu. + */ + case FIX_NO_FXSAVE_JMP: + { + uint32_t offTrg = *u.pu32++; + Assert(offTrg < pSwitcher->cbCode); + if (!CPUMSupportsXSave(pVM)) + { + *uSrc.pu8++ = 0xe9; /* jmp rel32 */ + *uSrc.pu32++ = offTrg - (offSrc + 5); + } + else + { + *uSrc.pu8++ = *((uint8_t *)pSwitcher->pvCode + offSrc); + *uSrc.pu32++ = *(uint32_t *)((uint8_t *)pSwitcher->pvCode + offSrc + 1); + } + break; + } +#endif + + /* + * Insert relative jump to specified target it SYSENTER isn't used by the host. + */ + case FIX_NO_SYSENTER_JMP: + { + uint32_t offTrg = *u.pu32++; + Assert(offTrg < pSwitcher->cbCode); + if (!CPUMIsHostUsingSysEnter(pVM)) + { + *uSrc.pu8++ = 0xe9; /* jmp rel32 */ + *uSrc.pu32++ = offTrg - (offSrc + 5); + } + else + { + *uSrc.pu8++ = *((uint8_t *)pSwitcher->pvCode + offSrc); + *uSrc.pu32++ = *(uint32_t *)((uint8_t *)pSwitcher->pvCode + offSrc + 1); + } + break; + } + + /* + * Insert relative jump to specified target it SYSCALL isn't used by the host. + */ + case FIX_NO_SYSCALL_JMP: + { + uint32_t offTrg = *u.pu32++; + Assert(offTrg < pSwitcher->cbCode); + if (!CPUMIsHostUsingSysCall(pVM)) + { + *uSrc.pu8++ = 0xe9; /* jmp rel32 */ + *uSrc.pu32++ = offTrg - (offSrc + 5); + } + else + { + *uSrc.pu8++ = *((uint8_t *)pSwitcher->pvCode + offSrc); + *uSrc.pu32++ = *(uint32_t *)((uint8_t *)pSwitcher->pvCode + offSrc + 1); + } + break; + } + + /* + * 32-bit HC pointer fixup to (HC) target within the code (32-bit offset). + */ + case FIX_HC_32BIT: + { + uint32_t offTrg = *u.pu32++; + Assert(offSrc < pSwitcher->cbCode); + Assert(offTrg - pSwitcher->offHCCode0 < pSwitcher->cbHCCode0 || offTrg - pSwitcher->offHCCode1 < pSwitcher->cbHCCode1); + *uSrc.pu32 = R0PtrCode + offTrg; + break; + } + +# if defined(RT_ARCH_AMD64) + /* + * 64-bit HC Code Selector (no argument). + */ + case FIX_HC_64BIT_CS: + { + Assert(offSrc < pSwitcher->cbCode); + AssertFatalMsgFailed(("FIX_HC_64BIT_CS not implemented for this host\n")); + break; + } + + /* + * 64-bit HC pointer to the CPUM instance data (no argument). + */ + case FIX_HC_64BIT_CPUM: + { + Assert(offSrc < pSwitcher->cbCode); + *uSrc.pu64 = pVM->pVMR0 + RT_UOFFSETOF(VM, cpum); + break; + } +# endif + /* + * 64-bit HC pointer fixup to (HC) target within the code (32-bit offset). + */ + case FIX_HC_64BIT: + { + uint32_t offTrg = *u.pu32++; + Assert(offSrc < pSwitcher->cbCode); + Assert(offTrg - pSwitcher->offHCCode0 < pSwitcher->cbHCCode0 || offTrg - pSwitcher->offHCCode1 < pSwitcher->cbHCCode1); + *uSrc.pu64 = R0PtrCode + offTrg; + break; + } + +# ifdef RT_ARCH_X86 + case FIX_GC_64_BIT_CPUM_OFF: + { + uint32_t offCPUM = *u.pu32++; + Assert(offCPUM < sizeof(pVM->cpum)); + *uSrc.pu64 = (uint32_t)(VM_RC_ADDR(pVM, &pVM->cpum) + offCPUM); + break; + } +# endif + + /* + * 32-bit ID pointer to (ID) target within the code (32-bit offset). + */ + case FIX_ID_32BIT: + { + uint32_t offTrg = *u.pu32++; + Assert(offSrc < pSwitcher->cbCode); + Assert(offTrg - pSwitcher->offIDCode0 < pSwitcher->cbIDCode0 || offTrg - pSwitcher->offIDCode1 < pSwitcher->cbIDCode1); + *uSrc.pu32 = u32IDCode + offTrg; + break; + } + + /* + * 64-bit ID pointer to (ID) target within the code (32-bit offset). + */ + case FIX_ID_64BIT: + case FIX_HC_64BIT_NOCHECK: + { + uint32_t offTrg = *u.pu32++; + Assert(offSrc < pSwitcher->cbCode); + Assert(u8 == FIX_HC_64BIT_NOCHECK || offTrg - pSwitcher->offIDCode0 < pSwitcher->cbIDCode0 || offTrg - pSwitcher->offIDCode1 < pSwitcher->cbIDCode1); + *uSrc.pu64 = u32IDCode + offTrg; + break; + } + + /* + * Far 16:32 ID pointer to 64-bit mode (ID) target within the code (32-bit offset). + */ + case FIX_ID_FAR32_TO_64BIT_MODE: + { + uint32_t offTrg = *u.pu32++; + Assert(offSrc < pSwitcher->cbCode); + Assert(offTrg - pSwitcher->offIDCode0 < pSwitcher->cbIDCode0 || offTrg - pSwitcher->offIDCode1 < pSwitcher->cbIDCode1); + *uSrc.pu32++ = u32IDCode + offTrg; + *uSrc.pu16 = SelCS64; + AssertRelease(SelCS64); + break; + } + +# ifdef VBOX_WITH_NMI + /* + * 32-bit address to the APIC base. + */ + case FIX_GC_APIC_BASE_32BIT: + { + *uSrc.pu32 = pVM->vmm.s.GCPtrApicBase; + break; + } +# endif + + default: + AssertReleaseMsgFailed(("Unknown fixup %d in switcher %s\n", u8, pSwitcher->pszDesc)); + break; + } + } + +# ifdef LOG_ENABLED + /* + * If Log2 is enabled disassemble the switcher code. + * + * The switcher code have 1-2 HC parts, 1 GC part and 0-2 ID parts. + */ + if (LogIs2Enabled()) + { + RTLogPrintf("*** Disassembly of switcher %d '%s' %#x bytes ***\n" + " R0PtrCode = %p\n" + " pu8CodeR3 = %p\n" + " GCPtrCode = %RGv\n" + " u32IDCode = %08x\n" + " pVMRC = %RRv\n" + " pCPUMRC = %RRv\n" + " pVMR3 = %p\n" + " pCPUMR3 = %p\n" + " GCPtrGDT = %RGv\n" + " InterCR3s = %08RHp, %08RHp, %08RHp (32-Bit, PAE, AMD64)\n" + " HyperCR3s = %08RHp (32-Bit, PAE & AMD64)\n" + " SelCS = %04x\n" + " SelDS = %04x\n" + " SelCS64 = %04x\n" + " SelTSS = %04x\n", + pSwitcher->enmType, pSwitcher->pszDesc, pSwitcher->cbCode, + R0PtrCode, + pu8CodeR3, + GCPtrCode, + u32IDCode, + VM_RC_ADDR(pVM, pVM), + VM_RC_ADDR(pVM, &pVM->cpum), + pVM, + &pVM->cpum, + GCPtrGDT, + PGMGetInter32BitCR3(pVM), PGMGetInterPaeCR3(pVM), PGMGetInterAmd64CR3(pVM), + PGMGetHyperCR3(VMMGetCpu(pVM)), + SelCS, SelDS, SelCS64, SelTSS); + + uint32_t offCode = 0; + while (offCode < pSwitcher->cbCode) + { + /* + * Figure out where this is. + */ + const char *pszDesc = NULL; + RTUINTPTR uBase; + uint32_t cbCode; + if (offCode - pSwitcher->offHCCode0 < pSwitcher->cbHCCode0) + { + pszDesc = "HCCode0"; + uBase = R0PtrCode; + offCode = pSwitcher->offHCCode0; + cbCode = pSwitcher->cbHCCode0; + } + else if (offCode - pSwitcher->offHCCode1 < pSwitcher->cbHCCode1) + { + pszDesc = "HCCode1"; + uBase = R0PtrCode; + offCode = pSwitcher->offHCCode1; + cbCode = pSwitcher->cbHCCode1; + } + else if (offCode - pSwitcher->offGCCode < pSwitcher->cbGCCode) + { + pszDesc = "GCCode"; + uBase = GCPtrCode; + offCode = pSwitcher->offGCCode; + cbCode = pSwitcher->cbGCCode; + } + else if (offCode - pSwitcher->offIDCode0 < pSwitcher->cbIDCode0) + { + pszDesc = "IDCode0"; + uBase = u32IDCode; + offCode = pSwitcher->offIDCode0; + cbCode = pSwitcher->cbIDCode0; + } + else if (offCode - pSwitcher->offIDCode1 < pSwitcher->cbIDCode1) + { + pszDesc = "IDCode1"; + uBase = u32IDCode; + offCode = pSwitcher->offIDCode1; + cbCode = pSwitcher->cbIDCode1; + } + else + { + RTLogPrintf(" %04x: %02x '%c' (nowhere)\n", + offCode, pu8CodeR3[offCode], RT_C_IS_PRINT(pu8CodeR3[offCode]) ? pu8CodeR3[offCode] : ' '); + offCode++; + continue; + } + + /* + * Disassemble it. + */ + RTLogPrintf(" %s: offCode=%#x cbCode=%#x\n", pszDesc, offCode, cbCode); + + while (cbCode > 0) + { + /* try label it */ + if (pSwitcher->offR0ToRawMode == offCode) + RTLogPrintf(" *R0ToRawMode:\n"); + if (pSwitcher->offRCToHost == offCode) + RTLogPrintf(" *RCToHost:\n"); + if (pSwitcher->offRCCallTrampoline == offCode) + RTLogPrintf(" *RCCallTrampoline:\n"); + if (pSwitcher->offRCToHostAsm == offCode) + RTLogPrintf(" *RCToHostAsm:\n"); + if (pSwitcher->offRCToHostAsmNoReturn == offCode) + RTLogPrintf(" *RCToHostAsmNoReturn:\n"); + + /* disas */ + uint32_t cbInstr = 0; + DISCPUSTATE Cpu; + char szDisas[256]; + int rc = DISInstr(pu8CodeR3 + offCode, DISCPUMODE_32BIT, &Cpu, &cbInstr); + if (RT_SUCCESS(rc)) + { + Cpu.uInstrAddr += uBase - (uintptr_t)pu8CodeR3; + DISFormatYasmEx(&Cpu, szDisas, sizeof(szDisas), + DIS_FMT_FLAGS_ADDR_LEFT | DIS_FMT_FLAGS_BYTES_LEFT | DIS_FMT_FLAGS_BYTES_SPACED + | DIS_FMT_FLAGS_RELATIVE_BRANCH, + NULL, NULL); + } + if (RT_SUCCESS(rc)) + RTLogPrintf(" %04x: %s\n", offCode, szDisas); + else + { + RTLogPrintf(" %04x: %02x '%c' (rc=%Rrc\n", + offCode, pu8CodeR3[offCode], RT_C_IS_PRINT(pu8CodeR3[offCode]) ? pu8CodeR3[offCode] : ' ', rc); + cbInstr = 1; + } + offCode += cbInstr; + cbCode -= RT_MIN(cbInstr, cbCode); + } + } + } +# endif +} + +/** + * Wrapper around SELMGetHyperGDT() that avoids calling it when raw-mode context + * is not initialized. + * + * @returns Raw-mode contet GDT address. Null pointer if not applicable. + * @param pVM The cross context VM structure. + */ +static RTRCPTR vmmR3SwitcherGetHyperGDT(PVM pVM) +{ + if (VM_IS_RAW_MODE_ENABLED(pVM) || HMIsRawModeCtxNeeded(pVM)) + return SELMGetHyperGDT(pVM); +# if HC_ARCH_BITS != 32 + AssertFailed(); /* This path is only applicable to some 32-bit hosts. */ +# endif + return NIL_RTRCPTR; +} + +/** + * Relocator for the 32-Bit to 32-Bit world switcher. + */ +DECLCALLBACK(void) vmmR3Switcher32BitTo32Bit_Relocate(PVM pVM, PVMMSWITCHERDEF pSwitcher, RTR0PTR R0PtrCode, uint8_t *pu8CodeR3, RTGCPTR GCPtrCode, uint32_t u32IDCode) +{ + vmmR3SwitcherGenericRelocate(pVM, pSwitcher, R0PtrCode, pu8CodeR3, GCPtrCode, u32IDCode, + SELMGetHyperCS(pVM), SELMGetHyperDS(pVM), SELMGetHyperTSS(pVM), SELMGetHyperGDT(pVM), 0); +} + + +/** + * Relocator for the 32-Bit to PAE world switcher. + */ +DECLCALLBACK(void) vmmR3Switcher32BitToPAE_Relocate(PVM pVM, PVMMSWITCHERDEF pSwitcher, RTR0PTR R0PtrCode, uint8_t *pu8CodeR3, RTGCPTR GCPtrCode, uint32_t u32IDCode) +{ + vmmR3SwitcherGenericRelocate(pVM, pSwitcher, R0PtrCode, pu8CodeR3, GCPtrCode, u32IDCode, + SELMGetHyperCS(pVM), SELMGetHyperDS(pVM), SELMGetHyperTSS(pVM), SELMGetHyperGDT(pVM), 0); +} + + +/** + * Relocator for the 32-Bit to AMD64 world switcher. + */ +DECLCALLBACK(void) vmmR3Switcher32BitToAMD64_Relocate(PVM pVM, PVMMSWITCHERDEF pSwitcher, RTR0PTR R0PtrCode, uint8_t *pu8CodeR3, RTGCPTR GCPtrCode, uint32_t u32IDCode) +{ + vmmR3SwitcherGenericRelocate(pVM, pSwitcher, R0PtrCode, pu8CodeR3, GCPtrCode, u32IDCode, + SELMGetHyperCS(pVM), SELMGetHyperDS(pVM), SELMGetHyperTSS(pVM), vmmR3SwitcherGetHyperGDT(pVM), SELMGetHyperCS64(pVM)); +} + + +/** + * Relocator for the PAE to 32-Bit world switcher. + */ +DECLCALLBACK(void) vmmR3SwitcherPAETo32Bit_Relocate(PVM pVM, PVMMSWITCHERDEF pSwitcher, RTR0PTR R0PtrCode, uint8_t *pu8CodeR3, RTGCPTR GCPtrCode, uint32_t u32IDCode) +{ + vmmR3SwitcherGenericRelocate(pVM, pSwitcher, R0PtrCode, pu8CodeR3, GCPtrCode, u32IDCode, + SELMGetHyperCS(pVM), SELMGetHyperDS(pVM), SELMGetHyperTSS(pVM), SELMGetHyperGDT(pVM), 0); +} + + +/** + * Relocator for the PAE to PAE world switcher. + */ +DECLCALLBACK(void) vmmR3SwitcherPAEToPAE_Relocate(PVM pVM, PVMMSWITCHERDEF pSwitcher, RTR0PTR R0PtrCode, uint8_t *pu8CodeR3, RTGCPTR GCPtrCode, uint32_t u32IDCode) +{ + vmmR3SwitcherGenericRelocate(pVM, pSwitcher, R0PtrCode, pu8CodeR3, GCPtrCode, u32IDCode, + SELMGetHyperCS(pVM), SELMGetHyperDS(pVM), SELMGetHyperTSS(pVM), SELMGetHyperGDT(pVM), 0); +} + +/** + * Relocator for the PAE to AMD64 world switcher. + */ +DECLCALLBACK(void) vmmR3SwitcherPAEToAMD64_Relocate(PVM pVM, PVMMSWITCHERDEF pSwitcher, RTR0PTR R0PtrCode, uint8_t *pu8CodeR3, RTGCPTR GCPtrCode, uint32_t u32IDCode) +{ + vmmR3SwitcherGenericRelocate(pVM, pSwitcher, R0PtrCode, pu8CodeR3, GCPtrCode, u32IDCode, + SELMGetHyperCS(pVM), SELMGetHyperDS(pVM), SELMGetHyperTSS(pVM), vmmR3SwitcherGetHyperGDT(pVM), SELMGetHyperCS64(pVM)); +} + + +/** + * Relocator for the AMD64 to 32-bit world switcher. + */ +DECLCALLBACK(void) vmmR3SwitcherAMD64To32Bit_Relocate(PVM pVM, PVMMSWITCHERDEF pSwitcher, RTR0PTR R0PtrCode, uint8_t *pu8CodeR3, RTGCPTR GCPtrCode, uint32_t u32IDCode) +{ + vmmR3SwitcherGenericRelocate(pVM, pSwitcher, R0PtrCode, pu8CodeR3, GCPtrCode, u32IDCode, + SELMGetHyperCS(pVM), SELMGetHyperDS(pVM), SELMGetHyperTSS(pVM), SELMGetHyperGDT(pVM), SELMGetHyperCS64(pVM)); +} + + +/** + * Relocator for the AMD64 to PAE world switcher. + */ +DECLCALLBACK(void) vmmR3SwitcherAMD64ToPAE_Relocate(PVM pVM, PVMMSWITCHERDEF pSwitcher, RTR0PTR R0PtrCode, uint8_t *pu8CodeR3, RTGCPTR GCPtrCode, uint32_t u32IDCode) +{ + vmmR3SwitcherGenericRelocate(pVM, pSwitcher, R0PtrCode, pu8CodeR3, GCPtrCode, u32IDCode, + SELMGetHyperCS(pVM), SELMGetHyperDS(pVM), SELMGetHyperTSS(pVM), SELMGetHyperGDT(pVM), SELMGetHyperCS64(pVM)); +} + + +/** + * Selects the switcher to be used for switching to raw-mode context. + * + * @returns VBox status code. + * @param pVM The cross context VM structure. + * @param enmSwitcher The new switcher. + * @remark This function may be called before the VMM is initialized. + */ +VMMR3_INT_DECL(int) VMMR3SelectSwitcher(PVM pVM, VMMSWITCHER enmSwitcher) +{ + /* + * Validate input. + */ + if ( enmSwitcher < VMMSWITCHER_INVALID + || enmSwitcher >= VMMSWITCHER_MAX) + { + AssertMsgFailed(("Invalid input enmSwitcher=%d\n", enmSwitcher)); + return VERR_INVALID_PARAMETER; + } + + /* + * Override it if HM is active. + */ + if (!VM_IS_RAW_MODE_ENABLED(pVM)) + pVM->vmm.s.enmSwitcher = HC_ARCH_BITS == 64 ? VMMSWITCHER_AMD64_STUB : VMMSWITCHER_X86_STUB; + + /* + * Select the new switcher. + */ + const PVMMSWITCHERDEF *papSwitchers = VM_IS_RAW_MODE_ENABLED(pVM) ? g_apRawModeSwitchers : g_apHmSwitchers; + PVMMSWITCHERDEF pSwitcher = papSwitchers[enmSwitcher]; + if (pSwitcher) + { + Log(("VMMR3SelectSwitcher: enmSwitcher %d -> %d %s\n", pVM->vmm.s.enmSwitcher, enmSwitcher, pSwitcher->pszDesc)); + pVM->vmm.s.enmSwitcher = enmSwitcher; + + RTR0PTR pbCodeR0 = (RTR0PTR)pVM->vmm.s.pvCoreCodeR0 + pVM->vmm.s.aoffSwitchers[enmSwitcher]; /** @todo fix the pvCoreCodeR0 type */ + pVM->vmm.s.pfnR0ToRawMode = pbCodeR0 + pSwitcher->offR0ToRawMode; + + RTRCPTR RCPtr = pVM->vmm.s.pvCoreCodeRC + pVM->vmm.s.aoffSwitchers[enmSwitcher]; + pVM->vmm.s.pfnRCToHost = RCPtr + pSwitcher->offRCToHost; + pVM->vmm.s.pfnCallTrampolineRC = RCPtr + pSwitcher->offRCCallTrampoline; + pVM->pfnVMMRCToHostAsm = RCPtr + pSwitcher->offRCToHostAsm; + pVM->pfnVMMRCToHostAsmNoReturn = RCPtr + pSwitcher->offRCToHostAsmNoReturn; + return VINF_SUCCESS; + } + + return VERR_NOT_IMPLEMENTED; +} + +#endif /* #defined(VBOX_WITH_RAW_MODE) || (HC_ARCH_BITS != 64) */ + + +#if HC_ARCH_BITS == 32 && defined(VBOX_WITH_64_BITS_GUESTS) +/** + * Gets the switcher to be used for switching to GC. + * + * This is for 64-on-32 with HM. Caller is HMR3Relocate(). + * + * @returns host to guest ring 0 switcher entrypoint + * @param pVM The cross context VM structure. + * @param enmSwitcher The new switcher. + */ +VMMR3_INT_DECL(RTR0PTR) VMMR3GetHostToGuestSwitcher(PVM pVM, VMMSWITCHER enmSwitcher) +{ + /* + * Validate input. + */ + AssertMsgReturn( enmSwitcher == VMMSWITCHER_32_TO_AMD64 + || enmSwitcher == VMMSWITCHER_PAE_TO_AMD64, + ("%d\n", enmSwitcher), + NIL_RTR0PTR); + AssertReturn(HMIsEnabled(pVM), NIL_RTR0PTR); + + /* + * Select the new switcher. + */ + const PVMMSWITCHERDEF *papSwitchers = g_apHmSwitchers; + PVMMSWITCHERDEF pSwitcher = papSwitchers[enmSwitcher]; + if (pSwitcher) + { + /** @todo fix the pvCoreCodeR0 type */ + RTR0PTR pbCodeR0 = (RTR0PTR)pVM->vmm.s.pvCoreCodeR0 + pVM->vmm.s.aoffSwitchers[enmSwitcher]; + return pbCodeR0 + pSwitcher->offR0ToRawMode; + } + return NIL_RTR0PTR; +} +#endif + |