summaryrefslogtreecommitdiffstats
path: root/src/VBox/VMM/include/EMHandleRCTmpl.h
diff options
context:
space:
mode:
Diffstat (limited to 'src/VBox/VMM/include/EMHandleRCTmpl.h')
-rw-r--r--src/VBox/VMM/include/EMHandleRCTmpl.h416
1 files changed, 416 insertions, 0 deletions
diff --git a/src/VBox/VMM/include/EMHandleRCTmpl.h b/src/VBox/VMM/include/EMHandleRCTmpl.h
new file mode 100644
index 00000000..02683b66
--- /dev/null
+++ b/src/VBox/VMM/include/EMHandleRCTmpl.h
@@ -0,0 +1,416 @@
+/* $Id: EMHandleRCTmpl.h $ */
+/** @file
+ * EM - emR3[Raw|Hm|Nem]HandleRC template.
+ */
+
+/*
+ * Copyright (C) 2006-2019 Oracle Corporation
+ *
+ * This file is part of VirtualBox Open Source Edition (OSE), as
+ * available from http://www.virtualbox.org. This file is free software;
+ * you can redistribute it and/or modify it under the terms of the GNU
+ * General Public License (GPL) as published by the Free Software
+ * Foundation, in version 2 as it comes in the "COPYING" file of the
+ * VirtualBox OSE distribution. VirtualBox OSE is distributed in the
+ * hope that it will be useful, but WITHOUT ANY WARRANTY of any kind.
+ */
+
+#ifndef VMM_INCLUDED_SRC_include_EMHandleRCTmpl_h
+#define VMM_INCLUDED_SRC_include_EMHandleRCTmpl_h
+#ifndef RT_WITHOUT_PRAGMA_ONCE
+# pragma once
+#endif
+
+#if defined(EMHANDLERC_WITH_PATM) + defined(EMHANDLERC_WITH_HM) + defined(EMHANDLERC_WITH_NEM) != 1
+# error "Exactly one of these must be defined: EMHANDLERC_WITH_PATM, EMHANDLERC_WITH_HM, EMHANDLERC_WITH_NEM"
+#endif
+
+
+/**
+ * Process a subset of the raw-mode, HM and NEM return codes.
+ *
+ * Since we have to share this with raw-mode single stepping, this inline
+ * function has been created to avoid code duplication.
+ *
+ * @returns VINF_SUCCESS if it's ok to continue raw mode.
+ * @returns VBox status code to return to the EM main loop.
+ *
+ * @param pVM The cross context VM structure.
+ * @param pVCpu The cross context virtual CPU structure.
+ * @param rc The return code.
+ */
+#ifdef EMHANDLERC_WITH_PATM
+int emR3RawHandleRC(PVM pVM, PVMCPU pVCpu, int rc)
+#elif defined(EMHANDLERC_WITH_HM) || defined(DOXYGEN_RUNNING)
+int emR3HmHandleRC(PVM pVM, PVMCPU pVCpu, int rc)
+#elif defined(EMHANDLERC_WITH_NEM)
+int emR3NemHandleRC(PVM pVM, PVMCPU pVCpu, int rc)
+#endif
+{
+ switch (rc)
+ {
+ /*
+ * Common & simple ones.
+ */
+ case VINF_SUCCESS:
+ break;
+ case VINF_EM_RESCHEDULE_RAW:
+ case VINF_EM_RESCHEDULE_HM:
+ case VINF_EM_RAW_INTERRUPT:
+ case VINF_EM_RAW_TO_R3:
+ case VINF_EM_RAW_TIMER_PENDING:
+ case VINF_EM_PENDING_REQUEST:
+ rc = VINF_SUCCESS;
+ break;
+
+#ifdef EMHANDLERC_WITH_PATM
+ /*
+ * Privileged instruction.
+ */
+ case VINF_EM_RAW_EXCEPTION_PRIVILEGED:
+ case VINF_PATM_PATCH_TRAP_GP:
+ rc = emR3RawPrivileged(pVM, pVCpu);
+ break;
+
+ case VINF_EM_RAW_GUEST_TRAP:
+ /*
+ * Got a trap which needs dispatching.
+ */
+ if (PATMR3IsInsidePatchJump(pVM, pVCpu->cpum.GstCtx.eip, NULL))
+ {
+ AssertReleaseMsgFailed(("FATAL ERROR: executing random instruction inside generated patch jump %08X\n", CPUMGetGuestEIP(pVCpu)));
+ rc = VERR_EM_RAW_PATCH_CONFLICT;
+ break;
+ }
+ rc = emR3RawGuestTrap(pVM, pVCpu);
+ break;
+
+ /*
+ * Trap in patch code.
+ */
+ case VINF_PATM_PATCH_TRAP_PF:
+ case VINF_PATM_PATCH_INT3:
+ rc = emR3RawPatchTrap(pVM, pVCpu, rc);
+ break;
+
+ case VINF_PATM_DUPLICATE_FUNCTION:
+ Assert(PATMIsPatchGCAddr(pVM, pVCpu->cpum.GstCtx.eip));
+ rc = PATMR3DuplicateFunctionRequest(pVM, &pVCpu->cpum.GstCtx);
+ AssertRC(rc);
+ rc = VINF_SUCCESS;
+ break;
+
+ case VINF_PATM_CHECK_PATCH_PAGE:
+ rc = PATMR3HandleMonitoredPage(pVM);
+ AssertRC(rc);
+ rc = VINF_SUCCESS;
+ break;
+
+ /*
+ * Patch manager.
+ */
+ case VERR_EM_RAW_PATCH_CONFLICT:
+ AssertReleaseMsgFailed(("%Rrc handling is not yet implemented\n", rc));
+ break;
+
+ /*
+ * Memory mapped I/O access - attempt to patch the instruction
+ */
+ case VINF_PATM_HC_MMIO_PATCH_READ:
+ rc = PATMR3InstallPatch(pVM, SELMToFlat(pVM, DISSELREG_CS, CPUMCTX2CORE(&pVCpu->cpum.GstCtx), pVCpu->cpum.GstCtx.eip),
+ PATMFL_MMIO_ACCESS
+ | (CPUMGetGuestCodeBits(pVCpu) == 32 ? PATMFL_CODE32 : 0));
+ if (RT_FAILURE(rc))
+ rc = emR3ExecuteInstruction(pVM, pVCpu, "MMIO");
+ break;
+
+ case VINF_PATM_HC_MMIO_PATCH_WRITE:
+ AssertFailed(); /* not yet implemented. */
+ rc = emR3ExecuteInstruction(pVM, pVCpu, "MMIO");
+ break;
+#endif /* EMHANDLERC_WITH_PATM */
+
+#ifndef EMHANDLERC_WITH_NEM
+ /*
+ * Conflict or out of page tables.
+ *
+ * VM_FF_PGM_SYNC_CR3 is set by the hypervisor and all we need to
+ * do here is to execute the pending forced actions.
+ */
+ case VINF_PGM_SYNC_CR3:
+ AssertMsg(VMCPU_FF_IS_ANY_SET(pVCpu, VMCPU_FF_PGM_SYNC_CR3 | VMCPU_FF_PGM_SYNC_CR3_NON_GLOBAL),
+ ("VINF_PGM_SYNC_CR3 and no VMCPU_FF_PGM_SYNC_CR3*!\n"));
+ rc = VINF_SUCCESS;
+ break;
+
+ /*
+ * PGM pool flush pending (guest SMP only).
+ */
+ /** @todo jumping back and forth between ring 0 and 3 can burn a lot of cycles
+ * if the EMT thread that's supposed to handle the flush is currently not active
+ * (e.g. waiting to be scheduled) -> fix this properly!
+ *
+ * bird: Since the clearing is global and done via a rendezvous any CPU can do
+ * it. They would have to choose who to call VMMR3EmtRendezvous and send
+ * the rest to VMMR3EmtRendezvousFF ... Hmm ... that's not going to work
+ * all that well since the latter will race the setup done by the
+ * first. Guess that means we need some new magic in that area for
+ * handling this case. :/
+ */
+ case VINF_PGM_POOL_FLUSH_PENDING:
+ rc = VINF_SUCCESS;
+ break;
+
+ /*
+ * Paging mode change.
+ */
+ case VINF_PGM_CHANGE_MODE:
+ CPUM_ASSERT_NOT_EXTRN(pVCpu, CPUMCTX_EXTRN_CR0 | CPUMCTX_EXTRN_CR3 | CPUMCTX_EXTRN_CR4 | CPUMCTX_EXTRN_EFER);
+ rc = PGMChangeMode(pVCpu, pVCpu->cpum.GstCtx.cr0, pVCpu->cpum.GstCtx.cr4, pVCpu->cpum.GstCtx.msrEFER);
+ if (rc == VINF_SUCCESS)
+ rc = VINF_EM_RESCHEDULE;
+ AssertMsg(RT_FAILURE(rc) || (rc >= VINF_EM_FIRST && rc <= VINF_EM_LAST), ("%Rrc\n", rc));
+ break;
+#endif /* !EMHANDLERC_WITH_NEM */
+
+#ifdef EMHANDLERC_WITH_PATM
+ /*
+ * CSAM wants to perform a task in ring-3. It has set an FF action flag.
+ */
+ case VINF_CSAM_PENDING_ACTION:
+ rc = VINF_SUCCESS;
+ break;
+
+ /*
+ * Invoked Interrupt gate - must directly (!) go to the recompiler.
+ */
+ case VINF_EM_RAW_INTERRUPT_PENDING:
+ case VINF_EM_RAW_RING_SWITCH_INT:
+ Assert(TRPMHasTrap(pVCpu));
+ Assert(!PATMIsPatchGCAddr(pVM, pVCpu->cpum.GstCtx.eip));
+
+ if (TRPMHasTrap(pVCpu))
+ {
+ /* If the guest gate is marked unpatched, then we will check again if we can patch it. */
+ uint8_t u8Interrupt = TRPMGetTrapNo(pVCpu);
+ if (TRPMR3GetGuestTrapHandler(pVM, u8Interrupt) == TRPM_INVALID_HANDLER)
+ {
+ CSAMR3CheckGates(pVM, u8Interrupt, 1);
+ Log(("emR3RawHandleRC: recheck gate %x -> valid=%d\n", u8Interrupt, TRPMR3GetGuestTrapHandler(pVM, u8Interrupt) != TRPM_INVALID_HANDLER));
+ /* Note: If it was successful, then we could go back to raw mode, but let's keep things simple for now. */
+ }
+ }
+ rc = VINF_EM_RESCHEDULE_REM;
+ break;
+
+ /*
+ * Other ring switch types.
+ */
+ case VINF_EM_RAW_RING_SWITCH:
+ rc = emR3RawRingSwitch(pVM, pVCpu);
+ break;
+#endif /* EMHANDLERC_WITH_PATM */
+
+ /*
+ * I/O Port access - emulate the instruction.
+ */
+ case VINF_IOM_R3_IOPORT_READ:
+ case VINF_IOM_R3_IOPORT_WRITE:
+ case VINF_EM_RESUME_R3_HISTORY_EXEC: /* Resume EMHistoryExec after VMCPU_FF_IOM. */
+ rc = emR3ExecuteIOInstruction(pVM, pVCpu);
+ break;
+
+ /*
+ * Execute pending I/O Port access.
+ */
+ case VINF_EM_PENDING_R3_IOPORT_WRITE:
+ rc = VBOXSTRICTRC_TODO(emR3ExecutePendingIoPortWrite(pVM, pVCpu));
+ break;
+ case VINF_EM_PENDING_R3_IOPORT_READ:
+ rc = VBOXSTRICTRC_TODO(emR3ExecutePendingIoPortRead(pVM, pVCpu));
+ break;
+
+ /*
+ * Memory mapped I/O access - emulate the instruction.
+ */
+ case VINF_IOM_R3_MMIO_READ:
+ case VINF_IOM_R3_MMIO_WRITE:
+ case VINF_IOM_R3_MMIO_READ_WRITE:
+ rc = emR3ExecuteInstruction(pVM, pVCpu, "MMIO");
+ break;
+
+ /*
+ * Machine specific register access - emulate the instruction.
+ */
+ case VINF_CPUM_R3_MSR_READ:
+ case VINF_CPUM_R3_MSR_WRITE:
+ rc = emR3ExecuteInstruction(pVM, pVCpu, "MSR");
+ break;
+
+ /*
+ * GIM hypercall.
+ */
+ case VINF_GIM_R3_HYPERCALL:
+ rc = emR3ExecuteInstruction(pVM, pVCpu, "Hypercall");
+ break;
+
+#ifdef EMHANDLERC_WITH_HM
+ case VINF_EM_HM_PATCH_TPR_INSTR:
+ rc = HMR3PatchTprInstr(pVM, pVCpu);
+ break;
+#endif
+
+#ifdef EMHANDLERC_WITH_PATM
+ /*
+ * Execute instruction.
+ */
+ case VINF_EM_RAW_EMULATE_INSTR_LDT_FAULT:
+ rc = emR3ExecuteInstruction(pVM, pVCpu, "LDT FAULT: ");
+ break;
+ case VINF_EM_RAW_EMULATE_INSTR_GDT_FAULT:
+ rc = emR3ExecuteInstruction(pVM, pVCpu, "GDT FAULT: ");
+ break;
+ case VINF_EM_RAW_EMULATE_INSTR_IDT_FAULT:
+ rc = emR3ExecuteInstruction(pVM, pVCpu, "IDT FAULT: ");
+ break;
+ case VINF_EM_RAW_EMULATE_INSTR_TSS_FAULT:
+ rc = emR3ExecuteInstruction(pVM, pVCpu, "TSS FAULT: ");
+ break;
+
+ case VINF_PATM_PENDING_IRQ_AFTER_IRET:
+ rc = emR3ExecuteInstruction(pVM, pVCpu, "EMUL: ", VINF_PATM_PENDING_IRQ_AFTER_IRET);
+ break;
+
+ case VINF_PATCH_EMULATE_INSTR:
+#else
+ case VINF_EM_RAW_GUEST_TRAP:
+#endif
+ case VINF_EM_RAW_EMULATE_INSTR:
+ rc = emR3ExecuteInstruction(pVM, pVCpu, "EMUL: ");
+ break;
+
+ case VINF_EM_RAW_INJECT_TRPM_EVENT:
+ CPUM_IMPORT_EXTRN_RET(pVCpu, IEM_CPUMCTX_EXTRN_XCPT_MASK);
+ rc = VBOXSTRICTRC_VAL(IEMInjectTrpmEvent(pVCpu));
+ /* The following condition should be removed when IEM_IMPLEMENTS_TASKSWITCH becomes true. */
+ if (rc == VERR_IEM_ASPECT_NOT_IMPLEMENTED)
+ rc = emR3ExecuteInstruction(pVM, pVCpu, "EVENT: ");
+ break;
+
+
+#ifdef EMHANDLERC_WITH_PATM
+ /*
+ * Stale selector and iret traps => REM.
+ */
+ case VINF_EM_RAW_STALE_SELECTOR:
+ case VINF_EM_RAW_IRET_TRAP:
+ /* We will not go to the recompiler if EIP points to patch code. */
+ if (PATMIsPatchGCAddr(pVM, pVCpu->cpum.GstCtx.eip))
+ {
+ pVCpu->cpum.GstCtx.eip = PATMR3PatchToGCPtr(pVM, (RTGCPTR)pVCpu->cpum.GstCtx.eip, 0);
+ }
+ LogFlow(("emR3RawHandleRC: %Rrc -> %Rrc\n", rc, VINF_EM_RESCHEDULE_REM));
+ rc = VINF_EM_RESCHEDULE_REM;
+ break;
+
+ /*
+ * Conflict in GDT, resync and continue.
+ */
+ case VINF_SELM_SYNC_GDT:
+ AssertMsg(VMCPU_FF_IS_ANY_SET(pVCpu, VMCPU_FF_SELM_SYNC_GDT | VMCPU_FF_SELM_SYNC_LDT | VMCPU_FF_SELM_SYNC_TSS),
+ ("VINF_SELM_SYNC_GDT without VMCPU_FF_SELM_SYNC_GDT/LDT/TSS!\n"));
+ rc = VINF_SUCCESS;
+ break;
+#endif
+
+ /*
+ * Up a level.
+ */
+ case VINF_EM_TERMINATE:
+ case VINF_EM_OFF:
+ case VINF_EM_RESET:
+ case VINF_EM_SUSPEND:
+ case VINF_EM_HALT:
+ case VINF_EM_RESUME:
+ case VINF_EM_NO_MEMORY:
+ case VINF_EM_RESCHEDULE:
+ case VINF_EM_RESCHEDULE_REM:
+ case VINF_EM_WAIT_SIPI:
+ break;
+
+ /*
+ * Up a level and invoke the debugger.
+ */
+ case VINF_EM_DBG_STEPPED:
+ case VINF_EM_DBG_BREAKPOINT:
+ case VINF_EM_DBG_STEP:
+ case VINF_EM_DBG_HYPER_BREAKPOINT:
+ case VINF_EM_DBG_HYPER_STEPPED:
+ case VINF_EM_DBG_HYPER_ASSERTION:
+ case VINF_EM_DBG_STOP:
+ case VINF_EM_DBG_EVENT:
+ break;
+
+ /*
+ * Up a level, dump and debug.
+ */
+ case VERR_TRPM_DONT_PANIC:
+ case VERR_TRPM_PANIC:
+ case VERR_VMM_RING0_ASSERTION:
+ case VINF_EM_TRIPLE_FAULT:
+ case VERR_VMM_HYPER_CR3_MISMATCH:
+ case VERR_VMM_RING3_CALL_DISABLED:
+ case VERR_IEM_INSTR_NOT_IMPLEMENTED:
+ case VERR_IEM_ASPECT_NOT_IMPLEMENTED:
+ case VERR_EM_GUEST_CPU_HANG:
+ break;
+
+#ifdef EMHANDLERC_WITH_HM
+ /*
+ * Up a level, after Hm have done some release logging.
+ */
+ case VERR_VMX_INVALID_VMCS_FIELD:
+ case VERR_VMX_INVALID_VMCS_PTR:
+ case VERR_VMX_INVALID_VMXON_PTR:
+ case VERR_VMX_UNEXPECTED_INTERRUPTION_EXIT_TYPE:
+ case VERR_VMX_UNEXPECTED_EXCEPTION:
+ case VERR_VMX_UNEXPECTED_EXIT:
+ case VERR_VMX_INVALID_GUEST_STATE:
+ case VERR_VMX_UNABLE_TO_START_VM:
+ case VERR_SVM_UNKNOWN_EXIT:
+ case VERR_SVM_UNEXPECTED_EXIT:
+ case VERR_SVM_UNEXPECTED_PATCH_TYPE:
+ case VERR_SVM_UNEXPECTED_XCPT_EXIT:
+ HMR3CheckError(pVM, rc);
+ break;
+
+ /* Up a level; fatal */
+ case VERR_VMX_IN_VMX_ROOT_MODE:
+ case VERR_SVM_IN_USE:
+ case VERR_SVM_UNABLE_TO_START_VM:
+ break;
+#endif
+
+ /*
+ * These two should be handled via the force flag already, but just in
+ * case they end up here deal with it.
+ */
+ case VINF_IOM_R3_IOPORT_COMMIT_WRITE:
+ case VINF_IOM_R3_MMIO_COMMIT_WRITE:
+ AssertFailed();
+ rc = VBOXSTRICTRC_TODO(IOMR3ProcessForceFlag(pVM, pVCpu, rc));
+ break;
+
+ /*
+ * Anything which is not known to us means an internal error
+ * and the termination of the VM!
+ */
+ default:
+ AssertMsgFailed(("Unknown GC return code: %Rra\n", rc));
+ break;
+ }
+ return rc;
+}
+
+#endif /* !VMM_INCLUDED_SRC_include_EMHandleRCTmpl_h */
+