summaryrefslogtreecommitdiffstats
path: root/src/VBox/VMM/VMMAll/IEMAllCImplStrInstr.cpp.h
diff options
context:
space:
mode:
Diffstat (limited to 'src/VBox/VMM/VMMAll/IEMAllCImplStrInstr.cpp.h')
-rw-r--r--src/VBox/VMM/VMMAll/IEMAllCImplStrInstr.cpp.h1762
1 files changed, 1762 insertions, 0 deletions
diff --git a/src/VBox/VMM/VMMAll/IEMAllCImplStrInstr.cpp.h b/src/VBox/VMM/VMMAll/IEMAllCImplStrInstr.cpp.h
new file mode 100644
index 00000000..d5dc1340
--- /dev/null
+++ b/src/VBox/VMM/VMMAll/IEMAllCImplStrInstr.cpp.h
@@ -0,0 +1,1762 @@
+/* $Id: IEMAllCImplStrInstr.cpp.h $ */
+/** @file
+ * IEM - String Instruction Implementation Code Template.
+ */
+
+/*
+ * Copyright (C) 2011-2020 Oracle Corporation
+ *
+ * This file is part of VirtualBox Open Source Edition (OSE), as
+ * available from http://www.virtualbox.org. This file is free software;
+ * you can redistribute it and/or modify it under the terms of the GNU
+ * General Public License (GPL) as published by the Free Software
+ * Foundation, in version 2 as it comes in the "COPYING" file of the
+ * VirtualBox OSE distribution. VirtualBox OSE is distributed in the
+ * hope that it will be useful, but WITHOUT ANY WARRANTY of any kind.
+ */
+
+
+/*******************************************************************************
+* Defined Constants And Macros *
+*******************************************************************************/
+#if OP_SIZE == 8
+# define OP_rAX al
+#elif OP_SIZE == 16
+# define OP_rAX ax
+#elif OP_SIZE == 32
+# define OP_rAX eax
+#elif OP_SIZE == 64
+# define OP_rAX rax
+#else
+# error "Bad OP_SIZE."
+#endif
+#define OP_TYPE RT_CONCAT3(uint,OP_SIZE,_t)
+
+#if ADDR_SIZE == 16
+# define ADDR_rDI di
+# define ADDR_rSI si
+# define ADDR_rCX cx
+# define ADDR2_TYPE uint32_t
+# define ADDR_VMXSTRIO 0
+#elif ADDR_SIZE == 32
+# define ADDR_rDI edi
+# define ADDR_rSI esi
+# define ADDR_rCX ecx
+# define ADDR2_TYPE uint32_t
+# define ADDR_VMXSTRIO 1
+#elif ADDR_SIZE == 64
+# define ADDR_rDI rdi
+# define ADDR_rSI rsi
+# define ADDR_rCX rcx
+# define ADDR2_TYPE uint64_t
+# define ADDR_VMXSTRIO 2
+# define IS_64_BIT_CODE(a_pVCpu) (true)
+#else
+# error "Bad ADDR_SIZE."
+#endif
+#define ADDR_TYPE RT_CONCAT3(uint,ADDR_SIZE,_t)
+
+#if ADDR_SIZE == 64 || OP_SIZE == 64
+# define IS_64_BIT_CODE(a_pVCpu) (true)
+#elif ADDR_SIZE == 32
+# define IS_64_BIT_CODE(a_pVCpu) ((a_pVCpu)->iem.s.enmCpuMode == IEMMODE_64BIT)
+#else
+# define IS_64_BIT_CODE(a_pVCpu) (false)
+#endif
+
+/** @def IEM_CHECK_FF_YIELD_REPSTR_MAYBE_RETURN
+ * Used in the outer (page-by-page) loop to check for reasons for returnning
+ * before completing the instruction. In raw-mode we temporarily enable
+ * interrupts to let the host interrupt us. We cannot let big string operations
+ * hog the CPU, especially not in raw-mode.
+ */
+#define IEM_CHECK_FF_YIELD_REPSTR_MAYBE_RETURN(a_pVM, a_pVCpu, a_fEflags) \
+ do { \
+ if (RT_LIKELY( !VMCPU_FF_IS_ANY_SET(a_pVCpu, (a_fEflags) & X86_EFL_IF ? VMCPU_FF_YIELD_REPSTR_MASK \
+ : VMCPU_FF_YIELD_REPSTR_NOINT_MASK) \
+ && !VM_FF_IS_ANY_SET(a_pVM, VM_FF_YIELD_REPSTR_MASK) \
+ )) \
+ { /* probable */ } \
+ else \
+ { \
+ LogFlow(("%s: Leaving early (outer)! ffcpu=%#RX64 ffvm=%#x\n", \
+ __FUNCTION__, (uint64_t)(a_pVCpu)->fLocalForcedActions, (a_pVM)->fGlobalForcedActions)); \
+ return VINF_SUCCESS; \
+ } \
+ } while (0)
+
+/** @def IEM_CHECK_FF_HIGH_PRIORITY_POST_REPSTR_MAYBE_RETURN
+ * This is used in some of the inner loops to make sure we respond immediately
+ * to VMCPU_FF_IOM as well as outside requests. Use this for expensive
+ * instructions. Use IEM_CHECK_FF_CPU_HIGH_PRIORITY_POST_REPSTR_MAYBE_RETURN for
+ * ones that are typically cheap. */
+#define IEM_CHECK_FF_HIGH_PRIORITY_POST_REPSTR_MAYBE_RETURN(a_pVM, a_pVCpu, a_fExitExpr) \
+ do { \
+ if (RT_LIKELY( ( !VMCPU_FF_IS_ANY_SET(a_pVCpu, VMCPU_FF_HIGH_PRIORITY_POST_REPSTR_MASK) \
+ && !VM_FF_IS_ANY_SET(a_pVM, VM_FF_HIGH_PRIORITY_POST_REPSTR_MASK)) \
+ || (a_fExitExpr) )) \
+ { /* very likely */ } \
+ else \
+ { \
+ LogFlow(("%s: Leaving early (inner)! ffcpu=%#RX64 ffvm=%#x\n", \
+ __FUNCTION__, (uint64_t)(a_pVCpu)->fLocalForcedActions, (a_pVM)->fGlobalForcedActions)); \
+ return VINF_SUCCESS; \
+ } \
+ } while (0)
+
+
+/** @def IEM_CHECK_FF_CPU_HIGH_PRIORITY_POST_REPSTR_MAYBE_RETURN
+ * This is used in the inner loops where
+ * IEM_CHECK_FF_HIGH_PRIORITY_POST_REPSTR_MAYBE_RETURN isn't used. It only
+ * checks the CPU FFs so that we respond immediately to the pending IOM FF
+ * (status code is hidden in IEMCPU::rcPassUp by IEM memory commit code).
+ */
+#define IEM_CHECK_FF_CPU_HIGH_PRIORITY_POST_REPSTR_MAYBE_RETURN(a_pVM, a_pVCpu, a_fExitExpr) \
+ do { \
+ if (RT_LIKELY( !VMCPU_FF_IS_ANY_SET(a_pVCpu, VMCPU_FF_HIGH_PRIORITY_POST_REPSTR_MASK) \
+ || (a_fExitExpr) )) \
+ { /* very likely */ } \
+ else \
+ { \
+ LogFlow(("%s: Leaving early (inner)! ffcpu=%#RX64 (ffvm=%#x)\n", \
+ __FUNCTION__, (uint64_t)(a_pVCpu)->fLocalForcedActions, (a_pVM)->fGlobalForcedActions)); \
+ return VINF_SUCCESS; \
+ } \
+ } while (0)
+
+
+/**
+ * Implements 'REPE CMPS'.
+ */
+IEM_CIMPL_DEF_1(RT_CONCAT4(iemCImpl_repe_cmps_op,OP_SIZE,_addr,ADDR_SIZE), uint8_t, iEffSeg)
+{
+ PVM pVM = pVCpu->CTX_SUFF(pVM);
+
+ /*
+ * Setup.
+ */
+ ADDR_TYPE uCounterReg = pVCpu->cpum.GstCtx.ADDR_rCX;
+ if (uCounterReg == 0)
+ {
+ iemRegAddToRipAndClearRF(pVCpu, cbInstr);
+ return VINF_SUCCESS;
+ }
+
+ IEM_CTX_IMPORT_RET(pVCpu, CPUMCTX_EXTRN_SREG_FROM_IDX(iEffSeg) | CPUMCTX_EXTRN_ES);
+
+ PCCPUMSELREGHID pSrc1Hid = iemSRegGetHid(pVCpu, iEffSeg);
+ uint64_t uSrc1Base;
+ VBOXSTRICTRC rcStrict = iemMemSegCheckReadAccessEx(pVCpu, pSrc1Hid, iEffSeg, &uSrc1Base);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ uint64_t uSrc2Base;
+ rcStrict = iemMemSegCheckReadAccessEx(pVCpu, iemSRegUpdateHid(pVCpu, &pVCpu->cpum.GstCtx.es), X86_SREG_ES, &uSrc2Base);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ int8_t const cbIncr = pVCpu->cpum.GstCtx.eflags.Bits.u1DF ? -(OP_SIZE / 8) : (OP_SIZE / 8);
+ ADDR_TYPE uSrc1AddrReg = pVCpu->cpum.GstCtx.ADDR_rSI;
+ ADDR_TYPE uSrc2AddrReg = pVCpu->cpum.GstCtx.ADDR_rDI;
+ uint32_t uEFlags = pVCpu->cpum.GstCtx.eflags.u;
+
+ /*
+ * The loop.
+ */
+ for (;;)
+ {
+ /*
+ * Do segmentation and virtual page stuff.
+ */
+ ADDR2_TYPE uVirtSrc1Addr = uSrc1AddrReg + (ADDR2_TYPE)uSrc1Base;
+ ADDR2_TYPE uVirtSrc2Addr = uSrc2AddrReg + (ADDR2_TYPE)uSrc2Base;
+ uint32_t cLeftSrc1Page = (PAGE_SIZE - (uVirtSrc1Addr & PAGE_OFFSET_MASK)) / (OP_SIZE / 8);
+ if (cLeftSrc1Page > uCounterReg)
+ cLeftSrc1Page = uCounterReg;
+ uint32_t cLeftSrc2Page = (PAGE_SIZE - (uVirtSrc2Addr & PAGE_OFFSET_MASK)) / (OP_SIZE / 8);
+ uint32_t cLeftPage = RT_MIN(cLeftSrc1Page, cLeftSrc2Page);
+
+ if ( cLeftPage > 0 /* can be null if unaligned, do one fallback round. */
+ && cbIncr > 0 /** @todo Optimize reverse direction string ops. */
+ && ( IS_64_BIT_CODE(pVCpu)
+ || ( uSrc1AddrReg < pSrc1Hid->u32Limit
+ && uSrc1AddrReg + (cLeftPage * (OP_SIZE / 8)) <= pSrc1Hid->u32Limit
+ && uSrc2AddrReg < pVCpu->cpum.GstCtx.es.u32Limit
+ && uSrc2AddrReg + (cLeftPage * (OP_SIZE / 8)) <= pVCpu->cpum.GstCtx.es.u32Limit)
+ )
+ )
+ {
+ RTGCPHYS GCPhysSrc1Mem;
+ rcStrict = iemMemPageTranslateAndCheckAccess(pVCpu, uVirtSrc1Addr, IEM_ACCESS_DATA_R, &GCPhysSrc1Mem);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ RTGCPHYS GCPhysSrc2Mem;
+ rcStrict = iemMemPageTranslateAndCheckAccess(pVCpu, uVirtSrc2Addr, IEM_ACCESS_DATA_R, &GCPhysSrc2Mem);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ /*
+ * If we can map the page without trouble, do a block processing
+ * until the end of the current page.
+ */
+ PGMPAGEMAPLOCK PgLockSrc2Mem;
+ OP_TYPE const *puSrc2Mem;
+ rcStrict = iemMemPageMap(pVCpu, GCPhysSrc2Mem, IEM_ACCESS_DATA_R, (void **)&puSrc2Mem, &PgLockSrc2Mem);
+ if (rcStrict == VINF_SUCCESS)
+ {
+ PGMPAGEMAPLOCK PgLockSrc1Mem;
+ OP_TYPE const *puSrc1Mem;
+ rcStrict = iemMemPageMap(pVCpu, GCPhysSrc1Mem, IEM_ACCESS_DATA_R, (void **)&puSrc1Mem, &PgLockSrc1Mem);
+ if (rcStrict == VINF_SUCCESS)
+ {
+ if (!memcmp(puSrc2Mem, puSrc1Mem, cLeftPage * (OP_SIZE / 8)))
+ {
+ /* All matches, only compare the last itme to get the right eflags. */
+ RT_CONCAT(iemAImpl_cmp_u,OP_SIZE)((OP_TYPE *)&puSrc1Mem[cLeftPage-1], puSrc2Mem[cLeftPage-1], &uEFlags);
+ uSrc1AddrReg += cLeftPage * cbIncr;
+ uSrc2AddrReg += cLeftPage * cbIncr;
+ uCounterReg -= cLeftPage;
+ }
+ else
+ {
+ /* Some mismatch, compare each item (and keep volatile
+ memory in mind). */
+ uint32_t off = 0;
+ do
+ {
+ RT_CONCAT(iemAImpl_cmp_u,OP_SIZE)((OP_TYPE *)&puSrc1Mem[off], puSrc2Mem[off], &uEFlags);
+ off++;
+ } while ( off < cLeftPage
+ && (uEFlags & X86_EFL_ZF));
+ uSrc1AddrReg += cbIncr * off;
+ uSrc2AddrReg += cbIncr * off;
+ uCounterReg -= off;
+ }
+
+ /* Update the registers before looping. */
+ pVCpu->cpum.GstCtx.ADDR_rCX = uCounterReg;
+ pVCpu->cpum.GstCtx.ADDR_rSI = uSrc1AddrReg;
+ pVCpu->cpum.GstCtx.ADDR_rDI = uSrc2AddrReg;
+ pVCpu->cpum.GstCtx.eflags.u = uEFlags;
+
+ iemMemPageUnmap(pVCpu, GCPhysSrc1Mem, IEM_ACCESS_DATA_R, puSrc1Mem, &PgLockSrc1Mem);
+ iemMemPageUnmap(pVCpu, GCPhysSrc2Mem, IEM_ACCESS_DATA_R, puSrc2Mem, &PgLockSrc2Mem);
+ if ( uCounterReg == 0
+ || !(uEFlags & X86_EFL_ZF))
+ break;
+ IEM_CHECK_FF_YIELD_REPSTR_MAYBE_RETURN(pVM, pVCpu, uEFlags);
+ continue;
+ }
+ iemMemPageUnmap(pVCpu, GCPhysSrc2Mem, IEM_ACCESS_DATA_R, puSrc2Mem, &PgLockSrc2Mem);
+ }
+ }
+
+ /*
+ * Fallback - slow processing till the end of the current page.
+ * In the cross page boundrary case we will end up here with cLeftPage
+ * as 0, we execute one loop then.
+ */
+ do
+ {
+ OP_TYPE uValue1;
+ rcStrict = RT_CONCAT(iemMemFetchDataU,OP_SIZE)(pVCpu, &uValue1, iEffSeg, uSrc1AddrReg);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+ OP_TYPE uValue2;
+ rcStrict = RT_CONCAT(iemMemFetchDataU,OP_SIZE)(pVCpu, &uValue2, X86_SREG_ES, uSrc2AddrReg);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+ RT_CONCAT(iemAImpl_cmp_u,OP_SIZE)(&uValue1, uValue2, &uEFlags);
+
+ pVCpu->cpum.GstCtx.ADDR_rSI = uSrc1AddrReg += cbIncr;
+ pVCpu->cpum.GstCtx.ADDR_rDI = uSrc2AddrReg += cbIncr;
+ pVCpu->cpum.GstCtx.ADDR_rCX = --uCounterReg;
+ pVCpu->cpum.GstCtx.eflags.u = uEFlags;
+ cLeftPage--;
+ IEM_CHECK_FF_CPU_HIGH_PRIORITY_POST_REPSTR_MAYBE_RETURN(pVM, pVCpu, uCounterReg == 0 || !(uEFlags & X86_EFL_ZF));
+ } while ( (int32_t)cLeftPage > 0
+ && (uEFlags & X86_EFL_ZF));
+
+ /*
+ * Next page? Must check for interrupts and stuff here.
+ */
+ if ( uCounterReg == 0
+ || !(uEFlags & X86_EFL_ZF))
+ break;
+ IEM_CHECK_FF_YIELD_REPSTR_MAYBE_RETURN(pVM, pVCpu, uEFlags);
+ }
+
+ /*
+ * Done.
+ */
+ iemRegAddToRipAndClearRF(pVCpu, cbInstr);
+ return VINF_SUCCESS;
+}
+
+
+/**
+ * Implements 'REPNE CMPS'.
+ */
+IEM_CIMPL_DEF_1(RT_CONCAT4(iemCImpl_repne_cmps_op,OP_SIZE,_addr,ADDR_SIZE), uint8_t, iEffSeg)
+{
+ PVM pVM = pVCpu->CTX_SUFF(pVM);
+
+ /*
+ * Setup.
+ */
+ ADDR_TYPE uCounterReg = pVCpu->cpum.GstCtx.ADDR_rCX;
+ if (uCounterReg == 0)
+ {
+ iemRegAddToRipAndClearRF(pVCpu, cbInstr);
+ return VINF_SUCCESS;
+ }
+
+ IEM_CTX_IMPORT_RET(pVCpu, CPUMCTX_EXTRN_SREG_FROM_IDX(iEffSeg) | CPUMCTX_EXTRN_ES);
+
+ PCCPUMSELREGHID pSrc1Hid = iemSRegGetHid(pVCpu, iEffSeg);
+ uint64_t uSrc1Base;
+ VBOXSTRICTRC rcStrict = iemMemSegCheckReadAccessEx(pVCpu, pSrc1Hid, iEffSeg, &uSrc1Base);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ uint64_t uSrc2Base;
+ rcStrict = iemMemSegCheckReadAccessEx(pVCpu, iemSRegUpdateHid(pVCpu, &pVCpu->cpum.GstCtx.es), X86_SREG_ES, &uSrc2Base);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ int8_t const cbIncr = pVCpu->cpum.GstCtx.eflags.Bits.u1DF ? -(OP_SIZE / 8) : (OP_SIZE / 8);
+ ADDR_TYPE uSrc1AddrReg = pVCpu->cpum.GstCtx.ADDR_rSI;
+ ADDR_TYPE uSrc2AddrReg = pVCpu->cpum.GstCtx.ADDR_rDI;
+ uint32_t uEFlags = pVCpu->cpum.GstCtx.eflags.u;
+
+ /*
+ * The loop.
+ */
+ for (;;)
+ {
+ /*
+ * Do segmentation and virtual page stuff.
+ */
+ ADDR2_TYPE uVirtSrc1Addr = uSrc1AddrReg + (ADDR2_TYPE)uSrc1Base;
+ ADDR2_TYPE uVirtSrc2Addr = uSrc2AddrReg + (ADDR2_TYPE)uSrc2Base;
+ uint32_t cLeftSrc1Page = (PAGE_SIZE - (uVirtSrc1Addr & PAGE_OFFSET_MASK)) / (OP_SIZE / 8);
+ if (cLeftSrc1Page > uCounterReg)
+ cLeftSrc1Page = uCounterReg;
+ uint32_t cLeftSrc2Page = (PAGE_SIZE - (uVirtSrc2Addr & PAGE_OFFSET_MASK)) / (OP_SIZE / 8);
+ uint32_t cLeftPage = RT_MIN(cLeftSrc1Page, cLeftSrc2Page);
+
+ if ( cLeftPage > 0 /* can be null if unaligned, do one fallback round. */
+ && cbIncr > 0 /** @todo Optimize reverse direction string ops. */
+ && ( IS_64_BIT_CODE(pVCpu)
+ || ( uSrc1AddrReg < pSrc1Hid->u32Limit
+ && uSrc1AddrReg + (cLeftPage * (OP_SIZE / 8)) <= pSrc1Hid->u32Limit
+ && uSrc2AddrReg < pVCpu->cpum.GstCtx.es.u32Limit
+ && uSrc2AddrReg + (cLeftPage * (OP_SIZE / 8)) <= pVCpu->cpum.GstCtx.es.u32Limit)
+ )
+ )
+ {
+ RTGCPHYS GCPhysSrc1Mem;
+ rcStrict = iemMemPageTranslateAndCheckAccess(pVCpu, uVirtSrc1Addr, IEM_ACCESS_DATA_R, &GCPhysSrc1Mem);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ RTGCPHYS GCPhysSrc2Mem;
+ rcStrict = iemMemPageTranslateAndCheckAccess(pVCpu, uVirtSrc2Addr, IEM_ACCESS_DATA_R, &GCPhysSrc2Mem);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ /*
+ * If we can map the page without trouble, do a block processing
+ * until the end of the current page.
+ */
+ OP_TYPE const *puSrc2Mem;
+ PGMPAGEMAPLOCK PgLockSrc2Mem;
+ rcStrict = iemMemPageMap(pVCpu, GCPhysSrc2Mem, IEM_ACCESS_DATA_R, (void **)&puSrc2Mem, &PgLockSrc2Mem);
+ if (rcStrict == VINF_SUCCESS)
+ {
+ OP_TYPE const *puSrc1Mem;
+ PGMPAGEMAPLOCK PgLockSrc1Mem;
+ rcStrict = iemMemPageMap(pVCpu, GCPhysSrc1Mem, IEM_ACCESS_DATA_R, (void **)&puSrc1Mem, &PgLockSrc1Mem);
+ if (rcStrict == VINF_SUCCESS)
+ {
+ if (memcmp(puSrc2Mem, puSrc1Mem, cLeftPage * (OP_SIZE / 8)))
+ {
+ /* All matches, only compare the last item to get the right eflags. */
+ RT_CONCAT(iemAImpl_cmp_u,OP_SIZE)((OP_TYPE *)&puSrc1Mem[cLeftPage-1], puSrc2Mem[cLeftPage-1], &uEFlags);
+ uSrc1AddrReg += cLeftPage * cbIncr;
+ uSrc2AddrReg += cLeftPage * cbIncr;
+ uCounterReg -= cLeftPage;
+ }
+ else
+ {
+ /* Some mismatch, compare each item (and keep volatile
+ memory in mind). */
+ uint32_t off = 0;
+ do
+ {
+ RT_CONCAT(iemAImpl_cmp_u,OP_SIZE)((OP_TYPE *)&puSrc1Mem[off], puSrc2Mem[off], &uEFlags);
+ off++;
+ } while ( off < cLeftPage
+ && !(uEFlags & X86_EFL_ZF));
+ uSrc1AddrReg += cbIncr * off;
+ uSrc2AddrReg += cbIncr * off;
+ uCounterReg -= off;
+ }
+
+ /* Update the registers before looping. */
+ pVCpu->cpum.GstCtx.ADDR_rCX = uCounterReg;
+ pVCpu->cpum.GstCtx.ADDR_rSI = uSrc1AddrReg;
+ pVCpu->cpum.GstCtx.ADDR_rDI = uSrc2AddrReg;
+ pVCpu->cpum.GstCtx.eflags.u = uEFlags;
+
+ iemMemPageUnmap(pVCpu, GCPhysSrc1Mem, IEM_ACCESS_DATA_R, puSrc1Mem, &PgLockSrc1Mem);
+ iemMemPageUnmap(pVCpu, GCPhysSrc2Mem, IEM_ACCESS_DATA_R, puSrc2Mem, &PgLockSrc2Mem);
+ if ( uCounterReg == 0
+ || (uEFlags & X86_EFL_ZF))
+ break;
+ IEM_CHECK_FF_YIELD_REPSTR_MAYBE_RETURN(pVM, pVCpu, uEFlags);
+ continue;
+ }
+ iemMemPageUnmap(pVCpu, GCPhysSrc2Mem, IEM_ACCESS_DATA_R, puSrc2Mem, &PgLockSrc2Mem);
+ }
+ }
+
+ /*
+ * Fallback - slow processing till the end of the current page.
+ * In the cross page boundrary case we will end up here with cLeftPage
+ * as 0, we execute one loop then.
+ */
+ do
+ {
+ OP_TYPE uValue1;
+ rcStrict = RT_CONCAT(iemMemFetchDataU,OP_SIZE)(pVCpu, &uValue1, iEffSeg, uSrc1AddrReg);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+ OP_TYPE uValue2;
+ rcStrict = RT_CONCAT(iemMemFetchDataU,OP_SIZE)(pVCpu, &uValue2, X86_SREG_ES, uSrc2AddrReg);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+ RT_CONCAT(iemAImpl_cmp_u,OP_SIZE)(&uValue1, uValue2, &uEFlags);
+
+ pVCpu->cpum.GstCtx.ADDR_rSI = uSrc1AddrReg += cbIncr;
+ pVCpu->cpum.GstCtx.ADDR_rDI = uSrc2AddrReg += cbIncr;
+ pVCpu->cpum.GstCtx.ADDR_rCX = --uCounterReg;
+ pVCpu->cpum.GstCtx.eflags.u = uEFlags;
+ cLeftPage--;
+ IEM_CHECK_FF_CPU_HIGH_PRIORITY_POST_REPSTR_MAYBE_RETURN(pVM, pVCpu, uCounterReg == 0 || (uEFlags & X86_EFL_ZF));
+ } while ( (int32_t)cLeftPage > 0
+ && !(uEFlags & X86_EFL_ZF));
+
+ /*
+ * Next page? Must check for interrupts and stuff here.
+ */
+ if ( uCounterReg == 0
+ || (uEFlags & X86_EFL_ZF))
+ break;
+ IEM_CHECK_FF_YIELD_REPSTR_MAYBE_RETURN(pVM, pVCpu, uEFlags);
+ }
+
+ /*
+ * Done.
+ */
+ iemRegAddToRipAndClearRF(pVCpu, cbInstr);
+ return VINF_SUCCESS;
+}
+
+
+/**
+ * Implements 'REPE SCAS'.
+ */
+IEM_CIMPL_DEF_0(RT_CONCAT4(iemCImpl_repe_scas_,OP_rAX,_m,ADDR_SIZE))
+{
+ PVM pVM = pVCpu->CTX_SUFF(pVM);
+
+ /*
+ * Setup.
+ */
+ ADDR_TYPE uCounterReg = pVCpu->cpum.GstCtx.ADDR_rCX;
+ if (uCounterReg == 0)
+ {
+ iemRegAddToRipAndClearRF(pVCpu, cbInstr);
+ return VINF_SUCCESS;
+ }
+
+ IEM_CTX_IMPORT_RET(pVCpu, CPUMCTX_EXTRN_ES);
+ uint64_t uBaseAddr;
+ VBOXSTRICTRC rcStrict = iemMemSegCheckReadAccessEx(pVCpu, iemSRegUpdateHid(pVCpu, &pVCpu->cpum.GstCtx.es), X86_SREG_ES, &uBaseAddr);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ int8_t const cbIncr = pVCpu->cpum.GstCtx.eflags.Bits.u1DF ? -(OP_SIZE / 8) : (OP_SIZE / 8);
+ OP_TYPE const uValueReg = pVCpu->cpum.GstCtx.OP_rAX;
+ ADDR_TYPE uAddrReg = pVCpu->cpum.GstCtx.ADDR_rDI;
+ uint32_t uEFlags = pVCpu->cpum.GstCtx.eflags.u;
+
+ /*
+ * The loop.
+ */
+ for (;;)
+ {
+ /*
+ * Do segmentation and virtual page stuff.
+ */
+ ADDR2_TYPE uVirtAddr = uAddrReg + (ADDR2_TYPE)uBaseAddr;
+ uint32_t cLeftPage = (PAGE_SIZE - (uVirtAddr & PAGE_OFFSET_MASK)) / (OP_SIZE / 8);
+ if (cLeftPage > uCounterReg)
+ cLeftPage = uCounterReg;
+ if ( cLeftPage > 0 /* can be null if unaligned, do one fallback round. */
+ && cbIncr > 0 /** @todo Implement reverse direction string ops. */
+ && ( IS_64_BIT_CODE(pVCpu)
+ || ( uAddrReg < pVCpu->cpum.GstCtx.es.u32Limit
+ && uAddrReg + (cLeftPage * (OP_SIZE / 8)) <= pVCpu->cpum.GstCtx.es.u32Limit)
+ )
+ )
+ {
+ RTGCPHYS GCPhysMem;
+ rcStrict = iemMemPageTranslateAndCheckAccess(pVCpu, uVirtAddr, IEM_ACCESS_DATA_R, &GCPhysMem);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ /*
+ * If we can map the page without trouble, do a block processing
+ * until the end of the current page.
+ */
+ PGMPAGEMAPLOCK PgLockMem;
+ OP_TYPE const *puMem;
+ rcStrict = iemMemPageMap(pVCpu, GCPhysMem, IEM_ACCESS_DATA_R, (void **)&puMem, &PgLockMem);
+ if (rcStrict == VINF_SUCCESS)
+ {
+ /* Search till we find a mismatching item. */
+ OP_TYPE uTmpValue;
+ bool fQuit;
+ uint32_t i = 0;
+ do
+ {
+ uTmpValue = puMem[i++];
+ fQuit = uTmpValue != uValueReg;
+ } while (i < cLeftPage && !fQuit);
+
+ /* Update the regs. */
+ RT_CONCAT(iemAImpl_cmp_u,OP_SIZE)((OP_TYPE *)&uValueReg, uTmpValue, &uEFlags);
+ pVCpu->cpum.GstCtx.ADDR_rCX = uCounterReg -= i;
+ pVCpu->cpum.GstCtx.ADDR_rDI = uAddrReg += i * cbIncr;
+ pVCpu->cpum.GstCtx.eflags.u = uEFlags;
+ Assert(!(uEFlags & X86_EFL_ZF) == fQuit);
+ iemMemPageUnmap(pVCpu, GCPhysMem, IEM_ACCESS_DATA_R, puMem, &PgLockMem);
+ if ( fQuit
+ || uCounterReg == 0)
+ break;
+
+ /* If unaligned, we drop thru and do the page crossing access
+ below. Otherwise, do the next page. */
+ if (!(uVirtAddr & (OP_SIZE / 8 - 1)))
+ {
+ IEM_CHECK_FF_YIELD_REPSTR_MAYBE_RETURN(pVM, pVCpu, uEFlags);
+ continue;
+ }
+ cLeftPage = 0;
+ }
+ }
+
+ /*
+ * Fallback - slow processing till the end of the current page.
+ * In the cross page boundrary case we will end up here with cLeftPage
+ * as 0, we execute one loop then.
+ */
+ do
+ {
+ OP_TYPE uTmpValue;
+ rcStrict = RT_CONCAT(iemMemFetchDataU,OP_SIZE)(pVCpu, &uTmpValue, X86_SREG_ES, uAddrReg);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+ RT_CONCAT(iemAImpl_cmp_u,OP_SIZE)((OP_TYPE *)&uValueReg, uTmpValue, &uEFlags);
+
+ pVCpu->cpum.GstCtx.ADDR_rDI = uAddrReg += cbIncr;
+ pVCpu->cpum.GstCtx.ADDR_rCX = --uCounterReg;
+ pVCpu->cpum.GstCtx.eflags.u = uEFlags;
+ cLeftPage--;
+ IEM_CHECK_FF_CPU_HIGH_PRIORITY_POST_REPSTR_MAYBE_RETURN(pVM, pVCpu, uCounterReg == 0 || !(uEFlags & X86_EFL_ZF));
+ } while ( (int32_t)cLeftPage > 0
+ && (uEFlags & X86_EFL_ZF));
+
+ /*
+ * Next page? Must check for interrupts and stuff here.
+ */
+ if ( uCounterReg == 0
+ || !(uEFlags & X86_EFL_ZF))
+ break;
+ IEM_CHECK_FF_YIELD_REPSTR_MAYBE_RETURN(pVM, pVCpu, uEFlags);
+ }
+
+ /*
+ * Done.
+ */
+ iemRegAddToRipAndClearRF(pVCpu, cbInstr);
+ return VINF_SUCCESS;
+}
+
+
+/**
+ * Implements 'REPNE SCAS'.
+ */
+IEM_CIMPL_DEF_0(RT_CONCAT4(iemCImpl_repne_scas_,OP_rAX,_m,ADDR_SIZE))
+{
+ PVM pVM = pVCpu->CTX_SUFF(pVM);
+
+ /*
+ * Setup.
+ */
+ ADDR_TYPE uCounterReg = pVCpu->cpum.GstCtx.ADDR_rCX;
+ if (uCounterReg == 0)
+ {
+ iemRegAddToRipAndClearRF(pVCpu, cbInstr);
+ return VINF_SUCCESS;
+ }
+
+ IEM_CTX_IMPORT_RET(pVCpu, CPUMCTX_EXTRN_ES);
+ uint64_t uBaseAddr;
+ VBOXSTRICTRC rcStrict = iemMemSegCheckReadAccessEx(pVCpu, iemSRegUpdateHid(pVCpu, &pVCpu->cpum.GstCtx.es), X86_SREG_ES, &uBaseAddr);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ int8_t const cbIncr = pVCpu->cpum.GstCtx.eflags.Bits.u1DF ? -(OP_SIZE / 8) : (OP_SIZE / 8);
+ OP_TYPE const uValueReg = pVCpu->cpum.GstCtx.OP_rAX;
+ ADDR_TYPE uAddrReg = pVCpu->cpum.GstCtx.ADDR_rDI;
+ uint32_t uEFlags = pVCpu->cpum.GstCtx.eflags.u;
+
+ /*
+ * The loop.
+ */
+ for (;;)
+ {
+ /*
+ * Do segmentation and virtual page stuff.
+ */
+ ADDR2_TYPE uVirtAddr = uAddrReg + (ADDR2_TYPE)uBaseAddr;
+ uint32_t cLeftPage = (PAGE_SIZE - (uVirtAddr & PAGE_OFFSET_MASK)) / (OP_SIZE / 8);
+ if (cLeftPage > uCounterReg)
+ cLeftPage = uCounterReg;
+ if ( cLeftPage > 0 /* can be null if unaligned, do one fallback round. */
+ && cbIncr > 0 /** @todo Implement reverse direction string ops. */
+ && ( IS_64_BIT_CODE(pVCpu)
+ || ( uAddrReg < pVCpu->cpum.GstCtx.es.u32Limit
+ && uAddrReg + (cLeftPage * (OP_SIZE / 8)) <= pVCpu->cpum.GstCtx.es.u32Limit)
+ )
+ )
+ {
+ RTGCPHYS GCPhysMem;
+ rcStrict = iemMemPageTranslateAndCheckAccess(pVCpu, uVirtAddr, IEM_ACCESS_DATA_R, &GCPhysMem);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ /*
+ * If we can map the page without trouble, do a block processing
+ * until the end of the current page.
+ */
+ PGMPAGEMAPLOCK PgLockMem;
+ OP_TYPE const *puMem;
+ rcStrict = iemMemPageMap(pVCpu, GCPhysMem, IEM_ACCESS_DATA_R, (void **)&puMem, &PgLockMem);
+ if (rcStrict == VINF_SUCCESS)
+ {
+ /* Search till we find a mismatching item. */
+ OP_TYPE uTmpValue;
+ bool fQuit;
+ uint32_t i = 0;
+ do
+ {
+ uTmpValue = puMem[i++];
+ fQuit = uTmpValue == uValueReg;
+ } while (i < cLeftPage && !fQuit);
+
+ /* Update the regs. */
+ RT_CONCAT(iemAImpl_cmp_u,OP_SIZE)((OP_TYPE *)&uValueReg, uTmpValue, &uEFlags);
+ pVCpu->cpum.GstCtx.ADDR_rCX = uCounterReg -= i;
+ pVCpu->cpum.GstCtx.ADDR_rDI = uAddrReg += i * cbIncr;
+ pVCpu->cpum.GstCtx.eflags.u = uEFlags;
+ Assert(!!(uEFlags & X86_EFL_ZF) == fQuit);
+ iemMemPageUnmap(pVCpu, GCPhysMem, IEM_ACCESS_DATA_R, puMem, &PgLockMem);
+ if ( fQuit
+ || uCounterReg == 0)
+ break;
+
+ /* If unaligned, we drop thru and do the page crossing access
+ below. Otherwise, do the next page. */
+ if (!(uVirtAddr & (OP_SIZE / 8 - 1)))
+ {
+ IEM_CHECK_FF_YIELD_REPSTR_MAYBE_RETURN(pVM, pVCpu, uEFlags);
+ continue;
+ }
+ cLeftPage = 0;
+ }
+ }
+
+ /*
+ * Fallback - slow processing till the end of the current page.
+ * In the cross page boundrary case we will end up here with cLeftPage
+ * as 0, we execute one loop then.
+ */
+ do
+ {
+ OP_TYPE uTmpValue;
+ rcStrict = RT_CONCAT(iemMemFetchDataU,OP_SIZE)(pVCpu, &uTmpValue, X86_SREG_ES, uAddrReg);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+ RT_CONCAT(iemAImpl_cmp_u,OP_SIZE)((OP_TYPE *)&uValueReg, uTmpValue, &uEFlags);
+ pVCpu->cpum.GstCtx.ADDR_rDI = uAddrReg += cbIncr;
+ pVCpu->cpum.GstCtx.ADDR_rCX = --uCounterReg;
+ pVCpu->cpum.GstCtx.eflags.u = uEFlags;
+ cLeftPage--;
+ IEM_CHECK_FF_CPU_HIGH_PRIORITY_POST_REPSTR_MAYBE_RETURN(pVM, pVCpu, uCounterReg == 0 || (uEFlags & X86_EFL_ZF));
+ } while ( (int32_t)cLeftPage > 0
+ && !(uEFlags & X86_EFL_ZF));
+
+ /*
+ * Next page? Must check for interrupts and stuff here.
+ */
+ if ( uCounterReg == 0
+ || (uEFlags & X86_EFL_ZF))
+ break;
+ IEM_CHECK_FF_YIELD_REPSTR_MAYBE_RETURN(pVM, pVCpu, uEFlags);
+ }
+
+ /*
+ * Done.
+ */
+ iemRegAddToRipAndClearRF(pVCpu, cbInstr);
+ return VINF_SUCCESS;
+}
+
+
+
+
+/**
+ * Implements 'REP MOVS'.
+ */
+IEM_CIMPL_DEF_1(RT_CONCAT4(iemCImpl_rep_movs_op,OP_SIZE,_addr,ADDR_SIZE), uint8_t, iEffSeg)
+{
+ PVM pVM = pVCpu->CTX_SUFF(pVM);
+
+ /*
+ * Setup.
+ */
+ ADDR_TYPE uCounterReg = pVCpu->cpum.GstCtx.ADDR_rCX;
+ if (uCounterReg == 0)
+ {
+ iemRegAddToRipAndClearRF(pVCpu, cbInstr);
+ return VINF_SUCCESS;
+ }
+
+ IEM_CTX_IMPORT_RET(pVCpu, CPUMCTX_EXTRN_SREG_FROM_IDX(iEffSeg) | CPUMCTX_EXTRN_ES);
+
+ PCCPUMSELREGHID pSrcHid = iemSRegGetHid(pVCpu, iEffSeg);
+ uint64_t uSrcBase;
+ VBOXSTRICTRC rcStrict = iemMemSegCheckReadAccessEx(pVCpu, pSrcHid, iEffSeg, &uSrcBase);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ uint64_t uDstBase;
+ rcStrict = iemMemSegCheckWriteAccessEx(pVCpu, iemSRegUpdateHid(pVCpu, &pVCpu->cpum.GstCtx.es), X86_SREG_ES, &uDstBase);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ int8_t const cbIncr = pVCpu->cpum.GstCtx.eflags.Bits.u1DF ? -(OP_SIZE / 8) : (OP_SIZE / 8);
+ ADDR_TYPE uSrcAddrReg = pVCpu->cpum.GstCtx.ADDR_rSI;
+ ADDR_TYPE uDstAddrReg = pVCpu->cpum.GstCtx.ADDR_rDI;
+
+ /*
+ * Be careful with handle bypassing.
+ */
+ if (pVCpu->iem.s.fBypassHandlers)
+ {
+ Log(("%s: declining because we're bypassing handlers\n", __FUNCTION__));
+ return VERR_IEM_ASPECT_NOT_IMPLEMENTED;
+ }
+
+ /*
+ * The loop.
+ */
+ for (;;)
+ {
+ /*
+ * Do segmentation and virtual page stuff.
+ */
+ ADDR2_TYPE uVirtSrcAddr = uSrcAddrReg + (ADDR2_TYPE)uSrcBase;
+ ADDR2_TYPE uVirtDstAddr = uDstAddrReg + (ADDR2_TYPE)uDstBase;
+ uint32_t cLeftSrcPage = (PAGE_SIZE - (uVirtSrcAddr & PAGE_OFFSET_MASK)) / (OP_SIZE / 8);
+ if (cLeftSrcPage > uCounterReg)
+ cLeftSrcPage = uCounterReg;
+ uint32_t cLeftDstPage = (PAGE_SIZE - (uVirtDstAddr & PAGE_OFFSET_MASK)) / (OP_SIZE / 8);
+ uint32_t cLeftPage = RT_MIN(cLeftSrcPage, cLeftDstPage);
+
+ if ( cLeftPage > 0 /* can be null if unaligned, do one fallback round. */
+ && cbIncr > 0 /** @todo Implement reverse direction string ops. */
+ && ( IS_64_BIT_CODE(pVCpu)
+ || ( uSrcAddrReg < pSrcHid->u32Limit
+ && uSrcAddrReg + (cLeftPage * (OP_SIZE / 8)) <= pSrcHid->u32Limit
+ && uDstAddrReg < pVCpu->cpum.GstCtx.es.u32Limit
+ && uDstAddrReg + (cLeftPage * (OP_SIZE / 8)) <= pVCpu->cpum.GstCtx.es.u32Limit)
+ )
+ )
+ {
+ RTGCPHYS GCPhysSrcMem;
+ rcStrict = iemMemPageTranslateAndCheckAccess(pVCpu, uVirtSrcAddr, IEM_ACCESS_DATA_R, &GCPhysSrcMem);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ RTGCPHYS GCPhysDstMem;
+ rcStrict = iemMemPageTranslateAndCheckAccess(pVCpu, uVirtDstAddr, IEM_ACCESS_DATA_W, &GCPhysDstMem);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ /*
+ * If we can map the page without trouble, do a block processing
+ * until the end of the current page.
+ */
+ PGMPAGEMAPLOCK PgLockDstMem;
+ OP_TYPE *puDstMem;
+ rcStrict = iemMemPageMap(pVCpu, GCPhysDstMem, IEM_ACCESS_DATA_W, (void **)&puDstMem, &PgLockDstMem);
+ if (rcStrict == VINF_SUCCESS)
+ {
+ PGMPAGEMAPLOCK PgLockSrcMem;
+ OP_TYPE const *puSrcMem;
+ rcStrict = iemMemPageMap(pVCpu, GCPhysSrcMem, IEM_ACCESS_DATA_R, (void **)&puSrcMem, &PgLockSrcMem);
+ if (rcStrict == VINF_SUCCESS)
+ {
+ Assert( (GCPhysSrcMem >> PAGE_SHIFT) != (GCPhysDstMem >> PAGE_SHIFT)
+ || ((uintptr_t)puSrcMem >> PAGE_SHIFT) == ((uintptr_t)puDstMem >> PAGE_SHIFT));
+
+ /* Perform the operation exactly (don't use memcpy to avoid
+ having to consider how its implementation would affect
+ any overlapping source and destination area). */
+ OP_TYPE const *puSrcCur = puSrcMem;
+ OP_TYPE *puDstCur = puDstMem;
+ uint32_t cTodo = cLeftPage;
+ while (cTodo-- > 0)
+ *puDstCur++ = *puSrcCur++;
+
+ /* Update the registers. */
+ pVCpu->cpum.GstCtx.ADDR_rSI = uSrcAddrReg += cLeftPage * cbIncr;
+ pVCpu->cpum.GstCtx.ADDR_rDI = uDstAddrReg += cLeftPage * cbIncr;
+ pVCpu->cpum.GstCtx.ADDR_rCX = uCounterReg -= cLeftPage;
+
+ iemMemPageUnmap(pVCpu, GCPhysSrcMem, IEM_ACCESS_DATA_R, puSrcMem, &PgLockSrcMem);
+ iemMemPageUnmap(pVCpu, GCPhysDstMem, IEM_ACCESS_DATA_W, puDstMem, &PgLockDstMem);
+
+ if (uCounterReg == 0)
+ break;
+ IEM_CHECK_FF_YIELD_REPSTR_MAYBE_RETURN(pVM, pVCpu, pVCpu->cpum.GstCtx.eflags.u);
+ continue;
+ }
+ iemMemPageUnmap(pVCpu, GCPhysDstMem, IEM_ACCESS_DATA_W, puDstMem, &PgLockDstMem);
+ }
+ }
+
+ /*
+ * Fallback - slow processing till the end of the current page.
+ * In the cross page boundrary case we will end up here with cLeftPage
+ * as 0, we execute one loop then.
+ */
+ do
+ {
+ OP_TYPE uValue;
+ rcStrict = RT_CONCAT(iemMemFetchDataU,OP_SIZE)(pVCpu, &uValue, iEffSeg, uSrcAddrReg);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+ rcStrict = RT_CONCAT(iemMemStoreDataU,OP_SIZE)(pVCpu, X86_SREG_ES, uDstAddrReg, uValue);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ pVCpu->cpum.GstCtx.ADDR_rSI = uSrcAddrReg += cbIncr;
+ pVCpu->cpum.GstCtx.ADDR_rDI = uDstAddrReg += cbIncr;
+ pVCpu->cpum.GstCtx.ADDR_rCX = --uCounterReg;
+ cLeftPage--;
+ IEM_CHECK_FF_HIGH_PRIORITY_POST_REPSTR_MAYBE_RETURN(pVM, pVCpu, uCounterReg == 0);
+ } while ((int32_t)cLeftPage > 0);
+
+ /*
+ * Next page. Must check for interrupts and stuff here.
+ */
+ if (uCounterReg == 0)
+ break;
+ IEM_CHECK_FF_YIELD_REPSTR_MAYBE_RETURN(pVM, pVCpu, pVCpu->cpum.GstCtx.eflags.u);
+ }
+
+ /*
+ * Done.
+ */
+ iemRegAddToRipAndClearRF(pVCpu, cbInstr);
+ return VINF_SUCCESS;
+}
+
+
+/**
+ * Implements 'REP STOS'.
+ */
+IEM_CIMPL_DEF_0(RT_CONCAT4(iemCImpl_stos_,OP_rAX,_m,ADDR_SIZE))
+{
+ PVM pVM = pVCpu->CTX_SUFF(pVM);
+
+ /*
+ * Setup.
+ */
+ ADDR_TYPE uCounterReg = pVCpu->cpum.GstCtx.ADDR_rCX;
+ if (uCounterReg == 0)
+ {
+ iemRegAddToRipAndClearRF(pVCpu, cbInstr);
+ return VINF_SUCCESS;
+ }
+
+ IEM_CTX_IMPORT_RET(pVCpu, CPUMCTX_EXTRN_ES);
+
+ uint64_t uBaseAddr;
+ VBOXSTRICTRC rcStrict = iemMemSegCheckWriteAccessEx(pVCpu, iemSRegUpdateHid(pVCpu, &pVCpu->cpum.GstCtx.es), X86_SREG_ES, &uBaseAddr);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ int8_t const cbIncr = pVCpu->cpum.GstCtx.eflags.Bits.u1DF ? -(OP_SIZE / 8) : (OP_SIZE / 8);
+ OP_TYPE const uValue = pVCpu->cpum.GstCtx.OP_rAX;
+ ADDR_TYPE uAddrReg = pVCpu->cpum.GstCtx.ADDR_rDI;
+
+ /*
+ * Be careful with handle bypassing.
+ */
+ /** @todo Permit doing a page if correctly aligned. */
+ if (pVCpu->iem.s.fBypassHandlers)
+ {
+ Log(("%s: declining because we're bypassing handlers\n", __FUNCTION__));
+ return VERR_IEM_ASPECT_NOT_IMPLEMENTED;
+ }
+
+ /*
+ * The loop.
+ */
+ for (;;)
+ {
+ /*
+ * Do segmentation and virtual page stuff.
+ */
+ ADDR2_TYPE uVirtAddr = uAddrReg + (ADDR2_TYPE)uBaseAddr;
+ uint32_t cLeftPage = (PAGE_SIZE - (uVirtAddr & PAGE_OFFSET_MASK)) / (OP_SIZE / 8);
+ if (cLeftPage > uCounterReg)
+ cLeftPage = uCounterReg;
+ if ( cLeftPage > 0 /* can be null if unaligned, do one fallback round. */
+ && cbIncr > 0 /** @todo Implement reverse direction string ops. */
+ && ( IS_64_BIT_CODE(pVCpu)
+ || ( uAddrReg < pVCpu->cpum.GstCtx.es.u32Limit
+ && uAddrReg + (cLeftPage * (OP_SIZE / 8)) <= pVCpu->cpum.GstCtx.es.u32Limit)
+ )
+ )
+ {
+ RTGCPHYS GCPhysMem;
+ rcStrict = iemMemPageTranslateAndCheckAccess(pVCpu, uVirtAddr, IEM_ACCESS_DATA_W, &GCPhysMem);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ /*
+ * If we can map the page without trouble, do a block processing
+ * until the end of the current page.
+ */
+ PGMPAGEMAPLOCK PgLockMem;
+ OP_TYPE *puMem;
+ rcStrict = iemMemPageMap(pVCpu, GCPhysMem, IEM_ACCESS_DATA_W, (void **)&puMem, &PgLockMem);
+ if (rcStrict == VINF_SUCCESS)
+ {
+ /* Update the regs first so we can loop on cLeftPage. */
+ pVCpu->cpum.GstCtx.ADDR_rCX = uCounterReg -= cLeftPage;
+ pVCpu->cpum.GstCtx.ADDR_rDI = uAddrReg += cLeftPage * cbIncr;
+
+ /* Do the memsetting. */
+#if OP_SIZE == 8
+ memset(puMem, uValue, cLeftPage);
+/*#elif OP_SIZE == 32
+ ASMMemFill32(puMem, cLeftPage * (OP_SIZE / 8), uValue);*/
+#else
+ while (cLeftPage-- > 0)
+ *puMem++ = uValue;
+#endif
+
+ iemMemPageUnmap(pVCpu, GCPhysMem, IEM_ACCESS_DATA_W, puMem, &PgLockMem);
+
+ if (uCounterReg == 0)
+ break;
+
+ /* If unaligned, we drop thru and do the page crossing access
+ below. Otherwise, do the next page. */
+ if (!(uVirtAddr & (OP_SIZE / 8 - 1)))
+ {
+ IEM_CHECK_FF_YIELD_REPSTR_MAYBE_RETURN(pVM, pVCpu, pVCpu->cpum.GstCtx.eflags.u);
+ continue;
+ }
+ cLeftPage = 0;
+ }
+ /* If we got an invalid physical address in the page table, just skip
+ ahead to the next page or the counter reaches zero. This crazy
+ optimization is for a buggy EFI firmware that's driving me nuts. */
+ else if (rcStrict == VERR_PGM_PHYS_TLB_UNASSIGNED)
+ {
+ pVCpu->cpum.GstCtx.ADDR_rCX = uCounterReg -= cLeftPage;
+ pVCpu->cpum.GstCtx.ADDR_rDI = uAddrReg += cLeftPage * cbIncr;
+ if (uCounterReg == 0)
+ break;
+ if (!(uVirtAddr & (OP_SIZE / 8 - 1)))
+ {
+ IEM_CHECK_FF_YIELD_REPSTR_MAYBE_RETURN(pVM, pVCpu, pVCpu->cpum.GstCtx.eflags.u);
+ continue;
+ }
+ }
+ }
+
+ /*
+ * Fallback - slow processing till the end of the current page.
+ * In the cross page boundrary case we will end up here with cLeftPage
+ * as 0, we execute one loop then.
+ */
+ do
+ {
+ rcStrict = RT_CONCAT(iemMemStoreDataU,OP_SIZE)(pVCpu, X86_SREG_ES, uAddrReg, uValue);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+ pVCpu->cpum.GstCtx.ADDR_rDI = uAddrReg += cbIncr;
+ pVCpu->cpum.GstCtx.ADDR_rCX = --uCounterReg;
+ cLeftPage--;
+ IEM_CHECK_FF_CPU_HIGH_PRIORITY_POST_REPSTR_MAYBE_RETURN(pVM, pVCpu, uCounterReg == 0);
+ } while ((int32_t)cLeftPage > 0);
+
+ /*
+ * Next page. Must check for interrupts and stuff here.
+ */
+ if (uCounterReg == 0)
+ break;
+ IEM_CHECK_FF_YIELD_REPSTR_MAYBE_RETURN(pVM, pVCpu, pVCpu->cpum.GstCtx.eflags.u);
+ }
+
+ /*
+ * Done.
+ */
+ iemRegAddToRipAndClearRF(pVCpu, cbInstr);
+ return VINF_SUCCESS;
+}
+
+
+/**
+ * Implements 'REP LODS'.
+ */
+IEM_CIMPL_DEF_1(RT_CONCAT4(iemCImpl_lods_,OP_rAX,_m,ADDR_SIZE), int8_t, iEffSeg)
+{
+ PVM pVM = pVCpu->CTX_SUFF(pVM);
+
+ /*
+ * Setup.
+ */
+ ADDR_TYPE uCounterReg = pVCpu->cpum.GstCtx.ADDR_rCX;
+ if (uCounterReg == 0)
+ {
+ iemRegAddToRipAndClearRF(pVCpu, cbInstr);
+ return VINF_SUCCESS;
+ }
+
+ IEM_CTX_IMPORT_RET(pVCpu, CPUMCTX_EXTRN_SREG_FROM_IDX(iEffSeg));
+ PCCPUMSELREGHID pSrcHid = iemSRegGetHid(pVCpu, iEffSeg);
+ uint64_t uBaseAddr;
+ VBOXSTRICTRC rcStrict = iemMemSegCheckReadAccessEx(pVCpu, pSrcHid, iEffSeg, &uBaseAddr);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ int8_t const cbIncr = pVCpu->cpum.GstCtx.eflags.Bits.u1DF ? -(OP_SIZE / 8) : (OP_SIZE / 8);
+ ADDR_TYPE uAddrReg = pVCpu->cpum.GstCtx.ADDR_rSI;
+
+ /*
+ * The loop.
+ */
+ for (;;)
+ {
+ /*
+ * Do segmentation and virtual page stuff.
+ */
+ ADDR2_TYPE uVirtAddr = uAddrReg + (ADDR2_TYPE)uBaseAddr;
+ uint32_t cLeftPage = (PAGE_SIZE - (uVirtAddr & PAGE_OFFSET_MASK)) / (OP_SIZE / 8);
+ if (cLeftPage > uCounterReg)
+ cLeftPage = uCounterReg;
+ if ( cLeftPage > 0 /* can be null if unaligned, do one fallback round. */
+ && cbIncr > 0 /** @todo Implement reverse direction string ops. */
+ && ( IS_64_BIT_CODE(pVCpu)
+ || ( uAddrReg < pSrcHid->u32Limit
+ && uAddrReg + (cLeftPage * (OP_SIZE / 8)) <= pSrcHid->u32Limit)
+ )
+ )
+ {
+ RTGCPHYS GCPhysMem;
+ rcStrict = iemMemPageTranslateAndCheckAccess(pVCpu, uVirtAddr, IEM_ACCESS_DATA_R, &GCPhysMem);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ /*
+ * If we can map the page without trouble, we can get away with
+ * just reading the last value on the page.
+ */
+ PGMPAGEMAPLOCK PgLockMem;
+ OP_TYPE const *puMem;
+ rcStrict = iemMemPageMap(pVCpu, GCPhysMem, IEM_ACCESS_DATA_R, (void **)&puMem, &PgLockMem);
+ if (rcStrict == VINF_SUCCESS)
+ {
+ /* Only get the last byte, the rest doesn't matter in direct access mode. */
+#if OP_SIZE == 32
+ pVCpu->cpum.GstCtx.rax = puMem[cLeftPage - 1];
+#else
+ pVCpu->cpum.GstCtx.OP_rAX = puMem[cLeftPage - 1];
+#endif
+ pVCpu->cpum.GstCtx.ADDR_rCX = uCounterReg -= cLeftPage;
+ pVCpu->cpum.GstCtx.ADDR_rSI = uAddrReg += cLeftPage * cbIncr;
+ iemMemPageUnmap(pVCpu, GCPhysMem, IEM_ACCESS_DATA_R, puMem, &PgLockMem);
+
+ if (uCounterReg == 0)
+ break;
+
+ /* If unaligned, we drop thru and do the page crossing access
+ below. Otherwise, do the next page. */
+ if (!(uVirtAddr & (OP_SIZE / 8 - 1)))
+ {
+ IEM_CHECK_FF_YIELD_REPSTR_MAYBE_RETURN(pVM, pVCpu, pVCpu->cpum.GstCtx.eflags.u);
+ continue;
+ }
+ cLeftPage = 0;
+ }
+ }
+
+ /*
+ * Fallback - slow processing till the end of the current page.
+ * In the cross page boundrary case we will end up here with cLeftPage
+ * as 0, we execute one loop then.
+ */
+ do
+ {
+ OP_TYPE uTmpValue;
+ rcStrict = RT_CONCAT(iemMemFetchDataU,OP_SIZE)(pVCpu, &uTmpValue, iEffSeg, uAddrReg);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+#if OP_SIZE == 32
+ pVCpu->cpum.GstCtx.rax = uTmpValue;
+#else
+ pVCpu->cpum.GstCtx.OP_rAX = uTmpValue;
+#endif
+ pVCpu->cpum.GstCtx.ADDR_rSI = uAddrReg += cbIncr;
+ pVCpu->cpum.GstCtx.ADDR_rCX = --uCounterReg;
+ cLeftPage--;
+ IEM_CHECK_FF_CPU_HIGH_PRIORITY_POST_REPSTR_MAYBE_RETURN(pVM, pVCpu, uCounterReg == 0);
+ } while ((int32_t)cLeftPage > 0);
+
+ if (rcStrict != VINF_SUCCESS)
+ break;
+
+ /*
+ * Next page. Must check for interrupts and stuff here.
+ */
+ if (uCounterReg == 0)
+ break;
+ IEM_CHECK_FF_YIELD_REPSTR_MAYBE_RETURN(pVM, pVCpu, pVCpu->cpum.GstCtx.eflags.u);
+ }
+
+ /*
+ * Done.
+ */
+ iemRegAddToRipAndClearRF(pVCpu, cbInstr);
+ return VINF_SUCCESS;
+}
+
+
+#if OP_SIZE != 64
+
+/**
+ * Implements 'INS' (no rep)
+ */
+IEM_CIMPL_DEF_1(RT_CONCAT4(iemCImpl_ins_op,OP_SIZE,_addr,ADDR_SIZE), bool, fIoChecked)
+{
+ PVMCC pVM = pVCpu->CTX_SUFF(pVM);
+ VBOXSTRICTRC rcStrict;
+
+ /*
+ * Be careful with handle bypassing.
+ */
+ if (pVCpu->iem.s.fBypassHandlers)
+ {
+ Log(("%s: declining because we're bypassing handlers\n", __FUNCTION__));
+ return VERR_IEM_ASPECT_NOT_IMPLEMENTED;
+ }
+
+ /*
+ * ASSUMES the #GP for I/O permission is taken first, then any #GP for
+ * segmentation and finally any #PF due to virtual address translation.
+ * ASSUMES nothing is read from the I/O port before traps are taken.
+ */
+ if (!fIoChecked)
+ {
+ rcStrict = iemHlpCheckPortIOPermission(pVCpu, pVCpu->cpum.GstCtx.dx, OP_SIZE / 8);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+ }
+
+ /*
+ * Check nested-guest I/O intercepts.
+ */
+#ifdef VBOX_WITH_NESTED_HWVIRT_VMX
+ if (IEM_VMX_IS_NON_ROOT_MODE(pVCpu))
+ {
+ VMXEXITINSTRINFO ExitInstrInfo;
+ ExitInstrInfo.u = 0;
+ ExitInstrInfo.StrIo.u3AddrSize = ADDR_VMXSTRIO;
+ ExitInstrInfo.StrIo.iSegReg = X86_SREG_ES;
+ rcStrict = iemVmxVmexitInstrStrIo(pVCpu, VMXINSTRID_IO_INS, pVCpu->cpum.GstCtx.dx, OP_SIZE / 8, false /* fRep */,
+ ExitInstrInfo, cbInstr);
+ if (rcStrict != VINF_VMX_INTERCEPT_NOT_ACTIVE)
+ return rcStrict;
+ }
+#endif
+
+#ifdef VBOX_WITH_NESTED_HWVIRT_SVM
+ if (IEM_SVM_IS_CTRL_INTERCEPT_SET(pVCpu, SVM_CTRL_INTERCEPT_IOIO_PROT))
+ {
+ rcStrict = iemSvmHandleIOIntercept(pVCpu, pVCpu->cpum.GstCtx.dx, SVMIOIOTYPE_IN, OP_SIZE / 8, ADDR_SIZE, X86_SREG_ES,
+ false /* fRep */, true /* fStrIo */, cbInstr);
+ if (rcStrict == VINF_SVM_VMEXIT)
+ return VINF_SUCCESS;
+ if (rcStrict != VINF_SVM_INTERCEPT_NOT_ACTIVE)
+ {
+ Log(("iemCImpl_ins_op: iemSvmHandleIOIntercept failed (u16Port=%#x, cbReg=%u) rc=%Rrc\n", pVCpu->cpum.GstCtx.dx,
+ OP_SIZE / 8, VBOXSTRICTRC_VAL(rcStrict)));
+ return rcStrict;
+ }
+ }
+#endif
+
+ OP_TYPE *puMem;
+ rcStrict = iemMemMap(pVCpu, (void **)&puMem, OP_SIZE / 8, X86_SREG_ES, pVCpu->cpum.GstCtx.ADDR_rDI, IEM_ACCESS_DATA_W);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ uint32_t u32Value = 0;
+ rcStrict = IOMIOPortRead(pVM, pVCpu, pVCpu->cpum.GstCtx.dx, &u32Value, OP_SIZE / 8);
+ if (IOM_SUCCESS(rcStrict))
+ {
+ *puMem = (OP_TYPE)u32Value;
+# ifdef IN_RING3
+ VBOXSTRICTRC rcStrict2 = iemMemCommitAndUnmap(pVCpu, puMem, IEM_ACCESS_DATA_W);
+# else
+ VBOXSTRICTRC rcStrict2 = iemMemCommitAndUnmapPostponeTroubleToR3(pVCpu, puMem, IEM_ACCESS_DATA_W);
+# endif
+ if (RT_LIKELY(rcStrict2 == VINF_SUCCESS))
+ {
+ if (!pVCpu->cpum.GstCtx.eflags.Bits.u1DF)
+ pVCpu->cpum.GstCtx.ADDR_rDI += OP_SIZE / 8;
+ else
+ pVCpu->cpum.GstCtx.ADDR_rDI -= OP_SIZE / 8;
+ iemRegAddToRipAndClearRF(pVCpu, cbInstr);
+ }
+ else
+ AssertLogRelMsgFailedReturn(("rcStrict2=%Rrc\n", VBOXSTRICTRC_VAL(rcStrict2)), RT_FAILURE_NP(rcStrict2) ? rcStrict2 : VERR_IEM_IPE_1);
+ }
+ return rcStrict;
+}
+
+
+/**
+ * Implements 'REP INS'.
+ */
+IEM_CIMPL_DEF_1(RT_CONCAT4(iemCImpl_rep_ins_op,OP_SIZE,_addr,ADDR_SIZE), bool, fIoChecked)
+{
+ PVMCC pVM = pVCpu->CTX_SUFF(pVM);
+
+ IEM_CTX_IMPORT_RET(pVCpu, CPUMCTX_EXTRN_ES | CPUMCTX_EXTRN_TR);
+
+ /*
+ * Setup.
+ */
+ uint16_t const u16Port = pVCpu->cpum.GstCtx.dx;
+ VBOXSTRICTRC rcStrict;
+ if (!fIoChecked)
+ {
+/** @todo check if this is too early for ecx=0. */
+ rcStrict = iemHlpCheckPortIOPermission(pVCpu, u16Port, OP_SIZE / 8);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+ }
+
+ /*
+ * Check nested-guest I/O intercepts.
+ */
+#ifdef VBOX_WITH_NESTED_HWVIRT_VMX
+ if (IEM_VMX_IS_NON_ROOT_MODE(pVCpu))
+ {
+ VMXEXITINSTRINFO ExitInstrInfo;
+ ExitInstrInfo.u = 0;
+ ExitInstrInfo.StrIo.u3AddrSize = ADDR_VMXSTRIO;
+ ExitInstrInfo.StrIo.iSegReg = X86_SREG_ES;
+ rcStrict = iemVmxVmexitInstrStrIo(pVCpu, VMXINSTRID_IO_INS, pVCpu->cpum.GstCtx.dx, OP_SIZE / 8, true /* fRep */,
+ ExitInstrInfo, cbInstr);
+ if (rcStrict != VINF_VMX_INTERCEPT_NOT_ACTIVE)
+ return rcStrict;
+ }
+#endif
+
+#ifdef VBOX_WITH_NESTED_HWVIRT_SVM
+ if (IEM_SVM_IS_CTRL_INTERCEPT_SET(pVCpu, SVM_CTRL_INTERCEPT_IOIO_PROT))
+ {
+ rcStrict = iemSvmHandleIOIntercept(pVCpu, u16Port, SVMIOIOTYPE_IN, OP_SIZE / 8, ADDR_SIZE, X86_SREG_ES, true /* fRep */,
+ true /* fStrIo */, cbInstr);
+ if (rcStrict == VINF_SVM_VMEXIT)
+ return VINF_SUCCESS;
+ if (rcStrict != VINF_SVM_INTERCEPT_NOT_ACTIVE)
+ {
+ Log(("iemCImpl_rep_ins_op: iemSvmHandleIOIntercept failed (u16Port=%#x, cbReg=%u) rc=%Rrc\n", u16Port, OP_SIZE / 8,
+ VBOXSTRICTRC_VAL(rcStrict)));
+ return rcStrict;
+ }
+ }
+#endif
+
+ ADDR_TYPE uCounterReg = pVCpu->cpum.GstCtx.ADDR_rCX;
+ if (uCounterReg == 0)
+ {
+ iemRegAddToRipAndClearRF(pVCpu, cbInstr);
+ return VINF_SUCCESS;
+ }
+
+ uint64_t uBaseAddr;
+ rcStrict = iemMemSegCheckWriteAccessEx(pVCpu, iemSRegUpdateHid(pVCpu, &pVCpu->cpum.GstCtx.es), X86_SREG_ES, &uBaseAddr);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ int8_t const cbIncr = pVCpu->cpum.GstCtx.eflags.Bits.u1DF ? -(OP_SIZE / 8) : (OP_SIZE / 8);
+ ADDR_TYPE uAddrReg = pVCpu->cpum.GstCtx.ADDR_rDI;
+
+ /*
+ * Be careful with handle bypassing.
+ */
+ if (pVCpu->iem.s.fBypassHandlers)
+ {
+ Log(("%s: declining because we're bypassing handlers\n", __FUNCTION__));
+ return VERR_IEM_ASPECT_NOT_IMPLEMENTED;
+ }
+
+ /*
+ * The loop.
+ */
+ for (;;)
+ {
+ /*
+ * Do segmentation and virtual page stuff.
+ */
+ ADDR2_TYPE uVirtAddr = uAddrReg + (ADDR2_TYPE)uBaseAddr;
+ uint32_t cLeftPage = (PAGE_SIZE - (uVirtAddr & PAGE_OFFSET_MASK)) / (OP_SIZE / 8);
+ if (cLeftPage > uCounterReg)
+ cLeftPage = uCounterReg;
+ if ( cLeftPage > 0 /* can be null if unaligned, do one fallback round. */
+ && cbIncr > 0 /** @todo Implement reverse direction string ops. */
+ && ( IS_64_BIT_CODE(pVCpu)
+ || ( uAddrReg < pVCpu->cpum.GstCtx.es.u32Limit
+ && uAddrReg + (cLeftPage * (OP_SIZE / 8)) <= pVCpu->cpum.GstCtx.es.u32Limit)
+ )
+ )
+ {
+ RTGCPHYS GCPhysMem;
+ rcStrict = iemMemPageTranslateAndCheckAccess(pVCpu, uVirtAddr, IEM_ACCESS_DATA_W, &GCPhysMem);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ /*
+ * If we can map the page without trouble, use the IOM
+ * string I/O interface to do the work.
+ */
+ PGMPAGEMAPLOCK PgLockMem;
+ OP_TYPE *puMem;
+ rcStrict = iemMemPageMap(pVCpu, GCPhysMem, IEM_ACCESS_DATA_W, (void **)&puMem, &PgLockMem);
+ if (rcStrict == VINF_SUCCESS)
+ {
+ uint32_t cTransfers = cLeftPage;
+ rcStrict = IOMIOPortReadString(pVM, pVCpu, u16Port, puMem, &cTransfers, OP_SIZE / 8);
+
+ uint32_t cActualTransfers = cLeftPage - cTransfers;
+ Assert(cActualTransfers <= cLeftPage);
+ pVCpu->cpum.GstCtx.ADDR_rDI = uAddrReg += cbIncr * cActualTransfers;
+ pVCpu->cpum.GstCtx.ADDR_rCX = uCounterReg -= cActualTransfers;
+ puMem += cActualTransfers;
+
+ iemMemPageUnmap(pVCpu, GCPhysMem, IEM_ACCESS_DATA_W, puMem, &PgLockMem);
+
+ if (rcStrict != VINF_SUCCESS)
+ {
+ if (IOM_SUCCESS(rcStrict))
+ {
+ rcStrict = iemSetPassUpStatus(pVCpu, rcStrict);
+ if (uCounterReg == 0)
+ iemRegAddToRipAndClearRF(pVCpu, cbInstr);
+ }
+ return rcStrict;
+ }
+
+ /* If unaligned, we drop thru and do the page crossing access
+ below. Otherwise, do the next page. */
+ if (uCounterReg == 0)
+ break;
+ if (!(uVirtAddr & (OP_SIZE / 8 - 1)))
+ {
+ IEM_CHECK_FF_YIELD_REPSTR_MAYBE_RETURN(pVM, pVCpu, pVCpu->cpum.GstCtx.eflags.u);
+ continue;
+ }
+ cLeftPage = 0;
+ }
+ }
+
+ /*
+ * Fallback - slow processing till the end of the current page.
+ * In the cross page boundrary case we will end up here with cLeftPage
+ * as 0, we execute one loop then.
+ *
+ * Note! We ASSUME the CPU will raise #PF or #GP before access the
+ * I/O port, otherwise it wouldn't really be restartable.
+ */
+ /** @todo investigate what the CPU actually does with \#PF/\#GP
+ * during INS. */
+ do
+ {
+ OP_TYPE *puMem;
+ rcStrict = iemMemMap(pVCpu, (void **)&puMem, OP_SIZE / 8, X86_SREG_ES, uAddrReg, IEM_ACCESS_DATA_W);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ uint32_t u32Value = 0;
+ rcStrict = IOMIOPortRead(pVM, pVCpu, u16Port, &u32Value, OP_SIZE / 8);
+ if (!IOM_SUCCESS(rcStrict))
+ {
+ iemMemRollback(pVCpu);
+ return rcStrict;
+ }
+
+ *puMem = (OP_TYPE)u32Value;
+# ifdef IN_RING3
+ VBOXSTRICTRC rcStrict2 = iemMemCommitAndUnmap(pVCpu, puMem, IEM_ACCESS_DATA_W);
+# else
+ VBOXSTRICTRC rcStrict2 = iemMemCommitAndUnmapPostponeTroubleToR3(pVCpu, puMem, IEM_ACCESS_DATA_W);
+# endif
+ if (rcStrict2 == VINF_SUCCESS)
+ { /* likely */ }
+ else
+ AssertLogRelMsgFailedReturn(("rcStrict2=%Rrc\n", VBOXSTRICTRC_VAL(rcStrict2)),
+ RT_FAILURE(rcStrict2) ? rcStrict2 : VERR_IEM_IPE_1);
+
+ pVCpu->cpum.GstCtx.ADDR_rDI = uAddrReg += cbIncr;
+ pVCpu->cpum.GstCtx.ADDR_rCX = --uCounterReg;
+
+ cLeftPage--;
+ if (rcStrict != VINF_SUCCESS)
+ {
+ if (uCounterReg == 0)
+ iemRegAddToRipAndClearRF(pVCpu, cbInstr);
+ rcStrict = iemSetPassUpStatus(pVCpu, rcStrict);
+ return rcStrict;
+ }
+
+ IEM_CHECK_FF_HIGH_PRIORITY_POST_REPSTR_MAYBE_RETURN(pVM, pVCpu, uCounterReg == 0);
+ } while ((int32_t)cLeftPage > 0);
+
+
+ /*
+ * Next page. Must check for interrupts and stuff here.
+ */
+ if (uCounterReg == 0)
+ break;
+ IEM_CHECK_FF_YIELD_REPSTR_MAYBE_RETURN(pVM, pVCpu, pVCpu->cpum.GstCtx.eflags.u);
+ }
+
+ /*
+ * Done.
+ */
+ iemRegAddToRipAndClearRF(pVCpu, cbInstr);
+ return VINF_SUCCESS;
+}
+
+
+/**
+ * Implements 'OUTS' (no rep)
+ */
+IEM_CIMPL_DEF_2(RT_CONCAT4(iemCImpl_outs_op,OP_SIZE,_addr,ADDR_SIZE), uint8_t, iEffSeg, bool, fIoChecked)
+{
+ PVMCC pVM = pVCpu->CTX_SUFF(pVM);
+ VBOXSTRICTRC rcStrict;
+
+ /*
+ * ASSUMES the #GP for I/O permission is taken first, then any #GP for
+ * segmentation and finally any #PF due to virtual address translation.
+ * ASSUMES nothing is read from the I/O port before traps are taken.
+ */
+ if (!fIoChecked)
+ {
+ rcStrict = iemHlpCheckPortIOPermission(pVCpu, pVCpu->cpum.GstCtx.dx, OP_SIZE / 8);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+ }
+
+ /*
+ * Check nested-guest I/O intercepts.
+ */
+#ifdef VBOX_WITH_NESTED_HWVIRT_VMX
+ if (IEM_VMX_IS_NON_ROOT_MODE(pVCpu))
+ {
+ VMXEXITINSTRINFO ExitInstrInfo;
+ ExitInstrInfo.u = 0;
+ ExitInstrInfo.StrIo.u3AddrSize = ADDR_VMXSTRIO;
+ ExitInstrInfo.StrIo.iSegReg = iEffSeg;
+ rcStrict = iemVmxVmexitInstrStrIo(pVCpu, VMXINSTRID_IO_OUTS, pVCpu->cpum.GstCtx.dx, OP_SIZE / 8, false /* fRep */,
+ ExitInstrInfo, cbInstr);
+ if (rcStrict != VINF_VMX_INTERCEPT_NOT_ACTIVE)
+ return rcStrict;
+ }
+#endif
+
+#ifdef VBOX_WITH_NESTED_HWVIRT_SVM
+ if (IEM_SVM_IS_CTRL_INTERCEPT_SET(pVCpu, SVM_CTRL_INTERCEPT_IOIO_PROT))
+ {
+ rcStrict = iemSvmHandleIOIntercept(pVCpu, pVCpu->cpum.GstCtx.dx, SVMIOIOTYPE_OUT, OP_SIZE / 8, ADDR_SIZE, iEffSeg,
+ false /* fRep */, true /* fStrIo */, cbInstr);
+ if (rcStrict == VINF_SVM_VMEXIT)
+ return VINF_SUCCESS;
+ if (rcStrict != VINF_SVM_INTERCEPT_NOT_ACTIVE)
+ {
+ Log(("iemCImpl_outs_op: iemSvmHandleIOIntercept failed (u16Port=%#x, cbReg=%u) rc=%Rrc\n", pVCpu->cpum.GstCtx.dx,
+ OP_SIZE / 8, VBOXSTRICTRC_VAL(rcStrict)));
+ return rcStrict;
+ }
+ }
+#endif
+
+ OP_TYPE uValue;
+ rcStrict = RT_CONCAT(iemMemFetchDataU,OP_SIZE)(pVCpu, &uValue, iEffSeg, pVCpu->cpum.GstCtx.ADDR_rSI);
+ if (rcStrict == VINF_SUCCESS)
+ {
+ rcStrict = IOMIOPortWrite(pVM, pVCpu, pVCpu->cpum.GstCtx.dx, uValue, OP_SIZE / 8);
+ if (IOM_SUCCESS(rcStrict))
+ {
+ if (!pVCpu->cpum.GstCtx.eflags.Bits.u1DF)
+ pVCpu->cpum.GstCtx.ADDR_rSI += OP_SIZE / 8;
+ else
+ pVCpu->cpum.GstCtx.ADDR_rSI -= OP_SIZE / 8;
+ iemRegAddToRipAndClearRF(pVCpu, cbInstr);
+ if (rcStrict != VINF_SUCCESS)
+ rcStrict = iemSetPassUpStatus(pVCpu, rcStrict);
+ }
+ }
+ return rcStrict;
+}
+
+
+/**
+ * Implements 'REP OUTS'.
+ */
+IEM_CIMPL_DEF_2(RT_CONCAT4(iemCImpl_rep_outs_op,OP_SIZE,_addr,ADDR_SIZE), uint8_t, iEffSeg, bool, fIoChecked)
+{
+ PVMCC pVM = pVCpu->CTX_SUFF(pVM);
+
+ /*
+ * Setup.
+ */
+ uint16_t const u16Port = pVCpu->cpum.GstCtx.dx;
+ VBOXSTRICTRC rcStrict;
+ if (!fIoChecked)
+ {
+/** @todo check if this is too early for ecx=0. */
+ rcStrict = iemHlpCheckPortIOPermission(pVCpu, u16Port, OP_SIZE / 8);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+ }
+
+ /*
+ * Check nested-guest I/O intercepts.
+ */
+#ifdef VBOX_WITH_NESTED_HWVIRT_VMX
+ if (IEM_VMX_IS_NON_ROOT_MODE(pVCpu))
+ {
+ VMXEXITINSTRINFO ExitInstrInfo;
+ ExitInstrInfo.u = 0;
+ ExitInstrInfo.StrIo.u3AddrSize = ADDR_VMXSTRIO;
+ ExitInstrInfo.StrIo.iSegReg = iEffSeg;
+ rcStrict = iemVmxVmexitInstrStrIo(pVCpu, VMXINSTRID_IO_OUTS, pVCpu->cpum.GstCtx.dx, OP_SIZE / 8, true /* fRep */,
+ ExitInstrInfo, cbInstr);
+ if (rcStrict != VINF_VMX_INTERCEPT_NOT_ACTIVE)
+ return rcStrict;
+ }
+#endif
+
+#ifdef VBOX_WITH_NESTED_HWVIRT_SVM
+ if (IEM_SVM_IS_CTRL_INTERCEPT_SET(pVCpu, SVM_CTRL_INTERCEPT_IOIO_PROT))
+ {
+ rcStrict = iemSvmHandleIOIntercept(pVCpu, u16Port, SVMIOIOTYPE_OUT, OP_SIZE / 8, ADDR_SIZE, iEffSeg, true /* fRep */,
+ true /* fStrIo */, cbInstr);
+ if (rcStrict == VINF_SVM_VMEXIT)
+ return VINF_SUCCESS;
+ if (rcStrict != VINF_SVM_INTERCEPT_NOT_ACTIVE)
+ {
+ Log(("iemCImpl_rep_outs_op: iemSvmHandleIOIntercept failed (u16Port=%#x, cbReg=%u) rc=%Rrc\n", u16Port, OP_SIZE / 8,
+ VBOXSTRICTRC_VAL(rcStrict)));
+ return rcStrict;
+ }
+ }
+#endif
+
+ ADDR_TYPE uCounterReg = pVCpu->cpum.GstCtx.ADDR_rCX;
+ if (uCounterReg == 0)
+ {
+ iemRegAddToRipAndClearRF(pVCpu, cbInstr);
+ return VINF_SUCCESS;
+ }
+
+ PCCPUMSELREGHID pHid = iemSRegGetHid(pVCpu, iEffSeg);
+ uint64_t uBaseAddr;
+ rcStrict = iemMemSegCheckReadAccessEx(pVCpu, pHid, iEffSeg, &uBaseAddr);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ int8_t const cbIncr = pVCpu->cpum.GstCtx.eflags.Bits.u1DF ? -(OP_SIZE / 8) : (OP_SIZE / 8);
+ ADDR_TYPE uAddrReg = pVCpu->cpum.GstCtx.ADDR_rSI;
+
+ /*
+ * The loop.
+ */
+ for (;;)
+ {
+ /*
+ * Do segmentation and virtual page stuff.
+ */
+ ADDR2_TYPE uVirtAddr = uAddrReg + (ADDR2_TYPE)uBaseAddr;
+ uint32_t cLeftPage = (PAGE_SIZE - (uVirtAddr & PAGE_OFFSET_MASK)) / (OP_SIZE / 8);
+ if (cLeftPage > uCounterReg)
+ cLeftPage = uCounterReg;
+ if ( cLeftPage > 0 /* can be null if unaligned, do one fallback round. */
+ && cbIncr > 0 /** @todo Implement reverse direction string ops. */
+ && ( IS_64_BIT_CODE(pVCpu)
+ || ( uAddrReg < pHid->u32Limit
+ && uAddrReg + (cLeftPage * (OP_SIZE / 8)) <= pHid->u32Limit)
+ )
+ )
+ {
+ RTGCPHYS GCPhysMem;
+ rcStrict = iemMemPageTranslateAndCheckAccess(pVCpu, uVirtAddr, IEM_ACCESS_DATA_R, &GCPhysMem);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ /*
+ * If we can map the page without trouble, we use the IOM
+ * string I/O interface to do the job.
+ */
+ PGMPAGEMAPLOCK PgLockMem;
+ OP_TYPE const *puMem;
+ rcStrict = iemMemPageMap(pVCpu, GCPhysMem, IEM_ACCESS_DATA_R, (void **)&puMem, &PgLockMem);
+ if (rcStrict == VINF_SUCCESS)
+ {
+ uint32_t cTransfers = cLeftPage;
+ rcStrict = IOMIOPortWriteString(pVM, pVCpu, u16Port, puMem, &cTransfers, OP_SIZE / 8);
+
+ uint32_t cActualTransfers = cLeftPage - cTransfers;
+ Assert(cActualTransfers <= cLeftPage);
+ pVCpu->cpum.GstCtx.ADDR_rSI = uAddrReg += cbIncr * cActualTransfers;
+ pVCpu->cpum.GstCtx.ADDR_rCX = uCounterReg -= cActualTransfers;
+ puMem += cActualTransfers;
+
+ iemMemPageUnmap(pVCpu, GCPhysMem, IEM_ACCESS_DATA_R, puMem, &PgLockMem);
+
+ if (rcStrict != VINF_SUCCESS)
+ {
+ if (IOM_SUCCESS(rcStrict))
+ {
+ rcStrict = iemSetPassUpStatus(pVCpu, rcStrict);
+ if (uCounterReg == 0)
+ iemRegAddToRipAndClearRF(pVCpu, cbInstr);
+ }
+ return rcStrict;
+ }
+
+ if (uCounterReg == 0)
+ break;
+
+ /* If unaligned, we drop thru and do the page crossing access
+ below. Otherwise, do the next page. */
+ if (!(uVirtAddr & (OP_SIZE / 8 - 1)))
+ {
+ IEM_CHECK_FF_YIELD_REPSTR_MAYBE_RETURN(pVM, pVCpu, pVCpu->cpum.GstCtx.eflags.u);
+ continue;
+ }
+ cLeftPage = 0;
+ }
+ }
+
+ /*
+ * Fallback - slow processing till the end of the current page.
+ * In the cross page boundrary case we will end up here with cLeftPage
+ * as 0, we execute one loop then.
+ *
+ * Note! We ASSUME the CPU will raise #PF or #GP before access the
+ * I/O port, otherwise it wouldn't really be restartable.
+ */
+ /** @todo investigate what the CPU actually does with \#PF/\#GP
+ * during INS. */
+ do
+ {
+ OP_TYPE uValue;
+ rcStrict = RT_CONCAT(iemMemFetchDataU,OP_SIZE)(pVCpu, &uValue, iEffSeg, uAddrReg);
+ if (rcStrict != VINF_SUCCESS)
+ return rcStrict;
+
+ rcStrict = IOMIOPortWrite(pVM, pVCpu, u16Port, uValue, OP_SIZE / 8);
+ if (IOM_SUCCESS(rcStrict))
+ {
+ pVCpu->cpum.GstCtx.ADDR_rSI = uAddrReg += cbIncr;
+ pVCpu->cpum.GstCtx.ADDR_rCX = --uCounterReg;
+ cLeftPage--;
+ }
+ if (rcStrict != VINF_SUCCESS)
+ {
+ if (IOM_SUCCESS(rcStrict))
+ {
+ if (uCounterReg == 0)
+ iemRegAddToRipAndClearRF(pVCpu, cbInstr);
+ rcStrict = iemSetPassUpStatus(pVCpu, rcStrict);
+ }
+ return rcStrict;
+ }
+ IEM_CHECK_FF_HIGH_PRIORITY_POST_REPSTR_MAYBE_RETURN(pVM, pVCpu, uCounterReg == 0);
+ } while ((int32_t)cLeftPage > 0);
+
+
+ /*
+ * Next page. Must check for interrupts and stuff here.
+ */
+ if (uCounterReg == 0)
+ break;
+ IEM_CHECK_FF_YIELD_REPSTR_MAYBE_RETURN(pVM, pVCpu, pVCpu->cpum.GstCtx.eflags.u);
+ }
+
+ /*
+ * Done.
+ */
+ iemRegAddToRipAndClearRF(pVCpu, cbInstr);
+ return VINF_SUCCESS;
+}
+
+#endif /* OP_SIZE != 64-bit */
+
+
+#undef OP_rAX
+#undef OP_SIZE
+#undef ADDR_SIZE
+#undef ADDR_rDI
+#undef ADDR_rSI
+#undef ADDR_rCX
+#undef ADDR_rIP
+#undef ADDR2_TYPE
+#undef ADDR_TYPE
+#undef ADDR2_TYPE
+#undef ADDR_VMXSTRIO
+#undef IS_64_BIT_CODE
+#undef IEM_CHECK_FF_YIELD_REPSTR_MAYBE_RETURN
+#undef IEM_CHECK_FF_HIGH_PRIORITY_POST_REPSTR_MAYBE_RETURN
+#undef IEM_CHECK_FF_CPU_HIGH_PRIORITY_POST_REPSTR_MAYBE_RETURN
+