diff options
Diffstat (limited to 'src/VBox/Runtime/testcase/tstRTInlineAsm.cpp')
-rw-r--r-- | src/VBox/Runtime/testcase/tstRTInlineAsm.cpp | 3355 |
1 files changed, 3355 insertions, 0 deletions
diff --git a/src/VBox/Runtime/testcase/tstRTInlineAsm.cpp b/src/VBox/Runtime/testcase/tstRTInlineAsm.cpp new file mode 100644 index 00000000..16e7d8db --- /dev/null +++ b/src/VBox/Runtime/testcase/tstRTInlineAsm.cpp @@ -0,0 +1,3355 @@ +/* $Id: tstRTInlineAsm.cpp $ */ +/** @file + * IPRT Testcase - inline assembly. + */ + +/* + * Copyright (C) 2006-2022 Oracle and/or its affiliates. + * + * This file is part of VirtualBox base platform packages, as + * available from https://www.virtualbox.org. + * + * This program is free software; you can redistribute it and/or + * modify it under the terms of the GNU General Public License + * as published by the Free Software Foundation, in version 3 of the + * License. + * + * This program is distributed in the hope that it will be useful, but + * WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * General Public License for more details. + * + * You should have received a copy of the GNU General Public License + * along with this program; if not, see <https://www.gnu.org/licenses>. + * + * The contents of this file may alternatively be used under the terms + * of the Common Development and Distribution License Version 1.0 + * (CDDL), a copy of it is provided in the "COPYING.CDDL" file included + * in the VirtualBox distribution, in which case the provisions of the + * CDDL are applicable instead of those of the GPL. + * + * You may elect to license modified versions of this file under the + * terms and conditions of either the GPL or the CDDL or both. + * + * SPDX-License-Identifier: GPL-3.0-only OR CDDL-1.0 + */ + + +/********************************************************************************************************************************* +* Header Files * +*********************************************************************************************************************************/ +#include <iprt/asm.h> +#include <iprt/asm-math.h> + +/* See http://gcc.gnu.org/bugzilla/show_bug.cgi?id=44018. Only gcc version 4.4 + * is affected. No harm for the VBox code: If the cpuid code compiles, it works + * fine. */ +#if defined(__GNUC__) && defined(RT_ARCH_X86) && defined(__PIC__) +# if __GNUC__ == 4 && __GNUC_MINOR__ == 4 +# define GCC44_32BIT_PIC +# endif +#endif + +#if !defined(GCC44_32BIT_PIC) && (defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86)) +# include <iprt/asm-amd64-x86.h> +# include <iprt/x86.h> +#elif defined(RT_ARCH_ARM64) || defined(RT_ARCH_ARM32) +# include <iprt/asm-arm.h> +# include <iprt/time.h> +#else +# include <iprt/time.h> +#endif +#include <iprt/mem.h> +#include <iprt/param.h> +#include <iprt/rand.h> +#include <iprt/stream.h> +#include <iprt/string.h> +#include <iprt/thread.h> +#include <iprt/test.h> +#include <iprt/time.h> + + + +/********************************************************************************************************************************* +* Defined Constants And Macros * +*********************************************************************************************************************************/ +#define CHECKVAL(val, expect, fmt) \ + do \ + { \ + if ((val) != (expect)) \ + { \ + RTTestFailed(g_hTest, "%s, %d: " #val ": expected " fmt " got " fmt "\n", __FUNCTION__, __LINE__, (expect), (val)); \ + } \ + } while (0) + +#define CHECKOP(op, expect, fmt, type) \ + do \ + { \ + type val = op; \ + if (val != (type)(expect)) \ + { \ + RTTestFailed(g_hTest, "%s, %d: " #op ": expected " fmt " got " fmt "\n", __FUNCTION__, __LINE__, (type)(expect), val); \ + } \ + } while (0) + +#define CHECK_OP_AND_VAL(a_Type, a_Fmt, a_pVar, a_Operation, a_ExpectRetVal, a_ExpectVarVal) \ + do { \ + CHECKOP(a_Operation, a_ExpectRetVal, a_Fmt, a_Type); \ + CHECKVAL(*a_pVar, a_ExpectVarVal, a_Fmt); \ + } while (0) + +#define CHECK_OP_AND_VAL_EX(a_TypeRet, a_FmtRet, a_FmtVar, a_pVar, a_Operation, a_ExpectRetVal, a_ExpectVarVal) \ + do { \ + CHECKOP(a_Operation, a_ExpectRetVal, a_FmtRet, a_TypeRet); \ + CHECKVAL(*a_pVar, a_ExpectVarVal, a_FmtVar); \ + } while (0) + +#define CHECK_OP_AND_VAL_EX2(a_TypeRet, a_FmtRet, a_FmtVar, a_pVar, a_uVar2, a_Operation, a_ExpectRetVal, a_ExpectVarVal, a_ExpectVarVal2) \ + do { \ + CHECKOP(a_Operation, a_ExpectRetVal, a_FmtRet, a_TypeRet); \ + CHECKVAL(*a_pVar, a_ExpectVarVal, a_FmtVar); \ + CHECKVAL(a_uVar2, a_ExpectVarVal2, a_FmtVar); \ + } while (0) + +#define CHECKVAL128(a_pu128Val, a_u64HiExpect, a_u64LoExpect) \ + do \ + { \ + if ((a_pu128Val)->s.Hi != (a_u64HiExpect) || (a_pu128Val)->s.Lo != (a_u64LoExpect)) \ + RTTestFailed(g_hTest, "%s, %d: " #a_pu128Val ": expected %#RX64'%016RX64 got %#RX64'%016RX64\n", \ + __FUNCTION__, __LINE__, (a_u64HiExpect), (a_u64LoExpect), (a_pu128Val)->s.Hi, (a_pu128Val)->s.Lo); \ + } while (0) +#define CHECKVAL128_C(a_pu128Val, a_u64HiExpect, a_u64LoExpect) \ + do \ + { \ + if ((a_pu128Val)->s.Hi != UINT64_C(a_u64HiExpect) || (a_pu128Val)->s.Lo != UINT64_C(a_u64LoExpect)) \ + RTTestFailed(g_hTest, "%s, %d: " #a_pu128Val ": expected %#RX64'%016RX64 got %#RX64'%016RX64\n", \ + __FUNCTION__, __LINE__, UINT64_C(a_u64HiExpect), UINT64_C(a_u64LoExpect), \ + (a_pu128Val)->s.Hi, (a_pu128Val)->s.Lo); \ + } while (0) +#define CHECK_OP_AND_VAL_128(a_TypeRet, a_FmtRet, a_pu128Val, a_Operation, a_ExpectRetVal, a_u64HiExpect, a_u64LoExpect) \ + do { \ + CHECKOP(a_Operation, a_ExpectRetVal, a_FmtRet, a_TypeRet); \ + CHECKVAL128(a_pu128Val, a_u64HiExpect, a_u64LoExpect); \ + } while (0) +#define CHECK_OP_AND_VAL_128_C(a_TypeRet, a_FmtRet, a_pu128Val, a_Operation, a_ExpectRetVal, a_u64HiExpect, a_u64LoExpect) \ + do { \ + CHECKOP(a_Operation, a_ExpectRetVal, a_FmtRet, a_TypeRet); \ + CHECKVAL128_C(a_pu128Val, a_u64HiExpect, a_u64LoExpect); \ + } while (0) + +/** + * Calls a worker function with different worker variable storage types. + */ +#define DO_SIMPLE_TEST_NO_SUB_NO_STACK(a_WorkerFunction, type) \ + do \ + { \ + type *pVar = (type *)RTTestGuardedAllocHead(g_hTest, sizeof(type)); \ + RTTEST_CHECK_BREAK(g_hTest, pVar); \ + a_WorkerFunction(pVar); \ + RTTestGuardedFree(g_hTest, pVar); \ + \ + pVar = (type *)RTTestGuardedAllocTail(g_hTest, sizeof(type)); \ + RTTEST_CHECK_BREAK(g_hTest, pVar); \ + a_WorkerFunction(pVar); \ + RTTestGuardedFree(g_hTest, pVar); \ + } while (0) + + +/** + * Calls a worker function with different worker variable storage types. + */ +#define DO_SIMPLE_TEST_NO_SUB(a_WorkerFunction, type) \ + do \ + { \ + type StackVar; \ + a_WorkerFunction(&StackVar); \ + DO_SIMPLE_TEST_NO_SUB_NO_STACK(a_WorkerFunction, type); \ + } while (0) + +/** + * Calls a worker function with different worker variable storage types. + */ +#define DO_SIMPLE_TEST(name, type) \ + do \ + { \ + RTTestISub(#name); \ + DO_SIMPLE_TEST_NO_SUB(tst ## name ## Worker, type); \ + } while (0) + + +/********************************************************************************************************************************* +* Global Variables * +*********************************************************************************************************************************/ +/** The test instance. */ +static RTTEST g_hTest; + + + +#if !defined(GCC44_32BIT_PIC) && (defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86)) + +const char *getCacheAss(unsigned u) +{ + if (u == 0) + return "res0 "; + if (u == 1) + return "direct"; + if (u >= 256) + return "???"; + + char *pszRet = NULL; + RTStrAPrintf(&pszRet, "%d way", u); + RTMEM_WILL_LEAK(pszRet); + return pszRet; +} + + +const char *getL2CacheAss(unsigned u) +{ + switch (u) + { + case 0: return "off "; + case 1: return "direct"; + case 2: return "2 way "; + case 3: return "res3 "; + case 4: return "4 way "; + case 5: return "res5 "; + case 6: return "8 way "; + case 7: return "res7 "; + case 8: return "16 way"; + case 9: return "res9 "; + case 10: return "res10 "; + case 11: return "res11 "; + case 12: return "res12 "; + case 13: return "res13 "; + case 14: return "res14 "; + case 15: return "fully "; + default: + return "????"; + } +} + + +/** + * Test and dump all possible info from the CPUID instruction. + * + * @remark Bits shared with the libc cpuid.c program. This all written by me, so no worries. + * @todo transform the dumping into a generic runtime function. We'll need it for logging! + */ +void tstASMCpuId(void) +{ + RTTestISub("ASMCpuId"); + + unsigned iBit; + struct + { + uint32_t uEBX, uEAX, uEDX, uECX; + } s; + if (!ASMHasCpuId()) + { + RTTestIPrintf(RTTESTLVL_ALWAYS, "warning! CPU doesn't support CPUID\n"); + return; + } + + /* + * Try the 0 function and use that for checking the ASMCpuId_* variants. + */ + ASMCpuId(0, &s.uEAX, &s.uEBX, &s.uECX, &s.uEDX); + + uint32_t u32; + + u32 = ASMCpuId_EAX(0); + CHECKVAL(u32, s.uEAX, "%x"); + u32 = ASMCpuId_EBX(0); + CHECKVAL(u32, s.uEBX, "%x"); + u32 = ASMCpuId_ECX(0); + CHECKVAL(u32, s.uECX, "%x"); + u32 = ASMCpuId_EDX(0); + CHECKVAL(u32, s.uEDX, "%x"); + + uint32_t uECX2 = s.uECX - 1; + uint32_t uEDX2 = s.uEDX - 1; + ASMCpuId_ECX_EDX(0, &uECX2, &uEDX2); + CHECKVAL(uECX2, s.uECX, "%x"); + CHECKVAL(uEDX2, s.uEDX, "%x"); + + uint32_t uEAX2 = s.uEAX - 1; + uint32_t uEBX2 = s.uEBX - 1; + uECX2 = s.uECX - 1; + uEDX2 = s.uEDX - 1; + ASMCpuIdExSlow(0, 0, 0, 0, &uEAX2, &uEBX2, &uECX2, &uEDX2); + CHECKVAL(uEAX2, s.uEAX, "%x"); + CHECKVAL(uEBX2, s.uEBX, "%x"); + CHECKVAL(uECX2, s.uECX, "%x"); + CHECKVAL(uEDX2, s.uEDX, "%x"); + + /* + * Check the extended APIC stuff. + */ + uint32_t idExtApic; + if (ASMCpuId_EAX(0) >= 0xb) + { + uint8_t idApic = ASMGetApicId(); + do + { + uEAX2 = uEBX2 = uECX2 = uEDX2 = UINT32_C(0x50486744); + ASMCpuIdExSlow(0xb, 0, 0, 0, &uEAX2, &uEBX2, &uECX2, &uEDX2); + idExtApic = ASMGetApicIdExt0B(); + } while (ASMGetApicId() != idApic); + + CHECKVAL(uEDX2, idExtApic, "%x"); + if (idApic != (uint8_t)idExtApic && uECX2 != 0) + RTTestIFailed("ASMGetApicIdExt0B() -> %#x vs ASMGetApicId() -> %#x", idExtApic, idApic); + } + if (ASMCpuId_EAX(UINT32_C(0x80000000)) >= UINT32_C(0x8000001E)) + { + uint8_t idApic = ASMGetApicId(); + do + { + uEAX2 = uEBX2 = uECX2 = uEDX2 = UINT32_C(0x50486744); + ASMCpuIdExSlow(0x8000001e, 0, 0, 0, &uEAX2, &uEBX2, &uECX2, &uEDX2); + idExtApic = ASMGetApicIdExt8000001E(); + } while (ASMGetApicId() != idApic); + CHECKVAL(uEAX2, idExtApic, "%x"); + if (idApic != (uint8_t)idExtApic) + RTTestIFailed("ASMGetApicIdExt8000001E() -> %#x vs ASMGetApicId() -> %#x", idExtApic, idApic); + } + + /* + * Done testing, dump the information. + */ + RTTestIPrintf(RTTESTLVL_ALWAYS, "CPUID Dump\n"); + ASMCpuId(0, &s.uEAX, &s.uEBX, &s.uECX, &s.uEDX); + const uint32_t cFunctions = s.uEAX; + + /* raw dump */ + RTTestIPrintf(RTTESTLVL_ALWAYS, + "\n" + " RAW Standard CPUIDs\n" + "Function eax ebx ecx edx\n"); + for (unsigned iStd = 0; iStd <= cFunctions + 3; iStd++) + { + ASMCpuId_Idx_ECX(iStd, 0, &s.uEAX, &s.uEBX, &s.uECX, &s.uEDX); + RTTestIPrintf(RTTESTLVL_ALWAYS, "%08x %08x %08x %08x %08x%s\n", + iStd, s.uEAX, s.uEBX, s.uECX, s.uEDX, iStd <= cFunctions ? "" : "*"); + + /* Some leafs output depend on the initial value of ECX. + * The same seems to apply to invalid standard functions */ + if (iStd > cFunctions) + continue; + if (iStd == 0x04) /* Deterministic Cache Parameters Leaf */ + for (uint32_t uECX = 1; s.uEAX & 0x1f; uECX++) + { + ASMCpuId_Idx_ECX(iStd, uECX, &s.uEAX, &s.uEBX, &s.uECX, &s.uEDX); + RTTestIPrintf(RTTESTLVL_ALWAYS, " [%02x] %08x %08x %08x %08x\n", uECX, s.uEAX, s.uEBX, s.uECX, s.uEDX); + RTTESTI_CHECK_BREAK(uECX < 128); + } + else if (iStd == 0x07) /* Structured Extended Feature Flags */ + { + uint32_t uMax = s.uEAX; + for (uint32_t uECX = 1; uECX < uMax; uECX++) + { + ASMCpuId_Idx_ECX(iStd, uECX, &s.uEAX, &s.uEBX, &s.uECX, &s.uEDX); + RTTestIPrintf(RTTESTLVL_ALWAYS, " [%02x] %08x %08x %08x %08x\n", uECX, s.uEAX, s.uEBX, s.uECX, s.uEDX); + RTTESTI_CHECK_BREAK(uECX < 128); + } + } + else if (iStd == 0x0b) /* Extended Topology Enumeration Leafs */ + for (uint32_t uECX = 1; (s.uEAX & 0x1f) && (s.uEBX & 0xffff); uECX++) + { + ASMCpuId_Idx_ECX(iStd, uECX, &s.uEAX, &s.uEBX, &s.uECX, &s.uEDX); + RTTestIPrintf(RTTESTLVL_ALWAYS, " [%02x] %08x %08x %08x %08x\n", uECX, s.uEAX, s.uEBX, s.uECX, s.uEDX); + RTTESTI_CHECK_BREAK(uECX < 128); + } + else if (iStd == 0x0d) /* Extended State Enumeration Leafs */ + for (uint32_t uECX = 1; s.uEAX != 0 || s.uEBX != 0 || s.uECX != 0 || s.uEDX != 0; uECX++) + { + ASMCpuId_Idx_ECX(iStd, uECX, &s.uEAX, &s.uEBX, &s.uECX, &s.uEDX); + RTTestIPrintf(RTTESTLVL_ALWAYS, " [%02x] %08x %08x %08x %08x\n", uECX, s.uEAX, s.uEBX, s.uECX, s.uEDX); + RTTESTI_CHECK_BREAK(uECX < 128); + } + else if ( iStd == 0x0f /* Platform quality of service monitoring (PQM) */ + || iStd == 0x10 /* Platform quality of service enforcement (PQE) */ + || iStd == 0x12 /* SGX Enumeration */ + || iStd == 0x14 /* Processor Trace Enumeration */ + || iStd == 0x17 /* SoC Vendor Attribute Enumeration */ + || iStd == 0x18 /* Deterministic Address Translation Parameters */) + { + /** @todo */ + } + else + { + u32 = ASMCpuId_EAX(iStd); + CHECKVAL(u32, s.uEAX, "%x"); + + uint32_t u32EbxMask = UINT32_MAX; + if (iStd == 1) + u32EbxMask = UINT32_C(0x00ffffff); /* Omit the local apic ID in case we're rescheduled. */ + u32 = ASMCpuId_EBX(iStd); + CHECKVAL(u32 & u32EbxMask, s.uEBX & u32EbxMask, "%x"); + + u32 = ASMCpuId_ECX(iStd); + CHECKVAL(u32, s.uECX, "%x"); + u32 = ASMCpuId_EDX(iStd); + CHECKVAL(u32, s.uEDX, "%x"); + + uECX2 = s.uECX - 1; + uEDX2 = s.uEDX - 1; + ASMCpuId_ECX_EDX(iStd, &uECX2, &uEDX2); + CHECKVAL(uECX2, s.uECX, "%x"); + CHECKVAL(uEDX2, s.uEDX, "%x"); + + uEAX2 = s.uEAX - 1; + uEBX2 = s.uEBX - 1; + uECX2 = s.uECX - 1; + uEDX2 = s.uEDX - 1; + ASMCpuId(iStd, &uEAX2, &uEBX2, &uECX2, &uEDX2); + CHECKVAL(uEAX2, s.uEAX, "%x"); + CHECKVAL(uEBX2 & u32EbxMask, s.uEBX & u32EbxMask, "%x"); + CHECKVAL(uECX2, s.uECX, "%x"); + CHECKVAL(uEDX2, s.uEDX, "%x"); + } + } + + /* + * Understandable output + */ + ASMCpuId(0, &s.uEAX, &s.uEBX, &s.uECX, &s.uEDX); + RTTestIPrintf(RTTESTLVL_ALWAYS, + "Name: %.04s%.04s%.04s\n" + "Support: 0-%u\n", + &s.uEBX, &s.uEDX, &s.uECX, s.uEAX); + bool const fIntel = RTX86IsIntelCpu(s.uEBX, s.uECX, s.uEDX); + + /* + * Get Features. + */ + if (cFunctions >= 1) + { + static const char * const s_apszTypes[4] = { "primary", "overdrive", "MP", "reserved" }; + ASMCpuId(1, &s.uEAX, &s.uEBX, &s.uECX, &s.uEDX); + RTTestIPrintf(RTTESTLVL_ALWAYS, + "Family: %#x \tExtended: %#x \tEffective: %#x\n" + "Model: %#x \tExtended: %#x \tEffective: %#x\n" + "Stepping: %d\n" + "Type: %d (%s)\n" + "APIC ID: %#04x\n" + "Logical CPUs: %d\n" + "CLFLUSH Size: %d\n" + "Brand ID: %#04x\n", + (s.uEAX >> 8) & 0xf, (s.uEAX >> 20) & 0x7f, RTX86GetCpuFamily(s.uEAX), + (s.uEAX >> 4) & 0xf, (s.uEAX >> 16) & 0x0f, RTX86GetCpuModel(s.uEAX, fIntel), + RTX86GetCpuStepping(s.uEAX), + (s.uEAX >> 12) & 0x3, s_apszTypes[(s.uEAX >> 12) & 0x3], + (s.uEBX >> 24) & 0xff, + (s.uEBX >> 16) & 0xff, + (s.uEBX >> 8) & 0xff, + (s.uEBX >> 0) & 0xff); + + RTTestIPrintf(RTTESTLVL_ALWAYS, "Features EDX: "); + if (s.uEDX & RT_BIT(0)) RTTestIPrintf(RTTESTLVL_ALWAYS, " FPU"); + if (s.uEDX & RT_BIT(1)) RTTestIPrintf(RTTESTLVL_ALWAYS, " VME"); + if (s.uEDX & RT_BIT(2)) RTTestIPrintf(RTTESTLVL_ALWAYS, " DE"); + if (s.uEDX & RT_BIT(3)) RTTestIPrintf(RTTESTLVL_ALWAYS, " PSE"); + if (s.uEDX & RT_BIT(4)) RTTestIPrintf(RTTESTLVL_ALWAYS, " TSC"); + if (s.uEDX & RT_BIT(5)) RTTestIPrintf(RTTESTLVL_ALWAYS, " MSR"); + if (s.uEDX & RT_BIT(6)) RTTestIPrintf(RTTESTLVL_ALWAYS, " PAE"); + if (s.uEDX & RT_BIT(7)) RTTestIPrintf(RTTESTLVL_ALWAYS, " MCE"); + if (s.uEDX & RT_BIT(8)) RTTestIPrintf(RTTESTLVL_ALWAYS, " CX8"); + if (s.uEDX & RT_BIT(9)) RTTestIPrintf(RTTESTLVL_ALWAYS, " APIC"); + if (s.uEDX & RT_BIT(10)) RTTestIPrintf(RTTESTLVL_ALWAYS, " 10"); + if (s.uEDX & RT_BIT(11)) RTTestIPrintf(RTTESTLVL_ALWAYS, " SEP"); + if (s.uEDX & RT_BIT(12)) RTTestIPrintf(RTTESTLVL_ALWAYS, " MTRR"); + if (s.uEDX & RT_BIT(13)) RTTestIPrintf(RTTESTLVL_ALWAYS, " PGE"); + if (s.uEDX & RT_BIT(14)) RTTestIPrintf(RTTESTLVL_ALWAYS, " MCA"); + if (s.uEDX & RT_BIT(15)) RTTestIPrintf(RTTESTLVL_ALWAYS, " CMOV"); + if (s.uEDX & RT_BIT(16)) RTTestIPrintf(RTTESTLVL_ALWAYS, " PAT"); + if (s.uEDX & RT_BIT(17)) RTTestIPrintf(RTTESTLVL_ALWAYS, " PSE36"); + if (s.uEDX & RT_BIT(18)) RTTestIPrintf(RTTESTLVL_ALWAYS, " PSN"); + if (s.uEDX & RT_BIT(19)) RTTestIPrintf(RTTESTLVL_ALWAYS, " CLFSH"); + if (s.uEDX & RT_BIT(20)) RTTestIPrintf(RTTESTLVL_ALWAYS, " 20"); + if (s.uEDX & RT_BIT(21)) RTTestIPrintf(RTTESTLVL_ALWAYS, " DS"); + if (s.uEDX & RT_BIT(22)) RTTestIPrintf(RTTESTLVL_ALWAYS, " ACPI"); + if (s.uEDX & RT_BIT(23)) RTTestIPrintf(RTTESTLVL_ALWAYS, " MMX"); + if (s.uEDX & RT_BIT(24)) RTTestIPrintf(RTTESTLVL_ALWAYS, " FXSR"); + if (s.uEDX & RT_BIT(25)) RTTestIPrintf(RTTESTLVL_ALWAYS, " SSE"); + if (s.uEDX & RT_BIT(26)) RTTestIPrintf(RTTESTLVL_ALWAYS, " SSE2"); + if (s.uEDX & RT_BIT(27)) RTTestIPrintf(RTTESTLVL_ALWAYS, " SS"); + if (s.uEDX & RT_BIT(28)) RTTestIPrintf(RTTESTLVL_ALWAYS, " HTT"); + if (s.uEDX & RT_BIT(29)) RTTestIPrintf(RTTESTLVL_ALWAYS, " 29"); + if (s.uEDX & RT_BIT(30)) RTTestIPrintf(RTTESTLVL_ALWAYS, " 30"); + if (s.uEDX & RT_BIT(31)) RTTestIPrintf(RTTESTLVL_ALWAYS, " 31"); + RTTestIPrintf(RTTESTLVL_ALWAYS, "\n"); + + /** @todo check intel docs. */ + RTTestIPrintf(RTTESTLVL_ALWAYS, "Features ECX: "); + if (s.uECX & RT_BIT(0)) RTTestIPrintf(RTTESTLVL_ALWAYS, " SSE3"); + for (iBit = 1; iBit < 13; iBit++) + if (s.uECX & RT_BIT(iBit)) + RTTestIPrintf(RTTESTLVL_ALWAYS, " %d", iBit); + if (s.uECX & RT_BIT(13)) RTTestIPrintf(RTTESTLVL_ALWAYS, " CX16"); + for (iBit = 14; iBit < 32; iBit++) + if (s.uECX & RT_BIT(iBit)) + RTTestIPrintf(RTTESTLVL_ALWAYS, " %d", iBit); + RTTestIPrintf(RTTESTLVL_ALWAYS, "\n"); + } + if (ASMCpuId_EAX(0) >= 0xb) + RTTestIPrintf(RTTESTLVL_ALWAYS, "APIC ID(Ext 0b): %#010x\n", ASMGetApicIdExt0B()); + + /* + * Extended. + * Implemented after AMD specs. + */ + /** @todo check out the intel specs. */ + ASMCpuId(0x80000000, &s.uEAX, &s.uEBX, &s.uECX, &s.uEDX); + if (!s.uEAX && !s.uEBX && !s.uECX && !s.uEDX) + { + RTTestIPrintf(RTTESTLVL_ALWAYS, "No extended CPUID info? Check the manual on how to detect this...\n"); + return; + } + const uint32_t cExtFunctions = s.uEAX | 0x80000000; + + /* raw dump */ + RTTestIPrintf(RTTESTLVL_ALWAYS, + "\n" + " RAW Extended CPUIDs\n" + "Function eax ebx ecx edx\n"); + for (unsigned iExt = 0x80000000; iExt <= cExtFunctions + 3; iExt++) + { + ASMCpuId(iExt, &s.uEAX, &s.uEBX, &s.uECX, &s.uEDX); + RTTestIPrintf(RTTESTLVL_ALWAYS, "%08x %08x %08x %08x %08x%s\n", + iExt, s.uEAX, s.uEBX, s.uECX, s.uEDX, iExt <= cExtFunctions ? "" : "*"); + + if (iExt > cExtFunctions) + continue; /* Invalid extended functions seems change the value if ECX changes */ + if (iExt == 0x8000001d) + continue; /* Takes cache level in ecx. */ + + u32 = ASMCpuId_EAX(iExt); + CHECKVAL(u32, s.uEAX, "%x"); + u32 = ASMCpuId_EBX(iExt); + CHECKVAL(u32, s.uEBX, "%x"); + u32 = ASMCpuId_ECX(iExt); + CHECKVAL(u32, s.uECX, "%x"); + u32 = ASMCpuId_EDX(iExt); + CHECKVAL(u32, s.uEDX, "%x"); + + uECX2 = s.uECX - 1; + uEDX2 = s.uEDX - 1; + ASMCpuId_ECX_EDX(iExt, &uECX2, &uEDX2); + CHECKVAL(uECX2, s.uECX, "%x"); + CHECKVAL(uEDX2, s.uEDX, "%x"); + + uEAX2 = s.uEAX - 1; + uEBX2 = s.uEBX - 1; + uECX2 = s.uECX - 1; + uEDX2 = s.uEDX - 1; + ASMCpuId(iExt, &uEAX2, &uEBX2, &uECX2, &uEDX2); + CHECKVAL(uEAX2, s.uEAX, "%x"); + CHECKVAL(uEBX2, s.uEBX, "%x"); + CHECKVAL(uECX2, s.uECX, "%x"); + CHECKVAL(uEDX2, s.uEDX, "%x"); + } + + /* + * Understandable output + */ + ASMCpuId(0x80000000, &s.uEAX, &s.uEBX, &s.uECX, &s.uEDX); + RTTestIPrintf(RTTESTLVL_ALWAYS, + "Ext Name: %.4s%.4s%.4s\n" + "Ext Supports: 0x80000000-%#010x\n", + &s.uEBX, &s.uEDX, &s.uECX, s.uEAX); + + if (cExtFunctions >= 0x80000001) + { + ASMCpuId(0x80000001, &s.uEAX, &s.uEBX, &s.uECX, &s.uEDX); + RTTestIPrintf(RTTESTLVL_ALWAYS, + "Family: %#x \tExtended: %#x \tEffective: %#x\n" + "Model: %#x \tExtended: %#x \tEffective: %#x\n" + "Stepping: %d\n" + "Brand ID: %#05x\n", + (s.uEAX >> 8) & 0xf, (s.uEAX >> 20) & 0x7f, RTX86GetCpuFamily(s.uEAX), + (s.uEAX >> 4) & 0xf, (s.uEAX >> 16) & 0x0f, RTX86GetCpuModel(s.uEAX, fIntel), + RTX86GetCpuStepping(s.uEAX), + s.uEBX & 0xfff); + + RTTestIPrintf(RTTESTLVL_ALWAYS, "Features EDX: "); + if (s.uEDX & RT_BIT(0)) RTTestIPrintf(RTTESTLVL_ALWAYS, " FPU"); + if (s.uEDX & RT_BIT(1)) RTTestIPrintf(RTTESTLVL_ALWAYS, " VME"); + if (s.uEDX & RT_BIT(2)) RTTestIPrintf(RTTESTLVL_ALWAYS, " DE"); + if (s.uEDX & RT_BIT(3)) RTTestIPrintf(RTTESTLVL_ALWAYS, " PSE"); + if (s.uEDX & RT_BIT(4)) RTTestIPrintf(RTTESTLVL_ALWAYS, " TSC"); + if (s.uEDX & RT_BIT(5)) RTTestIPrintf(RTTESTLVL_ALWAYS, " MSR"); + if (s.uEDX & RT_BIT(6)) RTTestIPrintf(RTTESTLVL_ALWAYS, " PAE"); + if (s.uEDX & RT_BIT(7)) RTTestIPrintf(RTTESTLVL_ALWAYS, " MCE"); + if (s.uEDX & RT_BIT(8)) RTTestIPrintf(RTTESTLVL_ALWAYS, " CMPXCHG8B"); + if (s.uEDX & RT_BIT(9)) RTTestIPrintf(RTTESTLVL_ALWAYS, " APIC"); + if (s.uEDX & RT_BIT(10)) RTTestIPrintf(RTTESTLVL_ALWAYS, " 10"); + if (s.uEDX & RT_BIT(11)) RTTestIPrintf(RTTESTLVL_ALWAYS, " SysCallSysRet"); + if (s.uEDX & RT_BIT(12)) RTTestIPrintf(RTTESTLVL_ALWAYS, " MTRR"); + if (s.uEDX & RT_BIT(13)) RTTestIPrintf(RTTESTLVL_ALWAYS, " PGE"); + if (s.uEDX & RT_BIT(14)) RTTestIPrintf(RTTESTLVL_ALWAYS, " MCA"); + if (s.uEDX & RT_BIT(15)) RTTestIPrintf(RTTESTLVL_ALWAYS, " CMOV"); + if (s.uEDX & RT_BIT(16)) RTTestIPrintf(RTTESTLVL_ALWAYS, " PAT"); + if (s.uEDX & RT_BIT(17)) RTTestIPrintf(RTTESTLVL_ALWAYS, " PSE36"); + if (s.uEDX & RT_BIT(18)) RTTestIPrintf(RTTESTLVL_ALWAYS, " 18"); + if (s.uEDX & RT_BIT(19)) RTTestIPrintf(RTTESTLVL_ALWAYS, " 19"); + if (s.uEDX & RT_BIT(20)) RTTestIPrintf(RTTESTLVL_ALWAYS, " NX"); + if (s.uEDX & RT_BIT(21)) RTTestIPrintf(RTTESTLVL_ALWAYS, " 21"); + if (s.uEDX & RT_BIT(22)) RTTestIPrintf(RTTESTLVL_ALWAYS, " MmxExt"); + if (s.uEDX & RT_BIT(23)) RTTestIPrintf(RTTESTLVL_ALWAYS, " MMX"); + if (s.uEDX & RT_BIT(24)) RTTestIPrintf(RTTESTLVL_ALWAYS, " FXSR"); + if (s.uEDX & RT_BIT(25)) RTTestIPrintf(RTTESTLVL_ALWAYS, " FastFXSR"); + if (s.uEDX & RT_BIT(26)) RTTestIPrintf(RTTESTLVL_ALWAYS, " 26"); + if (s.uEDX & RT_BIT(27)) RTTestIPrintf(RTTESTLVL_ALWAYS, " RDTSCP"); + if (s.uEDX & RT_BIT(28)) RTTestIPrintf(RTTESTLVL_ALWAYS, " 28"); + if (s.uEDX & RT_BIT(29)) RTTestIPrintf(RTTESTLVL_ALWAYS, " LongMode"); + if (s.uEDX & RT_BIT(30)) RTTestIPrintf(RTTESTLVL_ALWAYS, " 3DNowExt"); + if (s.uEDX & RT_BIT(31)) RTTestIPrintf(RTTESTLVL_ALWAYS, " 3DNow"); + RTTestIPrintf(RTTESTLVL_ALWAYS, "\n"); + + RTTestIPrintf(RTTESTLVL_ALWAYS, "Features ECX: "); + if (s.uECX & RT_BIT(0)) RTTestIPrintf(RTTESTLVL_ALWAYS, " LahfSahf"); + if (s.uECX & RT_BIT(1)) RTTestIPrintf(RTTESTLVL_ALWAYS, " CmpLegacy"); + if (s.uECX & RT_BIT(2)) RTTestIPrintf(RTTESTLVL_ALWAYS, " SVM"); + if (s.uECX & RT_BIT(3)) RTTestIPrintf(RTTESTLVL_ALWAYS, " 3"); + if (s.uECX & RT_BIT(4)) RTTestIPrintf(RTTESTLVL_ALWAYS, " AltMovCr8"); + for (iBit = 5; iBit < 32; iBit++) + if (s.uECX & RT_BIT(iBit)) + RTTestIPrintf(RTTESTLVL_ALWAYS, " %d", iBit); + RTTestIPrintf(RTTESTLVL_ALWAYS, "\n"); + } + + char szString[4*4*3+1] = {0}; + if (cExtFunctions >= 0x80000002) + ASMCpuId(0x80000002, &szString[0 + 0], &szString[0 + 4], &szString[0 + 8], &szString[0 + 12]); + if (cExtFunctions >= 0x80000003) + ASMCpuId(0x80000003, &szString[16 + 0], &szString[16 + 4], &szString[16 + 8], &szString[16 + 12]); + if (cExtFunctions >= 0x80000004) + ASMCpuId(0x80000004, &szString[32 + 0], &szString[32 + 4], &szString[32 + 8], &szString[32 + 12]); + if (cExtFunctions >= 0x80000002) + RTTestIPrintf(RTTESTLVL_ALWAYS, "Full Name: %s\n", szString); + + if (cExtFunctions >= 0x80000005) + { + ASMCpuId(0x80000005, &s.uEAX, &s.uEBX, &s.uECX, &s.uEDX); + RTTestIPrintf(RTTESTLVL_ALWAYS, + "TLB 2/4M Instr/Uni: %s %3d entries\n" + "TLB 2/4M Data: %s %3d entries\n", + getCacheAss((s.uEAX >> 8) & 0xff), (s.uEAX >> 0) & 0xff, + getCacheAss((s.uEAX >> 24) & 0xff), (s.uEAX >> 16) & 0xff); + RTTestIPrintf(RTTESTLVL_ALWAYS, + "TLB 4K Instr/Uni: %s %3d entries\n" + "TLB 4K Data: %s %3d entries\n", + getCacheAss((s.uEBX >> 8) & 0xff), (s.uEBX >> 0) & 0xff, + getCacheAss((s.uEBX >> 24) & 0xff), (s.uEBX >> 16) & 0xff); + RTTestIPrintf(RTTESTLVL_ALWAYS, + "L1 Instr Cache Line Size: %d bytes\n" + "L1 Instr Cache Lines Per Tag: %d\n" + "L1 Instr Cache Associativity: %s\n" + "L1 Instr Cache Size: %d KB\n", + (s.uEDX >> 0) & 0xff, + (s.uEDX >> 8) & 0xff, + getCacheAss((s.uEDX >> 16) & 0xff), + (s.uEDX >> 24) & 0xff); + RTTestIPrintf(RTTESTLVL_ALWAYS, + "L1 Data Cache Line Size: %d bytes\n" + "L1 Data Cache Lines Per Tag: %d\n" + "L1 Data Cache Associativity: %s\n" + "L1 Data Cache Size: %d KB\n", + (s.uECX >> 0) & 0xff, + (s.uECX >> 8) & 0xff, + getCacheAss((s.uECX >> 16) & 0xff), + (s.uECX >> 24) & 0xff); + } + + if (cExtFunctions >= 0x80000006) + { + ASMCpuId(0x80000006, &s.uEAX, &s.uEBX, &s.uECX, &s.uEDX); + RTTestIPrintf(RTTESTLVL_ALWAYS, + "L2 TLB 2/4M Instr/Uni: %s %4d entries\n" + "L2 TLB 2/4M Data: %s %4d entries\n", + getL2CacheAss((s.uEAX >> 12) & 0xf), (s.uEAX >> 0) & 0xfff, + getL2CacheAss((s.uEAX >> 28) & 0xf), (s.uEAX >> 16) & 0xfff); + RTTestIPrintf(RTTESTLVL_ALWAYS, + "L2 TLB 4K Instr/Uni: %s %4d entries\n" + "L2 TLB 4K Data: %s %4d entries\n", + getL2CacheAss((s.uEBX >> 12) & 0xf), (s.uEBX >> 0) & 0xfff, + getL2CacheAss((s.uEBX >> 28) & 0xf), (s.uEBX >> 16) & 0xfff); + RTTestIPrintf(RTTESTLVL_ALWAYS, + "L2 Cache Line Size: %d bytes\n" + "L2 Cache Lines Per Tag: %d\n" + "L2 Cache Associativity: %s\n" + "L2 Cache Size: %d KB\n", + (s.uEDX >> 0) & 0xff, + (s.uEDX >> 8) & 0xf, + getL2CacheAss((s.uEDX >> 12) & 0xf), + (s.uEDX >> 16) & 0xffff); + } + + if (cExtFunctions >= 0x80000007) + { + ASMCpuId(0x80000007, &s.uEAX, &s.uEBX, &s.uECX, &s.uEDX); + RTTestIPrintf(RTTESTLVL_ALWAYS, "APM Features: "); + if (s.uEDX & RT_BIT(0)) RTTestIPrintf(RTTESTLVL_ALWAYS, " TS"); + if (s.uEDX & RT_BIT(1)) RTTestIPrintf(RTTESTLVL_ALWAYS, " FID"); + if (s.uEDX & RT_BIT(2)) RTTestIPrintf(RTTESTLVL_ALWAYS, " VID"); + if (s.uEDX & RT_BIT(3)) RTTestIPrintf(RTTESTLVL_ALWAYS, " TTP"); + if (s.uEDX & RT_BIT(4)) RTTestIPrintf(RTTESTLVL_ALWAYS, " TM"); + if (s.uEDX & RT_BIT(5)) RTTestIPrintf(RTTESTLVL_ALWAYS, " STC"); + if (s.uEDX & RT_BIT(6)) RTTestIPrintf(RTTESTLVL_ALWAYS, " 6"); + if (s.uEDX & RT_BIT(7)) RTTestIPrintf(RTTESTLVL_ALWAYS, " 7"); + if (s.uEDX & RT_BIT(8)) RTTestIPrintf(RTTESTLVL_ALWAYS, " TscInvariant"); + for (iBit = 9; iBit < 32; iBit++) + if (s.uEDX & RT_BIT(iBit)) + RTTestIPrintf(RTTESTLVL_ALWAYS, " %d", iBit); + RTTestIPrintf(RTTESTLVL_ALWAYS, "\n"); + } + + if (cExtFunctions >= 0x80000008) + { + ASMCpuId(0x80000008, &s.uEAX, &s.uEBX, &s.uECX, &s.uEDX); + RTTestIPrintf(RTTESTLVL_ALWAYS, + "Physical Address Width: %d bits\n" + "Virtual Address Width: %d bits\n" + "Guest Physical Address Width: %d bits\n", + (s.uEAX >> 0) & 0xff, + (s.uEAX >> 8) & 0xff, + (s.uEAX >> 16) & 0xff); + RTTestIPrintf(RTTESTLVL_ALWAYS, + "Physical Core Count: %d\n", + ((s.uECX >> 0) & 0xff) + 1); + if ((s.uECX >> 12) & 0xf) + RTTestIPrintf(RTTESTLVL_ALWAYS, "ApicIdCoreIdSize: %d bits\n", (s.uECX >> 12) & 0xf); + } + + if (cExtFunctions >= 0x8000000a) + { + ASMCpuId(0x8000000a, &s.uEAX, &s.uEBX, &s.uECX, &s.uEDX); + RTTestIPrintf(RTTESTLVL_ALWAYS, + "SVM Revision: %d (%#x)\n" + "Number of Address Space IDs: %d (%#x)\n", + s.uEAX & 0xff, s.uEAX & 0xff, + s.uEBX, s.uEBX); + } + if (ASMCpuId_EAX(UINT32_C(0x80000000)) >= UINT32_C(0x8000001E)) + RTTestIPrintf(RTTESTLVL_ALWAYS, "APIC ID(Ext 8000001b): %#010x\n", ASMGetApicIdExt8000001E()); +} + +# if 0 +static void bruteForceCpuId(void) +{ + RTTestISub("brute force CPUID leafs"); + uint32_t auPrevValues[4] = { 0, 0, 0, 0}; + uint32_t uLeaf = 0; + do + { + uint32_t auValues[4]; + ASMCpuIdExSlow(uLeaf, 0, 0, 0, &auValues[0], &auValues[1], &auValues[2], &auValues[3]); + if ( (auValues[0] != auPrevValues[0] && auValues[0] != uLeaf) + || (auValues[1] != auPrevValues[1] && auValues[1] != 0) + || (auValues[2] != auPrevValues[2] && auValues[2] != 0) + || (auValues[3] != auPrevValues[3] && auValues[3] != 0) + || (uLeaf & (UINT32_C(0x08000000) - UINT32_C(1))) == 0) + { + RTTestIPrintf(RTTESTLVL_ALWAYS, + "%08x: %08x %08x %08x %08x\n", uLeaf, + auValues[0], auValues[1], auValues[2], auValues[3]); + } + auPrevValues[0] = auValues[0]; + auPrevValues[1] = auValues[1]; + auPrevValues[2] = auValues[2]; + auPrevValues[3] = auValues[3]; + + //uint32_t uSubLeaf = 0; + //do + //{ + // + // + //} while (false); + } while (uLeaf++ < UINT32_MAX); +} +# endif + +#endif /* AMD64 || X86 */ + +#define TEST_READ(a_pVar, a_Type, a_Fmt, a_Function, a_Val) \ + do { *a_pVar = a_Val; CHECKOP(a_Function(a_pVar), a_Val, a_Fmt, a_Type); CHECKVAL(*a_pVar, a_Val, a_Fmt); } while (0) + +DECLINLINE(void) tstASMAtomicReadU8Worker(uint8_t volatile *pu8) +{ + TEST_READ(pu8, uint8_t, "%#x", ASMAtomicReadU8, 0); + TEST_READ(pu8, uint8_t, "%#x", ASMAtomicReadU8, 1); + TEST_READ(pu8, uint8_t, "%#x", ASMAtomicReadU8, 2); + TEST_READ(pu8, uint8_t, "%#x", ASMAtomicReadU8, 16); + TEST_READ(pu8, uint8_t, "%#x", ASMAtomicReadU8, 32); + TEST_READ(pu8, uint8_t, "%#x", ASMAtomicReadU8, 32); + TEST_READ(pu8, uint8_t, "%#x", ASMAtomicReadU8, 127); + TEST_READ(pu8, uint8_t, "%#x", ASMAtomicReadU8, 128); + TEST_READ(pu8, uint8_t, "%#x", ASMAtomicReadU8, 169); + TEST_READ(pu8, uint8_t, "%#x", ASMAtomicReadU8, 239); + TEST_READ(pu8, uint8_t, "%#x", ASMAtomicReadU8, 254); + TEST_READ(pu8, uint8_t, "%#x", ASMAtomicReadU8, 255); + + int8_t volatile *pi8 = (int8_t volatile *)pu8; + TEST_READ(pi8, uint8_t, "%d", ASMAtomicReadS8, INT8_MAX); + TEST_READ(pi8, uint8_t, "%d", ASMAtomicReadS8, INT8_MIN); + TEST_READ(pi8, uint8_t, "%d", ASMAtomicReadS8, 42); + TEST_READ(pi8, uint8_t, "%d", ASMAtomicReadS8, -21); + + bool volatile *pf = (bool volatile *)pu8; + TEST_READ(pf, bool, "%d", ASMAtomicReadBool, true); + TEST_READ(pf, bool, "%d", ASMAtomicReadBool, false); +} + + +DECLINLINE(void) tstASMAtomicUoReadU8Worker(uint8_t volatile *pu8) +{ + TEST_READ(pu8, uint8_t, "%#x", ASMAtomicUoReadU8, 0); + TEST_READ(pu8, uint8_t, "%#x", ASMAtomicUoReadU8, 1); + TEST_READ(pu8, uint8_t, "%#x", ASMAtomicUoReadU8, 2); + TEST_READ(pu8, uint8_t, "%#x", ASMAtomicUoReadU8, 16); + TEST_READ(pu8, uint8_t, "%#x", ASMAtomicUoReadU8, 32); + TEST_READ(pu8, uint8_t, "%#x", ASMAtomicUoReadU8, 32); + TEST_READ(pu8, uint8_t, "%#x", ASMAtomicUoReadU8, 127); + TEST_READ(pu8, uint8_t, "%#x", ASMAtomicUoReadU8, 128); + TEST_READ(pu8, uint8_t, "%#x", ASMAtomicUoReadU8, 169); + TEST_READ(pu8, uint8_t, "%#x", ASMAtomicUoReadU8, 239); + TEST_READ(pu8, uint8_t, "%#x", ASMAtomicUoReadU8, 254); + TEST_READ(pu8, uint8_t, "%#x", ASMAtomicUoReadU8, 255); + + int8_t volatile *pi8 = (int8_t volatile *)pu8; + TEST_READ(pi8, uint8_t, "%d", ASMAtomicUoReadS8, INT8_MAX); + TEST_READ(pi8, uint8_t, "%d", ASMAtomicUoReadS8, INT8_MIN); + TEST_READ(pi8, uint8_t, "%d", ASMAtomicUoReadS8, 42); + TEST_READ(pi8, uint8_t, "%d", ASMAtomicUoReadS8, -21); + + bool volatile *pf = (bool volatile *)pu8; + TEST_READ(pf, bool, "%d", ASMAtomicUoReadBool, true); + TEST_READ(pf, bool, "%d", ASMAtomicUoReadBool, false); +} + + +DECLINLINE(void) tstASMAtomicReadU16Worker(uint16_t volatile *pu16) +{ + TEST_READ(pu16, uint16_t, "%#x", ASMAtomicReadU16, 0); + TEST_READ(pu16, uint16_t, "%#x", ASMAtomicReadU16, 19983); + TEST_READ(pu16, uint16_t, "%#x", ASMAtomicReadU16, INT16_MAX); + TEST_READ(pu16, uint16_t, "%#x", ASMAtomicReadU16, UINT16_MAX); + + int16_t volatile *pi16 = (int16_t volatile *)pu16; + TEST_READ(pi16, uint16_t, "%d", ASMAtomicReadS16, INT16_MAX); + TEST_READ(pi16, uint16_t, "%d", ASMAtomicReadS16, INT16_MIN); + TEST_READ(pi16, uint16_t, "%d", ASMAtomicReadS16, 42); + TEST_READ(pi16, uint16_t, "%d", ASMAtomicReadS16, -21); +} + + +DECLINLINE(void) tstASMAtomicUoReadU16Worker(uint16_t volatile *pu16) +{ + TEST_READ(pu16, uint16_t, "%#x", ASMAtomicUoReadU16, 0); + TEST_READ(pu16, uint16_t, "%#x", ASMAtomicUoReadU16, 19983); + TEST_READ(pu16, uint16_t, "%#x", ASMAtomicUoReadU16, INT16_MAX); + TEST_READ(pu16, uint16_t, "%#x", ASMAtomicUoReadU16, UINT16_MAX); + + int16_t volatile *pi16 = (int16_t volatile *)pu16; + TEST_READ(pi16, uint16_t, "%d", ASMAtomicUoReadS16, INT16_MAX); + TEST_READ(pi16, uint16_t, "%d", ASMAtomicUoReadS16, INT16_MIN); + TEST_READ(pi16, uint16_t, "%d", ASMAtomicUoReadS16, 42); + TEST_READ(pi16, uint16_t, "%d", ASMAtomicUoReadS16, -21); +} + + +DECLINLINE(void) tstASMAtomicReadU32Worker(uint32_t volatile *pu32) +{ + TEST_READ(pu32, uint32_t, "%#x", ASMAtomicReadU32, 0); + TEST_READ(pu32, uint32_t, "%#x", ASMAtomicReadU32, 19983); + TEST_READ(pu32, uint32_t, "%#x", ASMAtomicReadU32, INT16_MAX); + TEST_READ(pu32, uint32_t, "%#x", ASMAtomicReadU32, UINT16_MAX); + TEST_READ(pu32, uint32_t, "%#x", ASMAtomicReadU32, _1M-1); + TEST_READ(pu32, uint32_t, "%#x", ASMAtomicReadU32, _1M+1); + TEST_READ(pu32, uint32_t, "%#x", ASMAtomicReadU32, _1G-1); + TEST_READ(pu32, uint32_t, "%#x", ASMAtomicReadU32, _1G+1); + TEST_READ(pu32, uint32_t, "%#x", ASMAtomicReadU32, INT32_MAX); + TEST_READ(pu32, uint32_t, "%#x", ASMAtomicReadU32, UINT32_MAX); + + int32_t volatile *pi32 = (int32_t volatile *)pu32; + TEST_READ(pi32, uint32_t, "%d", ASMAtomicReadS32, INT32_MAX); + TEST_READ(pi32, uint32_t, "%d", ASMAtomicReadS32, INT32_MIN); + TEST_READ(pi32, uint32_t, "%d", ASMAtomicReadS32, 42); + TEST_READ(pi32, uint32_t, "%d", ASMAtomicReadS32, -21); + +#if ARCH_BITS == 32 + size_t volatile *pcb = (size_t volatile *)pu32; + TEST_READ(pcb, size_t, "%#llz", ASMAtomicReadZ, 0); + TEST_READ(pcb, size_t, "%#llz", ASMAtomicReadZ, ~(size_t)2); + TEST_READ(pcb, size_t, "%#llz", ASMAtomicReadZ, ~(size_t)0 / 4); + + void * volatile *ppv = (void * volatile *)pu32; + TEST_READ(ppv, void *, "%p", ASMAtomicReadPtr, NULL); + TEST_READ(ppv, void *, "%p", ASMAtomicReadPtr, (void *)~(uintptr_t)42); + + RTSEMEVENT volatile *phEvt = (RTSEMEVENT volatile *)pu32; + RTSEMEVENT hEvt = ASMAtomicReadPtrT(phEvt, RTSEMEVENT); + CHECKVAL(hEvt, (RTSEMEVENT)~(uintptr_t)42, "%p"); + + ASMAtomicReadHandle(phEvt, &hEvt); + CHECKVAL(hEvt, (RTSEMEVENT)~(uintptr_t)42, "%p"); +#endif +} + + +DECLINLINE(void) tstASMAtomicUoReadU32Worker(uint32_t volatile *pu32) +{ + TEST_READ(pu32, uint32_t, "%#x", ASMAtomicUoReadU32, 0); + TEST_READ(pu32, uint32_t, "%#x", ASMAtomicUoReadU32, 19983); + TEST_READ(pu32, uint32_t, "%#x", ASMAtomicUoReadU32, INT16_MAX); + TEST_READ(pu32, uint32_t, "%#x", ASMAtomicUoReadU32, UINT16_MAX); + TEST_READ(pu32, uint32_t, "%#x", ASMAtomicUoReadU32, _1M-1); + TEST_READ(pu32, uint32_t, "%#x", ASMAtomicUoReadU32, _1M+1); + TEST_READ(pu32, uint32_t, "%#x", ASMAtomicUoReadU32, _1G-1); + TEST_READ(pu32, uint32_t, "%#x", ASMAtomicUoReadU32, _1G+1); + TEST_READ(pu32, uint32_t, "%#x", ASMAtomicUoReadU32, INT32_MAX); + TEST_READ(pu32, uint32_t, "%#x", ASMAtomicUoReadU32, UINT32_MAX); + + int32_t volatile *pi32 = (int32_t volatile *)pu32; + TEST_READ(pi32, uint32_t, "%d", ASMAtomicUoReadS32, INT32_MAX); + TEST_READ(pi32, uint32_t, "%d", ASMAtomicUoReadS32, INT32_MIN); + TEST_READ(pi32, uint32_t, "%d", ASMAtomicUoReadS32, 42); + TEST_READ(pi32, uint32_t, "%d", ASMAtomicUoReadS32, -21); + +#if ARCH_BITS == 32 + size_t volatile *pcb = (size_t volatile *)pu32; + TEST_READ(pcb, size_t, "%#llz", ASMAtomicUoReadZ, 0); + TEST_READ(pcb, size_t, "%#llz", ASMAtomicUoReadZ, ~(size_t)2); + TEST_READ(pcb, size_t, "%#llz", ASMAtomicUoReadZ, ~(size_t)0 / 4); + + void * volatile *ppv = (void * volatile *)pu32; + TEST_READ(ppv, void *, "%p", ASMAtomicUoReadPtr, NULL); + TEST_READ(ppv, void *, "%p", ASMAtomicUoReadPtr, (void *)~(uintptr_t)42); + + RTSEMEVENT volatile *phEvt = (RTSEMEVENT volatile *)pu32; + RTSEMEVENT hEvt = ASMAtomicUoReadPtrT(phEvt, RTSEMEVENT); + CHECKVAL(hEvt, (RTSEMEVENT)~(uintptr_t)42, "%p"); + + ASMAtomicUoReadHandle(phEvt, &hEvt); + CHECKVAL(hEvt, (RTSEMEVENT)~(uintptr_t)42, "%p"); +#endif +} + + +DECLINLINE(void) tstASMAtomicReadU64Worker(uint64_t volatile *pu64) +{ + TEST_READ(pu64, uint64_t, "%#llx", ASMAtomicReadU64, 0); + TEST_READ(pu64, uint64_t, "%#llx", ASMAtomicReadU64, 19983); + TEST_READ(pu64, uint64_t, "%#llx", ASMAtomicReadU64, INT16_MAX); + TEST_READ(pu64, uint64_t, "%#llx", ASMAtomicReadU64, UINT16_MAX); + TEST_READ(pu64, uint64_t, "%#llx", ASMAtomicReadU64, _1M-1); + TEST_READ(pu64, uint64_t, "%#llx", ASMAtomicReadU64, _1M+1); + TEST_READ(pu64, uint64_t, "%#llx", ASMAtomicReadU64, _1G-1); + TEST_READ(pu64, uint64_t, "%#llx", ASMAtomicReadU64, _1G+1); + TEST_READ(pu64, uint64_t, "%#llx", ASMAtomicReadU64, INT32_MAX); + TEST_READ(pu64, uint64_t, "%#llx", ASMAtomicReadU64, UINT32_MAX); + TEST_READ(pu64, uint64_t, "%#llx", ASMAtomicReadU64, INT64_MAX); + TEST_READ(pu64, uint64_t, "%#llx", ASMAtomicReadU64, UINT64_MAX); + TEST_READ(pu64, uint64_t, "%#llx", ASMAtomicReadU64, UINT64_C(0x450872549687134)); + + int64_t volatile *pi64 = (int64_t volatile *)pu64; + TEST_READ(pi64, uint64_t, "%d", ASMAtomicReadS64, INT64_MAX); + TEST_READ(pi64, uint64_t, "%d", ASMAtomicReadS64, INT64_MIN); + TEST_READ(pi64, uint64_t, "%d", ASMAtomicReadS64, 42); + TEST_READ(pi64, uint64_t, "%d", ASMAtomicReadS64, -21); + +#if ARCH_BITS == 64 + size_t volatile *pcb = (size_t volatile *)pu64; + TEST_READ(pcb, size_t, "%#llz", ASMAtomicReadZ, 0); + TEST_READ(pcb, size_t, "%#llz", ASMAtomicReadZ, ~(size_t)2); + TEST_READ(pcb, size_t, "%#llz", ASMAtomicReadZ, ~(size_t)0 / 4); + + void * volatile *ppv = (void * volatile *)pu64; + TEST_READ(ppv, void *, "%p", ASMAtomicReadPtr, NULL); + TEST_READ(ppv, void *, "%p", ASMAtomicReadPtr, (void *)~(uintptr_t)42); + + RTSEMEVENT volatile *phEvt = (RTSEMEVENT volatile *)pu64; + RTSEMEVENT hEvt = ASMAtomicReadPtrT(phEvt, RTSEMEVENT); + CHECKVAL(hEvt, (RTSEMEVENT)~(uintptr_t)42, "%p"); + + ASMAtomicReadHandle(phEvt, &hEvt); + CHECKVAL(hEvt, (RTSEMEVENT)~(uintptr_t)42, "%p"); +#endif +} + + +DECLINLINE(void) tstASMAtomicUoReadU64Worker(uint64_t volatile *pu64) +{ + TEST_READ(pu64, uint64_t, "%#llx", ASMAtomicUoReadU64, 0); + TEST_READ(pu64, uint64_t, "%#llx", ASMAtomicUoReadU64, 19983); + TEST_READ(pu64, uint64_t, "%#llx", ASMAtomicUoReadU64, INT16_MAX); + TEST_READ(pu64, uint64_t, "%#llx", ASMAtomicUoReadU64, UINT16_MAX); + TEST_READ(pu64, uint64_t, "%#llx", ASMAtomicUoReadU64, _1M-1); + TEST_READ(pu64, uint64_t, "%#llx", ASMAtomicUoReadU64, _1M+1); + TEST_READ(pu64, uint64_t, "%#llx", ASMAtomicUoReadU64, _1G-1); + TEST_READ(pu64, uint64_t, "%#llx", ASMAtomicUoReadU64, _1G+1); + TEST_READ(pu64, uint64_t, "%#llx", ASMAtomicUoReadU64, INT32_MAX); + TEST_READ(pu64, uint64_t, "%#llx", ASMAtomicUoReadU64, UINT32_MAX); + TEST_READ(pu64, uint64_t, "%#llx", ASMAtomicUoReadU64, INT64_MAX); + TEST_READ(pu64, uint64_t, "%#llx", ASMAtomicUoReadU64, UINT64_MAX); + TEST_READ(pu64, uint64_t, "%#llx", ASMAtomicUoReadU64, UINT64_C(0x450872549687134)); + + int64_t volatile *pi64 = (int64_t volatile *)pu64; + TEST_READ(pi64, uint64_t, "%d", ASMAtomicUoReadS64, INT64_MAX); + TEST_READ(pi64, uint64_t, "%d", ASMAtomicUoReadS64, INT64_MIN); + TEST_READ(pi64, uint64_t, "%d", ASMAtomicUoReadS64, 42); + TEST_READ(pi64, uint64_t, "%d", ASMAtomicUoReadS64, -21); + +#if ARCH_BITS == 64 + size_t volatile *pcb = (size_t volatile *)pu64; + TEST_READ(pcb, size_t, "%#llz", ASMAtomicUoReadZ, 0); + TEST_READ(pcb, size_t, "%#llz", ASMAtomicUoReadZ, ~(size_t)2); + TEST_READ(pcb, size_t, "%#llz", ASMAtomicUoReadZ, ~(size_t)0 / 4); + + void * volatile *ppv = (void * volatile *)pu64; + TEST_READ(ppv, void *, "%p", ASMAtomicUoReadPtr, NULL); + TEST_READ(ppv, void *, "%p", ASMAtomicUoReadPtr, (void *)~(uintptr_t)42); + + RTSEMEVENT volatile *phEvt = (RTSEMEVENT volatile *)pu64; + RTSEMEVENT hEvt = ASMAtomicUoReadPtrT(phEvt, RTSEMEVENT); + CHECKVAL(hEvt, (RTSEMEVENT)~(uintptr_t)42, "%p"); + + ASMAtomicUoReadHandle(phEvt, &hEvt); + CHECKVAL(hEvt, (RTSEMEVENT)~(uintptr_t)42, "%p"); +#endif +} + + +static void tstASMAtomicRead(void) +{ + DO_SIMPLE_TEST(ASMAtomicReadU8, uint8_t); + DO_SIMPLE_TEST(ASMAtomicUoReadU8, uint8_t); + + DO_SIMPLE_TEST(ASMAtomicReadU16, uint16_t); + DO_SIMPLE_TEST(ASMAtomicUoReadU16, uint16_t); + + DO_SIMPLE_TEST(ASMAtomicReadU32, uint32_t); + DO_SIMPLE_TEST(ASMAtomicUoReadU32, uint32_t); + + DO_SIMPLE_TEST(ASMAtomicReadU64, uint64_t); + DO_SIMPLE_TEST(ASMAtomicUoReadU64, uint64_t); +} + + +#define TEST_WRITE(a_pVar, a_Type, a_Fmt, a_Function, a_Val) \ + do { a_Function(a_pVar, a_Val); CHECKVAL(*a_pVar, a_Val, a_Fmt); } while (0) + +DECLINLINE(void) tstASMAtomicWriteU8Worker(uint8_t volatile *pu8) +{ + TEST_WRITE(pu8, uint8_t, "%#x", ASMAtomicWriteU8, 0); + TEST_WRITE(pu8, uint8_t, "%#x", ASMAtomicWriteU8, 1); + TEST_WRITE(pu8, uint8_t, "%#x", ASMAtomicWriteU8, 2); + TEST_WRITE(pu8, uint8_t, "%#x", ASMAtomicWriteU8, 16); + TEST_WRITE(pu8, uint8_t, "%#x", ASMAtomicWriteU8, 32); + TEST_WRITE(pu8, uint8_t, "%#x", ASMAtomicWriteU8, 32); + TEST_WRITE(pu8, uint8_t, "%#x", ASMAtomicWriteU8, 127); + TEST_WRITE(pu8, uint8_t, "%#x", ASMAtomicWriteU8, 128); + TEST_WRITE(pu8, uint8_t, "%#x", ASMAtomicWriteU8, 169); + TEST_WRITE(pu8, uint8_t, "%#x", ASMAtomicWriteU8, 239); + TEST_WRITE(pu8, uint8_t, "%#x", ASMAtomicWriteU8, 254); + TEST_WRITE(pu8, uint8_t, "%#x", ASMAtomicWriteU8, 255); + + volatile int8_t *pi8 = (volatile int8_t *)pu8; + TEST_WRITE(pi8, int8_t, "%d", ASMAtomicWriteS8, INT8_MIN); + TEST_WRITE(pi8, int8_t, "%d", ASMAtomicWriteS8, INT8_MAX); + TEST_WRITE(pi8, int8_t, "%d", ASMAtomicWriteS8, 42); + TEST_WRITE(pi8, int8_t, "%d", ASMAtomicWriteS8, -41); + + volatile bool *pf = (volatile bool *)pu8; + TEST_WRITE(pf, bool, "%d", ASMAtomicWriteBool, true); + TEST_WRITE(pf, bool, "%d", ASMAtomicWriteBool, false); +} + + +DECLINLINE(void) tstASMAtomicUoWriteU8Worker(uint8_t volatile *pu8) +{ + TEST_WRITE(pu8, uint8_t, "%#x", ASMAtomicUoWriteU8, 0); + TEST_WRITE(pu8, uint8_t, "%#x", ASMAtomicUoWriteU8, 1); + TEST_WRITE(pu8, uint8_t, "%#x", ASMAtomicUoWriteU8, 2); + TEST_WRITE(pu8, uint8_t, "%#x", ASMAtomicUoWriteU8, 16); + TEST_WRITE(pu8, uint8_t, "%#x", ASMAtomicUoWriteU8, 32); + TEST_WRITE(pu8, uint8_t, "%#x", ASMAtomicUoWriteU8, 32); + TEST_WRITE(pu8, uint8_t, "%#x", ASMAtomicUoWriteU8, 127); + TEST_WRITE(pu8, uint8_t, "%#x", ASMAtomicUoWriteU8, 128); + TEST_WRITE(pu8, uint8_t, "%#x", ASMAtomicUoWriteU8, 169); + TEST_WRITE(pu8, uint8_t, "%#x", ASMAtomicUoWriteU8, 239); + TEST_WRITE(pu8, uint8_t, "%#x", ASMAtomicUoWriteU8, 254); + TEST_WRITE(pu8, uint8_t, "%#x", ASMAtomicUoWriteU8, 255); + + volatile int8_t *pi8 = (volatile int8_t *)pu8; + TEST_WRITE(pi8, int8_t, "%d", ASMAtomicUoWriteS8, INT8_MIN); + TEST_WRITE(pi8, int8_t, "%d", ASMAtomicUoWriteS8, INT8_MAX); + TEST_WRITE(pi8, int8_t, "%d", ASMAtomicUoWriteS8, 42); + TEST_WRITE(pi8, int8_t, "%d", ASMAtomicUoWriteS8, -41); + + volatile bool *pf = (volatile bool *)pu8; + TEST_WRITE(pf, bool, "%d", ASMAtomicUoWriteBool, true); + TEST_WRITE(pf, bool, "%d", ASMAtomicUoWriteBool, false); +} + + +DECLINLINE(void) tstASMAtomicWriteU16Worker(uint16_t volatile *pu16) +{ + TEST_WRITE(pu16, uint16_t, "%#x", ASMAtomicWriteU16, 0); + TEST_WRITE(pu16, uint16_t, "%#x", ASMAtomicWriteU16, 19983); + TEST_WRITE(pu16, uint16_t, "%#x", ASMAtomicWriteU16, INT16_MAX); + TEST_WRITE(pu16, uint16_t, "%#x", ASMAtomicWriteU16, UINT16_MAX); + + volatile int16_t *pi16 = (volatile int16_t *)pu16; + TEST_WRITE(pi16, int16_t, "%d", ASMAtomicWriteS16, INT16_MIN); + TEST_WRITE(pi16, int16_t, "%d", ASMAtomicWriteS16, INT16_MAX); + TEST_WRITE(pi16, int16_t, "%d", ASMAtomicWriteS16, 42); + TEST_WRITE(pi16, int16_t, "%d", ASMAtomicWriteS16, -41); +} + + +DECLINLINE(void) tstASMAtomicUoWriteU16Worker(uint16_t volatile *pu16) +{ + TEST_WRITE(pu16, uint16_t, "%#x", ASMAtomicUoWriteU16, 0); + TEST_WRITE(pu16, uint16_t, "%#x", ASMAtomicUoWriteU16, 19983); + TEST_WRITE(pu16, uint16_t, "%#x", ASMAtomicUoWriteU16, INT16_MAX); + TEST_WRITE(pu16, uint16_t, "%#x", ASMAtomicUoWriteU16, UINT16_MAX); + + volatile int16_t *pi16 = (volatile int16_t *)pu16; + TEST_WRITE(pi16, int16_t, "%d", ASMAtomicUoWriteS16, INT16_MIN); + TEST_WRITE(pi16, int16_t, "%d", ASMAtomicUoWriteS16, INT16_MAX); + TEST_WRITE(pi16, int16_t, "%d", ASMAtomicUoWriteS16, 42); + TEST_WRITE(pi16, int16_t, "%d", ASMAtomicUoWriteS16, -41); +} + + +DECLINLINE(void) tstASMAtomicWriteU32Worker(uint32_t volatile *pu32) +{ + TEST_WRITE(pu32, uint32_t, "%#x", ASMAtomicWriteU32, 0); + TEST_WRITE(pu32, uint32_t, "%#x", ASMAtomicWriteU32, 19983); + TEST_WRITE(pu32, uint32_t, "%#x", ASMAtomicWriteU32, INT16_MAX); + TEST_WRITE(pu32, uint32_t, "%#x", ASMAtomicWriteU32, UINT16_MAX); + TEST_WRITE(pu32, uint32_t, "%#x", ASMAtomicWriteU32, _1M-1); + TEST_WRITE(pu32, uint32_t, "%#x", ASMAtomicWriteU32, _1M+1); + TEST_WRITE(pu32, uint32_t, "%#x", ASMAtomicWriteU32, _1G-1); + TEST_WRITE(pu32, uint32_t, "%#x", ASMAtomicWriteU32, _1G+1); + TEST_WRITE(pu32, uint32_t, "%#x", ASMAtomicWriteU32, INT32_MAX); + TEST_WRITE(pu32, uint32_t, "%#x", ASMAtomicWriteU32, UINT32_MAX); + + volatile int32_t *pi32 = (volatile int32_t *)pu32; + TEST_WRITE(pi32, int32_t, "%d", ASMAtomicWriteS32, INT32_MIN); + TEST_WRITE(pi32, int32_t, "%d", ASMAtomicWriteS32, INT32_MAX); + TEST_WRITE(pi32, int32_t, "%d", ASMAtomicWriteS32, 42); + TEST_WRITE(pi32, int32_t, "%d", ASMAtomicWriteS32, -41); + +#if ARCH_BITS == 32 + size_t volatile *pcb = (size_t volatile *)pu32; + TEST_WRITE(pcb, size_t, "%#zx", ASMAtomicWriteZ, ~(size_t)42); + TEST_WRITE(pcb, size_t, "%#zx", ASMAtomicWriteZ, 42); + + void * volatile *ppv = (void * volatile *)pu32; + TEST_WRITE(ppv, void *, "%#zx", ASMAtomicWritePtrVoid, NULL); + TEST_WRITE(ppv, void *, "%#zx", ASMAtomicWritePtrVoid, (void *)~(uintptr_t)12938754); + + ASMAtomicWriteNullPtr(ppv); CHECKVAL(*ppv, NULL, "%p"); + ASMAtomicWritePtr(ppv, (void *)~(intptr_t)2322434); CHECKVAL(*ppv, (void *)~(intptr_t)2322434, "%p"); + + RTSEMEVENT volatile *phEvt = (RTSEMEVENT volatile *)pu32; + ASMAtomicWriteHandle(phEvt, (RTSEMEVENT)(uintptr_t)99753456); CHECKVAL(*phEvt, (RTSEMEVENT)(uintptr_t)99753456, "%p"); +#endif +} + + +DECLINLINE(void) tstASMAtomicUoWriteU32Worker(uint32_t volatile *pu32) +{ + TEST_WRITE(pu32, uint32_t, "%#x", ASMAtomicUoWriteU32, 0); + TEST_WRITE(pu32, uint32_t, "%#x", ASMAtomicUoWriteU32, 19983); + TEST_WRITE(pu32, uint32_t, "%#x", ASMAtomicUoWriteU32, INT16_MAX); + TEST_WRITE(pu32, uint32_t, "%#x", ASMAtomicUoWriteU32, UINT16_MAX); + TEST_WRITE(pu32, uint32_t, "%#x", ASMAtomicUoWriteU32, _1M-1); + TEST_WRITE(pu32, uint32_t, "%#x", ASMAtomicUoWriteU32, _1M+1); + TEST_WRITE(pu32, uint32_t, "%#x", ASMAtomicUoWriteU32, _1G-1); + TEST_WRITE(pu32, uint32_t, "%#x", ASMAtomicUoWriteU32, _1G+1); + TEST_WRITE(pu32, uint32_t, "%#x", ASMAtomicUoWriteU32, INT32_MAX); + TEST_WRITE(pu32, uint32_t, "%#x", ASMAtomicUoWriteU32, UINT32_MAX); + + volatile int32_t *pi32 = (volatile int32_t *)pu32; + TEST_WRITE(pi32, int32_t, "%d", ASMAtomicUoWriteS32, INT32_MIN); + TEST_WRITE(pi32, int32_t, "%d", ASMAtomicUoWriteS32, INT32_MAX); + TEST_WRITE(pi32, int32_t, "%d", ASMAtomicUoWriteS32, 42); + TEST_WRITE(pi32, int32_t, "%d", ASMAtomicUoWriteS32, -41); + +#if ARCH_BITS == 32 + size_t volatile *pcb = (size_t volatile *)pu32; + TEST_WRITE(pcb, size_t, "%#zx", ASMAtomicUoWriteZ, ~(size_t)42); + TEST_WRITE(pcb, size_t, "%#zx", ASMAtomicUoWriteZ, 42); + + void * volatile *ppv = (void * volatile *)pu32; + TEST_WRITE(ppv, void *, "%#zx", ASMAtomicUoWritePtrVoid, NULL); + TEST_WRITE(ppv, void *, "%#zx", ASMAtomicUoWritePtrVoid, (void *)~(uintptr_t)12938754); + + ASMAtomicUoWriteNullPtr(ppv); CHECKVAL(*ppv, NULL, "%p"); + ASMAtomicUoWritePtr(ppv, (void *)~(intptr_t)2322434); CHECKVAL(*ppv, (void *)~(intptr_t)2322434, "%p"); + + RTSEMEVENT volatile *phEvt = (RTSEMEVENT volatile *)pu32; + ASMAtomicUoWriteHandle(phEvt, (RTSEMEVENT)(uintptr_t)99753456); CHECKVAL(*phEvt, (RTSEMEVENT)(uintptr_t)99753456, "%p"); +#endif +} + + +DECLINLINE(void) tstASMAtomicWriteU64Worker(uint64_t volatile *pu64) +{ + TEST_WRITE(pu64, uint64_t, "%#llx", ASMAtomicWriteU64, 0); + TEST_WRITE(pu64, uint64_t, "%#llx", ASMAtomicWriteU64, 19983); + TEST_WRITE(pu64, uint64_t, "%#llx", ASMAtomicWriteU64, INT16_MAX); + TEST_WRITE(pu64, uint64_t, "%#llx", ASMAtomicWriteU64, UINT16_MAX); + TEST_WRITE(pu64, uint64_t, "%#llx", ASMAtomicWriteU64, _1M-1); + TEST_WRITE(pu64, uint64_t, "%#llx", ASMAtomicWriteU64, _1M+1); + TEST_WRITE(pu64, uint64_t, "%#llx", ASMAtomicWriteU64, _1G-1); + TEST_WRITE(pu64, uint64_t, "%#llx", ASMAtomicWriteU64, _1G+1); + TEST_WRITE(pu64, uint64_t, "%#llx", ASMAtomicWriteU64, INT32_MAX); + TEST_WRITE(pu64, uint64_t, "%#llx", ASMAtomicWriteU64, UINT32_MAX); + TEST_WRITE(pu64, uint64_t, "%#llx", ASMAtomicWriteU64, INT64_MAX); + TEST_WRITE(pu64, uint64_t, "%#llx", ASMAtomicWriteU64, UINT64_MAX); + TEST_WRITE(pu64, uint64_t, "%#llx", ASMAtomicWriteU64, UINT64_C(0x450872549687134)); + + volatile int64_t *pi64 = (volatile int64_t *)pu64; + TEST_WRITE(pi64, int64_t, "%d", ASMAtomicWriteS64, INT64_MIN); + TEST_WRITE(pi64, int64_t, "%d", ASMAtomicWriteS64, INT64_MAX); + TEST_WRITE(pi64, int64_t, "%d", ASMAtomicWriteS64, 42); + +#if ARCH_BITS == 64 + size_t volatile *pcb = (size_t volatile *)pu64; + TEST_WRITE(pcb, size_t, "%#zx", ASMAtomicWriteZ, ~(size_t)42); + TEST_WRITE(pcb, size_t, "%#zx", ASMAtomicWriteZ, 42); + + void * volatile *ppv = (void * volatile *)pu64; + TEST_WRITE(ppv, void *, "%#zx", ASMAtomicWritePtrVoid, NULL); + TEST_WRITE(ppv, void *, "%#zx", ASMAtomicWritePtrVoid, (void *)~(uintptr_t)12938754); + + ASMAtomicWriteNullPtr(ppv); CHECKVAL(*ppv, NULL, "%p"); + ASMAtomicWritePtr(ppv, (void *)~(intptr_t)2322434); CHECKVAL(*ppv, (void *)~(intptr_t)2322434, "%p"); + + RTSEMEVENT volatile *phEvt = (RTSEMEVENT volatile *)pu64; + ASMAtomicWriteHandle(phEvt, (RTSEMEVENT)(uintptr_t)99753456); CHECKVAL(*phEvt, (RTSEMEVENT)(uintptr_t)99753456, "%p"); +#endif +} + + +DECLINLINE(void) tstASMAtomicUoWriteU64Worker(uint64_t volatile *pu64) +{ + TEST_WRITE(pu64, uint64_t, "%#llx", ASMAtomicUoWriteU64, 0); + TEST_WRITE(pu64, uint64_t, "%#llx", ASMAtomicUoWriteU64, 19983); + TEST_WRITE(pu64, uint64_t, "%#llx", ASMAtomicUoWriteU64, INT16_MAX); + TEST_WRITE(pu64, uint64_t, "%#llx", ASMAtomicUoWriteU64, UINT16_MAX); + TEST_WRITE(pu64, uint64_t, "%#llx", ASMAtomicUoWriteU64, _1M-1); + TEST_WRITE(pu64, uint64_t, "%#llx", ASMAtomicUoWriteU64, _1M+1); + TEST_WRITE(pu64, uint64_t, "%#llx", ASMAtomicUoWriteU64, _1G-1); + TEST_WRITE(pu64, uint64_t, "%#llx", ASMAtomicUoWriteU64, _1G+1); + TEST_WRITE(pu64, uint64_t, "%#llx", ASMAtomicUoWriteU64, INT32_MAX); + TEST_WRITE(pu64, uint64_t, "%#llx", ASMAtomicUoWriteU64, UINT32_MAX); + TEST_WRITE(pu64, uint64_t, "%#llx", ASMAtomicUoWriteU64, INT64_MAX); + TEST_WRITE(pu64, uint64_t, "%#llx", ASMAtomicUoWriteU64, UINT64_MAX); + TEST_WRITE(pu64, uint64_t, "%#llx", ASMAtomicUoWriteU64, UINT64_C(0x450872549687134)); + + volatile int64_t *pi64 = (volatile int64_t *)pu64; + TEST_WRITE(pi64, int64_t, "%d", ASMAtomicUoWriteS64, INT64_MIN); + TEST_WRITE(pi64, int64_t, "%d", ASMAtomicUoWriteS64, INT64_MAX); + TEST_WRITE(pi64, int64_t, "%d", ASMAtomicUoWriteS64, 42); + +#if ARCH_BITS == 64 + size_t volatile *pcb = (size_t volatile *)pu64; + TEST_WRITE(pcb, size_t, "%#zx", ASMAtomicUoWriteZ, ~(size_t)42); + TEST_WRITE(pcb, size_t, "%#zx", ASMAtomicUoWriteZ, 42); + + void * volatile *ppv = (void * volatile *)pu64; + TEST_WRITE(ppv, void *, "%#zx", ASMAtomicUoWritePtrVoid, NULL); + TEST_WRITE(ppv, void *, "%#zx", ASMAtomicUoWritePtrVoid, (void *)~(uintptr_t)12938754); + + ASMAtomicUoWriteNullPtr(ppv); CHECKVAL(*ppv, NULL, "%p"); + ASMAtomicUoWritePtr(ppv, (void *)~(intptr_t)2322434); CHECKVAL(*ppv, (void *)~(intptr_t)2322434, "%p"); + + RTSEMEVENT volatile *phEvt = (RTSEMEVENT volatile *)pu64; + ASMAtomicUoWriteHandle(phEvt, (RTSEMEVENT)(uintptr_t)99753456); CHECKVAL(*phEvt, (RTSEMEVENT)(uintptr_t)99753456, "%p"); +#endif +} + +static void tstASMAtomicWrite(void) +{ + DO_SIMPLE_TEST(ASMAtomicWriteU8, uint8_t); + DO_SIMPLE_TEST(ASMAtomicUoWriteU8, uint8_t); + + DO_SIMPLE_TEST(ASMAtomicWriteU16, uint16_t); + DO_SIMPLE_TEST(ASMAtomicUoWriteU16, uint16_t); + + DO_SIMPLE_TEST(ASMAtomicWriteU32, uint32_t); + DO_SIMPLE_TEST(ASMAtomicUoWriteU32, uint32_t); + + DO_SIMPLE_TEST(ASMAtomicWriteU64, uint64_t); + DO_SIMPLE_TEST(ASMAtomicUoWriteU64, uint64_t); +} + + +DECLINLINE(void) tstASMAtomicXchgU8Worker(uint8_t volatile *pu8) +{ + *pu8 = 0; + CHECK_OP_AND_VAL(uint8_t, "%#x", pu8, ASMAtomicXchgU8(pu8, 1), 0, 1); + CHECK_OP_AND_VAL(uint8_t, "%#x", pu8, ASMAtomicXchgU8(pu8, UINT8_C(0xff)), 1, UINT8_C(0xff)); + CHECK_OP_AND_VAL(uint8_t, "%#x", pu8, ASMAtomicXchgU8(pu8, UINT8_C(0x87)), UINT8_C(0xff), UINT8_C(0x87)); + CHECK_OP_AND_VAL(uint8_t, "%#x", pu8, ASMAtomicXchgU8(pu8, UINT8_C(0xfe)), UINT8_C(0x87), UINT8_C(0xfe)); + + int8_t volatile *pi8 = (int8_t volatile *)pu8; + CHECK_OP_AND_VAL(int8_t, "%d", pi8, ASMAtomicXchgS8(pi8, INT8_C(-4)), INT8_C(-2), INT8_C(-4)); + CHECK_OP_AND_VAL(int8_t, "%d", pi8, ASMAtomicXchgS8(pi8, INT8_C(4)), INT8_C(-4), INT8_C(4)); + CHECK_OP_AND_VAL(int8_t, "%d", pi8, ASMAtomicXchgS8(pi8, INT8_MAX), INT8_C(4), INT8_MAX); + CHECK_OP_AND_VAL(int8_t, "%d", pi8, ASMAtomicXchgS8(pi8, INT8_MIN), INT8_MAX, INT8_MIN); + CHECK_OP_AND_VAL(int8_t, "%d", pi8, ASMAtomicXchgS8(pi8, 1), INT8_MIN, 1); + + bool volatile *pf = (bool volatile *)pu8; + CHECK_OP_AND_VAL(bool, "%d", pf, ASMAtomicXchgBool(pf, false), true, false); + CHECK_OP_AND_VAL(bool, "%d", pf, ASMAtomicXchgBool(pf, false), false, false); + CHECK_OP_AND_VAL(bool, "%d", pf, ASMAtomicXchgBool(pf, true), false, true); +} + + +DECLINLINE(void) tstASMAtomicXchgU16Worker(uint16_t volatile *pu16) +{ + *pu16 = 0; + CHECK_OP_AND_VAL(uint16_t, "%#x", pu16, ASMAtomicXchgU16(pu16, 1), 0, 1); + CHECK_OP_AND_VAL(uint16_t, "%#x", pu16, ASMAtomicXchgU16(pu16, 0), 1, 0); + CHECK_OP_AND_VAL(uint16_t, "%#x", pu16, ASMAtomicXchgU16(pu16, UINT16_MAX), 0, UINT16_MAX); + CHECK_OP_AND_VAL(uint16_t, "%#x", pu16, ASMAtomicXchgU16(pu16, UINT16_C(0x7fff)), UINT16_MAX, UINT16_C(0x7fff)); + CHECK_OP_AND_VAL(uint16_t, "%#x", pu16, ASMAtomicXchgU16(pu16, UINT16_C(0x8765)), UINT16_C(0x7fff), UINT16_C(0x8765)); + CHECK_OP_AND_VAL(uint16_t, "%#x", pu16, ASMAtomicXchgU16(pu16, UINT16_C(0xfffe)), UINT16_C(0x8765), UINT16_C(0xfffe)); + + int16_t volatile *pi16 = (int16_t volatile *)pu16; + CHECK_OP_AND_VAL(int16_t, "%d", pi16, ASMAtomicXchgS16(pi16, INT16_MIN), INT16_C(-2), INT16_MIN); + CHECK_OP_AND_VAL(int16_t, "%d", pi16, ASMAtomicXchgS16(pi16, INT16_MAX), INT16_MIN, INT16_MAX); + CHECK_OP_AND_VAL(int16_t, "%d", pi16, ASMAtomicXchgS16(pi16, -8), INT16_MAX, -8); + CHECK_OP_AND_VAL(int16_t, "%d", pi16, ASMAtomicXchgS16(pi16, 8), -8, 8); +} + + +DECLINLINE(void) tstASMAtomicXchgU32Worker(uint32_t volatile *pu32) +{ + *pu32 = 0; + CHECK_OP_AND_VAL(uint32_t, "%#x", pu32, ASMAtomicXchgU32(pu32, 1), 0, 1); + CHECK_OP_AND_VAL(uint32_t, "%#x", pu32, ASMAtomicXchgU32(pu32, 0), 1, 0); + CHECK_OP_AND_VAL(uint32_t, "%#x", pu32, ASMAtomicXchgU32(pu32, UINT32_MAX), 0, UINT32_MAX); + CHECK_OP_AND_VAL(uint32_t, "%#x", pu32, ASMAtomicXchgU32(pu32, UINT32_C(0x87654321)), UINT32_MAX, UINT32_C(0x87654321)); + CHECK_OP_AND_VAL(uint32_t, "%#x", pu32, ASMAtomicXchgU32(pu32, UINT32_C(0xfffffffe)), UINT32_C(0x87654321), UINT32_C(0xfffffffe)); + + int32_t volatile *pi32 = (int32_t volatile *)pu32; + CHECK_OP_AND_VAL(int32_t, "%d", pi32, ASMAtomicXchgS32(pi32, INT32_MIN), INT32_C(-2), INT32_MIN); + CHECK_OP_AND_VAL(int32_t, "%d", pi32, ASMAtomicXchgS32(pi32, INT32_MAX), INT32_MIN, INT32_MAX); + CHECK_OP_AND_VAL(int32_t, "%d", pi32, ASMAtomicXchgS32(pi32, -16), INT32_MAX, -16); + CHECK_OP_AND_VAL(int32_t, "%d", pi32, ASMAtomicXchgS32(pi32, 16), -16, 16); + +#if ARCH_BITS == 32 + size_t volatile *pcb = (size_t volatile *)pu32; + CHECK_OP_AND_VAL(size_t, "%#zx", pcb, ASMAtomicXchgZ(pcb, UINT32_C(0x9481239b)), 0x10, UINT32_C(0x9481239b)); + CHECK_OP_AND_VAL(size_t, "%#zx", pcb, ASMAtomicXchgZ(pcb, UINT32_C(0xcdef1234)), UINT32_C(0x9481239b), UINT32_C(0xcdef1234)); +#endif + +#if R0_ARCH_BITS == 32 + RTR0PTR volatile *pR0Ptr = (RTR0PTR volatile *)pu32; + CHECK_OP_AND_VAL(size_t, "%#llx", pcb, ASMAtomicXchgR0Ptr(pR0Ptr, UINT32_C(0x80341237)), UINT32_C(0xcdef1234), UINT32_C(0x80341237)); +#endif +} + + +DECLINLINE(void) tstASMAtomicXchgU64Worker(uint64_t volatile *pu64) +{ + *pu64 = 0; + CHECK_OP_AND_VAL(uint64_t, "%#llx", pu64, ASMAtomicXchgU64(pu64, 1), 0, 1); + CHECK_OP_AND_VAL(uint64_t, "%#llx", pu64, ASMAtomicXchgU64(pu64, 0), 1, 0); + CHECK_OP_AND_VAL(uint64_t, "%#llx", pu64, ASMAtomicXchgU64(pu64, UINT64_MAX), 0, UINT64_MAX); + CHECK_OP_AND_VAL(uint64_t, "%#llx", pu64, ASMAtomicXchgU64(pu64, UINT64_C(0xfedcba0987654321)), UINT64_MAX, UINT64_C(0xfedcba0987654321)); + CHECK_OP_AND_VAL(uint64_t, "%#llx", pu64, ASMAtomicXchgU64(pu64, UINT64_C(0xfffffffffffffffe)), UINT64_C(0xfedcba0987654321), UINT64_C(0xfffffffffffffffe)); + + int64_t volatile *pi64 = (int64_t volatile *)pu64; + CHECK_OP_AND_VAL(int64_t, "%lld", pi64, ASMAtomicXchgS64(pi64, INT64_MAX), -2, INT64_MAX); + CHECK_OP_AND_VAL(int64_t, "%lld", pi64, ASMAtomicXchgS64(pi64, INT64_MIN), INT64_MAX, INT64_MIN); + CHECK_OP_AND_VAL(int64_t, "%lld", pi64, ASMAtomicXchgS64(pi64, -32), INT64_MIN, -32); + CHECK_OP_AND_VAL(int64_t, "%lld", pi64, ASMAtomicXchgS64(pi64, 32), -32, 32); + +#if ARCH_BITS == 64 + size_t volatile *pcb = (size_t volatile *)pu64; + CHECK_OP_AND_VAL(size_t, "%#zx", pcb, ASMAtomicXchgZ(pcb, UINT64_C(0x94812396759)), 0x20, UINT64_C(0x94812396759)); + CHECK_OP_AND_VAL(size_t, "%#zx", pcb, ASMAtomicXchgZ(pcb, UINT64_C(0xcdef1234abdf7896)), UINT64_C(0x94812396759), UINT64_C(0xcdef1234abdf7896)); +#endif + +#if R0_ARCH_BITS == 64 + RTR0PTR volatile *pR0Ptr = (RTR0PTR volatile *)pu64; + CHECK_OP_AND_VAL(size_t, "%#llx", pcb, ASMAtomicXchgR0Ptr(pR0Ptr, UINT64_C(0xfedc1234567890ab)), UINT64_C(0xcdef1234abdf7896), UINT64_C(0xfedc1234567890ab)); +#endif +} + + +DECLINLINE(void) tstASMAtomicXchgPtrWorker(void * volatile *ppv) +{ + *ppv = NULL; + CHECK_OP_AND_VAL(void *, "%p", ppv, ASMAtomicXchgPtr(ppv, (void *)(~(uintptr_t)0)), NULL, (void *)(~(uintptr_t)0)); + CHECK_OP_AND_VAL(void *, "%p", ppv, ASMAtomicXchgPtr(ppv, (void *)(~(uintptr_t)0x87654321)), (void *)(~(uintptr_t)0), (void *)(~(uintptr_t)0x87654321)); + CHECK_OP_AND_VAL(void *, "%p", ppv, ASMAtomicXchgPtr(ppv, NULL), (void *)(~(uintptr_t)0x87654321), NULL); + + CHECK_OP_AND_VAL(void *, "%p", ppv, ASMAtomicXchgR3Ptr(ppv, (void *)ppv), NULL, (void *)ppv); + + RTSEMEVENT volatile *phEvt = (RTSEMEVENT volatile *)ppv; + RTSEMEVENT hRet; + ASMAtomicXchgHandle(phEvt, (RTSEMEVENT)(~(uintptr_t)12345), &hRet); + CHECKVAL(hRet, (RTSEMEVENT)ppv, "%p"); + CHECKVAL(*phEvt, (RTSEMEVENT)(~(uintptr_t)12345), "%p"); +} + + +static void tstASMAtomicXchg(void) +{ + DO_SIMPLE_TEST(ASMAtomicXchgU8, uint8_t); + DO_SIMPLE_TEST(ASMAtomicXchgU16, uint16_t); + DO_SIMPLE_TEST(ASMAtomicXchgU32, uint32_t); + DO_SIMPLE_TEST(ASMAtomicXchgU64, uint64_t); + DO_SIMPLE_TEST(ASMAtomicXchgPtr, void *); +} + + +DECLINLINE(void) tstASMAtomicCmpXchgU8Worker(uint8_t volatile *pu8) +{ + *pu8 = 0xff; + CHECK_OP_AND_VAL_EX(bool, "%d", "%#x", pu8, ASMAtomicCmpXchgU8(pu8, 0, 0), false, 0xff); + CHECK_OP_AND_VAL_EX(bool, "%d", "%#x", pu8, ASMAtomicCmpXchgU8(pu8, 0, 0xff), true, 0); + CHECK_OP_AND_VAL_EX(bool, "%d", "%#x", pu8, ASMAtomicCmpXchgU8(pu8, 0x97, 0), true, 0x97); + CHECK_OP_AND_VAL_EX(bool, "%d", "%#x", pu8, ASMAtomicCmpXchgU8(pu8, 0x97, 0), false, 0x97); + CHECK_OP_AND_VAL_EX(bool, "%d", "%#x", pu8, ASMAtomicCmpXchgU8(pu8, 0x7f, 0x97), true, 0x7f); + + int8_t volatile *pi8 = (int8_t volatile *)pu8; + CHECK_OP_AND_VAL(bool, "%d", pi8, ASMAtomicCmpXchgS8(pi8, -2, 0x7f), true, -2); + CHECK_OP_AND_VAL(bool, "%d", pi8, ASMAtomicCmpXchgS8(pi8, INT8_MAX, -2), true, INT8_MAX); + CHECK_OP_AND_VAL(bool, "%d", pi8, ASMAtomicCmpXchgS8(pi8, INT8_MAX, INT8_MIN), false, INT8_MAX); + CHECK_OP_AND_VAL(bool, "%d", pi8, ASMAtomicCmpXchgS8(pi8, INT8_MIN, INT8_MAX), true, INT8_MIN); + CHECK_OP_AND_VAL(bool, "%d", pi8, ASMAtomicCmpXchgS8(pi8, 1, INT8_MIN), true, 1); + + bool volatile *pf = (bool volatile *)pu8; + CHECK_OP_AND_VAL(bool, "%d", pf, ASMAtomicCmpXchgBool(pf, true, true), true, true); + CHECK_OP_AND_VAL(bool, "%d", pf, ASMAtomicCmpXchgBool(pf, false, true), true, false); + CHECK_OP_AND_VAL(bool, "%d", pf, ASMAtomicCmpXchgBool(pf, false, true), false, false); + CHECK_OP_AND_VAL(bool, "%d", pf, ASMAtomicCmpXchgBool(pf, false, false), true, false); +} + + +DECLINLINE(void) tstASMAtomicCmpXchgU32Worker(uint32_t volatile *pu32) +{ + *pu32 = UINT32_C(0xffffffff); + CHECK_OP_AND_VAL_EX(bool, "%d", "%#x", pu32, ASMAtomicCmpXchgU32(pu32, 0, 0), false, UINT32_C(0xffffffff)); + CHECK_OP_AND_VAL_EX(bool, "%d", "%#x", pu32, ASMAtomicCmpXchgU32(pu32, 0, UINT32_C(0xffffffff)), true, 0); + CHECK_OP_AND_VAL_EX(bool, "%d", "%#x", pu32, ASMAtomicCmpXchgU32(pu32, UINT32_C(0x80088efd), UINT32_C(0x12345678)), false, 0); + CHECK_OP_AND_VAL_EX(bool, "%d", "%#x", pu32, ASMAtomicCmpXchgU32(pu32, UINT32_C(0x80088efd), 0), true, UINT32_C(0x80088efd)); + CHECK_OP_AND_VAL_EX(bool, "%d", "%#x", pu32, ASMAtomicCmpXchgU32(pu32, UINT32_C(0xfffffffe), UINT32_C(0x80088efd)), true, UINT32_C(0xfffffffe)); + + int32_t volatile *pi32 = (int32_t volatile *)pu32; + CHECK_OP_AND_VAL_EX(bool, "%d", "%d", pi32, ASMAtomicCmpXchgS32(pi32, INT32_MIN, 2), false, -2); + CHECK_OP_AND_VAL_EX(bool, "%d", "%d", pi32, ASMAtomicCmpXchgS32(pi32, INT32_MIN, -2), true, INT32_MIN); + CHECK_OP_AND_VAL_EX(bool, "%d", "%d", pi32, ASMAtomicCmpXchgS32(pi32, -19, -2), false, INT32_MIN); + CHECK_OP_AND_VAL_EX(bool, "%d", "%d", pi32, ASMAtomicCmpXchgS32(pi32, -19, INT32_MIN), true, -19); + CHECK_OP_AND_VAL_EX(bool, "%d", "%d", pi32, ASMAtomicCmpXchgS32(pi32, -19, INT32_MIN), false, -19); + CHECK_OP_AND_VAL_EX(bool, "%d", "%d", pi32, ASMAtomicCmpXchgS32(pi32, 19, -19), true, 19); + CHECK_OP_AND_VAL_EX(bool, "%d", "%d", pi32, ASMAtomicCmpXchgS32(pi32, INT32_MAX, -234), false, 19); + CHECK_OP_AND_VAL_EX(bool, "%d", "%d", pi32, ASMAtomicCmpXchgS32(pi32, INT32_MAX, 19), true, INT32_MAX); + +#if ARCH_BITS == 32 + *pu32 = 29; + void * volatile *ppv = (void * volatile *)pu32; + CHECK_OP_AND_VAL_EX(bool, "%d", "%p", ppv, ASMAtomicCmpXchgPtrVoid(ppv, NULL, (void *)(intptr_t)-29), false, (void *)(intptr_t)29); + CHECK_OP_AND_VAL_EX(bool, "%d", "%p", ppv, ASMAtomicCmpXchgPtrVoid(ppv, NULL, (void *)(intptr_t)29), true, NULL); + CHECK_OP_AND_VAL_EX(bool, "%d", "%p", ppv, ASMAtomicCmpXchgPtrVoid(ppv, NULL, (void *)(intptr_t)29), false, NULL); + CHECK_OP_AND_VAL_EX(bool, "%d", "%p", ppv, ASMAtomicCmpXchgPtrVoid(ppv, (void *)~(uintptr_t)42, NULL), true, (void *)~(uintptr_t)42); + + bool fRc; + RTSEMEVENT volatile *phEvt = (RTSEMEVENT volatile *)pu32; + ASMAtomicCmpXchgHandle(phEvt, (RTSEMEVENT)~(uintptr_t)0x12356389, (RTSEMEVENT)NULL, fRc); + CHECKVAL(fRc, false, "%d"); + CHECKVAL(*phEvt, (RTSEMEVENT)~(uintptr_t)42, "%p"); + + ASMAtomicCmpXchgHandle(phEvt, (RTSEMEVENT)~(uintptr_t)0x12356389, (RTSEMEVENT)~(uintptr_t)42, fRc); + CHECKVAL(fRc, true, "%d"); + CHECKVAL(*phEvt, (RTSEMEVENT)~(uintptr_t)0x12356389, "%p"); +#endif +} + + +DECLINLINE(void) tstASMAtomicCmpXchgU64Worker(uint64_t volatile *pu64) +{ + *pu64 = UINT64_C(0xffffffffffffff); + CHECK_OP_AND_VAL_EX(bool, "%d", "%#llx", pu64, ASMAtomicCmpXchgU64(pu64, 0, 0), false, UINT64_C(0xffffffffffffff)); + CHECK_OP_AND_VAL_EX(bool, "%d", "%#llx", pu64, ASMAtomicCmpXchgU64(pu64, 0, UINT64_C(0xffffffffffffff)), true, 0); + CHECK_OP_AND_VAL_EX(bool, "%d", "%#llx", pu64, ASMAtomicCmpXchgU64(pu64, UINT64_C(0x80040008008efd), 1), false, 0); + CHECK_OP_AND_VAL_EX(bool, "%d", "%#llx", pu64, ASMAtomicCmpXchgU64(pu64, UINT64_C(0x80040008008efd), 0), true, UINT64_C(0x80040008008efd)); + CHECK_OP_AND_VAL_EX(bool, "%d", "%#llx", pu64, ASMAtomicCmpXchgU64(pu64, UINT64_C(0x80040008008efd), 0), false, UINT64_C(0x80040008008efd)); + CHECK_OP_AND_VAL_EX(bool, "%d", "%#llx", pu64, ASMAtomicCmpXchgU64(pu64, UINT64_C(0xfffffffffffffffd), UINT64_C(0x80040008008efd)), true, UINT64_C(0xfffffffffffffffd)); + + int64_t volatile *pi64 = (int64_t volatile *)pu64; + CHECK_OP_AND_VAL_EX(bool, "%d", "%#lld", pi64, ASMAtomicCmpXchgS64(pi64, INT64_MAX, 0), false, -3); + CHECK_OP_AND_VAL_EX(bool, "%d", "%#lld", pi64, ASMAtomicCmpXchgS64(pi64, INT64_MAX, -3), true, INT64_MAX); + CHECK_OP_AND_VAL_EX(bool, "%d", "%#lld", pi64, ASMAtomicCmpXchgS64(pi64, INT64_MIN, INT64_MIN), false, INT64_MAX); + CHECK_OP_AND_VAL_EX(bool, "%d", "%#lld", pi64, ASMAtomicCmpXchgS64(pi64, INT64_MIN, INT64_MAX), true, INT64_MIN); + CHECK_OP_AND_VAL_EX(bool, "%d", "%#lld", pi64, ASMAtomicCmpXchgS64(pi64, -29, -29), false, INT64_MIN); + CHECK_OP_AND_VAL_EX(bool, "%d", "%#lld", pi64, ASMAtomicCmpXchgS64(pi64, -29, INT64_MIN), true, -29); + CHECK_OP_AND_VAL_EX(bool, "%d", "%#lld", pi64, ASMAtomicCmpXchgS64(pi64, -29, INT64_MIN), false, -29); + CHECK_OP_AND_VAL_EX(bool, "%d", "%#lld", pi64, ASMAtomicCmpXchgS64(pi64, 29, -29), true, 29); + +#if ARCH_BITS == 64 + void * volatile *ppv = (void * volatile *)pu64; + CHECK_OP_AND_VAL_EX(bool, "%d", "%p", ppv, ASMAtomicCmpXchgPtrVoid(ppv, NULL, (void *)(intptr_t)-29), false, (void *)(intptr_t)29); + CHECK_OP_AND_VAL_EX(bool, "%d", "%p", ppv, ASMAtomicCmpXchgPtrVoid(ppv, NULL, (void *)(intptr_t)29), true, NULL); + CHECK_OP_AND_VAL_EX(bool, "%d", "%p", ppv, ASMAtomicCmpXchgPtrVoid(ppv, NULL, (void *)(intptr_t)29), false, NULL); + CHECK_OP_AND_VAL_EX(bool, "%d", "%p", ppv, ASMAtomicCmpXchgPtrVoid(ppv, (void *)~(uintptr_t)42, NULL), true, (void *)~(uintptr_t)42); + + bool fRc; + RTSEMEVENT volatile *phEvt = (RTSEMEVENT volatile *)pu64; + ASMAtomicCmpXchgHandle(phEvt, (RTSEMEVENT)~(uintptr_t)0x12356389, (RTSEMEVENT)NULL, fRc); + CHECKVAL(fRc, false, "%d"); + CHECKVAL(*phEvt, (RTSEMEVENT)~(uintptr_t)42, "%p"); + + ASMAtomicCmpXchgHandle(phEvt, (RTSEMEVENT)~(uintptr_t)0x12356389, (RTSEMEVENT)~(uintptr_t)42, fRc); + CHECKVAL(fRc, true, "%d"); + CHECKVAL(*phEvt, (RTSEMEVENT)~(uintptr_t)0x12356389, "%p"); +#endif +} + + +#ifdef RTASM_HAVE_CMP_WRITE_U128 +DECLINLINE(void) tstASMAtomicCmpWriteU128Worker(RTUINT128U volatile *pu128) +{ + pu128->s.Lo = UINT64_C(0xffffffffffffff); + pu128->s.Hi = UINT64_C(0xffffffffffffff); + + RTUINT128U u128A, u128B; + CHECK_OP_AND_VAL_128_C(bool, "%d", pu128, ASMAtomicCmpWriteU128U(pu128, + u128A = RTUINT128_INIT_C(0, 0), + u128B = RTUINT128_INIT_C(0, 0)), + false, 0xffffffffffffff, 0xffffffffffffff); + CHECK_OP_AND_VAL_128_C(bool, "%d", pu128, ASMAtomicCmpWriteU128U(pu128, + u128A = RTUINT128_INIT_C(0, 0), + u128B = RTUINT128_INIT_C(0xffffffffffffff, 0xffffffffffffff)), + true, 0, 0); + + CHECK_OP_AND_VAL_128_C(bool, "%d", pu128, ASMAtomicCmpWriteU128U(pu128, + u128A = RTUINT128_INIT_C(0x80040008008efd, 0x40080004004def), + u128B = RTUINT128_INIT_C(0, 1)), + false, 0, 0); + CHECK_OP_AND_VAL_128_C(bool, "%d", pu128, ASMAtomicCmpWriteU128U(pu128, + u128A = RTUINT128_INIT_C(0x80040008008efd, 0x40080004004def), + u128B = RTUINT128_INIT_C(1, 0)), + false, 0, 0); + CHECK_OP_AND_VAL_128_C(bool, "%d", pu128, ASMAtomicCmpWriteU128U(pu128, + u128A = RTUINT128_INIT_C(0x80040008008efd, 0x40080004004def), + u128B = RTUINT128_INIT_C(0, 0)), + true, 0x80040008008efd, 0x40080004004def); + + CHECK_OP_AND_VAL_128_C(bool, "%d", pu128, ASMAtomicCmpWriteU128U(pu128, + u128A = RTUINT128_INIT_C(0xfff40ff8f08ef3, 0x4ee8ee04cc4de4), + u128B = RTUINT128_INIT_C(0x80040008008efd, 0)), + false, 0x80040008008efd, 0x40080004004def); + CHECK_OP_AND_VAL_128_C(bool, "%d", pu128, ASMAtomicCmpWriteU128U(pu128, + u128A = RTUINT128_INIT_C(0xfff40ff8f08ef3, 0x4ee8ee04cc4de4), + u128B = RTUINT128_INIT_C(0, 0x40080004004def)), + false, 0x80040008008efd, 0x40080004004def); + CHECK_OP_AND_VAL_128_C(bool, "%d", pu128, ASMAtomicCmpWriteU128U(pu128, + u128A = RTUINT128_INIT_C(0xfff40ff8f08ef3, 0x4ee8ee04cc4de4), + u128B = RTUINT128_INIT_C(0x80040008008efd, 0x40080004004def)), + true, 0xfff40ff8f08ef3, 0x4ee8ee04cc4de4); + + /* Make sure the v2 version works too (arm) */ + CHECK_OP_AND_VAL_128_C(bool, "%d", pu128, ASMAtomicCmpWriteU128v2(&pu128->u, + UINT64_C(0x95487930069587), UINT64_C(0x89958490385964), + UINT64_C(0xfff40ff8f08ef3), UINT64_C(0x4ee8ee04cc4de4)), + true, 0x95487930069587, 0x89958490385964); + CHECK_OP_AND_VAL_128_C(bool, "%d", pu128, ASMAtomicCmpWriteU128v2(&pu128->u, + UINT64_C(0x99969404869434), UINT64_C(0x11049309994567), + UINT64_C(0x33f40ff8f08eff), UINT64_C(0x99e8ee04cc4dee)), + false, 0x95487930069587, 0x89958490385964); +} +#endif /* RTASM_HAVE_CMP_WRITE_U128 */ + + +static void tstASMAtomicCmpXchg(void) +{ + DO_SIMPLE_TEST(ASMAtomicCmpXchgU8, uint8_t); + DO_SIMPLE_TEST(ASMAtomicCmpXchgU32, uint32_t); + DO_SIMPLE_TEST(ASMAtomicCmpXchgU64, uint64_t); +#ifdef RTASM_HAVE_CMP_WRITE_U128 +# ifdef RT_ARCH_AMD64 + if (ASMCpuId_ECX(1) & X86_CPUID_FEATURE_ECX_CX16) +# endif + { + RTTestISub("ASMAtomicCmpWriteU128U"); + DO_SIMPLE_TEST_NO_SUB_NO_STACK(tstASMAtomicCmpWriteU128Worker, RTUINT128U); + } +#endif +} + + +DECLINLINE(void) tstASMAtomicCmpXchgExU8Worker(uint8_t volatile *pu8) +{ + *pu8 = UINT8_C(0xff); + uint8_t u8Old = UINT8_C(0x11); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#x", pu8, u8Old, ASMAtomicCmpXchgExU8(pu8, 0, 0, &u8Old), false, UINT8_C(0xff), UINT8_C(0xff)); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#x", pu8, u8Old, ASMAtomicCmpXchgExU8(pu8, 0, UINT8_C(0xff), &u8Old), true, 0, UINT8_C(0xff)); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#x", pu8, u8Old, ASMAtomicCmpXchgExU8(pu8, 0, UINT8_C(0xff), &u8Old), false, 0, UINT8_C(0x00)); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#x", pu8, u8Old, ASMAtomicCmpXchgExU8(pu8, UINT8_C(0xfd), 0, &u8Old), true, UINT8_C(0xfd), 0); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#x", pu8, u8Old, ASMAtomicCmpXchgExU8(pu8, UINT8_C(0xfd), 0, &u8Old), false, UINT8_C(0xfd), UINT8_C(0xfd)); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#x", pu8, u8Old, ASMAtomicCmpXchgExU8(pu8, UINT8_C(0xe0), UINT8_C(0xfd), &u8Old), true, UINT8_C(0xe0), UINT8_C(0xfd)); + + int8_t volatile *pi8 = (int8_t volatile *)pu8; + int8_t i8Old = 0; + CHECK_OP_AND_VAL_EX2(bool, "%d", "%d", pi8, i8Old, ASMAtomicCmpXchgExS8(pi8, 32, 32, &i8Old), false, -32, -32); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%d", pi8, i8Old, ASMAtomicCmpXchgExS8(pi8, 32, -32, &i8Old), true, 32, -32); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%d", pi8, i8Old, ASMAtomicCmpXchgExS8(pi8, INT8_MIN, 32, &i8Old), true, INT8_MIN, 32); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%d", pi8, i8Old, ASMAtomicCmpXchgExS8(pi8, INT8_MIN, 32, &i8Old), false, INT8_MIN, INT8_MIN); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%d", pi8, i8Old, ASMAtomicCmpXchgExS8(pi8, INT8_MAX, INT8_MAX, &i8Old), false, INT8_MIN, INT8_MIN); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%d", pi8, i8Old, ASMAtomicCmpXchgExS8(pi8, INT8_MAX, INT8_MIN, &i8Old), true, INT8_MAX, INT8_MIN); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%d", pi8, i8Old, ASMAtomicCmpXchgExS8(pi8, 42, INT8_MAX, &i8Old), true, 42, INT8_MAX); +} + + +DECLINLINE(void) tstASMAtomicCmpXchgExU16Worker(uint16_t volatile *pu16) +{ + *pu16 = UINT16_C(0xffff); + uint16_t u16Old = UINT16_C(0x5111); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#x", pu16, u16Old, ASMAtomicCmpXchgExU16(pu16, 0, 0, &u16Old), false, UINT16_C(0xffff), UINT16_C(0xffff)); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#x", pu16, u16Old, ASMAtomicCmpXchgExU16(pu16, 0, UINT16_C(0xffff), &u16Old), true, 0, UINT16_C(0xffff)); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#x", pu16, u16Old, ASMAtomicCmpXchgExU16(pu16, 0, UINT16_C(0xffff), &u16Old), false, 0, UINT16_C(0x0000)); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#x", pu16, u16Old, ASMAtomicCmpXchgExU16(pu16, UINT16_C(0x8efd), 0, &u16Old), true, UINT16_C(0x8efd), 0); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#x", pu16, u16Old, ASMAtomicCmpXchgExU16(pu16, UINT16_C(0x8efd), 0, &u16Old), false, UINT16_C(0x8efd), UINT16_C(0x8efd)); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#x", pu16, u16Old, ASMAtomicCmpXchgExU16(pu16, UINT16_C(0xffe0), UINT16_C(0x8efd), &u16Old), true, UINT16_C(0xffe0), UINT16_C(0x8efd)); + + int16_t volatile *pi16 = (int16_t volatile *)pu16; + int16_t i16Old = 0; + CHECK_OP_AND_VAL_EX2(bool, "%d", "%d", pi16, i16Old, ASMAtomicCmpXchgExS16(pi16, 32, 32, &i16Old), false, -32, -32); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%d", pi16, i16Old, ASMAtomicCmpXchgExS16(pi16, 32, -32, &i16Old), true, 32, -32); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%d", pi16, i16Old, ASMAtomicCmpXchgExS16(pi16, INT16_MIN, 32, &i16Old), true, INT16_MIN, 32); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%d", pi16, i16Old, ASMAtomicCmpXchgExS16(pi16, INT16_MIN, 32, &i16Old), false, INT16_MIN, INT16_MIN); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%d", pi16, i16Old, ASMAtomicCmpXchgExS16(pi16, INT16_MAX, INT16_MAX, &i16Old), false, INT16_MIN, INT16_MIN); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%d", pi16, i16Old, ASMAtomicCmpXchgExS16(pi16, INT16_MAX, INT16_MIN, &i16Old), true, INT16_MAX, INT16_MIN); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%d", pi16, i16Old, ASMAtomicCmpXchgExS16(pi16, 42, INT16_MAX, &i16Old), true, 42, INT16_MAX); +} + + +DECLINLINE(void) tstASMAtomicCmpXchgExU32Worker(uint32_t volatile *pu32) +{ + *pu32 = UINT32_C(0xffffffff); + uint32_t u32Old = UINT32_C(0x80005111); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#x", pu32, u32Old, ASMAtomicCmpXchgExU32(pu32, 0, 0, &u32Old), false, UINT32_C(0xffffffff), UINT32_C(0xffffffff)); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#x", pu32, u32Old, ASMAtomicCmpXchgExU32(pu32, 0, UINT32_C(0xffffffff), &u32Old), true, 0, UINT32_C(0xffffffff)); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#x", pu32, u32Old, ASMAtomicCmpXchgExU32(pu32, 0, UINT32_C(0xffffffff), &u32Old), false, 0, UINT32_C(0x00000000)); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#x", pu32, u32Old, ASMAtomicCmpXchgExU32(pu32, UINT32_C(0x80088efd), 0, &u32Old), true, UINT32_C(0x80088efd), 0); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#x", pu32, u32Old, ASMAtomicCmpXchgExU32(pu32, UINT32_C(0x80088efd), 0, &u32Old), false, UINT32_C(0x80088efd), UINT32_C(0x80088efd)); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#x", pu32, u32Old, ASMAtomicCmpXchgExU32(pu32, UINT32_C(0xffffffe0), UINT32_C(0x80088efd), &u32Old), true, UINT32_C(0xffffffe0), UINT32_C(0x80088efd)); + + int32_t volatile *pi32 = (int32_t volatile *)pu32; + int32_t i32Old = 0; + CHECK_OP_AND_VAL_EX2(bool, "%d", "%d", pi32, i32Old, ASMAtomicCmpXchgExS32(pi32, 32, 32, &i32Old), false, -32, -32); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%d", pi32, i32Old, ASMAtomicCmpXchgExS32(pi32, 32, -32, &i32Old), true, 32, -32); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%d", pi32, i32Old, ASMAtomicCmpXchgExS32(pi32, INT32_MIN, 32, &i32Old), true, INT32_MIN, 32); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%d", pi32, i32Old, ASMAtomicCmpXchgExS32(pi32, INT32_MIN, 32, &i32Old), false, INT32_MIN, INT32_MIN); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%d", pi32, i32Old, ASMAtomicCmpXchgExS32(pi32, INT32_MAX, INT32_MAX, &i32Old), false, INT32_MIN, INT32_MIN); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%d", pi32, i32Old, ASMAtomicCmpXchgExS32(pi32, INT32_MAX, INT32_MIN, &i32Old), true, INT32_MAX, INT32_MIN); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%d", pi32, i32Old, ASMAtomicCmpXchgExS32(pi32, 42, INT32_MAX, &i32Old), true, 42, INT32_MAX); + +#if ARCH_BITS == 32 + RTSEMEVENT volatile *phEvt = (RTSEMEVENT volatile *)pu32; + RTSEMEVENT hEvtOld = (RTSEMEVENT)~(uintptr_t)31; + bool fRc = true; + ASMAtomicCmpXchgExHandle(phEvt, (RTSEMEVENT)~(uintptr_t)0x12380964, (RTSEMEVENT)~(uintptr_t)0, fRc, &hEvtOld); + CHECKVAL(fRc, false, "%d"); + CHECKVAL(*phEvt, (RTSEMEVENT)(uintptr_t)42, "%p"); + CHECKVAL(hEvtOld, (RTSEMEVENT)(uintptr_t)42, "%p"); + + ASMAtomicCmpXchgExHandle(phEvt, (RTSEMEVENT)~(uintptr_t)0x12380964, (RTSEMEVENT)(uintptr_t)42, fRc, &hEvtOld); + CHECKVAL(fRc, true, "%d"); + CHECKVAL(*phEvt, (RTSEMEVENT)~(uintptr_t)0x12380964, "%p"); + CHECKVAL(hEvtOld, (RTSEMEVENT)(uintptr_t)42, "%p"); +#endif +} + + +DECLINLINE(void) tstASMAtomicCmpXchgExU64Worker(uint64_t volatile *pu64) +{ + *pu64 = UINT64_C(0xffffffffffffffff); + uint64_t u64Old = UINT64_C(0x8000000051111111); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#llx", pu64, u64Old, ASMAtomicCmpXchgExU64(pu64, 0, 0, &u64Old), false, UINT64_C(0xffffffffffffffff), UINT64_C(0xffffffffffffffff)); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#llx", pu64, u64Old, ASMAtomicCmpXchgExU64(pu64, 0, UINT64_C(0xffffffffffffffff), &u64Old), true, 0, UINT64_C(0xffffffffffffffff)); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#llx", pu64, u64Old, ASMAtomicCmpXchgExU64(pu64, UINT64_C(0x0080040008008efd), 0x342, &u64Old), false, 0, 0); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#llx", pu64, u64Old, ASMAtomicCmpXchgExU64(pu64, UINT64_C(0x0080040008008efd), 0, &u64Old), true, UINT64_C(0x0080040008008efd), 0); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#llx", pu64, u64Old, ASMAtomicCmpXchgExU64(pu64, UINT64_C(0xffffffffffffffc0), UINT64_C(0x0080040008008efd), &u64Old), true, UINT64_C(0xffffffffffffffc0), UINT64_C(0x0080040008008efd)); + + int64_t volatile *pi64 = (int64_t volatile *)pu64; + int64_t i64Old = -3; + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#lld", pi64, i64Old, ASMAtomicCmpXchgExS64(pi64, 64, 64, &i64Old), false, -64, -64); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#lld", pi64, i64Old, ASMAtomicCmpXchgExS64(pi64, 64, -64, &i64Old), true, 64, -64); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#lld", pi64, i64Old, ASMAtomicCmpXchgExS64(pi64, 64, -64, &i64Old), false, 64, 64); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#lld", pi64, i64Old, ASMAtomicCmpXchgExS64(pi64, INT64_MIN, -64, &i64Old), false, 64, 64); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#lld", pi64, i64Old, ASMAtomicCmpXchgExS64(pi64, INT64_MIN, 64, &i64Old), true, INT64_MIN, 64); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#lld", pi64, i64Old, ASMAtomicCmpXchgExS64(pi64, INT64_MAX, INT64_MIN, &i64Old), true, INT64_MAX, INT64_MIN); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%#lld", pi64, i64Old, ASMAtomicCmpXchgExS64(pi64, 42, INT64_MAX, &i64Old), true, 42, INT64_MAX); + +#if ARCH_BITS == 64 + RTSEMEVENT volatile *phEvt = (RTSEMEVENT volatile *)pu64; + RTSEMEVENT hEvtOld = (RTSEMEVENT)~(uintptr_t)31; + bool fRc = true; + ASMAtomicCmpXchgExHandle(phEvt, (RTSEMEVENT)~(uintptr_t)0x12380964, (RTSEMEVENT)~(uintptr_t)0, fRc, &hEvtOld); + CHECKVAL(fRc, false, "%d"); + CHECKVAL(*phEvt, (RTSEMEVENT)(uintptr_t)42, "%p"); + CHECKVAL(hEvtOld, (RTSEMEVENT)(uintptr_t)42, "%p"); + + ASMAtomicCmpXchgExHandle(phEvt, (RTSEMEVENT)~(uintptr_t)0x12380964, (RTSEMEVENT)(uintptr_t)42, fRc, &hEvtOld); + CHECKVAL(fRc, true, "%d"); + CHECKVAL(*phEvt, (RTSEMEVENT)~(uintptr_t)0x12380964, "%p"); + CHECKVAL(hEvtOld, (RTSEMEVENT)(uintptr_t)42, "%p"); + + void * volatile *ppv = (void * volatile *)pu64; + void *pvOld; + CHECK_OP_AND_VAL_EX2(bool, "%d", "%p", ppv, pvOld, ASMAtomicCmpXchgExPtrVoid(ppv, (void *)(intptr_t)12345678, NULL, &pvOld), false, (void *)~(uintptr_t)0x12380964, (void *)~(uintptr_t)0x12380964); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%p", ppv, pvOld, ASMAtomicCmpXchgExPtrVoid(ppv, (void *)(intptr_t)12345678, (void *)~(uintptr_t)0x12380964, &pvOld), true, (void *)(intptr_t)12345678, (void *)~(uintptr_t)0x12380964); + + CHECK_OP_AND_VAL_EX2(bool, "%d", "%p", ppv, pvOld, ASMAtomicCmpXchgExPtr(ppv, (void *)~(uintptr_t)99, (void *)~(uintptr_t)99, &pvOld), false, (void *)(intptr_t)12345678, (void *)(intptr_t)12345678); + CHECK_OP_AND_VAL_EX2(bool, "%d", "%p", ppv, pvOld, ASMAtomicCmpXchgExPtr(ppv, (void *)~(uintptr_t)99, (void *)(intptr_t)12345678, &pvOld), true, (void *)~(intptr_t)99, (void *)(intptr_t)12345678); +#endif +} + + +#ifdef RTASM_HAVE_CMP_XCHG_U128 +DECLINLINE(void) tstASMAtomicCmpXchgU128Worker(RTUINT128U volatile *pu128) +{ + pu128->s.Lo = UINT64_C(0xffffffffffffff); + pu128->s.Hi = UINT64_C(0xffffffffffffff); + + RTUINT128U u128A, u128B; + RTUINT128U const u128OldInit = RTUINT128_INIT_C(0x4242424242424242, 0x2222222222222222); + RTUINT128U u128Old = u128OldInit; + CHECK_OP_AND_VAL_128_C(bool, "%d", pu128, ASMAtomicCmpXchgU128U(pu128, + u128A = RTUINT128_INIT_C(0, 0), + u128B = RTUINT128_INIT_C(0, 0), + &u128Old), + false, 0xffffffffffffff, 0xffffffffffffff); + CHECKVAL128_C(&u128Old, 0xffffffffffffff, 0xffffffffffffff); + + u128Old = u128OldInit; + CHECK_OP_AND_VAL_128_C(bool, "%d", pu128, ASMAtomicCmpXchgU128U(pu128, + u128A = RTUINT128_INIT_C(0, 0), + u128B = RTUINT128_INIT_C(0xffffffffffffff, 0xffffffffffffff), + &u128Old), + true, 0, 0); + CHECKVAL128_C(&u128Old, 0xffffffffffffff, 0xffffffffffffff); + + u128Old = u128OldInit; + CHECK_OP_AND_VAL_128_C(bool, "%d", pu128, ASMAtomicCmpXchgU128U(pu128, + u128A = RTUINT128_INIT_C(0x80040008008efd, 0x40080004004def), + u128B = RTUINT128_INIT_C(0, 1), + &u128Old), + false, 0, 0); + CHECKVAL128_C(&u128Old, 0, 0); + + u128Old = u128OldInit; + CHECK_OP_AND_VAL_128_C(bool, "%d", pu128, ASMAtomicCmpXchgU128U(pu128, + u128A = RTUINT128_INIT_C(0x80040008008efd, 0x40080004004def), + u128B = RTUINT128_INIT_C(1, 0), + &u128Old), + false, 0, 0); + CHECKVAL128_C(&u128Old, 0, 0); + + u128Old = u128OldInit; + CHECK_OP_AND_VAL_128_C(bool, "%d", pu128, ASMAtomicCmpXchgU128U(pu128, + u128A = RTUINT128_INIT_C(0x80040008008efd, 0x40080004004def), + u128B = RTUINT128_INIT_C(0, 0), + &u128Old), + true, 0x80040008008efd, 0x40080004004def); + CHECKVAL128_C(&u128Old, 0, 0); + + u128Old = u128OldInit; + CHECK_OP_AND_VAL_128_C(bool, "%d", pu128, ASMAtomicCmpXchgU128U(pu128, + u128A = RTUINT128_INIT_C(0xfff40ff8f08ef3, 0x4ee8ee04cc4de4), + u128B = RTUINT128_INIT_C(0x80040008008efd, 0), + &u128Old), + false, 0x80040008008efd, 0x40080004004def); + CHECKVAL128_C(&u128Old, 0x80040008008efd, 0x40080004004def); + + u128Old = u128OldInit; + CHECK_OP_AND_VAL_128_C(bool, "%d", pu128, ASMAtomicCmpXchgU128U(pu128, + u128A = RTUINT128_INIT_C(0xfff40ff8f08ef3, 0x4ee8ee04cc4de4), + u128B = RTUINT128_INIT_C(0, 0x40080004004def), + &u128Old), + false, 0x80040008008efd, 0x40080004004def); + CHECKVAL128_C(&u128Old, 0x80040008008efd, 0x40080004004def); + + u128Old = u128OldInit; + CHECK_OP_AND_VAL_128_C(bool, "%d", pu128, ASMAtomicCmpXchgU128U(pu128, + u128A = RTUINT128_INIT_C(0xfff40ff8f08ef3, 0x4ee8ee04cc4de4), + u128B = RTUINT128_INIT_C(0x80040008008efd, 0x40080004004def), + &u128Old), + true, 0xfff40ff8f08ef3, 0x4ee8ee04cc4de4); + CHECKVAL128_C(&u128Old, 0x80040008008efd, 0x40080004004def); + + /* Make sure the v2 version works too (arm) */ + u128Old = u128OldInit; + CHECK_OP_AND_VAL_128_C(bool, "%d", pu128, ASMAtomicCmpXchgU128v2(&pu128->u, + UINT64_C(0x78039485960543), UINT64_C(0x97058437294586), + UINT64_C(0xfff40ff8f08ef3), UINT64_C(0x4ee8ee04cc4de4), + &u128Old.u), + true, 0x78039485960543, 0x97058437294586); + CHECKVAL128_C(&u128Old, 0xfff40ff8f08ef3, 0x4ee8ee04cc4de4); + + u128Old = u128OldInit; + CHECK_OP_AND_VAL_128_C(bool, "%d", pu128, ASMAtomicCmpXchgU128v2(&pu128->u, + UINT64_C(0x13495874560495), UINT64_C(0x12304896098597), + UINT64_C(0xfff40ff8f08ef3), UINT64_C(0x4ee8ee04cc4de4), + &u128Old.u), + false, 0x78039485960543, 0x97058437294586); + CHECKVAL128_C(&u128Old, 0x78039485960543, 0x97058437294586); +} +#endif /* RTASM_HAVE_CMP_XCHG_U128 */ + + +static void tstASMAtomicCmpXchgEx(void) +{ + DO_SIMPLE_TEST(ASMAtomicCmpXchgExU8, uint8_t); + DO_SIMPLE_TEST(ASMAtomicCmpXchgExU16, uint16_t); + DO_SIMPLE_TEST(ASMAtomicCmpXchgExU32, uint32_t); + DO_SIMPLE_TEST(ASMAtomicCmpXchgExU64, uint64_t); +#ifdef RTASM_HAVE_CMP_XCHG_U128 +# ifdef RT_ARCH_AMD64 + if (ASMCpuId_ECX(1) & X86_CPUID_FEATURE_ECX_CX16) +# endif + { + RTTestISub("ASMAtomicCmpXchgU128"); + DO_SIMPLE_TEST_NO_SUB_NO_STACK(tstASMAtomicCmpXchgU128Worker, RTUINT128U); + } +#endif +} + + +#define TEST_RET_OLD(a_Type, a_Fmt, a_pVar, a_Function, a_uVal, a_VarExpect) do { \ + a_Type const uOldExpect = *(a_pVar); \ + a_Type uOldRet = a_Function(a_pVar, a_uVal); \ + if (RT_LIKELY( uOldRet == (uOldExpect) && *(a_pVar) == (a_VarExpect) )) { } \ + else RTTestFailed(g_hTest, "%s, %d: FAILURE: %s(%s," a_Fmt ") -> " a_Fmt ", expected " a_Fmt "; %s=" a_Fmt ", expected " a_Fmt "\n", \ + __FUNCTION__, __LINE__, #a_Function, #a_pVar, a_uVal, uOldRet, uOldExpect, #a_pVar, *(a_pVar), (a_VarExpect)); \ + } while (0) + + +DECLINLINE(void) tstASMAtomicAddU32Worker(uint32_t *pu32) +{ + *pu32 = 10; + TEST_RET_OLD(uint32_t, "%#x", pu32, ASMAtomicAddU32, 1, 11); + TEST_RET_OLD(uint32_t, "%#x", pu32, ASMAtomicAddU32, UINT32_C(0xfffffffe), 9); + TEST_RET_OLD(uint32_t, "%#x", pu32, ASMAtomicAddU32, UINT32_C(0xfffffff7), 0); + TEST_RET_OLD(uint32_t, "%#x", pu32, ASMAtomicAddU32, UINT32_C(0x7fffffff), UINT32_C(0x7fffffff)); + TEST_RET_OLD(uint32_t, "%#x", pu32, ASMAtomicAddU32, 1, UINT32_C(0x80000000)); + TEST_RET_OLD(uint32_t, "%#x", pu32, ASMAtomicAddU32, 1, UINT32_C(0x80000001)); + TEST_RET_OLD(uint32_t, "%#x", pu32, ASMAtomicAddU32, UINT32_C(0x7fffffff), 0); + TEST_RET_OLD(uint32_t, "%#x", pu32, ASMAtomicAddU32, 0, 0); + + TEST_RET_OLD(uint32_t, "%#x", pu32, ASMAtomicSubU32, 0, 0); + TEST_RET_OLD(uint32_t, "%#x", pu32, ASMAtomicSubU32, 32, UINT32_C(0xffffffe0)); + TEST_RET_OLD(uint32_t, "%#x", pu32, ASMAtomicSubU32, UINT32_C(0x7fffffff), UINT32_C(0x7fffffe1)); + TEST_RET_OLD(uint32_t, "%#x", pu32, ASMAtomicSubU32, UINT32_C(0x7fffffde), UINT32_C(0x00000003)); +} + + +DECLINLINE(void) tstASMAtomicAddS32Worker(int32_t *pi32) +{ + *pi32 = 10; + TEST_RET_OLD(int32_t, "%d", pi32, ASMAtomicAddS32, 1, 11); + TEST_RET_OLD(int32_t, "%d", pi32, ASMAtomicAddS32, -2, 9); + TEST_RET_OLD(int32_t, "%d", pi32, ASMAtomicAddS32, -9, 0); + TEST_RET_OLD(int32_t, "%d", pi32, ASMAtomicAddS32, -0x7fffffff, -0x7fffffff); + TEST_RET_OLD(int32_t, "%d", pi32, ASMAtomicAddS32, 0, -0x7fffffff); + TEST_RET_OLD(int32_t, "%d", pi32, ASMAtomicAddS32, 0x7fffffff, 0); + TEST_RET_OLD(int32_t, "%d", pi32, ASMAtomicAddS32, 0, 0); + + TEST_RET_OLD(int32_t, "%d", pi32, ASMAtomicSubS32, 0, 0); + TEST_RET_OLD(int32_t, "%d", pi32, ASMAtomicSubS32, 1, -1); + TEST_RET_OLD(int32_t, "%d", pi32, ASMAtomicSubS32, INT32_MIN, INT32_MAX); +} + + +DECLINLINE(void) tstASMAtomicAddU64Worker(uint64_t volatile *pu64) +{ + *pu64 = 10; + TEST_RET_OLD(uint64_t, "%llx", pu64, ASMAtomicAddU64, 1, 11); + TEST_RET_OLD(uint64_t, "%llx", pu64, ASMAtomicAddU64, UINT64_C(0xfffffffffffffffe), UINT64_C(0x0000000000000009)); + TEST_RET_OLD(uint64_t, "%llx", pu64, ASMAtomicAddU64, UINT64_C(0xfffffffffffffff7), UINT64_C(0x0000000000000000)); + TEST_RET_OLD(uint64_t, "%llx", pu64, ASMAtomicAddU64, UINT64_C(0x7ffffffffffffff0), UINT64_C(0x7ffffffffffffff0)); + TEST_RET_OLD(uint64_t, "%llx", pu64, ASMAtomicAddU64, UINT64_C(0x7ffffffffffffff0), UINT64_C(0xffffffffffffffe0)); + TEST_RET_OLD(uint64_t, "%llx", pu64, ASMAtomicAddU64, UINT64_C(0x0000000000000000), UINT64_C(0xffffffffffffffe0)); + TEST_RET_OLD(uint64_t, "%llx", pu64, ASMAtomicAddU64, UINT64_C(0x000000000000001f), UINT64_C(0xffffffffffffffff)); + TEST_RET_OLD(uint64_t, "%llx", pu64, ASMAtomicAddU64, UINT64_C(0x0000000000000001), UINT64_C(0x0000000000000000)); + + TEST_RET_OLD(uint64_t, "%llx", pu64, ASMAtomicSubU64, UINT64_C(0x0000000000000000), UINT64_C(0x0000000000000000)); + TEST_RET_OLD(uint64_t, "%llx", pu64, ASMAtomicSubU64, UINT64_C(0x0000000000000020), UINT64_C(0xffffffffffffffe0)); + TEST_RET_OLD(uint64_t, "%llx", pu64, ASMAtomicSubU64, UINT64_C(0x7fffffffffffffff), UINT64_C(0x7fffffffffffffe1)); + TEST_RET_OLD(uint64_t, "%llx", pu64, ASMAtomicSubU64, UINT64_C(0x7fffffffffffffdd), UINT64_C(0x0000000000000004)); +} + + +DECLINLINE(void) tstASMAtomicAddS64Worker(int64_t volatile *pi64) +{ + *pi64 = 10; + TEST_RET_OLD(int64_t, "%lld", pi64, ASMAtomicAddS64, 1, 11); + TEST_RET_OLD(int64_t, "%lld", pi64, ASMAtomicAddS64, -2, 9); + TEST_RET_OLD(int64_t, "%lld", pi64, ASMAtomicAddS64, -9, 0); + TEST_RET_OLD(int64_t, "%lld", pi64, ASMAtomicAddS64, -INT64_MAX, -INT64_MAX); + TEST_RET_OLD(int64_t, "%lld", pi64, ASMAtomicAddS64, 0, -INT64_MAX); + TEST_RET_OLD(int64_t, "%lld", pi64, ASMAtomicAddS64, -1, INT64_MIN); + TEST_RET_OLD(int64_t, "%lld", pi64, ASMAtomicAddS64, INT64_MAX, -1); + TEST_RET_OLD(int64_t, "%lld", pi64, ASMAtomicAddS64, 1, 0); + TEST_RET_OLD(int64_t, "%lld", pi64, ASMAtomicAddS64, 0, 0); + + TEST_RET_OLD(int64_t, "%d", pi64, ASMAtomicSubS64, 0, 0); + TEST_RET_OLD(int64_t, "%d", pi64, ASMAtomicSubS64, 1, -1); + TEST_RET_OLD(int64_t, "%d", pi64, ASMAtomicSubS64, INT64_MIN, INT64_MAX); +} + + + +DECLINLINE(void) tstASMAtomicAddZWorker(size_t volatile *pcb) +{ + *pcb = 10; + TEST_RET_OLD(size_t, "%zx", pcb, ASMAtomicAddZ, 1, 11); + TEST_RET_OLD(size_t, "%zx", pcb, ASMAtomicAddZ, ~(size_t)1, 9); + TEST_RET_OLD(size_t, "%zx", pcb, ASMAtomicAddZ, ~(size_t)8, 0); + + TEST_RET_OLD(size_t, "%zx", pcb, ASMAtomicSubZ, 0, 0); + TEST_RET_OLD(size_t, "%zx", pcb, ASMAtomicSubZ, 10, ~(size_t)9); +} + +static void tstASMAtomicAdd(void) +{ + DO_SIMPLE_TEST(ASMAtomicAddU32, uint32_t); + DO_SIMPLE_TEST(ASMAtomicAddS32, int32_t); + DO_SIMPLE_TEST(ASMAtomicAddU64, uint64_t); + DO_SIMPLE_TEST(ASMAtomicAddS64, int64_t); + DO_SIMPLE_TEST(ASMAtomicAddZ, size_t); +} + + +#define TEST_RET_NEW_NV(a_Type, a_Fmt, a_pVar, a_Function, a_VarExpect) do { \ + a_Type uNewRet = a_Function(a_pVar); \ + if (RT_LIKELY( uNewRet == (a_VarExpect) && *(a_pVar) == (a_VarExpect) )) { } \ + else RTTestFailed(g_hTest, "%s, %d: FAILURE: %s(%s) -> " a_Fmt " and %s=" a_Fmt ", expected both " a_Fmt "\n", \ + __FUNCTION__, __LINE__, #a_Function, #a_pVar, uNewRet, #a_pVar, *(a_pVar), (a_VarExpect)); \ + } while (0) + + +DECLINLINE(void) tstASMAtomicDecIncU32Worker(uint32_t volatile *pu32) +{ + *pu32 = 3; + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicDecU32, 2); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicDecU32, 1); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicDecU32, 0); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicDecU32, UINT32_MAX); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicDecU32, UINT32_MAX - 1); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicDecU32, UINT32_MAX - 2); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicIncU32, UINT32_MAX - 1); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicIncU32, UINT32_MAX); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicIncU32, 0); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicIncU32, 1); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicIncU32, 2); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicDecU32, 1); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicIncU32, 2); + *pu32 = _1M; + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicDecU32, _1M - 1); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicIncU32, _1M); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicIncU32, _1M + 1); +} + +DECLINLINE(void) tstASMAtomicUoDecIncU32Worker(uint32_t volatile *pu32) +{ + *pu32 = 3; + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicUoDecU32, 2); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicUoDecU32, 1); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicUoDecU32, 0); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicUoDecU32, UINT32_MAX); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicUoDecU32, UINT32_MAX - 1); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicUoDecU32, UINT32_MAX - 2); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicUoIncU32, UINT32_MAX - 1); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicUoIncU32, UINT32_MAX); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicUoIncU32, 0); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicUoIncU32, 1); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicUoIncU32, 2); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicUoDecU32, 1); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicUoIncU32, 2); + *pu32 = _1M; + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicUoDecU32, _1M - 1); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicUoIncU32, _1M); + TEST_RET_NEW_NV(uint32_t, "%#x", pu32, ASMAtomicUoIncU32, _1M + 1); +} + + +DECLINLINE(void) tstASMAtomicDecIncS32Worker(int32_t volatile *pi32) +{ + *pi32 = 10; + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicDecS32, 9); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicDecS32, 8); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicDecS32, 7); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicDecS32, 6); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicDecS32, 5); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicDecS32, 4); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicDecS32, 3); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicDecS32, 2); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicDecS32, 1); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicDecS32, 0); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicDecS32, -1); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicDecS32, -2); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicIncS32, -1); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicIncS32, 0); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicIncS32, 1); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicIncS32, 2); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicIncS32, 3); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicDecS32, 2); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicIncS32, 3); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicDecS32, 2); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicIncS32, 3); + *pi32 = INT32_MAX; + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicDecS32, INT32_MAX - 1); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicIncS32, INT32_MAX); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicIncS32, INT32_MIN); +} + + +#if 0 +DECLINLINE(void) tstASMAtomicUoDecIncS32Worker(int32_t volatile *pi32) +{ + *pi32 = 10; + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicUoDecS32, 9); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicUoDecS32, 8); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicUoDecS32, 7); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicUoDecS32, 6); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicUoDecS32, 5); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicUoDecS32, 4); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicUoDecS32, 3); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicUoDecS32, 2); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicUoDecS32, 1); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicUoDecS32, 0); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicUoDecS32, -1); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicUoDecS32, -2); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicUoIncS32, -1); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicUoIncS32, 0); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicUoIncS32, 1); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicUoIncS32, 2); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicUoIncS32, 3); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicUoDecS32, 2); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicUoIncS32, 3); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicUoDecS32, 2); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicUoIncS32, 3); + *pi32 = INT32_MAX; + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicUoDecS32, INT32_MAX - 1); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicUoIncS32, INT32_MAX); + TEST_RET_NEW_NV(int32_t, "%d", pi32, ASMAtomicUoIncS32, INT32_MIN); +} +#endif + + +DECLINLINE(void) tstASMAtomicDecIncU64Worker(uint64_t volatile *pu64) +{ + *pu64 = 3; + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicDecU64, 2); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicDecU64, 1); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicDecU64, 0); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicDecU64, UINT64_MAX); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicDecU64, UINT64_MAX - 1); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicDecU64, UINT64_MAX - 2); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicIncU64, UINT64_MAX - 1); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicIncU64, UINT64_MAX); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicIncU64, 0); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicIncU64, 1); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicIncU64, 2); + *pu64 = _4G - 1; + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicDecU64, _4G - 2); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicIncU64, _4G - 1); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicIncU64, _4G); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicIncU64, _4G + 1); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicDecU64, _4G); +} + + +#if 0 +DECLINLINE(void) tstASMAtomicUoDecIncU64Worker(uint64_t volatile *pu64) +{ + *pu64 = 3; + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicUoDecU64, 2); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicUoDecU64, 1); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicUoDecU64, 0); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicUoDecU64, UINT64_MAX); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicUoDecU64, UINT64_MAX - 1); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicUoDecU64, UINT64_MAX - 2); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicUoIncU64, UINT64_MAX - 1); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicUoIncU64, UINT64_MAX); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicUoIncU64, 0); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicUoIncU64, 1); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicUoIncU64, 2); + *pu64 = _4G - 1; + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicUoDecU64, _4G - 2); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicUoIncU64, _4G - 1); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicUoIncU64, _4G); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicUoIncU64, _4G + 1); + TEST_RET_NEW_NV(uint64_t, "%lld", pu64, ASMAtomicUoDecU64, _4G); +} +#endif + + +DECLINLINE(void) tstASMAtomicDecIncS64Worker(int64_t volatile *pi64) +{ + *pi64 = 10; + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicDecS64, 9); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicDecS64, 8); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicDecS64, 7); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicDecS64, 6); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicDecS64, 5); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicDecS64, 4); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicDecS64, 3); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicDecS64, 2); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicDecS64, 1); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicDecS64, 0); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicDecS64, -1); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicDecS64, -2); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicIncS64, -1); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicIncS64, 0); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicIncS64, 1); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicIncS64, 2); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicIncS64, 3); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicDecS64, 2); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicIncS64, 3); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicDecS64, 2); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicIncS64, 3); + *pi64 = INT64_MAX; + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicDecS64, INT64_MAX - 1); +} + + +#if 0 +DECLINLINE(void) tstASMAtomicUoDecIncS64Worker(int64_t volatile *pi64) +{ + *pi64 = 10; + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicUoDecS64, 9); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicUoDecS64, 8); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicUoDecS64, 7); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicUoDecS64, 6); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicUoDecS64, 5); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicUoDecS64, 4); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicUoDecS64, 3); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicUoDecS64, 2); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicUoDecS64, 1); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicUoDecS64, 0); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicUoDecS64, -1); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicUoDecS64, -2); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicUoIncS64, -1); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicUoIncS64, 0); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicUoIncS64, 1); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicUoIncS64, 2); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicUoIncS64, 3); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicUoDecS64, 2); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicUoIncS64, 3); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicUoDecS64, 2); + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicUoIncS64, 3); + *pi64 = INT64_MAX; + TEST_RET_NEW_NV(int64_t, "%lld", pi64, ASMAtomicUoDecS64, INT64_MAX - 1); +} +#endif + + +DECLINLINE(void) tstASMAtomicDecIncZWorker(size_t volatile *pcb) +{ + size_t const uBaseVal = ~(size_t)0 >> 7; + *pcb = uBaseVal; + TEST_RET_NEW_NV(size_t, "%zx", pcb, ASMAtomicDecZ, uBaseVal - 1); + TEST_RET_NEW_NV(size_t, "%zx", pcb, ASMAtomicDecZ, uBaseVal - 2); + TEST_RET_NEW_NV(size_t, "%zx", pcb, ASMAtomicDecZ, uBaseVal - 3); + TEST_RET_NEW_NV(size_t, "%zx", pcb, ASMAtomicIncZ, uBaseVal - 2); + TEST_RET_NEW_NV(size_t, "%zx", pcb, ASMAtomicIncZ, uBaseVal - 1); + TEST_RET_NEW_NV(size_t, "%zx", pcb, ASMAtomicIncZ, uBaseVal); + TEST_RET_NEW_NV(size_t, "%zx", pcb, ASMAtomicIncZ, uBaseVal + 1); + TEST_RET_NEW_NV(size_t, "%zx", pcb, ASMAtomicDecZ, uBaseVal); + TEST_RET_NEW_NV(size_t, "%zx", pcb, ASMAtomicDecZ, uBaseVal - 1); + TEST_RET_NEW_NV(size_t, "%zx", pcb, ASMAtomicIncZ, uBaseVal); +} + + +static void tstASMAtomicDecInc(void) +{ + DO_SIMPLE_TEST(ASMAtomicDecIncU32, uint32_t); + DO_SIMPLE_TEST(ASMAtomicUoDecIncU32, uint32_t); + DO_SIMPLE_TEST(ASMAtomicDecIncS32, int32_t); + //DO_SIMPLE_TEST(ASMAtomicUoDecIncS32, int32_t); + DO_SIMPLE_TEST(ASMAtomicDecIncU64, uint64_t); + //DO_SIMPLE_TEST(ASMAtomicUoDecIncU64, uint64_t); + DO_SIMPLE_TEST(ASMAtomicDecIncS64, int64_t); + //DO_SIMPLE_TEST(ASMAtomicUoDecIncS64, int64_t); + DO_SIMPLE_TEST(ASMAtomicDecIncZ, size_t); +} + + +#define TEST_RET_VOID(a_Type, a_Fmt, a_pVar, a_Function, a_uVal, a_VarExpect) do { \ + a_Function(a_pVar, a_uVal); \ + if (RT_LIKELY( *(a_pVar) == (a_VarExpect) )) { } \ + else RTTestFailed(g_hTest, "%s, %d: FAILURE: %s(%s, " a_Fmt ") -> %s=" a_Fmt ", expected " a_Fmt "\n", \ + __FUNCTION__, __LINE__, #a_Function, #a_pVar, a_uVal, #a_pVar, *(a_pVar), (a_VarExpect)); \ + } while (0) + +#define TEST_RET_NEW(a_Type, a_Fmt, a_pVar, a_Function, a_uVal, a_VarExpect) do { \ + a_Type uNewRet = a_Function(a_pVar, a_uVal); \ + if (RT_LIKELY( uNewRet == (a_VarExpect) && *(a_pVar) == (a_VarExpect) )) { } \ + else RTTestFailed(g_hTest, "%s, %d: FAILURE: %s(%s, " a_Fmt ") -> " a_Fmt " and %s=" a_Fmt ", expected both " a_Fmt "\n", \ + __FUNCTION__, __LINE__, #a_Function, #a_pVar, a_uVal, uNewRet, #a_pVar, *(a_pVar), (a_VarExpect)); \ + } while (0) + + +DECLINLINE(void) tstASMAtomicAndOrXorU32Worker(uint32_t volatile *pu32) +{ + *pu32 = UINT32_C(0xffffffff); + TEST_RET_VOID(uint32_t, "%#x", pu32, ASMAtomicOrU32, UINT32_C(0xffffffff), UINT32_C(0xffffffff)); + TEST_RET_VOID(uint32_t, "%#x", pu32, ASMAtomicAndU32, UINT32_C(0xffffffff), UINT32_C(0xffffffff)); + TEST_RET_VOID(uint32_t, "%#x", pu32, ASMAtomicAndU32, UINT32_C(0x8f8f8f8f), UINT32_C(0x8f8f8f8f)); + TEST_RET_VOID(uint32_t, "%#x", pu32, ASMAtomicOrU32, UINT32_C(0x70707070), UINT32_C(0xffffffff)); + TEST_RET_VOID(uint32_t, "%#x", pu32, ASMAtomicAndU32, UINT32_C(1), UINT32_C(1)); + TEST_RET_VOID(uint32_t, "%#x", pu32, ASMAtomicOrU32, UINT32_C(0x80000000), UINT32_C(0x80000001)); + TEST_RET_VOID(uint32_t, "%#x", pu32, ASMAtomicAndU32, UINT32_C(0x80000000), UINT32_C(0x80000000)); + TEST_RET_VOID(uint32_t, "%#x", pu32, ASMAtomicAndU32, UINT32_C(0), UINT32_C(0)); + TEST_RET_VOID(uint32_t, "%#x", pu32, ASMAtomicOrU32, UINT32_C(0x42424242), UINT32_C(0x42424242)); + TEST_RET_VOID(uint32_t, "%#x", pu32, ASMAtomicAndU32, UINT32_C(0x00ff0f00), UINT32_C(0x00420200)); + TEST_RET_VOID(uint32_t, "%#x", pu32, ASMAtomicXorU32, UINT32_C(0x42004042), UINT32_C(0x42424242)); + TEST_RET_VOID(uint32_t, "%#x", pu32, ASMAtomicXorU32, UINT32_C(0xff024200), UINT32_C(0xbd400042)); + TEST_RET_VOID(uint32_t, "%#x", pu32, ASMAtomicXorU32, UINT32_C(0x00000000), UINT32_C(0xbd400042)); +} + + +DECLINLINE(void) tstASMAtomicUoAndOrXorU32Worker(uint32_t volatile *pu32) +{ + *pu32 = UINT32_C(0xffffffff); + TEST_RET_VOID(uint32_t, "%#x", pu32, ASMAtomicUoOrU32, UINT32_C(0xffffffff), UINT32_C(0xffffffff)); + TEST_RET_VOID(uint32_t, "%#x", pu32, ASMAtomicUoAndU32, UINT32_C(0xffffffff), UINT32_C(0xffffffff)); + TEST_RET_VOID(uint32_t, "%#x", pu32, ASMAtomicUoAndU32, UINT32_C(0x8f8f8f8f), UINT32_C(0x8f8f8f8f)); + TEST_RET_VOID(uint32_t, "%#x", pu32, ASMAtomicUoOrU32, UINT32_C(0x70707070), UINT32_C(0xffffffff)); + TEST_RET_VOID(uint32_t, "%#x", pu32, ASMAtomicUoAndU32, UINT32_C(1), UINT32_C(1)); + TEST_RET_VOID(uint32_t, "%#x", pu32, ASMAtomicUoOrU32, UINT32_C(0x80000000), UINT32_C(0x80000001)); + TEST_RET_VOID(uint32_t, "%#x", pu32, ASMAtomicUoAndU32, UINT32_C(0x80000000), UINT32_C(0x80000000)); + TEST_RET_VOID(uint32_t, "%#x", pu32, ASMAtomicUoAndU32, UINT32_C(0), UINT32_C(0)); + TEST_RET_VOID(uint32_t, "%#x", pu32, ASMAtomicUoOrU32, UINT32_C(0x42424242), UINT32_C(0x42424242)); + TEST_RET_VOID(uint32_t, "%#x", pu32, ASMAtomicUoAndU32, UINT32_C(0x00ff0f00), UINT32_C(0x00420200)); + TEST_RET_VOID(uint32_t, "%#x", pu32, ASMAtomicUoXorU32, UINT32_C(0x42004042), UINT32_C(0x42424242)); + TEST_RET_VOID(uint32_t, "%#x", pu32, ASMAtomicUoXorU32, UINT32_C(0xff024200), UINT32_C(0xbd400042)); + TEST_RET_VOID(uint32_t, "%#x", pu32, ASMAtomicUoXorU32, UINT32_C(0x00000000), UINT32_C(0xbd400042)); +} + + +DECLINLINE(void) tstASMAtomicAndOrXorExU32Worker(uint32_t volatile *pu32) +{ + *pu32 = UINT32_C(0xffffffff); + TEST_RET_OLD(uint32_t, "%#x", pu32, ASMAtomicOrExU32, UINT32_C(0xffffffff), UINT32_C(0xffffffff)); + TEST_RET_OLD(uint32_t, "%#x", pu32, ASMAtomicAndExU32, UINT32_C(0xffffffff), UINT32_C(0xffffffff)); + TEST_RET_OLD(uint32_t, "%#x", pu32, ASMAtomicAndExU32, UINT32_C(0x8f8f8f8f), UINT32_C(0x8f8f8f8f)); + TEST_RET_OLD(uint32_t, "%#x", pu32, ASMAtomicOrExU32, UINT32_C(0x70707070), UINT32_C(0xffffffff)); + TEST_RET_OLD(uint32_t, "%#x", pu32, ASMAtomicAndExU32, UINT32_C(1), UINT32_C(1)); + TEST_RET_OLD(uint32_t, "%#x", pu32, ASMAtomicOrExU32, UINT32_C(0x80000000), UINT32_C(0x80000001)); + TEST_RET_OLD(uint32_t, "%#x", pu32, ASMAtomicAndExU32, UINT32_C(0x80000000), UINT32_C(0x80000000)); + TEST_RET_OLD(uint32_t, "%#x", pu32, ASMAtomicAndExU32, UINT32_C(0), UINT32_C(0)); + TEST_RET_OLD(uint32_t, "%#x", pu32, ASMAtomicOrExU32, UINT32_C(0x42424242), UINT32_C(0x42424242)); + TEST_RET_OLD(uint32_t, "%#x", pu32, ASMAtomicAndExU32, UINT32_C(0x00ff0f00), UINT32_C(0x00420200)); + TEST_RET_OLD(uint32_t, "%#x", pu32, ASMAtomicXorExU32, UINT32_C(0x42004042), UINT32_C(0x42424242)); + TEST_RET_OLD(uint32_t, "%#x", pu32, ASMAtomicXorExU32, UINT32_C(0xff024200), UINT32_C(0xbd400042)); + TEST_RET_OLD(uint32_t, "%#x", pu32, ASMAtomicXorExU32, UINT32_C(0x00000000), UINT32_C(0xbd400042)); +} + + +DECLINLINE(void) tstASMAtomicAndOrXorU64Worker(uint64_t volatile *pu64) +{ + *pu64 = UINT64_C(0xffffffff); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicOrU64, UINT64_C(0xffffffff), UINT64_C(0xffffffff)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicAndU64, UINT64_C(0xffffffff), UINT64_C(0xffffffff)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicAndU64, UINT64_C(0x8f8f8f8f), UINT64_C(0x8f8f8f8f)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicOrU64, UINT64_C(0x70707070), UINT64_C(0xffffffff)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicAndU64, UINT64_C(1), UINT64_C(1)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicOrU64, UINT64_C(0x80000000), UINT64_C(0x80000001)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicAndU64, UINT64_C(0x80000000), UINT64_C(0x80000000)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicAndU64, UINT64_C(0), UINT64_C(0)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicOrU64, UINT64_C(0x42424242), UINT64_C(0x42424242)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicAndU64, UINT64_C(0x00ff0f00), UINT64_C(0x00420200)); + //TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicXorU64, UINT64_C(0x42004042), UINT64_C(0x42424242)); + //TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicXorU64, UINT64_C(0xff024200), UINT64_C(0xbd400042)); + //TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicXorU64, UINT64_C(0x00000000), UINT64_C(0xbd400042)); + + /* full 64-bit */ + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicAndU64, UINT64_C(0x0000000000000000), UINT64_C(0x0000000000000000)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicOrU64, UINT64_C(0xffffffffffffffff), UINT64_C(0xffffffffffffffff)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicAndU64, UINT64_C(0xffffffffffffffff), UINT64_C(0xffffffffffffffff)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicAndU64, UINT64_C(0x8f8f8f8f8f8f8f8f), UINT64_C(0x8f8f8f8f8f8f8f8f)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicOrU64, UINT64_C(0x7070707070707070), UINT64_C(0xffffffffffffffff)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicAndU64, UINT64_C(0x0000000000000001), UINT64_C(0x0000000000000001)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicOrU64, UINT64_C(0x8000000000000000), UINT64_C(0x8000000000000001)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicAndU64, UINT64_C(0x8000000000000000), UINT64_C(0x8000000000000000)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicAndU64, UINT64_C(0), UINT64_C(0)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicOrU64, UINT64_C(0x4242424242424242), UINT64_C(0x4242424242424242)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicAndU64, UINT64_C(0x00ff0f00ff0f0000), UINT64_C(0x0042020042020000)); + //TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicXorU64, UINT64_C(0x4200404242040000), UINT64_C(0x4242424242420000)); + //TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicXorU64, UINT64_C(0xff02420000ff2127), UINT64_C(0xbd40004242bd2127)); + //TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicXorU64, UINT64_C(0x0000000000000000), UINT64_C(0xbd40004242bd2127)); +} + + +DECLINLINE(void) tstASMAtomicUoAndOrXorU64Worker(uint64_t volatile *pu64) +{ + *pu64 = UINT64_C(0xffffffff); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicUoOrU64, UINT64_C(0xffffffff), UINT64_C(0xffffffff)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicUoAndU64, UINT64_C(0xffffffff), UINT64_C(0xffffffff)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicUoAndU64, UINT64_C(0x8f8f8f8f), UINT64_C(0x8f8f8f8f)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicUoOrU64, UINT64_C(0x70707070), UINT64_C(0xffffffff)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicUoAndU64, UINT64_C(1), UINT64_C(1)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicUoOrU64, UINT64_C(0x80000000), UINT64_C(0x80000001)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicUoAndU64, UINT64_C(0x80000000), UINT64_C(0x80000000)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicUoAndU64, UINT64_C(0), UINT64_C(0)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicUoOrU64, UINT64_C(0x42424242), UINT64_C(0x42424242)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicUoAndU64, UINT64_C(0x00ff0f00), UINT64_C(0x00420200)); + //TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicUoXorU64, UINT64_C(0x42004042), UINT64_C(0x42424242)); + //TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicUoXorU64, UINT64_C(0xff024200), UINT64_C(0xbd400042)); + //TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicUoXorU64, UINT64_C(0x00000000), UINT64_C(0xbd400042)); + + /* full 64-bit */ + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicUoAndU64, UINT64_C(0x0000000000000000), UINT64_C(0x0000000000000000)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicUoOrU64, UINT64_C(0xffffffffffffffff), UINT64_C(0xffffffffffffffff)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicUoAndU64, UINT64_C(0xffffffffffffffff), UINT64_C(0xffffffffffffffff)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicUoAndU64, UINT64_C(0x8f8f8f8f8f8f8f8f), UINT64_C(0x8f8f8f8f8f8f8f8f)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicUoOrU64, UINT64_C(0x7070707070707070), UINT64_C(0xffffffffffffffff)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicUoAndU64, UINT64_C(0x0000000000000001), UINT64_C(0x0000000000000001)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicUoOrU64, UINT64_C(0x8000000000000000), UINT64_C(0x8000000000000001)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicUoAndU64, UINT64_C(0x8000000000000000), UINT64_C(0x8000000000000000)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicUoAndU64, UINT64_C(0), UINT64_C(0)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicUoOrU64, UINT64_C(0x4242424242424242), UINT64_C(0x4242424242424242)); + TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicUoAndU64, UINT64_C(0x00ff0f00ff0f0000), UINT64_C(0x0042020042020000)); + //TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicUoXorU64, UINT64_C(0x4200404242040000), UINT64_C(0x4242424242420000)); + //TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicUoXorU64, UINT64_C(0xff02420000ff2127), UINT64_C(0xbd40004242bd2127)); + //TEST_RET_VOID(uint64_t, "%#llx", pu64, ASMAtomicUoXorU64, UINT64_C(0x0000000000000000), UINT64_C(0xbd40004242bd2127)); +} + + +#if 0 +DECLINLINE(void) tstASMAtomicAndOrXorExU64Worker(uint64_t volatile *pu64) +{ + *pu64 = UINT64_C(0xffffffff); + TEST_RET_OLD(uint64_t, "%#llx", pu64, ASMAtomicOrExU64, UINT64_C(0xffffffff), UINT64_C(0xffffffff)); + TEST_RET_OLD(uint64_t, "%#llx", pu64, ASMAtomicAndExU64, UINT64_C(0xffffffff), UINT64_C(0xffffffff)); + TEST_RET_OLD(uint64_t, "%#llx", pu64, ASMAtomicAndExU64, UINT64_C(0x8f8f8f8f), UINT64_C(0x8f8f8f8f)); + TEST_RET_OLD(uint64_t, "%#llx", pu64, ASMAtomicOrExU64, UINT64_C(0x70707070), UINT64_C(0xffffffff)); + TEST_RET_OLD(uint64_t, "%#llx", pu64, ASMAtomicAndExU64, UINT64_C(1), UINT64_C(1)); + TEST_RET_OLD(uint64_t, "%#llx", pu64, ASMAtomicOrExU64, UINT64_C(0x80000000), UINT64_C(0x80000001)); + TEST_RET_OLD(uint64_t, "%#llx", pu64, ASMAtomicAndExU64, UINT64_C(0x80000000), UINT64_C(0x80000000)); + TEST_RET_OLD(uint64_t, "%#llx", pu64, ASMAtomicAndExU64, UINT64_C(0), UINT64_C(0)); + TEST_RET_OLD(uint64_t, "%#llx", pu64, ASMAtomicOrExU64, UINT64_C(0x42424242), UINT64_C(0x42424242)); + TEST_RET_OLD(uint64_t, "%#llx", pu64, ASMAtomicAndExU64, UINT64_C(0x00ff0f00), UINT64_C(0x00420200)); + //TEST_RET_OLD(uint64_t, "%#llx", pu64, ASMAtomicXorExU64, UINT64_C(0x42004042), UINT64_C(0x42424242)); + //TEST_RET_OLD(uint64_t, "%#llx", pu64, ASMAtomicXorExU64, UINT64_C(0xff024200), UINT64_C(0xbd400042)); + //TEST_RET_OLD(uint64_t, "%#llx", pu64, ASMAtomicXorExU64, UINT64_C(0x00000000), UINT64_C(0xbd400042)); + + /* full 64-bit */ + TEST_RET_OLD(uint64_t, "%#llx", pu64, ASMAtomicAndExU64, UINT64_C(0x0000000000000000), UINT64_C(0x0000000000000000)); + TEST_RET_OLD(uint64_t, "%#llx", pu64, ASMAtomicOrExU64, UINT64_C(0xffffffffffffffff), UINT64_C(0xffffffffffffffff)); + TEST_RET_OLD(uint64_t, "%#llx", pu64, ASMAtomicAndExU64, UINT64_C(0xffffffffffffffff), UINT64_C(0xffffffffffffffff)); + TEST_RET_OLD(uint64_t, "%#llx", pu64, ASMAtomicAndExU64, UINT64_C(0x8f8f8f8f8f8f8f8f), UINT64_C(0x8f8f8f8f8f8f8f8f)); + TEST_RET_OLD(uint64_t, "%#llx", pu64, ASMAtomicOrExU64, UINT64_C(0x7070707070707070), UINT64_C(0xffffffffffffffff)); + TEST_RET_OLD(uint64_t, "%#llx", pu64, ASMAtomicAndExU64, UINT64_C(0x0000000000000001), UINT64_C(0x0000000000000001)); + TEST_RET_OLD(uint64_t, "%#llx", pu64, ASMAtomicOrExU64, UINT64_C(0x8000000000000000), UINT64_C(0x8000000000000001)); + TEST_RET_OLD(uint64_t, "%#llx", pu64, ASMAtomicAndExU64, UINT64_C(0x8000000000000000), UINT64_C(0x8000000000000000)); + TEST_RET_OLD(uint64_t, "%#llx", pu64, ASMAtomicAndExU64, UINT64_C(0), UINT64_C(0)); + TEST_RET_OLD(uint64_t, "%#llx", pu64, ASMAtomicOrExU64, UINT64_C(0x4242424242424242), UINT64_C(0x4242424242424242)); + TEST_RET_OLD(uint64_t, "%#llx", pu64, ASMAtomicAndExU64, UINT64_C(0x00ff0f00ff0f0000), UINT64_C(0x0042020042020000)); + //TEST_RET_OLD(uint64_t, "%#llx", pu64, ASMAtomicXorExU64, UINT64_C(0x4200404242040000), UINT64_C(0x4242424242420000)); + //TEST_RET_OLD(uint64_t, "%#llx", pu64, ASMAtomicXorExU64, UINT64_C(0xff02420000ff2127), UINT64_C(0xbd40004242bd2127)); + //TEST_RET_OLD(uint64_t, "%#llx", pu64, ASMAtomicXorExU64, UINT64_C(0x0000000000000000), UINT64_C(0xbd40004242bd2127)); +} +#endif + + +static void tstASMAtomicAndOrXor(void) +{ + DO_SIMPLE_TEST(ASMAtomicAndOrXorU32, uint32_t); + DO_SIMPLE_TEST(ASMAtomicUoAndOrXorU32, uint32_t); + DO_SIMPLE_TEST(ASMAtomicAndOrXorExU32, uint32_t); + DO_SIMPLE_TEST(ASMAtomicAndOrXorU64, uint64_t); + DO_SIMPLE_TEST(ASMAtomicUoAndOrXorU64, uint64_t); + //DO_SIMPLE_TEST(ASMAtomicAndOrXorExU64, uint64_t); +} + + +typedef struct +{ + uint8_t ab[PAGE_SIZE]; +} TSTPAGE; + + +DECLINLINE(void) tstASMMemZeroPageWorker(TSTPAGE *pPage) +{ + for (unsigned j = 0; j < 16; j++) + { + memset(pPage, 0x11 * j, sizeof(*pPage)); + ASMMemZeroPage(pPage); + for (unsigned i = 0; i < sizeof(pPage->ab); i++) + if (pPage->ab[i]) + RTTestFailed(g_hTest, "ASMMemZeroPage didn't clear byte at offset %#x!\n", i); + if (ASMMemIsZeroPage(pPage) != true) + RTTestFailed(g_hTest, "ASMMemIsZeroPage returns false after ASMMemZeroPage!\n"); + if (ASMMemFirstMismatchingU32(pPage, sizeof(pPage), 0) != NULL) + RTTestFailed(g_hTest, "ASMMemFirstMismatchingU32(,,0) returns non-NULL after ASMMemZeroPage!\n"); + } +} + + +static void tstASMMemZeroPage(void) +{ + RTTestISub("ASMMemZeroPage"); + DO_SIMPLE_TEST_NO_SUB_NO_STACK(tstASMMemZeroPageWorker, TSTPAGE); +} + + +void tstASMMemIsZeroPage(RTTEST hTest) +{ + RTTestSub(hTest, "ASMMemIsZeroPage"); + + void *pvPage1 = RTTestGuardedAllocHead(hTest, PAGE_SIZE); + void *pvPage2 = RTTestGuardedAllocTail(hTest, PAGE_SIZE); + RTTESTI_CHECK_RETV(pvPage1 && pvPage2); + + memset(pvPage1, 0, PAGE_SIZE); + memset(pvPage2, 0, PAGE_SIZE); + RTTESTI_CHECK(ASMMemIsZeroPage(pvPage1)); + RTTESTI_CHECK(ASMMemIsZeroPage(pvPage2)); + + memset(pvPage1, 0xff, PAGE_SIZE); + memset(pvPage2, 0xff, PAGE_SIZE); + RTTESTI_CHECK(!ASMMemIsZeroPage(pvPage1)); + RTTESTI_CHECK(!ASMMemIsZeroPage(pvPage2)); + + memset(pvPage1, 0, PAGE_SIZE); + memset(pvPage2, 0, PAGE_SIZE); + for (unsigned off = 0; off < PAGE_SIZE; off++) + { + ((uint8_t *)pvPage1)[off] = 1; + RTTESTI_CHECK(!ASMMemIsZeroPage(pvPage1)); + ((uint8_t *)pvPage1)[off] = 0; + + ((uint8_t *)pvPage2)[off] = 0x80; + RTTESTI_CHECK(!ASMMemIsZeroPage(pvPage2)); + ((uint8_t *)pvPage2)[off] = 0; + } + + RTTestSubDone(hTest); +} + + +void tstASMMemFirstMismatchingU8(RTTEST hTest) +{ + RTTestSub(hTest, "ASMMemFirstMismatchingU8"); + + uint8_t *pbPage1 = (uint8_t *)RTTestGuardedAllocHead(hTest, PAGE_SIZE); + uint8_t *pbPage2 = (uint8_t *)RTTestGuardedAllocTail(hTest, PAGE_SIZE); + RTTESTI_CHECK_RETV(pbPage1 && pbPage2); + + memset(pbPage1, 0, PAGE_SIZE); + memset(pbPage2, 0, PAGE_SIZE); + RTTESTI_CHECK(ASMMemFirstMismatchingU8(pbPage1, PAGE_SIZE, 0) == NULL); + RTTESTI_CHECK(ASMMemFirstMismatchingU8(pbPage2, PAGE_SIZE, 0) == NULL); + RTTESTI_CHECK(ASMMemFirstMismatchingU8(pbPage1, PAGE_SIZE, 1) == pbPage1); + RTTESTI_CHECK(ASMMemFirstMismatchingU8(pbPage2, PAGE_SIZE, 1) == pbPage2); + RTTESTI_CHECK(ASMMemFirstMismatchingU8(pbPage1, PAGE_SIZE, 0x87) == pbPage1); + RTTESTI_CHECK(ASMMemFirstMismatchingU8(pbPage2, PAGE_SIZE, 0x87) == pbPage2); + RTTESTI_CHECK(ASMMemIsZero(pbPage1, PAGE_SIZE)); + RTTESTI_CHECK(ASMMemIsZero(pbPage2, PAGE_SIZE)); + RTTESTI_CHECK(ASMMemIsAllU8(pbPage1, PAGE_SIZE, 0)); + RTTESTI_CHECK(ASMMemIsAllU8(pbPage2, PAGE_SIZE, 0)); + RTTESTI_CHECK(!ASMMemIsAllU8(pbPage1, PAGE_SIZE, 0x34)); + RTTESTI_CHECK(!ASMMemIsAllU8(pbPage2, PAGE_SIZE, 0x88)); + unsigned cbSub = 32; + while (cbSub-- > 0) + { + RTTESTI_CHECK(ASMMemFirstMismatchingU8(&pbPage1[PAGE_SIZE - cbSub], cbSub, 0) == NULL); + RTTESTI_CHECK(ASMMemFirstMismatchingU8(&pbPage2[PAGE_SIZE - cbSub], cbSub, 0) == NULL); + RTTESTI_CHECK(ASMMemFirstMismatchingU8(pbPage1, cbSub, 0) == NULL); + RTTESTI_CHECK(ASMMemFirstMismatchingU8(pbPage2, cbSub, 0) == NULL); + + RTTESTI_CHECK(ASMMemFirstMismatchingU8(&pbPage1[PAGE_SIZE - cbSub], cbSub, 0x34) == &pbPage1[PAGE_SIZE - cbSub] || !cbSub); + RTTESTI_CHECK(ASMMemFirstMismatchingU8(&pbPage2[PAGE_SIZE - cbSub], cbSub, 0x99) == &pbPage2[PAGE_SIZE - cbSub] || !cbSub); + RTTESTI_CHECK(ASMMemFirstMismatchingU8(pbPage1, cbSub, 0x42) == pbPage1 || !cbSub); + RTTESTI_CHECK(ASMMemFirstMismatchingU8(pbPage2, cbSub, 0x88) == pbPage2 || !cbSub); + } + + memset(pbPage1, 0xff, PAGE_SIZE); + memset(pbPage2, 0xff, PAGE_SIZE); + RTTESTI_CHECK(ASMMemFirstMismatchingU8(pbPage1, PAGE_SIZE, 0xff) == NULL); + RTTESTI_CHECK(ASMMemFirstMismatchingU8(pbPage2, PAGE_SIZE, 0xff) == NULL); + RTTESTI_CHECK(ASMMemFirstMismatchingU8(pbPage1, PAGE_SIZE, 0xfe) == pbPage1); + RTTESTI_CHECK(ASMMemFirstMismatchingU8(pbPage2, PAGE_SIZE, 0xfe) == pbPage2); + RTTESTI_CHECK(!ASMMemIsZero(pbPage1, PAGE_SIZE)); + RTTESTI_CHECK(!ASMMemIsZero(pbPage2, PAGE_SIZE)); + RTTESTI_CHECK(ASMMemIsAllU8(pbPage1, PAGE_SIZE, 0xff)); + RTTESTI_CHECK(ASMMemIsAllU8(pbPage2, PAGE_SIZE, 0xff)); + RTTESTI_CHECK(!ASMMemIsAllU8(pbPage1, PAGE_SIZE, 0)); + RTTESTI_CHECK(!ASMMemIsAllU8(pbPage2, PAGE_SIZE, 0)); + cbSub = 32; + while (cbSub-- > 0) + { + RTTESTI_CHECK(ASMMemFirstMismatchingU8(&pbPage1[PAGE_SIZE - cbSub], cbSub, 0xff) == NULL); + RTTESTI_CHECK(ASMMemFirstMismatchingU8(&pbPage2[PAGE_SIZE - cbSub], cbSub, 0xff) == NULL); + RTTESTI_CHECK(ASMMemFirstMismatchingU8(pbPage1, cbSub, 0xff) == NULL); + RTTESTI_CHECK(ASMMemFirstMismatchingU8(pbPage2, cbSub, 0xff) == NULL); + + RTTESTI_CHECK(ASMMemFirstMismatchingU8(&pbPage1[PAGE_SIZE - cbSub], cbSub, 0xfe) == &pbPage1[PAGE_SIZE - cbSub] || !cbSub); + RTTESTI_CHECK(ASMMemFirstMismatchingU8(&pbPage2[PAGE_SIZE - cbSub], cbSub, 0xfe) == &pbPage2[PAGE_SIZE - cbSub] || !cbSub); + RTTESTI_CHECK(ASMMemFirstMismatchingU8(pbPage1, cbSub, 0xfe) == pbPage1 || !cbSub); + RTTESTI_CHECK(ASMMemFirstMismatchingU8(pbPage2, cbSub, 0xfe) == pbPage2 || !cbSub); + } + + + /* + * Various alignments and sizes. + */ + uint8_t const bFiller1 = 0x00; + uint8_t const bFiller2 = 0xf6; + size_t const cbBuf = 128; + uint8_t *pbBuf1 = pbPage1; + uint8_t *pbBuf2 = &pbPage2[PAGE_SIZE - cbBuf]; /* Put it up against the tail guard */ + memset(pbPage1, ~bFiller1, PAGE_SIZE); + memset(pbPage2, ~bFiller2, PAGE_SIZE); + memset(pbBuf1, bFiller1, cbBuf); + memset(pbBuf2, bFiller2, cbBuf); + for (size_t offNonZero = 0; offNonZero < cbBuf; offNonZero++) + { + uint8_t bRand = (uint8_t)RTRandU32(); + pbBuf1[offNonZero] = bRand | 1; + pbBuf2[offNonZero] = (0x80 | bRand) ^ 0xf6; + + for (size_t offStart = 0; offStart < 32; offStart++) + { + size_t const cbMax = cbBuf - offStart; + for (size_t cb = 0; cb < cbMax; cb++) + { + size_t const offEnd = offStart + cb; + uint8_t bSaved1, bSaved2; + if (offEnd < PAGE_SIZE) + { + bSaved1 = pbBuf1[offEnd]; + bSaved2 = pbBuf2[offEnd]; + pbBuf1[offEnd] = 0xff; + pbBuf2[offEnd] = 0xff; + } +#ifdef _MSC_VER /* simple stupid compiler warnings */ + else + bSaved1 = bSaved2 = 0; +#endif + + uint8_t *pbRet = (uint8_t *)ASMMemFirstMismatchingU8(pbBuf1 + offStart, cb, bFiller1); + RTTESTI_CHECK(offNonZero - offStart < cb ? pbRet == &pbBuf1[offNonZero] : pbRet == NULL); + + pbRet = (uint8_t *)ASMMemFirstMismatchingU8(pbBuf2 + offStart, cb, bFiller2); + RTTESTI_CHECK(offNonZero - offStart < cb ? pbRet == &pbBuf2[offNonZero] : pbRet == NULL); + + if (offEnd < PAGE_SIZE) + { + pbBuf1[offEnd] = bSaved1; + pbBuf2[offEnd] = bSaved2; + } + } + } + + pbBuf1[offNonZero] = 0; + pbBuf2[offNonZero] = 0xf6; + } + + RTTestSubDone(hTest); +} + + +typedef struct TSTBUF32 { uint32_t au32[384]; } TSTBUF32; + +DECLINLINE(void) tstASMMemZero32Worker(TSTBUF32 *pBuf) +{ + ASMMemZero32(pBuf, sizeof(*pBuf)); + for (unsigned i = 0; i < RT_ELEMENTS(pBuf->au32); i++) + if (pBuf->au32[i]) + RTTestFailed(g_hTest, "ASMMemZero32 didn't clear dword at index %#x!\n", i); + if (ASMMemFirstNonZero(pBuf, sizeof(*pBuf)) != NULL) + RTTestFailed(g_hTest, "ASMMemFirstNonZero return non-NULL after ASMMemZero32\n"); + if (!ASMMemIsZero(pBuf, sizeof(*pBuf))) + RTTestFailed(g_hTest, "ASMMemIsZero return false after ASMMemZero32\n"); + + memset(pBuf, 0xfe, sizeof(*pBuf)); + ASMMemZero32(pBuf, sizeof(*pBuf)); + for (unsigned i = 0; i < RT_ELEMENTS(pBuf->au32); i++) + if (pBuf->au32[i]) + RTTestFailed(g_hTest, "ASMMemZero32 didn't clear dword at index %#x!\n", i); + if (ASMMemFirstNonZero(pBuf, sizeof(*pBuf)) != NULL) + RTTestFailed(g_hTest, "ASMMemFirstNonZero return non-NULL after ASMMemZero32\n"); + if (!ASMMemIsZero(pBuf, sizeof(*pBuf))) + RTTestFailed(g_hTest, "ASMMemIsZero return false after ASMMemZero32\n"); +} + + +void tstASMMemZero32(void) +{ + RTTestSub(g_hTest, "ASMMemZero32"); + + struct + { + uint64_t u64Magic1; + uint8_t abPage[PAGE_SIZE - 32]; + uint64_t u64Magic2; + } Buf1, Buf2, Buf3; + + Buf1.u64Magic1 = UINT64_C(0xffffffffffffffff); + memset(Buf1.abPage, 0x55, sizeof(Buf1.abPage)); + Buf1.u64Magic2 = UINT64_C(0xffffffffffffffff); + Buf2.u64Magic1 = UINT64_C(0xffffffffffffffff); + memset(Buf2.abPage, 0x77, sizeof(Buf2.abPage)); + Buf2.u64Magic2 = UINT64_C(0xffffffffffffffff); + Buf3.u64Magic1 = UINT64_C(0xffffffffffffffff); + memset(Buf3.abPage, 0x99, sizeof(Buf3.abPage)); + Buf3.u64Magic2 = UINT64_C(0xffffffffffffffff); + ASMMemZero32(Buf1.abPage, sizeof(Buf1.abPage)); + ASMMemZero32(Buf2.abPage, sizeof(Buf2.abPage)); + ASMMemZero32(Buf3.abPage, sizeof(Buf3.abPage)); + if ( Buf1.u64Magic1 != UINT64_C(0xffffffffffffffff) + || Buf1.u64Magic2 != UINT64_C(0xffffffffffffffff) + || Buf2.u64Magic1 != UINT64_C(0xffffffffffffffff) + || Buf2.u64Magic2 != UINT64_C(0xffffffffffffffff) + || Buf3.u64Magic1 != UINT64_C(0xffffffffffffffff) + || Buf3.u64Magic2 != UINT64_C(0xffffffffffffffff)) + { + RTTestFailed(g_hTest, "ASMMemZero32 violated one/both magic(s)!\n"); + } + for (unsigned i = 0; i < RT_ELEMENTS(Buf1.abPage); i++) + if (Buf1.abPage[i]) + RTTestFailed(g_hTest, "ASMMemZero32 didn't clear byte at offset %#x!\n", i); + for (unsigned i = 0; i < RT_ELEMENTS(Buf2.abPage); i++) + if (Buf2.abPage[i]) + RTTestFailed(g_hTest, "ASMMemZero32 didn't clear byte at offset %#x!\n", i); + for (unsigned i = 0; i < RT_ELEMENTS(Buf3.abPage); i++) + if (Buf3.abPage[i]) + RTTestFailed(g_hTest, "ASMMemZero32 didn't clear byte at offset %#x!\n", i); + + DO_SIMPLE_TEST_NO_SUB(tstASMMemZero32Worker, TSTBUF32); +} + + +DECLINLINE(void) tstASMMemFill32Worker(TSTBUF32 *pBuf) +{ + ASMMemFill32(pBuf, sizeof(*pBuf), UINT32_C(0xf629bce1)); + for (unsigned i = 0; i < RT_ELEMENTS(pBuf->au32); i++) + if (pBuf->au32[i] != UINT32_C(0xf629bce1)) + RTTestFailed(g_hTest, "ASMMemFill32 didn't set dword at index %#x correctly!\n", i); + if (ASMMemFirstMismatchingU32(pBuf, sizeof(*pBuf), UINT32_C(0xf629bce1)) != NULL) + RTTestFailed(g_hTest, "ASMMemFirstMismatchingU32(,,UINT32_C(0xf629bce1)) returns non-NULL after ASMMemFill32!\n"); + + memset(pBuf, 0xfe, sizeof(*pBuf)); + ASMMemFill32(pBuf, sizeof(*pBuf), UINT32_C(0x12345678)); + for (unsigned i = 0; i < RT_ELEMENTS(pBuf->au32); i++) + if (pBuf->au32[i] != UINT32_C(0x12345678)) + RTTestFailed(g_hTest, "ASMMemFill32 didn't set dword at index %#x correctly!\n", i); + if (ASMMemFirstMismatchingU32(pBuf, sizeof(*pBuf), UINT32_C(0x12345678)) != NULL) + RTTestFailed(g_hTest, "ASMMemFirstMismatchingU32(,,UINT32_C(0x12345678)) returns non-NULL after ASMMemFill32!\n"); +} + +void tstASMMemFill32(void) +{ + RTTestSub(g_hTest, "ASMMemFill32"); + + struct + { + uint64_t u64Magic1; + uint32_t au32Page[PAGE_SIZE / 4]; + uint64_t u64Magic2; + } Buf1; + struct + { + uint64_t u64Magic1; + uint32_t au32Page[(PAGE_SIZE / 4) - 3]; + uint64_t u64Magic2; + } Buf2; + struct + { + uint64_t u64Magic1; + uint32_t au32Page[(PAGE_SIZE / 4) - 1]; + uint64_t u64Magic2; + } Buf3; + + Buf1.u64Magic1 = UINT64_C(0xffffffffffffffff); + memset(Buf1.au32Page, 0x55, sizeof(Buf1.au32Page)); + Buf1.u64Magic2 = UINT64_C(0xffffffffffffffff); + Buf2.u64Magic1 = UINT64_C(0xffffffffffffffff); + memset(Buf2.au32Page, 0x77, sizeof(Buf2.au32Page)); + Buf2.u64Magic2 = UINT64_C(0xffffffffffffffff); + Buf3.u64Magic1 = UINT64_C(0xffffffffffffffff); + memset(Buf3.au32Page, 0x99, sizeof(Buf3.au32Page)); + Buf3.u64Magic2 = UINT64_C(0xffffffffffffffff); + ASMMemFill32(Buf1.au32Page, sizeof(Buf1.au32Page), 0xdeadbeef); + ASMMemFill32(Buf2.au32Page, sizeof(Buf2.au32Page), 0xcafeff01); + ASMMemFill32(Buf3.au32Page, sizeof(Buf3.au32Page), 0xf00dd00f); + if ( Buf1.u64Magic1 != UINT64_C(0xffffffffffffffff) + || Buf1.u64Magic2 != UINT64_C(0xffffffffffffffff) + || Buf2.u64Magic1 != UINT64_C(0xffffffffffffffff) + || Buf2.u64Magic2 != UINT64_C(0xffffffffffffffff) + || Buf3.u64Magic1 != UINT64_C(0xffffffffffffffff) + || Buf3.u64Magic2 != UINT64_C(0xffffffffffffffff)) + RTTestFailed(g_hTest, "ASMMemFill32 violated one/both magic(s)!\n"); + for (unsigned i = 0; i < RT_ELEMENTS(Buf1.au32Page); i++) + if (Buf1.au32Page[i] != 0xdeadbeef) + RTTestFailed(g_hTest, "ASMMemFill32 %#x: %#x exepcted %#x\n", i, Buf1.au32Page[i], 0xdeadbeef); + for (unsigned i = 0; i < RT_ELEMENTS(Buf2.au32Page); i++) + if (Buf2.au32Page[i] != 0xcafeff01) + RTTestFailed(g_hTest, "ASMMemFill32 %#x: %#x exepcted %#x\n", i, Buf2.au32Page[i], 0xcafeff01); + for (unsigned i = 0; i < RT_ELEMENTS(Buf3.au32Page); i++) + if (Buf3.au32Page[i] != 0xf00dd00f) + RTTestFailed(g_hTest, "ASMMemFill32 %#x: %#x exepcted %#x\n", i, Buf3.au32Page[i], 0xf00dd00f); + + DO_SIMPLE_TEST_NO_SUB(tstASMMemFill32Worker, TSTBUF32); +} + + +void tstASMProbe(RTTEST hTest) +{ + RTTestSub(hTest, "ASMProbeReadByte/Buffer"); + + uint8_t b = 42; + RTTESTI_CHECK(ASMProbeReadByte(&b) == 42); + ASMProbeReadBuffer(&b, sizeof(b)); + + for (uint32_t cPages = 1; cPages < 16; cPages++) + { + uint8_t *pbBuf1 = (uint8_t *)RTTestGuardedAllocHead(hTest, cPages * PAGE_SIZE); + uint8_t *pbBuf2 = (uint8_t *)RTTestGuardedAllocTail(hTest, cPages * PAGE_SIZE); + RTTESTI_CHECK_RETV(pbBuf1 && pbBuf2); + + memset(pbBuf1, 0xf6, cPages * PAGE_SIZE); + memset(pbBuf2, 0x42, cPages * PAGE_SIZE); + + RTTESTI_CHECK(ASMProbeReadByte(&pbBuf1[cPages * PAGE_SIZE - 1]) == 0xf6); + RTTESTI_CHECK(ASMProbeReadByte(&pbBuf2[cPages * PAGE_SIZE - 1]) == 0x42); + RTTESTI_CHECK(ASMProbeReadByte(&pbBuf1[0]) == 0xf6); + RTTESTI_CHECK(ASMProbeReadByte(&pbBuf2[0]) == 0x42); + + ASMProbeReadBuffer(pbBuf1, cPages * PAGE_SIZE); + ASMProbeReadBuffer(pbBuf2, cPages * PAGE_SIZE); + } +} + + +void tstASMMisc(void) +{ + RTTestSub(g_hTest, "Misc"); + for (uint32_t i = 0; i < 20; i++) + { + ASMWriteFence(); + ASMCompilerBarrier(); + ASMReadFence(); + ASMNopPause(); + ASMSerializeInstruction(); + ASMMemoryFence(); + } +} + + +void tstASMBit(void) +{ + RTTestSub(g_hTest, "ASMBitFirstSetU16"); + RTTESTI_CHECK(ASMBitFirstSetU16(0x0000) == 0); + RTTESTI_CHECK(ASMBitFirstSetU16(0x0001) == 1); + RTTESTI_CHECK(ASMBitFirstSetU16(0x8000) == 16); + RTTESTI_CHECK(ASMBitFirstSetU16(0x0ef0) == 5); + for (unsigned iBit = 0; iBit < 16; iBit++) + { + RTTESTI_CHECK(ASMBitFirstSetU16((uint16_t)1 << iBit) == iBit + 1); + RTTESTI_CHECK(ASMBitFirstSetU16(UINT16_MAX << iBit) == iBit + 1); + } + + RTTestSub(g_hTest, "ASMBitFirstSetU32"); + RTTESTI_CHECK(ASMBitFirstSetU32(UINT32_C(0x00000000)) == 0); + RTTESTI_CHECK(ASMBitFirstSetU32(UINT32_C(0x00000001)) == 1); + RTTESTI_CHECK(ASMBitFirstSetU32(UINT32_C(0x80000000)) == 32); + RTTESTI_CHECK(ASMBitFirstSetU32(UINT32_C(0x0efff0f0)) == 5); + for (unsigned iBit = 0; iBit < 32; iBit++) + { + RTTESTI_CHECK(ASMBitFirstSetU32((uint32_t)1 << iBit) == iBit + 1); + RTTESTI_CHECK(ASMBitFirstSetU32(UINT32_MAX << iBit) == iBit + 1); + } + + RTTestSub(g_hTest, "ASMBitFirstSetU64"); + RTTESTI_CHECK(ASMBitFirstSetU64(UINT64_C(0x0000000000000000)) == 0); + RTTESTI_CHECK(ASMBitFirstSetU64(UINT64_C(0x0000000000000001)) == 1); + RTTESTI_CHECK(ASMBitFirstSetU64(UINT64_C(0x8000000000000000)) == 64); + RTTESTI_CHECK(ASMBitFirstSetU64(UINT64_C(0x0effffff0ffff0f0)) == 5); + for (unsigned iBit = 0; iBit < 64; iBit++) + { + RTTESTI_CHECK(ASMBitFirstSetU64((uint64_t)1 << iBit) == iBit + 1); + RTTESTI_CHECK(ASMBitFirstSetU64(UINT64_MAX << iBit) == iBit + 1); + } + + RTTestSub(g_hTest, "ASMBitLastSetU16"); + RTTESTI_CHECK(ASMBitLastSetU16(0x0000) == 0); + RTTESTI_CHECK(ASMBitLastSetU16(0x0001) == 1); + RTTESTI_CHECK(ASMBitLastSetU16(0x8000) == 16); + RTTESTI_CHECK(ASMBitLastSetU16(0x0fe0) == 12); + for (unsigned iBit = 0; iBit < 16; iBit++) + { + RTTESTI_CHECK(ASMBitLastSetU16(UINT16_C(0x8000) >> (15 - iBit)) == iBit + 1); + RTTESTI_CHECK(ASMBitLastSetU16(UINT16_MAX >> (15 - iBit)) == iBit + 1); + } + + RTTestSub(g_hTest, "ASMBitLastSetU32"); + RTTESTI_CHECK(ASMBitLastSetU32(UINT32_C(0x00000000)) == 0); + RTTESTI_CHECK(ASMBitLastSetU32(UINT32_C(0x00000001)) == 1); + RTTESTI_CHECK(ASMBitLastSetU32(UINT32_C(0x80000000)) == 32); + RTTESTI_CHECK(ASMBitLastSetU32(UINT32_C(0x0fffffe0)) == 28); + for (unsigned iBit = 0; iBit < 32; iBit++) + { + RTTESTI_CHECK(ASMBitLastSetU32(UINT32_C(0x80000000) >> (31 - iBit)) == iBit + 1); + RTTESTI_CHECK(ASMBitLastSetU32(UINT32_MAX >> (31 - iBit)) == iBit + 1); + } + + RTTestSub(g_hTest, "ASMBitLastSetU64"); + RTTESTI_CHECK(ASMBitLastSetU64(UINT64_C(0x0000000000000000)) == 0); + RTTESTI_CHECK(ASMBitLastSetU64(UINT64_C(0x0000000000000001)) == 1); + RTTESTI_CHECK(ASMBitLastSetU64(UINT64_C(0x8000000000000000)) == 64); + RTTESTI_CHECK(ASMBitLastSetU64(UINT64_C(0x0ffffefff0ffffe0)) == 60); + for (unsigned iBit = 0; iBit < 64; iBit++) + { + RTTESTI_CHECK(ASMBitLastSetU64(UINT64_C(0x8000000000000000) >> (63 - iBit)) == iBit + 1); + RTTESTI_CHECK(ASMBitLastSetU64(UINT64_MAX >> (63 - iBit)) == iBit + 1); + } + + RTTestSub(g_hTest, "ASMCountLeadingZerosU16"); + RTTESTI_CHECK(ASMCountLeadingZerosU16(0x0000) == 16); + RTTESTI_CHECK(ASMCountLeadingZerosU16(0x0001) == 15); + RTTESTI_CHECK(ASMCountLeadingZerosU16(0x8000) == 0); + RTTESTI_CHECK(ASMCountLeadingZerosU16(0x0fe0) == 4); + for (unsigned iBit = 0; iBit < 16; iBit++) + { + RTTESTI_CHECK(ASMCountLeadingZerosU16(UINT16_C(0x8000) >> iBit) == iBit); + RTTESTI_CHECK(ASMCountLeadingZerosU16(UINT16_MAX >> iBit) == iBit); + } + + RTTestSub(g_hTest, "ASMCountLeadingZerosU32"); + RTTESTI_CHECK(ASMCountLeadingZerosU32(UINT32_C(0x00000000)) == 32); + RTTESTI_CHECK(ASMCountLeadingZerosU32(UINT32_C(0x00000001)) == 31); + RTTESTI_CHECK(ASMCountLeadingZerosU32(UINT32_C(0x80000000)) == 0); + RTTESTI_CHECK(ASMCountLeadingZerosU32(UINT32_C(0x0fffffe0)) == 4); + for (unsigned iBit = 0; iBit < 32; iBit++) + { + RTTESTI_CHECK(ASMCountLeadingZerosU32(UINT32_C(0x80000000) >> iBit) == iBit); + RTTESTI_CHECK(ASMCountLeadingZerosU32(UINT32_MAX >> iBit) == iBit); + } + + RTTestSub(g_hTest, "ASMCountLeadingZerosU64"); + RTTESTI_CHECK(ASMCountLeadingZerosU64(UINT64_C(0x0000000000000000)) == 64); + RTTESTI_CHECK(ASMCountLeadingZerosU64(UINT64_C(0x0000000000000001)) == 63); + RTTESTI_CHECK(ASMCountLeadingZerosU64(UINT64_C(0x8000000000000000)) == 0); + RTTESTI_CHECK(ASMCountLeadingZerosU64(UINT64_C(0x0fffffff0f0fffe0)) == 4); + for (unsigned iBit = 0; iBit < 64; iBit++) + { + RTTESTI_CHECK(ASMCountLeadingZerosU64(UINT64_C(0x8000000000000000) >> iBit) == iBit); + RTTESTI_CHECK(ASMCountLeadingZerosU64(UINT64_MAX >> iBit) == iBit); + } + + RTTestSub(g_hTest, "ASMCountTrailingZerosU16"); + RTTESTI_CHECK(ASMCountTrailingZerosU16(0x0000) == 16); + RTTESTI_CHECK(ASMCountTrailingZerosU16(0x0001) == 0); + RTTESTI_CHECK(ASMCountTrailingZerosU16(0x8000) == 15); + RTTESTI_CHECK(ASMCountTrailingZerosU16(0x0ef0) == 4); + for (unsigned iBit = 0; iBit < 16; iBit++) + { + RTTESTI_CHECK(ASMCountTrailingZerosU16((uint16_t)1 << iBit) == iBit); + RTTESTI_CHECK(ASMCountTrailingZerosU16(UINT16_MAX << iBit) == iBit); + } + + RTTestSub(g_hTest, "ASMCountTrailingZerosU32"); + RTTESTI_CHECK(ASMCountTrailingZerosU32(UINT32_C(0x00000000)) == 32); + RTTESTI_CHECK(ASMCountTrailingZerosU32(UINT32_C(0x00000001)) == 0); + RTTESTI_CHECK(ASMCountTrailingZerosU32(UINT32_C(0x80000000)) == 31); + RTTESTI_CHECK(ASMCountTrailingZerosU32(UINT32_C(0x0efffff0)) == 4); + for (unsigned iBit = 0; iBit < 32; iBit++) + { + RTTESTI_CHECK(ASMCountTrailingZerosU32((uint32_t)1 << iBit) == iBit); + RTTESTI_CHECK(ASMCountTrailingZerosU32(UINT32_MAX << iBit) == iBit); + } + + RTTestSub(g_hTest, "ASMCountTrailingZerosU64"); + RTTESTI_CHECK(ASMCountTrailingZerosU64(UINT64_C(0x0000000000000000)) == 64); + RTTESTI_CHECK(ASMCountTrailingZerosU64(UINT64_C(0x0000000000000001)) == 0); + RTTESTI_CHECK(ASMCountTrailingZerosU64(UINT64_C(0x8000000000000000)) == 63); + RTTESTI_CHECK(ASMCountTrailingZerosU64(UINT64_C(0x0effff0fefef0ff0)) == 4); + for (unsigned iBit = 0; iBit < 64; iBit++) + { + RTTESTI_CHECK(ASMCountTrailingZerosU64((uint64_t)1 << iBit) == iBit); + RTTESTI_CHECK(ASMCountTrailingZerosU64(UINT64_MAX << iBit) == iBit); + } +} + + +void tstASMMath(void) +{ + RTTestSub(g_hTest, "Math"); + + uint64_t u64 = ASMMult2xU32RetU64(UINT32_C(0x80000000), UINT32_C(0x10000000)); + CHECKVAL(u64, UINT64_C(0x0800000000000000), "%#018RX64"); + + uint32_t u32 = ASMDivU64ByU32RetU32(UINT64_C(0x0800000000000000), UINT32_C(0x10000000)); + CHECKVAL(u32, UINT32_C(0x80000000), "%#010RX32"); + + u32 = ASMMultU32ByU32DivByU32(UINT32_C(0x00000001), UINT32_C(0x00000001), UINT32_C(0x00000001)); + CHECKVAL(u32, UINT32_C(0x00000001), "%#018RX32"); + u32 = ASMMultU32ByU32DivByU32(UINT32_C(0x10000000), UINT32_C(0x80000000), UINT32_C(0x20000000)); + CHECKVAL(u32, UINT32_C(0x40000000), "%#018RX32"); + u32 = ASMMultU32ByU32DivByU32(UINT32_C(0x76543210), UINT32_C(0xffffffff), UINT32_C(0xffffffff)); + CHECKVAL(u32, UINT32_C(0x76543210), "%#018RX32"); + u32 = ASMMultU32ByU32DivByU32(UINT32_C(0xffffffff), UINT32_C(0xffffffff), UINT32_C(0xffffffff)); + CHECKVAL(u32, UINT32_C(0xffffffff), "%#018RX32"); + u32 = ASMMultU32ByU32DivByU32(UINT32_C(0xffffffff), UINT32_C(0xfffffff0), UINT32_C(0xffffffff)); + CHECKVAL(u32, UINT32_C(0xfffffff0), "%#018RX32"); + u32 = ASMMultU32ByU32DivByU32(UINT32_C(0x10359583), UINT32_C(0x58734981), UINT32_C(0xf8694045)); + CHECKVAL(u32, UINT32_C(0x05c584ce), "%#018RX32"); + u32 = ASMMultU32ByU32DivByU32(UINT32_C(0x10359583), UINT32_C(0xf8694045), UINT32_C(0x58734981)); + CHECKVAL(u32, UINT32_C(0x2d860795), "%#018RX32"); + +#if defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86) + u64 = ASMMultU64ByU32DivByU32(UINT64_C(0x0000000000000001), UINT32_C(0x00000001), UINT32_C(0x00000001)); + CHECKVAL(u64, UINT64_C(0x0000000000000001), "%#018RX64"); + u64 = ASMMultU64ByU32DivByU32(UINT64_C(0x0000000100000000), UINT32_C(0x80000000), UINT32_C(0x00000002)); + CHECKVAL(u64, UINT64_C(0x4000000000000000), "%#018RX64"); + u64 = ASMMultU64ByU32DivByU32(UINT64_C(0xfedcba9876543210), UINT32_C(0xffffffff), UINT32_C(0xffffffff)); + CHECKVAL(u64, UINT64_C(0xfedcba9876543210), "%#018RX64"); + u64 = ASMMultU64ByU32DivByU32(UINT64_C(0xffffffffffffffff), UINT32_C(0xffffffff), UINT32_C(0xffffffff)); + CHECKVAL(u64, UINT64_C(0xffffffffffffffff), "%#018RX64"); + u64 = ASMMultU64ByU32DivByU32(UINT64_C(0xffffffffffffffff), UINT32_C(0xfffffff0), UINT32_C(0xffffffff)); + CHECKVAL(u64, UINT64_C(0xfffffff0fffffff0), "%#018RX64"); + u64 = ASMMultU64ByU32DivByU32(UINT64_C(0x3415934810359583), UINT32_C(0x58734981), UINT32_C(0xf8694045)); + CHECKVAL(u64, UINT64_C(0x128b9c3d43184763), "%#018RX64"); + u64 = ASMMultU64ByU32DivByU32(UINT64_C(0x3415934810359583), UINT32_C(0xf8694045), UINT32_C(0x58734981)); + CHECKVAL(u64, UINT64_C(0x924719355cd35a27), "%#018RX64"); + +# if 0 /* bird: question is whether this should trap or not: + * + * frank: Of course it must trap: + * + * 0xfffffff8 * 0x77d7daf8 = 0x77d7daf441412840 + * + * During the following division, the quotient must fit into a 32-bit register. + * Therefore the smallest valid divisor is + * + * (0x77d7daf441412840 >> 32) + 1 = 0x77d7daf5 + * + * which is definitely greater than 0x3b9aca00. + * + * bird: No, the C version does *not* crash. So, the question is whether there's any + * code depending on it not crashing. + * + * Of course the assembly versions of the code crash right now for the reasons you've + * given, but the 32-bit MSC version does not crash. + * + * frank: The C version does not crash but delivers incorrect results for this case. + * The reason is + * + * u.s.Hi = (unsigned long)(u64Hi / u32C); + * + * Here the division is actually 64-bit by 64-bit but the 64-bit result is truncated + * to 32 bit. If using this (optimized and fast) function we should just be sure that + * the operands are in a valid range. + */ + u64 = ASMMultU64ByU32DivByU32(UINT64_C(0xfffffff8c65d6731), UINT32_C(0x77d7daf8), UINT32_C(0x3b9aca00)); + CHECKVAL(u64, UINT64_C(0x02b8f9a2aa74e3dc), "%#018RX64"); +# endif +#endif /* AMD64 || X86 */ + + u32 = ASMModU64ByU32RetU32(UINT64_C(0x0ffffff8c65d6731), UINT32_C(0x77d7daf8)); + CHECKVAL(u32, UINT32_C(0x3B642451), "%#010RX32"); + + int32_t i32; + i32 = ASMModS64ByS32RetS32(INT64_C(-11), INT32_C(-2)); + CHECKVAL(i32, INT32_C(-1), "%010RI32"); + i32 = ASMModS64ByS32RetS32(INT64_C(-11), INT32_C(2)); + CHECKVAL(i32, INT32_C(-1), "%010RI32"); + i32 = ASMModS64ByS32RetS32(INT64_C(11), INT32_C(-2)); + CHECKVAL(i32, INT32_C(1), "%010RI32"); + + i32 = ASMModS64ByS32RetS32(INT64_C(92233720368547758), INT32_C(2147483647)); + CHECKVAL(i32, INT32_C(2104533974), "%010RI32"); + i32 = ASMModS64ByS32RetS32(INT64_C(-92233720368547758), INT32_C(2147483647)); + CHECKVAL(i32, INT32_C(-2104533974), "%010RI32"); +} + + +void tstASMByteSwap(void) +{ + RTTestSub(g_hTest, "ASMByteSwap*"); + + uint64_t u64In = UINT64_C(0x0011223344556677); + uint64_t u64Out = ASMByteSwapU64(u64In); + CHECKVAL(u64In, UINT64_C(0x0011223344556677), "%#018RX64"); + CHECKVAL(u64Out, UINT64_C(0x7766554433221100), "%#018RX64"); + u64Out = ASMByteSwapU64(u64Out); + CHECKVAL(u64Out, u64In, "%#018RX64"); + u64In = UINT64_C(0x0123456789abcdef); + u64Out = ASMByteSwapU64(u64In); + CHECKVAL(u64In, UINT64_C(0x0123456789abcdef), "%#018RX64"); + CHECKVAL(u64Out, UINT64_C(0xefcdab8967452301), "%#018RX64"); + u64Out = ASMByteSwapU64(u64Out); + CHECKVAL(u64Out, u64In, "%#018RX64"); + u64In = 0; + u64Out = ASMByteSwapU64(u64In); + CHECKVAL(u64Out, u64In, "%#018RX64"); + u64In = UINT64_MAX; + u64Out = ASMByteSwapU64(u64In); + CHECKVAL(u64Out, u64In, "%#018RX64"); + + uint32_t u32In = UINT32_C(0x00112233); + uint32_t u32Out = ASMByteSwapU32(u32In); + CHECKVAL(u32In, UINT32_C(0x00112233), "%#010RX32"); + CHECKVAL(u32Out, UINT32_C(0x33221100), "%#010RX32"); + u32Out = ASMByteSwapU32(u32Out); + CHECKVAL(u32Out, u32In, "%#010RX32"); + u32In = UINT32_C(0x12345678); + u32Out = ASMByteSwapU32(u32In); + CHECKVAL(u32In, UINT32_C(0x12345678), "%#010RX32"); + CHECKVAL(u32Out, UINT32_C(0x78563412), "%#010RX32"); + u32Out = ASMByteSwapU32(u32Out); + CHECKVAL(u32Out, u32In, "%#010RX32"); + u32In = 0; + u32Out = ASMByteSwapU32(u32In); + CHECKVAL(u32Out, u32In, "%#010RX32"); + u32In = UINT32_MAX; + u32Out = ASMByteSwapU32(u32In); + CHECKVAL(u32Out, u32In, "%#010RX32"); + + uint16_t u16In = UINT16_C(0x0011); + uint16_t u16Out = ASMByteSwapU16(u16In); + CHECKVAL(u16In, UINT16_C(0x0011), "%#06RX16"); + CHECKVAL(u16Out, UINT16_C(0x1100), "%#06RX16"); + u16Out = ASMByteSwapU16(u16Out); + CHECKVAL(u16Out, u16In, "%#06RX16"); + u16In = UINT16_C(0x1234); + u16Out = ASMByteSwapU16(u16In); + CHECKVAL(u16In, UINT16_C(0x1234), "%#06RX16"); + CHECKVAL(u16Out, UINT16_C(0x3412), "%#06RX16"); + u16Out = ASMByteSwapU16(u16Out); + CHECKVAL(u16Out, u16In, "%#06RX16"); + u16In = 0; + u16Out = ASMByteSwapU16(u16In); + CHECKVAL(u16Out, u16In, "%#06RX16"); + u16In = UINT16_MAX; + u16Out = ASMByteSwapU16(u16In); + CHECKVAL(u16Out, u16In, "%#06RX16"); +} + + +void tstASMBench(void) +{ + /* + * Make this static. We don't want to have this located on the stack. + */ + static uint8_t volatile s_u8; + static int8_t volatile s_i8; + static uint16_t volatile s_u16; + static int16_t volatile s_i16; + static uint32_t volatile s_u32; + static int32_t volatile s_i32; + static uint64_t volatile s_u64; + static int64_t volatile s_i64; +#if defined(RTASM_HAVE_CMP_WRITE_U128) || defined(RTASM_HAVE_CMP_XCHG_U128) + static RTUINT128U volatile s_u128; +#endif + static uint8_t s_u8Old; + static int8_t s_i8Old; + static uint16_t s_u16Old; + static int16_t s_i16Old; + static uint32_t s_u32Old; + static int32_t s_i32Old; + static uint64_t s_u64Old; + static int64_t s_i64Old; +#if defined(RTASM_HAVE_CMP_WRITE_U128) || defined(RTASM_HAVE_CMP_XCHG_U128) + static RTUINT128U s_u128Old; + RTUINT128U u128Tmp1, u128Tmp2; +# ifdef RT_ARCH_AMD64 + bool const fHaveCmpXchg128 = RT_BOOL(ASMCpuId_ECX(1) & X86_CPUID_FEATURE_ECX_CX16); +# else + bool const fHaveCmpXchg128 = true; +# endif +#endif + unsigned i; + const unsigned cRounds = _16M; /* Must be multiple of 8 */ + uint64_t u64Elapsed; + + RTTestSub(g_hTest, "Benchmarking"); + +#if 0 && !defined(GCC44_32BIT_PIC) && (defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86) || defined(RT_ARCH_ARM64) || defined(RT_ARCH_ARM32)) +# define BENCH(op, str) \ + do { \ + RTThreadYield(); \ + u64Elapsed = ASMReadTSC(); \ + for (i = cRounds; i > 0; i--) \ + op; \ + u64Elapsed = ASMReadTSC() - u64Elapsed; \ + RTTestValue(g_hTest, str, u64Elapsed / cRounds, RTTESTUNIT_TICKS_PER_CALL); \ + } while (0) +#else +# define BENCH(op, str) \ + do { \ + RTThreadYield(); \ + u64Elapsed = RTTimeNanoTS(); \ + for (i = cRounds / 8; i > 0; i--) \ + { \ + op; \ + op; \ + op; \ + op; \ + op; \ + op; \ + op; \ + op; \ + } \ + u64Elapsed = RTTimeNanoTS() - u64Elapsed; \ + RTTestValue(g_hTest, str, u64Elapsed * 1000 / cRounds, RTTESTUNIT_PS_PER_CALL); \ + } while (0) +#endif +#if (defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86) || defined(RT_ARCH_ARM64) || defined(RT_ARCH_ARM32)) && !defined(GCC44_32BIT_PIC) +# define BENCH_TSC(op, str) \ + do { \ + RTThreadYield(); \ + u64Elapsed = ASMReadTSC(); \ + for (i = cRounds / 8; i > 0; i--) \ + { \ + op; \ + op; \ + op; \ + op; \ + op; \ + op; \ + op; \ + op; \ + } \ + u64Elapsed = ASMReadTSC() - u64Elapsed; \ + RTTestValue(g_hTest, str, u64Elapsed / cRounds, RTTESTUNIT_TICKS_PER_CALL); \ + } while (0) +#else +# define BENCH_TSC(op, str) BENCH(op, str) +#endif + + BENCH(s_u32 = 0, "s_u32 = 0"); + BENCH(ASMAtomicUoReadU8(&s_u8), "ASMAtomicUoReadU8"); + BENCH(ASMAtomicUoReadS8(&s_i8), "ASMAtomicUoReadS8"); + BENCH(ASMAtomicUoReadU16(&s_u16), "ASMAtomicUoReadU16"); + BENCH(ASMAtomicUoReadS16(&s_i16), "ASMAtomicUoReadS16"); + BENCH(ASMAtomicUoReadU32(&s_u32), "ASMAtomicUoReadU32"); + BENCH(ASMAtomicUoReadS32(&s_i32), "ASMAtomicUoReadS32"); + BENCH(ASMAtomicUoReadU64(&s_u64), "ASMAtomicUoReadU64"); + BENCH(ASMAtomicUoReadS64(&s_i64), "ASMAtomicUoReadS64"); + BENCH(ASMAtomicReadU8(&s_u8), "ASMAtomicReadU8"); + BENCH(ASMAtomicReadS8(&s_i8), "ASMAtomicReadS8"); + BENCH(ASMAtomicReadU16(&s_u16), "ASMAtomicReadU16"); + BENCH(ASMAtomicReadS16(&s_i16), "ASMAtomicReadS16"); + BENCH(ASMAtomicReadU32(&s_u32), "ASMAtomicReadU32"); + BENCH(ASMAtomicReadS32(&s_i32), "ASMAtomicReadS32"); + BENCH(ASMAtomicReadU64(&s_u64), "ASMAtomicReadU64"); + BENCH(ASMAtomicReadS64(&s_i64), "ASMAtomicReadS64"); + BENCH(ASMAtomicUoWriteU8(&s_u8, 0), "ASMAtomicUoWriteU8"); + BENCH(ASMAtomicUoWriteS8(&s_i8, 0), "ASMAtomicUoWriteS8"); + BENCH(ASMAtomicUoWriteU16(&s_u16, 0), "ASMAtomicUoWriteU16"); + BENCH(ASMAtomicUoWriteS16(&s_i16, 0), "ASMAtomicUoWriteS16"); + BENCH(ASMAtomicUoWriteU32(&s_u32, 0), "ASMAtomicUoWriteU32"); + BENCH(ASMAtomicUoWriteS32(&s_i32, 0), "ASMAtomicUoWriteS32"); + BENCH(ASMAtomicUoWriteU64(&s_u64, 0), "ASMAtomicUoWriteU64"); + BENCH(ASMAtomicUoWriteS64(&s_i64, 0), "ASMAtomicUoWriteS64"); + BENCH(ASMAtomicWriteU8(&s_u8, 0), "ASMAtomicWriteU8"); + BENCH(ASMAtomicWriteS8(&s_i8, 0), "ASMAtomicWriteS8"); + BENCH(ASMAtomicWriteU16(&s_u16, 0), "ASMAtomicWriteU16"); + BENCH(ASMAtomicWriteS16(&s_i16, 0), "ASMAtomicWriteS16"); + BENCH(ASMAtomicWriteU32(&s_u32, 0), "ASMAtomicWriteU32"); + BENCH(ASMAtomicWriteS32(&s_i32, 0), "ASMAtomicWriteS32"); + BENCH(ASMAtomicWriteU64(&s_u64, 0), "ASMAtomicWriteU64"); + BENCH(ASMAtomicWriteS64(&s_i64, 0), "ASMAtomicWriteS64"); + BENCH(ASMAtomicXchgU8(&s_u8, 0), "ASMAtomicXchgU8"); + BENCH(ASMAtomicXchgS8(&s_i8, 0), "ASMAtomicXchgS8"); + BENCH(ASMAtomicXchgU16(&s_u16, 0), "ASMAtomicXchgU16"); + BENCH(ASMAtomicXchgS16(&s_i16, 0), "ASMAtomicXchgS16"); + BENCH(ASMAtomicXchgU32(&s_u32, 0), "ASMAtomicXchgU32"); + BENCH(ASMAtomicXchgS32(&s_i32, 0), "ASMAtomicXchgS32"); + BENCH(ASMAtomicXchgU64(&s_u64, 0), "ASMAtomicXchgU64"); + BENCH(ASMAtomicXchgS64(&s_i64, 0), "ASMAtomicXchgS64"); + BENCH(ASMAtomicCmpXchgU8(&s_u8, 0, 0), "ASMAtomicCmpXchgU8"); + BENCH(ASMAtomicCmpXchgS8(&s_i8, 0, 0), "ASMAtomicCmpXchgS8"); + //BENCH(ASMAtomicCmpXchgU16(&s_u16, 0, 0), "ASMAtomicCmpXchgU16"); + //BENCH(ASMAtomicCmpXchgS16(&s_i16, 0, 0), "ASMAtomicCmpXchgS16"); + BENCH(ASMAtomicCmpXchgU32(&s_u32, 0, 0), "ASMAtomicCmpXchgU32"); + BENCH(ASMAtomicCmpXchgS32(&s_i32, 0, 0), "ASMAtomicCmpXchgS32"); + BENCH(ASMAtomicCmpXchgU64(&s_u64, 0, 0), "ASMAtomicCmpXchgU64"); + BENCH(ASMAtomicCmpXchgS64(&s_i64, 0, 0), "ASMAtomicCmpXchgS64"); +#ifdef RTASM_HAVE_CMP_WRITE_U128 + if (fHaveCmpXchg128) + BENCH(ASMAtomicCmpWriteU128U(&s_u128, u128Tmp1 = RTUINT128_INIT_C(0, 0), u128Tmp2 = RTUINT128_INIT_C(0, 0)), + "ASMAtomicCmpWriteU128U"); +#endif + BENCH(ASMAtomicCmpXchgU8(&s_u8, 0, 1), "ASMAtomicCmpXchgU8/neg"); + BENCH(ASMAtomicCmpXchgS8(&s_i8, 0, 1), "ASMAtomicCmpXchgS8/neg"); + //BENCH(ASMAtomicCmpXchgU16(&s_u16, 0, 1), "ASMAtomicCmpXchgU16/neg"); + //BENCH(ASMAtomicCmpXchgS16(&s_s16, 0, 1), "ASMAtomicCmpXchgS16/neg"); + BENCH(ASMAtomicCmpXchgU32(&s_u32, 0, 1), "ASMAtomicCmpXchgU32/neg"); + BENCH(ASMAtomicCmpXchgS32(&s_i32, 0, 1), "ASMAtomicCmpXchgS32/neg"); + BENCH(ASMAtomicCmpXchgU64(&s_u64, 0, 1), "ASMAtomicCmpXchgU64/neg"); + BENCH(ASMAtomicCmpXchgS64(&s_i64, 0, 1), "ASMAtomicCmpXchgS64/neg"); +#ifdef RTASM_HAVE_CMP_WRITE_U128 + if (fHaveCmpXchg128) + BENCH(ASMAtomicCmpWriteU128U(&s_u128, u128Tmp1 = RTUINT128_INIT_C(0, 0), u128Tmp2 = RTUINT128_INIT_C(0, 1)), + "ASMAtomicCmpWriteU128U/neg"); +#endif + BENCH(ASMAtomicCmpXchgExU8(&s_u8, 0, 0, &s_u8Old), "ASMAtomicCmpXchgExU8"); + BENCH(ASMAtomicCmpXchgExS8(&s_i8, 0, 0, &s_i8Old), "ASMAtomicCmpXchgExS8"); + BENCH(ASMAtomicCmpXchgExU16(&s_u16, 0, 0, &s_u16Old), "ASMAtomicCmpXchgExU16"); + BENCH(ASMAtomicCmpXchgExS16(&s_i16, 0, 0, &s_i16Old), "ASMAtomicCmpXchgExS16"); + BENCH(ASMAtomicCmpXchgExU32(&s_u32, 0, 0, &s_u32Old), "ASMAtomicCmpXchgExU32"); + BENCH(ASMAtomicCmpXchgExS32(&s_i32, 0, 0, &s_i32Old), "ASMAtomicCmpXchgExS32"); + BENCH(ASMAtomicCmpXchgExU64(&s_u64, 0, 0, &s_u64Old), "ASMAtomicCmpXchgExU64"); + BENCH(ASMAtomicCmpXchgExS64(&s_i64, 0, 0, &s_i64Old), "ASMAtomicCmpXchgExS64"); +#ifdef RTASM_HAVE_CMP_XCHG_U128 + if (fHaveCmpXchg128) + BENCH(ASMAtomicCmpXchgU128U(&s_u128, u128Tmp1 = RTUINT128_INIT_C(0, 0), u128Tmp2 = RTUINT128_INIT_C(0, 0), &s_u128Old), + "ASMAtomicCmpXchgU128U"); +#endif + BENCH(ASMAtomicCmpXchgExU8(&s_u8, 0, 1, &s_u8Old), "ASMAtomicCmpXchgExU8/neg"); + BENCH(ASMAtomicCmpXchgExS8(&s_i8, 0, 1, &s_i8Old), "ASMAtomicCmpXchgExS8/neg"); + BENCH(ASMAtomicCmpXchgExU16(&s_u16, 0, 1, &s_u16Old), "ASMAtomicCmpXchgExU16/neg"); + BENCH(ASMAtomicCmpXchgExS16(&s_i16, 0, 1, &s_i16Old), "ASMAtomicCmpXchgExS16/neg"); + BENCH(ASMAtomicCmpXchgExU32(&s_u32, 0, 1, &s_u32Old), "ASMAtomicCmpXchgExU32/neg"); + BENCH(ASMAtomicCmpXchgExS32(&s_i32, 0, 1, &s_i32Old), "ASMAtomicCmpXchgExS32/neg"); + BENCH(ASMAtomicCmpXchgExU64(&s_u64, 0, 1, &s_u64Old), "ASMAtomicCmpXchgExU64/neg"); + BENCH(ASMAtomicCmpXchgExS64(&s_i64, 0, 1, &s_i64Old), "ASMAtomicCmpXchgExS64/neg"); +#ifdef RTASM_HAVE_CMP_XCHG_U128 + if (fHaveCmpXchg128) + BENCH(ASMAtomicCmpXchgU128U(&s_u128, u128Tmp1 = RTUINT128_INIT_C(0, 0), u128Tmp2 = RTUINT128_INIT_C(0, 1), &s_u128Old), + "ASMAtomicCmpXchgU128U/neg"); +#endif + BENCH(ASMAtomicIncU32(&s_u32), "ASMAtomicIncU32"); + BENCH(ASMAtomicIncS32(&s_i32), "ASMAtomicIncS32"); + BENCH(ASMAtomicDecU32(&s_u32), "ASMAtomicDecU32"); + BENCH(ASMAtomicDecS32(&s_i32), "ASMAtomicDecS32"); + BENCH(ASMAtomicAddU32(&s_u32, 5), "ASMAtomicAddU32"); + BENCH(ASMAtomicAddS32(&s_i32, 5), "ASMAtomicAddS32"); + BENCH(ASMAtomicUoIncU32(&s_u32), "ASMAtomicUoIncU32"); + BENCH(ASMAtomicUoDecU32(&s_u32), "ASMAtomicUoDecU32"); + BENCH(ASMAtomicUoAndU32(&s_u32, 0xffffffff), "ASMAtomicUoAndU32"); + BENCH(ASMAtomicUoOrU32(&s_u32, 0xffffffff), "ASMAtomicUoOrU32"); +#if defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86) + BENCH_TSC(ASMSerializeInstructionCpuId(), "ASMSerializeInstructionCpuId"); + BENCH_TSC(ASMSerializeInstructionIRet(), "ASMSerializeInstructionIRet"); +#endif + BENCH(ASMReadFence(), "ASMReadFence"); + BENCH(ASMWriteFence(), "ASMWriteFence"); + BENCH(ASMMemoryFence(), "ASMMemoryFence"); + BENCH(ASMSerializeInstruction(), "ASMSerializeInstruction"); + BENCH(ASMNopPause(), "ASMNopPause"); + + BENCH(ASMBitFirstSetU16(s_u16), "ASMBitFirstSetU16"); + BENCH(ASMBitFirstSetU32(s_u32), "ASMBitFirstSetU32"); + BENCH(ASMBitFirstSetU64(s_u32), "ASMBitFirstSetU64"); + BENCH(ASMBitLastSetU16(s_u16), "ASMBitLastSetU16"); + BENCH(ASMBitLastSetU32(s_u32), "ASMBitLastSetU32"); + BENCH(ASMBitLastSetU64(s_u32), "ASMBitLastSetU64"); + BENCH(ASMCountLeadingZerosU16(s_u16), "ASMCountLeadingZerosU16"); + BENCH(ASMCountLeadingZerosU32(s_u32), "ASMCountLeadingZerosU32"); + BENCH(ASMCountLeadingZerosU64(s_u64), "ASMCountLeadingZerosU64"); + BENCH(ASMCountTrailingZerosU16(s_u16), "ASMCountTrailingZerosU16"); + BENCH(ASMCountTrailingZerosU32(s_u32), "ASMCountTrailingZerosU32"); + BENCH(ASMCountTrailingZerosU64(s_u64), "ASMCountTrailingZerosU64"); + + /* The Darwin gcc does not like this ... */ +#if !defined(RT_OS_DARWIN) && !defined(GCC44_32BIT_PIC) && (defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86)) + BENCH(s_u8 = ASMGetApicId(), "ASMGetApicId"); + BENCH(s_u32 = ASMGetApicIdExt0B(), "ASMGetApicIdExt0B"); + BENCH(s_u32 = ASMGetApicIdExt8000001E(), "ASMGetApicIdExt8000001E"); +#endif +#if !defined(GCC44_32BIT_PIC) && (defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86) || defined(RT_ARCH_ARM64) || defined(RT_ARCH_ARM32)) + BENCH(s_u64 = ASMReadTSC(), "ASMReadTSC"); +#endif +#if !defined(GCC44_32BIT_PIC) && (defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86)) + uint32_t uAux; + if ( ASMHasCpuId() + && RTX86IsValidExtRange(ASMCpuId_EAX(0x80000000)) + && (ASMCpuId_EDX(0x80000001) & X86_CPUID_EXT_FEATURE_EDX_RDTSCP) ) + { + BENCH_TSC(ASMSerializeInstructionRdTscp(), "ASMSerializeInstructionRdTscp"); + BENCH(s_u64 = ASMReadTscWithAux(&uAux), "ASMReadTscWithAux"); + } + union + { + uint64_t u64[2]; + RTIDTR Unaligned; + struct + { + uint16_t abPadding[3]; + RTIDTR Aligned; + } s; + } uBuf; + Assert(((uintptr_t)&uBuf.Unaligned.pIdt & (sizeof(uintptr_t) - 1)) != 0); + BENCH(ASMGetIDTR(&uBuf.Unaligned), "ASMGetIDTR/unaligned"); + Assert(((uintptr_t)&uBuf.s.Aligned.pIdt & (sizeof(uintptr_t) - 1)) == 0); + BENCH(ASMGetIDTR(&uBuf.s.Aligned), "ASMGetIDTR/aligned"); +#endif + +#undef BENCH +} + + +int main(int argc, char **argv) +{ + RT_NOREF_PV(argc); RT_NOREF_PV(argv); + + int rc = RTTestInitAndCreate("tstRTInlineAsm", &g_hTest); + if (rc) + return rc; + RTTestBanner(g_hTest); + + /* + * Execute the tests. + */ +#if !defined(GCC44_32BIT_PIC) && (defined(RT_ARCH_AMD64) || defined(RT_ARCH_X86)) + tstASMCpuId(); + //bruteForceCpuId(); +#endif +#if 1 + tstASMAtomicRead(); + tstASMAtomicWrite(); + tstASMAtomicXchg(); + tstASMAtomicCmpXchg(); + tstASMAtomicCmpXchgEx(); + + tstASMAtomicAdd(); + tstASMAtomicDecInc(); + tstASMAtomicAndOrXor(); + + tstASMMemZeroPage(); + tstASMMemIsZeroPage(g_hTest); + tstASMMemFirstMismatchingU8(g_hTest); + tstASMMemZero32(); + tstASMMemFill32(); + tstASMProbe(g_hTest); + + tstASMMisc(); + + tstASMBit(); + + tstASMMath(); + + tstASMByteSwap(); + + tstASMBench(); +#endif + + /* + * Show the result. + */ + return RTTestSummaryAndDestroy(g_hTest); +} + |