summaryrefslogtreecommitdiffstats
path: root/src/VBox/Devices/Bus/MsiCommon.cpp
diff options
context:
space:
mode:
Diffstat (limited to '')
-rw-r--r--src/VBox/Devices/Bus/MsiCommon.cpp336
1 files changed, 336 insertions, 0 deletions
diff --git a/src/VBox/Devices/Bus/MsiCommon.cpp b/src/VBox/Devices/Bus/MsiCommon.cpp
new file mode 100644
index 00000000..4ed91c58
--- /dev/null
+++ b/src/VBox/Devices/Bus/MsiCommon.cpp
@@ -0,0 +1,336 @@
+/* $Id: MsiCommon.cpp $ */
+/** @file
+ * MSI support routines
+ *
+ * @todo Straighten up this file!!
+ */
+
+/*
+ * Copyright (C) 2010-2019 Oracle Corporation
+ *
+ * This file is part of VirtualBox Open Source Edition (OSE), as
+ * available from http://www.virtualbox.org. This file is free software;
+ * you can redistribute it and/or modify it under the terms of the GNU
+ * General Public License (GPL) as published by the Free Software
+ * Foundation, in version 2 as it comes in the "COPYING" file of the
+ * VirtualBox OSE distribution. VirtualBox OSE is distributed in the
+ * hope that it will be useful, but WITHOUT ANY WARRANTY of any kind.
+ */
+
+#define LOG_GROUP LOG_GROUP_DEV_PCI
+#define PDMPCIDEV_INCLUDE_PRIVATE /* Hack to get pdmpcidevint.h included at the right point. */
+#include <VBox/pci.h>
+#include <VBox/msi.h>
+#include <VBox/vmm/pdmdev.h>
+#include <VBox/log.h>
+
+#include "MsiCommon.h"
+#include "PciInline.h"
+
+
+DECLINLINE(uint16_t) msiGetMessageControl(PPDMPCIDEV pDev)
+{
+ uint32_t idxMessageControl = pDev->Int.s.u8MsiCapOffset + VBOX_MSI_CAP_MESSAGE_CONTROL;
+#ifdef IN_RING3
+ if (pciDevIsPassthrough(pDev))
+ return pDev->Int.s.pfnConfigRead(pDev->Int.s.CTX_SUFF(pDevIns), pDev, idxMessageControl, 2);
+#endif
+ return PCIDevGetWord(pDev, idxMessageControl);
+}
+
+DECLINLINE(bool) msiIs64Bit(PPDMPCIDEV pDev)
+{
+ return pciDevIsMsi64Capable(pDev);
+}
+
+/** @todo r=klaus This design assumes that the config space cache is always
+ * up to date, which is a wrong assumption for the "emulate passthrough" case
+ * where only the callbacks give the correct data. */
+DECLINLINE(uint32_t *) msiGetMaskBits(PPDMPCIDEV pDev)
+{
+ uint8_t iOff = msiIs64Bit(pDev) ? VBOX_MSI_CAP_MASK_BITS_64 : VBOX_MSI_CAP_MASK_BITS_32;
+ /* devices may have no masked/pending support */
+ if (iOff >= pDev->Int.s.u8MsiCapSize)
+ return NULL;
+ iOff += pDev->Int.s.u8MsiCapOffset;
+ return (uint32_t*)(pDev->abConfig + iOff);
+}
+
+/** @todo r=klaus This design assumes that the config space cache is always
+ * up to date, which is a wrong assumption for the "emulate passthrough" case
+ * where only the callbacks give the correct data. */
+DECLINLINE(uint32_t*) msiGetPendingBits(PPDMPCIDEV pDev)
+{
+ uint8_t iOff = msiIs64Bit(pDev) ? VBOX_MSI_CAP_PENDING_BITS_64 : VBOX_MSI_CAP_PENDING_BITS_32;
+ /* devices may have no masked/pending support */
+ if (iOff >= pDev->Int.s.u8MsiCapSize)
+ return NULL;
+ iOff += pDev->Int.s.u8MsiCapOffset;
+ return (uint32_t*)(pDev->abConfig + iOff);
+}
+
+DECLINLINE(bool) msiIsEnabled(PPDMPCIDEV pDev)
+{
+ return (msiGetMessageControl(pDev) & VBOX_PCI_MSI_FLAGS_ENABLE) != 0;
+}
+
+DECLINLINE(uint8_t) msiGetMme(PPDMPCIDEV pDev)
+{
+ return (msiGetMessageControl(pDev) & VBOX_PCI_MSI_FLAGS_QSIZE) >> 4;
+}
+
+DECLINLINE(RTGCPHYS) msiGetMsiAddress(PPDMPCIDEV pDev)
+{
+ if (msiIs64Bit(pDev))
+ {
+ uint32_t lo = PCIDevGetDWord(pDev, pDev->Int.s.u8MsiCapOffset + VBOX_MSI_CAP_MESSAGE_ADDRESS_LO);
+ uint32_t hi = PCIDevGetDWord(pDev, pDev->Int.s.u8MsiCapOffset + VBOX_MSI_CAP_MESSAGE_ADDRESS_HI);
+ return RT_MAKE_U64(lo, hi);
+ }
+ return PCIDevGetDWord(pDev, pDev->Int.s.u8MsiCapOffset + VBOX_MSI_CAP_MESSAGE_ADDRESS_32);
+}
+
+DECLINLINE(uint32_t) msiGetMsiData(PPDMPCIDEV pDev, int32_t iVector)
+{
+ int16_t iOff = msiIs64Bit(pDev) ? VBOX_MSI_CAP_MESSAGE_DATA_64 : VBOX_MSI_CAP_MESSAGE_DATA_32;
+ uint16_t lo = PCIDevGetWord(pDev, pDev->Int.s.u8MsiCapOffset + iOff);
+
+ // vector encoding into lower bits of message data
+ uint8_t bits = msiGetMme(pDev);
+ uint16_t uMask = ((1 << bits) - 1);
+ lo &= ~uMask;
+ lo |= iVector & uMask;
+
+ return RT_MAKE_U32(lo, 0);
+}
+
+#ifdef IN_RING3
+
+DECLINLINE(bool) msiR3BitJustCleared(uint32_t uOldValue, uint32_t uNewValue, uint32_t uMask)
+{
+ return !!(uOldValue & uMask) && !(uNewValue & uMask);
+}
+
+DECLINLINE(bool) msiR3BitJustSet(uint32_t uOldValue, uint32_t uNewValue, uint32_t uMask)
+{
+ return !(uOldValue & uMask) && !!(uNewValue & uMask);
+}
+
+/**
+ * PCI config space accessors for MSI registers.
+ */
+void MsiR3PciConfigWrite(PPDMDEVINS pDevIns, PCPDMPCIHLP pPciHlp, PPDMPCIDEV pDev,
+ uint32_t u32Address, uint32_t val, unsigned len)
+{
+ int32_t iOff = u32Address - pDev->Int.s.u8MsiCapOffset;
+ Assert(iOff >= 0 && (pciDevIsMsiCapable(pDev) && iOff < pDev->Int.s.u8MsiCapSize));
+
+ Log2(("MsiR3PciConfigWrite: %d <- %x (%d)\n", iOff, val, len));
+
+ uint32_t uAddr = u32Address;
+ bool f64Bit = msiIs64Bit(pDev);
+
+ for (uint32_t i = 0; i < len; i++)
+ {
+ uint32_t reg = i + iOff;
+ uint8_t u8Val = (uint8_t)val;
+ switch (reg)
+ {
+ case 0: /* Capability ID, ro */
+ case 1: /* Next pointer, ro */
+ break;
+ case VBOX_MSI_CAP_MESSAGE_CONTROL:
+ /* don't change read-only bits: 1-3,7 */
+ u8Val &= UINT8_C(~0x8e);
+ pDev->abConfig[uAddr] = u8Val | (pDev->abConfig[uAddr] & UINT8_C(0x8e));
+ break;
+ case VBOX_MSI_CAP_MESSAGE_CONTROL + 1:
+ /* don't change read-only bit 8, and reserved 9-15 */
+ break;
+ default:
+ if (pDev->abConfig[uAddr] != u8Val)
+ {
+ int32_t maskUpdated = -1;
+
+ /* If we're enabling masked vector, and have pending messages
+ for this vector, we have to send this message now */
+ if ( !f64Bit
+ && (reg >= VBOX_MSI_CAP_MASK_BITS_32)
+ && (reg < VBOX_MSI_CAP_MASK_BITS_32 + 4)
+ )
+ {
+ maskUpdated = reg - VBOX_MSI_CAP_MASK_BITS_32;
+ }
+ if ( f64Bit
+ && (reg >= VBOX_MSI_CAP_MASK_BITS_64)
+ && (reg < VBOX_MSI_CAP_MASK_BITS_64 + 4)
+ )
+ {
+ maskUpdated = reg - VBOX_MSI_CAP_MASK_BITS_64;
+ }
+
+ if (maskUpdated != -1 && msiIsEnabled(pDev))
+ {
+ uint32_t* puPending = msiGetPendingBits(pDev);
+ for (int iBitNum = 0; iBitNum < 8; iBitNum++)
+ {
+ int32_t iBit = 1 << iBitNum;
+ uint32_t uVector = maskUpdated*8 + iBitNum;
+
+ if (msiR3BitJustCleared(pDev->abConfig[uAddr], u8Val, iBit))
+ {
+ Log(("msi: mask updated bit %d@%x (%d)\n", iBitNum, uAddr, maskUpdated));
+
+ /* To ensure that we're no longer masked */
+ pDev->abConfig[uAddr] &= ~iBit;
+ if ((*puPending & (1 << uVector)) != 0)
+ {
+ Log(("msi: notify earlier masked pending vector: %d\n", uVector));
+ MsiNotify(pDevIns, pPciHlp, pDev, uVector, PDM_IRQ_LEVEL_HIGH, 0 /*uTagSrc*/);
+ }
+ }
+ if (msiR3BitJustSet(pDev->abConfig[uAddr], u8Val, iBit))
+ {
+ Log(("msi: mask vector: %d\n", uVector));
+ }
+ }
+ }
+
+ pDev->abConfig[uAddr] = u8Val;
+ }
+ }
+ uAddr++;
+ val >>= 8;
+ }
+}
+
+/**
+ * Initializes MSI support for the given PCI device.
+ */
+int MsiR3Init(PPDMPCIDEV pDev, PPDMMSIREG pMsiReg)
+{
+ if (pMsiReg->cMsiVectors == 0)
+ return VINF_SUCCESS;
+
+ /* XXX: done in pcirawAnalyzePciCaps() */
+ if (pciDevIsPassthrough(pDev))
+ return VINF_SUCCESS;
+
+ uint16_t cVectors = pMsiReg->cMsiVectors;
+ uint8_t iCapOffset = pMsiReg->iMsiCapOffset;
+ uint8_t iNextOffset = pMsiReg->iMsiNextOffset;
+ bool f64bit = pMsiReg->fMsi64bit;
+ bool fNoMasking = pMsiReg->fMsiNoMasking;
+ uint16_t iFlags = 0;
+
+ Assert(iCapOffset != 0 && iCapOffset < 0xff && iNextOffset < 0xff);
+
+ if (!fNoMasking)
+ {
+ int iMmc;
+
+ /* Compute multiple-message capable bitfield */
+ for (iMmc = 0; iMmc < 6; iMmc++)
+ {
+ if ((1 << iMmc) >= cVectors)
+ break;
+ }
+
+ if ((cVectors > VBOX_MSI_MAX_ENTRIES) || (1 << iMmc) < cVectors)
+ return VERR_TOO_MUCH_DATA;
+
+ /* We support per-vector masking */
+ iFlags |= VBOX_PCI_MSI_FLAGS_MASKBIT;
+ /* How many vectors we're capable of */
+ iFlags |= iMmc;
+ }
+
+ if (f64bit)
+ iFlags |= VBOX_PCI_MSI_FLAGS_64BIT;
+
+ pDev->Int.s.u8MsiCapOffset = iCapOffset;
+ pDev->Int.s.u8MsiCapSize = f64bit ? VBOX_MSI_CAP_SIZE_64 : VBOX_MSI_CAP_SIZE_32;
+
+ PCIDevSetByte(pDev, iCapOffset + 0, VBOX_PCI_CAP_ID_MSI);
+ PCIDevSetByte(pDev, iCapOffset + 1, iNextOffset); /* next */
+ PCIDevSetWord(pDev, iCapOffset + VBOX_MSI_CAP_MESSAGE_CONTROL, iFlags);
+
+ if (!fNoMasking)
+ {
+ *msiGetMaskBits(pDev) = 0;
+ *msiGetPendingBits(pDev) = 0;
+ }
+
+ pciDevSetMsiCapable(pDev);
+ if (f64bit)
+ pciDevSetMsi64Capable(pDev);
+
+ return VINF_SUCCESS;
+}
+
+#endif /* IN_RING3 */
+
+
+/**
+ * Checks if MSI is enabled for the given PCI device.
+ *
+ * (Must use MSINotify() for notifications when true.)
+ */
+bool MsiIsEnabled(PPDMPCIDEV pDev)
+{
+ return pciDevIsMsiCapable(pDev) && msiIsEnabled(pDev);
+}
+
+/**
+ * Device notification (aka interrupt).
+ */
+void MsiNotify(PPDMDEVINS pDevIns, PCPDMPCIHLP pPciHlp, PPDMPCIDEV pDev, int iVector, int iLevel, uint32_t uTagSrc)
+{
+ AssertMsg(msiIsEnabled(pDev), ("Must be enabled to use that"));
+
+ uint32_t uMask;
+ uint32_t *puPending = msiGetPendingBits(pDev);
+ if (puPending)
+ {
+ uint32_t *puMask = msiGetMaskBits(pDev);
+ AssertPtr(puMask);
+ uMask = *puMask;
+ LogFlow(("MsiNotify: %d pending=%x mask=%x\n", iVector, *puPending, uMask));
+ }
+ else
+ {
+ uMask = 0;
+ LogFlow(("MsiNotify: %d\n", iVector));
+ }
+
+ /* We only trigger MSI on level up */
+ if ((iLevel & PDM_IRQ_LEVEL_HIGH) == 0)
+ {
+ /** @todo maybe clear pending interrupts on level down? */
+#if 0
+ if (puPending)
+ {
+ *puPending &= ~(1<<iVector);
+ LogFlow(("msi: clear pending %d, now %x\n", iVector, *puPending));
+ }
+#endif
+ return;
+ }
+
+ if ((uMask & (1<<iVector)) != 0)
+ {
+ *puPending |= (1<<iVector);
+ LogFlow(("msi: %d is masked, mark pending, now %x\n", iVector, *puPending));
+ return;
+ }
+
+ RTGCPHYS GCAddr = msiGetMsiAddress(pDev);
+ uint32_t u32Value = msiGetMsiData(pDev, iVector);
+
+ if (puPending)
+ *puPending &= ~(1<<iVector);
+
+ Assert(pPciHlp->pfnIoApicSendMsi != NULL);
+ pPciHlp->pfnIoApicSendMsi(pDevIns, GCAddr, u32Value, uTagSrc);
+}
+