summaryrefslogtreecommitdiffstats
path: root/sys-utils/lscpu-virt.c
diff options
context:
space:
mode:
authorDaniel Baumann <daniel.baumann@progress-linux.org>2024-04-14 19:10:49 +0000
committerDaniel Baumann <daniel.baumann@progress-linux.org>2024-04-14 19:10:49 +0000
commitcfe5e3905201349e9cf3f95d52ff4bd100bde37d (patch)
treed0baf160cbee3195249d095f85e52d20c21acf02 /sys-utils/lscpu-virt.c
parentInitial commit. (diff)
downloadutil-linux-cfe5e3905201349e9cf3f95d52ff4bd100bde37d.tar.xz
util-linux-cfe5e3905201349e9cf3f95d52ff4bd100bde37d.zip
Adding upstream version 2.39.3.upstream/2.39.3
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'sys-utils/lscpu-virt.c')
-rw-r--r--sys-utils/lscpu-virt.c686
1 files changed, 686 insertions, 0 deletions
diff --git a/sys-utils/lscpu-virt.c b/sys-utils/lscpu-virt.c
new file mode 100644
index 0000000..6b6deb8
--- /dev/null
+++ b/sys-utils/lscpu-virt.c
@@ -0,0 +1,686 @@
+#include <errno.h>
+#include <stdlib.h>
+#include <sys/types.h>
+#include <sys/stat.h>
+#include <fcntl.h>
+#include <unistd.h>
+#include <string.h>
+#include <stdio.h>
+
+#include "lscpu.h"
+
+#if (defined(__x86_64__) || defined(__i386__))
+# define INCLUDE_VMWARE_BDOOR
+#endif
+
+#ifdef INCLUDE_VMWARE_BDOOR
+# include <stdint.h>
+# include <signal.h>
+# include <strings.h>
+# include <setjmp.h>
+# ifdef HAVE_SYS_IO_H
+# include <sys/io.h>
+# endif
+#endif
+
+/* Xen Domain feature flag used for /sys/hypervisor/properties/features */
+#define XENFEAT_supervisor_mode_kernel 3
+#define XENFEAT_mmu_pt_update_preserve_ad 5
+#define XENFEAT_hvm_callback_vector 8
+
+#define XEN_FEATURES_PV_MASK (1U << XENFEAT_mmu_pt_update_preserve_ad)
+#define XEN_FEATURES_PVH_MASK ( (1U << XENFEAT_supervisor_mode_kernel) \
+ | (1U << XENFEAT_hvm_callback_vector) )
+static const int hv_vendor_pci[] = {
+ [VIRT_VENDOR_NONE] = 0x0000,
+ [VIRT_VENDOR_XEN] = 0x5853,
+ [VIRT_VENDOR_KVM] = 0x0000,
+ [VIRT_VENDOR_MSHV] = 0x1414,
+ [VIRT_VENDOR_VMWARE] = 0x15ad,
+ [VIRT_VENDOR_VBOX] = 0x80ee,
+};
+
+static const int hv_graphics_pci[] = {
+ [VIRT_VENDOR_NONE] = 0x0000,
+ [VIRT_VENDOR_XEN] = 0x0001,
+ [VIRT_VENDOR_KVM] = 0x0000,
+ [VIRT_VENDOR_MSHV] = 0x5353,
+ [VIRT_VENDOR_VMWARE] = 0x0710,
+ [VIRT_VENDOR_VBOX] = 0xbeef,
+};
+
+#define WORD(x) (uint16_t)(*(const uint16_t *)(x))
+#define DWORD(x) (uint32_t)(*(const uint32_t *)(x))
+
+void *get_mem_chunk(size_t base, size_t len, const char *devmem)
+{
+ void *p = NULL;
+ int fd;
+
+ if ((fd = open(devmem, O_RDONLY)) < 0)
+ return NULL;
+
+ if (!(p = malloc(len)))
+ goto nothing;
+ if (lseek(fd, base, SEEK_SET) == -1)
+ goto nothing;
+ if (read_all(fd, p, len) == -1)
+ goto nothing;
+
+ close(fd);
+ return p;
+
+nothing:
+ free(p);
+ close(fd);
+ return NULL;
+}
+
+static int hypervisor_from_dmi_table(uint32_t base, uint16_t len,
+ uint16_t num, const char *devmem)
+{
+ uint8_t *data;
+ int rc = VIRT_VENDOR_NONE;
+ struct dmi_info di;
+
+ data = get_mem_chunk(base, len, devmem);
+ if (!data)
+ return rc;
+
+ memset(&di, 0, sizeof(struct dmi_info));
+ rc = parse_dmi_table(len, num, data, &di);
+ if (rc < 0)
+ goto done;
+
+ if (di.manufacturer && !strcmp(di.manufacturer, "innotek GmbH"))
+ rc = VIRT_VENDOR_INNOTEK;
+ else if (di.manufacturer && strstr(di.manufacturer, "HITACHI") &&
+ di.product && strstr(di.product, "LPAR"))
+ rc = VIRT_VENDOR_HITACHI;
+ else if (di.vendor && !strcmp(di.vendor, "Parallels"))
+ rc = VIRT_VENDOR_PARALLELS;
+done:
+ free(data);
+ return rc;
+}
+
+static int checksum(const uint8_t *buf, size_t len)
+{
+ uint8_t sum = 0;
+ size_t a;
+
+ for (a = 0; a < len; a++)
+ sum += buf[a];
+ return (sum == 0);
+}
+
+#if defined(__x86_64__) || defined(__i386__)
+static int hypervisor_decode_legacy(uint8_t *buf, const char *devmem)
+{
+ if (!checksum(buf, 0x0F))
+ return -1;
+
+ return hypervisor_from_dmi_table(DWORD(buf + 0x08), WORD(buf + 0x06),
+ WORD(buf + 0x0C),
+ devmem);
+}
+#endif
+
+static int hypervisor_decode_smbios(uint8_t *buf, const char *devmem)
+{
+ if (!checksum(buf, buf[0x05])
+ || memcmp(buf + 0x10, "_DMI_", 5) != 0
+ || !checksum(buf + 0x10, 0x0F))
+ return -1;
+
+ return hypervisor_from_dmi_table(DWORD(buf + 0x18), WORD(buf + 0x16),
+ WORD(buf + 0x1C),
+ devmem);
+}
+
+/*
+ * Probe for EFI interface
+ */
+#define EFI_NOT_FOUND (-1)
+#define EFI_NO_SMBIOS (-2)
+static int address_from_efi(size_t *address)
+{
+ FILE *tab;
+ char linebuf[64];
+ int ret;
+
+ *address = 0; /* Prevent compiler warning */
+
+ /*
+ * Linux up to 2.6.6: /proc/efi/systab
+ * Linux 2.6.7 and up: /sys/firmware/efi/systab
+ */
+ if (!(tab = fopen("/sys/firmware/efi/systab", "r")) &&
+ !(tab = fopen("/proc/efi/systab", "r")))
+ return EFI_NOT_FOUND; /* No EFI interface */
+
+ ret = EFI_NO_SMBIOS;
+ while ((fgets(linebuf, sizeof(linebuf) - 1, tab)) != NULL) {
+ char *addrp = strchr(linebuf, '=');
+ if (!addrp)
+ continue;
+ *(addrp++) = '\0';
+ if (strcmp(linebuf, "SMBIOS") == 0) {
+ errno = 0;
+ *address = strtoul(addrp, NULL, 0);
+ if (errno)
+ continue;
+ ret = 0;
+ break;
+ }
+ }
+
+ fclose(tab);
+ return ret;
+}
+
+static int read_hypervisor_dmi_from_devmem(void)
+{
+ int rc = VIRT_VENDOR_NONE;
+ uint8_t *buf = NULL;
+ size_t fp = 0;
+
+ /* First try EFI (ia64, Intel-based Mac) */
+ switch (address_from_efi(&fp)) {
+ case EFI_NOT_FOUND:
+ goto memory_scan;
+ case EFI_NO_SMBIOS:
+ goto done;
+ }
+
+ buf = get_mem_chunk(fp, 0x20, _PATH_DEV_MEM);
+ if (!buf)
+ goto done;
+
+ rc = hypervisor_decode_smbios(buf, _PATH_DEV_MEM);
+ if (rc >= VIRT_VENDOR_NONE)
+ goto done;
+
+ free(buf);
+ buf = NULL;
+memory_scan:
+#if defined(__x86_64__) || defined(__i386__)
+ /* Fallback to memory scan (x86, x86_64) */
+ buf = get_mem_chunk(0xF0000, 0x10000, _PATH_DEV_MEM);
+ if (!buf)
+ goto done;
+
+ for (fp = 0; fp <= 0xFFF0; fp += 16) {
+ if (memcmp(buf + fp, "_SM_", 4) == 0 && fp <= 0xFFE0) {
+ rc = hypervisor_decode_smbios(buf + fp, _PATH_DEV_MEM);
+ if (rc < 0)
+ fp += 16;
+
+ } else if (memcmp(buf + fp, "_DMI_", 5) == 0)
+ rc = hypervisor_decode_legacy(buf + fp, _PATH_DEV_MEM);
+
+ if (rc >= VIRT_VENDOR_NONE)
+ break;
+ }
+#endif
+done:
+ free(buf);
+ return rc;
+}
+
+static int read_hypervisor_dmi_from_sysfw(void)
+{
+ static char const sys_fw_dmi_tables[] = _PATH_SYS_DMI;
+ struct stat st;
+
+ if (stat(sys_fw_dmi_tables, &st))
+ return -1;
+
+ return hypervisor_from_dmi_table(0, st.st_size, st.st_size / 4,
+ sys_fw_dmi_tables);
+}
+
+static int read_hypervisor_dmi(void)
+{
+ int rc;
+
+ if (sizeof(uint8_t) != 1
+ || sizeof(uint16_t) != 2
+ || sizeof(uint32_t) != 4
+ || '\0' != 0)
+ return VIRT_VENDOR_NONE;
+
+ /* -1 : no DMI in /sys,
+ * 0 : DMI exist, nothing detected (VIRT_VENDOR_NONE)
+ * >0 : hypervisor detected
+ */
+ rc = read_hypervisor_dmi_from_sysfw();
+ if (rc < 0)
+ rc = read_hypervisor_dmi_from_devmem();
+
+ return rc < 0 ? VIRT_VENDOR_NONE : rc;
+}
+
+static int has_pci_device(struct lscpu_cxt *cxt,
+ unsigned int vendor, unsigned int device)
+{
+ FILE *f;
+ unsigned int num, fn, ven, dev;
+ int res = 1;
+
+ f = ul_path_fopen(cxt->procfs, "r", "bus/pci/devices");
+ if (!f)
+ return 0;
+
+ /* for more details about bus/pci/devices format see
+ * drivers/pci/proc.c in linux kernel
+ */
+ while(fscanf(f, "%02x%02x\t%04x%04x\t%*[^\n]",
+ &num, &fn, &ven, &dev) == 4) {
+
+ if (ven == vendor && dev == device)
+ goto found;
+ }
+
+ res = 0;
+found:
+ fclose(f);
+ return res;
+}
+
+#if defined(__x86_64__) || defined(__i386__)
+/*
+ * This CPUID leaf returns the information about the hypervisor.
+ * EAX : maximum input value for CPUID supported by the hypervisor.
+ * EBX, ECX, EDX : Hypervisor vendor ID signature. E.g. VMwareVMware.
+ */
+#define HYPERVISOR_INFO_LEAF 0x40000000
+
+static inline void cpuid(unsigned int op, unsigned int *eax,
+ unsigned int *ebx, unsigned int *ecx,
+ unsigned int *edx)
+{
+ __asm__(
+#if defined(__PIC__) && defined(__i386__)
+ /* x86 PIC cannot clobber ebx -- gcc bitches */
+ "xchg %%ebx, %%esi;"
+ "cpuid;"
+ "xchg %%esi, %%ebx;"
+ : "=S" (*ebx),
+#else
+ "cpuid;"
+ : "=b" (*ebx),
+#endif
+ "=a" (*eax),
+ "=c" (*ecx),
+ "=d" (*edx)
+ : "1" (op), "c"(0));
+}
+
+static int read_hypervisor_cpuid(void)
+{
+ unsigned int eax = 0, ebx = 0, ecx = 0, edx = 0;
+ char hyper_vendor_id[13] = { 0 };
+
+ cpuid(HYPERVISOR_INFO_LEAF, &eax, &ebx, &ecx, &edx);
+ memcpy(hyper_vendor_id + 0, &ebx, 4);
+ memcpy(hyper_vendor_id + 4, &ecx, 4);
+ memcpy(hyper_vendor_id + 8, &edx, 4);
+ hyper_vendor_id[12] = '\0';
+
+ if (!hyper_vendor_id[0])
+ goto none;
+
+ if (!strncmp("XenVMMXenVMM", hyper_vendor_id, 12))
+ return VIRT_VENDOR_XEN;
+ else if (!strncmp("KVMKVMKVM", hyper_vendor_id, 9))
+ return VIRT_VENDOR_KVM;
+ else if (!strncmp("Microsoft Hv", hyper_vendor_id, 12))
+ return VIRT_VENDOR_MSHV;
+ else if (!strncmp("VMwareVMware", hyper_vendor_id, 12))
+ return VIRT_VENDOR_VMWARE;
+ else if (!strncmp("UnisysSpar64", hyper_vendor_id, 12))
+ return VIRT_VENDOR_SPAR;
+none:
+ return VIRT_VENDOR_NONE;
+}
+
+#else /* ! (__x86_64__ || __i386__) */
+static int read_hypervisor_cpuid(void)
+{
+ return 0;
+}
+#endif
+
+static int is_devtree_compatible(struct lscpu_cxt *cxt, const char *str)
+{
+ FILE *fd = ul_path_fopen(cxt->procfs, "r", "device-tree/compatible");
+
+ if (fd) {
+ char buf[256];
+ size_t i, len;
+
+ memset(buf, 0, sizeof(buf));
+ len = fread(buf, 1, sizeof(buf) - 1, fd);
+ fclose(fd);
+
+ for (i = 0; i < len;) {
+ if (!strcmp(&buf[i], str))
+ return 1;
+ i += strlen(&buf[i]);
+ i++;
+ }
+ }
+
+ return 0;
+}
+
+static int read_hypervisor_powerpc(struct lscpu_cxt *cxt, int *type)
+{
+ int vendor = VIRT_VENDOR_NONE;
+
+ *type = VIRT_TYPE_NONE;
+
+ /* IBM iSeries: legacy, para-virtualized on top of OS/400 */
+ if (ul_path_access(cxt->procfs, F_OK, "iSeries") == 0) {
+ vendor = VIRT_VENDOR_OS400;
+ *type = VIRT_TYPE_PARA;
+
+ /* PowerNV (POWER Non-Virtualized, bare-metal) */
+ } else if (is_devtree_compatible(cxt, "ibm,powernv") != 0) {
+ ;
+
+ /* PowerVM (IBM's proprietary hypervisor, aka pHyp) */
+ } else if (ul_path_access(cxt->procfs, F_OK, "device-tree/ibm,partition-name") == 0
+ && ul_path_access(cxt->procfs, F_OK, "device-tree/hmc-managed?") == 0
+ && ul_path_access(cxt->procfs, F_OK, "device-tree/chosen/qemu,graphic-width") != 0) {
+
+ FILE *fd;
+ vendor = VIRT_VENDOR_PHYP;
+ *type = VIRT_TYPE_PARA;
+
+ fd = ul_path_fopen(cxt->procfs, "r", "device-tree/ibm,partition-name");
+ if (fd) {
+ char buf[256];
+ if (fscanf(fd, "%255s", buf) == 1 && !strcmp(buf, "full"))
+ *type = VIRT_TYPE_NONE;
+ fclose(fd);
+ }
+
+ /* Qemu */
+ } else if (is_devtree_compatible(cxt, "qemu,pseries")) {
+ vendor = VIRT_VENDOR_KVM;
+ *type = VIRT_TYPE_PARA;
+ }
+
+ return vendor;
+}
+
+#ifdef INCLUDE_VMWARE_BDOOR
+
+#define VMWARE_BDOOR_MAGIC 0x564D5868
+#define VMWARE_BDOOR_PORT 0x5658
+#define VMWARE_BDOOR_CMD_GETVERSION 10
+
+static UL_ASAN_BLACKLIST
+void vmware_bdoor(uint32_t *eax, uint32_t *ebx, uint32_t *ecx, uint32_t *edx)
+{
+ __asm__(
+#if defined(__PIC__) && defined(__i386__)
+ /* x86 PIC cannot clobber ebx -- gcc bitches */
+ "xchg %%ebx, %%esi;"
+ "inl (%%dx), %%eax;"
+ "xchg %%esi, %%ebx;"
+ : "=S" (*ebx),
+#else
+ "inl (%%dx), %%eax;"
+ : "=b" (*ebx),
+#endif
+ "=a" (*eax),
+ "=c" (*ecx),
+ "=d" (*edx)
+ : "0" (VMWARE_BDOOR_MAGIC),
+ "1" (VMWARE_BDOOR_CMD_GETVERSION),
+ "2" (VMWARE_BDOOR_PORT),
+ "3" (0)
+ : "memory");
+}
+
+static jmp_buf segv_handler_env;
+
+static void
+segv_handler(__attribute__((__unused__)) int sig,
+ __attribute__((__unused__)) siginfo_t *info,
+ __attribute__((__unused__)) void *ignored)
+{
+ siglongjmp(segv_handler_env, 1);
+}
+
+static int is_vmware_platform(void)
+{
+ uint32_t eax, ebx, ecx, edx;
+ struct sigaction act, oact;
+
+ /*
+ * FIXME: Not reliable for non-root users. Note it works as expected if
+ * vmware_bdoor() is not optimized for PIE, but then it fails to build
+ * on 32bit x86 systems. See lscpu git log for more details (commit
+ * 7845b91dbc7690064a2be6df690e4aaba728fb04). kzak [3-Nov-2016]
+ */
+ if (getuid() != 0)
+ return 0;
+
+ /*
+ * The assembly routine for vmware detection works
+ * fine under vmware, even if ran as regular user. But
+ * on real HW or under other hypervisors, it segfaults (which is
+ * expected). So we temporarily install SIGSEGV handler to catch
+ * the signal. All this magic is needed because lscpu
+ * isn't supposed to require root privileges.
+ */
+ if (sigsetjmp(segv_handler_env, 1))
+ return 0;
+
+ memset(&act, 0, sizeof(act));
+ act.sa_sigaction = segv_handler;
+ act.sa_flags = SA_SIGINFO;
+
+ if (sigaction(SIGSEGV, &act, &oact))
+ err(EXIT_FAILURE, _("cannot set signal handler"));
+
+ vmware_bdoor(&eax, &ebx, &ecx, &edx);
+
+ if (sigaction(SIGSEGV, &oact, NULL))
+ err(EXIT_FAILURE, _("cannot restore signal handler"));
+
+ return eax != (uint32_t)-1 && ebx == VMWARE_BDOOR_MAGIC;
+}
+
+#else /* ! INCLUDE_VMWARE_BDOOR */
+
+static int is_vmware_platform(void)
+{
+ return 0;
+}
+
+#endif /* INCLUDE_VMWARE_BDOOR */
+struct lscpu_virt *lscpu_read_virtualization(struct lscpu_cxt *cxt)
+{
+ char buf[BUFSIZ];
+ struct lscpu_cputype *ct;
+ struct lscpu_virt *virt;
+ FILE *fd;
+
+ DBG(VIRT, ul_debug("reading virtualization"));
+ virt = xcalloc(1, sizeof(*virt));
+
+ /* CPU flags */
+ ct = lscpu_cputype_get_default(cxt);
+ if (ct && ct->flags) {
+ snprintf(buf, sizeof(buf), " %s ", ct->flags);
+ if (strstr(buf, " svm "))
+ virt->cpuflag = xstrdup("svm");
+ else if (strstr(buf, " vmx "))
+ virt->cpuflag = xstrdup("vmx");
+ }
+
+
+ /* We have to detect WSL first. is_vmware_platform() crashes on Windows 10. */
+ fd = ul_path_fopen(cxt->procfs, "r", "sys/kernel/osrelease");
+ if (fd) {
+ if (fgets(buf, sizeof(buf), fd) && strstr(buf, "Microsoft")) {
+ virt->vendor = VIRT_VENDOR_WSL;
+ virt->type = VIRT_TYPE_CONTAINER;
+ }
+ fclose(fd);
+ if (virt->type)
+ goto done;
+ }
+
+ if (!cxt->noalive) {
+ virt->vendor = read_hypervisor_cpuid();
+ if (!virt->vendor)
+ virt->vendor = read_hypervisor_dmi();
+ if (!virt->vendor && is_vmware_platform())
+ virt->vendor = VIRT_VENDOR_VMWARE;
+ }
+
+ if (virt->vendor) {
+ virt->type = VIRT_TYPE_FULL;
+
+ if (virt->vendor == VIRT_VENDOR_XEN) {
+ uint32_t features;
+
+ fd = ul_prefix_fopen(cxt->prefix, "r", _PATH_SYS_HYP_FEATURES);
+
+ if (fd && fscanf(fd, "%x", &features) == 1) {
+ /* Xen PV domain */
+ if (features & XEN_FEATURES_PV_MASK)
+ virt->type = VIRT_TYPE_PARA;
+ /* Xen PVH domain */
+ else if ((features & XEN_FEATURES_PVH_MASK)
+ == XEN_FEATURES_PVH_MASK)
+ virt->type = VIRT_TYPE_PARA;
+ }
+ if (fd)
+ fclose(fd);
+ }
+ } else if ((virt->vendor = read_hypervisor_powerpc(cxt, &virt->type))) {
+ ;
+
+ /* Xen para-virt or dom0 */
+ } else if (ul_path_access(cxt->procfs, F_OK, "xen") == 0) {
+ int dom0 = 0;
+
+ fd = ul_path_fopen(cxt->procfs, "r", "xen/capabilities");
+ if (fd) {
+ char xenbuf[256];
+
+ if (fscanf(fd, "%255s", xenbuf) == 1 &&
+ !strcmp(xenbuf, "control_d"))
+ dom0 = 1;
+ fclose(fd);
+ }
+ virt->type = dom0 ? VIRT_TYPE_NONE : VIRT_TYPE_PARA;
+ virt->vendor = VIRT_VENDOR_XEN;
+
+ /* Xen full-virt on non-x86_64 */
+ } else if (has_pci_device(cxt, hv_vendor_pci[VIRT_VENDOR_XEN], hv_graphics_pci[VIRT_VENDOR_XEN])) {
+ virt->vendor = VIRT_VENDOR_XEN;
+ virt->type = VIRT_TYPE_FULL;
+ } else if (has_pci_device(cxt, hv_vendor_pci[VIRT_VENDOR_VMWARE], hv_graphics_pci[VIRT_VENDOR_VMWARE])) {
+ virt->vendor = VIRT_VENDOR_VMWARE;
+ virt->type = VIRT_TYPE_FULL;
+ } else if (has_pci_device(cxt, hv_vendor_pci[VIRT_VENDOR_VBOX], hv_graphics_pci[VIRT_VENDOR_VBOX])) {
+ virt->vendor = VIRT_VENDOR_VBOX;
+ virt->type = VIRT_TYPE_FULL;
+
+ /* IBM PR/SM */
+ } else if ((fd = ul_path_fopen(cxt->procfs, "r", "sysinfo"))) {
+
+ virt->vendor = VIRT_VENDOR_IBM;
+ virt->hypervisor = "PR/SM";
+ virt->type = VIRT_TYPE_FULL;
+
+ while (fgets(buf, sizeof(buf), fd) != NULL) {
+ if (!strstr(buf, "Control Program:"))
+ continue;
+ virt->vendor = strstr(buf, "KVM") ? VIRT_VENDOR_KVM : VIRT_VENDOR_IBM;
+ virt->hypervisor = strchr(buf, ':');
+
+ if (virt->hypervisor) {
+ virt->hypervisor++;
+ normalize_whitespace((unsigned char *) virt->hypervisor);
+ break;
+ }
+ }
+ if (virt->hypervisor)
+ virt->hypervisor = xstrdup(virt->hypervisor);
+ fclose(fd);
+ }
+
+ /* OpenVZ/Virtuozzo - /proc/vz dir should exist
+ * /proc/bc should not */
+ else if (ul_path_access(cxt->procfs, F_OK, "vz") == 0 &&
+ ul_path_access(cxt->procfs, F_OK, "bc") != 0) {
+ virt->vendor = VIRT_VENDOR_PARALLELS;
+ virt->type = VIRT_TYPE_CONTAINER;
+
+ /* IBM */
+ } else if (virt->hypervisor &&
+ (strcmp(virt->hypervisor, "PowerVM Lx86") == 0 ||
+ strcmp(virt->hypervisor, "IBM/S390") == 0)) {
+ virt->vendor = VIRT_VENDOR_IBM;
+ virt->type = VIRT_TYPE_FULL;
+
+ /* User-mode-linux */
+ } else if (ct && ct->modelname && strstr(ct->modelname, "UML")) {
+ virt->vendor = VIRT_VENDOR_UML;
+ virt->type = VIRT_TYPE_PARA;
+
+ /* Linux-VServer */
+ } else if ((fd = ul_path_fopen(cxt->procfs, "r", "self/status"))) {
+ char *val = NULL;
+
+ while (fgets(buf, sizeof(buf), fd) != NULL) {
+ if (lookup(buf, "VxID", &val))
+ break;
+ }
+ fclose(fd);
+
+ if (val) {
+ char *org = val;
+
+ while (isdigit(*val))
+ ++val;
+ if (!*val) {
+ virt->vendor = VIRT_VENDOR_VSERVER;
+ virt->type = VIRT_TYPE_CONTAINER;
+ }
+ free(org);
+ }
+ }
+done:
+ DBG(VIRT, ul_debugobj(virt, "virt: cpu='%s' hypervisor='%s' vendor=%d type=%d",
+ virt->cpuflag,
+ virt->hypervisor,
+ virt->vendor,
+ virt->type));
+
+ if (!virt->cpuflag && !virt->hypervisor && !virt->vendor && !virt->type) {
+ lscpu_free_virtualization(virt);
+ virt = NULL;
+ }
+ return virt;
+}
+
+void lscpu_free_virtualization(struct lscpu_virt *virt)
+{
+ if (!virt)
+ return;
+
+ free(virt->cpuflag);
+ free(virt->hypervisor);
+ free(virt);
+}
+