summaryrefslogtreecommitdiffstats
path: root/sys-utils/lscpu-cputype.c
diff options
context:
space:
mode:
Diffstat (limited to 'sys-utils/lscpu-cputype.c')
-rw-r--r--sys-utils/lscpu-cputype.c911
1 files changed, 911 insertions, 0 deletions
diff --git a/sys-utils/lscpu-cputype.c b/sys-utils/lscpu-cputype.c
new file mode 100644
index 0000000..31f4009
--- /dev/null
+++ b/sys-utils/lscpu-cputype.c
@@ -0,0 +1,911 @@
+
+#include <sys/utsname.h>
+#include <sys/personality.h>
+
+#if defined(HAVE_LIBRTAS)
+# include <librtas.h>
+#endif
+
+#include "lscpu.h"
+
+#include "fileutils.h"
+#include "c_strtod.h"
+
+/* Lookup a pattern and get the value for format "<pattern> : <key>"
+ */
+int lookup(char *line, char *pattern, char **value)
+{
+ char *p, *v;
+ int len = strlen(pattern);
+
+ /* don't re-fill already found tags, first one wins */
+ if (!*line || *value)
+ return 0;
+ /* pattern */
+ if (strncmp(line, pattern, len))
+ return 0;
+ /* white spaces */
+ for (p = line + len; isspace(*p); p++);
+
+ /* separator */
+ if (*p != ':')
+ return 0;
+ /* white spaces */
+ for (++p; isspace(*p); p++);
+
+ /* value */
+ if (!*p)
+ return 0;
+ v = p;
+
+ /* end of value */
+ len = strlen(line) - 1;
+ for (p = line + len; isspace(*(p-1)); p--);
+ *p = '\0';
+
+ *value = xstrdup(v);
+ return 1;
+}
+
+struct lscpu_cputype *lscpu_new_cputype(void)
+{
+ struct lscpu_cputype *ct;
+
+ ct = xcalloc(1, sizeof(struct lscpu_cputype));
+ ct->refcount = 1;
+ ct->dispatching = -1;
+ ct->freqboost = -1;
+
+ DBG(TYPE, ul_debugobj(ct, "alloc"));
+ return ct;
+}
+
+void lscpu_ref_cputype(struct lscpu_cputype *ct)
+{
+ if (ct) {
+ ct->refcount++;
+ DBG(TYPE, ul_debugobj(ct, ">>> ref %d", ct->refcount));
+ }
+}
+
+void lscpu_unref_cputype(struct lscpu_cputype *ct)
+{
+ if (!ct)
+ return;
+
+ /*DBG(TYPE, ul_debugobj(ct, ">>> unref %d", ct->refcount - 1));*/
+
+ if (--ct->refcount <= 0) {
+ DBG(TYPE, ul_debugobj(ct, " freeing %s/%s", ct->vendor, ct->model));
+ lscpu_cputype_free_topology(ct);
+ free(ct->vendor);
+ free(ct->bios_vendor);
+ free(ct->machinetype); /* s390 */
+ free(ct->family);
+ free(ct->model);
+ free(ct->modelname);
+ free(ct->bios_modelname);
+ free(ct->bios_family);
+ free(ct->revision); /* alternative for model (ppc) */
+ free(ct->stepping);
+ free(ct->bogomips);
+ free(ct->flags);
+ free(ct->mtid); /* maximum thread id (s390) */
+ free(ct->addrsz); /* address sizes */
+ free(ct->static_mhz);
+ free(ct->dynamic_mhz);
+ free(ct);
+ }
+}
+
+struct lscpu_cputype *lscpu_cputype_get_default(struct lscpu_cxt *cxt)
+{
+ return cxt->cputypes ? cxt->cputypes[0] : NULL;
+}
+
+#define match(astr, bstr) \
+ ((!astr && !bstr) || (astr && bstr && strcmp(astr, bstr) == 0))
+
+struct lscpu_cputype *lscpu_add_cputype(struct lscpu_cxt *cxt, struct lscpu_cputype *ct)
+{
+ DBG(TYPE, ul_debugobj(ct, "add new"));
+ cxt->cputypes = xrealloc(cxt->cputypes, (cxt->ncputypes + 1)
+ * sizeof(struct lscpu_cputype *));
+ cxt->cputypes[cxt->ncputypes] = ct;
+ cxt->ncputypes++;
+ lscpu_ref_cputype(ct);
+ return ct;
+}
+
+static void fprintf_cputypes(FILE *f, struct lscpu_cxt *cxt)
+{
+ size_t i;
+
+ for (i = 0; i < cxt->ncputypes; i++) {
+ struct lscpu_cputype *ct = cxt->cputypes[i];
+
+ fprintf(f, "\n vendor: %s\n", ct->vendor);
+ fprintf(f, " machinetype: %s\n", ct->machinetype);
+ fprintf(f, " family: %s\n", ct->family);
+ fprintf(f, " model: %s\n", ct->model);
+ fprintf(f, " modelname: %s\n", ct->modelname);
+ fprintf(f, " revision: %s\n", ct->revision);
+ fprintf(f, " stepping: %s\n", ct->stepping);
+ fprintf(f, " mtid: %s\n", ct->mtid);
+ fprintf(f, " addrsz: %s\n", ct->addrsz);
+ }
+}
+
+enum {
+ CPUINFO_LINE_UNKNOWN, /* unknown line */
+ CPUINFO_LINE_CPUTYPE, /* line found in type_patterns[] */
+ CPUINFO_LINE_CPU, /* line found in cpu_patterns[] */
+ CPUINFO_LINE_CACHE /* line found in cache_pattern[] */
+};
+
+/* Describes /proc/cpuinfo fields */
+struct cpuinfo_pattern {
+ int id; /* field ID */
+ int domain; /* CPUINFO_LINE_* */
+ const char *pattern; /* field name as used in /proc/cpuinfo */
+ size_t offset; /* offset in lscpu_cputype or lscpu_cpu struct */
+};
+
+/* field identifiers (field name may be different on different archs) */
+enum {
+ PAT_ADDRESS_SIZES,
+ PAT_BOGOMIPS, /* global */
+ PAT_BOGOMIPS_CPU, /* per-cpu */
+ PAT_CPU,
+ PAT_FAMILY,
+ PAT_FEATURES,
+ PAT_FLAGS,
+ PAT_IMPLEMENTER,
+ PAT_MAX_THREAD_ID,
+ PAT_MHZ,
+ PAT_MHZ_DYNAMIC,
+ PAT_MHZ_STATIC,
+ PAT_MODEL,
+ PAT_MODEL_NAME,
+ PAT_PART,
+ PAT_PROCESSOR,
+ PAT_REVISION,
+ PAT_STEPPING,
+ PAT_TYPE,
+ PAT_VARIANT,
+ PAT_VENDOR,
+ PAT_CACHE
+};
+
+/*
+ * /proc/cpuinfo to lscpu_cputype conversion
+ */
+#define DEF_PAT_CPUTYPE(_str, _id, _member) \
+ { \
+ .id = (_id), \
+ .domain = CPUINFO_LINE_CPUTYPE, \
+ .pattern = (_str), \
+ .offset = offsetof(struct lscpu_cputype, _member), \
+ }
+
+static const struct cpuinfo_pattern type_patterns[] =
+{
+ /* Sort by fields name! */
+ DEF_PAT_CPUTYPE( "ASEs implemented", PAT_FLAGS, flags), /* mips */
+ DEF_PAT_CPUTYPE( "BogoMIPS", PAT_BOGOMIPS, bogomips), /* aarch64 */
+ DEF_PAT_CPUTYPE( "CPU implementer", PAT_IMPLEMENTER,vendor), /* ARM and aarch64 */
+ DEF_PAT_CPUTYPE( "CPU part", PAT_PART, model), /* ARM and aarch64 */
+ DEF_PAT_CPUTYPE( "CPU revision", PAT_REVISION, revision), /* aarch64 */
+ DEF_PAT_CPUTYPE( "CPU variant", PAT_VARIANT, stepping), /* aarch64 */
+ DEF_PAT_CPUTYPE( "Features", PAT_FEATURES, flags), /* aarch64 */
+ DEF_PAT_CPUTYPE( "address sizes", PAT_ADDRESS_SIZES, addrsz),/* x86 */
+ DEF_PAT_CPUTYPE( "bogomips per cpu", PAT_BOGOMIPS, bogomips), /* s390 */
+ DEF_PAT_CPUTYPE( "cpu", PAT_CPU, modelname), /* ppc, sparc */
+ DEF_PAT_CPUTYPE( "cpu family", PAT_FAMILY, family),
+ DEF_PAT_CPUTYPE( "cpu model", PAT_MODEL, model), /* mips */
+ DEF_PAT_CPUTYPE( "family", PAT_FAMILY, family),
+ DEF_PAT_CPUTYPE( "features", PAT_FEATURES, flags), /* s390 */
+ DEF_PAT_CPUTYPE( "flags", PAT_FLAGS, flags), /* x86 */
+ DEF_PAT_CPUTYPE( "max thread id", PAT_MAX_THREAD_ID, mtid), /* s390 */
+ DEF_PAT_CPUTYPE( "model", PAT_MODEL, model),
+ DEF_PAT_CPUTYPE( "model name", PAT_MODEL_NAME, modelname),
+ DEF_PAT_CPUTYPE( "revision", PAT_REVISION, revision),
+ DEF_PAT_CPUTYPE( "stepping", PAT_STEPPING, stepping),
+ DEF_PAT_CPUTYPE( "type", PAT_TYPE, flags), /* sparc64 */
+ DEF_PAT_CPUTYPE( "vendor", PAT_VENDOR, vendor),
+ DEF_PAT_CPUTYPE( "vendor_id", PAT_VENDOR, vendor), /* s390 */
+};
+
+/*
+ * /proc/cpuinfo to lscpu_cpu conversion
+ */
+#define DEF_PAT_CPU(_str, _id, _member) \
+ { \
+ .id = (_id), \
+ .domain = CPUINFO_LINE_CPU, \
+ .pattern = (_str), \
+ .offset = offsetof(struct lscpu_cpu, _member), \
+ }
+
+static const struct cpuinfo_pattern cpu_patterns[] =
+{
+ /* Sort by fields name! */
+ DEF_PAT_CPU( "bogomips", PAT_BOGOMIPS_CPU, bogomips),
+ DEF_PAT_CPU( "cpu MHz", PAT_MHZ, mhz),
+ DEF_PAT_CPU( "cpu MHz dynamic", PAT_MHZ_DYNAMIC, dynamic_mhz), /* s390 */
+ DEF_PAT_CPU( "cpu MHz static", PAT_MHZ_STATIC, static_mhz), /* s390 */
+ DEF_PAT_CPU( "cpu number", PAT_PROCESSOR, logical_id), /* s390 */
+ DEF_PAT_CPU( "processor", PAT_PROCESSOR, logical_id),
+
+};
+
+/*
+ * /proc/cpuinfo to lscpu_cache conversion
+ */
+#define DEF_PAT_CACHE(_str, _id) \
+ { \
+ .id = (_id), \
+ .domain = CPUINFO_LINE_CACHE, \
+ .pattern = (_str) \
+ }
+
+static const struct cpuinfo_pattern cache_patterns[] =
+{
+ /* Sort by fields name! */
+ DEF_PAT_CACHE("cache", PAT_CACHE),
+};
+
+#define CPUTYPE_PATTERN_BUFSZ 32
+
+static int cmp_pattern(const void *a0, const void *b0)
+{
+ const struct cpuinfo_pattern
+ *a = (const struct cpuinfo_pattern *) a0,
+ *b = (const struct cpuinfo_pattern *) b0;
+ return strcmp(a->pattern, b->pattern);
+}
+
+struct cpuinfo_parser {
+ struct lscpu_cxt *cxt;
+ struct lscpu_cpu *curr_cpu;
+ struct lscpu_cputype *curr_type;
+ unsigned int curr_type_added : 1;
+};
+
+static int is_different_cputype(struct lscpu_cputype *ct, size_t offset, const char *value)
+{
+ switch (offset) {
+ case offsetof(struct lscpu_cputype, vendor):
+ return ct->vendor && value && strcmp(ct->vendor, value) != 0;
+ case offsetof(struct lscpu_cputype, model):
+ return ct->model && value && strcmp(ct->model, value) != 0;
+ case offsetof(struct lscpu_cputype, modelname):
+ return ct->modelname && value && strcmp(ct->modelname, value) != 0;
+ case offsetof(struct lscpu_cputype, stepping):
+ return ct->stepping && value && strcmp(ct->stepping, value) != 0;
+ }
+ return 0;
+}
+
+/* canonicalize @str -- remove number at the end return the
+ * number by @keynum. This is usable for example for "processor 5" or "cache1"
+ * cpuinfo lines */
+static char *key_cleanup(char *str, int *keynum)
+{
+ size_t sz = rtrim_whitespace((unsigned char *)str);
+ size_t i;
+
+ if (!sz)
+ return str;
+
+ for (i = sz; i > 0; i--) {
+ if (!isdigit(str[i - 1]))
+ break;
+ }
+
+ if (i < sz) {
+ char *end = NULL, *p = str + i;
+ int n;
+
+ errno = 0;
+ n = strtol(p, &end, 10);
+ if (errno || !end || end == p)
+ return str;
+
+ *keynum = n;
+ str[i] = '\0';
+ rtrim_whitespace((unsigned char *)str);
+ }
+ return str;
+}
+
+static const struct cpuinfo_pattern *cpuinfo_parse_line(char *str, char **value, int *keynum)
+{
+ struct cpuinfo_pattern key = { .id = 0 }, *pat;
+ char *p, *v;
+ char buf[CPUTYPE_PATTERN_BUFSZ] = { 0 };
+
+ DBG(GATHER, ul_debug("parse \"%s\"", str));
+
+ if (!str || !*str)
+ return NULL;
+ p = (char *) skip_blank(str);
+ if (!p || !*p)
+ return NULL;
+
+ v = strchr(p, ':');
+ if (!v || !*v)
+ return NULL;
+
+ /* prepare name of the field */
+ xstrncpy(buf, p, sizeof(buf));
+ buf[v - p] = '\0';
+ v++;
+
+ /* prepare value */
+ v = (char *) skip_space(v);
+ if (!v || !*v)
+ return NULL;
+
+ key.pattern = key_cleanup(buf, keynum);
+ /* CPU-type */
+ if ((pat = bsearch(&key, type_patterns,
+ ARRAY_SIZE(type_patterns),
+ sizeof(struct cpuinfo_pattern),
+ cmp_pattern)))
+ goto found;
+
+ /* CPU */
+ if ((pat = bsearch(&key, cpu_patterns,
+ ARRAY_SIZE(cpu_patterns),
+ sizeof(struct cpuinfo_pattern),
+ cmp_pattern)))
+ goto found;
+
+ /* CACHE */
+ if ((pat = bsearch(&key, cache_patterns,
+ ARRAY_SIZE(cache_patterns),
+ sizeof(struct cpuinfo_pattern),
+ cmp_pattern)))
+ goto found;
+
+ return NULL;
+found:
+ rtrim_whitespace((unsigned char *) v);
+ *value = v;
+ return pat;
+}
+
+/* Parse extra cache lines contained within /proc/cpuinfo but which are not
+ * part of the cache topology information within the sysfs filesystem. This is
+ * true for all shared caches on e.g. s390. When there are layers of
+ * hypervisors in between it is not knows which CPUs share which caches.
+ * Therefore information about shared caches is only available in
+ * /proc/cpuinfo. Format is:
+ *
+ * cache<nr> : level=<lvl> type=<type> scope=<scope> size=<size> line_size=<lsz> associativity=<as>
+ *
+ * the cache<nr> part is parsed in cpuinfo_parse_line, in this function parses part after ":".
+ */
+static int cpuinfo_parse_cache(struct lscpu_cxt *cxt, int keynum, char *data)
+{
+ struct lscpu_cache *cache;
+ long long size;
+ char *p, type;
+ int level;
+ unsigned int line_size, associativity;
+
+ DBG(GATHER, ul_debugobj(cxt, " parse cpuinfo cache '%s'", data));
+
+ p = strstr(data, "scope=") + 6;
+ /* Skip private caches, also present in sysfs */
+ if (!p || strncmp(p, "Private", 7) == 0)
+ return 0;
+ p = strstr(data, "level=");
+ if (!p || sscanf(p, "level=%d", &level) != 1)
+ return 0;
+ p = strstr(data, "type=") + 5;
+ if (!p || !*p)
+ return 0;
+ type = 0;
+ if (strncmp(p, "Data", 4) == 0)
+ type = 'd';
+ else if (strncmp(p, "Instruction", 11) == 0)
+ type = 'i';
+ else if (strncmp(p, "Unified", 7) == 0)
+ type = 'u';
+ p = strstr(data, "size=");
+ if (!p || sscanf(p, "size=%lld", &size) != 1)
+ return 0;
+
+ p = strstr(data, "line_size=");
+ if (!p || sscanf(p, "line_size=%u", &line_size) != 1)
+ return 0;
+
+ p = strstr(data, "associativity=");
+ if (!p || sscanf(p, "associativity=%u", &associativity) != 1)
+ return 0;
+
+ cxt->necaches++;
+ cxt->ecaches = xrealloc(cxt->ecaches,
+ cxt->necaches * sizeof(struct lscpu_cache));
+ cache = &cxt->ecaches[cxt->necaches - 1];
+ memset(cache, 0 , sizeof(*cache));
+
+ if (type == 'i' || type == 'd')
+ xasprintf(&cache->name, "L%d%c", level, type);
+ else
+ xasprintf(&cache->name, "L%d", level);
+
+ cache->nth = keynum;
+ cache->level = level;
+ cache->size = size * 1024;
+ cache->ways_of_associativity = associativity;
+ cache->coherency_line_size = line_size;
+ /* Number of sets for s390. For safety, just check divide by zero */
+ cache->number_of_sets = line_size ? (cache->size / line_size): 0;
+ cache->number_of_sets = associativity ? (cache->number_of_sets / associativity) : 0;
+
+ cache->type = type == 'i' ? xstrdup("Instruction") :
+ type == 'd' ? xstrdup("Data") :
+ type == 'u' ? xstrdup("Unified") : NULL;
+ return 1;
+}
+
+int lscpu_read_cpuinfo(struct lscpu_cxt *cxt)
+{
+ FILE *fp;
+ char buf[BUFSIZ];
+ size_t i;
+ struct lscpu_cputype *ct;
+ struct cpuinfo_parser _pr = { .cxt = cxt }, *pr = &_pr;
+
+ assert(cxt->npossibles); /* lscpu_create_cpus() required */
+ assert(cxt->cpus);
+
+ DBG(GATHER, ul_debugobj(cxt, "reading cpuinfo"));
+
+ fp = ul_path_fopen(cxt->procfs, "r", "cpuinfo");
+ if (!fp)
+ err(EXIT_FAILURE, _("cannot open %s"), "/proc/cpuinfo");
+
+ do {
+ int keynum = -1;
+ char *p = NULL, *value = NULL;
+ const struct cpuinfo_pattern *pattern;
+
+ if (fgets(buf, sizeof(buf), fp) != NULL)
+ p = (char *) skip_space(buf);
+
+ if (p == NULL || (*buf && !*p)) {
+ /* Blank line separates information */
+ if (p == NULL)
+ break; /* fgets() returns nothing; EOF */
+ continue;
+ }
+
+ rtrim_whitespace((unsigned char *) buf);
+
+ /* parse */
+ pattern = cpuinfo_parse_line(p, &value, &keynum);
+ if (!pattern) {
+ DBG(GATHER, ul_debug("'%s' not found", buf));
+ continue;
+ }
+
+ /* set data */
+ switch (pattern->domain) {
+ case CPUINFO_LINE_CPU:
+ if (pattern->id == PAT_PROCESSOR) {
+ /* switch CPU */
+ int id = 0;
+
+ if (keynum >= 0)
+ id = keynum;
+ else {
+ uint32_t n;
+ if (ul_strtou32(value, &n, 10) == 0)
+ id = n;
+ }
+
+ if (pr->curr_cpu && pr->curr_type)
+ lscpu_cpu_set_type(pr->curr_cpu, pr->curr_type);
+
+ lscpu_unref_cpu(pr->curr_cpu);
+ pr->curr_cpu = lscpu_get_cpu(cxt, id);
+
+ if (!pr->curr_cpu)
+ DBG(GATHER, ul_debug("*** cpu ID '%d' undefined", id));
+ else
+ DBG(GATHER, ul_debug(" switch to CPU %d", id));
+ lscpu_ref_cpu(pr->curr_cpu);
+ break;
+ }
+ if (!pr->curr_cpu)
+ DBG(GATHER, ul_debug("*** cpu data before cpu ID"));
+ else
+ strdup_to_offset(pr->curr_cpu, pattern->offset, value);
+
+ if (pattern->id == PAT_MHZ_DYNAMIC && pr->curr_type && !pr->curr_type->dynamic_mhz)
+ pr->curr_type->dynamic_mhz = xstrdup(value);
+ if (pattern->id == PAT_MHZ_STATIC && pr->curr_type && !pr->curr_type->static_mhz)
+ pr->curr_type->static_mhz = xstrdup(value);
+ if (pattern->id == PAT_BOGOMIPS_CPU && pr->curr_type && !pr->curr_type->bogomips)
+ pr->curr_type->bogomips = xstrdup(value);
+ if (pattern->id == PAT_MHZ && pr->curr_cpu && value) {
+ errno = 0;
+ pr->curr_cpu->mhz_cur_freq = (float) c_strtod(value, NULL);
+ if (errno)
+ pr->curr_cpu->mhz_cur_freq = 0;
+ }
+ break;
+ case CPUINFO_LINE_CPUTYPE:
+ if (pr->curr_type && is_different_cputype(pr->curr_type, pattern->offset, value)) {
+ lscpu_unref_cputype(pr->curr_type);
+ pr->curr_type = NULL;
+ }
+ if (!pr->curr_type) {
+ pr->curr_type = lscpu_new_cputype();
+ lscpu_add_cputype(cxt, pr->curr_type);
+ }
+
+ strdup_to_offset(pr->curr_type, pattern->offset, value);
+ break;
+ case CPUINFO_LINE_CACHE:
+ if (pattern->id != PAT_CACHE)
+ break;
+ cpuinfo_parse_cache(cxt, keynum, value);
+ break;
+ }
+ } while (1);
+
+ DBG(GATHER, fprintf_cputypes(stderr, cxt));
+
+ if (pr->curr_cpu && !pr->curr_cpu->type)
+ lscpu_cpu_set_type(pr->curr_cpu, pr->curr_type);
+
+ lscpu_unref_cputype(pr->curr_type);
+ lscpu_unref_cpu(pr->curr_cpu);
+
+ fclose(fp);
+ lscpu_sort_caches(cxt->ecaches, cxt->necaches);
+
+ /* Set the default type to CPUs which are missing (or not parsed)
+ * in cpuinfo */
+ ct = lscpu_cputype_get_default(cxt);
+ for (i = 0; ct && i < cxt->npossibles; i++) {
+ struct lscpu_cpu *cpu = cxt->cpus[i];
+
+ if (cpu && !cpu->type)
+ lscpu_cpu_set_type(cpu, ct);
+ }
+
+ return 0;
+}
+
+struct lscpu_arch *lscpu_read_architecture(struct lscpu_cxt *cxt)
+{
+ struct utsname utsbuf;
+ struct lscpu_arch *ar;
+ struct lscpu_cputype *ct;
+
+ assert(cxt);
+
+ DBG(GATHER, ul_debug("reading architecture"));
+
+ if (uname(&utsbuf) == -1)
+ err(EXIT_FAILURE, _("error: uname failed"));
+
+ ar = xcalloc(1, sizeof(*cxt->arch));
+ ar->name = xstrdup(utsbuf.machine);
+
+ if (cxt->noalive)
+ /* reading info from any /{sys,proc} dump, don't mix it with
+ * information about our real CPU */
+ ;
+ else {
+#if defined(__alpha__) || defined(__ia64__)
+ ar->bit64 = 1; /* 64bit platforms only */
+#endif
+ /* platforms with 64bit flag in /proc/cpuinfo, define
+ * 32bit default here */
+#if defined(__i386__) || defined(__x86_64__) || \
+ defined(__s390x__) || defined(__s390__) || defined(__sparc_v9__)
+ ar->bit32 = 1;
+#endif
+
+#if defined(__aarch64__)
+ {
+ /* personality() is the most reliable way (since 4.7)
+ * to determine aarch32 support */
+ int pers = personality(PER_LINUX32);
+ if (pers != -1) {
+ personality(pers);
+ ar->bit32 = 1;
+ }
+ ar->bit64 = 1;
+ }
+#endif
+ }
+
+ ct = lscpu_cputype_get_default(cxt);
+ if (ct && ct->flags) {
+ char buf[BUFSIZ];
+
+ snprintf(buf, sizeof(buf), " %s ", ct->flags);
+ if (strstr(buf, " lm "))
+ ar->bit32 = 1, ar->bit64 = 1; /* x86_64 */
+ if (strstr(buf, " zarch "))
+ ar->bit32 = 1, ar->bit64 = 1; /* s390x */
+ if (strstr(buf, " sun4v ") || strstr(buf, " sun4u "))
+ ar->bit32 = 1, ar->bit64 = 1; /* sparc64 */
+ }
+
+ if (ar->name && !cxt->noalive) {
+ if (strcmp(ar->name, "ppc64") == 0)
+ ar->bit32 = 1, ar->bit64 = 1;
+ else if (strcmp(ar->name, "ppc") == 0)
+ ar->bit32 = 1;
+ }
+
+ DBG(GATHER, ul_debugobj(ar, "arch: name=%s %s %s",
+ ar->name,
+ ar->bit64 ? "64-bit" : "",
+ ar->bit64 ? "32-bit" : ""));
+ return ar;
+}
+
+void lscpu_free_architecture(struct lscpu_arch *ar)
+{
+ if (!ar)
+ return;
+ free(ar->name);
+ free(ar);
+}
+
+int lscpu_read_cpulists(struct lscpu_cxt *cxt)
+{
+ cpu_set_t *cpuset = NULL;
+
+ assert(cxt);
+ DBG(GATHER, ul_debugobj(cxt, "reading cpulists"));
+
+ if (ul_path_read_s32(cxt->syscpu, &cxt->maxcpus, "kernel_max") == 0)
+ /* note that kernel_max is maximum index [NR_CPUS-1] */
+ cxt->maxcpus += 1;
+
+ else if (!cxt->noalive)
+ /* the root is '/' so we are working with data from the current kernel */
+ cxt->maxcpus = get_max_number_of_cpus();
+
+ if (cxt->maxcpus <= 0)
+ /* error or we are reading some /sys snapshot instead of the
+ * real /sys, let's use any crazy number... */
+ cxt->maxcpus = 2048;
+
+ cxt->setsize = CPU_ALLOC_SIZE(cxt->maxcpus);
+
+ /* create CPUs from possible mask */
+ if (ul_path_readf_cpulist(cxt->syscpu, &cpuset, cxt->maxcpus, "possible") == 0) {
+ lscpu_create_cpus(cxt, cpuset, cxt->setsize);
+ cpuset_free(cpuset);
+ cpuset = NULL;
+ } else
+ err(EXIT_FAILURE, _("failed to determine number of CPUs: %s"),
+ _PATH_SYS_CPU "/possible");
+
+
+ /* get mask for present CPUs */
+ if (ul_path_readf_cpulist(cxt->syscpu, &cxt->present, cxt->maxcpus, "present") == 0)
+ cxt->npresents = CPU_COUNT_S(cxt->setsize, cxt->present);
+
+ /* get mask for online CPUs */
+ if (ul_path_readf_cpulist(cxt->syscpu, &cxt->online, cxt->maxcpus, "online") == 0)
+ cxt->nonlines = CPU_COUNT_S(cxt->setsize, cxt->online);
+
+ return 0;
+}
+
+#if defined(HAVE_LIBRTAS)
+# define PROCESSOR_MODULE_INFO 43
+static int strbe16toh(const char *buf, int offset)
+{
+ return (buf[offset] << 8) + buf[offset+1];
+}
+#endif
+
+/* some extra information for the default CPU type */
+int lscpu_read_archext(struct lscpu_cxt *cxt)
+{
+ FILE *f;
+ char buf[BUFSIZ];
+ struct lscpu_cputype *ct;
+
+ DBG(GATHER, ul_debugobj(cxt, "reading extra arch info"));
+
+ assert(cxt);
+ ct = lscpu_cputype_get_default(cxt);
+ if (!ct)
+ return -EINVAL;
+
+ /* get dispatching mode */
+ if (ul_path_read_s32(cxt->syscpu, &ct->dispatching, "dispatching") != 0)
+ ct->dispatching = -1;
+
+ /* get cpufreq boost mode */
+ if (ul_path_read_s32(cxt->syscpu, &ct->freqboost, "cpufreq/boost") != 0)
+ ct->freqboost = -1;
+
+ if ((f = ul_path_fopen(cxt->procfs, "r", "sysinfo"))) {
+ while (fgets(buf, sizeof(buf), f) != NULL) {
+ if (lookup(buf, "Type", &ct->machinetype))
+ break;
+ }
+ fclose(f);
+ }
+
+#if defined(HAVE_LIBRTAS)
+ /* Get PowerPC specific info */
+ if (!cxt->noalive) {
+ int rc, len, ntypes;
+
+ ct->physsockets = ct->physchips = ct->physcoresperchip = 0;
+
+ rc = rtas_get_sysparm(PROCESSOR_MODULE_INFO, sizeof(buf), buf);
+ if (rc < 0)
+ goto nortas;
+
+ len = strbe16toh(buf, 0);
+ if (len < 8)
+ goto nortas;
+
+ ntypes = strbe16toh(buf, 2);
+ if (!ntypes)
+ goto nortas;
+
+ ct->physsockets = strbe16toh(buf, 4);
+ ct->physchips = strbe16toh(buf, 6);
+ ct->physcoresperchip = strbe16toh(buf, 8);
+ }
+nortas:
+#endif
+ return 0;
+}
+
+static int cmp_vulnerability_name(const void *a0, const void *b0)
+{
+ const struct lscpu_vulnerability
+ *a = (const struct lscpu_vulnerability *) a0,
+ *b = (const struct lscpu_vulnerability *) b0;
+ return strcmp(a->name, b->name);
+}
+
+int lscpu_read_vulnerabilities(struct lscpu_cxt *cxt)
+{
+ struct dirent *d;
+ DIR *dir;
+ size_t n = 0;
+
+ assert(cxt);
+
+ DBG(GATHER, ul_debugobj(cxt, "reading vulnerabilities"));
+
+ dir = ul_path_opendir(cxt->syscpu, "vulnerabilities");
+ if (!dir)
+ return 0;
+
+ cxt->nvuls = n = 0;
+ while (xreaddir(dir))
+ n++;
+ if (!n) {
+ closedir(dir);
+ return 0;
+ }
+
+ rewinddir(dir);
+ cxt->vuls = xcalloc(n, sizeof(struct lscpu_vulnerability));
+
+ while (cxt->nvuls < n && (d = xreaddir(dir))) {
+ char *str, *p;
+ struct lscpu_vulnerability *vu;
+
+#ifdef _DIRENT_HAVE_D_TYPE
+ if (d->d_type == DT_DIR || d->d_type == DT_UNKNOWN)
+ continue;
+#endif
+ if (ul_path_readf_string(cxt->syscpu, &str,
+ "vulnerabilities/%s", d->d_name) <= 0)
+ continue;
+
+ vu = &cxt->vuls[cxt->nvuls++];
+
+ /* Name */
+ vu->name = xstrdup(d->d_name);
+ *vu->name = toupper(*vu->name);
+ strrep(vu->name, '_', ' ');
+
+ /* Description */
+ vu->text = str;
+ p = (char *) startswith(vu->text, "Mitigation");
+ if (p) {
+ *p = ';';
+ strrem(vu->text, ':');
+ }
+ }
+ closedir(dir);
+
+ qsort(cxt->vuls, cxt->nvuls,
+ sizeof(struct lscpu_vulnerability), cmp_vulnerability_name);
+
+ return 0;
+}
+
+static inline int is_node_dirent(struct dirent *d)
+{
+ return
+ d &&
+#ifdef _DIRENT_HAVE_D_TYPE
+ (d->d_type == DT_DIR || d->d_type == DT_UNKNOWN) &&
+#endif
+ strncmp(d->d_name, "node", 4) == 0 &&
+ isdigit_string(d->d_name + 4);
+}
+
+static int nodecmp(const void *ap, const void *bp)
+{
+ int *a = (int *) ap, *b = (int *) bp;
+ return *a - *b;
+}
+
+int lscpu_read_numas(struct lscpu_cxt *cxt)
+{
+ size_t i = 0;
+ DIR *dir;
+ struct dirent *d;
+ struct path_cxt *sys;
+
+ assert(!cxt->nnodes);
+
+
+ sys = ul_new_path(_PATH_SYS_NODE);
+ if (!sys)
+ err(EXIT_FAILURE, _("failed to initialize %s handler"), _PATH_SYS_NODE);
+
+ ul_path_set_prefix(sys, cxt->prefix);
+
+ dir = ul_path_opendir(sys, NULL);
+ if (!dir)
+ goto done;
+
+ while ((d = readdir(dir))) {
+ if (is_node_dirent(d))
+ cxt->nnodes++;
+ }
+
+ if (!cxt->nnodes) {
+ closedir(dir);
+ goto done;
+ }
+
+ cxt->nodemaps = xcalloc(cxt->nnodes, sizeof(cpu_set_t *));
+ cxt->idx2nodenum = xmalloc(cxt->nnodes * sizeof(int));
+
+ rewinddir(dir);
+ for (i = 0; (d = readdir(dir)) && i < cxt->nnodes;) {
+ if (is_node_dirent(d))
+ cxt->idx2nodenum[i++] = strtol_or_err(((d->d_name) + 4),
+ _("Failed to extract the node number"));
+ }
+ closedir(dir);
+ qsort(cxt->idx2nodenum, cxt->nnodes, sizeof(int), nodecmp);
+
+ /* information about how nodes share different CPUs */
+ for (i = 0; i < cxt->nnodes; i++)
+ ul_path_readf_cpuset(sys, &cxt->nodemaps[i], cxt->maxcpus,
+ "node%d/cpumap", cxt->idx2nodenum[i]);
+done:
+ DBG(GATHER, ul_debugobj(cxt, "read %zu numas", cxt->nnodes));
+
+ ul_unref_path(sys);
+ return 0;
+}