/*-
* BSD LICENSE
- *
- * Copyright(c) 2010-2013 Intel Corporation. All rights reserved.
+ *
+ * Copyright(c) 2010-2016 Intel Corporation. All rights reserved.
+ * Copyright(c) 2012-2014 6WIND S.A.
* All rights reserved.
- *
+ *
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
- *
+ *
* * Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* * Redistributions in binary form must reproduce the above copyright
* * Neither the name of Intel Corporation nor the names of its
* contributors may be used to endorse or promote products derived
* from this software without specific prior written permission.
- *
+ *
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
#include <syslog.h>
#include <getopt.h>
#include <sys/file.h>
+#include <fcntl.h>
#include <stddef.h>
#include <errno.h>
#include <limits.h>
#include <errno.h>
#include <sys/mman.h>
#include <sys/queue.h>
+#include <sys/stat.h>
+#if defined(RTE_ARCH_X86)
#include <sys/io.h>
-#include <sys/user.h>
-#include <linux/binfmts.h>
+#endif
#include <rte_common.h>
#include <rte_debug.h>
#include <rte_memory.h>
#include <rte_memzone.h>
#include <rte_launch.h>
-#include <rte_tailq.h>
#include <rte_eal.h>
#include <rte_eal_memconfig.h>
+#include <rte_errno.h>
#include <rte_per_lcore.h>
#include <rte_lcore.h>
#include <rte_log.h>
#include <rte_string_fns.h>
#include <rte_cpuflags.h>
#include <rte_interrupts.h>
+#include <rte_bus.h>
#include <rte_pci.h>
+#include <rte_dev.h>
+#include <rte_devargs.h>
#include <rte_common.h>
#include <rte_version.h>
#include <rte_atomic.h>
#include <malloc_heap.h>
-#include <rte_eth_ring.h>
#include "eal_private.h"
#include "eal_thread.h"
#include "eal_internal_cfg.h"
#include "eal_filesystem.h"
#include "eal_hugepages.h"
-
-#define OPT_HUGE_DIR "huge-dir"
-#define OPT_PROC_TYPE "proc-type"
-#define OPT_NO_SHCONF "no-shconf"
-#define OPT_NO_HPET "no-hpet"
-#define OPT_VMWARE_TSC_MAP "vmware-tsc-map"
-#define OPT_NO_PCI "no-pci"
-#define OPT_NO_HUGE "no-huge"
-#define OPT_FILE_PREFIX "file-prefix"
-#define OPT_SOCKET_MEM "socket-mem"
-#define OPT_USE_DEVICE "use-device"
-#define OPT_SYSLOG "syslog"
-
-#define RTE_EAL_BLACKLIST_SIZE 0x100
+#include "eal_options.h"
+#include "eal_vfio.h"
#define MEMSIZE_IF_NO_HUGE_PAGE (64ULL * 1024ULL * 1024ULL)
#define SOCKET_MEM_STRLEN (RTE_MAX_NUMA_NODES * 10)
-#define HIGHEST_RPL 3
-
-#define BITS_PER_HEX 4
-
-#define GET_BLACKLIST_FIELD(in, fd, lim, dlm) \
-{ \
- unsigned long val; \
- char *end; \
- errno = 0; \
- val = strtoul((in), &end, 16); \
- if (errno != 0 || end[0] != (dlm) || val > (lim)) \
- return (-EINVAL); \
- (fd) = (typeof (fd))val; \
- (in) = end + 1; \
-}
-
/* Allow the application to print its usage message too if set */
static rte_usage_hook_t rte_application_usage_hook = NULL;
+
/* early configuration structure, when memory config is not mmapped */
static struct rte_mem_config early_mem_config;
.mem_config = &early_mem_config,
};
-static struct rte_pci_addr eal_dev_blacklist[RTE_EAL_BLACKLIST_SIZE];
-
/* internal configuration (per-core) */
struct lcore_config lcore_config[RTE_MAX_LCORE];
if (internal_config.no_shconf)
return;
+ /* map the config before hugepage address so that we don't waste a page */
+ if (internal_config.base_virtaddr != 0)
+ rte_mem_cfg_addr = (void *)
+ RTE_ALIGN_FLOOR(internal_config.base_virtaddr -
+ sizeof(struct rte_mem_config), sysconf(_SC_PAGE_SIZE));
+ else
+ rte_mem_cfg_addr = NULL;
+
if (mem_cfg_fd < 0){
mem_cfg_fd = open(pathname, O_RDWR | O_CREAT, 0660);
if (mem_cfg_fd < 0)
"process running?\n", pathname);
}
- rte_mem_cfg_addr = mmap(NULL, sizeof(*rte_config.mem_config),
+ rte_mem_cfg_addr = mmap(rte_mem_cfg_addr, sizeof(*rte_config.mem_config),
PROT_READ | PROT_WRITE, MAP_SHARED, mem_cfg_fd, 0);
if (rte_mem_cfg_addr == MAP_FAILED){
rte_panic("Cannot mmap memory for rte_config\n");
}
memcpy(rte_mem_cfg_addr, &early_mem_config, sizeof(early_mem_config));
- rte_config.mem_config = (struct rte_mem_config *) rte_mem_cfg_addr;
+ rte_config.mem_config = rte_mem_cfg_addr;
+
+ /* store address of the config in the config itself so that secondary
+ * processes could later map the config into this exact location */
+ rte_config.mem_config->mem_cfg_addr = (uintptr_t) rte_mem_cfg_addr;
+
}
/* attach to an existing shared memory config */
static void
rte_eal_config_attach(void)
{
- void *rte_mem_cfg_addr;
+ struct rte_mem_config *mem_config;
+
const char *pathname = eal_runtime_config_path();
if (internal_config.no_shconf)
rte_panic("Cannot open '%s' for rte_mem_config\n", pathname);
}
- rte_mem_cfg_addr = mmap(NULL, sizeof(*rte_config.mem_config),
- PROT_READ | PROT_WRITE, MAP_SHARED, mem_cfg_fd, 0);
+ /* map it as read-only first */
+ mem_config = (struct rte_mem_config *) mmap(NULL, sizeof(*mem_config),
+ PROT_READ, MAP_SHARED, mem_cfg_fd, 0);
+ if (mem_config == MAP_FAILED)
+ rte_panic("Cannot mmap memory for rte_config! error %i (%s)\n",
+ errno, strerror(errno));
+
+ rte_config.mem_config = mem_config;
+}
+
+/* reattach the shared config at exact memory location primary process has it */
+static void
+rte_eal_config_reattach(void)
+{
+ struct rte_mem_config *mem_config;
+ void *rte_mem_cfg_addr;
+
+ if (internal_config.no_shconf)
+ return;
+
+ /* save the address primary process has mapped shared config to */
+ rte_mem_cfg_addr = (void *) (uintptr_t) rte_config.mem_config->mem_cfg_addr;
+
+ /* unmap original config */
+ munmap(rte_config.mem_config, sizeof(struct rte_mem_config));
+
+ /* remap the config at proper address */
+ mem_config = (struct rte_mem_config *) mmap(rte_mem_cfg_addr,
+ sizeof(*mem_config), PROT_READ | PROT_WRITE, MAP_SHARED,
+ mem_cfg_fd, 0);
+ if (mem_config == MAP_FAILED || mem_config != rte_mem_cfg_addr) {
+ if (mem_config != MAP_FAILED)
+ /* errno is stale, don't use */
+ rte_panic("Cannot mmap memory for rte_config at [%p], got [%p]"
+ " - please use '--base-virtaddr' option\n",
+ rte_mem_cfg_addr, mem_config);
+ else
+ rte_panic("Cannot mmap memory for rte_config! error %i (%s)\n",
+ errno, strerror(errno));
+ }
close(mem_cfg_fd);
- if (rte_mem_cfg_addr == MAP_FAILED)
- rte_panic("Cannot mmap memory for rte_config\n");
- rte_config.mem_config = (struct rte_mem_config *) rte_mem_cfg_addr;
+ rte_config.mem_config = mem_config;
}
/* Detect if we are a primary or a secondary process */
-static enum rte_proc_type_t
+enum rte_proc_type_t
eal_proc_type_detect(void)
{
enum rte_proc_type_t ptype = RTE_PROC_PRIMARY;
static void
rte_config_init(void)
{
- /* set the magic in configuration structure */
- rte_config.magic = RTE_MAGIC;
- rte_config.process_type = (internal_config.process_type == RTE_PROC_AUTO) ?
- eal_proc_type_detect() : /* for auto, detect the type */
- internal_config.process_type; /* otherwise use what's already set */
+ rte_config.process_type = internal_config.process_type;
switch (rte_config.process_type){
case RTE_PROC_PRIMARY:
case RTE_PROC_SECONDARY:
rte_eal_config_attach();
rte_eal_mcfg_wait_complete(rte_config.mem_config);
+ rte_eal_config_reattach();
break;
case RTE_PROC_AUTO:
case RTE_PROC_INVALID:
for (i = 0; i < MAX_HUGEPAGE_SIZES; i++)
{
/* skip uninitialized */
- if (internal_config.hugepage_info[i].lock_descriptor == 0)
+ if (internal_config.hugepage_info[i].lock_descriptor < 0)
continue;
/* unlock hugepage file */
flock(internal_config.hugepage_info[i].lock_descriptor, LOCK_UN);
close(internal_config.hugepage_info[i].lock_descriptor);
/* reset the field */
- internal_config.hugepage_info[i].lock_descriptor = 0;
+ internal_config.hugepage_info[i].lock_descriptor = -1;
}
}
static void
eal_usage(const char *prgname)
{
- printf("\nUsage: %s -c COREMASK -n NUM [-m NB] [-r NUM] [-b <domain:bus:devid.func>]"
- "[--proc-type primary|secondary|auto] \n\n"
- "EAL options:\n"
- " -c COREMASK : A hexadecimal bitmask of cores to run on\n"
- " -n NUM : Number of memory channels\n"
- " -v : Display version information on startup\n"
- " -b <domain:bus:devid.func>: to prevent EAL from using specified "
- "PCI device\n"
- " (multiple -b options are allowed)\n"
- " -m MB : memory to allocate (see also --"OPT_SOCKET_MEM")\n"
- " -r NUM : force number of memory ranks (don't detect)\n"
- " --"OPT_SYSLOG" : set syslog facility\n"
- " --"OPT_SOCKET_MEM" : memory to allocate on specific \n"
- " sockets (use comma separated values)\n"
- " --"OPT_HUGE_DIR" : directory where hugetlbfs is mounted\n"
- " --"OPT_PROC_TYPE" : type of this process\n"
- " --"OPT_FILE_PREFIX": prefix for hugepage filenames\n"
- " --"OPT_USE_DEVICE": use the specified ethernet device(s) only."
- "Use comma-separate <[domain:]bus:devid.func> values.\n"
- " [NOTE: Cannot be used with -b option]\n"
- " --"OPT_VMWARE_TSC_MAP": use VMware TSC map instead of "
- "native RDTSC\n"
- "\nEAL options for DEBUG use only:\n"
- " --"OPT_NO_HUGE" : use malloc instead of hugetlbfs\n"
- " --"OPT_NO_PCI" : disable pci\n"
- " --"OPT_NO_HPET" : disable hpet\n"
- " --"OPT_NO_SHCONF": no shared config (mmap'd files)\n"
- "\n",
- prgname);
+ printf("\nUsage: %s ", prgname);
+ eal_common_usage();
+ printf("EAL Linux options:\n"
+ " --"OPT_SOCKET_MEM" Memory to allocate on sockets (comma separated values)\n"
+ " --"OPT_HUGE_DIR" Directory where hugetlbfs is mounted\n"
+ " --"OPT_FILE_PREFIX" Prefix for hugepage filenames\n"
+ " --"OPT_BASE_VIRTADDR" Base virtual address\n"
+ " --"OPT_CREATE_UIO_DEV" Create /dev/uioX (usually done by hotplug)\n"
+ " --"OPT_VFIO_INTR" Interrupt mode for VFIO (legacy|msi|msix)\n"
+ " --"OPT_XEN_DOM0" Support running on Xen dom0 without hugetlbfs\n"
+ "\n");
/* Allow the application to print its usage message too if hook is set */
if ( rte_application_usage_hook ) {
printf("===== Application Usage =====\n\n");
return old_func;
}
-/*
- * Parse the coremask given as argument (hexadecimal string) and fill
- * the global configuration (core role and core count) with the parsed
- * value.
- */
-static int xdigit2val(unsigned char c)
-{
- int val;
- if(isdigit(c))
- val = c - '0';
- else if(isupper(c))
- val = c - 'A' + 10;
- else
- val = c - 'a' + 10;
- return val;
-}
-static int
-eal_parse_coremask(const char *coremask)
-{
- struct rte_config *cfg = rte_eal_get_configuration();
- int i, j, idx = 0 ;
- unsigned count = 0;
- char c;
- int val;
-
- if (coremask == NULL)
- return -1;
- /* Remove all blank characters ahead and after .
- * Remove 0x/0X if exists.
- */
- while (isblank(*coremask))
- coremask++;
- if (coremask[0] == '0' && ((coremask[1] == 'x')
- || (coremask[1] == 'X')) )
- coremask += 2;
- i = strnlen(coremask, MAX_ARG_STRLEN);
- while ((i > 0) && isblank(coremask[i - 1]))
- i--;
- if (i == 0)
- return -1;
-
- for (i = i - 1; i >= 0 && idx < RTE_MAX_LCORE; i--) {
- c = coremask[i];
- if (isxdigit(c) == 0) {
- /* invalid characters */
- return (-1);
- }
- val = xdigit2val(c);
- for(j = 0; j < BITS_PER_HEX && idx < RTE_MAX_LCORE; j++, idx++) {
- if((1 << j) & val) {
- cfg->lcore_role[idx] = ROLE_RTE;
- if(count == 0)
- cfg->master_lcore = idx;
- count++;
- } else {
- cfg->lcore_role[idx] = ROLE_OFF;
- }
- }
- }
- for(; i >= 0; i--)
- if(coremask[i] != '0')
- return -1;
- for(; idx < RTE_MAX_LCORE; idx++)
- cfg->lcore_role[idx] = ROLE_OFF;
- if(count == 0)
- return -1;
- return 0;
-}
-
-static int
-eal_parse_syslog(const char *facility)
-{
- int i;
- static struct {
- const char *name;
- int value;
- } map[] = {
- { "auth", LOG_AUTH },
- { "cron", LOG_CRON },
- { "daemon", LOG_DAEMON },
- { "ftp", LOG_FTP },
- { "kern", LOG_KERN },
- { "lpr", LOG_LPR },
- { "mail", LOG_MAIL },
- { "news", LOG_NEWS },
- { "syslog", LOG_SYSLOG },
- { "user", LOG_USER },
- { "uucp", LOG_UUCP },
- { "local0", LOG_LOCAL0 },
- { "local1", LOG_LOCAL1 },
- { "local2", LOG_LOCAL2 },
- { "local3", LOG_LOCAL3 },
- { "local4", LOG_LOCAL4 },
- { "local5", LOG_LOCAL5 },
- { "local6", LOG_LOCAL6 },
- { "local7", LOG_LOCAL7 },
- { NULL, 0 }
- };
-
- for (i = 0; map[i].name; i++) {
- if (!strcmp(facility, map[i].name)) {
- internal_config.syslog_facility = map[i].value;
- return 0;
- }
- }
- return -1;
-}
-
static int
eal_parse_socket_mem(char *socket_mem)
{
return 0;
}
-static inline size_t
-eal_get_hugepage_mem_size(void)
+static int
+eal_parse_base_virtaddr(const char *arg)
{
- uint64_t size = 0;
- unsigned i, j;
-
- for (i = 0; i < internal_config.num_hugepage_sizes; i++) {
- struct hugepage_info *hpi = &internal_config.hugepage_info[i];
- if (hpi->hugedir != NULL) {
- for (j = 0; j < RTE_MAX_NUMA_NODES; j++) {
- size += hpi->hugepage_sz * hpi->num_pages[j];
- }
- }
- }
+ char *end;
+ uint64_t addr;
+
+ errno = 0;
+ addr = strtoull(arg, &end, 16);
+
+ /* check for errors */
+ if ((errno != 0) || (arg[0] == '\0') || end == NULL || (*end != '\0'))
+ return -1;
+
+ /* make sure we don't exceed 32-bit boundary on 32-bit target */
+#ifndef RTE_ARCH_64
+ if (addr >= UINTPTR_MAX)
+ return -1;
+#endif
- return (size < SIZE_MAX) ? (size_t)(size) : SIZE_MAX;
+ /* align the addr on 16M boundary, 16MB is the minimum huge page
+ * size on IBM Power architecture. If the addr is aligned to 16MB,
+ * it can align to 2MB for x86. So this alignment can also be used
+ * on x86 */
+ internal_config.base_virtaddr =
+ RTE_PTR_ALIGN_CEIL((uintptr_t)addr, (size_t)RTE_PGSIZE_16M);
+
+ return 0;
}
-static enum rte_proc_type_t
-eal_parse_proc_type(const char *arg)
+static int
+eal_parse_vfio_intr(const char *mode)
{
- if (strncasecmp(arg, "primary", sizeof("primary")) == 0)
- return RTE_PROC_PRIMARY;
- if (strncasecmp(arg, "secondary", sizeof("secondary")) == 0)
- return RTE_PROC_SECONDARY;
- if (strncasecmp(arg, "auto", sizeof("auto")) == 0)
- return RTE_PROC_AUTO;
-
- return RTE_PROC_INVALID;
+ unsigned i;
+ static struct {
+ const char *name;
+ enum rte_intr_mode value;
+ } map[] = {
+ { "legacy", RTE_INTR_MODE_LEGACY },
+ { "msi", RTE_INTR_MODE_MSI },
+ { "msix", RTE_INTR_MODE_MSIX },
+ };
+
+ for (i = 0; i < RTE_DIM(map); i++) {
+ if (!strcmp(mode, map[i].name)) {
+ internal_config.vfio_intr_mode = map[i].value;
+ return 0;
+ }
+ }
+ return -1;
}
-static ssize_t
-eal_parse_blacklist_opt(const char *optarg, size_t idx)
+/* Parse the arguments for --log-level only */
+static void
+eal_log_level_parse(int argc, char **argv)
{
- if (idx >= sizeof (eal_dev_blacklist) / sizeof (eal_dev_blacklist[0])) {
- RTE_LOG(ERR, EAL, "%s - too many devices to blacklist...\n", optarg);
- return (-EINVAL);
- } else if (eal_parse_pci_DomBDF(optarg, eal_dev_blacklist + idx) < 0 &&
- eal_parse_pci_BDF(optarg, eal_dev_blacklist + idx) < 0) {
- RTE_LOG(ERR, EAL, "%s - invalid device to blacklist...\n", optarg);
- return (-EINVAL);
+ int opt;
+ char **argvopt;
+ int option_index;
+ const int old_optind = optind;
+ const int old_optopt = optopt;
+ char * const old_optarg = optarg;
+
+ argvopt = argv;
+ optind = 1;
+
+ eal_reset_internal_config(&internal_config);
+
+ while ((opt = getopt_long(argc, argvopt, eal_short_options,
+ eal_long_options, &option_index)) != EOF) {
+
+ int ret;
+
+ /* getopt is not happy, stop right now */
+ if (opt == '?')
+ break;
+
+ ret = (opt == OPT_LOG_LEVEL_NUM) ?
+ eal_parse_common_option(opt, optarg, &internal_config) : 0;
+
+ /* common parser is not happy */
+ if (ret < 0)
+ break;
}
- idx += 1;
- return (idx);
+ /* restore getopt lib */
+ optind = old_optind;
+ optopt = old_optopt;
+ optarg = old_optarg;
}
/* Parse the argument given in the command line of the application */
static int
eal_parse_args(int argc, char **argv)
{
- int opt, ret, i;
+ int opt, ret;
char **argvopt;
int option_index;
- int coremask_ok = 0;
- ssize_t blacklist_index = 0;
char *prgname = argv[0];
- static struct option lgopts[] = {
- {OPT_NO_HUGE, 0, 0, 0},
- {OPT_NO_PCI, 0, 0, 0},
- {OPT_NO_HPET, 0, 0, 0},
- {OPT_VMWARE_TSC_MAP, 0, 0, 0},
- {OPT_HUGE_DIR, 1, 0, 0},
- {OPT_NO_SHCONF, 0, 0, 0},
- {OPT_PROC_TYPE, 1, 0, 0},
- {OPT_FILE_PREFIX, 1, 0, 0},
- {OPT_SOCKET_MEM, 1, 0, 0},
- {OPT_USE_DEVICE, 1, 0, 0},
- {OPT_SYSLOG, 1, NULL, 0},
- {0, 0, 0, 0}
- };
+ const int old_optind = optind;
+ const int old_optopt = optopt;
+ char * const old_optarg = optarg;
argvopt = argv;
+ optind = 1;
- internal_config.memory = 0;
- internal_config.force_nrank = 0;
- internal_config.force_nchannel = 0;
- internal_config.hugefile_prefix = HUGEFILE_PREFIX_DEFAULT;
- internal_config.hugepage_dir = NULL;
- internal_config.force_sockets = 0;
- internal_config.syslog_facility = LOG_DAEMON;
-#ifdef RTE_LIBEAL_USE_HPET
- internal_config.no_hpet = 0;
-#else
- internal_config.no_hpet = 1;
-#endif
- /* zero out the NUMA config */
- for (i = 0; i < RTE_MAX_NUMA_NODES; i++)
- internal_config.socket_mem[i] = 0;
-
- /* zero out hugedir descriptors */
- for (i = 0; i < MAX_HUGEPAGE_SIZES; i++)
- internal_config.hugepage_info[i].lock_descriptor = 0;
+ while ((opt = getopt_long(argc, argvopt, eal_short_options,
+ eal_long_options, &option_index)) != EOF) {
- internal_config.vmware_tsc_map = 0;
+ /* getopt is not happy, stop right now */
+ if (opt == '?') {
+ eal_usage(prgname);
+ ret = -1;
+ goto out;
+ }
- while ((opt = getopt_long(argc, argvopt, "b:c:m:n:r:v",
- lgopts, &option_index)) != EOF) {
+ ret = eal_parse_common_option(opt, optarg, &internal_config);
+ /* common parser is not happy */
+ if (ret < 0) {
+ eal_usage(prgname);
+ ret = -1;
+ goto out;
+ }
+ /* common parser handled this option */
+ if (ret == 0)
+ continue;
switch (opt) {
- /* blacklist */
- case 'b':
- if ((blacklist_index = eal_parse_blacklist_opt(optarg,
- blacklist_index)) < 0) {
- eal_usage(prgname);
- return (-1);
- }
+ case 'h':
+ eal_usage(prgname);
+ exit(EXIT_SUCCESS);
+
+ /* long options */
+ case OPT_XEN_DOM0_NUM:
+#ifdef RTE_LIBRTE_XEN_DOM0
+ internal_config.xen_dom0_support = 1;
+#else
+ RTE_LOG(ERR, EAL, "Can't support DPDK app "
+ "running on Dom0, please configure"
+ " RTE_LIBRTE_XEN_DOM0=y\n");
+ ret = -1;
+ goto out;
+#endif
break;
- /* coremask */
- case 'c':
- if (eal_parse_coremask(optarg) < 0) {
- RTE_LOG(ERR, EAL, "invalid coremask\n");
- eal_usage(prgname);
- return -1;
- }
- coremask_ok = 1;
+
+ case OPT_HUGE_DIR_NUM:
+ internal_config.hugepage_dir = optarg;
break;
- /* size of memory */
- case 'm':
- internal_config.memory = atoi(optarg);
- internal_config.memory *= 1024ULL;
- internal_config.memory *= 1024ULL;
+
+ case OPT_FILE_PREFIX_NUM:
+ internal_config.hugefile_prefix = optarg;
break;
- /* force number of channels */
- case 'n':
- internal_config.force_nchannel = atoi(optarg);
- if (internal_config.force_nchannel == 0 ||
- internal_config.force_nchannel > 4) {
- RTE_LOG(ERR, EAL, "invalid channel number\n");
+
+ case OPT_SOCKET_MEM_NUM:
+ if (eal_parse_socket_mem(optarg) < 0) {
+ RTE_LOG(ERR, EAL, "invalid parameters for --"
+ OPT_SOCKET_MEM "\n");
eal_usage(prgname);
- return -1;
+ ret = -1;
+ goto out;
}
break;
- /* force number of ranks */
- case 'r':
- internal_config.force_nrank = atoi(optarg);
- if (internal_config.force_nrank == 0 ||
- internal_config.force_nrank > 16) {
- RTE_LOG(ERR, EAL, "invalid rank number\n");
+
+ case OPT_BASE_VIRTADDR_NUM:
+ if (eal_parse_base_virtaddr(optarg) < 0) {
+ RTE_LOG(ERR, EAL, "invalid parameter for --"
+ OPT_BASE_VIRTADDR "\n");
eal_usage(prgname);
- return -1;
+ ret = -1;
+ goto out;
}
break;
- case 'v':
- /* since message is explicitly requested by user, we
- * write message at highest log level so it can always be seen
- * even if info or warning messages are disabled */
- RTE_LOG(CRIT, EAL, "RTE Version: '%s'\n", rte_version());
- break;
- /* long options */
- case 0:
- if (!strcmp(lgopts[option_index].name, OPT_NO_HUGE)) {
- internal_config.no_hugetlbfs = 1;
- }
- else if (!strcmp(lgopts[option_index].name, OPT_NO_PCI)) {
- internal_config.no_pci = 1;
- }
- else if (!strcmp(lgopts[option_index].name, OPT_NO_HPET)) {
- internal_config.no_hpet = 1;
- }
- else if (!strcmp(lgopts[option_index].name, OPT_VMWARE_TSC_MAP)) {
- internal_config.vmware_tsc_map = 1;
- }
- else if (!strcmp(lgopts[option_index].name, OPT_NO_SHCONF)) {
- internal_config.no_shconf = 1;
- }
- else if (!strcmp(lgopts[option_index].name, OPT_HUGE_DIR)) {
- internal_config.hugepage_dir = optarg;
- }
- else if (!strcmp(lgopts[option_index].name, OPT_PROC_TYPE)) {
- internal_config.process_type = eal_parse_proc_type(optarg);
- }
- else if (!strcmp(lgopts[option_index].name, OPT_FILE_PREFIX)) {
- internal_config.hugefile_prefix = optarg;
- }
- else if (!strcmp(lgopts[option_index].name, OPT_SOCKET_MEM)) {
- if (eal_parse_socket_mem(optarg) < 0) {
- RTE_LOG(ERR, EAL, "invalid parameters for --"
- OPT_SOCKET_MEM "\n");
- eal_usage(prgname);
- return -1;
- }
- }
- else if (!strcmp(lgopts[option_index].name, OPT_USE_DEVICE)) {
- eal_dev_whitelist_add_entry(optarg);
- }
- else if (!strcmp(lgopts[option_index].name, OPT_SYSLOG)) {
- if (eal_parse_syslog(optarg) < 0) {
- RTE_LOG(ERR, EAL, "invalid parameters for --"
- OPT_SYSLOG "\n");
- eal_usage(prgname);
- return -1;
- }
+ case OPT_VFIO_INTR_NUM:
+ if (eal_parse_vfio_intr(optarg) < 0) {
+ RTE_LOG(ERR, EAL, "invalid parameters for --"
+ OPT_VFIO_INTR "\n");
+ eal_usage(prgname);
+ ret = -1;
+ goto out;
}
break;
+ case OPT_CREATE_UIO_DEV_NUM:
+ internal_config.create_uio_dev = 1;
+ break;
+
default:
+ if (opt < OPT_LONG_MIN_NUM && isprint(opt)) {
+ RTE_LOG(ERR, EAL, "Option %c is not supported "
+ "on Linux\n", opt);
+ } else if (opt >= OPT_LONG_MIN_NUM &&
+ opt < OPT_LONG_MAX_NUM) {
+ RTE_LOG(ERR, EAL, "Option %s is not supported "
+ "on Linux\n",
+ eal_long_options[option_index].name);
+ } else {
+ RTE_LOG(ERR, EAL, "Option %d is not supported "
+ "on Linux\n", opt);
+ }
eal_usage(prgname);
- return -1;
+ ret = -1;
+ goto out;
}
}
- /* sanity checks */
- if (!coremask_ok) {
- RTE_LOG(ERR, EAL, "coremask not specified\n");
- eal_usage(prgname);
- return -1;
- }
- if (internal_config.process_type == RTE_PROC_AUTO){
- internal_config.process_type = eal_proc_type_detect();
- }
- if (internal_config.process_type == RTE_PROC_INVALID){
- RTE_LOG(ERR, EAL, "Invalid process type specified\n");
- eal_usage(prgname);
- return -1;
- }
- if (internal_config.process_type == RTE_PROC_PRIMARY &&
- internal_config.force_nchannel == 0) {
- RTE_LOG(ERR, EAL, "Number of memory channels (-n) not specified\n");
- eal_usage(prgname);
- return -1;
+ if (eal_adjust_config(&internal_config) != 0) {
+ ret = -1;
+ goto out;
}
- if (index(internal_config.hugefile_prefix,'%') != NULL){
- RTE_LOG(ERR, EAL, "Invalid char, '%%', in '"OPT_FILE_PREFIX"' option\n");
- eal_usage(prgname);
- return -1;
- }
- if (internal_config.memory > 0 && internal_config.force_sockets == 1) {
- RTE_LOG(ERR, EAL, "Options -m and --socket-mem cannot be specified "
- "at the same time\n");
+
+ /* sanity checks */
+ if (eal_check_common_options(&internal_config) != 0) {
eal_usage(prgname);
- return -1;
- }
- /* --no-huge doesn't make sense with either -m or --socket-mem */
- if (internal_config.no_hugetlbfs &&
- (internal_config.memory > 0 ||
- internal_config.force_sockets == 1)) {
- RTE_LOG(ERR, EAL, "Options -m or --socket-mem cannot be specified "
- "together with --no-huge!\n");
- eal_usage(prgname);
- return -1;
+ ret = -1;
+ goto out;
}
- /* if no blacklist, parse a whitelist */
- if (blacklist_index > 0) {
- if (eal_dev_whitelist_exists()) {
- RTE_LOG(ERR, EAL, "Error: blacklist [-b] and whitelist "
- "[--use-device] options cannot be used at the same time\n");
- eal_usage(prgname);
- return -1;
- }
- rte_eal_pci_set_blacklist(eal_dev_blacklist, blacklist_index);
- } else {
- if (eal_dev_whitelist_exists() && eal_dev_whitelist_parse() < 0) {
- RTE_LOG(ERR,EAL, "Error parsing whitelist[--use-device] options\n");
- return -1;
- }
+ /* --xen-dom0 doesn't make sense with --socket-mem */
+ if (internal_config.xen_dom0_support && internal_config.force_sockets == 1) {
+ RTE_LOG(ERR, EAL, "Options --"OPT_SOCKET_MEM" cannot be specified "
+ "together with --"OPT_XEN_DOM0"\n");
+ eal_usage(prgname);
+ ret = -1;
+ goto out;
}
if (optind >= 0)
argv[optind-1] = prgname;
+ ret = optind-1;
- /* if no memory amounts were requested, this will result in 0 and
- * will be overriden later, right after eal_hugepage_info_init() */
- for (i = 0; i < RTE_MAX_NUMA_NODES; i++)
- internal_config.memory += internal_config.socket_mem[i];
+out:
+ /* restore getopt lib */
+ optind = old_optind;
+ optopt = old_optopt;
+ optarg = old_optarg;
- ret = optind-1;
- optind = 0; /* reset getopt lib */
return ret;
}
return 0;
}
-inline static void
+inline static void
rte_eal_mcfg_complete(void)
{
/* ALL shared mem_config related INIT DONE */
}
/*
- * Request iopl priviledge for all RPL, returns 0 on success
+ * Request iopl privilege for all RPL, returns 0 on success
+ * iopl() call is mostly for the i386 architecture. For other architectures,
+ * return -1 to indicate IO privilege can't be changed in this way.
*/
-static int
+int
rte_eal_iopl_init(void)
{
- return iopl(HIGHEST_RPL);
+#if defined(RTE_ARCH_X86)
+ if (iopl(3) != 0)
+ return -1;
+#endif
+ return 0;
+}
+
+#ifdef VFIO_PRESENT
+static int rte_eal_vfio_setup(void)
+{
+ int vfio_enabled = 0;
+
+ if (!internal_config.no_pci) {
+ pci_vfio_enable();
+ vfio_enabled |= pci_vfio_is_enabled();
+ }
+
+ if (vfio_enabled) {
+
+ /* if we are primary process, create a thread to communicate with
+ * secondary processes. the thread will use a socket to wait for
+ * requests from secondary process to send open file descriptors,
+ * because VFIO does not allow multiple open descriptors on a group or
+ * VFIO container.
+ */
+ if (internal_config.process_type == RTE_PROC_PRIMARY &&
+ vfio_mp_sync_setup() < 0)
+ return -1;
+ }
+
+ return 0;
+}
+#endif
+
+static void rte_eal_init_alert(const char *msg)
+{
+ fprintf(stderr, "EAL: FATAL: %s\n", msg);
+ RTE_LOG(ERR, EAL, "%s\n", msg);
}
/* Launch threads, called at application init(). */
int i, fctret, ret;
pthread_t thread_id;
static rte_atomic32_t run_once = RTE_ATOMIC32_INIT(0);
+ const char *logid;
+ char cpuset[RTE_CPU_AFFINITY_STR_LEN];
+ char thread_name[RTE_MAX_THREAD_NAME_LEN];
+
+ /* checks if the machine is adequate */
+ if (!rte_cpu_is_supported()) {
+ rte_eal_init_alert("unsupported cpu type.");
+ rte_errno = ENOTSUP;
+ return -1;
+ }
- if (!rte_atomic32_test_and_set(&run_once))
+ if (!rte_atomic32_test_and_set(&run_once)) {
+ rte_eal_init_alert("already called initialization.");
+ rte_errno = EALREADY;
return -1;
+ }
+
+ logid = strrchr(argv[0], '/');
+ logid = strdup(logid ? logid + 1: argv[0]);
thread_id = pthread_self();
- if (rte_eal_log_early_init() < 0)
- rte_panic("Cannot init early logs\n");
+ /* set log level as early as possible */
+ eal_log_level_parse(argc, argv);
+
+ if (rte_eal_cpu_init() < 0) {
+ rte_eal_init_alert("Cannot detect lcores.");
+ rte_errno = ENOTSUP;
+ return -1;
+ }
fctret = eal_parse_args(argc, argv);
- if (fctret < 0)
- exit(1);
+ if (fctret < 0) {
+ rte_eal_init_alert("Invalid 'command line' arguments.");
+ rte_errno = EINVAL;
+ rte_atomic32_clear(&run_once);
+ return -1;
+ }
if (internal_config.no_hugetlbfs == 0 &&
internal_config.process_type != RTE_PROC_SECONDARY &&
- eal_hugepage_info_init() < 0)
- rte_panic("Cannot get hugepage information\n");
+ internal_config.xen_dom0_support == 0 &&
+ eal_hugepage_info_init() < 0) {
+ rte_eal_init_alert("Cannot get hugepage information.");
+ rte_errno = EACCES;
+ rte_atomic32_clear(&run_once);
+ return -1;
+ }
if (internal_config.memory == 0 && internal_config.force_sockets == 0) {
if (internal_config.no_hugetlbfs)
internal_config.memory = MEMSIZE_IF_NO_HUGE_PAGE;
- else
- internal_config.memory = eal_get_hugepage_mem_size();
}
if (internal_config.vmware_tsc_map == 1) {
rte_config_init();
- if (rte_eal_iopl_init() == 0)
- rte_config.flags |= EAL_FLG_HIGH_IOPL;
-
- if (rte_eal_cpu_init() < 0)
- rte_panic("Cannot detect lcores\n");
+ if (rte_eal_log_init(logid, internal_config.syslog_facility) < 0) {
+ rte_eal_init_alert("Cannot init logging.");
+ rte_errno = ENOMEM;
+ rte_atomic32_clear(&run_once);
+ return -1;
+ }
+
+#ifdef VFIO_PRESENT
+ if (rte_eal_vfio_setup() < 0) {
+ rte_eal_init_alert("Cannot init VFIO\n");
+ rte_errno = EAGAIN;
+ rte_atomic32_clear(&run_once);
+ return -1;
+ }
+#endif
- if (rte_eal_memory_init() < 0)
- rte_panic("Cannot init memory\n");
+ if (rte_eal_memory_init() < 0) {
+ rte_eal_init_alert("Cannot init memory\n");
+ rte_errno = ENOMEM;
+ return -1;
+ }
/* the directories are locked during eal_hugepage_info_init */
eal_hugedirs_unlock();
-
- if (rte_eal_memzone_init() < 0)
- rte_panic("Cannot init memzone\n");
- if (rte_eal_tailqs_init() < 0)
- rte_panic("Cannot init tail queues for objects\n");
+ if (rte_eal_memzone_init() < 0) {
+ rte_eal_init_alert("Cannot init memzone\n");
+ rte_errno = ENODEV;
+ return -1;
+ }
- if (rte_eal_log_init(argv[0], internal_config.syslog_facility) < 0)
- rte_panic("Cannot init logs\n");
+ if (rte_eal_tailqs_init() < 0) {
+ rte_eal_init_alert("Cannot init tail queues for objects\n");
+ rte_errno = EFAULT;
+ return -1;
+ }
- if (rte_eal_alarm_init() < 0)
- rte_panic("Cannot init interrupt-handling thread\n");
+ if (rte_eal_alarm_init() < 0) {
+ rte_eal_init_alert("Cannot init interrupt-handling thread\n");
+ /* rte_eal_alarm_init sets rte_errno on failure. */
+ return -1;
+ }
- if (rte_eal_intr_init() < 0)
- rte_panic("Cannot init interrupt-handling thread\n");
+ if (rte_eal_timer_init() < 0) {
+ rte_eal_init_alert("Cannot init HPET or TSC timers\n");
+ rte_errno = ENOTSUP;
+ return -1;
+ }
- if (rte_eal_timer_init() < 0)
- rte_panic("Cannot init HPET or TSC timers\n");
+ eal_check_mem_on_local_socket();
- if (rte_eal_pci_init() < 0)
- rte_panic("Cannot init PCI\n");
+ if (eal_plugins_init() < 0)
+ rte_eal_init_alert("Cannot init plugins\n");
- RTE_LOG(DEBUG, EAL, "Master core %u is ready (tid=%x)\n",
- rte_config.master_lcore, (int)thread_id);
+ eal_thread_init_master(rte_config.master_lcore);
- eal_check_mem_on_local_socket();
+ ret = eal_thread_dump_affinity(cpuset, RTE_CPU_AFFINITY_STR_LEN);
- rte_eal_mcfg_complete();
+ RTE_LOG(DEBUG, EAL, "Master lcore %u is ready (tid=%x;cpuset=[%s%s])\n",
+ rte_config.master_lcore, (int)thread_id, cpuset,
+ ret == 0 ? "" : "...");
- if (rte_eal_non_pci_ethdev_init() < 0)
- rte_panic("Cannot init non-PCI eth_devs\n");
+ if (rte_eal_intr_init() < 0) {
+ rte_eal_init_alert("Cannot init interrupt-handling thread\n");
+ return -1;
+ }
+
+ if (rte_bus_scan()) {
+ rte_eal_init_alert("Cannot scan the buses for devices\n");
+ rte_errno = ENODEV;
+ return -1;
+ }
RTE_LCORE_FOREACH_SLAVE(i) {
eal_thread_loop, NULL);
if (ret != 0)
rte_panic("Cannot create thread\n");
- }
- eal_thread_init_master(rte_config.master_lcore);
+ /* Set thread_name for aid in debugging. */
+ snprintf(thread_name, RTE_MAX_THREAD_NAME_LEN,
+ "lcore-slave-%d", i);
+ ret = rte_thread_setname(lcore_config[i].thread_id,
+ thread_name);
+ if (ret != 0)
+ RTE_LOG(DEBUG, EAL,
+ "Cannot set name for lcore thread\n");
+ }
/*
* Launch a dummy function on all slave lcores, so that master lcore
rte_eal_mp_remote_launch(sync_func, NULL, SKIP_MASTER);
rte_eal_mp_wait_lcore();
+ /* Probe all the buses and devices/drivers on them */
+ if (rte_bus_probe()) {
+ rte_eal_init_alert("Cannot probe devices\n");
+ rte_errno = ENOTSUP;
+ return -1;
+ }
+
+ if (rte_eal_dev_init() < 0)
+ rte_eal_init_alert("Cannot init pmd devices\n");
+
+ rte_eal_mcfg_complete();
+
return fctret;
}
enum rte_lcore_role_t
rte_eal_lcore_role(unsigned lcore_id)
{
- return (rte_config.lcore_role[lcore_id]);
+ return rte_config.lcore_role[lcore_id];
}
enum rte_proc_type_t
rte_eal_process_type(void)
{
- return (rte_config.process_type);
+ return rte_config.process_type;
}
+int rte_eal_has_hugepages(void)
+{
+ return ! internal_config.no_hugetlbfs;
+}
+
+int
+rte_eal_check_module(const char *module_name)
+{
+ char sysfs_mod_name[PATH_MAX];
+ struct stat st;
+ int n;
+
+ if (NULL == module_name)
+ return -1;
+
+ /* Check if there is sysfs mounted */
+ if (stat("/sys/module", &st) != 0) {
+ RTE_LOG(DEBUG, EAL, "sysfs is not mounted! error %i (%s)\n",
+ errno, strerror(errno));
+ return -1;
+ }
+
+ /* A module might be built-in, therefore try sysfs */
+ n = snprintf(sysfs_mod_name, PATH_MAX, "/sys/module/%s", module_name);
+ if (n < 0 || n > PATH_MAX) {
+ RTE_LOG(DEBUG, EAL, "Could not format module path\n");
+ return -1;
+ }
+
+ if (stat(sysfs_mod_name, &st) != 0) {
+ RTE_LOG(DEBUG, EAL, "Module %s not found! error %i (%s)\n",
+ sysfs_mod_name, errno, strerror(errno));
+ return 0;
+ }
+
+ /* Module has been found */
+ return 1;
+}