/*-
* BSD LICENSE
*
- * Copyright(c) 2010-2013 Intel Corporation. All rights reserved.
+ * Copyright(c) 2010-2014 Intel Corporation. All rights reserved.
* All rights reserved.
*
- * Redistribution and use in source and binary forms, with or without
- * modification, are permitted provided that the following conditions
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions
* are met:
*
- * * Redistributions of source code must retain the above copyright
+ * * Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
- * * Redistributions in binary form must reproduce the above copyright
- * notice, this list of conditions and the following disclaimer in
- * the documentation and/or other materials provided with the
+ * * Redistributions in binary form must reproduce the above copyright
+ * notice, this list of conditions and the following disclaimer in
+ * the documentation and/or other materials provided with the
* distribution.
- * * Neither the name of Intel Corporation nor the names of its
- * contributors may be used to endorse or promote products derived
+ * * Neither the name of Intel Corporation nor the names of its
+ * contributors may be used to endorse or promote products derived
* from this software without specific prior written permission.
*
- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
- * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
- * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
- * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
- * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
- * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
- * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
- * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
- * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
- * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
+ * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
+ * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
+ * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
+ * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
+ * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
+ * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
+ * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
+ * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
- *
*/
#ifndef RTE_EXEC_ENV_LINUXAPP
#include <fcntl.h>
#include <unistd.h>
#include <sys/ioctl.h>
-#include <netinet/in.h>
-#include <linux/if.h>
#include <rte_string_fns.h>
#include <rte_ethdev.h>
* KNI context
*/
struct rte_kni {
- char name[IFNAMSIZ]; /**< KNI interface name */
- uint8_t port_id; /**< Port id KNI associate with */
+ char name[RTE_KNI_NAMESIZE]; /**< KNI interface name */
+ uint16_t group_id; /**< Group ID of KNI devices */
struct rte_mempool *pktmbuf_pool; /**< pkt mbuf mempool */
unsigned mbuf_size; /**< mbuf size */
void * sync_addr; /**< Req/Resp Mem address */
struct rte_kni_ops ops; /**< operations for request */
- uint8_t port_in_use : 1; /**< kni creation flag */
+ uint8_t in_use : 1; /**< kni in use */
};
enum kni_ops_status {
return mz;
}
+/* It is deprecated and just for backward compatibility */
struct rte_kni *
rte_kni_create(uint8_t port_id,
- unsigned mbuf_size,
- struct rte_mempool *pktmbuf_pool,
- struct rte_kni_ops *ops)
+ unsigned mbuf_size,
+ struct rte_mempool *pktmbuf_pool,
+ struct rte_kni_ops *ops)
+{
+ struct rte_kni_conf conf;
+ struct rte_eth_dev_info info;
+
+ memset(&info, 0, sizeof(info));
+ memset(&conf, 0, sizeof(conf));
+ rte_eth_dev_info_get(port_id, &info);
+
+ rte_snprintf(conf.name, sizeof(conf.name), "vEth%u", port_id);
+ conf.addr = info.pci_dev->addr;
+ conf.id = info.pci_dev->id;
+ conf.group_id = (uint16_t)port_id;
+ conf.mbuf_size = mbuf_size;
+
+ /* Save the port id for request handling */
+ ops->port_id = port_id;
+
+ return rte_kni_alloc(pktmbuf_pool, &conf, ops);
+}
+
+struct rte_kni *
+rte_kni_alloc(struct rte_mempool *pktmbuf_pool,
+ const struct rte_kni_conf *conf,
+ struct rte_kni_ops *ops)
{
int ret;
struct rte_kni_device_info dev_info;
- struct rte_eth_dev_info eth_dev_info;
struct rte_kni *ctx;
- char itf_name[IFNAMSIZ];
+ char intf_name[RTE_KNI_NAMESIZE];
#define OBJNAMSIZ 32
char obj_name[OBJNAMSIZ];
char mz_name[RTE_MEMZONE_NAMESIZE];
const struct rte_memzone *mz;
- if (port_id >= RTE_MAX_ETHPORTS || pktmbuf_pool == NULL)
+ if (!pktmbuf_pool || !conf || !conf->name[0])
return NULL;
/* Check FD and open once */
}
}
- rte_eth_dev_info_get(port_id, ð_dev_info);
- RTE_LOG(INFO, KNI, "pci: %02x:%02x:%02x \t %02x:%02x\n",
- eth_dev_info.pci_dev->addr.bus,
- eth_dev_info.pci_dev->addr.devid,
- eth_dev_info.pci_dev->addr.function,
- eth_dev_info.pci_dev->id.vendor_id,
- eth_dev_info.pci_dev->id.device_id);
- dev_info.bus = eth_dev_info.pci_dev->addr.bus;
- dev_info.devid = eth_dev_info.pci_dev->addr.devid;
- dev_info.function = eth_dev_info.pci_dev->addr.function;
- dev_info.vendor_id = eth_dev_info.pci_dev->id.vendor_id;
- dev_info.device_id = eth_dev_info.pci_dev->id.device_id;
- dev_info.port_id = port_id;
-
- rte_snprintf(mz_name, RTE_MEMZONE_NAMESIZE, "KNI_INFO_%d", port_id);
+ rte_snprintf(intf_name, RTE_KNI_NAMESIZE, conf->name);
+ rte_snprintf(mz_name, RTE_MEMZONE_NAMESIZE, "KNI_INFO_%s", intf_name);
mz = kni_memzone_reserve(mz_name, sizeof(struct rte_kni),
SOCKET_ID_ANY, 0);
KNI_MZ_CHECK(mz == NULL);
ctx = mz->addr;
- if (ctx->port_in_use != 0) {
- RTE_LOG(ERR, KNI, "Port %d has been used\n", port_id);
+ if (ctx->in_use) {
+ RTE_LOG(ERR, KNI, "KNI %s is in use\n", ctx->name);
goto fail;
}
memset(ctx, 0, sizeof(struct rte_kni));
if (ops)
memcpy(&ctx->ops, ops, sizeof(struct rte_kni_ops));
- rte_snprintf(itf_name, IFNAMSIZ, "vEth%u", port_id);
- rte_snprintf(ctx->name, IFNAMSIZ, itf_name);
- rte_snprintf(dev_info.name, IFNAMSIZ, itf_name);
+ memset(&dev_info, 0, sizeof(dev_info));
+ dev_info.bus = conf->addr.bus;
+ dev_info.devid = conf->addr.devid;
+ dev_info.function = conf->addr.function;
+ dev_info.vendor_id = conf->id.vendor_id;
+ dev_info.device_id = conf->id.device_id;
+ dev_info.core_id = conf->core_id;
+ dev_info.force_bind = conf->force_bind;
+ dev_info.group_id = conf->group_id;
+ dev_info.mbuf_size = conf->mbuf_size;
+
+ rte_snprintf(ctx->name, RTE_KNI_NAMESIZE, intf_name);
+ rte_snprintf(dev_info.name, RTE_KNI_NAMESIZE, intf_name);
+
+ RTE_LOG(INFO, KNI, "pci: %02x:%02x:%02x \t %02x:%02x\n",
+ dev_info.bus, dev_info.devid, dev_info.function,
+ dev_info.vendor_id, dev_info.device_id);
/* TX RING */
- rte_snprintf(obj_name, OBJNAMSIZ, "kni_tx_%d", port_id);
+ rte_snprintf(obj_name, OBJNAMSIZ, "kni_tx_%s", intf_name);
mz = kni_memzone_reserve(obj_name, KNI_FIFO_SIZE, SOCKET_ID_ANY, 0);
KNI_MZ_CHECK(mz == NULL);
ctx->tx_q = mz->addr;
dev_info.tx_phys = mz->phys_addr;
/* RX RING */
- rte_snprintf(obj_name, OBJNAMSIZ, "kni_rx_%d", port_id);
+ rte_snprintf(obj_name, OBJNAMSIZ, "kni_rx_%s", intf_name);
mz = kni_memzone_reserve(obj_name, KNI_FIFO_SIZE, SOCKET_ID_ANY, 0);
KNI_MZ_CHECK(mz == NULL);
ctx->rx_q = mz->addr;
dev_info.rx_phys = mz->phys_addr;
/* ALLOC RING */
- rte_snprintf(obj_name, OBJNAMSIZ, "kni_alloc_%d", port_id);
+ rte_snprintf(obj_name, OBJNAMSIZ, "kni_alloc_%s", intf_name);
mz = kni_memzone_reserve(obj_name, KNI_FIFO_SIZE, SOCKET_ID_ANY, 0);
KNI_MZ_CHECK(mz == NULL);
ctx->alloc_q = mz->addr;
dev_info.alloc_phys = mz->phys_addr;
/* FREE RING */
- rte_snprintf(obj_name, OBJNAMSIZ, "kni_free_%d", port_id);
+ rte_snprintf(obj_name, OBJNAMSIZ, "kni_free_%s", intf_name);
mz = kni_memzone_reserve(obj_name, KNI_FIFO_SIZE, SOCKET_ID_ANY, 0);
KNI_MZ_CHECK(mz == NULL);
ctx->free_q = mz->addr;
dev_info.free_phys = mz->phys_addr;
/* Request RING */
- rte_snprintf(obj_name, OBJNAMSIZ, "kni_req_%d", port_id);
+ rte_snprintf(obj_name, OBJNAMSIZ, "kni_req_%s", intf_name);
mz = kni_memzone_reserve(obj_name, KNI_FIFO_SIZE, SOCKET_ID_ANY, 0);
KNI_MZ_CHECK(mz == NULL);
ctx->req_q = mz->addr;
dev_info.req_phys = mz->phys_addr;
/* Response RING */
- rte_snprintf(obj_name, OBJNAMSIZ, "kni_resp_%d", port_id);
+ rte_snprintf(obj_name, OBJNAMSIZ, "kni_resp_%s", intf_name);
mz = kni_memzone_reserve(obj_name, KNI_FIFO_SIZE, SOCKET_ID_ANY, 0);
KNI_MZ_CHECK(mz == NULL);
ctx->resp_q = mz->addr;
dev_info.resp_phys = mz->phys_addr;
/* Req/Resp sync mem area */
- rte_snprintf(obj_name, OBJNAMSIZ, "kni_sync_%d", port_id);
+ rte_snprintf(obj_name, OBJNAMSIZ, "kni_sync_%s", intf_name);
mz = kni_memzone_reserve(obj_name, KNI_FIFO_SIZE, SOCKET_ID_ANY, 0);
KNI_MZ_CHECK(mz == NULL);
ctx->sync_addr = mz->addr;
dev_info.sync_phys = mz->phys_addr;
/* MBUF mempool */
- rte_snprintf(mz_name, sizeof(mz_name), "MP_%s", pktmbuf_pool->name);
+ rte_snprintf(mz_name, sizeof(mz_name), RTE_MEMPOOL_OBJ_NAME,
+ pktmbuf_pool->name);
mz = rte_memzone_lookup(mz_name);
KNI_MZ_CHECK(mz == NULL);
dev_info.mbuf_va = mz->addr;
dev_info.mbuf_phys = mz->phys_addr;
ctx->pktmbuf_pool = pktmbuf_pool;
- ctx->port_id = port_id;
- ctx->mbuf_size = mbuf_size;
-
- /* Configure the buffer size which will be checked in kernel module */
- dev_info.mbuf_size = ctx->mbuf_size;
+ ctx->group_id = conf->group_id;
+ ctx->mbuf_size = conf->mbuf_size;
ret = ioctl(kni_fd, RTE_KNI_IOCTL_CREATE, &dev_info);
KNI_MZ_CHECK(ret < 0);
- ctx->port_in_use = 1;
+ ctx->in_use = 1;
return ctx;
int
rte_kni_release(struct rte_kni *kni)
{
- if (!kni || kni->port_in_use == 0)
+ struct rte_kni_device_info dev_info;
+
+ if (!kni || !kni->in_use)
return -1;
- if (ioctl(kni_fd, RTE_KNI_IOCTL_RELEASE, &kni->port_id) < 0) {
+ rte_snprintf(dev_info.name, sizeof(dev_info.name), kni->name);
+ if (ioctl(kni_fd, RTE_KNI_IOCTL_RELEASE, &dev_info) < 0) {
RTE_LOG(ERR, KNI, "Fail to release kni device\n");
return -1;
}
switch (req->req_id) {
case RTE_KNI_REQ_CHANGE_MTU: /* Change MTU */
if (kni->ops.change_mtu)
- req->result = kni->ops.change_mtu(kni->port_id,
+ req->result = kni->ops.change_mtu(kni->ops.port_id,
req->new_mtu);
break;
case RTE_KNI_REQ_CFG_NETWORK_IF: /* Set network interface up/down */
if (kni->ops.config_network_if)
- req->result = kni->ops.config_network_if(kni->port_id,
- req->if_up);
+ req->result = kni->ops.config_network_if(\
+ kni->ops.port_id, req->if_up);
break;
default:
RTE_LOG(ERR, KNI, "Unknown request id %u\n", req->req_id);
}
}
+/* It is deprecated and just for backward compatibility */
uint8_t
rte_kni_get_port_id(struct rte_kni *kni)
{
- if (kni == NULL)
+ if (!kni)
return ~0x0;
- return kni->port_id;
+ return kni->ops.port_id;
}
struct rte_kni *
-rte_kni_info_get(uint8_t port_id)
+rte_kni_get(const char *name)
{
struct rte_kni *kni;
const struct rte_memzone *mz;
char mz_name[RTE_MEMZONE_NAMESIZE];
- if(port_id >= RTE_MAX_ETHPORTS)
+ if (!name || !name[0])
return NULL;
- rte_snprintf(mz_name, RTE_MEMZONE_NAMESIZE, "KNI_INFO_%d", port_id);
+ rte_snprintf(mz_name, RTE_MEMZONE_NAMESIZE, "KNI_INFO_%s", name);
mz = rte_memzone_lookup(mz_name);
- if (NULL == mz)
+ if (!mz)
return NULL;
kni = mz->addr;
- if (0 == kni->port_in_use)
+ if (!kni->in_use)
return NULL;
-
+
return kni;
}
+/*
+ * It is deprecated and just for backward compatibility.
+ */
+struct rte_kni *
+rte_kni_info_get(uint8_t port_id)
+{
+ char name[RTE_MEMZONE_NAMESIZE];
+
+ if (port_id >= RTE_MAX_ETHPORTS)
+ return NULL;
+
+ rte_snprintf(name, RTE_MEMZONE_NAMESIZE, "vEth%u", port_id);
+
+ return rte_kni_get(name);
+}
+
static enum kni_ops_status
kni_check_request_register(struct rte_kni_ops *ops)
{
return -1;
}
- if (NULL == &kni->ops) {
- RTE_LOG(ERR, KNI, "The invalid KNI unregister operation.\n");
- return -1;
- }
-
kni->ops.change_mtu = NULL;
kni->ops.config_network_if = NULL;
return 0;
}
+void
+rte_kni_close(void)
+{
+ if (kni_fd < 0)
+ return;
+
+ close(kni_fd);
+ kni_fd = -1;
+}