{
struct rx_queue *rxq;
uint32_t i, len, next_hop;
- uint8_t ipv6;
- uint16_t port_out;
+ uint16_t port_out, ether_type;
int32_t len2;
+ uint64_t ol_flags;
+ const struct rte_ether_hdr *eth;
- ipv6 = 0;
+ ol_flags = 0;
rxq = &qconf->rx_queue_list[queueid];
/* by default, send everything back to the source port */
port_out = port_in;
+ /* save ether type of the incoming packet */
+ eth = rte_pktmbuf_mtod(m, const struct rte_ether_hdr *);
+ ether_type = eth->ether_type;
+
/* Remove the Ethernet header and trailer from the input packet */
rte_pktmbuf_adj(m, (uint16_t)sizeof(struct rte_ether_hdr));
/* Free input packet */
rte_pktmbuf_free(m);
+ /* request HW to regenerate IPv4 cksum */
+ ol_flags |= (PKT_TX_IPV4 | PKT_TX_IP_CKSUM);
+
/* If we fail to fragment the packet */
if (unlikely (len2 < 0))
return;
/* if this is an IPv6 packet */
struct rte_ipv6_hdr *ip_hdr;
- ipv6 = 1;
-
/* Read the lookup key (i.e. ip_dst) from the input packet */
ip_hdr = rte_pktmbuf_mtod(m, struct rte_ipv6_hdr *);
rte_panic("No headroom in mbuf.\n");
}
+ m->ol_flags |= ol_flags;
m->l2_len = sizeof(struct rte_ether_hdr);
/* 02:00:00:00:00:xx */
d_addr_bytes = ð_hdr->d_addr.addr_bytes[0];
- *((uint64_t *)d_addr_bytes) = 0x000000000002 + ((uint64_t)port_out << 40);
+ *((uint64_t *)d_addr_bytes) = 0x000000000002 +
+ ((uint64_t)port_out << 40);
/* src addr */
rte_ether_addr_copy(&ports_eth_addr[port_out],
ð_hdr->s_addr);
- if (ipv6) {
- eth_hdr->ether_type =
- rte_be_to_cpu_16(RTE_ETHER_TYPE_IPV6);
- } else {
- eth_hdr->ether_type =
- rte_be_to_cpu_16(RTE_ETHER_TYPE_IPV4);
- m->ol_flags |= (PKT_TX_IPV4 | PKT_TX_IP_CKSUM);
- }
+ eth_hdr->ether_type = ether_type;
}
len += len2;
n_tx_queue = nb_lcores;
if (n_tx_queue > MAX_TX_QUEUE_PER_PORT)
n_tx_queue = MAX_TX_QUEUE_PER_PORT;
- if (dev_info.tx_offload_capa & DEV_TX_OFFLOAD_MBUF_FAST_FREE)
- local_port_conf.txmode.offloads |=
- DEV_TX_OFFLOAD_MBUF_FAST_FREE;
ret = rte_eth_dev_configure(portid, 1, (uint16_t)n_tx_queue,
&local_port_conf);
if (ret < 0) {
printf("\n");
/* init one TX queue per couple (lcore,port) */
+ rte_eth_dev_info_get(portid, &dev_info);
queueid = 0;
for (lcore_id = 0; lcore_id < RTE_MAX_LCORE; lcore_id++) {
if (rte_lcore_is_enabled(lcore_id) == 0)
continue;
- if (queueid >= rte_eth_devices[portid].data->nb_tx_queues)
+ if (queueid >= dev_info.nb_tx_queues)
break;
socket = (int) rte_lcore_to_socket_id(lcore_id);