static struct rte_eth_conf port_conf = {
.rxmode = {
.mq_mode = ETH_MQ_RX_NONE,
- .max_rx_pkt_len = ETHER_MAX_LEN,
+ .max_rx_pkt_len = RTE_ETHER_MAX_LEN,
.split_hdr_size = 0,
},
.rx_adv_conf = {
{
size_t vlan_offset = 0;
- if (rte_cpu_to_be_16(ETHER_TYPE_VLAN) == *proto) {
+ if (rte_cpu_to_be_16(RTE_ETHER_TYPE_VLAN) == *proto) {
struct rte_vlan_hdr *vlan_hdr =
(struct rte_vlan_hdr *)(eth_hdr + 1);
vlan_offset = sizeof(struct rte_vlan_hdr);
*proto = vlan_hdr->eth_proto;
- if (rte_cpu_to_be_16(ETHER_TYPE_VLAN) == *proto) {
+ if (rte_cpu_to_be_16(RTE_ETHER_TYPE_VLAN) == *proto) {
vlan_hdr = vlan_hdr + 1;
*proto = vlan_hdr->eth_proto;
struct rte_ether_hdr *eth_hdr;
struct rte_arp_hdr *arp_hdr;
- struct ipv4_hdr *ipv4_hdr;
+ struct rte_ipv4_hdr *ipv4_hdr;
uint16_t ether_type, offset;
uint16_t rx_cnt;
eth_hdr = rte_pktmbuf_mtod(pkts[i],
struct rte_ether_hdr *);
ether_type = eth_hdr->ether_type;
- if (ether_type == rte_cpu_to_be_16(ETHER_TYPE_VLAN))
+ if (ether_type == rte_cpu_to_be_16(RTE_ETHER_TYPE_VLAN))
printf("VLAN taged frame, offset:");
offset = get_vlan_offset(eth_hdr, ðer_type);
if (offset > 0)
printf("%d\n", offset);
- if (ether_type == rte_cpu_to_be_16(ETHER_TYPE_ARP)) {
+ if (ether_type == rte_cpu_to_be_16(RTE_ETHER_TYPE_ARP)) {
if (rte_spinlock_trylock(&global_flag_stru_p->lock) == 1) {
global_flag_stru_p->port_packets[1]++;
rte_spinlock_unlock(&global_flag_stru_p->lock);
if (arp_hdr->arp_opcode == rte_cpu_to_be_16(RTE_ARP_OP_REQUEST)) {
arp_hdr->arp_opcode = rte_cpu_to_be_16(RTE_ARP_OP_REPLY);
/* Switch src and dst data and set bonding MAC */
- ether_addr_copy(ð_hdr->s_addr, ð_hdr->d_addr);
+ rte_ether_addr_copy(ð_hdr->s_addr, ð_hdr->d_addr);
rte_eth_macaddr_get(BOND_PORT, ð_hdr->s_addr);
- ether_addr_copy(&arp_hdr->arp_data.arp_sha, &arp_hdr->arp_data.arp_tha);
+ rte_ether_addr_copy(&arp_hdr->arp_data.arp_sha,
+ &arp_hdr->arp_data.arp_tha);
arp_hdr->arp_data.arp_tip = arp_hdr->arp_data.arp_sip;
rte_eth_macaddr_get(BOND_PORT, &d_addr);
- ether_addr_copy(&d_addr, &arp_hdr->arp_data.arp_sha);
+ rte_ether_addr_copy(&d_addr, &arp_hdr->arp_data.arp_sha);
arp_hdr->arp_data.arp_sip = bond_ip;
rte_eth_tx_burst(BOND_PORT, 0, &pkts[i], 1);
is_free = 1;
rte_eth_tx_burst(BOND_PORT, 0, NULL, 0);
}
}
- } else if (ether_type == rte_cpu_to_be_16(ETHER_TYPE_IPv4)) {
+ } else if (ether_type == rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV4)) {
if (rte_spinlock_trylock(&global_flag_stru_p->lock) == 1) {
global_flag_stru_p->port_packets[2]++;
rte_spinlock_unlock(&global_flag_stru_p->lock);
}
- ipv4_hdr = (struct ipv4_hdr *)((char *)(eth_hdr + 1) + offset);
+ ipv4_hdr = (struct rte_ipv4_hdr *)((char *)(eth_hdr + 1) + offset);
if (ipv4_hdr->dst_addr == bond_ip) {
- ether_addr_copy(ð_hdr->s_addr, ð_hdr->d_addr);
+ rte_ether_addr_copy(ð_hdr->s_addr, ð_hdr->d_addr);
rte_eth_macaddr_get(BOND_PORT, ð_hdr->s_addr);
ipv4_hdr->dst_addr = ipv4_hdr->src_addr;
ipv4_hdr->src_addr = bond_ip;
eth_hdr = rte_pktmbuf_mtod(created_pkt, struct rte_ether_hdr *);
rte_eth_macaddr_get(BOND_PORT, ð_hdr->s_addr);
- memset(ð_hdr->d_addr, 0xFF, ETHER_ADDR_LEN);
- eth_hdr->ether_type = rte_cpu_to_be_16(ETHER_TYPE_ARP);
+ memset(ð_hdr->d_addr, 0xFF, RTE_ETHER_ADDR_LEN);
+ eth_hdr->ether_type = rte_cpu_to_be_16(RTE_ETHER_TYPE_ARP);
arp_hdr = (struct rte_arp_hdr *)(
(char *)eth_hdr + sizeof(struct rte_ether_hdr));
arp_hdr->arp_hardware = rte_cpu_to_be_16(RTE_ARP_HRD_ETHER);
- arp_hdr->arp_protocol = rte_cpu_to_be_16(ETHER_TYPE_IPv4);
- arp_hdr->arp_hlen = ETHER_ADDR_LEN;
+ arp_hdr->arp_protocol = rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV4);
+ arp_hdr->arp_hlen = RTE_ETHER_ADDR_LEN;
arp_hdr->arp_plen = sizeof(uint32_t);
arp_hdr->arp_opcode = rte_cpu_to_be_16(RTE_ARP_OP_REQUEST);
rte_eth_macaddr_get(BOND_PORT, &arp_hdr->arp_data.arp_sha);
arp_hdr->arp_data.arp_sip = bond_ip;
- memset(&arp_hdr->arp_data.arp_tha, 0, ETHER_ADDR_LEN);
+ memset(&arp_hdr->arp_data.arp_tha, 0, RTE_ETHER_ADDR_LEN);
arp_hdr->arp_data.arp_tip =
((unsigned char *)&res->ip.addr.ipv4)[0] |
(((unsigned char *)&res->ip.addr.ipv4)[1] << 8) |
int slave_core_id = rte_lcore_id();
rte_spinlock_trylock(&global_flag_stru_p->lock);
- if (global_flag_stru_p->LcoreMainIsRunning == 0) {
- if (lcore_config[global_flag_stru_p->LcoreMainCore].state != WAIT) {
+ if (global_flag_stru_p->LcoreMainIsRunning == 0) {
+ if (rte_eal_get_lcore_state(global_flag_stru_p->LcoreMainCore)
+ != WAIT) {
rte_spinlock_unlock(&global_flag_stru_p->lock);
return;
}
int
main(int argc, char *argv[])
{
- int ret;
+ int ret, slave_core_id;
uint16_t nb_ports, i;
/* init EAL */
bond_port_init(mbuf_pool);
rte_spinlock_init(&global_flag_stru_p->lock);
- int slave_core_id = rte_lcore_id();
/* check state of lcores */
RTE_LCORE_FOREACH_SLAVE(slave_core_id) {
- if (lcore_config[slave_core_id].state != WAIT)
- return -EBUSY;
+ if (rte_eal_get_lcore_state(slave_core_id) != WAIT)
+ return -EBUSY;
}
+
/* start lcore main on core != master_core - ARP response thread */
slave_core_id = rte_get_next_lcore(rte_lcore_id(), 1, 0);
if ((slave_core_id >= RTE_MAX_LCORE) || (slave_core_id == 0))