examples/l3fwd: add event em main loop
[dpdk.git] / examples / l3fwd / l3fwd_em_hlm.h
index ad8b9ce..7981271 100644 (file)
@@ -75,6 +75,60 @@ em_get_dst_port_ipv6xN(struct lcore_conf *qconf, struct rte_mbuf *m[],
        }
 }
 
+static __rte_always_inline void
+em_get_dst_port_ipv4xN_events(struct lcore_conf *qconf, struct rte_mbuf *m[],
+                             uint16_t dst_port[])
+{
+       int i;
+       int32_t ret[EM_HASH_LOOKUP_COUNT];
+       union ipv4_5tuple_host key[EM_HASH_LOOKUP_COUNT];
+       const void *key_array[EM_HASH_LOOKUP_COUNT];
+
+       for (i = 0; i < EM_HASH_LOOKUP_COUNT; i++) {
+               get_ipv4_5tuple(m[i], mask0.x, &key[i]);
+               key_array[i] = &key[i];
+       }
+
+       rte_hash_lookup_bulk(qconf->ipv4_lookup_struct, &key_array[0],
+                            EM_HASH_LOOKUP_COUNT, ret);
+
+       for (i = 0; i < EM_HASH_LOOKUP_COUNT; i++) {
+               dst_port[i] = ((ret[i] < 0) ?
+                               m[i]->port : ipv4_l3fwd_out_if[ret[i]]);
+
+               if (dst_port[i] >= RTE_MAX_ETHPORTS ||
+                               (enabled_port_mask & 1 << dst_port[i]) == 0)
+                       dst_port[i] = m[i]->port;
+       }
+}
+
+static __rte_always_inline void
+em_get_dst_port_ipv6xN_events(struct lcore_conf *qconf, struct rte_mbuf *m[],
+                             uint16_t dst_port[])
+{
+       int i;
+       int32_t ret[EM_HASH_LOOKUP_COUNT];
+       union ipv6_5tuple_host key[EM_HASH_LOOKUP_COUNT];
+       const void *key_array[EM_HASH_LOOKUP_COUNT];
+
+       for (i = 0; i < EM_HASH_LOOKUP_COUNT; i++) {
+               get_ipv6_5tuple(m[i], mask1.x, mask2.x, &key[i]);
+               key_array[i] = &key[i];
+       }
+
+       rte_hash_lookup_bulk(qconf->ipv6_lookup_struct, &key_array[0],
+                            EM_HASH_LOOKUP_COUNT, ret);
+
+       for (i = 0; i < EM_HASH_LOOKUP_COUNT; i++) {
+               dst_port[i] = ((ret[i] < 0) ?
+                               m[i]->port : ipv6_l3fwd_out_if[ret[i]]);
+
+               if (dst_port[i] >= RTE_MAX_ETHPORTS ||
+                               (enabled_port_mask & 1 << dst_port[i]) == 0)
+                       dst_port[i] = m[i]->port;
+       }
+}
+
 static __rte_always_inline uint16_t
 em_get_dst_port(const struct lcore_conf *qconf, struct rte_mbuf *pkt,
                uint16_t portid)
@@ -187,4 +241,81 @@ l3fwd_em_send_packets(int nb_rx, struct rte_mbuf **pkts_burst,
        send_packets_multi(qconf, pkts_burst, dst_port, nb_rx);
 
 }
+
+/*
+ * Buffer optimized handling of events, invoked
+ * from main_loop.
+ */
+static inline void
+l3fwd_em_process_events(int nb_rx, struct rte_event **ev,
+                    struct lcore_conf *qconf)
+{
+       int32_t i, j, pos;
+       uint16_t dst_port[MAX_PKT_BURST];
+       struct rte_mbuf *pkts_burst[MAX_PKT_BURST];
+
+       /*
+        * Send nb_rx - nb_rx % EM_HASH_LOOKUP_COUNT packets
+        * in groups of EM_HASH_LOOKUP_COUNT.
+        */
+       int32_t n = RTE_ALIGN_FLOOR(nb_rx, EM_HASH_LOOKUP_COUNT);
+
+       for (j = 0; j < EM_HASH_LOOKUP_COUNT && j < nb_rx; j++) {
+               pkts_burst[j] = ev[j]->mbuf;
+               rte_prefetch0(rte_pktmbuf_mtod(pkts_burst[j],
+                                              struct rte_ether_hdr *) + 1);
+       }
+
+       for (j = 0; j < n; j += EM_HASH_LOOKUP_COUNT) {
+
+               uint32_t pkt_type = RTE_PTYPE_L3_MASK |
+                                   RTE_PTYPE_L4_TCP | RTE_PTYPE_L4_UDP;
+               uint32_t l3_type, tcp_or_udp;
+
+               for (i = 0; i < EM_HASH_LOOKUP_COUNT; i++)
+                       pkt_type &= pkts_burst[j + i]->packet_type;
+
+               l3_type = pkt_type & RTE_PTYPE_L3_MASK;
+               tcp_or_udp = pkt_type & (RTE_PTYPE_L4_TCP | RTE_PTYPE_L4_UDP);
+
+               for (i = 0, pos = j + EM_HASH_LOOKUP_COUNT;
+                    i < EM_HASH_LOOKUP_COUNT && pos < nb_rx; i++, pos++) {
+                       rte_prefetch0(rte_pktmbuf_mtod(
+                                       pkts_burst[pos],
+                                       struct rte_ether_hdr *) + 1);
+               }
+
+               if (tcp_or_udp && (l3_type == RTE_PTYPE_L3_IPV4)) {
+
+                       em_get_dst_port_ipv4xN_events(qconf, &pkts_burst[j],
+                                              &dst_port[j]);
+
+               } else if (tcp_or_udp && (l3_type == RTE_PTYPE_L3_IPV6)) {
+
+                       em_get_dst_port_ipv6xN_events(qconf, &pkts_burst[j],
+                                              &dst_port[j]);
+
+               } else {
+                       for (i = 0; i < EM_HASH_LOOKUP_COUNT; i++) {
+                               pkts_burst[j + i]->port = em_get_dst_port(qconf,
+                                               pkts_burst[j + i],
+                                               pkts_burst[j + i]->port);
+                               process_packet(pkts_burst[j + i],
+                                               &pkts_burst[j + i]->port);
+                       }
+                       continue;
+               }
+               processx4_step3(&pkts_burst[j], &dst_port[j]);
+
+               for (i = 0; i < EM_HASH_LOOKUP_COUNT; i++)
+                       pkts_burst[j + i]->port = dst_port[j + i];
+
+       }
+
+       for (; j < nb_rx; j++) {
+               pkts_burst[j]->port = em_get_dst_port(qconf, pkts_burst[j],
+                                                     pkts_burst[j]->port);
+               process_packet(pkts_burst[j], &pkts_burst[j]->port);
+       }
+}
 #endif /* __L3FWD_EM_HLM_H__ */