net/bnx2x: get cache line size from build configuration
[dpdk.git] / drivers / net / virtio / virtio_rxtx.c
index fc0d84b..b4c4aa4 100644 (file)
@@ -53,6 +53,8 @@
 #include <rte_cpuflags.h>
 #include <rte_net.h>
 #include <rte_ip.h>
+#include <rte_udp.h>
+#include <rte_tcp.h>
 
 #include "virtio_logs.h"
 #include "virtio_ethdev.h"
@@ -207,18 +209,70 @@ virtqueue_enqueue_recv_refill(struct virtqueue *vq, struct rte_mbuf *cookie)
        return 0;
 }
 
+/* When doing TSO, the IP length is not included in the pseudo header
+ * checksum of the packet given to the PMD, but for virtio it is
+ * expected.
+ */
+static void
+virtio_tso_fix_cksum(struct rte_mbuf *m)
+{
+       /* common case: header is not fragmented */
+       if (likely(rte_pktmbuf_data_len(m) >= m->l2_len + m->l3_len +
+                       m->l4_len)) {
+               struct ipv4_hdr *iph;
+               struct ipv6_hdr *ip6h;
+               struct tcp_hdr *th;
+               uint16_t prev_cksum, new_cksum, ip_len, ip_paylen;
+               uint32_t tmp;
+
+               iph = rte_pktmbuf_mtod_offset(m, struct ipv4_hdr *, m->l2_len);
+               th = RTE_PTR_ADD(iph, m->l3_len);
+               if ((iph->version_ihl >> 4) == 4) {
+                       iph->hdr_checksum = 0;
+                       iph->hdr_checksum = rte_ipv4_cksum(iph);
+                       ip_len = iph->total_length;
+                       ip_paylen = rte_cpu_to_be_16(rte_be_to_cpu_16(ip_len) -
+                               m->l3_len);
+               } else {
+                       ip6h = (struct ipv6_hdr *)iph;
+                       ip_paylen = ip6h->payload_len;
+               }
+
+               /* calculate the new phdr checksum not including ip_paylen */
+               prev_cksum = th->cksum;
+               tmp = prev_cksum;
+               tmp += ip_paylen;
+               tmp = (tmp & 0xffff) + (tmp >> 16);
+               new_cksum = tmp;
+
+               /* replace it in the packet */
+               th->cksum = new_cksum;
+       }
+}
+
+static inline int
+tx_offload_enabled(struct virtio_hw *hw)
+{
+       return vtpci_with_feature(hw, VIRTIO_NET_F_CSUM) ||
+               vtpci_with_feature(hw, VIRTIO_NET_F_HOST_TSO4) ||
+               vtpci_with_feature(hw, VIRTIO_NET_F_HOST_TSO6);
+}
+
 static inline void
 virtqueue_enqueue_xmit(struct virtnet_tx *txvq, struct rte_mbuf *cookie,
                       uint16_t needed, int use_indirect, int can_push)
 {
+       struct virtio_tx_region *txr = txvq->virtio_net_hdr_mz->addr;
        struct vq_desc_extra *dxp;
        struct virtqueue *vq = txvq->vq;
        struct vring_desc *start_dp;
        uint16_t seg_num = cookie->nb_segs;
        uint16_t head_idx, idx;
        uint16_t head_size = vq->hw->vtnet_hdr_size;
-       unsigned long offs;
+       struct virtio_net_hdr *hdr;
+       int offload;
 
+       offload = tx_offload_enabled(vq->hw);
        head_idx = vq->vq_desc_head_idx;
        idx = head_idx;
        dxp = &vq->vq_descx[idx];
@@ -228,10 +282,12 @@ virtqueue_enqueue_xmit(struct virtnet_tx *txvq, struct rte_mbuf *cookie,
        start_dp = vq->vq_ring.desc;
 
        if (can_push) {
-               /* put on zero'd transmit header (no offloads) */
-               void *hdr = rte_pktmbuf_prepend(cookie, head_size);
-
-               memset(hdr, 0, head_size);
+               /* prepend cannot fail, checked by caller */
+               hdr = (struct virtio_net_hdr *)
+                       rte_pktmbuf_prepend(cookie, head_size);
+               /* if offload disabled, it is not zeroed below, do it now */
+               if (offload == 0)
+                       memset(hdr, 0, head_size);
        } else if (use_indirect) {
                /* setup tx ring slot to point to indirect
                 * descriptor list stored in reserved region.
@@ -239,14 +295,11 @@ virtqueue_enqueue_xmit(struct virtnet_tx *txvq, struct rte_mbuf *cookie,
                 * the first slot in indirect ring is already preset
                 * to point to the header in reserved region
                 */
-               struct virtio_tx_region *txr = txvq->virtio_net_hdr_mz->addr;
-
-               offs = idx * sizeof(struct virtio_tx_region)
-                       + offsetof(struct virtio_tx_region, tx_indir);
-
-               start_dp[idx].addr  = txvq->virtio_net_hdr_mem + offs;
+               start_dp[idx].addr  = txvq->virtio_net_hdr_mem +
+                       RTE_PTR_DIFF(&txr[idx].tx_indir, txr);
                start_dp[idx].len   = (seg_num + 1) * sizeof(struct vring_desc);
                start_dp[idx].flags = VRING_DESC_F_INDIRECT;
+               hdr = (struct virtio_net_hdr *)&txr[idx].tx_hdr;
 
                /* loop below will fill in rest of the indirect elements */
                start_dp = txr[idx].tx_indir;
@@ -255,15 +308,59 @@ virtqueue_enqueue_xmit(struct virtnet_tx *txvq, struct rte_mbuf *cookie,
                /* setup first tx ring slot to point to header
                 * stored in reserved region.
                 */
-               offs = idx * sizeof(struct virtio_tx_region)
-                       + offsetof(struct virtio_tx_region, tx_hdr);
-
-               start_dp[idx].addr  = txvq->virtio_net_hdr_mem + offs;
+               start_dp[idx].addr  = txvq->virtio_net_hdr_mem +
+                       RTE_PTR_DIFF(&txr[idx].tx_hdr, txr);
                start_dp[idx].len   = vq->hw->vtnet_hdr_size;
                start_dp[idx].flags = VRING_DESC_F_NEXT;
+               hdr = (struct virtio_net_hdr *)&txr[idx].tx_hdr;
+
                idx = start_dp[idx].next;
        }
 
+       /* Checksum Offload / TSO */
+       if (offload) {
+               if (cookie->ol_flags & PKT_TX_TCP_SEG)
+                       cookie->ol_flags |= PKT_TX_TCP_CKSUM;
+
+               switch (cookie->ol_flags & PKT_TX_L4_MASK) {
+               case PKT_TX_UDP_CKSUM:
+                       hdr->csum_start = cookie->l2_len + cookie->l3_len;
+                       hdr->csum_offset = offsetof(struct udp_hdr,
+                               dgram_cksum);
+                       hdr->flags = VIRTIO_NET_HDR_F_NEEDS_CSUM;
+                       break;
+
+               case PKT_TX_TCP_CKSUM:
+                       hdr->csum_start = cookie->l2_len + cookie->l3_len;
+                       hdr->csum_offset = offsetof(struct tcp_hdr, cksum);
+                       hdr->flags = VIRTIO_NET_HDR_F_NEEDS_CSUM;
+                       break;
+
+               default:
+                       hdr->csum_start = 0;
+                       hdr->csum_offset = 0;
+                       hdr->flags = 0;
+                       break;
+               }
+
+               /* TCP Segmentation Offload */
+               if (cookie->ol_flags & PKT_TX_TCP_SEG) {
+                       virtio_tso_fix_cksum(cookie);
+                       hdr->gso_type = (cookie->ol_flags & PKT_TX_IPV6) ?
+                               VIRTIO_NET_HDR_GSO_TCPV6 :
+                               VIRTIO_NET_HDR_GSO_TCPV4;
+                       hdr->gso_size = cookie->tso_segsz;
+                       hdr->hdr_len =
+                               cookie->l2_len +
+                               cookie->l3_len +
+                               cookie->l4_len;
+               } else {
+                       hdr->gso_type = 0;
+                       hdr->gso_size = 0;
+                       hdr->hdr_len = 0;
+               }
+       }
+
        do {
                start_dp[idx].addr  = VIRTIO_MBUF_DATA_DMA_ADDR(cookie, vq);
                start_dp[idx].len   = cookie->data_len;
@@ -527,11 +624,6 @@ virtio_dev_tx_queue_setup(struct rte_eth_dev *dev,
 
        PMD_INIT_FUNC_TRACE();
 
-       if ((tx_conf->txq_flags & ETH_TXQ_FLAGS_NOXSUMS)
-           != ETH_TXQ_FLAGS_NOXSUMS) {
-               PMD_INIT_LOG(ERR, "TX checksum offload not supported\n");
-               return -EINVAL;
-       }
 
        virtio_update_rxtx_handler(dev, tx_conf);
 
@@ -682,13 +774,36 @@ virtio_rx_offload(struct rte_mbuf *m, struct virtio_net_hdr *hdr)
                m->ol_flags |= PKT_RX_L4_CKSUM_GOOD;
        }
 
+       /* GSO request, save required information in mbuf */
+       if (hdr->gso_type != VIRTIO_NET_HDR_GSO_NONE) {
+               /* Check unsupported modes */
+               if ((hdr->gso_type & VIRTIO_NET_HDR_GSO_ECN) ||
+                   (hdr->gso_size == 0)) {
+                       return -EINVAL;
+               }
+
+               /* Update mss lengthes in mbuf */
+               m->tso_segsz = hdr->gso_size;
+               switch (hdr->gso_type & ~VIRTIO_NET_HDR_GSO_ECN) {
+                       case VIRTIO_NET_HDR_GSO_TCPV4:
+                       case VIRTIO_NET_HDR_GSO_TCPV6:
+                               m->ol_flags |= PKT_RX_LRO | \
+                                       PKT_RX_L4_CKSUM_NONE;
+                               break;
+                       default:
+                               return -EINVAL;
+               }
+       }
+
        return 0;
 }
 
 static inline int
 rx_offload_enabled(struct virtio_hw *hw)
 {
-       return vtpci_with_feature(hw, VIRTIO_NET_F_GUEST_CSUM);
+       return vtpci_with_feature(hw, VIRTIO_NET_F_GUEST_CSUM) ||
+               vtpci_with_feature(hw, VIRTIO_NET_F_GUEST_TSO4) ||
+               vtpci_with_feature(hw, VIRTIO_NET_F_GUEST_TSO6);
 }
 
 #define VIRTIO_MBUF_BURST_SZ 64