mbuf: add raw allocation function
[dpdk.git] / drivers / net / virtio / virtio_rxtx.c
index 2b88efd..04f6832 100644 (file)
@@ -271,8 +271,6 @@ virtqueue_enqueue_xmit(struct virtqueue *txvq, struct rte_mbuf *cookie,
                idx = start_dp[idx].next;
        } while ((cookie = cookie->next) != NULL);
 
-       start_dp[idx].flags &= ~VRING_DESC_F_NEXT;
-
        if (use_indirect)
                idx = txvq->vq_ring.desc[head_idx].next;
 
@@ -283,17 +281,6 @@ virtqueue_enqueue_xmit(struct virtqueue *txvq, struct rte_mbuf *cookie,
        vq_update_avail_ring(txvq, head_idx);
 }
 
-static inline struct rte_mbuf *
-rte_rxmbuf_alloc(struct rte_mempool *mp)
-{
-       struct rte_mbuf *m;
-
-       m = __rte_mbuf_raw_alloc(mp);
-       __rte_mbuf_sanity_check_raw(m, 0);
-
-       return m;
-}
-
 static void
 virtio_dev_vring_start(struct virtqueue *vq, int queue_type)
 {
@@ -345,7 +332,7 @@ virtio_dev_vring_start(struct virtqueue *vq, int queue_type)
                        vq->sw_ring[vq->vq_nentries + i] = &vq->fake_mbuf;
 
                while (!virtqueue_full(vq)) {
-                       m = rte_rxmbuf_alloc(vq->mpool);
+                       m = rte_mbuf_raw_alloc(vq->mpool);
                        if (m == NULL)
                                break;
 
@@ -377,7 +364,7 @@ virtio_dev_vring_start(struct virtqueue *vq, int queue_type)
                                vq->vq_ring.desc[i + mid_idx].next = i;
                                vq->vq_ring.desc[i + mid_idx].addr =
                                        vq->virtio_net_hdr_mem +
-                                               i * vq->hw->vtnet_hdr_size;
+                                       offsetof(struct virtio_tx_region, tx_hdr);
                                vq->vq_ring.desc[i + mid_idx].len =
                                        vq->hw->vtnet_hdr_size;
                                vq->vq_ring.desc[i + mid_idx].flags =
@@ -660,7 +647,7 @@ virtio_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts, uint16_t nb_pkts)
        /* Allocate new mbuf for the used descriptor */
        error = ENOSPC;
        while (likely(!virtqueue_full(rxvq))) {
-               new_mbuf = rte_rxmbuf_alloc(rxvq->mpool);
+               new_mbuf = rte_mbuf_raw_alloc(rxvq->mpool);
                if (unlikely(new_mbuf == NULL)) {
                        struct rte_eth_dev *dev
                                = &rte_eth_devices[rxvq->port_id];
@@ -680,7 +667,7 @@ virtio_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts, uint16_t nb_pkts)
 
                if (unlikely(virtqueue_kick_prepare(rxvq))) {
                        virtqueue_notify(rxvq);
-                       PMD_RX_LOG(DEBUG, "Notified\n");
+                       PMD_RX_LOG(DEBUG, "Notified");
                }
        }
 
@@ -710,7 +697,7 @@ virtio_recv_mergeable_pkts(void *rx_queue,
 
        virtio_rmb();
 
-       PMD_RX_LOG(DEBUG, "used:%d\n", nb_used);
+       PMD_RX_LOG(DEBUG, "used:%d", nb_used);
 
        hw = rxvq->hw;
        nb_rx = 0;
@@ -733,13 +720,13 @@ virtio_recv_mergeable_pkts(void *rx_queue,
 
                i++;
 
-               PMD_RX_LOG(DEBUG, "dequeue:%d\n", num);
-               PMD_RX_LOG(DEBUG, "packet len:%d\n", len[0]);
+               PMD_RX_LOG(DEBUG, "dequeue:%d", num);
+               PMD_RX_LOG(DEBUG, "packet len:%d", len[0]);
 
                rxm = rcv_pkts[0];
 
                if (unlikely(len[0] < hdr_size + ETHER_HDR_LEN)) {
-                       PMD_RX_LOG(ERR, "Packet drop\n");
+                       PMD_RX_LOG(ERR, "Packet drop");
                        nb_enqueued++;
                        virtio_discard_rxbuf(rxvq, rxm);
                        rxvq->errors++;
@@ -781,7 +768,7 @@ virtio_recv_mergeable_pkts(void *rx_queue,
                                rcv_cnt = rx_num;
                        } else {
                                PMD_RX_LOG(ERR,
-                                       "No enough segments for packet.\n");
+                                          "No enough segments for packet.");
                                nb_enqueued++;
                                virtio_discard_rxbuf(rxvq, rxm);
                                rxvq->errors++;
@@ -824,7 +811,7 @@ virtio_recv_mergeable_pkts(void *rx_queue,
        /* Allocate new mbuf for the used descriptor */
        error = ENOSPC;
        while (likely(!virtqueue_full(rxvq))) {
-               new_mbuf = rte_rxmbuf_alloc(rxvq->mpool);
+               new_mbuf = rte_mbuf_raw_alloc(rxvq->mpool);
                if (unlikely(new_mbuf == NULL)) {
                        struct rte_eth_dev *dev
                                = &rte_eth_devices[rxvq->port_id];