dxp->next = VQ_RING_DESC_CHAIN_END;
 }
 
-static inline void
+void
 virtio_update_packet_stats(struct virtnet_stats *stats, struct rte_mbuf *mbuf)
 {
        uint32_t s = mbuf->pkt_len;
 
 };
 
 int virtio_rxq_vec_setup(struct virtnet_rx *rxvq);
+void virtio_update_packet_stats(struct virtnet_stats *stats,
+                               struct rte_mbuf *mbuf);
 
 #endif /* _VIRTIO_RXTX_H_ */
 
        struct vring_used_elem *rused;
        struct rte_mbuf **sw_ring;
        struct rte_mbuf **sw_ring_end;
+       struct rte_mbuf **ref_rx_pkts;
        uint16_t nb_pkts_received = 0;
 
        uint8x16_t shuf_msk1 = {
                        virtqueue_notify(vq);
        }
 
+       ref_rx_pkts = rx_pkts;
        for (nb_pkts_received = 0;
                nb_pkts_received < nb_used;) {
                uint64x2_t desc[RTE_VIRTIO_DESC_PER_LOOP / 2];
        vq->vq_used_cons_idx += nb_pkts_received;
        vq->vq_free_cnt += nb_pkts_received;
        rxvq->stats.packets += nb_pkts_received;
+       for (nb_used = 0; nb_used < nb_pkts_received; nb_used++)
+               virtio_update_packet_stats(&rxvq->stats, ref_rx_pkts[nb_used]);
+
        return nb_pkts_received;
 }
 
        struct vring_used_elem *rused;
        struct rte_mbuf **sw_ring;
        struct rte_mbuf **sw_ring_end;
+       struct rte_mbuf **ref_rx_pkts;
        uint16_t nb_pkts_received = 0;
        __m128i shuf_msk1, shuf_msk2, len_adjust;
 
                        virtqueue_notify(vq);
        }
 
+       ref_rx_pkts = rx_pkts;
        for (nb_pkts_received = 0;
                nb_pkts_received < nb_used;) {
                __m128i desc[RTE_VIRTIO_DESC_PER_LOOP / 2];
        vq->vq_used_cons_idx += nb_pkts_received;
        vq->vq_free_cnt += nb_pkts_received;
        rxvq->stats.packets += nb_pkts_received;
+       for (nb_used = 0; nb_used < nb_pkts_received; nb_used++)
+               virtio_update_packet_stats(&rxvq->stats, ref_rx_pkts[nb_used]);
+
        return nb_pkts_received;
 }