X-Git-Url: http://git.droids-corp.org/?a=blobdiff_plain;f=lib%2Fip_frag%2Frte_ipv4_fragmentation.c;h=fead5a985a88fbd9ac4f1d69f3d6ddd99688623d;hb=cf8f6aa12a339c3140f71acde18b3dbe91cdb1b3;hp=e9de335ae20109ceaadff903d7eb346d12fc990a;hpb=99a2dd955fba6e4cc23b77d590a033650ced9c45;p=dpdk.git diff --git a/lib/ip_frag/rte_ipv4_fragmentation.c b/lib/ip_frag/rte_ipv4_fragmentation.c index e9de335ae2..fead5a985a 100644 --- a/lib/ip_frag/rte_ipv4_fragmentation.c +++ b/lib/ip_frag/rte_ipv4_fragmentation.c @@ -23,10 +23,10 @@ #define IPV4_HDR_FO_ALIGN (1 << RTE_IPV4_HDR_FO_SHIFT) static inline void __fill_ipv4hdr_frag(struct rte_ipv4_hdr *dst, - const struct rte_ipv4_hdr *src, uint16_t len, uint16_t fofs, - uint16_t dofs, uint32_t mf) + const struct rte_ipv4_hdr *src, uint16_t header_len, + uint16_t len, uint16_t fofs, uint16_t dofs, uint32_t mf) { - rte_memcpy(dst, src, sizeof(*dst)); + rte_memcpy(dst, src, header_len); fofs = (uint16_t)(fofs + (dofs >> RTE_IPV4_HDR_FO_SHIFT)); fofs = (uint16_t)(fofs | mf << RTE_IPV4_HDR_MF_SHIFT); dst->fragment_offset = rte_cpu_to_be_16(fofs); @@ -74,8 +74,8 @@ rte_ipv4_fragment_packet(struct rte_mbuf *pkt_in, struct rte_ipv4_hdr *in_hdr; uint32_t out_pkt_pos, in_seg_data_pos; uint32_t more_in_segs; - uint16_t fragment_offset, flag_offset, frag_size; - uint16_t frag_bytes_remaining; + uint16_t fragment_offset, flag_offset, frag_size, header_len; + uint16_t frag_bytes_remaining, not_last_frag; /* * Formal parameter checking. @@ -86,14 +86,22 @@ rte_ipv4_fragment_packet(struct rte_mbuf *pkt_in, unlikely(mtu_size < RTE_ETHER_MIN_MTU)) return -EINVAL; + in_hdr = rte_pktmbuf_mtod(pkt_in, struct rte_ipv4_hdr *); + header_len = (in_hdr->version_ihl & RTE_IPV4_HDR_IHL_MASK) * + RTE_IPV4_IHL_MULTIPLIER; + + /* Check IP header length */ + if (unlikely(pkt_in->data_len < header_len) || + unlikely(mtu_size < header_len)) + return -EINVAL; + /* * Ensure the IP payload length of all fragments is aligned to a * multiple of 8 bytes as per RFC791 section 2.3. */ - frag_size = RTE_ALIGN_FLOOR((mtu_size - sizeof(struct rte_ipv4_hdr)), + frag_size = RTE_ALIGN_FLOOR((mtu_size - header_len), IPV4_HDR_FO_ALIGN); - in_hdr = rte_pktmbuf_mtod(pkt_in, struct rte_ipv4_hdr *); flag_offset = rte_cpu_to_be_16(in_hdr->fragment_offset); /* If Don't Fragment flag is set */ @@ -102,13 +110,15 @@ rte_ipv4_fragment_packet(struct rte_mbuf *pkt_in, /* Check that pkts_out is big enough to hold all fragments */ if (unlikely(frag_size * nb_pkts_out < - (uint16_t)(pkt_in->pkt_len - sizeof(struct rte_ipv4_hdr)))) + (uint16_t)(pkt_in->pkt_len - header_len))) return -EINVAL; in_seg = pkt_in; - in_seg_data_pos = sizeof(struct rte_ipv4_hdr); + in_seg_data_pos = header_len; out_pkt_pos = 0; - fragment_offset = 0; + fragment_offset = (uint16_t)((flag_offset & + RTE_IPV4_HDR_OFFSET_MASK) << RTE_IPV4_HDR_FO_SHIFT); + not_last_frag = (uint16_t)(flag_offset & IPV4_HDR_MF_MASK); more_in_segs = 1; while (likely(more_in_segs)) { @@ -124,8 +134,8 @@ rte_ipv4_fragment_packet(struct rte_mbuf *pkt_in, } /* Reserve space for the IP header that will be built later */ - out_pkt->data_len = sizeof(struct rte_ipv4_hdr); - out_pkt->pkt_len = sizeof(struct rte_ipv4_hdr); + out_pkt->data_len = header_len; + out_pkt->pkt_len = header_len; frag_bytes_remaining = frag_size; out_seg_prev = out_pkt; @@ -176,14 +186,15 @@ rte_ipv4_fragment_packet(struct rte_mbuf *pkt_in, out_hdr = rte_pktmbuf_mtod(out_pkt, struct rte_ipv4_hdr *); - __fill_ipv4hdr_frag(out_hdr, in_hdr, + __fill_ipv4hdr_frag(out_hdr, in_hdr, header_len, (uint16_t)out_pkt->pkt_len, - flag_offset, fragment_offset, more_in_segs); + flag_offset, fragment_offset, + not_last_frag || more_in_segs); fragment_offset = (uint16_t)(fragment_offset + - out_pkt->pkt_len - sizeof(struct rte_ipv4_hdr)); + out_pkt->pkt_len - header_len); - out_pkt->l3_len = sizeof(struct rte_ipv4_hdr); + out_pkt->l3_len = header_len; /* Write the fragment to the output list */ pkts_out[out_pkt_pos] = out_pkt;