4 * Copyright (c) 2015 - 2016 CESNET
7 * Redistribution and use in source and binary forms, with or without
8 * modification, are permitted provided that the following conditions
11 * * Redistributions of source code must retain the above copyright
12 * notice, this list of conditions and the following disclaimer.
13 * * Redistributions in binary form must reproduce the above copyright
14 * notice, this list of conditions and the following disclaimer in
15 * the documentation and/or other materials provided with the
17 * * Neither the name of CESNET nor the names of its
18 * contributors may be used to endorse or promote products derived
19 * from this software without specific prior written permission.
21 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
22 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
23 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
24 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
25 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
26 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
27 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
28 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
29 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
30 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
31 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
38 #include <sys/types.h>
47 #include <rte_ethdev.h>
48 #include <rte_ethdev_pci.h>
49 #include <rte_malloc.h>
50 #include <rte_memcpy.h>
51 #include <rte_kvargs.h>
53 #include <rte_atomic.h>
55 #include "rte_eth_szedata2.h"
56 #include "szedata2_iobuf.h"
58 #define RTE_ETH_SZEDATA2_MAX_RX_QUEUES 32
59 #define RTE_ETH_SZEDATA2_MAX_TX_QUEUES 32
60 #define RTE_ETH_SZEDATA2_TX_LOCK_SIZE (32 * 1024 * 1024)
63 * size of szedata2_packet header with alignment
65 #define RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED 8
67 #define RTE_SZEDATA2_DRIVER_NAME net_szedata2
69 #define SZEDATA2_DEV_PATH_FMT "/dev/szedataII%u"
71 struct szedata2_rx_queue {
75 struct rte_mempool *mb_pool;
76 volatile uint64_t rx_pkts;
77 volatile uint64_t rx_bytes;
78 volatile uint64_t err_pkts;
81 struct szedata2_tx_queue {
84 volatile uint64_t tx_pkts;
85 volatile uint64_t tx_bytes;
86 volatile uint64_t err_pkts;
89 struct pmd_internals {
90 struct szedata2_rx_queue rx_queue[RTE_ETH_SZEDATA2_MAX_RX_QUEUES];
91 struct szedata2_tx_queue tx_queue[RTE_ETH_SZEDATA2_MAX_TX_QUEUES];
92 uint16_t max_rx_queues;
93 uint16_t max_tx_queues;
94 char sze_dev[PATH_MAX];
95 struct rte_mem_resource *pci_rsc;
98 static struct ether_addr eth_addr = {
99 .addr_bytes = { 0x00, 0x11, 0x17, 0x00, 0x00, 0x00 }
103 eth_szedata2_rx(void *queue,
104 struct rte_mbuf **bufs,
108 struct rte_mbuf *mbuf;
109 struct szedata2_rx_queue *sze_q = queue;
110 struct rte_pktmbuf_pool_private *mbp_priv;
115 uint16_t packet_size;
116 uint64_t num_bytes = 0;
117 struct szedata *sze = sze_q->sze;
118 uint8_t *header_ptr = NULL; /* header of packet */
119 uint8_t *packet_ptr1 = NULL;
120 uint8_t *packet_ptr2 = NULL;
121 uint16_t packet_len1 = 0;
122 uint16_t packet_len2 = 0;
123 uint16_t hw_data_align;
125 if (unlikely(sze_q->sze == NULL || nb_pkts == 0))
129 * Reads the given number of packets from szedata2 channel given
130 * by queue and copies the packet data into a newly allocated mbuf
133 for (i = 0; i < nb_pkts; i++) {
134 mbuf = rte_pktmbuf_alloc(sze_q->mb_pool);
136 if (unlikely(mbuf == NULL))
139 /* get the next sze packet */
140 if (sze->ct_rx_lck != NULL && !sze->ct_rx_rem_bytes &&
141 sze->ct_rx_lck->next == NULL) {
142 /* unlock old data */
143 szedata_rx_unlock_data(sze_q->sze, sze->ct_rx_lck_orig);
144 sze->ct_rx_lck_orig = NULL;
145 sze->ct_rx_lck = NULL;
148 if (!sze->ct_rx_rem_bytes && sze->ct_rx_lck_orig == NULL) {
149 /* nothing to read, lock new data */
150 sze->ct_rx_lck = szedata_rx_lock_data(sze_q->sze, ~0U);
151 sze->ct_rx_lck_orig = sze->ct_rx_lck;
153 if (sze->ct_rx_lck == NULL) {
154 /* nothing to lock */
155 rte_pktmbuf_free(mbuf);
159 sze->ct_rx_cur_ptr = sze->ct_rx_lck->start;
160 sze->ct_rx_rem_bytes = sze->ct_rx_lck->len;
162 if (!sze->ct_rx_rem_bytes) {
163 rte_pktmbuf_free(mbuf);
168 if (sze->ct_rx_rem_bytes < RTE_SZE2_PACKET_HEADER_SIZE) {
171 * copy parts of header to merge buffer
173 if (sze->ct_rx_lck->next == NULL) {
174 rte_pktmbuf_free(mbuf);
178 /* copy first part of header */
179 rte_memcpy(sze->ct_rx_buffer, sze->ct_rx_cur_ptr,
180 sze->ct_rx_rem_bytes);
182 /* copy second part of header */
183 sze->ct_rx_lck = sze->ct_rx_lck->next;
184 sze->ct_rx_cur_ptr = sze->ct_rx_lck->start;
185 rte_memcpy(sze->ct_rx_buffer + sze->ct_rx_rem_bytes,
187 RTE_SZE2_PACKET_HEADER_SIZE -
188 sze->ct_rx_rem_bytes);
190 sze->ct_rx_cur_ptr += RTE_SZE2_PACKET_HEADER_SIZE -
191 sze->ct_rx_rem_bytes;
192 sze->ct_rx_rem_bytes = sze->ct_rx_lck->len -
193 RTE_SZE2_PACKET_HEADER_SIZE +
194 sze->ct_rx_rem_bytes;
196 header_ptr = (uint8_t *)sze->ct_rx_buffer;
199 header_ptr = (uint8_t *)sze->ct_rx_cur_ptr;
200 sze->ct_rx_cur_ptr += RTE_SZE2_PACKET_HEADER_SIZE;
201 sze->ct_rx_rem_bytes -= RTE_SZE2_PACKET_HEADER_SIZE;
204 sg_size = le16toh(*((uint16_t *)header_ptr));
205 hw_size = le16toh(*(((uint16_t *)header_ptr) + 1));
206 packet_size = sg_size -
207 RTE_SZE2_ALIGN8(RTE_SZE2_PACKET_HEADER_SIZE + hw_size);
210 /* checks if packet all right */
212 errx(5, "Zero segsize");
214 /* check sg_size and hwsize */
215 if (hw_size > sg_size - RTE_SZE2_PACKET_HEADER_SIZE) {
216 errx(10, "Hwsize bigger than expected. Segsize: %d, "
217 "hwsize: %d", sg_size, hw_size);
221 RTE_SZE2_ALIGN8(RTE_SZE2_PACKET_HEADER_SIZE + hw_size) -
222 RTE_SZE2_PACKET_HEADER_SIZE;
224 if (sze->ct_rx_rem_bytes >=
226 RTE_SZE2_PACKET_HEADER_SIZE)) {
228 /* one packet ready - go to another */
229 packet_ptr1 = sze->ct_rx_cur_ptr + hw_data_align;
230 packet_len1 = packet_size;
234 sze->ct_rx_cur_ptr += RTE_SZE2_ALIGN8(sg_size) -
235 RTE_SZE2_PACKET_HEADER_SIZE;
236 sze->ct_rx_rem_bytes -= RTE_SZE2_ALIGN8(sg_size) -
237 RTE_SZE2_PACKET_HEADER_SIZE;
240 if (sze->ct_rx_lck->next == NULL) {
241 errx(6, "Need \"next\" lock, "
242 "but it is missing: %u",
243 sze->ct_rx_rem_bytes);
247 if (sze->ct_rx_rem_bytes <= hw_data_align) {
248 uint16_t rem_size = hw_data_align -
249 sze->ct_rx_rem_bytes;
251 /* MOVE to next lock */
252 sze->ct_rx_lck = sze->ct_rx_lck->next;
254 (void *)(((uint8_t *)
255 (sze->ct_rx_lck->start)) + rem_size);
257 packet_ptr1 = sze->ct_rx_cur_ptr;
258 packet_len1 = packet_size;
262 sze->ct_rx_cur_ptr +=
263 RTE_SZE2_ALIGN8(packet_size);
264 sze->ct_rx_rem_bytes = sze->ct_rx_lck->len -
265 rem_size - RTE_SZE2_ALIGN8(packet_size);
267 /* get pointer and length from first part */
268 packet_ptr1 = sze->ct_rx_cur_ptr +
270 packet_len1 = sze->ct_rx_rem_bytes -
273 /* MOVE to next lock */
274 sze->ct_rx_lck = sze->ct_rx_lck->next;
275 sze->ct_rx_cur_ptr = sze->ct_rx_lck->start;
277 /* get pointer and length from second part */
278 packet_ptr2 = sze->ct_rx_cur_ptr;
279 packet_len2 = packet_size - packet_len1;
281 sze->ct_rx_cur_ptr +=
282 RTE_SZE2_ALIGN8(packet_size) -
284 sze->ct_rx_rem_bytes = sze->ct_rx_lck->len -
285 (RTE_SZE2_ALIGN8(packet_size) -
290 if (unlikely(packet_ptr1 == NULL)) {
291 rte_pktmbuf_free(mbuf);
295 /* get the space available for data in the mbuf */
296 mbp_priv = rte_mempool_get_priv(sze_q->mb_pool);
297 buf_size = (uint16_t)(mbp_priv->mbuf_data_room_size -
298 RTE_PKTMBUF_HEADROOM);
300 if (packet_size <= buf_size) {
301 /* sze packet will fit in one mbuf, go ahead and copy */
302 rte_memcpy(rte_pktmbuf_mtod(mbuf, void *),
303 packet_ptr1, packet_len1);
304 if (packet_ptr2 != NULL) {
305 rte_memcpy((void *)(rte_pktmbuf_mtod(mbuf,
306 uint8_t *) + packet_len1),
307 packet_ptr2, packet_len2);
309 mbuf->data_len = (uint16_t)packet_size;
311 mbuf->pkt_len = packet_size;
312 mbuf->port = sze_q->in_port;
315 num_bytes += packet_size;
318 * sze packet will not fit in one mbuf,
319 * scattered mode is not enabled, drop packet
322 "SZE segment %d bytes will not fit in one mbuf "
323 "(%d bytes), scattered mode is not enabled, "
325 packet_size, buf_size);
326 rte_pktmbuf_free(mbuf);
330 sze_q->rx_pkts += num_rx;
331 sze_q->rx_bytes += num_bytes;
336 eth_szedata2_rx_scattered(void *queue,
337 struct rte_mbuf **bufs,
341 struct rte_mbuf *mbuf;
342 struct szedata2_rx_queue *sze_q = queue;
343 struct rte_pktmbuf_pool_private *mbp_priv;
348 uint16_t packet_size;
349 uint64_t num_bytes = 0;
350 struct szedata *sze = sze_q->sze;
351 uint8_t *header_ptr = NULL; /* header of packet */
352 uint8_t *packet_ptr1 = NULL;
353 uint8_t *packet_ptr2 = NULL;
354 uint16_t packet_len1 = 0;
355 uint16_t packet_len2 = 0;
356 uint16_t hw_data_align;
358 if (unlikely(sze_q->sze == NULL || nb_pkts == 0))
362 * Reads the given number of packets from szedata2 channel given
363 * by queue and copies the packet data into a newly allocated mbuf
366 for (i = 0; i < nb_pkts; i++) {
367 const struct szedata_lock *ct_rx_lck_backup;
368 unsigned int ct_rx_rem_bytes_backup;
369 unsigned char *ct_rx_cur_ptr_backup;
371 /* get the next sze packet */
372 if (sze->ct_rx_lck != NULL && !sze->ct_rx_rem_bytes &&
373 sze->ct_rx_lck->next == NULL) {
374 /* unlock old data */
375 szedata_rx_unlock_data(sze_q->sze, sze->ct_rx_lck_orig);
376 sze->ct_rx_lck_orig = NULL;
377 sze->ct_rx_lck = NULL;
381 * Store items from sze structure which can be changed
382 * before mbuf allocating. Use these items in case of mbuf
383 * allocating failure.
385 ct_rx_lck_backup = sze->ct_rx_lck;
386 ct_rx_rem_bytes_backup = sze->ct_rx_rem_bytes;
387 ct_rx_cur_ptr_backup = sze->ct_rx_cur_ptr;
389 if (!sze->ct_rx_rem_bytes && sze->ct_rx_lck_orig == NULL) {
390 /* nothing to read, lock new data */
391 sze->ct_rx_lck = szedata_rx_lock_data(sze_q->sze, ~0U);
392 sze->ct_rx_lck_orig = sze->ct_rx_lck;
395 * Backup items from sze structure must be updated
396 * after locking to contain pointers to new locks.
398 ct_rx_lck_backup = sze->ct_rx_lck;
399 ct_rx_rem_bytes_backup = sze->ct_rx_rem_bytes;
400 ct_rx_cur_ptr_backup = sze->ct_rx_cur_ptr;
402 if (sze->ct_rx_lck == NULL)
403 /* nothing to lock */
406 sze->ct_rx_cur_ptr = sze->ct_rx_lck->start;
407 sze->ct_rx_rem_bytes = sze->ct_rx_lck->len;
409 if (!sze->ct_rx_rem_bytes)
413 if (sze->ct_rx_rem_bytes < RTE_SZE2_PACKET_HEADER_SIZE) {
415 * cut in header - copy parts of header to merge buffer
417 if (sze->ct_rx_lck->next == NULL)
420 /* copy first part of header */
421 rte_memcpy(sze->ct_rx_buffer, sze->ct_rx_cur_ptr,
422 sze->ct_rx_rem_bytes);
424 /* copy second part of header */
425 sze->ct_rx_lck = sze->ct_rx_lck->next;
426 sze->ct_rx_cur_ptr = sze->ct_rx_lck->start;
427 rte_memcpy(sze->ct_rx_buffer + sze->ct_rx_rem_bytes,
429 RTE_SZE2_PACKET_HEADER_SIZE -
430 sze->ct_rx_rem_bytes);
432 sze->ct_rx_cur_ptr += RTE_SZE2_PACKET_HEADER_SIZE -
433 sze->ct_rx_rem_bytes;
434 sze->ct_rx_rem_bytes = sze->ct_rx_lck->len -
435 RTE_SZE2_PACKET_HEADER_SIZE +
436 sze->ct_rx_rem_bytes;
438 header_ptr = (uint8_t *)sze->ct_rx_buffer;
441 header_ptr = (uint8_t *)sze->ct_rx_cur_ptr;
442 sze->ct_rx_cur_ptr += RTE_SZE2_PACKET_HEADER_SIZE;
443 sze->ct_rx_rem_bytes -= RTE_SZE2_PACKET_HEADER_SIZE;
446 sg_size = le16toh(*((uint16_t *)header_ptr));
447 hw_size = le16toh(*(((uint16_t *)header_ptr) + 1));
448 packet_size = sg_size -
449 RTE_SZE2_ALIGN8(RTE_SZE2_PACKET_HEADER_SIZE + hw_size);
452 /* checks if packet all right */
454 errx(5, "Zero segsize");
456 /* check sg_size and hwsize */
457 if (hw_size > sg_size - RTE_SZE2_PACKET_HEADER_SIZE) {
458 errx(10, "Hwsize bigger than expected. Segsize: %d, "
459 "hwsize: %d", sg_size, hw_size);
463 RTE_SZE2_ALIGN8((RTE_SZE2_PACKET_HEADER_SIZE +
464 hw_size)) - RTE_SZE2_PACKET_HEADER_SIZE;
466 if (sze->ct_rx_rem_bytes >=
468 RTE_SZE2_PACKET_HEADER_SIZE)) {
470 /* one packet ready - go to another */
471 packet_ptr1 = sze->ct_rx_cur_ptr + hw_data_align;
472 packet_len1 = packet_size;
476 sze->ct_rx_cur_ptr += RTE_SZE2_ALIGN8(sg_size) -
477 RTE_SZE2_PACKET_HEADER_SIZE;
478 sze->ct_rx_rem_bytes -= RTE_SZE2_ALIGN8(sg_size) -
479 RTE_SZE2_PACKET_HEADER_SIZE;
482 if (sze->ct_rx_lck->next == NULL) {
483 errx(6, "Need \"next\" lock, but it is "
484 "missing: %u", sze->ct_rx_rem_bytes);
488 if (sze->ct_rx_rem_bytes <= hw_data_align) {
489 uint16_t rem_size = hw_data_align -
490 sze->ct_rx_rem_bytes;
492 /* MOVE to next lock */
493 sze->ct_rx_lck = sze->ct_rx_lck->next;
495 (void *)(((uint8_t *)
496 (sze->ct_rx_lck->start)) + rem_size);
498 packet_ptr1 = sze->ct_rx_cur_ptr;
499 packet_len1 = packet_size;
503 sze->ct_rx_cur_ptr +=
504 RTE_SZE2_ALIGN8(packet_size);
505 sze->ct_rx_rem_bytes = sze->ct_rx_lck->len -
506 rem_size - RTE_SZE2_ALIGN8(packet_size);
508 /* get pointer and length from first part */
509 packet_ptr1 = sze->ct_rx_cur_ptr +
511 packet_len1 = sze->ct_rx_rem_bytes -
514 /* MOVE to next lock */
515 sze->ct_rx_lck = sze->ct_rx_lck->next;
516 sze->ct_rx_cur_ptr = sze->ct_rx_lck->start;
518 /* get pointer and length from second part */
519 packet_ptr2 = sze->ct_rx_cur_ptr;
520 packet_len2 = packet_size - packet_len1;
522 sze->ct_rx_cur_ptr +=
523 RTE_SZE2_ALIGN8(packet_size) -
525 sze->ct_rx_rem_bytes = sze->ct_rx_lck->len -
526 (RTE_SZE2_ALIGN8(packet_size) -
531 if (unlikely(packet_ptr1 == NULL))
534 mbuf = rte_pktmbuf_alloc(sze_q->mb_pool);
536 if (unlikely(mbuf == NULL)) {
538 * Restore items from sze structure to state after
539 * unlocking (eventually locking).
541 sze->ct_rx_lck = ct_rx_lck_backup;
542 sze->ct_rx_rem_bytes = ct_rx_rem_bytes_backup;
543 sze->ct_rx_cur_ptr = ct_rx_cur_ptr_backup;
547 /* get the space available for data in the mbuf */
548 mbp_priv = rte_mempool_get_priv(sze_q->mb_pool);
549 buf_size = (uint16_t)(mbp_priv->mbuf_data_room_size -
550 RTE_PKTMBUF_HEADROOM);
552 if (packet_size <= buf_size) {
553 /* sze packet will fit in one mbuf, go ahead and copy */
554 rte_memcpy(rte_pktmbuf_mtod(mbuf, void *),
555 packet_ptr1, packet_len1);
556 if (packet_ptr2 != NULL) {
558 (rte_pktmbuf_mtod(mbuf, uint8_t *) +
559 packet_len1), packet_ptr2, packet_len2);
561 mbuf->data_len = (uint16_t)packet_size;
564 * sze packet will not fit in one mbuf,
565 * scatter packet into more mbufs
567 struct rte_mbuf *m = mbuf;
568 uint16_t len = rte_pktmbuf_tailroom(mbuf);
570 /* copy first part of packet */
571 /* fill first mbuf */
572 rte_memcpy(rte_pktmbuf_append(mbuf, len), packet_ptr1,
575 packet_ptr1 = ((uint8_t *)packet_ptr1) + len;
577 while (packet_len1 > 0) {
579 m->next = rte_pktmbuf_alloc(sze_q->mb_pool);
581 if (unlikely(m->next == NULL)) {
582 rte_pktmbuf_free(mbuf);
584 * Restore items from sze structure
585 * to state after unlocking (eventually
588 sze->ct_rx_lck = ct_rx_lck_backup;
589 sze->ct_rx_rem_bytes =
590 ct_rx_rem_bytes_backup;
592 ct_rx_cur_ptr_backup;
598 len = RTE_MIN(rte_pktmbuf_tailroom(m),
600 rte_memcpy(rte_pktmbuf_append(mbuf, len),
605 packet_ptr1 = ((uint8_t *)packet_ptr1) + len;
608 if (packet_ptr2 != NULL) {
609 /* copy second part of packet, if exists */
610 /* fill the rest of currently last mbuf */
611 len = rte_pktmbuf_tailroom(m);
612 rte_memcpy(rte_pktmbuf_append(mbuf, len),
615 packet_ptr2 = ((uint8_t *)packet_ptr2) + len;
617 while (packet_len2 > 0) {
619 m->next = rte_pktmbuf_alloc(
622 if (unlikely(m->next == NULL)) {
623 rte_pktmbuf_free(mbuf);
625 * Restore items from sze
626 * structure to state after
627 * unlocking (eventually
632 sze->ct_rx_rem_bytes =
633 ct_rx_rem_bytes_backup;
635 ct_rx_cur_ptr_backup;
641 len = RTE_MIN(rte_pktmbuf_tailroom(m),
644 rte_pktmbuf_append(mbuf, len),
649 packet_ptr2 = ((uint8_t *)packet_ptr2) +
654 mbuf->pkt_len = packet_size;
655 mbuf->port = sze_q->in_port;
658 num_bytes += packet_size;
662 sze_q->rx_pkts += num_rx;
663 sze_q->rx_bytes += num_bytes;
668 eth_szedata2_tx(void *queue,
669 struct rte_mbuf **bufs,
672 struct rte_mbuf *mbuf;
673 struct szedata2_tx_queue *sze_q = queue;
675 uint64_t num_bytes = 0;
677 const struct szedata_lock *lck;
683 uint32_t unlock_size;
686 uint16_t pkt_left = nb_pkts;
688 if (sze_q->sze == NULL || nb_pkts == 0)
691 while (pkt_left > 0) {
693 lck = szedata_tx_lock_data(sze_q->sze,
694 RTE_ETH_SZEDATA2_TX_LOCK_SIZE,
700 lock_size = lck->len;
701 lock_size2 = lck->next ? lck->next->len : 0;
704 mbuf = bufs[nb_pkts - pkt_left];
706 pkt_len = mbuf->pkt_len;
707 mbuf_segs = mbuf->nb_segs;
709 hwpkt_len = RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED +
710 RTE_SZE2_ALIGN8(pkt_len);
712 if (lock_size + lock_size2 < hwpkt_len) {
713 szedata_tx_unlock_data(sze_q->sze, lck, unlock_size);
717 num_bytes += pkt_len;
719 if (lock_size > hwpkt_len) {
724 /* write packet length at first 2 bytes in 8B header */
725 *((uint16_t *)dst) = htole16(
726 RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED +
728 *(((uint16_t *)dst) + 1) = htole16(0);
730 /* copy packet from mbuf */
731 tmp_dst = ((uint8_t *)(dst)) +
732 RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED;
733 if (mbuf_segs == 1) {
735 * non-scattered packet,
736 * transmit from one mbuf
739 rte_pktmbuf_mtod(mbuf, const void *),
742 /* scattered packet, transmit from more mbufs */
743 struct rte_mbuf *m = mbuf;
749 tmp_dst = ((uint8_t *)(tmp_dst)) +
756 dst = ((uint8_t *)dst) + hwpkt_len;
757 unlock_size += hwpkt_len;
758 lock_size -= hwpkt_len;
760 rte_pktmbuf_free(mbuf);
764 szedata_tx_unlock_data(sze_q->sze, lck,
769 } else if (lock_size + lock_size2 >= hwpkt_len) {
773 /* write packet length at first 2 bytes in 8B header */
775 htole16(RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED +
777 *(((uint16_t *)dst) + 1) = htole16(0);
780 * If the raw packet (pkt_len) is smaller than lock_size
781 * get the correct length for memcpy
784 pkt_len < lock_size -
785 RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED ?
787 lock_size - RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED;
789 rem_len = hwpkt_len - lock_size;
791 tmp_dst = ((uint8_t *)(dst)) +
792 RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED;
793 if (mbuf_segs == 1) {
795 * non-scattered packet,
796 * transmit from one mbuf
798 /* copy part of packet to first area */
800 rte_pktmbuf_mtod(mbuf, const void *),
804 dst = lck->next->start;
806 /* copy part of packet to second area */
808 (const void *)(rte_pktmbuf_mtod(mbuf,
810 write_len), pkt_len - write_len);
812 /* scattered packet, transmit from more mbufs */
813 struct rte_mbuf *m = mbuf;
814 uint16_t written = 0;
815 uint16_t to_write = 0;
816 bool new_mbuf = true;
817 uint16_t write_off = 0;
819 /* copy part of packet to first area */
820 while (m && written < write_len) {
821 to_write = RTE_MIN(m->data_len,
822 write_len - written);
828 tmp_dst = ((uint8_t *)(tmp_dst)) +
830 if (m->data_len <= write_len -
841 dst = lck->next->start;
845 write_off = new_mbuf ? 0 : to_write;
847 /* copy part of packet to second area */
848 while (m && written < pkt_len - write_len) {
849 rte_memcpy(tmp_dst, (const void *)
851 uint8_t *) + write_off),
852 m->data_len - write_off);
854 tmp_dst = ((uint8_t *)(tmp_dst)) +
855 (m->data_len - write_off);
856 written += m->data_len - write_off;
862 dst = ((uint8_t *)dst) + rem_len;
863 unlock_size += hwpkt_len;
864 lock_size = lock_size2 - rem_len;
867 rte_pktmbuf_free(mbuf);
871 szedata_tx_unlock_data(sze_q->sze, lck, unlock_size);
875 sze_q->tx_pkts += num_tx;
876 sze_q->err_pkts += nb_pkts - num_tx;
877 sze_q->tx_bytes += num_bytes;
882 eth_rx_queue_start(struct rte_eth_dev *dev, uint16_t rxq_id)
884 struct szedata2_rx_queue *rxq = dev->data->rx_queues[rxq_id];
886 struct pmd_internals *internals = (struct pmd_internals *)
887 dev->data->dev_private;
889 if (rxq->sze == NULL) {
890 uint32_t rx = 1 << rxq->rx_channel;
892 rxq->sze = szedata_open(internals->sze_dev);
893 if (rxq->sze == NULL)
895 ret = szedata_subscribe3(rxq->sze, &rx, &tx);
896 if (ret != 0 || rx == 0)
900 ret = szedata_start(rxq->sze);
903 dev->data->rx_queue_state[rxq_id] = RTE_ETH_QUEUE_STATE_STARTED;
907 szedata_close(rxq->sze);
913 eth_rx_queue_stop(struct rte_eth_dev *dev, uint16_t rxq_id)
915 struct szedata2_rx_queue *rxq = dev->data->rx_queues[rxq_id];
917 if (rxq->sze != NULL) {
918 szedata_close(rxq->sze);
922 dev->data->rx_queue_state[rxq_id] = RTE_ETH_QUEUE_STATE_STOPPED;
927 eth_tx_queue_start(struct rte_eth_dev *dev, uint16_t txq_id)
929 struct szedata2_tx_queue *txq = dev->data->tx_queues[txq_id];
931 struct pmd_internals *internals = (struct pmd_internals *)
932 dev->data->dev_private;
934 if (txq->sze == NULL) {
936 uint32_t tx = 1 << txq->tx_channel;
937 txq->sze = szedata_open(internals->sze_dev);
938 if (txq->sze == NULL)
940 ret = szedata_subscribe3(txq->sze, &rx, &tx);
941 if (ret != 0 || tx == 0)
945 ret = szedata_start(txq->sze);
948 dev->data->tx_queue_state[txq_id] = RTE_ETH_QUEUE_STATE_STARTED;
952 szedata_close(txq->sze);
958 eth_tx_queue_stop(struct rte_eth_dev *dev, uint16_t txq_id)
960 struct szedata2_tx_queue *txq = dev->data->tx_queues[txq_id];
962 if (txq->sze != NULL) {
963 szedata_close(txq->sze);
967 dev->data->tx_queue_state[txq_id] = RTE_ETH_QUEUE_STATE_STOPPED;
972 eth_dev_start(struct rte_eth_dev *dev)
976 uint16_t nb_rx = dev->data->nb_rx_queues;
977 uint16_t nb_tx = dev->data->nb_tx_queues;
979 for (i = 0; i < nb_rx; i++) {
980 ret = eth_rx_queue_start(dev, i);
985 for (i = 0; i < nb_tx; i++) {
986 ret = eth_tx_queue_start(dev, i);
994 for (i = 0; i < nb_tx; i++)
995 eth_tx_queue_stop(dev, i);
997 for (i = 0; i < nb_rx; i++)
998 eth_rx_queue_stop(dev, i);
1003 eth_dev_stop(struct rte_eth_dev *dev)
1006 uint16_t nb_rx = dev->data->nb_rx_queues;
1007 uint16_t nb_tx = dev->data->nb_tx_queues;
1009 for (i = 0; i < nb_tx; i++)
1010 eth_tx_queue_stop(dev, i);
1012 for (i = 0; i < nb_rx; i++)
1013 eth_rx_queue_stop(dev, i);
1017 eth_dev_configure(struct rte_eth_dev *dev)
1019 struct rte_eth_dev_data *data = dev->data;
1020 if (data->dev_conf.rxmode.enable_scatter == 1) {
1021 dev->rx_pkt_burst = eth_szedata2_rx_scattered;
1022 data->scattered_rx = 1;
1024 dev->rx_pkt_burst = eth_szedata2_rx;
1025 data->scattered_rx = 0;
1031 eth_dev_info(struct rte_eth_dev *dev,
1032 struct rte_eth_dev_info *dev_info)
1034 struct pmd_internals *internals = dev->data->dev_private;
1035 dev_info->pci_dev = RTE_ETH_DEV_TO_PCI(dev);
1036 dev_info->if_index = 0;
1037 dev_info->max_mac_addrs = 1;
1038 dev_info->max_rx_pktlen = (uint32_t)-1;
1039 dev_info->max_rx_queues = internals->max_rx_queues;
1040 dev_info->max_tx_queues = internals->max_tx_queues;
1041 dev_info->min_rx_bufsize = 0;
1042 dev_info->speed_capa = ETH_LINK_SPEED_100G;
1046 eth_stats_get(struct rte_eth_dev *dev,
1047 struct rte_eth_stats *stats)
1050 uint16_t nb_rx = dev->data->nb_rx_queues;
1051 uint16_t nb_tx = dev->data->nb_tx_queues;
1052 uint64_t rx_total = 0;
1053 uint64_t tx_total = 0;
1054 uint64_t tx_err_total = 0;
1055 uint64_t rx_total_bytes = 0;
1056 uint64_t tx_total_bytes = 0;
1057 const struct pmd_internals *internals = dev->data->dev_private;
1059 for (i = 0; i < RTE_ETHDEV_QUEUE_STAT_CNTRS && i < nb_rx; i++) {
1060 stats->q_ipackets[i] = internals->rx_queue[i].rx_pkts;
1061 stats->q_ibytes[i] = internals->rx_queue[i].rx_bytes;
1062 rx_total += stats->q_ipackets[i];
1063 rx_total_bytes += stats->q_ibytes[i];
1066 for (i = 0; i < RTE_ETHDEV_QUEUE_STAT_CNTRS && i < nb_tx; i++) {
1067 stats->q_opackets[i] = internals->tx_queue[i].tx_pkts;
1068 stats->q_obytes[i] = internals->tx_queue[i].tx_bytes;
1069 stats->q_errors[i] = internals->tx_queue[i].err_pkts;
1070 tx_total += stats->q_opackets[i];
1071 tx_total_bytes += stats->q_obytes[i];
1072 tx_err_total += stats->q_errors[i];
1075 stats->ipackets = rx_total;
1076 stats->opackets = tx_total;
1077 stats->ibytes = rx_total_bytes;
1078 stats->obytes = tx_total_bytes;
1079 stats->oerrors = tx_err_total;
1085 eth_stats_reset(struct rte_eth_dev *dev)
1088 uint16_t nb_rx = dev->data->nb_rx_queues;
1089 uint16_t nb_tx = dev->data->nb_tx_queues;
1090 struct pmd_internals *internals = dev->data->dev_private;
1092 for (i = 0; i < nb_rx; i++) {
1093 internals->rx_queue[i].rx_pkts = 0;
1094 internals->rx_queue[i].rx_bytes = 0;
1095 internals->rx_queue[i].err_pkts = 0;
1097 for (i = 0; i < nb_tx; i++) {
1098 internals->tx_queue[i].tx_pkts = 0;
1099 internals->tx_queue[i].tx_bytes = 0;
1100 internals->tx_queue[i].err_pkts = 0;
1105 eth_rx_queue_release(void *q)
1107 struct szedata2_rx_queue *rxq = (struct szedata2_rx_queue *)q;
1108 if (rxq->sze != NULL) {
1109 szedata_close(rxq->sze);
1115 eth_tx_queue_release(void *q)
1117 struct szedata2_tx_queue *txq = (struct szedata2_tx_queue *)q;
1118 if (txq->sze != NULL) {
1119 szedata_close(txq->sze);
1125 eth_dev_close(struct rte_eth_dev *dev)
1128 uint16_t nb_rx = dev->data->nb_rx_queues;
1129 uint16_t nb_tx = dev->data->nb_tx_queues;
1133 for (i = 0; i < nb_rx; i++) {
1134 eth_rx_queue_release(dev->data->rx_queues[i]);
1135 dev->data->rx_queues[i] = NULL;
1137 dev->data->nb_rx_queues = 0;
1138 for (i = 0; i < nb_tx; i++) {
1139 eth_tx_queue_release(dev->data->tx_queues[i]);
1140 dev->data->tx_queues[i] = NULL;
1142 dev->data->nb_tx_queues = 0;
1146 * Function takes value from first IBUF status register.
1147 * Values in IBUF and OBUF should be same.
1150 * Pointer to device private structure.
1152 * Link speed constant.
1154 static inline enum szedata2_link_speed
1155 get_link_speed(const struct pmd_internals *internals)
1157 const volatile struct szedata2_ibuf *ibuf =
1158 ibuf_ptr_by_index(internals->pci_rsc, 0);
1159 uint32_t speed = (szedata2_read32(&ibuf->ibuf_st) & 0x70) >> 4;
1162 return SZEDATA2_LINK_SPEED_10G;
1164 return SZEDATA2_LINK_SPEED_40G;
1166 return SZEDATA2_LINK_SPEED_100G;
1168 return SZEDATA2_LINK_SPEED_DEFAULT;
1173 eth_link_update(struct rte_eth_dev *dev,
1174 int wait_to_complete __rte_unused)
1176 struct rte_eth_link link;
1177 struct rte_eth_link *link_ptr = &link;
1178 struct rte_eth_link *dev_link = &dev->data->dev_link;
1179 struct pmd_internals *internals = (struct pmd_internals *)
1180 dev->data->dev_private;
1181 const volatile struct szedata2_ibuf *ibuf;
1183 bool link_is_up = false;
1185 switch (get_link_speed(internals)) {
1186 case SZEDATA2_LINK_SPEED_10G:
1187 link.link_speed = ETH_SPEED_NUM_10G;
1189 case SZEDATA2_LINK_SPEED_40G:
1190 link.link_speed = ETH_SPEED_NUM_40G;
1192 case SZEDATA2_LINK_SPEED_100G:
1193 link.link_speed = ETH_SPEED_NUM_100G;
1196 link.link_speed = ETH_SPEED_NUM_10G;
1200 /* szedata2 uses only full duplex */
1201 link.link_duplex = ETH_LINK_FULL_DUPLEX;
1203 for (i = 0; i < szedata2_ibuf_count; i++) {
1204 ibuf = ibuf_ptr_by_index(internals->pci_rsc, i);
1206 * Link is considered up if at least one ibuf is enabled
1209 if (ibuf_is_enabled(ibuf) && ibuf_is_link_up(ibuf)) {
1215 link.link_status = (link_is_up) ? ETH_LINK_UP : ETH_LINK_DOWN;
1217 link.link_autoneg = ETH_LINK_SPEED_FIXED;
1219 rte_atomic64_cmpset((uint64_t *)dev_link, *(uint64_t *)dev_link,
1220 *(uint64_t *)link_ptr);
1226 eth_dev_set_link_up(struct rte_eth_dev *dev)
1228 struct pmd_internals *internals = (struct pmd_internals *)
1229 dev->data->dev_private;
1232 for (i = 0; i < szedata2_ibuf_count; i++)
1233 ibuf_enable(ibuf_ptr_by_index(internals->pci_rsc, i));
1234 for (i = 0; i < szedata2_obuf_count; i++)
1235 obuf_enable(obuf_ptr_by_index(internals->pci_rsc, i));
1240 eth_dev_set_link_down(struct rte_eth_dev *dev)
1242 struct pmd_internals *internals = (struct pmd_internals *)
1243 dev->data->dev_private;
1246 for (i = 0; i < szedata2_ibuf_count; i++)
1247 ibuf_disable(ibuf_ptr_by_index(internals->pci_rsc, i));
1248 for (i = 0; i < szedata2_obuf_count; i++)
1249 obuf_disable(obuf_ptr_by_index(internals->pci_rsc, i));
1254 eth_rx_queue_setup(struct rte_eth_dev *dev,
1255 uint16_t rx_queue_id,
1256 uint16_t nb_rx_desc __rte_unused,
1257 unsigned int socket_id __rte_unused,
1258 const struct rte_eth_rxconf *rx_conf __rte_unused,
1259 struct rte_mempool *mb_pool)
1261 struct pmd_internals *internals = dev->data->dev_private;
1262 struct szedata2_rx_queue *rxq = &internals->rx_queue[rx_queue_id];
1264 uint32_t rx = 1 << rx_queue_id;
1267 rxq->sze = szedata_open(internals->sze_dev);
1268 if (rxq->sze == NULL)
1270 ret = szedata_subscribe3(rxq->sze, &rx, &tx);
1271 if (ret != 0 || rx == 0) {
1272 szedata_close(rxq->sze);
1276 rxq->rx_channel = rx_queue_id;
1277 rxq->in_port = dev->data->port_id;
1278 rxq->mb_pool = mb_pool;
1283 dev->data->rx_queues[rx_queue_id] = rxq;
1288 eth_tx_queue_setup(struct rte_eth_dev *dev,
1289 uint16_t tx_queue_id,
1290 uint16_t nb_tx_desc __rte_unused,
1291 unsigned int socket_id __rte_unused,
1292 const struct rte_eth_txconf *tx_conf __rte_unused)
1294 struct pmd_internals *internals = dev->data->dev_private;
1295 struct szedata2_tx_queue *txq = &internals->tx_queue[tx_queue_id];
1298 uint32_t tx = 1 << tx_queue_id;
1300 txq->sze = szedata_open(internals->sze_dev);
1301 if (txq->sze == NULL)
1303 ret = szedata_subscribe3(txq->sze, &rx, &tx);
1304 if (ret != 0 || tx == 0) {
1305 szedata_close(txq->sze);
1309 txq->tx_channel = tx_queue_id;
1314 dev->data->tx_queues[tx_queue_id] = txq;
1319 eth_mac_addr_set(struct rte_eth_dev *dev __rte_unused,
1320 struct ether_addr *mac_addr __rte_unused)
1325 eth_promiscuous_enable(struct rte_eth_dev *dev)
1327 struct pmd_internals *internals = (struct pmd_internals *)
1328 dev->data->dev_private;
1331 for (i = 0; i < szedata2_ibuf_count; i++) {
1332 ibuf_mac_mode_write(ibuf_ptr_by_index(internals->pci_rsc, i),
1333 SZEDATA2_MAC_CHMODE_PROMISC);
1338 eth_promiscuous_disable(struct rte_eth_dev *dev)
1340 struct pmd_internals *internals = (struct pmd_internals *)
1341 dev->data->dev_private;
1344 for (i = 0; i < szedata2_ibuf_count; i++) {
1345 ibuf_mac_mode_write(ibuf_ptr_by_index(internals->pci_rsc, i),
1346 SZEDATA2_MAC_CHMODE_ONLY_VALID);
1351 eth_allmulticast_enable(struct rte_eth_dev *dev)
1353 struct pmd_internals *internals = (struct pmd_internals *)
1354 dev->data->dev_private;
1357 for (i = 0; i < szedata2_ibuf_count; i++) {
1358 ibuf_mac_mode_write(ibuf_ptr_by_index(internals->pci_rsc, i),
1359 SZEDATA2_MAC_CHMODE_ALL_MULTICAST);
1364 eth_allmulticast_disable(struct rte_eth_dev *dev)
1366 struct pmd_internals *internals = (struct pmd_internals *)
1367 dev->data->dev_private;
1370 for (i = 0; i < szedata2_ibuf_count; i++) {
1371 ibuf_mac_mode_write(ibuf_ptr_by_index(internals->pci_rsc, i),
1372 SZEDATA2_MAC_CHMODE_ONLY_VALID);
1376 static const struct eth_dev_ops ops = {
1377 .dev_start = eth_dev_start,
1378 .dev_stop = eth_dev_stop,
1379 .dev_set_link_up = eth_dev_set_link_up,
1380 .dev_set_link_down = eth_dev_set_link_down,
1381 .dev_close = eth_dev_close,
1382 .dev_configure = eth_dev_configure,
1383 .dev_infos_get = eth_dev_info,
1384 .promiscuous_enable = eth_promiscuous_enable,
1385 .promiscuous_disable = eth_promiscuous_disable,
1386 .allmulticast_enable = eth_allmulticast_enable,
1387 .allmulticast_disable = eth_allmulticast_disable,
1388 .rx_queue_start = eth_rx_queue_start,
1389 .rx_queue_stop = eth_rx_queue_stop,
1390 .tx_queue_start = eth_tx_queue_start,
1391 .tx_queue_stop = eth_tx_queue_stop,
1392 .rx_queue_setup = eth_rx_queue_setup,
1393 .tx_queue_setup = eth_tx_queue_setup,
1394 .rx_queue_release = eth_rx_queue_release,
1395 .tx_queue_release = eth_tx_queue_release,
1396 .link_update = eth_link_update,
1397 .stats_get = eth_stats_get,
1398 .stats_reset = eth_stats_reset,
1399 .mac_addr_set = eth_mac_addr_set,
1403 * This function goes through sysfs and looks for an index of szedata2
1404 * device file (/dev/szedataIIX, where X is the index).
1411 get_szedata2_index(const struct rte_pci_addr *pcislot_addr, uint32_t *index)
1414 struct dirent *entry;
1418 char pcislot_path[PATH_MAX];
1424 dir = opendir("/sys/class/combo");
1429 * Iterate through all combosixX directories.
1430 * When the value in /sys/class/combo/combosixX/device/pcislot
1431 * file is the location of the ethernet device dev, "X" is the
1432 * index of the device.
1434 while ((entry = readdir(dir)) != NULL) {
1435 ret = sscanf(entry->d_name, "combosix%u", &tmp_index);
1439 snprintf(pcislot_path, PATH_MAX,
1440 "/sys/class/combo/combosix%u/device/pcislot",
1443 fd = fopen(pcislot_path, "r");
1447 ret = fscanf(fd, "%4" PRIx16 ":%2" PRIx8 ":%2" PRIx8 ".%" PRIx8,
1448 &domain, &bus, &devid, &function);
1453 if (pcislot_addr->domain == domain &&
1454 pcislot_addr->bus == bus &&
1455 pcislot_addr->devid == devid &&
1456 pcislot_addr->function == function) {
1468 rte_szedata2_eth_dev_init(struct rte_eth_dev *dev)
1470 struct rte_eth_dev_data *data = dev->data;
1471 struct pmd_internals *internals = (struct pmd_internals *)
1473 struct szedata *szedata_temp;
1475 uint32_t szedata2_index;
1476 struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(dev);
1477 struct rte_pci_addr *pci_addr = &pci_dev->addr;
1478 struct rte_mem_resource *pci_rsc =
1479 &pci_dev->mem_resource[PCI_RESOURCE_NUMBER];
1480 char rsc_filename[PATH_MAX];
1481 void *pci_resource_ptr = NULL;
1484 RTE_LOG(INFO, PMD, "Initializing szedata2 device (" PCI_PRI_FMT ")\n",
1485 pci_addr->domain, pci_addr->bus, pci_addr->devid,
1486 pci_addr->function);
1488 /* Get index of szedata2 device file and create path to device file */
1489 ret = get_szedata2_index(pci_addr, &szedata2_index);
1491 RTE_LOG(ERR, PMD, "Failed to get szedata2 device index!\n");
1494 snprintf(internals->sze_dev, PATH_MAX, SZEDATA2_DEV_PATH_FMT,
1497 RTE_LOG(INFO, PMD, "SZEDATA2 path: %s\n", internals->sze_dev);
1500 * Get number of available DMA RX and TX channels, which is maximum
1501 * number of queues that can be created and store it in private device
1504 szedata_temp = szedata_open(internals->sze_dev);
1505 if (szedata_temp == NULL) {
1506 RTE_LOG(ERR, PMD, "szedata_open(): failed to open %s",
1507 internals->sze_dev);
1510 internals->max_rx_queues = szedata_ifaces_available(szedata_temp,
1512 internals->max_tx_queues = szedata_ifaces_available(szedata_temp,
1514 szedata_close(szedata_temp);
1516 RTE_LOG(INFO, PMD, "Available DMA channels RX: %u TX: %u\n",
1517 internals->max_rx_queues, internals->max_tx_queues);
1519 /* Set rx, tx burst functions */
1520 if (data->dev_conf.rxmode.enable_scatter == 1 ||
1521 data->scattered_rx == 1) {
1522 dev->rx_pkt_burst = eth_szedata2_rx_scattered;
1523 data->scattered_rx = 1;
1525 dev->rx_pkt_burst = eth_szedata2_rx;
1526 data->scattered_rx = 0;
1528 dev->tx_pkt_burst = eth_szedata2_tx;
1530 /* Set function callbacks for Ethernet API */
1531 dev->dev_ops = &ops;
1533 rte_eth_copy_pci_info(dev, pci_dev);
1535 /* mmap pci resource0 file to rte_mem_resource structure */
1536 if (pci_dev->mem_resource[PCI_RESOURCE_NUMBER].phys_addr ==
1538 RTE_LOG(ERR, PMD, "Missing resource%u file\n",
1539 PCI_RESOURCE_NUMBER);
1542 snprintf(rsc_filename, PATH_MAX,
1543 "%s/" PCI_PRI_FMT "/resource%u", pci_get_sysfs_path(),
1544 pci_addr->domain, pci_addr->bus,
1545 pci_addr->devid, pci_addr->function, PCI_RESOURCE_NUMBER);
1546 fd = open(rsc_filename, O_RDWR);
1548 RTE_LOG(ERR, PMD, "Could not open file %s\n", rsc_filename);
1552 pci_resource_ptr = mmap(0,
1553 pci_dev->mem_resource[PCI_RESOURCE_NUMBER].len,
1554 PROT_READ | PROT_WRITE, MAP_SHARED, fd, 0);
1556 if (pci_resource_ptr == NULL) {
1557 RTE_LOG(ERR, PMD, "Could not mmap file %s (fd = %d)\n",
1561 pci_dev->mem_resource[PCI_RESOURCE_NUMBER].addr = pci_resource_ptr;
1562 internals->pci_rsc = pci_rsc;
1564 RTE_LOG(DEBUG, PMD, "resource%u phys_addr = 0x%llx len = %llu "
1565 "virt addr = %llx\n", PCI_RESOURCE_NUMBER,
1566 (unsigned long long)pci_rsc->phys_addr,
1567 (unsigned long long)pci_rsc->len,
1568 (unsigned long long)pci_rsc->addr);
1570 /* Get link state */
1571 eth_link_update(dev, 0);
1573 /* Allocate space for one mac address */
1574 data->mac_addrs = rte_zmalloc(data->name, sizeof(struct ether_addr),
1575 RTE_CACHE_LINE_SIZE);
1576 if (data->mac_addrs == NULL) {
1577 RTE_LOG(ERR, PMD, "Could not alloc space for MAC address!\n");
1578 munmap(pci_dev->mem_resource[PCI_RESOURCE_NUMBER].addr,
1579 pci_dev->mem_resource[PCI_RESOURCE_NUMBER].len);
1583 ether_addr_copy(ð_addr, data->mac_addrs);
1585 /* At initial state COMBO card is in promiscuous mode so disable it */
1586 eth_promiscuous_disable(dev);
1588 RTE_LOG(INFO, PMD, "szedata2 device ("
1589 PCI_PRI_FMT ") successfully initialized\n",
1590 pci_addr->domain, pci_addr->bus, pci_addr->devid,
1591 pci_addr->function);
1597 rte_szedata2_eth_dev_uninit(struct rte_eth_dev *dev)
1599 struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(dev);
1600 struct rte_pci_addr *pci_addr = &pci_dev->addr;
1602 rte_free(dev->data->mac_addrs);
1603 dev->data->mac_addrs = NULL;
1604 munmap(pci_dev->mem_resource[PCI_RESOURCE_NUMBER].addr,
1605 pci_dev->mem_resource[PCI_RESOURCE_NUMBER].len);
1607 RTE_LOG(INFO, PMD, "szedata2 device ("
1608 PCI_PRI_FMT ") successfully uninitialized\n",
1609 pci_addr->domain, pci_addr->bus, pci_addr->devid,
1610 pci_addr->function);
1615 static const struct rte_pci_id rte_szedata2_pci_id_table[] = {
1617 RTE_PCI_DEVICE(PCI_VENDOR_ID_NETCOPE,
1618 PCI_DEVICE_ID_NETCOPE_COMBO80G)
1621 RTE_PCI_DEVICE(PCI_VENDOR_ID_NETCOPE,
1622 PCI_DEVICE_ID_NETCOPE_COMBO100G)
1625 RTE_PCI_DEVICE(PCI_VENDOR_ID_NETCOPE,
1626 PCI_DEVICE_ID_NETCOPE_COMBO100G2)
1633 static int szedata2_eth_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
1634 struct rte_pci_device *pci_dev)
1636 return rte_eth_dev_pci_generic_probe(pci_dev,
1637 sizeof(struct pmd_internals), rte_szedata2_eth_dev_init);
1640 static int szedata2_eth_pci_remove(struct rte_pci_device *pci_dev)
1642 return rte_eth_dev_pci_generic_remove(pci_dev,
1643 rte_szedata2_eth_dev_uninit);
1646 static struct rte_pci_driver szedata2_eth_driver = {
1647 .id_table = rte_szedata2_pci_id_table,
1648 .probe = szedata2_eth_pci_probe,
1649 .remove = szedata2_eth_pci_remove,
1652 RTE_PMD_REGISTER_PCI(RTE_SZEDATA2_DRIVER_NAME, szedata2_eth_driver);
1653 RTE_PMD_REGISTER_PCI_TABLE(RTE_SZEDATA2_DRIVER_NAME, rte_szedata2_pci_id_table);
1654 RTE_PMD_REGISTER_KMOD_DEP(RTE_SZEDATA2_DRIVER_NAME,
1655 "* combo6core & combov3 & szedata2 & szedata2_cv3");