4 * Copyright (c) 2015 - 2016 CESNET
7 * Redistribution and use in source and binary forms, with or without
8 * modification, are permitted provided that the following conditions
11 * * Redistributions of source code must retain the above copyright
12 * notice, this list of conditions and the following disclaimer.
13 * * Redistributions in binary form must reproduce the above copyright
14 * notice, this list of conditions and the following disclaimer in
15 * the documentation and/or other materials provided with the
17 * * Neither the name of CESNET nor the names of its
18 * contributors may be used to endorse or promote products derived
19 * from this software without specific prior written permission.
21 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
22 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
23 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
24 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
25 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
26 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
27 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
28 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
29 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
30 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
31 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
38 #include <sys/types.h>
47 #include <rte_ethdev_driver.h>
48 #include <rte_ethdev_pci.h>
49 #include <rte_malloc.h>
50 #include <rte_memcpy.h>
51 #include <rte_kvargs.h>
54 #include "rte_eth_szedata2.h"
55 #include "szedata2_iobuf.h"
57 #define RTE_ETH_SZEDATA2_MAX_RX_QUEUES 32
58 #define RTE_ETH_SZEDATA2_MAX_TX_QUEUES 32
59 #define RTE_ETH_SZEDATA2_TX_LOCK_SIZE (32 * 1024 * 1024)
62 * size of szedata2_packet header with alignment
64 #define RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED 8
66 #define RTE_SZEDATA2_DRIVER_NAME net_szedata2
68 #define SZEDATA2_DEV_PATH_FMT "/dev/szedataII%u"
70 struct szedata2_rx_queue {
74 struct rte_mempool *mb_pool;
75 volatile uint64_t rx_pkts;
76 volatile uint64_t rx_bytes;
77 volatile uint64_t err_pkts;
80 struct szedata2_tx_queue {
83 volatile uint64_t tx_pkts;
84 volatile uint64_t tx_bytes;
85 volatile uint64_t err_pkts;
88 struct pmd_internals {
89 uint16_t max_rx_queues;
90 uint16_t max_tx_queues;
91 char sze_dev[PATH_MAX];
92 struct rte_mem_resource *pci_rsc;
95 static struct ether_addr eth_addr = {
96 .addr_bytes = { 0x00, 0x11, 0x17, 0x00, 0x00, 0x00 }
100 eth_szedata2_rx(void *queue,
101 struct rte_mbuf **bufs,
105 struct rte_mbuf *mbuf;
106 struct szedata2_rx_queue *sze_q = queue;
107 struct rte_pktmbuf_pool_private *mbp_priv;
112 uint16_t packet_size;
113 uint64_t num_bytes = 0;
114 struct szedata *sze = sze_q->sze;
115 uint8_t *header_ptr = NULL; /* header of packet */
116 uint8_t *packet_ptr1 = NULL;
117 uint8_t *packet_ptr2 = NULL;
118 uint16_t packet_len1 = 0;
119 uint16_t packet_len2 = 0;
120 uint16_t hw_data_align;
122 if (unlikely(sze_q->sze == NULL || nb_pkts == 0))
126 * Reads the given number of packets from szedata2 channel given
127 * by queue and copies the packet data into a newly allocated mbuf
130 for (i = 0; i < nb_pkts; i++) {
131 mbuf = rte_pktmbuf_alloc(sze_q->mb_pool);
133 if (unlikely(mbuf == NULL))
136 /* get the next sze packet */
137 if (sze->ct_rx_lck != NULL && !sze->ct_rx_rem_bytes &&
138 sze->ct_rx_lck->next == NULL) {
139 /* unlock old data */
140 szedata_rx_unlock_data(sze_q->sze, sze->ct_rx_lck_orig);
141 sze->ct_rx_lck_orig = NULL;
142 sze->ct_rx_lck = NULL;
145 if (!sze->ct_rx_rem_bytes && sze->ct_rx_lck_orig == NULL) {
146 /* nothing to read, lock new data */
147 sze->ct_rx_lck = szedata_rx_lock_data(sze_q->sze, ~0U);
148 sze->ct_rx_lck_orig = sze->ct_rx_lck;
150 if (sze->ct_rx_lck == NULL) {
151 /* nothing to lock */
152 rte_pktmbuf_free(mbuf);
156 sze->ct_rx_cur_ptr = sze->ct_rx_lck->start;
157 sze->ct_rx_rem_bytes = sze->ct_rx_lck->len;
159 if (!sze->ct_rx_rem_bytes) {
160 rte_pktmbuf_free(mbuf);
165 if (sze->ct_rx_rem_bytes < RTE_SZE2_PACKET_HEADER_SIZE) {
168 * copy parts of header to merge buffer
170 if (sze->ct_rx_lck->next == NULL) {
171 rte_pktmbuf_free(mbuf);
175 /* copy first part of header */
176 rte_memcpy(sze->ct_rx_buffer, sze->ct_rx_cur_ptr,
177 sze->ct_rx_rem_bytes);
179 /* copy second part of header */
180 sze->ct_rx_lck = sze->ct_rx_lck->next;
181 sze->ct_rx_cur_ptr = sze->ct_rx_lck->start;
182 rte_memcpy(sze->ct_rx_buffer + sze->ct_rx_rem_bytes,
184 RTE_SZE2_PACKET_HEADER_SIZE -
185 sze->ct_rx_rem_bytes);
187 sze->ct_rx_cur_ptr += RTE_SZE2_PACKET_HEADER_SIZE -
188 sze->ct_rx_rem_bytes;
189 sze->ct_rx_rem_bytes = sze->ct_rx_lck->len -
190 RTE_SZE2_PACKET_HEADER_SIZE +
191 sze->ct_rx_rem_bytes;
193 header_ptr = (uint8_t *)sze->ct_rx_buffer;
196 header_ptr = (uint8_t *)sze->ct_rx_cur_ptr;
197 sze->ct_rx_cur_ptr += RTE_SZE2_PACKET_HEADER_SIZE;
198 sze->ct_rx_rem_bytes -= RTE_SZE2_PACKET_HEADER_SIZE;
201 sg_size = le16toh(*((uint16_t *)header_ptr));
202 hw_size = le16toh(*(((uint16_t *)header_ptr) + 1));
203 packet_size = sg_size -
204 RTE_SZE2_ALIGN8(RTE_SZE2_PACKET_HEADER_SIZE + hw_size);
207 /* checks if packet all right */
209 errx(5, "Zero segsize");
211 /* check sg_size and hwsize */
212 if (hw_size > sg_size - RTE_SZE2_PACKET_HEADER_SIZE) {
213 errx(10, "Hwsize bigger than expected. Segsize: %d, "
214 "hwsize: %d", sg_size, hw_size);
218 RTE_SZE2_ALIGN8(RTE_SZE2_PACKET_HEADER_SIZE + hw_size) -
219 RTE_SZE2_PACKET_HEADER_SIZE;
221 if (sze->ct_rx_rem_bytes >=
223 RTE_SZE2_PACKET_HEADER_SIZE)) {
225 /* one packet ready - go to another */
226 packet_ptr1 = sze->ct_rx_cur_ptr + hw_data_align;
227 packet_len1 = packet_size;
231 sze->ct_rx_cur_ptr += RTE_SZE2_ALIGN8(sg_size) -
232 RTE_SZE2_PACKET_HEADER_SIZE;
233 sze->ct_rx_rem_bytes -= RTE_SZE2_ALIGN8(sg_size) -
234 RTE_SZE2_PACKET_HEADER_SIZE;
237 if (sze->ct_rx_lck->next == NULL) {
238 errx(6, "Need \"next\" lock, "
239 "but it is missing: %u",
240 sze->ct_rx_rem_bytes);
244 if (sze->ct_rx_rem_bytes <= hw_data_align) {
245 uint16_t rem_size = hw_data_align -
246 sze->ct_rx_rem_bytes;
248 /* MOVE to next lock */
249 sze->ct_rx_lck = sze->ct_rx_lck->next;
251 (void *)(((uint8_t *)
252 (sze->ct_rx_lck->start)) + rem_size);
254 packet_ptr1 = sze->ct_rx_cur_ptr;
255 packet_len1 = packet_size;
259 sze->ct_rx_cur_ptr +=
260 RTE_SZE2_ALIGN8(packet_size);
261 sze->ct_rx_rem_bytes = sze->ct_rx_lck->len -
262 rem_size - RTE_SZE2_ALIGN8(packet_size);
264 /* get pointer and length from first part */
265 packet_ptr1 = sze->ct_rx_cur_ptr +
267 packet_len1 = sze->ct_rx_rem_bytes -
270 /* MOVE to next lock */
271 sze->ct_rx_lck = sze->ct_rx_lck->next;
272 sze->ct_rx_cur_ptr = sze->ct_rx_lck->start;
274 /* get pointer and length from second part */
275 packet_ptr2 = sze->ct_rx_cur_ptr;
276 packet_len2 = packet_size - packet_len1;
278 sze->ct_rx_cur_ptr +=
279 RTE_SZE2_ALIGN8(packet_size) -
281 sze->ct_rx_rem_bytes = sze->ct_rx_lck->len -
282 (RTE_SZE2_ALIGN8(packet_size) -
287 if (unlikely(packet_ptr1 == NULL)) {
288 rte_pktmbuf_free(mbuf);
292 /* get the space available for data in the mbuf */
293 mbp_priv = rte_mempool_get_priv(sze_q->mb_pool);
294 buf_size = (uint16_t)(mbp_priv->mbuf_data_room_size -
295 RTE_PKTMBUF_HEADROOM);
297 if (packet_size <= buf_size) {
298 /* sze packet will fit in one mbuf, go ahead and copy */
299 rte_memcpy(rte_pktmbuf_mtod(mbuf, void *),
300 packet_ptr1, packet_len1);
301 if (packet_ptr2 != NULL) {
302 rte_memcpy((void *)(rte_pktmbuf_mtod(mbuf,
303 uint8_t *) + packet_len1),
304 packet_ptr2, packet_len2);
306 mbuf->data_len = (uint16_t)packet_size;
308 mbuf->pkt_len = packet_size;
309 mbuf->port = sze_q->in_port;
312 num_bytes += packet_size;
315 * sze packet will not fit in one mbuf,
316 * scattered mode is not enabled, drop packet
319 "SZE segment %d bytes will not fit in one mbuf "
320 "(%d bytes), scattered mode is not enabled, "
322 packet_size, buf_size);
323 rte_pktmbuf_free(mbuf);
327 sze_q->rx_pkts += num_rx;
328 sze_q->rx_bytes += num_bytes;
333 eth_szedata2_rx_scattered(void *queue,
334 struct rte_mbuf **bufs,
338 struct rte_mbuf *mbuf;
339 struct szedata2_rx_queue *sze_q = queue;
340 struct rte_pktmbuf_pool_private *mbp_priv;
345 uint16_t packet_size;
346 uint64_t num_bytes = 0;
347 struct szedata *sze = sze_q->sze;
348 uint8_t *header_ptr = NULL; /* header of packet */
349 uint8_t *packet_ptr1 = NULL;
350 uint8_t *packet_ptr2 = NULL;
351 uint16_t packet_len1 = 0;
352 uint16_t packet_len2 = 0;
353 uint16_t hw_data_align;
355 if (unlikely(sze_q->sze == NULL || nb_pkts == 0))
359 * Reads the given number of packets from szedata2 channel given
360 * by queue and copies the packet data into a newly allocated mbuf
363 for (i = 0; i < nb_pkts; i++) {
364 const struct szedata_lock *ct_rx_lck_backup;
365 unsigned int ct_rx_rem_bytes_backup;
366 unsigned char *ct_rx_cur_ptr_backup;
368 /* get the next sze packet */
369 if (sze->ct_rx_lck != NULL && !sze->ct_rx_rem_bytes &&
370 sze->ct_rx_lck->next == NULL) {
371 /* unlock old data */
372 szedata_rx_unlock_data(sze_q->sze, sze->ct_rx_lck_orig);
373 sze->ct_rx_lck_orig = NULL;
374 sze->ct_rx_lck = NULL;
378 * Store items from sze structure which can be changed
379 * before mbuf allocating. Use these items in case of mbuf
380 * allocating failure.
382 ct_rx_lck_backup = sze->ct_rx_lck;
383 ct_rx_rem_bytes_backup = sze->ct_rx_rem_bytes;
384 ct_rx_cur_ptr_backup = sze->ct_rx_cur_ptr;
386 if (!sze->ct_rx_rem_bytes && sze->ct_rx_lck_orig == NULL) {
387 /* nothing to read, lock new data */
388 sze->ct_rx_lck = szedata_rx_lock_data(sze_q->sze, ~0U);
389 sze->ct_rx_lck_orig = sze->ct_rx_lck;
392 * Backup items from sze structure must be updated
393 * after locking to contain pointers to new locks.
395 ct_rx_lck_backup = sze->ct_rx_lck;
396 ct_rx_rem_bytes_backup = sze->ct_rx_rem_bytes;
397 ct_rx_cur_ptr_backup = sze->ct_rx_cur_ptr;
399 if (sze->ct_rx_lck == NULL)
400 /* nothing to lock */
403 sze->ct_rx_cur_ptr = sze->ct_rx_lck->start;
404 sze->ct_rx_rem_bytes = sze->ct_rx_lck->len;
406 if (!sze->ct_rx_rem_bytes)
410 if (sze->ct_rx_rem_bytes < RTE_SZE2_PACKET_HEADER_SIZE) {
412 * cut in header - copy parts of header to merge buffer
414 if (sze->ct_rx_lck->next == NULL)
417 /* copy first part of header */
418 rte_memcpy(sze->ct_rx_buffer, sze->ct_rx_cur_ptr,
419 sze->ct_rx_rem_bytes);
421 /* copy second part of header */
422 sze->ct_rx_lck = sze->ct_rx_lck->next;
423 sze->ct_rx_cur_ptr = sze->ct_rx_lck->start;
424 rte_memcpy(sze->ct_rx_buffer + sze->ct_rx_rem_bytes,
426 RTE_SZE2_PACKET_HEADER_SIZE -
427 sze->ct_rx_rem_bytes);
429 sze->ct_rx_cur_ptr += RTE_SZE2_PACKET_HEADER_SIZE -
430 sze->ct_rx_rem_bytes;
431 sze->ct_rx_rem_bytes = sze->ct_rx_lck->len -
432 RTE_SZE2_PACKET_HEADER_SIZE +
433 sze->ct_rx_rem_bytes;
435 header_ptr = (uint8_t *)sze->ct_rx_buffer;
438 header_ptr = (uint8_t *)sze->ct_rx_cur_ptr;
439 sze->ct_rx_cur_ptr += RTE_SZE2_PACKET_HEADER_SIZE;
440 sze->ct_rx_rem_bytes -= RTE_SZE2_PACKET_HEADER_SIZE;
443 sg_size = le16toh(*((uint16_t *)header_ptr));
444 hw_size = le16toh(*(((uint16_t *)header_ptr) + 1));
445 packet_size = sg_size -
446 RTE_SZE2_ALIGN8(RTE_SZE2_PACKET_HEADER_SIZE + hw_size);
449 /* checks if packet all right */
451 errx(5, "Zero segsize");
453 /* check sg_size and hwsize */
454 if (hw_size > sg_size - RTE_SZE2_PACKET_HEADER_SIZE) {
455 errx(10, "Hwsize bigger than expected. Segsize: %d, "
456 "hwsize: %d", sg_size, hw_size);
460 RTE_SZE2_ALIGN8((RTE_SZE2_PACKET_HEADER_SIZE +
461 hw_size)) - RTE_SZE2_PACKET_HEADER_SIZE;
463 if (sze->ct_rx_rem_bytes >=
465 RTE_SZE2_PACKET_HEADER_SIZE)) {
467 /* one packet ready - go to another */
468 packet_ptr1 = sze->ct_rx_cur_ptr + hw_data_align;
469 packet_len1 = packet_size;
473 sze->ct_rx_cur_ptr += RTE_SZE2_ALIGN8(sg_size) -
474 RTE_SZE2_PACKET_HEADER_SIZE;
475 sze->ct_rx_rem_bytes -= RTE_SZE2_ALIGN8(sg_size) -
476 RTE_SZE2_PACKET_HEADER_SIZE;
479 if (sze->ct_rx_lck->next == NULL) {
480 errx(6, "Need \"next\" lock, but it is "
481 "missing: %u", sze->ct_rx_rem_bytes);
485 if (sze->ct_rx_rem_bytes <= hw_data_align) {
486 uint16_t rem_size = hw_data_align -
487 sze->ct_rx_rem_bytes;
489 /* MOVE to next lock */
490 sze->ct_rx_lck = sze->ct_rx_lck->next;
492 (void *)(((uint8_t *)
493 (sze->ct_rx_lck->start)) + rem_size);
495 packet_ptr1 = sze->ct_rx_cur_ptr;
496 packet_len1 = packet_size;
500 sze->ct_rx_cur_ptr +=
501 RTE_SZE2_ALIGN8(packet_size);
502 sze->ct_rx_rem_bytes = sze->ct_rx_lck->len -
503 rem_size - RTE_SZE2_ALIGN8(packet_size);
505 /* get pointer and length from first part */
506 packet_ptr1 = sze->ct_rx_cur_ptr +
508 packet_len1 = sze->ct_rx_rem_bytes -
511 /* MOVE to next lock */
512 sze->ct_rx_lck = sze->ct_rx_lck->next;
513 sze->ct_rx_cur_ptr = sze->ct_rx_lck->start;
515 /* get pointer and length from second part */
516 packet_ptr2 = sze->ct_rx_cur_ptr;
517 packet_len2 = packet_size - packet_len1;
519 sze->ct_rx_cur_ptr +=
520 RTE_SZE2_ALIGN8(packet_size) -
522 sze->ct_rx_rem_bytes = sze->ct_rx_lck->len -
523 (RTE_SZE2_ALIGN8(packet_size) -
528 if (unlikely(packet_ptr1 == NULL))
531 mbuf = rte_pktmbuf_alloc(sze_q->mb_pool);
533 if (unlikely(mbuf == NULL)) {
535 * Restore items from sze structure to state after
536 * unlocking (eventually locking).
538 sze->ct_rx_lck = ct_rx_lck_backup;
539 sze->ct_rx_rem_bytes = ct_rx_rem_bytes_backup;
540 sze->ct_rx_cur_ptr = ct_rx_cur_ptr_backup;
544 /* get the space available for data in the mbuf */
545 mbp_priv = rte_mempool_get_priv(sze_q->mb_pool);
546 buf_size = (uint16_t)(mbp_priv->mbuf_data_room_size -
547 RTE_PKTMBUF_HEADROOM);
549 if (packet_size <= buf_size) {
550 /* sze packet will fit in one mbuf, go ahead and copy */
551 rte_memcpy(rte_pktmbuf_mtod(mbuf, void *),
552 packet_ptr1, packet_len1);
553 if (packet_ptr2 != NULL) {
555 (rte_pktmbuf_mtod(mbuf, uint8_t *) +
556 packet_len1), packet_ptr2, packet_len2);
558 mbuf->data_len = (uint16_t)packet_size;
561 * sze packet will not fit in one mbuf,
562 * scatter packet into more mbufs
564 struct rte_mbuf *m = mbuf;
565 uint16_t len = rte_pktmbuf_tailroom(mbuf);
567 /* copy first part of packet */
568 /* fill first mbuf */
569 rte_memcpy(rte_pktmbuf_append(mbuf, len), packet_ptr1,
572 packet_ptr1 = ((uint8_t *)packet_ptr1) + len;
574 while (packet_len1 > 0) {
576 m->next = rte_pktmbuf_alloc(sze_q->mb_pool);
578 if (unlikely(m->next == NULL)) {
579 rte_pktmbuf_free(mbuf);
581 * Restore items from sze structure
582 * to state after unlocking (eventually
585 sze->ct_rx_lck = ct_rx_lck_backup;
586 sze->ct_rx_rem_bytes =
587 ct_rx_rem_bytes_backup;
589 ct_rx_cur_ptr_backup;
595 len = RTE_MIN(rte_pktmbuf_tailroom(m),
597 rte_memcpy(rte_pktmbuf_append(mbuf, len),
602 packet_ptr1 = ((uint8_t *)packet_ptr1) + len;
605 if (packet_ptr2 != NULL) {
606 /* copy second part of packet, if exists */
607 /* fill the rest of currently last mbuf */
608 len = rte_pktmbuf_tailroom(m);
609 rte_memcpy(rte_pktmbuf_append(mbuf, len),
612 packet_ptr2 = ((uint8_t *)packet_ptr2) + len;
614 while (packet_len2 > 0) {
616 m->next = rte_pktmbuf_alloc(
619 if (unlikely(m->next == NULL)) {
620 rte_pktmbuf_free(mbuf);
622 * Restore items from sze
623 * structure to state after
624 * unlocking (eventually
629 sze->ct_rx_rem_bytes =
630 ct_rx_rem_bytes_backup;
632 ct_rx_cur_ptr_backup;
638 len = RTE_MIN(rte_pktmbuf_tailroom(m),
641 rte_pktmbuf_append(mbuf, len),
646 packet_ptr2 = ((uint8_t *)packet_ptr2) +
651 mbuf->pkt_len = packet_size;
652 mbuf->port = sze_q->in_port;
655 num_bytes += packet_size;
659 sze_q->rx_pkts += num_rx;
660 sze_q->rx_bytes += num_bytes;
665 eth_szedata2_tx(void *queue,
666 struct rte_mbuf **bufs,
669 struct rte_mbuf *mbuf;
670 struct szedata2_tx_queue *sze_q = queue;
672 uint64_t num_bytes = 0;
674 const struct szedata_lock *lck;
680 uint32_t unlock_size;
683 uint16_t pkt_left = nb_pkts;
685 if (sze_q->sze == NULL || nb_pkts == 0)
688 while (pkt_left > 0) {
690 lck = szedata_tx_lock_data(sze_q->sze,
691 RTE_ETH_SZEDATA2_TX_LOCK_SIZE,
697 lock_size = lck->len;
698 lock_size2 = lck->next ? lck->next->len : 0;
701 mbuf = bufs[nb_pkts - pkt_left];
703 pkt_len = mbuf->pkt_len;
704 mbuf_segs = mbuf->nb_segs;
706 hwpkt_len = RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED +
707 RTE_SZE2_ALIGN8(pkt_len);
709 if (lock_size + lock_size2 < hwpkt_len) {
710 szedata_tx_unlock_data(sze_q->sze, lck, unlock_size);
714 num_bytes += pkt_len;
716 if (lock_size > hwpkt_len) {
721 /* write packet length at first 2 bytes in 8B header */
722 *((uint16_t *)dst) = htole16(
723 RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED +
725 *(((uint16_t *)dst) + 1) = htole16(0);
727 /* copy packet from mbuf */
728 tmp_dst = ((uint8_t *)(dst)) +
729 RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED;
730 if (mbuf_segs == 1) {
732 * non-scattered packet,
733 * transmit from one mbuf
736 rte_pktmbuf_mtod(mbuf, const void *),
739 /* scattered packet, transmit from more mbufs */
740 struct rte_mbuf *m = mbuf;
746 tmp_dst = ((uint8_t *)(tmp_dst)) +
753 dst = ((uint8_t *)dst) + hwpkt_len;
754 unlock_size += hwpkt_len;
755 lock_size -= hwpkt_len;
757 rte_pktmbuf_free(mbuf);
761 szedata_tx_unlock_data(sze_q->sze, lck,
766 } else if (lock_size + lock_size2 >= hwpkt_len) {
770 /* write packet length at first 2 bytes in 8B header */
772 htole16(RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED +
774 *(((uint16_t *)dst) + 1) = htole16(0);
777 * If the raw packet (pkt_len) is smaller than lock_size
778 * get the correct length for memcpy
781 pkt_len < lock_size -
782 RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED ?
784 lock_size - RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED;
786 rem_len = hwpkt_len - lock_size;
788 tmp_dst = ((uint8_t *)(dst)) +
789 RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED;
790 if (mbuf_segs == 1) {
792 * non-scattered packet,
793 * transmit from one mbuf
795 /* copy part of packet to first area */
797 rte_pktmbuf_mtod(mbuf, const void *),
801 dst = lck->next->start;
803 /* copy part of packet to second area */
805 (const void *)(rte_pktmbuf_mtod(mbuf,
807 write_len), pkt_len - write_len);
809 /* scattered packet, transmit from more mbufs */
810 struct rte_mbuf *m = mbuf;
811 uint16_t written = 0;
812 uint16_t to_write = 0;
813 bool new_mbuf = true;
814 uint16_t write_off = 0;
816 /* copy part of packet to first area */
817 while (m && written < write_len) {
818 to_write = RTE_MIN(m->data_len,
819 write_len - written);
825 tmp_dst = ((uint8_t *)(tmp_dst)) +
827 if (m->data_len <= write_len -
838 dst = lck->next->start;
842 write_off = new_mbuf ? 0 : to_write;
844 /* copy part of packet to second area */
845 while (m && written < pkt_len - write_len) {
846 rte_memcpy(tmp_dst, (const void *)
848 uint8_t *) + write_off),
849 m->data_len - write_off);
851 tmp_dst = ((uint8_t *)(tmp_dst)) +
852 (m->data_len - write_off);
853 written += m->data_len - write_off;
859 dst = ((uint8_t *)dst) + rem_len;
860 unlock_size += hwpkt_len;
861 lock_size = lock_size2 - rem_len;
864 rte_pktmbuf_free(mbuf);
868 szedata_tx_unlock_data(sze_q->sze, lck, unlock_size);
872 sze_q->tx_pkts += num_tx;
873 sze_q->err_pkts += nb_pkts - num_tx;
874 sze_q->tx_bytes += num_bytes;
879 eth_rx_queue_start(struct rte_eth_dev *dev, uint16_t rxq_id)
881 struct szedata2_rx_queue *rxq = dev->data->rx_queues[rxq_id];
883 struct pmd_internals *internals = (struct pmd_internals *)
884 dev->data->dev_private;
886 if (rxq->sze == NULL) {
887 uint32_t rx = 1 << rxq->rx_channel;
889 rxq->sze = szedata_open(internals->sze_dev);
890 if (rxq->sze == NULL)
892 ret = szedata_subscribe3(rxq->sze, &rx, &tx);
893 if (ret != 0 || rx == 0)
897 ret = szedata_start(rxq->sze);
900 dev->data->rx_queue_state[rxq_id] = RTE_ETH_QUEUE_STATE_STARTED;
904 szedata_close(rxq->sze);
910 eth_rx_queue_stop(struct rte_eth_dev *dev, uint16_t rxq_id)
912 struct szedata2_rx_queue *rxq = dev->data->rx_queues[rxq_id];
914 if (rxq->sze != NULL) {
915 szedata_close(rxq->sze);
919 dev->data->rx_queue_state[rxq_id] = RTE_ETH_QUEUE_STATE_STOPPED;
924 eth_tx_queue_start(struct rte_eth_dev *dev, uint16_t txq_id)
926 struct szedata2_tx_queue *txq = dev->data->tx_queues[txq_id];
928 struct pmd_internals *internals = (struct pmd_internals *)
929 dev->data->dev_private;
931 if (txq->sze == NULL) {
933 uint32_t tx = 1 << txq->tx_channel;
934 txq->sze = szedata_open(internals->sze_dev);
935 if (txq->sze == NULL)
937 ret = szedata_subscribe3(txq->sze, &rx, &tx);
938 if (ret != 0 || tx == 0)
942 ret = szedata_start(txq->sze);
945 dev->data->tx_queue_state[txq_id] = RTE_ETH_QUEUE_STATE_STARTED;
949 szedata_close(txq->sze);
955 eth_tx_queue_stop(struct rte_eth_dev *dev, uint16_t txq_id)
957 struct szedata2_tx_queue *txq = dev->data->tx_queues[txq_id];
959 if (txq->sze != NULL) {
960 szedata_close(txq->sze);
964 dev->data->tx_queue_state[txq_id] = RTE_ETH_QUEUE_STATE_STOPPED;
969 eth_dev_start(struct rte_eth_dev *dev)
973 uint16_t nb_rx = dev->data->nb_rx_queues;
974 uint16_t nb_tx = dev->data->nb_tx_queues;
976 for (i = 0; i < nb_rx; i++) {
977 ret = eth_rx_queue_start(dev, i);
982 for (i = 0; i < nb_tx; i++) {
983 ret = eth_tx_queue_start(dev, i);
991 for (i = 0; i < nb_tx; i++)
992 eth_tx_queue_stop(dev, i);
994 for (i = 0; i < nb_rx; i++)
995 eth_rx_queue_stop(dev, i);
1000 eth_dev_stop(struct rte_eth_dev *dev)
1003 uint16_t nb_rx = dev->data->nb_rx_queues;
1004 uint16_t nb_tx = dev->data->nb_tx_queues;
1006 for (i = 0; i < nb_tx; i++)
1007 eth_tx_queue_stop(dev, i);
1009 for (i = 0; i < nb_rx; i++)
1010 eth_rx_queue_stop(dev, i);
1014 eth_dev_configure(struct rte_eth_dev *dev)
1016 struct rte_eth_dev_data *data = dev->data;
1017 if (data->dev_conf.rxmode.offloads & DEV_RX_OFFLOAD_SCATTER) {
1018 dev->rx_pkt_burst = eth_szedata2_rx_scattered;
1019 data->scattered_rx = 1;
1021 dev->rx_pkt_burst = eth_szedata2_rx;
1022 data->scattered_rx = 0;
1028 eth_dev_info(struct rte_eth_dev *dev,
1029 struct rte_eth_dev_info *dev_info)
1031 struct pmd_internals *internals = dev->data->dev_private;
1033 dev_info->pci_dev = RTE_ETH_DEV_TO_PCI(dev);
1034 dev_info->if_index = 0;
1035 dev_info->max_mac_addrs = 1;
1036 dev_info->max_rx_pktlen = (uint32_t)-1;
1037 dev_info->max_rx_queues = internals->max_rx_queues;
1038 dev_info->max_tx_queues = internals->max_tx_queues;
1039 dev_info->min_rx_bufsize = 0;
1040 dev_info->rx_offload_capa = DEV_RX_OFFLOAD_SCATTER;
1041 dev_info->tx_offload_capa = 0;
1042 dev_info->rx_queue_offload_capa = 0;
1043 dev_info->tx_queue_offload_capa = 0;
1044 dev_info->speed_capa = ETH_LINK_SPEED_100G;
1048 eth_stats_get(struct rte_eth_dev *dev,
1049 struct rte_eth_stats *stats)
1052 uint16_t nb_rx = dev->data->nb_rx_queues;
1053 uint16_t nb_tx = dev->data->nb_tx_queues;
1054 uint64_t rx_total = 0;
1055 uint64_t tx_total = 0;
1056 uint64_t tx_err_total = 0;
1057 uint64_t rx_total_bytes = 0;
1058 uint64_t tx_total_bytes = 0;
1060 for (i = 0; i < nb_rx; i++) {
1061 struct szedata2_rx_queue *rxq = dev->data->rx_queues[i];
1063 if (i < RTE_ETHDEV_QUEUE_STAT_CNTRS) {
1064 stats->q_ipackets[i] = rxq->rx_pkts;
1065 stats->q_ibytes[i] = rxq->rx_bytes;
1067 rx_total += rxq->rx_pkts;
1068 rx_total_bytes += rxq->rx_bytes;
1071 for (i = 0; i < nb_tx; i++) {
1072 struct szedata2_tx_queue *txq = dev->data->tx_queues[i];
1074 if (i < RTE_ETHDEV_QUEUE_STAT_CNTRS) {
1075 stats->q_opackets[i] = txq->tx_pkts;
1076 stats->q_obytes[i] = txq->tx_bytes;
1077 stats->q_errors[i] = txq->err_pkts;
1079 tx_total += txq->tx_pkts;
1080 tx_total_bytes += txq->tx_bytes;
1081 tx_err_total += txq->err_pkts;
1084 stats->ipackets = rx_total;
1085 stats->opackets = tx_total;
1086 stats->ibytes = rx_total_bytes;
1087 stats->obytes = tx_total_bytes;
1088 stats->oerrors = tx_err_total;
1094 eth_stats_reset(struct rte_eth_dev *dev)
1097 uint16_t nb_rx = dev->data->nb_rx_queues;
1098 uint16_t nb_tx = dev->data->nb_tx_queues;
1100 for (i = 0; i < nb_rx; i++) {
1101 struct szedata2_rx_queue *rxq = dev->data->rx_queues[i];
1106 for (i = 0; i < nb_tx; i++) {
1107 struct szedata2_tx_queue *txq = dev->data->tx_queues[i];
1115 eth_rx_queue_release(void *q)
1117 struct szedata2_rx_queue *rxq = (struct szedata2_rx_queue *)q;
1120 if (rxq->sze != NULL)
1121 szedata_close(rxq->sze);
1127 eth_tx_queue_release(void *q)
1129 struct szedata2_tx_queue *txq = (struct szedata2_tx_queue *)q;
1132 if (txq->sze != NULL)
1133 szedata_close(txq->sze);
1139 eth_dev_close(struct rte_eth_dev *dev)
1142 uint16_t nb_rx = dev->data->nb_rx_queues;
1143 uint16_t nb_tx = dev->data->nb_tx_queues;
1147 for (i = 0; i < nb_rx; i++) {
1148 eth_rx_queue_release(dev->data->rx_queues[i]);
1149 dev->data->rx_queues[i] = NULL;
1151 dev->data->nb_rx_queues = 0;
1152 for (i = 0; i < nb_tx; i++) {
1153 eth_tx_queue_release(dev->data->tx_queues[i]);
1154 dev->data->tx_queues[i] = NULL;
1156 dev->data->nb_tx_queues = 0;
1160 * Function takes value from first IBUF status register.
1161 * Values in IBUF and OBUF should be same.
1164 * Pointer to device private structure.
1166 * Link speed constant.
1168 static inline enum szedata2_link_speed
1169 get_link_speed(const struct pmd_internals *internals)
1171 const volatile struct szedata2_ibuf *ibuf =
1172 ibuf_ptr_by_index(internals->pci_rsc, 0);
1173 uint32_t speed = (szedata2_read32(&ibuf->ibuf_st) & 0x70) >> 4;
1176 return SZEDATA2_LINK_SPEED_10G;
1178 return SZEDATA2_LINK_SPEED_40G;
1180 return SZEDATA2_LINK_SPEED_100G;
1182 return SZEDATA2_LINK_SPEED_DEFAULT;
1187 eth_link_update(struct rte_eth_dev *dev,
1188 int wait_to_complete __rte_unused)
1190 struct rte_eth_link link;
1191 struct pmd_internals *internals = (struct pmd_internals *)
1192 dev->data->dev_private;
1193 const volatile struct szedata2_ibuf *ibuf;
1195 bool link_is_up = false;
1197 memset(&link, 0, sizeof(link));
1199 switch (get_link_speed(internals)) {
1200 case SZEDATA2_LINK_SPEED_10G:
1201 link.link_speed = ETH_SPEED_NUM_10G;
1203 case SZEDATA2_LINK_SPEED_40G:
1204 link.link_speed = ETH_SPEED_NUM_40G;
1206 case SZEDATA2_LINK_SPEED_100G:
1207 link.link_speed = ETH_SPEED_NUM_100G;
1210 link.link_speed = ETH_SPEED_NUM_10G;
1214 /* szedata2 uses only full duplex */
1215 link.link_duplex = ETH_LINK_FULL_DUPLEX;
1217 for (i = 0; i < szedata2_ibuf_count; i++) {
1218 ibuf = ibuf_ptr_by_index(internals->pci_rsc, i);
1220 * Link is considered up if at least one ibuf is enabled
1223 if (ibuf_is_enabled(ibuf) && ibuf_is_link_up(ibuf)) {
1229 link.link_status = link_is_up ? ETH_LINK_UP : ETH_LINK_DOWN;
1231 link.link_autoneg = ETH_LINK_FIXED;
1233 rte_eth_linkstatus_set(dev, &link);
1238 eth_dev_set_link_up(struct rte_eth_dev *dev)
1240 struct pmd_internals *internals = (struct pmd_internals *)
1241 dev->data->dev_private;
1244 for (i = 0; i < szedata2_ibuf_count; i++)
1245 ibuf_enable(ibuf_ptr_by_index(internals->pci_rsc, i));
1246 for (i = 0; i < szedata2_obuf_count; i++)
1247 obuf_enable(obuf_ptr_by_index(internals->pci_rsc, i));
1252 eth_dev_set_link_down(struct rte_eth_dev *dev)
1254 struct pmd_internals *internals = (struct pmd_internals *)
1255 dev->data->dev_private;
1258 for (i = 0; i < szedata2_ibuf_count; i++)
1259 ibuf_disable(ibuf_ptr_by_index(internals->pci_rsc, i));
1260 for (i = 0; i < szedata2_obuf_count; i++)
1261 obuf_disable(obuf_ptr_by_index(internals->pci_rsc, i));
1266 eth_rx_queue_setup(struct rte_eth_dev *dev,
1267 uint16_t rx_queue_id,
1268 uint16_t nb_rx_desc __rte_unused,
1269 unsigned int socket_id,
1270 const struct rte_eth_rxconf *rx_conf __rte_unused,
1271 struct rte_mempool *mb_pool)
1273 struct pmd_internals *internals = dev->data->dev_private;
1274 struct szedata2_rx_queue *rxq;
1276 uint32_t rx = 1 << rx_queue_id;
1279 if (dev->data->rx_queues[rx_queue_id] != NULL) {
1280 eth_rx_queue_release(dev->data->rx_queues[rx_queue_id]);
1281 dev->data->rx_queues[rx_queue_id] = NULL;
1284 rxq = rte_zmalloc_socket("szedata2 rx queue",
1285 sizeof(struct szedata2_rx_queue),
1286 RTE_CACHE_LINE_SIZE, socket_id);
1288 RTE_LOG(ERR, PMD, "rte_zmalloc_socket() failed for rx queue id "
1289 "%" PRIu16 "!\n", rx_queue_id);
1293 rxq->sze = szedata_open(internals->sze_dev);
1294 if (rxq->sze == NULL) {
1295 RTE_LOG(ERR, PMD, "szedata_open() failed for rx queue id "
1296 "%" PRIu16 "!\n", rx_queue_id);
1297 eth_rx_queue_release(rxq);
1300 ret = szedata_subscribe3(rxq->sze, &rx, &tx);
1301 if (ret != 0 || rx == 0) {
1302 RTE_LOG(ERR, PMD, "szedata_subscribe3() failed for rx queue id "
1303 "%" PRIu16 "!\n", rx_queue_id);
1304 eth_rx_queue_release(rxq);
1307 rxq->rx_channel = rx_queue_id;
1308 rxq->in_port = dev->data->port_id;
1309 rxq->mb_pool = mb_pool;
1314 dev->data->rx_queues[rx_queue_id] = rxq;
1316 RTE_LOG(DEBUG, PMD, "Configured rx queue id %" PRIu16 " on socket "
1317 "%u.\n", rx_queue_id, socket_id);
1323 eth_tx_queue_setup(struct rte_eth_dev *dev,
1324 uint16_t tx_queue_id,
1325 uint16_t nb_tx_desc __rte_unused,
1326 unsigned int socket_id,
1327 const struct rte_eth_txconf *tx_conf __rte_unused)
1329 struct pmd_internals *internals = dev->data->dev_private;
1330 struct szedata2_tx_queue *txq;
1333 uint32_t tx = 1 << tx_queue_id;
1335 if (dev->data->tx_queues[tx_queue_id] != NULL) {
1336 eth_tx_queue_release(dev->data->tx_queues[tx_queue_id]);
1337 dev->data->tx_queues[tx_queue_id] = NULL;
1340 txq = rte_zmalloc_socket("szedata2 tx queue",
1341 sizeof(struct szedata2_tx_queue),
1342 RTE_CACHE_LINE_SIZE, socket_id);
1344 RTE_LOG(ERR, PMD, "rte_zmalloc_socket() failed for tx queue id "
1345 "%" PRIu16 "!\n", tx_queue_id);
1349 txq->sze = szedata_open(internals->sze_dev);
1350 if (txq->sze == NULL) {
1351 RTE_LOG(ERR, PMD, "szedata_open() failed for tx queue id "
1352 "%" PRIu16 "!\n", tx_queue_id);
1353 eth_tx_queue_release(txq);
1356 ret = szedata_subscribe3(txq->sze, &rx, &tx);
1357 if (ret != 0 || tx == 0) {
1358 RTE_LOG(ERR, PMD, "szedata_subscribe3() failed for tx queue id "
1359 "%" PRIu16 "!\n", tx_queue_id);
1360 eth_tx_queue_release(txq);
1363 txq->tx_channel = tx_queue_id;
1368 dev->data->tx_queues[tx_queue_id] = txq;
1370 RTE_LOG(DEBUG, PMD, "Configured tx queue id %" PRIu16 " on socket "
1371 "%u.\n", tx_queue_id, socket_id);
1377 eth_mac_addr_set(struct rte_eth_dev *dev __rte_unused,
1378 struct ether_addr *mac_addr __rte_unused)
1383 eth_promiscuous_enable(struct rte_eth_dev *dev)
1385 struct pmd_internals *internals = (struct pmd_internals *)
1386 dev->data->dev_private;
1389 for (i = 0; i < szedata2_ibuf_count; i++) {
1390 ibuf_mac_mode_write(ibuf_ptr_by_index(internals->pci_rsc, i),
1391 SZEDATA2_MAC_CHMODE_PROMISC);
1396 eth_promiscuous_disable(struct rte_eth_dev *dev)
1398 struct pmd_internals *internals = (struct pmd_internals *)
1399 dev->data->dev_private;
1402 for (i = 0; i < szedata2_ibuf_count; i++) {
1403 ibuf_mac_mode_write(ibuf_ptr_by_index(internals->pci_rsc, i),
1404 SZEDATA2_MAC_CHMODE_ONLY_VALID);
1409 eth_allmulticast_enable(struct rte_eth_dev *dev)
1411 struct pmd_internals *internals = (struct pmd_internals *)
1412 dev->data->dev_private;
1415 for (i = 0; i < szedata2_ibuf_count; i++) {
1416 ibuf_mac_mode_write(ibuf_ptr_by_index(internals->pci_rsc, i),
1417 SZEDATA2_MAC_CHMODE_ALL_MULTICAST);
1422 eth_allmulticast_disable(struct rte_eth_dev *dev)
1424 struct pmd_internals *internals = (struct pmd_internals *)
1425 dev->data->dev_private;
1428 for (i = 0; i < szedata2_ibuf_count; i++) {
1429 ibuf_mac_mode_write(ibuf_ptr_by_index(internals->pci_rsc, i),
1430 SZEDATA2_MAC_CHMODE_ONLY_VALID);
1434 static const struct eth_dev_ops ops = {
1435 .dev_start = eth_dev_start,
1436 .dev_stop = eth_dev_stop,
1437 .dev_set_link_up = eth_dev_set_link_up,
1438 .dev_set_link_down = eth_dev_set_link_down,
1439 .dev_close = eth_dev_close,
1440 .dev_configure = eth_dev_configure,
1441 .dev_infos_get = eth_dev_info,
1442 .promiscuous_enable = eth_promiscuous_enable,
1443 .promiscuous_disable = eth_promiscuous_disable,
1444 .allmulticast_enable = eth_allmulticast_enable,
1445 .allmulticast_disable = eth_allmulticast_disable,
1446 .rx_queue_start = eth_rx_queue_start,
1447 .rx_queue_stop = eth_rx_queue_stop,
1448 .tx_queue_start = eth_tx_queue_start,
1449 .tx_queue_stop = eth_tx_queue_stop,
1450 .rx_queue_setup = eth_rx_queue_setup,
1451 .tx_queue_setup = eth_tx_queue_setup,
1452 .rx_queue_release = eth_rx_queue_release,
1453 .tx_queue_release = eth_tx_queue_release,
1454 .link_update = eth_link_update,
1455 .stats_get = eth_stats_get,
1456 .stats_reset = eth_stats_reset,
1457 .mac_addr_set = eth_mac_addr_set,
1461 * This function goes through sysfs and looks for an index of szedata2
1462 * device file (/dev/szedataIIX, where X is the index).
1469 get_szedata2_index(const struct rte_pci_addr *pcislot_addr, uint32_t *index)
1472 struct dirent *entry;
1476 char pcislot_path[PATH_MAX];
1482 dir = opendir("/sys/class/combo");
1487 * Iterate through all combosixX directories.
1488 * When the value in /sys/class/combo/combosixX/device/pcislot
1489 * file is the location of the ethernet device dev, "X" is the
1490 * index of the device.
1492 while ((entry = readdir(dir)) != NULL) {
1493 ret = sscanf(entry->d_name, "combosix%u", &tmp_index);
1497 snprintf(pcislot_path, PATH_MAX,
1498 "/sys/class/combo/combosix%u/device/pcislot",
1501 fd = fopen(pcislot_path, "r");
1505 ret = fscanf(fd, "%4" PRIx16 ":%2" PRIx8 ":%2" PRIx8 ".%" PRIx8,
1506 &domain, &bus, &devid, &function);
1511 if (pcislot_addr->domain == domain &&
1512 pcislot_addr->bus == bus &&
1513 pcislot_addr->devid == devid &&
1514 pcislot_addr->function == function) {
1526 rte_szedata2_eth_dev_init(struct rte_eth_dev *dev)
1528 struct rte_eth_dev_data *data = dev->data;
1529 struct pmd_internals *internals = (struct pmd_internals *)
1531 struct szedata *szedata_temp;
1533 uint32_t szedata2_index;
1534 struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(dev);
1535 struct rte_pci_addr *pci_addr = &pci_dev->addr;
1536 struct rte_mem_resource *pci_rsc =
1537 &pci_dev->mem_resource[PCI_RESOURCE_NUMBER];
1538 char rsc_filename[PATH_MAX];
1539 void *pci_resource_ptr = NULL;
1542 RTE_LOG(INFO, PMD, "Initializing szedata2 device (" PCI_PRI_FMT ")\n",
1543 pci_addr->domain, pci_addr->bus, pci_addr->devid,
1544 pci_addr->function);
1546 /* Get index of szedata2 device file and create path to device file */
1547 ret = get_szedata2_index(pci_addr, &szedata2_index);
1549 RTE_LOG(ERR, PMD, "Failed to get szedata2 device index!\n");
1552 snprintf(internals->sze_dev, PATH_MAX, SZEDATA2_DEV_PATH_FMT,
1555 RTE_LOG(INFO, PMD, "SZEDATA2 path: %s\n", internals->sze_dev);
1558 * Get number of available DMA RX and TX channels, which is maximum
1559 * number of queues that can be created and store it in private device
1562 szedata_temp = szedata_open(internals->sze_dev);
1563 if (szedata_temp == NULL) {
1564 RTE_LOG(ERR, PMD, "szedata_open(): failed to open %s",
1565 internals->sze_dev);
1568 internals->max_rx_queues = szedata_ifaces_available(szedata_temp,
1570 internals->max_tx_queues = szedata_ifaces_available(szedata_temp,
1572 szedata_close(szedata_temp);
1574 RTE_LOG(INFO, PMD, "Available DMA channels RX: %u TX: %u\n",
1575 internals->max_rx_queues, internals->max_tx_queues);
1577 /* Set rx, tx burst functions */
1578 if (data->scattered_rx == 1)
1579 dev->rx_pkt_burst = eth_szedata2_rx_scattered;
1581 dev->rx_pkt_burst = eth_szedata2_rx;
1582 dev->tx_pkt_burst = eth_szedata2_tx;
1584 /* Set function callbacks for Ethernet API */
1585 dev->dev_ops = &ops;
1587 rte_eth_copy_pci_info(dev, pci_dev);
1589 /* mmap pci resource0 file to rte_mem_resource structure */
1590 if (pci_dev->mem_resource[PCI_RESOURCE_NUMBER].phys_addr ==
1592 RTE_LOG(ERR, PMD, "Missing resource%u file\n",
1593 PCI_RESOURCE_NUMBER);
1596 snprintf(rsc_filename, PATH_MAX,
1597 "%s/" PCI_PRI_FMT "/resource%u", rte_pci_get_sysfs_path(),
1598 pci_addr->domain, pci_addr->bus,
1599 pci_addr->devid, pci_addr->function, PCI_RESOURCE_NUMBER);
1600 fd = open(rsc_filename, O_RDWR);
1602 RTE_LOG(ERR, PMD, "Could not open file %s\n", rsc_filename);
1606 pci_resource_ptr = mmap(0,
1607 pci_dev->mem_resource[PCI_RESOURCE_NUMBER].len,
1608 PROT_READ | PROT_WRITE, MAP_SHARED, fd, 0);
1610 if (pci_resource_ptr == MAP_FAILED) {
1611 RTE_LOG(ERR, PMD, "Could not mmap file %s (fd = %d)\n",
1615 pci_dev->mem_resource[PCI_RESOURCE_NUMBER].addr = pci_resource_ptr;
1616 internals->pci_rsc = pci_rsc;
1618 RTE_LOG(DEBUG, PMD, "resource%u phys_addr = 0x%llx len = %llu "
1619 "virt addr = %llx\n", PCI_RESOURCE_NUMBER,
1620 (unsigned long long)pci_rsc->phys_addr,
1621 (unsigned long long)pci_rsc->len,
1622 (unsigned long long)pci_rsc->addr);
1624 /* Get link state */
1625 eth_link_update(dev, 0);
1627 /* Allocate space for one mac address */
1628 data->mac_addrs = rte_zmalloc(data->name, sizeof(struct ether_addr),
1629 RTE_CACHE_LINE_SIZE);
1630 if (data->mac_addrs == NULL) {
1631 RTE_LOG(ERR, PMD, "Could not alloc space for MAC address!\n");
1632 munmap(pci_dev->mem_resource[PCI_RESOURCE_NUMBER].addr,
1633 pci_dev->mem_resource[PCI_RESOURCE_NUMBER].len);
1637 ether_addr_copy(ð_addr, data->mac_addrs);
1639 /* At initial state COMBO card is in promiscuous mode so disable it */
1640 eth_promiscuous_disable(dev);
1642 RTE_LOG(INFO, PMD, "szedata2 device ("
1643 PCI_PRI_FMT ") successfully initialized\n",
1644 pci_addr->domain, pci_addr->bus, pci_addr->devid,
1645 pci_addr->function);
1651 rte_szedata2_eth_dev_uninit(struct rte_eth_dev *dev)
1653 struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(dev);
1654 struct rte_pci_addr *pci_addr = &pci_dev->addr;
1656 rte_free(dev->data->mac_addrs);
1657 dev->data->mac_addrs = NULL;
1658 munmap(pci_dev->mem_resource[PCI_RESOURCE_NUMBER].addr,
1659 pci_dev->mem_resource[PCI_RESOURCE_NUMBER].len);
1661 RTE_LOG(INFO, PMD, "szedata2 device ("
1662 PCI_PRI_FMT ") successfully uninitialized\n",
1663 pci_addr->domain, pci_addr->bus, pci_addr->devid,
1664 pci_addr->function);
1669 static const struct rte_pci_id rte_szedata2_pci_id_table[] = {
1671 RTE_PCI_DEVICE(PCI_VENDOR_ID_NETCOPE,
1672 PCI_DEVICE_ID_NETCOPE_COMBO80G)
1675 RTE_PCI_DEVICE(PCI_VENDOR_ID_NETCOPE,
1676 PCI_DEVICE_ID_NETCOPE_COMBO100G)
1679 RTE_PCI_DEVICE(PCI_VENDOR_ID_NETCOPE,
1680 PCI_DEVICE_ID_NETCOPE_COMBO100G2)
1687 static int szedata2_eth_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
1688 struct rte_pci_device *pci_dev)
1690 return rte_eth_dev_pci_generic_probe(pci_dev,
1691 sizeof(struct pmd_internals), rte_szedata2_eth_dev_init);
1694 static int szedata2_eth_pci_remove(struct rte_pci_device *pci_dev)
1696 return rte_eth_dev_pci_generic_remove(pci_dev,
1697 rte_szedata2_eth_dev_uninit);
1700 static struct rte_pci_driver szedata2_eth_driver = {
1701 .id_table = rte_szedata2_pci_id_table,
1702 .probe = szedata2_eth_pci_probe,
1703 .remove = szedata2_eth_pci_remove,
1706 RTE_PMD_REGISTER_PCI(RTE_SZEDATA2_DRIVER_NAME, szedata2_eth_driver);
1707 RTE_PMD_REGISTER_PCI_TABLE(RTE_SZEDATA2_DRIVER_NAME, rte_szedata2_pci_id_table);
1708 RTE_PMD_REGISTER_KMOD_DEP(RTE_SZEDATA2_DRIVER_NAME,
1709 "* combo6core & combov3 & szedata2 & szedata2_cv3");