4 * Copyright (c) 2015 - 2016 CESNET
7 * Redistribution and use in source and binary forms, with or without
8 * modification, are permitted provided that the following conditions
11 * * Redistributions of source code must retain the above copyright
12 * notice, this list of conditions and the following disclaimer.
13 * * Redistributions in binary form must reproduce the above copyright
14 * notice, this list of conditions and the following disclaimer in
15 * the documentation and/or other materials provided with the
17 * * Neither the name of CESNET nor the names of its
18 * contributors may be used to endorse or promote products derived
19 * from this software without specific prior written permission.
21 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
22 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
23 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
24 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
25 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
26 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
27 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
28 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
29 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
30 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
31 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
38 #include <sys/types.h>
47 #include <rte_ethdev_driver.h>
48 #include <rte_ethdev_pci.h>
49 #include <rte_malloc.h>
50 #include <rte_memcpy.h>
51 #include <rte_kvargs.h>
54 #include "rte_eth_szedata2.h"
55 #include "szedata2_iobuf.h"
57 #define RTE_ETH_SZEDATA2_MAX_RX_QUEUES 32
58 #define RTE_ETH_SZEDATA2_MAX_TX_QUEUES 32
59 #define RTE_ETH_SZEDATA2_TX_LOCK_SIZE (32 * 1024 * 1024)
62 * size of szedata2_packet header with alignment
64 #define RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED 8
66 #define RTE_SZEDATA2_DRIVER_NAME net_szedata2
68 #define SZEDATA2_DEV_PATH_FMT "/dev/szedataII%u"
70 struct pmd_internals {
71 struct rte_eth_dev *dev;
72 uint16_t max_rx_queues;
73 uint16_t max_tx_queues;
74 char sze_dev[PATH_MAX];
75 struct rte_mem_resource *pci_rsc;
78 struct szedata2_rx_queue {
79 struct pmd_internals *priv;
83 struct rte_mempool *mb_pool;
84 volatile uint64_t rx_pkts;
85 volatile uint64_t rx_bytes;
86 volatile uint64_t err_pkts;
89 struct szedata2_tx_queue {
90 struct pmd_internals *priv;
93 volatile uint64_t tx_pkts;
94 volatile uint64_t tx_bytes;
95 volatile uint64_t err_pkts;
98 static struct ether_addr eth_addr = {
99 .addr_bytes = { 0x00, 0x11, 0x17, 0x00, 0x00, 0x00 }
103 eth_szedata2_rx(void *queue,
104 struct rte_mbuf **bufs,
108 struct rte_mbuf *mbuf;
109 struct szedata2_rx_queue *sze_q = queue;
110 struct rte_pktmbuf_pool_private *mbp_priv;
115 uint16_t packet_size;
116 uint64_t num_bytes = 0;
117 struct szedata *sze = sze_q->sze;
118 uint8_t *header_ptr = NULL; /* header of packet */
119 uint8_t *packet_ptr1 = NULL;
120 uint8_t *packet_ptr2 = NULL;
121 uint16_t packet_len1 = 0;
122 uint16_t packet_len2 = 0;
123 uint16_t hw_data_align;
125 if (unlikely(sze_q->sze == NULL || nb_pkts == 0))
129 * Reads the given number of packets from szedata2 channel given
130 * by queue and copies the packet data into a newly allocated mbuf
133 for (i = 0; i < nb_pkts; i++) {
134 mbuf = rte_pktmbuf_alloc(sze_q->mb_pool);
136 if (unlikely(mbuf == NULL)) {
137 sze_q->priv->dev->data->rx_mbuf_alloc_failed++;
141 /* get the next sze packet */
142 if (sze->ct_rx_lck != NULL && !sze->ct_rx_rem_bytes &&
143 sze->ct_rx_lck->next == NULL) {
144 /* unlock old data */
145 szedata_rx_unlock_data(sze_q->sze, sze->ct_rx_lck_orig);
146 sze->ct_rx_lck_orig = NULL;
147 sze->ct_rx_lck = NULL;
150 if (!sze->ct_rx_rem_bytes && sze->ct_rx_lck_orig == NULL) {
151 /* nothing to read, lock new data */
152 sze->ct_rx_lck = szedata_rx_lock_data(sze_q->sze, ~0U);
153 sze->ct_rx_lck_orig = sze->ct_rx_lck;
155 if (sze->ct_rx_lck == NULL) {
156 /* nothing to lock */
157 rte_pktmbuf_free(mbuf);
161 sze->ct_rx_cur_ptr = sze->ct_rx_lck->start;
162 sze->ct_rx_rem_bytes = sze->ct_rx_lck->len;
164 if (!sze->ct_rx_rem_bytes) {
165 rte_pktmbuf_free(mbuf);
170 if (sze->ct_rx_rem_bytes < RTE_SZE2_PACKET_HEADER_SIZE) {
173 * copy parts of header to merge buffer
175 if (sze->ct_rx_lck->next == NULL) {
176 rte_pktmbuf_free(mbuf);
180 /* copy first part of header */
181 rte_memcpy(sze->ct_rx_buffer, sze->ct_rx_cur_ptr,
182 sze->ct_rx_rem_bytes);
184 /* copy second part of header */
185 sze->ct_rx_lck = sze->ct_rx_lck->next;
186 sze->ct_rx_cur_ptr = sze->ct_rx_lck->start;
187 rte_memcpy(sze->ct_rx_buffer + sze->ct_rx_rem_bytes,
189 RTE_SZE2_PACKET_HEADER_SIZE -
190 sze->ct_rx_rem_bytes);
192 sze->ct_rx_cur_ptr += RTE_SZE2_PACKET_HEADER_SIZE -
193 sze->ct_rx_rem_bytes;
194 sze->ct_rx_rem_bytes = sze->ct_rx_lck->len -
195 RTE_SZE2_PACKET_HEADER_SIZE +
196 sze->ct_rx_rem_bytes;
198 header_ptr = (uint8_t *)sze->ct_rx_buffer;
201 header_ptr = (uint8_t *)sze->ct_rx_cur_ptr;
202 sze->ct_rx_cur_ptr += RTE_SZE2_PACKET_HEADER_SIZE;
203 sze->ct_rx_rem_bytes -= RTE_SZE2_PACKET_HEADER_SIZE;
206 sg_size = le16toh(*((uint16_t *)header_ptr));
207 hw_size = le16toh(*(((uint16_t *)header_ptr) + 1));
208 packet_size = sg_size -
209 RTE_SZE2_ALIGN8(RTE_SZE2_PACKET_HEADER_SIZE + hw_size);
212 /* checks if packet all right */
214 errx(5, "Zero segsize");
216 /* check sg_size and hwsize */
217 if (hw_size > sg_size - RTE_SZE2_PACKET_HEADER_SIZE) {
218 errx(10, "Hwsize bigger than expected. Segsize: %d, "
219 "hwsize: %d", sg_size, hw_size);
223 RTE_SZE2_ALIGN8(RTE_SZE2_PACKET_HEADER_SIZE + hw_size) -
224 RTE_SZE2_PACKET_HEADER_SIZE;
226 if (sze->ct_rx_rem_bytes >=
228 RTE_SZE2_PACKET_HEADER_SIZE)) {
230 /* one packet ready - go to another */
231 packet_ptr1 = sze->ct_rx_cur_ptr + hw_data_align;
232 packet_len1 = packet_size;
236 sze->ct_rx_cur_ptr += RTE_SZE2_ALIGN8(sg_size) -
237 RTE_SZE2_PACKET_HEADER_SIZE;
238 sze->ct_rx_rem_bytes -= RTE_SZE2_ALIGN8(sg_size) -
239 RTE_SZE2_PACKET_HEADER_SIZE;
242 if (sze->ct_rx_lck->next == NULL) {
243 errx(6, "Need \"next\" lock, "
244 "but it is missing: %u",
245 sze->ct_rx_rem_bytes);
249 if (sze->ct_rx_rem_bytes <= hw_data_align) {
250 uint16_t rem_size = hw_data_align -
251 sze->ct_rx_rem_bytes;
253 /* MOVE to next lock */
254 sze->ct_rx_lck = sze->ct_rx_lck->next;
256 (void *)(((uint8_t *)
257 (sze->ct_rx_lck->start)) + rem_size);
259 packet_ptr1 = sze->ct_rx_cur_ptr;
260 packet_len1 = packet_size;
264 sze->ct_rx_cur_ptr +=
265 RTE_SZE2_ALIGN8(packet_size);
266 sze->ct_rx_rem_bytes = sze->ct_rx_lck->len -
267 rem_size - RTE_SZE2_ALIGN8(packet_size);
269 /* get pointer and length from first part */
270 packet_ptr1 = sze->ct_rx_cur_ptr +
272 packet_len1 = sze->ct_rx_rem_bytes -
275 /* MOVE to next lock */
276 sze->ct_rx_lck = sze->ct_rx_lck->next;
277 sze->ct_rx_cur_ptr = sze->ct_rx_lck->start;
279 /* get pointer and length from second part */
280 packet_ptr2 = sze->ct_rx_cur_ptr;
281 packet_len2 = packet_size - packet_len1;
283 sze->ct_rx_cur_ptr +=
284 RTE_SZE2_ALIGN8(packet_size) -
286 sze->ct_rx_rem_bytes = sze->ct_rx_lck->len -
287 (RTE_SZE2_ALIGN8(packet_size) -
292 if (unlikely(packet_ptr1 == NULL)) {
293 rte_pktmbuf_free(mbuf);
297 /* get the space available for data in the mbuf */
298 mbp_priv = rte_mempool_get_priv(sze_q->mb_pool);
299 buf_size = (uint16_t)(mbp_priv->mbuf_data_room_size -
300 RTE_PKTMBUF_HEADROOM);
302 if (packet_size <= buf_size) {
303 /* sze packet will fit in one mbuf, go ahead and copy */
304 rte_memcpy(rte_pktmbuf_mtod(mbuf, void *),
305 packet_ptr1, packet_len1);
306 if (packet_ptr2 != NULL) {
307 rte_memcpy((void *)(rte_pktmbuf_mtod(mbuf,
308 uint8_t *) + packet_len1),
309 packet_ptr2, packet_len2);
311 mbuf->data_len = (uint16_t)packet_size;
313 mbuf->pkt_len = packet_size;
314 mbuf->port = sze_q->in_port;
317 num_bytes += packet_size;
320 * sze packet will not fit in one mbuf,
321 * scattered mode is not enabled, drop packet
324 "SZE segment %d bytes will not fit in one mbuf "
325 "(%d bytes), scattered mode is not enabled, "
327 packet_size, buf_size);
328 rte_pktmbuf_free(mbuf);
332 sze_q->rx_pkts += num_rx;
333 sze_q->rx_bytes += num_bytes;
338 eth_szedata2_rx_scattered(void *queue,
339 struct rte_mbuf **bufs,
343 struct rte_mbuf *mbuf;
344 struct szedata2_rx_queue *sze_q = queue;
345 struct rte_pktmbuf_pool_private *mbp_priv;
350 uint16_t packet_size;
351 uint64_t num_bytes = 0;
352 struct szedata *sze = sze_q->sze;
353 uint8_t *header_ptr = NULL; /* header of packet */
354 uint8_t *packet_ptr1 = NULL;
355 uint8_t *packet_ptr2 = NULL;
356 uint16_t packet_len1 = 0;
357 uint16_t packet_len2 = 0;
358 uint16_t hw_data_align;
359 uint64_t *mbuf_failed_ptr =
360 &sze_q->priv->dev->data->rx_mbuf_alloc_failed;
362 if (unlikely(sze_q->sze == NULL || nb_pkts == 0))
366 * Reads the given number of packets from szedata2 channel given
367 * by queue and copies the packet data into a newly allocated mbuf
370 for (i = 0; i < nb_pkts; i++) {
371 const struct szedata_lock *ct_rx_lck_backup;
372 unsigned int ct_rx_rem_bytes_backup;
373 unsigned char *ct_rx_cur_ptr_backup;
375 /* get the next sze packet */
376 if (sze->ct_rx_lck != NULL && !sze->ct_rx_rem_bytes &&
377 sze->ct_rx_lck->next == NULL) {
378 /* unlock old data */
379 szedata_rx_unlock_data(sze_q->sze, sze->ct_rx_lck_orig);
380 sze->ct_rx_lck_orig = NULL;
381 sze->ct_rx_lck = NULL;
385 * Store items from sze structure which can be changed
386 * before mbuf allocating. Use these items in case of mbuf
387 * allocating failure.
389 ct_rx_lck_backup = sze->ct_rx_lck;
390 ct_rx_rem_bytes_backup = sze->ct_rx_rem_bytes;
391 ct_rx_cur_ptr_backup = sze->ct_rx_cur_ptr;
393 if (!sze->ct_rx_rem_bytes && sze->ct_rx_lck_orig == NULL) {
394 /* nothing to read, lock new data */
395 sze->ct_rx_lck = szedata_rx_lock_data(sze_q->sze, ~0U);
396 sze->ct_rx_lck_orig = sze->ct_rx_lck;
399 * Backup items from sze structure must be updated
400 * after locking to contain pointers to new locks.
402 ct_rx_lck_backup = sze->ct_rx_lck;
403 ct_rx_rem_bytes_backup = sze->ct_rx_rem_bytes;
404 ct_rx_cur_ptr_backup = sze->ct_rx_cur_ptr;
406 if (sze->ct_rx_lck == NULL)
407 /* nothing to lock */
410 sze->ct_rx_cur_ptr = sze->ct_rx_lck->start;
411 sze->ct_rx_rem_bytes = sze->ct_rx_lck->len;
413 if (!sze->ct_rx_rem_bytes)
417 if (sze->ct_rx_rem_bytes < RTE_SZE2_PACKET_HEADER_SIZE) {
419 * cut in header - copy parts of header to merge buffer
421 if (sze->ct_rx_lck->next == NULL)
424 /* copy first part of header */
425 rte_memcpy(sze->ct_rx_buffer, sze->ct_rx_cur_ptr,
426 sze->ct_rx_rem_bytes);
428 /* copy second part of header */
429 sze->ct_rx_lck = sze->ct_rx_lck->next;
430 sze->ct_rx_cur_ptr = sze->ct_rx_lck->start;
431 rte_memcpy(sze->ct_rx_buffer + sze->ct_rx_rem_bytes,
433 RTE_SZE2_PACKET_HEADER_SIZE -
434 sze->ct_rx_rem_bytes);
436 sze->ct_rx_cur_ptr += RTE_SZE2_PACKET_HEADER_SIZE -
437 sze->ct_rx_rem_bytes;
438 sze->ct_rx_rem_bytes = sze->ct_rx_lck->len -
439 RTE_SZE2_PACKET_HEADER_SIZE +
440 sze->ct_rx_rem_bytes;
442 header_ptr = (uint8_t *)sze->ct_rx_buffer;
445 header_ptr = (uint8_t *)sze->ct_rx_cur_ptr;
446 sze->ct_rx_cur_ptr += RTE_SZE2_PACKET_HEADER_SIZE;
447 sze->ct_rx_rem_bytes -= RTE_SZE2_PACKET_HEADER_SIZE;
450 sg_size = le16toh(*((uint16_t *)header_ptr));
451 hw_size = le16toh(*(((uint16_t *)header_ptr) + 1));
452 packet_size = sg_size -
453 RTE_SZE2_ALIGN8(RTE_SZE2_PACKET_HEADER_SIZE + hw_size);
456 /* checks if packet all right */
458 errx(5, "Zero segsize");
460 /* check sg_size and hwsize */
461 if (hw_size > sg_size - RTE_SZE2_PACKET_HEADER_SIZE) {
462 errx(10, "Hwsize bigger than expected. Segsize: %d, "
463 "hwsize: %d", sg_size, hw_size);
467 RTE_SZE2_ALIGN8((RTE_SZE2_PACKET_HEADER_SIZE +
468 hw_size)) - RTE_SZE2_PACKET_HEADER_SIZE;
470 if (sze->ct_rx_rem_bytes >=
472 RTE_SZE2_PACKET_HEADER_SIZE)) {
474 /* one packet ready - go to another */
475 packet_ptr1 = sze->ct_rx_cur_ptr + hw_data_align;
476 packet_len1 = packet_size;
480 sze->ct_rx_cur_ptr += RTE_SZE2_ALIGN8(sg_size) -
481 RTE_SZE2_PACKET_HEADER_SIZE;
482 sze->ct_rx_rem_bytes -= RTE_SZE2_ALIGN8(sg_size) -
483 RTE_SZE2_PACKET_HEADER_SIZE;
486 if (sze->ct_rx_lck->next == NULL) {
487 errx(6, "Need \"next\" lock, but it is "
488 "missing: %u", sze->ct_rx_rem_bytes);
492 if (sze->ct_rx_rem_bytes <= hw_data_align) {
493 uint16_t rem_size = hw_data_align -
494 sze->ct_rx_rem_bytes;
496 /* MOVE to next lock */
497 sze->ct_rx_lck = sze->ct_rx_lck->next;
499 (void *)(((uint8_t *)
500 (sze->ct_rx_lck->start)) + rem_size);
502 packet_ptr1 = sze->ct_rx_cur_ptr;
503 packet_len1 = packet_size;
507 sze->ct_rx_cur_ptr +=
508 RTE_SZE2_ALIGN8(packet_size);
509 sze->ct_rx_rem_bytes = sze->ct_rx_lck->len -
510 rem_size - RTE_SZE2_ALIGN8(packet_size);
512 /* get pointer and length from first part */
513 packet_ptr1 = sze->ct_rx_cur_ptr +
515 packet_len1 = sze->ct_rx_rem_bytes -
518 /* MOVE to next lock */
519 sze->ct_rx_lck = sze->ct_rx_lck->next;
520 sze->ct_rx_cur_ptr = sze->ct_rx_lck->start;
522 /* get pointer and length from second part */
523 packet_ptr2 = sze->ct_rx_cur_ptr;
524 packet_len2 = packet_size - packet_len1;
526 sze->ct_rx_cur_ptr +=
527 RTE_SZE2_ALIGN8(packet_size) -
529 sze->ct_rx_rem_bytes = sze->ct_rx_lck->len -
530 (RTE_SZE2_ALIGN8(packet_size) -
535 if (unlikely(packet_ptr1 == NULL))
538 mbuf = rte_pktmbuf_alloc(sze_q->mb_pool);
540 if (unlikely(mbuf == NULL)) {
542 * Restore items from sze structure to state after
543 * unlocking (eventually locking).
545 sze->ct_rx_lck = ct_rx_lck_backup;
546 sze->ct_rx_rem_bytes = ct_rx_rem_bytes_backup;
547 sze->ct_rx_cur_ptr = ct_rx_cur_ptr_backup;
548 sze_q->priv->dev->data->rx_mbuf_alloc_failed++;
552 /* get the space available for data in the mbuf */
553 mbp_priv = rte_mempool_get_priv(sze_q->mb_pool);
554 buf_size = (uint16_t)(mbp_priv->mbuf_data_room_size -
555 RTE_PKTMBUF_HEADROOM);
557 if (packet_size <= buf_size) {
558 /* sze packet will fit in one mbuf, go ahead and copy */
559 rte_memcpy(rte_pktmbuf_mtod(mbuf, void *),
560 packet_ptr1, packet_len1);
561 if (packet_ptr2 != NULL) {
563 (rte_pktmbuf_mtod(mbuf, uint8_t *) +
564 packet_len1), packet_ptr2, packet_len2);
566 mbuf->data_len = (uint16_t)packet_size;
569 * sze packet will not fit in one mbuf,
570 * scatter packet into more mbufs
572 struct rte_mbuf *m = mbuf;
573 uint16_t len = rte_pktmbuf_tailroom(mbuf);
575 /* copy first part of packet */
576 /* fill first mbuf */
577 rte_memcpy(rte_pktmbuf_append(mbuf, len), packet_ptr1,
580 packet_ptr1 = ((uint8_t *)packet_ptr1) + len;
582 while (packet_len1 > 0) {
584 m->next = rte_pktmbuf_alloc(sze_q->mb_pool);
586 if (unlikely(m->next == NULL)) {
587 rte_pktmbuf_free(mbuf);
589 * Restore items from sze structure
590 * to state after unlocking (eventually
593 sze->ct_rx_lck = ct_rx_lck_backup;
594 sze->ct_rx_rem_bytes =
595 ct_rx_rem_bytes_backup;
597 ct_rx_cur_ptr_backup;
598 (*mbuf_failed_ptr)++;
604 len = RTE_MIN(rte_pktmbuf_tailroom(m),
606 rte_memcpy(rte_pktmbuf_append(mbuf, len),
611 packet_ptr1 = ((uint8_t *)packet_ptr1) + len;
614 if (packet_ptr2 != NULL) {
615 /* copy second part of packet, if exists */
616 /* fill the rest of currently last mbuf */
617 len = rte_pktmbuf_tailroom(m);
618 rte_memcpy(rte_pktmbuf_append(mbuf, len),
621 packet_ptr2 = ((uint8_t *)packet_ptr2) + len;
623 while (packet_len2 > 0) {
625 m->next = rte_pktmbuf_alloc(
628 if (unlikely(m->next == NULL)) {
629 rte_pktmbuf_free(mbuf);
631 * Restore items from sze
632 * structure to state after
633 * unlocking (eventually
638 sze->ct_rx_rem_bytes =
639 ct_rx_rem_bytes_backup;
641 ct_rx_cur_ptr_backup;
642 (*mbuf_failed_ptr)++;
648 len = RTE_MIN(rte_pktmbuf_tailroom(m),
651 rte_pktmbuf_append(mbuf, len),
656 packet_ptr2 = ((uint8_t *)packet_ptr2) +
661 mbuf->pkt_len = packet_size;
662 mbuf->port = sze_q->in_port;
665 num_bytes += packet_size;
669 sze_q->rx_pkts += num_rx;
670 sze_q->rx_bytes += num_bytes;
675 eth_szedata2_tx(void *queue,
676 struct rte_mbuf **bufs,
679 struct rte_mbuf *mbuf;
680 struct szedata2_tx_queue *sze_q = queue;
682 uint64_t num_bytes = 0;
684 const struct szedata_lock *lck;
690 uint32_t unlock_size;
693 uint16_t pkt_left = nb_pkts;
695 if (sze_q->sze == NULL || nb_pkts == 0)
698 while (pkt_left > 0) {
700 lck = szedata_tx_lock_data(sze_q->sze,
701 RTE_ETH_SZEDATA2_TX_LOCK_SIZE,
707 lock_size = lck->len;
708 lock_size2 = lck->next ? lck->next->len : 0;
711 mbuf = bufs[nb_pkts - pkt_left];
713 pkt_len = mbuf->pkt_len;
714 mbuf_segs = mbuf->nb_segs;
716 hwpkt_len = RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED +
717 RTE_SZE2_ALIGN8(pkt_len);
719 if (lock_size + lock_size2 < hwpkt_len) {
720 szedata_tx_unlock_data(sze_q->sze, lck, unlock_size);
724 num_bytes += pkt_len;
726 if (lock_size > hwpkt_len) {
731 /* write packet length at first 2 bytes in 8B header */
732 *((uint16_t *)dst) = htole16(
733 RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED +
735 *(((uint16_t *)dst) + 1) = htole16(0);
737 /* copy packet from mbuf */
738 tmp_dst = ((uint8_t *)(dst)) +
739 RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED;
740 if (mbuf_segs == 1) {
742 * non-scattered packet,
743 * transmit from one mbuf
746 rte_pktmbuf_mtod(mbuf, const void *),
749 /* scattered packet, transmit from more mbufs */
750 struct rte_mbuf *m = mbuf;
756 tmp_dst = ((uint8_t *)(tmp_dst)) +
763 dst = ((uint8_t *)dst) + hwpkt_len;
764 unlock_size += hwpkt_len;
765 lock_size -= hwpkt_len;
767 rte_pktmbuf_free(mbuf);
771 szedata_tx_unlock_data(sze_q->sze, lck,
776 } else if (lock_size + lock_size2 >= hwpkt_len) {
780 /* write packet length at first 2 bytes in 8B header */
782 htole16(RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED +
784 *(((uint16_t *)dst) + 1) = htole16(0);
787 * If the raw packet (pkt_len) is smaller than lock_size
788 * get the correct length for memcpy
791 pkt_len < lock_size -
792 RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED ?
794 lock_size - RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED;
796 rem_len = hwpkt_len - lock_size;
798 tmp_dst = ((uint8_t *)(dst)) +
799 RTE_SZE2_PACKET_HEADER_SIZE_ALIGNED;
800 if (mbuf_segs == 1) {
802 * non-scattered packet,
803 * transmit from one mbuf
805 /* copy part of packet to first area */
807 rte_pktmbuf_mtod(mbuf, const void *),
811 dst = lck->next->start;
813 /* copy part of packet to second area */
815 (const void *)(rte_pktmbuf_mtod(mbuf,
817 write_len), pkt_len - write_len);
819 /* scattered packet, transmit from more mbufs */
820 struct rte_mbuf *m = mbuf;
821 uint16_t written = 0;
822 uint16_t to_write = 0;
823 bool new_mbuf = true;
824 uint16_t write_off = 0;
826 /* copy part of packet to first area */
827 while (m && written < write_len) {
828 to_write = RTE_MIN(m->data_len,
829 write_len - written);
835 tmp_dst = ((uint8_t *)(tmp_dst)) +
837 if (m->data_len <= write_len -
848 dst = lck->next->start;
852 write_off = new_mbuf ? 0 : to_write;
854 /* copy part of packet to second area */
855 while (m && written < pkt_len - write_len) {
856 rte_memcpy(tmp_dst, (const void *)
858 uint8_t *) + write_off),
859 m->data_len - write_off);
861 tmp_dst = ((uint8_t *)(tmp_dst)) +
862 (m->data_len - write_off);
863 written += m->data_len - write_off;
869 dst = ((uint8_t *)dst) + rem_len;
870 unlock_size += hwpkt_len;
871 lock_size = lock_size2 - rem_len;
874 rte_pktmbuf_free(mbuf);
878 szedata_tx_unlock_data(sze_q->sze, lck, unlock_size);
882 sze_q->tx_pkts += num_tx;
883 sze_q->err_pkts += nb_pkts - num_tx;
884 sze_q->tx_bytes += num_bytes;
889 eth_rx_queue_start(struct rte_eth_dev *dev, uint16_t rxq_id)
891 struct szedata2_rx_queue *rxq = dev->data->rx_queues[rxq_id];
893 struct pmd_internals *internals = (struct pmd_internals *)
894 dev->data->dev_private;
896 if (rxq->sze == NULL) {
897 uint32_t rx = 1 << rxq->rx_channel;
899 rxq->sze = szedata_open(internals->sze_dev);
900 if (rxq->sze == NULL)
902 ret = szedata_subscribe3(rxq->sze, &rx, &tx);
903 if (ret != 0 || rx == 0)
907 ret = szedata_start(rxq->sze);
910 dev->data->rx_queue_state[rxq_id] = RTE_ETH_QUEUE_STATE_STARTED;
914 szedata_close(rxq->sze);
920 eth_rx_queue_stop(struct rte_eth_dev *dev, uint16_t rxq_id)
922 struct szedata2_rx_queue *rxq = dev->data->rx_queues[rxq_id];
924 if (rxq->sze != NULL) {
925 szedata_close(rxq->sze);
929 dev->data->rx_queue_state[rxq_id] = RTE_ETH_QUEUE_STATE_STOPPED;
934 eth_tx_queue_start(struct rte_eth_dev *dev, uint16_t txq_id)
936 struct szedata2_tx_queue *txq = dev->data->tx_queues[txq_id];
938 struct pmd_internals *internals = (struct pmd_internals *)
939 dev->data->dev_private;
941 if (txq->sze == NULL) {
943 uint32_t tx = 1 << txq->tx_channel;
944 txq->sze = szedata_open(internals->sze_dev);
945 if (txq->sze == NULL)
947 ret = szedata_subscribe3(txq->sze, &rx, &tx);
948 if (ret != 0 || tx == 0)
952 ret = szedata_start(txq->sze);
955 dev->data->tx_queue_state[txq_id] = RTE_ETH_QUEUE_STATE_STARTED;
959 szedata_close(txq->sze);
965 eth_tx_queue_stop(struct rte_eth_dev *dev, uint16_t txq_id)
967 struct szedata2_tx_queue *txq = dev->data->tx_queues[txq_id];
969 if (txq->sze != NULL) {
970 szedata_close(txq->sze);
974 dev->data->tx_queue_state[txq_id] = RTE_ETH_QUEUE_STATE_STOPPED;
979 eth_dev_start(struct rte_eth_dev *dev)
983 uint16_t nb_rx = dev->data->nb_rx_queues;
984 uint16_t nb_tx = dev->data->nb_tx_queues;
986 for (i = 0; i < nb_rx; i++) {
987 ret = eth_rx_queue_start(dev, i);
992 for (i = 0; i < nb_tx; i++) {
993 ret = eth_tx_queue_start(dev, i);
1001 for (i = 0; i < nb_tx; i++)
1002 eth_tx_queue_stop(dev, i);
1004 for (i = 0; i < nb_rx; i++)
1005 eth_rx_queue_stop(dev, i);
1010 eth_dev_stop(struct rte_eth_dev *dev)
1013 uint16_t nb_rx = dev->data->nb_rx_queues;
1014 uint16_t nb_tx = dev->data->nb_tx_queues;
1016 for (i = 0; i < nb_tx; i++)
1017 eth_tx_queue_stop(dev, i);
1019 for (i = 0; i < nb_rx; i++)
1020 eth_rx_queue_stop(dev, i);
1024 eth_dev_configure(struct rte_eth_dev *dev)
1026 struct rte_eth_dev_data *data = dev->data;
1027 if (data->dev_conf.rxmode.offloads & DEV_RX_OFFLOAD_SCATTER) {
1028 dev->rx_pkt_burst = eth_szedata2_rx_scattered;
1029 data->scattered_rx = 1;
1031 dev->rx_pkt_burst = eth_szedata2_rx;
1032 data->scattered_rx = 0;
1038 eth_dev_info(struct rte_eth_dev *dev,
1039 struct rte_eth_dev_info *dev_info)
1041 struct pmd_internals *internals = dev->data->dev_private;
1043 dev_info->pci_dev = RTE_ETH_DEV_TO_PCI(dev);
1044 dev_info->if_index = 0;
1045 dev_info->max_mac_addrs = 1;
1046 dev_info->max_rx_pktlen = (uint32_t)-1;
1047 dev_info->max_rx_queues = internals->max_rx_queues;
1048 dev_info->max_tx_queues = internals->max_tx_queues;
1049 dev_info->min_rx_bufsize = 0;
1050 dev_info->rx_offload_capa = DEV_RX_OFFLOAD_SCATTER;
1051 dev_info->tx_offload_capa = 0;
1052 dev_info->rx_queue_offload_capa = 0;
1053 dev_info->tx_queue_offload_capa = 0;
1054 dev_info->speed_capa = ETH_LINK_SPEED_100G;
1058 eth_stats_get(struct rte_eth_dev *dev,
1059 struct rte_eth_stats *stats)
1062 uint16_t nb_rx = dev->data->nb_rx_queues;
1063 uint16_t nb_tx = dev->data->nb_tx_queues;
1064 uint64_t rx_total = 0;
1065 uint64_t tx_total = 0;
1066 uint64_t tx_err_total = 0;
1067 uint64_t rx_total_bytes = 0;
1068 uint64_t tx_total_bytes = 0;
1070 for (i = 0; i < nb_rx; i++) {
1071 struct szedata2_rx_queue *rxq = dev->data->rx_queues[i];
1073 if (i < RTE_ETHDEV_QUEUE_STAT_CNTRS) {
1074 stats->q_ipackets[i] = rxq->rx_pkts;
1075 stats->q_ibytes[i] = rxq->rx_bytes;
1077 rx_total += rxq->rx_pkts;
1078 rx_total_bytes += rxq->rx_bytes;
1081 for (i = 0; i < nb_tx; i++) {
1082 struct szedata2_tx_queue *txq = dev->data->tx_queues[i];
1084 if (i < RTE_ETHDEV_QUEUE_STAT_CNTRS) {
1085 stats->q_opackets[i] = txq->tx_pkts;
1086 stats->q_obytes[i] = txq->tx_bytes;
1087 stats->q_errors[i] = txq->err_pkts;
1089 tx_total += txq->tx_pkts;
1090 tx_total_bytes += txq->tx_bytes;
1091 tx_err_total += txq->err_pkts;
1094 stats->ipackets = rx_total;
1095 stats->opackets = tx_total;
1096 stats->ibytes = rx_total_bytes;
1097 stats->obytes = tx_total_bytes;
1098 stats->oerrors = tx_err_total;
1099 stats->rx_nombuf = dev->data->rx_mbuf_alloc_failed;
1105 eth_stats_reset(struct rte_eth_dev *dev)
1108 uint16_t nb_rx = dev->data->nb_rx_queues;
1109 uint16_t nb_tx = dev->data->nb_tx_queues;
1111 for (i = 0; i < nb_rx; i++) {
1112 struct szedata2_rx_queue *rxq = dev->data->rx_queues[i];
1117 for (i = 0; i < nb_tx; i++) {
1118 struct szedata2_tx_queue *txq = dev->data->tx_queues[i];
1126 eth_rx_queue_release(void *q)
1128 struct szedata2_rx_queue *rxq = (struct szedata2_rx_queue *)q;
1131 if (rxq->sze != NULL)
1132 szedata_close(rxq->sze);
1138 eth_tx_queue_release(void *q)
1140 struct szedata2_tx_queue *txq = (struct szedata2_tx_queue *)q;
1143 if (txq->sze != NULL)
1144 szedata_close(txq->sze);
1150 eth_dev_close(struct rte_eth_dev *dev)
1153 uint16_t nb_rx = dev->data->nb_rx_queues;
1154 uint16_t nb_tx = dev->data->nb_tx_queues;
1158 for (i = 0; i < nb_rx; i++) {
1159 eth_rx_queue_release(dev->data->rx_queues[i]);
1160 dev->data->rx_queues[i] = NULL;
1162 dev->data->nb_rx_queues = 0;
1163 for (i = 0; i < nb_tx; i++) {
1164 eth_tx_queue_release(dev->data->tx_queues[i]);
1165 dev->data->tx_queues[i] = NULL;
1167 dev->data->nb_tx_queues = 0;
1171 * Function takes value from first IBUF status register.
1172 * Values in IBUF and OBUF should be same.
1175 * Pointer to device private structure.
1177 * Link speed constant.
1179 static inline enum szedata2_link_speed
1180 get_link_speed(const struct pmd_internals *internals)
1182 const volatile struct szedata2_ibuf *ibuf =
1183 ibuf_ptr_by_index(internals->pci_rsc, 0);
1184 uint32_t speed = (szedata2_read32(&ibuf->ibuf_st) & 0x70) >> 4;
1187 return SZEDATA2_LINK_SPEED_10G;
1189 return SZEDATA2_LINK_SPEED_40G;
1191 return SZEDATA2_LINK_SPEED_100G;
1193 return SZEDATA2_LINK_SPEED_DEFAULT;
1198 eth_link_update(struct rte_eth_dev *dev,
1199 int wait_to_complete __rte_unused)
1201 struct rte_eth_link link;
1202 struct pmd_internals *internals = (struct pmd_internals *)
1203 dev->data->dev_private;
1204 const volatile struct szedata2_ibuf *ibuf;
1206 bool link_is_up = false;
1208 memset(&link, 0, sizeof(link));
1210 switch (get_link_speed(internals)) {
1211 case SZEDATA2_LINK_SPEED_10G:
1212 link.link_speed = ETH_SPEED_NUM_10G;
1214 case SZEDATA2_LINK_SPEED_40G:
1215 link.link_speed = ETH_SPEED_NUM_40G;
1217 case SZEDATA2_LINK_SPEED_100G:
1218 link.link_speed = ETH_SPEED_NUM_100G;
1221 link.link_speed = ETH_SPEED_NUM_10G;
1225 /* szedata2 uses only full duplex */
1226 link.link_duplex = ETH_LINK_FULL_DUPLEX;
1228 for (i = 0; i < szedata2_ibuf_count; i++) {
1229 ibuf = ibuf_ptr_by_index(internals->pci_rsc, i);
1231 * Link is considered up if at least one ibuf is enabled
1234 if (ibuf_is_enabled(ibuf) && ibuf_is_link_up(ibuf)) {
1240 link.link_status = link_is_up ? ETH_LINK_UP : ETH_LINK_DOWN;
1242 link.link_autoneg = ETH_LINK_FIXED;
1244 rte_eth_linkstatus_set(dev, &link);
1249 eth_dev_set_link_up(struct rte_eth_dev *dev)
1251 struct pmd_internals *internals = (struct pmd_internals *)
1252 dev->data->dev_private;
1255 for (i = 0; i < szedata2_ibuf_count; i++)
1256 ibuf_enable(ibuf_ptr_by_index(internals->pci_rsc, i));
1257 for (i = 0; i < szedata2_obuf_count; i++)
1258 obuf_enable(obuf_ptr_by_index(internals->pci_rsc, i));
1263 eth_dev_set_link_down(struct rte_eth_dev *dev)
1265 struct pmd_internals *internals = (struct pmd_internals *)
1266 dev->data->dev_private;
1269 for (i = 0; i < szedata2_ibuf_count; i++)
1270 ibuf_disable(ibuf_ptr_by_index(internals->pci_rsc, i));
1271 for (i = 0; i < szedata2_obuf_count; i++)
1272 obuf_disable(obuf_ptr_by_index(internals->pci_rsc, i));
1277 eth_rx_queue_setup(struct rte_eth_dev *dev,
1278 uint16_t rx_queue_id,
1279 uint16_t nb_rx_desc __rte_unused,
1280 unsigned int socket_id,
1281 const struct rte_eth_rxconf *rx_conf __rte_unused,
1282 struct rte_mempool *mb_pool)
1284 struct pmd_internals *internals = dev->data->dev_private;
1285 struct szedata2_rx_queue *rxq;
1287 uint32_t rx = 1 << rx_queue_id;
1290 if (dev->data->rx_queues[rx_queue_id] != NULL) {
1291 eth_rx_queue_release(dev->data->rx_queues[rx_queue_id]);
1292 dev->data->rx_queues[rx_queue_id] = NULL;
1295 rxq = rte_zmalloc_socket("szedata2 rx queue",
1296 sizeof(struct szedata2_rx_queue),
1297 RTE_CACHE_LINE_SIZE, socket_id);
1299 RTE_LOG(ERR, PMD, "rte_zmalloc_socket() failed for rx queue id "
1300 "%" PRIu16 "!\n", rx_queue_id);
1304 rxq->priv = internals;
1305 rxq->sze = szedata_open(internals->sze_dev);
1306 if (rxq->sze == NULL) {
1307 RTE_LOG(ERR, PMD, "szedata_open() failed for rx queue id "
1308 "%" PRIu16 "!\n", rx_queue_id);
1309 eth_rx_queue_release(rxq);
1312 ret = szedata_subscribe3(rxq->sze, &rx, &tx);
1313 if (ret != 0 || rx == 0) {
1314 RTE_LOG(ERR, PMD, "szedata_subscribe3() failed for rx queue id "
1315 "%" PRIu16 "!\n", rx_queue_id);
1316 eth_rx_queue_release(rxq);
1319 rxq->rx_channel = rx_queue_id;
1320 rxq->in_port = dev->data->port_id;
1321 rxq->mb_pool = mb_pool;
1326 dev->data->rx_queues[rx_queue_id] = rxq;
1328 RTE_LOG(DEBUG, PMD, "Configured rx queue id %" PRIu16 " on socket "
1329 "%u.\n", rx_queue_id, socket_id);
1335 eth_tx_queue_setup(struct rte_eth_dev *dev,
1336 uint16_t tx_queue_id,
1337 uint16_t nb_tx_desc __rte_unused,
1338 unsigned int socket_id,
1339 const struct rte_eth_txconf *tx_conf __rte_unused)
1341 struct pmd_internals *internals = dev->data->dev_private;
1342 struct szedata2_tx_queue *txq;
1345 uint32_t tx = 1 << tx_queue_id;
1347 if (dev->data->tx_queues[tx_queue_id] != NULL) {
1348 eth_tx_queue_release(dev->data->tx_queues[tx_queue_id]);
1349 dev->data->tx_queues[tx_queue_id] = NULL;
1352 txq = rte_zmalloc_socket("szedata2 tx queue",
1353 sizeof(struct szedata2_tx_queue),
1354 RTE_CACHE_LINE_SIZE, socket_id);
1356 RTE_LOG(ERR, PMD, "rte_zmalloc_socket() failed for tx queue id "
1357 "%" PRIu16 "!\n", tx_queue_id);
1361 txq->priv = internals;
1362 txq->sze = szedata_open(internals->sze_dev);
1363 if (txq->sze == NULL) {
1364 RTE_LOG(ERR, PMD, "szedata_open() failed for tx queue id "
1365 "%" PRIu16 "!\n", tx_queue_id);
1366 eth_tx_queue_release(txq);
1369 ret = szedata_subscribe3(txq->sze, &rx, &tx);
1370 if (ret != 0 || tx == 0) {
1371 RTE_LOG(ERR, PMD, "szedata_subscribe3() failed for tx queue id "
1372 "%" PRIu16 "!\n", tx_queue_id);
1373 eth_tx_queue_release(txq);
1376 txq->tx_channel = tx_queue_id;
1381 dev->data->tx_queues[tx_queue_id] = txq;
1383 RTE_LOG(DEBUG, PMD, "Configured tx queue id %" PRIu16 " on socket "
1384 "%u.\n", tx_queue_id, socket_id);
1390 eth_mac_addr_set(struct rte_eth_dev *dev __rte_unused,
1391 struct ether_addr *mac_addr __rte_unused)
1396 eth_promiscuous_enable(struct rte_eth_dev *dev)
1398 struct pmd_internals *internals = (struct pmd_internals *)
1399 dev->data->dev_private;
1402 for (i = 0; i < szedata2_ibuf_count; i++) {
1403 ibuf_mac_mode_write(ibuf_ptr_by_index(internals->pci_rsc, i),
1404 SZEDATA2_MAC_CHMODE_PROMISC);
1409 eth_promiscuous_disable(struct rte_eth_dev *dev)
1411 struct pmd_internals *internals = (struct pmd_internals *)
1412 dev->data->dev_private;
1415 for (i = 0; i < szedata2_ibuf_count; i++) {
1416 ibuf_mac_mode_write(ibuf_ptr_by_index(internals->pci_rsc, i),
1417 SZEDATA2_MAC_CHMODE_ONLY_VALID);
1422 eth_allmulticast_enable(struct rte_eth_dev *dev)
1424 struct pmd_internals *internals = (struct pmd_internals *)
1425 dev->data->dev_private;
1428 for (i = 0; i < szedata2_ibuf_count; i++) {
1429 ibuf_mac_mode_write(ibuf_ptr_by_index(internals->pci_rsc, i),
1430 SZEDATA2_MAC_CHMODE_ALL_MULTICAST);
1435 eth_allmulticast_disable(struct rte_eth_dev *dev)
1437 struct pmd_internals *internals = (struct pmd_internals *)
1438 dev->data->dev_private;
1441 for (i = 0; i < szedata2_ibuf_count; i++) {
1442 ibuf_mac_mode_write(ibuf_ptr_by_index(internals->pci_rsc, i),
1443 SZEDATA2_MAC_CHMODE_ONLY_VALID);
1447 static const struct eth_dev_ops ops = {
1448 .dev_start = eth_dev_start,
1449 .dev_stop = eth_dev_stop,
1450 .dev_set_link_up = eth_dev_set_link_up,
1451 .dev_set_link_down = eth_dev_set_link_down,
1452 .dev_close = eth_dev_close,
1453 .dev_configure = eth_dev_configure,
1454 .dev_infos_get = eth_dev_info,
1455 .promiscuous_enable = eth_promiscuous_enable,
1456 .promiscuous_disable = eth_promiscuous_disable,
1457 .allmulticast_enable = eth_allmulticast_enable,
1458 .allmulticast_disable = eth_allmulticast_disable,
1459 .rx_queue_start = eth_rx_queue_start,
1460 .rx_queue_stop = eth_rx_queue_stop,
1461 .tx_queue_start = eth_tx_queue_start,
1462 .tx_queue_stop = eth_tx_queue_stop,
1463 .rx_queue_setup = eth_rx_queue_setup,
1464 .tx_queue_setup = eth_tx_queue_setup,
1465 .rx_queue_release = eth_rx_queue_release,
1466 .tx_queue_release = eth_tx_queue_release,
1467 .link_update = eth_link_update,
1468 .stats_get = eth_stats_get,
1469 .stats_reset = eth_stats_reset,
1470 .mac_addr_set = eth_mac_addr_set,
1474 * This function goes through sysfs and looks for an index of szedata2
1475 * device file (/dev/szedataIIX, where X is the index).
1482 get_szedata2_index(const struct rte_pci_addr *pcislot_addr, uint32_t *index)
1485 struct dirent *entry;
1489 char pcislot_path[PATH_MAX];
1495 dir = opendir("/sys/class/combo");
1500 * Iterate through all combosixX directories.
1501 * When the value in /sys/class/combo/combosixX/device/pcislot
1502 * file is the location of the ethernet device dev, "X" is the
1503 * index of the device.
1505 while ((entry = readdir(dir)) != NULL) {
1506 ret = sscanf(entry->d_name, "combosix%u", &tmp_index);
1510 snprintf(pcislot_path, PATH_MAX,
1511 "/sys/class/combo/combosix%u/device/pcislot",
1514 fd = fopen(pcislot_path, "r");
1518 ret = fscanf(fd, "%4" PRIx16 ":%2" PRIx8 ":%2" PRIx8 ".%" PRIx8,
1519 &domain, &bus, &devid, &function);
1524 if (pcislot_addr->domain == domain &&
1525 pcislot_addr->bus == bus &&
1526 pcislot_addr->devid == devid &&
1527 pcislot_addr->function == function) {
1539 rte_szedata2_eth_dev_init(struct rte_eth_dev *dev)
1541 struct rte_eth_dev_data *data = dev->data;
1542 struct pmd_internals *internals = (struct pmd_internals *)
1544 struct szedata *szedata_temp;
1546 uint32_t szedata2_index;
1547 struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(dev);
1548 struct rte_pci_addr *pci_addr = &pci_dev->addr;
1549 struct rte_mem_resource *pci_rsc =
1550 &pci_dev->mem_resource[PCI_RESOURCE_NUMBER];
1551 char rsc_filename[PATH_MAX];
1552 void *pci_resource_ptr = NULL;
1555 RTE_LOG(INFO, PMD, "Initializing szedata2 device (" PCI_PRI_FMT ")\n",
1556 pci_addr->domain, pci_addr->bus, pci_addr->devid,
1557 pci_addr->function);
1559 internals->dev = dev;
1561 /* Get index of szedata2 device file and create path to device file */
1562 ret = get_szedata2_index(pci_addr, &szedata2_index);
1564 RTE_LOG(ERR, PMD, "Failed to get szedata2 device index!\n");
1567 snprintf(internals->sze_dev, PATH_MAX, SZEDATA2_DEV_PATH_FMT,
1570 RTE_LOG(INFO, PMD, "SZEDATA2 path: %s\n", internals->sze_dev);
1573 * Get number of available DMA RX and TX channels, which is maximum
1574 * number of queues that can be created and store it in private device
1577 szedata_temp = szedata_open(internals->sze_dev);
1578 if (szedata_temp == NULL) {
1579 RTE_LOG(ERR, PMD, "szedata_open(): failed to open %s",
1580 internals->sze_dev);
1583 internals->max_rx_queues = szedata_ifaces_available(szedata_temp,
1585 internals->max_tx_queues = szedata_ifaces_available(szedata_temp,
1587 szedata_close(szedata_temp);
1589 RTE_LOG(INFO, PMD, "Available DMA channels RX: %u TX: %u\n",
1590 internals->max_rx_queues, internals->max_tx_queues);
1592 /* Set rx, tx burst functions */
1593 if (data->scattered_rx == 1)
1594 dev->rx_pkt_burst = eth_szedata2_rx_scattered;
1596 dev->rx_pkt_burst = eth_szedata2_rx;
1597 dev->tx_pkt_burst = eth_szedata2_tx;
1599 /* Set function callbacks for Ethernet API */
1600 dev->dev_ops = &ops;
1602 rte_eth_copy_pci_info(dev, pci_dev);
1604 /* mmap pci resource0 file to rte_mem_resource structure */
1605 if (pci_dev->mem_resource[PCI_RESOURCE_NUMBER].phys_addr ==
1607 RTE_LOG(ERR, PMD, "Missing resource%u file\n",
1608 PCI_RESOURCE_NUMBER);
1611 snprintf(rsc_filename, PATH_MAX,
1612 "%s/" PCI_PRI_FMT "/resource%u", rte_pci_get_sysfs_path(),
1613 pci_addr->domain, pci_addr->bus,
1614 pci_addr->devid, pci_addr->function, PCI_RESOURCE_NUMBER);
1615 fd = open(rsc_filename, O_RDWR);
1617 RTE_LOG(ERR, PMD, "Could not open file %s\n", rsc_filename);
1621 pci_resource_ptr = mmap(0,
1622 pci_dev->mem_resource[PCI_RESOURCE_NUMBER].len,
1623 PROT_READ | PROT_WRITE, MAP_SHARED, fd, 0);
1625 if (pci_resource_ptr == MAP_FAILED) {
1626 RTE_LOG(ERR, PMD, "Could not mmap file %s (fd = %d)\n",
1630 pci_dev->mem_resource[PCI_RESOURCE_NUMBER].addr = pci_resource_ptr;
1631 internals->pci_rsc = pci_rsc;
1633 RTE_LOG(DEBUG, PMD, "resource%u phys_addr = 0x%llx len = %llu "
1634 "virt addr = %llx\n", PCI_RESOURCE_NUMBER,
1635 (unsigned long long)pci_rsc->phys_addr,
1636 (unsigned long long)pci_rsc->len,
1637 (unsigned long long)pci_rsc->addr);
1639 /* Get link state */
1640 eth_link_update(dev, 0);
1642 /* Allocate space for one mac address */
1643 data->mac_addrs = rte_zmalloc(data->name, sizeof(struct ether_addr),
1644 RTE_CACHE_LINE_SIZE);
1645 if (data->mac_addrs == NULL) {
1646 RTE_LOG(ERR, PMD, "Could not alloc space for MAC address!\n");
1647 munmap(pci_dev->mem_resource[PCI_RESOURCE_NUMBER].addr,
1648 pci_dev->mem_resource[PCI_RESOURCE_NUMBER].len);
1652 ether_addr_copy(ð_addr, data->mac_addrs);
1654 /* At initial state COMBO card is in promiscuous mode so disable it */
1655 eth_promiscuous_disable(dev);
1657 RTE_LOG(INFO, PMD, "szedata2 device ("
1658 PCI_PRI_FMT ") successfully initialized\n",
1659 pci_addr->domain, pci_addr->bus, pci_addr->devid,
1660 pci_addr->function);
1666 rte_szedata2_eth_dev_uninit(struct rte_eth_dev *dev)
1668 struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(dev);
1669 struct rte_pci_addr *pci_addr = &pci_dev->addr;
1671 rte_free(dev->data->mac_addrs);
1672 dev->data->mac_addrs = NULL;
1673 munmap(pci_dev->mem_resource[PCI_RESOURCE_NUMBER].addr,
1674 pci_dev->mem_resource[PCI_RESOURCE_NUMBER].len);
1676 RTE_LOG(INFO, PMD, "szedata2 device ("
1677 PCI_PRI_FMT ") successfully uninitialized\n",
1678 pci_addr->domain, pci_addr->bus, pci_addr->devid,
1679 pci_addr->function);
1684 static const struct rte_pci_id rte_szedata2_pci_id_table[] = {
1686 RTE_PCI_DEVICE(PCI_VENDOR_ID_NETCOPE,
1687 PCI_DEVICE_ID_NETCOPE_COMBO80G)
1690 RTE_PCI_DEVICE(PCI_VENDOR_ID_NETCOPE,
1691 PCI_DEVICE_ID_NETCOPE_COMBO100G)
1694 RTE_PCI_DEVICE(PCI_VENDOR_ID_NETCOPE,
1695 PCI_DEVICE_ID_NETCOPE_COMBO100G2)
1702 static int szedata2_eth_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
1703 struct rte_pci_device *pci_dev)
1705 return rte_eth_dev_pci_generic_probe(pci_dev,
1706 sizeof(struct pmd_internals), rte_szedata2_eth_dev_init);
1709 static int szedata2_eth_pci_remove(struct rte_pci_device *pci_dev)
1711 return rte_eth_dev_pci_generic_remove(pci_dev,
1712 rte_szedata2_eth_dev_uninit);
1715 static struct rte_pci_driver szedata2_eth_driver = {
1716 .id_table = rte_szedata2_pci_id_table,
1717 .probe = szedata2_eth_pci_probe,
1718 .remove = szedata2_eth_pci_remove,
1721 RTE_PMD_REGISTER_PCI(RTE_SZEDATA2_DRIVER_NAME, szedata2_eth_driver);
1722 RTE_PMD_REGISTER_PCI_TABLE(RTE_SZEDATA2_DRIVER_NAME, rte_szedata2_pci_id_table);
1723 RTE_PMD_REGISTER_KMOD_DEP(RTE_SZEDATA2_DRIVER_NAME,
1724 "* combo6core & combov3 & szedata2 & szedata2_cv3");