net/sfc: do not enforce hash offload in RSS multi-queue
[dpdk.git] / drivers / net / bonding / rte_eth_bond_api.c
1 /* SPDX-License-Identifier: BSD-3-Clause
2  * Copyright(c) 2010-2017 Intel Corporation
3  */
4
5 #include <string.h>
6
7 #include <rte_mbuf.h>
8 #include <rte_malloc.h>
9 #include <rte_ethdev_driver.h>
10 #include <rte_tcp.h>
11 #include <rte_bus_vdev.h>
12 #include <rte_kvargs.h>
13
14 #include "rte_eth_bond.h"
15 #include "eth_bond_private.h"
16 #include "eth_bond_8023ad_private.h"
17
18 int
19 check_for_bonded_ethdev(const struct rte_eth_dev *eth_dev)
20 {
21         /* Check valid pointer */
22         if (eth_dev == NULL ||
23                 eth_dev->device == NULL ||
24                 eth_dev->device->driver == NULL ||
25                 eth_dev->device->driver->name == NULL)
26                 return -1;
27
28         /* return 0 if driver name matches */
29         return eth_dev->device->driver->name != pmd_bond_drv.driver.name;
30 }
31
32 int
33 valid_bonded_port_id(uint16_t port_id)
34 {
35         RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -1);
36         return check_for_bonded_ethdev(&rte_eth_devices[port_id]);
37 }
38
39 int
40 check_for_master_bonded_ethdev(const struct rte_eth_dev *eth_dev)
41 {
42         int i;
43         struct bond_dev_private *internals;
44
45         if (check_for_bonded_ethdev(eth_dev) != 0)
46                 return 0;
47
48         internals = eth_dev->data->dev_private;
49
50         /* Check if any of slave devices is a bonded device */
51         for (i = 0; i < internals->slave_count; i++)
52                 if (valid_bonded_port_id(internals->slaves[i].port_id) == 0)
53                         return 1;
54
55         return 0;
56 }
57
58 int
59 valid_slave_port_id(uint16_t port_id, uint8_t mode)
60 {
61         RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -1);
62
63         /* Verify that port_id refers to a non bonded port */
64         if (check_for_bonded_ethdev(&rte_eth_devices[port_id]) == 0 &&
65                         mode == BONDING_MODE_8023AD) {
66                 RTE_BOND_LOG(ERR, "Cannot add slave to bonded device in 802.3ad"
67                                 " mode as slave is also a bonded device, only "
68                                 "physical devices can be support in this mode.");
69                 return -1;
70         }
71
72         return 0;
73 }
74
75 void
76 activate_slave(struct rte_eth_dev *eth_dev, uint16_t port_id)
77 {
78         struct bond_dev_private *internals = eth_dev->data->dev_private;
79         uint16_t active_count = internals->active_slave_count;
80
81         if (internals->mode == BONDING_MODE_8023AD)
82                 bond_mode_8023ad_activate_slave(eth_dev, port_id);
83
84         if (internals->mode == BONDING_MODE_TLB
85                         || internals->mode == BONDING_MODE_ALB) {
86
87                 internals->tlb_slaves_order[active_count] = port_id;
88         }
89
90         RTE_ASSERT(internals->active_slave_count <
91                         (RTE_DIM(internals->active_slaves) - 1));
92
93         internals->active_slaves[internals->active_slave_count] = port_id;
94         internals->active_slave_count++;
95
96         if (internals->mode == BONDING_MODE_TLB)
97                 bond_tlb_activate_slave(internals);
98         if (internals->mode == BONDING_MODE_ALB)
99                 bond_mode_alb_client_list_upd(eth_dev);
100 }
101
102 void
103 deactivate_slave(struct rte_eth_dev *eth_dev, uint16_t port_id)
104 {
105         uint16_t slave_pos;
106         struct bond_dev_private *internals = eth_dev->data->dev_private;
107         uint16_t active_count = internals->active_slave_count;
108
109         if (internals->mode == BONDING_MODE_8023AD) {
110                 bond_mode_8023ad_stop(eth_dev);
111                 bond_mode_8023ad_deactivate_slave(eth_dev, port_id);
112         } else if (internals->mode == BONDING_MODE_TLB
113                         || internals->mode == BONDING_MODE_ALB)
114                 bond_tlb_disable(internals);
115
116         slave_pos = find_slave_by_id(internals->active_slaves, active_count,
117                         port_id);
118
119         /* If slave was not at the end of the list
120          * shift active slaves up active array list */
121         if (slave_pos < active_count) {
122                 active_count--;
123                 memmove(internals->active_slaves + slave_pos,
124                                 internals->active_slaves + slave_pos + 1,
125                                 (active_count - slave_pos) *
126                                         sizeof(internals->active_slaves[0]));
127         }
128
129         RTE_ASSERT(active_count < RTE_DIM(internals->active_slaves));
130         internals->active_slave_count = active_count;
131
132         /* Resetting active_slave when reaches to max
133          * no of slaves in active list
134          */
135         if (internals->active_slave >= active_count)
136                 internals->active_slave = 0;
137
138         if (eth_dev->data->dev_started) {
139                 if (internals->mode == BONDING_MODE_8023AD) {
140                         bond_mode_8023ad_start(eth_dev);
141                 } else if (internals->mode == BONDING_MODE_TLB) {
142                         bond_tlb_enable(internals);
143                 } else if (internals->mode == BONDING_MODE_ALB) {
144                         bond_tlb_enable(internals);
145                         bond_mode_alb_client_list_upd(eth_dev);
146                 }
147         }
148 }
149
150 int
151 rte_eth_bond_create(const char *name, uint8_t mode, uint8_t socket_id)
152 {
153         struct bond_dev_private *internals;
154         char devargs[52];
155         uint16_t port_id;
156         int ret;
157
158         if (name == NULL) {
159                 RTE_BOND_LOG(ERR, "Invalid name specified");
160                 return -EINVAL;
161         }
162
163         ret = snprintf(devargs, sizeof(devargs),
164                 "driver=net_bonding,mode=%d,socket_id=%d", mode, socket_id);
165         if (ret < 0 || ret >= (int)sizeof(devargs))
166                 return -ENOMEM;
167
168         ret = rte_vdev_init(name, devargs);
169         if (ret)
170                 return ret;
171
172         ret = rte_eth_dev_get_port_by_name(name, &port_id);
173         RTE_ASSERT(!ret);
174
175         /*
176          * To make bond_ethdev_configure() happy we need to free the
177          * internals->kvlist here.
178          *
179          * Also see comment in bond_ethdev_configure().
180          */
181         internals = rte_eth_devices[port_id].data->dev_private;
182         rte_kvargs_free(internals->kvlist);
183         internals->kvlist = NULL;
184
185         return port_id;
186 }
187
188 int
189 rte_eth_bond_free(const char *name)
190 {
191         return rte_vdev_uninit(name);
192 }
193
194 static int
195 slave_vlan_filter_set(uint16_t bonded_port_id, uint16_t slave_port_id)
196 {
197         struct rte_eth_dev *bonded_eth_dev;
198         struct bond_dev_private *internals;
199         int found;
200         int res = 0;
201         uint64_t slab = 0;
202         uint32_t pos = 0;
203         uint16_t first;
204
205         bonded_eth_dev = &rte_eth_devices[bonded_port_id];
206         if ((bonded_eth_dev->data->dev_conf.rxmode.offloads &
207                         DEV_RX_OFFLOAD_VLAN_FILTER) == 0)
208                 return 0;
209
210         internals = bonded_eth_dev->data->dev_private;
211         found = rte_bitmap_scan(internals->vlan_filter_bmp, &pos, &slab);
212         first = pos;
213
214         if (!found)
215                 return 0;
216
217         do {
218                 uint32_t i;
219                 uint64_t mask;
220
221                 for (i = 0, mask = 1;
222                      i < RTE_BITMAP_SLAB_BIT_SIZE;
223                      i ++, mask <<= 1) {
224                         if (unlikely(slab & mask)) {
225                                 uint16_t vlan_id = pos + i;
226
227                                 res = rte_eth_dev_vlan_filter(slave_port_id,
228                                                               vlan_id, 1);
229                         }
230                 }
231                 found = rte_bitmap_scan(internals->vlan_filter_bmp,
232                                         &pos, &slab);
233         } while (found && first != pos && res == 0);
234
235         return res;
236 }
237
238 static int
239 slave_rte_flow_prepare(uint16_t slave_id, struct bond_dev_private *internals)
240 {
241         struct rte_flow *flow;
242         struct rte_flow_error ferror;
243         uint16_t slave_port_id = internals->slaves[slave_id].port_id;
244
245         if (internals->flow_isolated_valid != 0) {
246                 rte_eth_dev_stop(slave_port_id);
247                 if (rte_flow_isolate(slave_port_id, internals->flow_isolated,
248                     &ferror)) {
249                         RTE_BOND_LOG(ERR, "rte_flow_isolate failed for slave"
250                                      " %d: %s", slave_id, ferror.message ?
251                                      ferror.message : "(no stated reason)");
252                         return -1;
253                 }
254         }
255         TAILQ_FOREACH(flow, &internals->flow_list, next) {
256                 flow->flows[slave_id] = rte_flow_create(slave_port_id,
257                                                         flow->rule.attr,
258                                                         flow->rule.pattern,
259                                                         flow->rule.actions,
260                                                         &ferror);
261                 if (flow->flows[slave_id] == NULL) {
262                         RTE_BOND_LOG(ERR, "Cannot create flow for slave"
263                                      " %d: %s", slave_id,
264                                      ferror.message ? ferror.message :
265                                      "(no stated reason)");
266                         /* Destroy successful bond flows from the slave */
267                         TAILQ_FOREACH(flow, &internals->flow_list, next) {
268                                 if (flow->flows[slave_id] != NULL) {
269                                         rte_flow_destroy(slave_port_id,
270                                                          flow->flows[slave_id],
271                                                          &ferror);
272                                         flow->flows[slave_id] = NULL;
273                                 }
274                         }
275                         return -1;
276                 }
277         }
278         return 0;
279 }
280
281 static void
282 eth_bond_slave_inherit_dev_info_rx_first(struct bond_dev_private *internals,
283                                          const struct rte_eth_dev_info *di)
284 {
285         struct rte_eth_rxconf *rxconf_i = &internals->default_rxconf;
286
287         internals->reta_size = di->reta_size;
288
289         /* Inherit Rx offload capabilities from the first slave device */
290         internals->rx_offload_capa = di->rx_offload_capa;
291         internals->rx_queue_offload_capa = di->rx_queue_offload_capa;
292         internals->flow_type_rss_offloads = di->flow_type_rss_offloads;
293
294         /* Inherit maximum Rx packet size from the first slave device */
295         internals->candidate_max_rx_pktlen = di->max_rx_pktlen;
296
297         /* Inherit default Rx queue settings from the first slave device */
298         memcpy(rxconf_i, &di->default_rxconf, sizeof(*rxconf_i));
299
300         /*
301          * Turn off descriptor prefetch and writeback by default for all
302          * slave devices. Applications may tweak this setting if need be.
303          */
304         rxconf_i->rx_thresh.pthresh = 0;
305         rxconf_i->rx_thresh.hthresh = 0;
306         rxconf_i->rx_thresh.wthresh = 0;
307
308         /* Setting this to zero should effectively enable default values */
309         rxconf_i->rx_free_thresh = 0;
310
311         /* Disable deferred start by default for all slave devices */
312         rxconf_i->rx_deferred_start = 0;
313 }
314
315 static void
316 eth_bond_slave_inherit_dev_info_tx_first(struct bond_dev_private *internals,
317                                          const struct rte_eth_dev_info *di)
318 {
319         struct rte_eth_txconf *txconf_i = &internals->default_txconf;
320
321         /* Inherit Tx offload capabilities from the first slave device */
322         internals->tx_offload_capa = di->tx_offload_capa;
323         internals->tx_queue_offload_capa = di->tx_queue_offload_capa;
324
325         /* Inherit default Tx queue settings from the first slave device */
326         memcpy(txconf_i, &di->default_txconf, sizeof(*txconf_i));
327
328         /*
329          * Turn off descriptor prefetch and writeback by default for all
330          * slave devices. Applications may tweak this setting if need be.
331          */
332         txconf_i->tx_thresh.pthresh = 0;
333         txconf_i->tx_thresh.hthresh = 0;
334         txconf_i->tx_thresh.wthresh = 0;
335
336         /*
337          * Setting these parameters to zero assumes that default
338          * values will be configured implicitly by slave devices.
339          */
340         txconf_i->tx_free_thresh = 0;
341         txconf_i->tx_rs_thresh = 0;
342
343         /* Disable deferred start by default for all slave devices */
344         txconf_i->tx_deferred_start = 0;
345 }
346
347 static void
348 eth_bond_slave_inherit_dev_info_rx_next(struct bond_dev_private *internals,
349                                         const struct rte_eth_dev_info *di)
350 {
351         struct rte_eth_rxconf *rxconf_i = &internals->default_rxconf;
352         const struct rte_eth_rxconf *rxconf = &di->default_rxconf;
353
354         internals->rx_offload_capa &= di->rx_offload_capa;
355         internals->rx_queue_offload_capa &= di->rx_queue_offload_capa;
356         internals->flow_type_rss_offloads &= di->flow_type_rss_offloads;
357
358         /*
359          * If at least one slave device suggests enabling this
360          * setting by default, enable it for all slave devices
361          * since disabling it may not be necessarily supported.
362          */
363         if (rxconf->rx_drop_en == 1)
364                 rxconf_i->rx_drop_en = 1;
365
366         /*
367          * Adding a new slave device may cause some of previously inherited
368          * offloads to be withdrawn from the internal rx_queue_offload_capa
369          * value. Thus, the new internal value of default Rx queue offloads
370          * has to be masked by rx_queue_offload_capa to make sure that only
371          * commonly supported offloads are preserved from both the previous
372          * value and the value being inhereted from the new slave device.
373          */
374         rxconf_i->offloads = (rxconf_i->offloads | rxconf->offloads) &
375                              internals->rx_queue_offload_capa;
376
377         /*
378          * RETA size is GCD of all slaves RETA sizes, so, if all sizes will be
379          * the power of 2, the lower one is GCD
380          */
381         if (internals->reta_size > di->reta_size)
382                 internals->reta_size = di->reta_size;
383
384         if (!internals->max_rx_pktlen &&
385             di->max_rx_pktlen < internals->candidate_max_rx_pktlen)
386                 internals->candidate_max_rx_pktlen = di->max_rx_pktlen;
387 }
388
389 static void
390 eth_bond_slave_inherit_dev_info_tx_next(struct bond_dev_private *internals,
391                                         const struct rte_eth_dev_info *di)
392 {
393         struct rte_eth_txconf *txconf_i = &internals->default_txconf;
394         const struct rte_eth_txconf *txconf = &di->default_txconf;
395
396         internals->tx_offload_capa &= di->tx_offload_capa;
397         internals->tx_queue_offload_capa &= di->tx_queue_offload_capa;
398
399         /*
400          * Adding a new slave device may cause some of previously inherited
401          * offloads to be withdrawn from the internal tx_queue_offload_capa
402          * value. Thus, the new internal value of default Tx queue offloads
403          * has to be masked by tx_queue_offload_capa to make sure that only
404          * commonly supported offloads are preserved from both the previous
405          * value and the value being inhereted from the new slave device.
406          */
407         txconf_i->offloads = (txconf_i->offloads | txconf->offloads) &
408                              internals->tx_queue_offload_capa;
409 }
410
411 static void
412 eth_bond_slave_inherit_desc_lim_first(struct rte_eth_desc_lim *bond_desc_lim,
413                 const struct rte_eth_desc_lim *slave_desc_lim)
414 {
415         memcpy(bond_desc_lim, slave_desc_lim, sizeof(*bond_desc_lim));
416 }
417
418 static int
419 eth_bond_slave_inherit_desc_lim_next(struct rte_eth_desc_lim *bond_desc_lim,
420                 const struct rte_eth_desc_lim *slave_desc_lim)
421 {
422         bond_desc_lim->nb_max = RTE_MIN(bond_desc_lim->nb_max,
423                                         slave_desc_lim->nb_max);
424         bond_desc_lim->nb_min = RTE_MAX(bond_desc_lim->nb_min,
425                                         slave_desc_lim->nb_min);
426         bond_desc_lim->nb_align = RTE_MAX(bond_desc_lim->nb_align,
427                                           slave_desc_lim->nb_align);
428
429         if (bond_desc_lim->nb_min > bond_desc_lim->nb_max ||
430             bond_desc_lim->nb_align > bond_desc_lim->nb_max) {
431                 RTE_BOND_LOG(ERR, "Failed to inherit descriptor limits");
432                 return -EINVAL;
433         }
434
435         /* Treat maximum number of segments equal to 0 as unspecified */
436         if (slave_desc_lim->nb_seg_max != 0 &&
437             (bond_desc_lim->nb_seg_max == 0 ||
438              slave_desc_lim->nb_seg_max < bond_desc_lim->nb_seg_max))
439                 bond_desc_lim->nb_seg_max = slave_desc_lim->nb_seg_max;
440         if (slave_desc_lim->nb_mtu_seg_max != 0 &&
441             (bond_desc_lim->nb_mtu_seg_max == 0 ||
442              slave_desc_lim->nb_mtu_seg_max < bond_desc_lim->nb_mtu_seg_max))
443                 bond_desc_lim->nb_mtu_seg_max = slave_desc_lim->nb_mtu_seg_max;
444
445         return 0;
446 }
447
448 static int
449 __eth_bond_slave_add_lock_free(uint16_t bonded_port_id, uint16_t slave_port_id)
450 {
451         struct rte_eth_dev *bonded_eth_dev, *slave_eth_dev;
452         struct bond_dev_private *internals;
453         struct rte_eth_link link_props;
454         struct rte_eth_dev_info dev_info;
455         int ret;
456
457         bonded_eth_dev = &rte_eth_devices[bonded_port_id];
458         internals = bonded_eth_dev->data->dev_private;
459
460         if (valid_slave_port_id(slave_port_id, internals->mode) != 0)
461                 return -1;
462
463         slave_eth_dev = &rte_eth_devices[slave_port_id];
464         if (slave_eth_dev->data->dev_flags & RTE_ETH_DEV_BONDED_SLAVE) {
465                 RTE_BOND_LOG(ERR, "Slave device is already a slave of a bonded device");
466                 return -1;
467         }
468
469         ret = rte_eth_dev_info_get(slave_port_id, &dev_info);
470         if (ret != 0) {
471                 RTE_BOND_LOG(ERR,
472                         "%s: Error during getting device (port %u) info: %s\n",
473                         __func__, slave_port_id, strerror(-ret));
474
475                 return ret;
476         }
477         if (dev_info.max_rx_pktlen < internals->max_rx_pktlen) {
478                 RTE_BOND_LOG(ERR, "Slave (port %u) max_rx_pktlen too small",
479                              slave_port_id);
480                 return -1;
481         }
482
483         slave_add(internals, slave_eth_dev);
484
485         /* We need to store slaves reta_size to be able to synchronize RETA for all
486          * slave devices even if its sizes are different.
487          */
488         internals->slaves[internals->slave_count].reta_size = dev_info.reta_size;
489
490         if (internals->slave_count < 1) {
491                 /* if MAC is not user defined then use MAC of first slave add to
492                  * bonded device */
493                 if (!internals->user_defined_mac) {
494                         if (mac_address_set(bonded_eth_dev,
495                                             slave_eth_dev->data->mac_addrs)) {
496                                 RTE_BOND_LOG(ERR, "Failed to set MAC address");
497                                 return -1;
498                         }
499                 }
500
501                 /* Make primary slave */
502                 internals->primary_port = slave_port_id;
503                 internals->current_primary_port = slave_port_id;
504
505                 /* Inherit queues settings from first slave */
506                 internals->nb_rx_queues = slave_eth_dev->data->nb_rx_queues;
507                 internals->nb_tx_queues = slave_eth_dev->data->nb_tx_queues;
508
509                 eth_bond_slave_inherit_dev_info_rx_first(internals, &dev_info);
510                 eth_bond_slave_inherit_dev_info_tx_first(internals, &dev_info);
511
512                 eth_bond_slave_inherit_desc_lim_first(&internals->rx_desc_lim,
513                                                       &dev_info.rx_desc_lim);
514                 eth_bond_slave_inherit_desc_lim_first(&internals->tx_desc_lim,
515                                                       &dev_info.tx_desc_lim);
516         } else {
517                 int ret;
518
519                 eth_bond_slave_inherit_dev_info_rx_next(internals, &dev_info);
520                 eth_bond_slave_inherit_dev_info_tx_next(internals, &dev_info);
521
522                 ret = eth_bond_slave_inherit_desc_lim_next(
523                                 &internals->rx_desc_lim, &dev_info.rx_desc_lim);
524                 if (ret != 0)
525                         return ret;
526
527                 ret = eth_bond_slave_inherit_desc_lim_next(
528                                 &internals->tx_desc_lim, &dev_info.tx_desc_lim);
529                 if (ret != 0)
530                         return ret;
531         }
532
533         bonded_eth_dev->data->dev_conf.rx_adv_conf.rss_conf.rss_hf &=
534                         internals->flow_type_rss_offloads;
535
536         if (slave_rte_flow_prepare(internals->slave_count, internals) != 0) {
537                 RTE_BOND_LOG(ERR, "Failed to prepare new slave flows: port=%d",
538                              slave_port_id);
539                 return -1;
540         }
541
542         /* Add additional MAC addresses to the slave */
543         if (slave_add_mac_addresses(bonded_eth_dev, slave_port_id) != 0) {
544                 RTE_BOND_LOG(ERR, "Failed to add mac address(es) to slave %hu",
545                                 slave_port_id);
546                 return -1;
547         }
548
549         internals->slave_count++;
550
551         if (bonded_eth_dev->data->dev_started) {
552                 if (slave_configure(bonded_eth_dev, slave_eth_dev) != 0) {
553                         internals->slave_count--;
554                         RTE_BOND_LOG(ERR, "rte_bond_slaves_configure: port=%d",
555                                         slave_port_id);
556                         return -1;
557                 }
558         }
559
560         /* Update all slave devices MACs */
561         mac_address_slaves_update(bonded_eth_dev);
562
563         /* Register link status change callback with bonded device pointer as
564          * argument*/
565         rte_eth_dev_callback_register(slave_port_id, RTE_ETH_EVENT_INTR_LSC,
566                         bond_ethdev_lsc_event_callback, &bonded_eth_dev->data->port_id);
567
568         /* If bonded device is started then we can add the slave to our active
569          * slave array */
570         if (bonded_eth_dev->data->dev_started) {
571                 ret = rte_eth_link_get_nowait(slave_port_id, &link_props);
572                 if (ret < 0) {
573                         rte_eth_dev_callback_unregister(slave_port_id,
574                                         RTE_ETH_EVENT_INTR_LSC,
575                                         bond_ethdev_lsc_event_callback,
576                                         &bonded_eth_dev->data->port_id);
577                         internals->slave_count--;
578                         RTE_BOND_LOG(ERR,
579                                 "Slave (port %u) link get failed: %s\n",
580                                 slave_port_id, rte_strerror(-ret));
581                         return -1;
582                 }
583
584                  if (link_props.link_status == ETH_LINK_UP) {
585                         if (internals->active_slave_count == 0 &&
586                             !internals->user_defined_primary_port)
587                                 bond_ethdev_primary_set(internals,
588                                                         slave_port_id);
589                 }
590         }
591
592         /* Add slave details to bonded device */
593         slave_eth_dev->data->dev_flags |= RTE_ETH_DEV_BONDED_SLAVE;
594
595         slave_vlan_filter_set(bonded_port_id, slave_port_id);
596
597         return 0;
598
599 }
600
601 int
602 rte_eth_bond_slave_add(uint16_t bonded_port_id, uint16_t slave_port_id)
603 {
604         struct rte_eth_dev *bonded_eth_dev;
605         struct bond_dev_private *internals;
606
607         int retval;
608
609         /* Verify that port id's are valid bonded and slave ports */
610         if (valid_bonded_port_id(bonded_port_id) != 0)
611                 return -1;
612
613         bonded_eth_dev = &rte_eth_devices[bonded_port_id];
614         internals = bonded_eth_dev->data->dev_private;
615
616         rte_spinlock_lock(&internals->lock);
617
618         retval = __eth_bond_slave_add_lock_free(bonded_port_id, slave_port_id);
619
620         rte_spinlock_unlock(&internals->lock);
621
622         return retval;
623 }
624
625 static int
626 __eth_bond_slave_remove_lock_free(uint16_t bonded_port_id,
627                                    uint16_t slave_port_id)
628 {
629         struct rte_eth_dev *bonded_eth_dev;
630         struct bond_dev_private *internals;
631         struct rte_eth_dev *slave_eth_dev;
632         struct rte_flow_error flow_error;
633         struct rte_flow *flow;
634         int i, slave_idx;
635
636         bonded_eth_dev = &rte_eth_devices[bonded_port_id];
637         internals = bonded_eth_dev->data->dev_private;
638
639         if (valid_slave_port_id(slave_port_id, internals->mode) < 0)
640                 return -1;
641
642         /* first remove from active slave list */
643         slave_idx = find_slave_by_id(internals->active_slaves,
644                 internals->active_slave_count, slave_port_id);
645
646         if (slave_idx < internals->active_slave_count)
647                 deactivate_slave(bonded_eth_dev, slave_port_id);
648
649         slave_idx = -1;
650         /* now find in slave list */
651         for (i = 0; i < internals->slave_count; i++)
652                 if (internals->slaves[i].port_id == slave_port_id) {
653                         slave_idx = i;
654                         break;
655                 }
656
657         if (slave_idx < 0) {
658                 RTE_BOND_LOG(ERR, "Couldn't find slave in port list, slave count %d",
659                                 internals->slave_count);
660                 return -1;
661         }
662
663         /* Un-register link status change callback with bonded device pointer as
664          * argument*/
665         rte_eth_dev_callback_unregister(slave_port_id, RTE_ETH_EVENT_INTR_LSC,
666                         bond_ethdev_lsc_event_callback,
667                         &rte_eth_devices[bonded_port_id].data->port_id);
668
669         /* Restore original MAC address of slave device */
670         rte_eth_dev_default_mac_addr_set(slave_port_id,
671                         &(internals->slaves[slave_idx].persisted_mac_addr));
672
673         /* remove additional MAC addresses from the slave */
674         slave_remove_mac_addresses(bonded_eth_dev, slave_port_id);
675
676         /*
677          * Remove bond device flows from slave device.
678          * Note: don't restore flow isolate mode.
679          */
680         TAILQ_FOREACH(flow, &internals->flow_list, next) {
681                 if (flow->flows[slave_idx] != NULL) {
682                         rte_flow_destroy(slave_port_id, flow->flows[slave_idx],
683                                          &flow_error);
684                         flow->flows[slave_idx] = NULL;
685                 }
686         }
687
688         slave_eth_dev = &rte_eth_devices[slave_port_id];
689         slave_remove(internals, slave_eth_dev);
690         slave_eth_dev->data->dev_flags &= (~RTE_ETH_DEV_BONDED_SLAVE);
691
692         /*  first slave in the active list will be the primary by default,
693          *  otherwise use first device in list */
694         if (internals->current_primary_port == slave_port_id) {
695                 if (internals->active_slave_count > 0)
696                         internals->current_primary_port = internals->active_slaves[0];
697                 else if (internals->slave_count > 0)
698                         internals->current_primary_port = internals->slaves[0].port_id;
699                 else
700                         internals->primary_port = 0;
701                 mac_address_slaves_update(bonded_eth_dev);
702         }
703
704         if (internals->active_slave_count < 1) {
705                 /* if no slaves are any longer attached to bonded device and MAC is not
706                  * user defined then clear MAC of bonded device as it will be reset
707                  * when a new slave is added */
708                 if (internals->slave_count < 1 && !internals->user_defined_mac)
709                         memset(rte_eth_devices[bonded_port_id].data->mac_addrs, 0,
710                                         sizeof(*(rte_eth_devices[bonded_port_id].data->mac_addrs)));
711         }
712         if (internals->slave_count == 0) {
713                 internals->rx_offload_capa = 0;
714                 internals->tx_offload_capa = 0;
715                 internals->rx_queue_offload_capa = 0;
716                 internals->tx_queue_offload_capa = 0;
717                 internals->flow_type_rss_offloads = ETH_RSS_PROTO_MASK;
718                 internals->reta_size = 0;
719                 internals->candidate_max_rx_pktlen = 0;
720                 internals->max_rx_pktlen = 0;
721         }
722         return 0;
723 }
724
725 int
726 rte_eth_bond_slave_remove(uint16_t bonded_port_id, uint16_t slave_port_id)
727 {
728         struct rte_eth_dev *bonded_eth_dev;
729         struct bond_dev_private *internals;
730         int retval;
731
732         if (valid_bonded_port_id(bonded_port_id) != 0)
733                 return -1;
734
735         bonded_eth_dev = &rte_eth_devices[bonded_port_id];
736         internals = bonded_eth_dev->data->dev_private;
737
738         rte_spinlock_lock(&internals->lock);
739
740         retval = __eth_bond_slave_remove_lock_free(bonded_port_id, slave_port_id);
741
742         rte_spinlock_unlock(&internals->lock);
743
744         return retval;
745 }
746
747 int
748 rte_eth_bond_mode_set(uint16_t bonded_port_id, uint8_t mode)
749 {
750         struct rte_eth_dev *bonded_eth_dev;
751
752         if (valid_bonded_port_id(bonded_port_id) != 0)
753                 return -1;
754
755         bonded_eth_dev = &rte_eth_devices[bonded_port_id];
756
757         if (check_for_master_bonded_ethdev(bonded_eth_dev) != 0 &&
758                         mode == BONDING_MODE_8023AD)
759                 return -1;
760
761         return bond_ethdev_mode_set(bonded_eth_dev, mode);
762 }
763
764 int
765 rte_eth_bond_mode_get(uint16_t bonded_port_id)
766 {
767         struct bond_dev_private *internals;
768
769         if (valid_bonded_port_id(bonded_port_id) != 0)
770                 return -1;
771
772         internals = rte_eth_devices[bonded_port_id].data->dev_private;
773
774         return internals->mode;
775 }
776
777 int
778 rte_eth_bond_primary_set(uint16_t bonded_port_id, uint16_t slave_port_id)
779 {
780         struct bond_dev_private *internals;
781
782         if (valid_bonded_port_id(bonded_port_id) != 0)
783                 return -1;
784
785         internals = rte_eth_devices[bonded_port_id].data->dev_private;
786
787         if (valid_slave_port_id(slave_port_id, internals->mode) != 0)
788                 return -1;
789
790         internals->user_defined_primary_port = 1;
791         internals->primary_port = slave_port_id;
792
793         bond_ethdev_primary_set(internals, slave_port_id);
794
795         return 0;
796 }
797
798 int
799 rte_eth_bond_primary_get(uint16_t bonded_port_id)
800 {
801         struct bond_dev_private *internals;
802
803         if (valid_bonded_port_id(bonded_port_id) != 0)
804                 return -1;
805
806         internals = rte_eth_devices[bonded_port_id].data->dev_private;
807
808         if (internals->slave_count < 1)
809                 return -1;
810
811         return internals->current_primary_port;
812 }
813
814 int
815 rte_eth_bond_slaves_get(uint16_t bonded_port_id, uint16_t slaves[],
816                         uint16_t len)
817 {
818         struct bond_dev_private *internals;
819         uint16_t i;
820
821         if (valid_bonded_port_id(bonded_port_id) != 0)
822                 return -1;
823
824         if (slaves == NULL)
825                 return -1;
826
827         internals = rte_eth_devices[bonded_port_id].data->dev_private;
828
829         if (internals->slave_count > len)
830                 return -1;
831
832         for (i = 0; i < internals->slave_count; i++)
833                 slaves[i] = internals->slaves[i].port_id;
834
835         return internals->slave_count;
836 }
837
838 int
839 rte_eth_bond_active_slaves_get(uint16_t bonded_port_id, uint16_t slaves[],
840                 uint16_t len)
841 {
842         struct bond_dev_private *internals;
843
844         if (valid_bonded_port_id(bonded_port_id) != 0)
845                 return -1;
846
847         if (slaves == NULL)
848                 return -1;
849
850         internals = rte_eth_devices[bonded_port_id].data->dev_private;
851
852         if (internals->active_slave_count > len)
853                 return -1;
854
855         memcpy(slaves, internals->active_slaves,
856         internals->active_slave_count * sizeof(internals->active_slaves[0]));
857
858         return internals->active_slave_count;
859 }
860
861 int
862 rte_eth_bond_mac_address_set(uint16_t bonded_port_id,
863                 struct rte_ether_addr *mac_addr)
864 {
865         struct rte_eth_dev *bonded_eth_dev;
866         struct bond_dev_private *internals;
867
868         if (valid_bonded_port_id(bonded_port_id) != 0)
869                 return -1;
870
871         bonded_eth_dev = &rte_eth_devices[bonded_port_id];
872         internals = bonded_eth_dev->data->dev_private;
873
874         /* Set MAC Address of Bonded Device */
875         if (mac_address_set(bonded_eth_dev, mac_addr))
876                 return -1;
877
878         internals->user_defined_mac = 1;
879
880         /* Update all slave devices MACs*/
881         if (internals->slave_count > 0)
882                 return mac_address_slaves_update(bonded_eth_dev);
883
884         return 0;
885 }
886
887 int
888 rte_eth_bond_mac_address_reset(uint16_t bonded_port_id)
889 {
890         struct rte_eth_dev *bonded_eth_dev;
891         struct bond_dev_private *internals;
892
893         if (valid_bonded_port_id(bonded_port_id) != 0)
894                 return -1;
895
896         bonded_eth_dev = &rte_eth_devices[bonded_port_id];
897         internals = bonded_eth_dev->data->dev_private;
898
899         internals->user_defined_mac = 0;
900
901         if (internals->slave_count > 0) {
902                 int slave_port;
903                 /* Get the primary slave location based on the primary port
904                  * number as, while slave_add(), we will keep the primary
905                  * slave based on slave_count,but not based on the primary port.
906                  */
907                 for (slave_port = 0; slave_port < internals->slave_count;
908                      slave_port++) {
909                         if (internals->slaves[slave_port].port_id ==
910                             internals->primary_port)
911                                 break;
912                 }
913
914                 /* Set MAC Address of Bonded Device */
915                 if (mac_address_set(bonded_eth_dev,
916                         &internals->slaves[slave_port].persisted_mac_addr)
917                                 != 0) {
918                         RTE_BOND_LOG(ERR, "Failed to set MAC address on bonded device");
919                         return -1;
920                 }
921                 /* Update all slave devices MAC addresses */
922                 return mac_address_slaves_update(bonded_eth_dev);
923         }
924         /* No need to update anything as no slaves present */
925         return 0;
926 }
927
928 int
929 rte_eth_bond_xmit_policy_set(uint16_t bonded_port_id, uint8_t policy)
930 {
931         struct bond_dev_private *internals;
932
933         if (valid_bonded_port_id(bonded_port_id) != 0)
934                 return -1;
935
936         internals = rte_eth_devices[bonded_port_id].data->dev_private;
937
938         switch (policy) {
939         case BALANCE_XMIT_POLICY_LAYER2:
940                 internals->balance_xmit_policy = policy;
941                 internals->burst_xmit_hash = burst_xmit_l2_hash;
942                 break;
943         case BALANCE_XMIT_POLICY_LAYER23:
944                 internals->balance_xmit_policy = policy;
945                 internals->burst_xmit_hash = burst_xmit_l23_hash;
946                 break;
947         case BALANCE_XMIT_POLICY_LAYER34:
948                 internals->balance_xmit_policy = policy;
949                 internals->burst_xmit_hash = burst_xmit_l34_hash;
950                 break;
951
952         default:
953                 return -1;
954         }
955         return 0;
956 }
957
958 int
959 rte_eth_bond_xmit_policy_get(uint16_t bonded_port_id)
960 {
961         struct bond_dev_private *internals;
962
963         if (valid_bonded_port_id(bonded_port_id) != 0)
964                 return -1;
965
966         internals = rte_eth_devices[bonded_port_id].data->dev_private;
967
968         return internals->balance_xmit_policy;
969 }
970
971 int
972 rte_eth_bond_link_monitoring_set(uint16_t bonded_port_id, uint32_t internal_ms)
973 {
974         struct bond_dev_private *internals;
975
976         if (valid_bonded_port_id(bonded_port_id) != 0)
977                 return -1;
978
979         internals = rte_eth_devices[bonded_port_id].data->dev_private;
980         internals->link_status_polling_interval_ms = internal_ms;
981
982         return 0;
983 }
984
985 int
986 rte_eth_bond_link_monitoring_get(uint16_t bonded_port_id)
987 {
988         struct bond_dev_private *internals;
989
990         if (valid_bonded_port_id(bonded_port_id) != 0)
991                 return -1;
992
993         internals = rte_eth_devices[bonded_port_id].data->dev_private;
994
995         return internals->link_status_polling_interval_ms;
996 }
997
998 int
999 rte_eth_bond_link_down_prop_delay_set(uint16_t bonded_port_id,
1000                                        uint32_t delay_ms)
1001
1002 {
1003         struct bond_dev_private *internals;
1004
1005         if (valid_bonded_port_id(bonded_port_id) != 0)
1006                 return -1;
1007
1008         internals = rte_eth_devices[bonded_port_id].data->dev_private;
1009         internals->link_down_delay_ms = delay_ms;
1010
1011         return 0;
1012 }
1013
1014 int
1015 rte_eth_bond_link_down_prop_delay_get(uint16_t bonded_port_id)
1016 {
1017         struct bond_dev_private *internals;
1018
1019         if (valid_bonded_port_id(bonded_port_id) != 0)
1020                 return -1;
1021
1022         internals = rte_eth_devices[bonded_port_id].data->dev_private;
1023
1024         return internals->link_down_delay_ms;
1025 }
1026
1027 int
1028 rte_eth_bond_link_up_prop_delay_set(uint16_t bonded_port_id, uint32_t delay_ms)
1029
1030 {
1031         struct bond_dev_private *internals;
1032
1033         if (valid_bonded_port_id(bonded_port_id) != 0)
1034                 return -1;
1035
1036         internals = rte_eth_devices[bonded_port_id].data->dev_private;
1037         internals->link_up_delay_ms = delay_ms;
1038
1039         return 0;
1040 }
1041
1042 int
1043 rte_eth_bond_link_up_prop_delay_get(uint16_t bonded_port_id)
1044 {
1045         struct bond_dev_private *internals;
1046
1047         if (valid_bonded_port_id(bonded_port_id) != 0)
1048                 return -1;
1049
1050         internals = rte_eth_devices[bonded_port_id].data->dev_private;
1051
1052         return internals->link_up_delay_ms;
1053 }