* reference CYCLES to be used to
* measure core busyness based on poll count
*/
-#define MIN_CYCLES 1500000ULL
-#define MAX_CYCLES 2500000ULL
+#define MIN_CYCLES 1500000ULL
+#define MAX_CYCLES 22000000ULL
/* (500ms) */
#define TELEMETRY_INTERVALS_PER_SEC 2
br = FULL;
} else if (diff_tsc > MIN_CYCLES &&
diff_tsc < MAX_CYCLES) {
- br = PARTIAL;
+ br = (diff_tsc * 100) / MAX_CYCLES;
} else {
br = ZERO;
}
/* init RX queues */
for(queue = 0; queue < qconf->n_rx_queue; ++queue) {
struct rte_eth_rxconf rxq_conf;
- struct rte_eth_dev *dev;
- struct rte_eth_conf *conf;
portid = qconf->rx_queue_list[queue].port_id;
queueid = qconf->rx_queue_list[queue].queue_id;
- dev = &rte_eth_devices[portid];
- conf = &dev->data->dev_conf;
if (numa_on)
socketid = \
rte_eth_dev_info_get(portid, &dev_info);
rxq_conf = dev_info.default_rxconf;
- rxq_conf.offloads = conf->rxmode.offloads;
+ rxq_conf.offloads = port_conf.rxmode.offloads;
ret = rte_eth_rx_queue_setup(portid, queueid, nb_rxd,
socketid, &rxq_conf,
pktmbuf_pool[socketid]);