X-Git-Url: http://git.droids-corp.org/?a=blobdiff_plain;f=app%2Ftest-compress-perf%2Fcomp_perf_test_common.c;h=1b8985b4305425ce707445dd2ff7c6daae97d09d;hb=94a24aaf6c5bd0a03c2828e7411d30a4fc0ac075;hp=dc9d0b0f49cc29c66adcc34269964ee39d0a5be7;hpb=424dd6c8c1a86fbc73c3e8f739bafa0324640ca8;p=dpdk.git diff --git a/app/test-compress-perf/comp_perf_test_common.c b/app/test-compress-perf/comp_perf_test_common.c index dc9d0b0f49..1b8985b430 100644 --- a/app/test-compress-perf/comp_perf_test_common.c +++ b/app/test-compress-perf/comp_perf_test_common.c @@ -7,14 +7,27 @@ #include #include +#include "comp_perf.h" #include "comp_perf_options.h" -#include "comp_perf_test_verify.h" #include "comp_perf_test_benchmark.h" -#include "comp_perf.h" #include "comp_perf_test_common.h" +#include "comp_perf_test_verify.h" + #define DIV_CEIL(a, b) ((a) / (b) + ((a) % (b) != 0)) +struct cperf_buffer_info { + uint16_t total_segments; + uint16_t segment_sz; + uint16_t last_segment_sz; + uint32_t total_buffs; /*number of buffers = number of ops*/ + uint16_t segments_per_buff; + uint16_t segments_per_last_buff; + size_t input_data_sz; +}; + +static struct cperf_buffer_info buffer_info; + int param_range_check(uint16_t size, const struct rte_param_log2_range *range) { @@ -49,7 +62,7 @@ find_buf_size(uint32_t input_size) * power of 2 but also should be enough to store incompressible data */ - /* We're looking for nearest power of 2 buffer size, which is greather + /* We're looking for nearest power of 2 buffer size, which is greater * than input_size */ uint32_t size = @@ -64,14 +77,18 @@ find_buf_size(uint32_t input_size) } void -comp_perf_free_memory(struct cperf_mem_resources *mem) +comp_perf_free_memory(struct comp_test_data *test_data, + struct cperf_mem_resources *mem) { uint32_t i; - for (i = 0; i < mem->total_bufs; i++) { - rte_pktmbuf_free(mem->comp_bufs[i]); - rte_pktmbuf_free(mem->decomp_bufs[i]); - } + if (mem->decomp_bufs != NULL) + for (i = 0; i < mem->total_bufs; i++) + rte_pktmbuf_free(mem->decomp_bufs[i]); + + if (mem->comp_bufs != NULL) + for (i = 0; i < mem->total_bufs; i++) + rte_pktmbuf_free(mem->comp_bufs[i]); rte_free(mem->decomp_bufs); rte_free(mem->comp_bufs); @@ -80,26 +97,162 @@ comp_perf_free_memory(struct cperf_mem_resources *mem) rte_mempool_free(mem->op_pool); rte_mempool_free(mem->decomp_buf_pool); rte_mempool_free(mem->comp_buf_pool); + + /* external mbuf support */ + if (mem->decomp_memzones != NULL) { + for (i = 0; i < test_data->total_segs; i++) + rte_memzone_free(mem->decomp_memzones[i]); + rte_free(mem->decomp_memzones); + } + if (mem->comp_memzones != NULL) { + for (i = 0; i < test_data->total_segs; i++) + rte_memzone_free(mem->comp_memzones[i]); + rte_free(mem->comp_memzones); + } + rte_free(mem->decomp_buf_infos); + rte_free(mem->comp_buf_infos); +} + +static void +comp_perf_extbuf_free_cb(void *addr __rte_unused, void *opaque __rte_unused) +{ +} + +static const struct rte_memzone * +comp_perf_make_memzone(const char *name, struct cperf_mem_resources *mem, + unsigned int number, size_t size) +{ + unsigned int socket_id = rte_socket_id(); + char mz_name[RTE_MEMZONE_NAMESIZE]; + const struct rte_memzone *memzone; + + snprintf(mz_name, RTE_MEMZONE_NAMESIZE, "%s_s%u_d%u_q%u_%d", name, + socket_id, mem->dev_id, mem->qp_id, number); + memzone = rte_memzone_lookup(mz_name); + if (memzone != NULL && memzone->len != size) { + rte_memzone_free(memzone); + memzone = NULL; + } + if (memzone == NULL) { + memzone = rte_memzone_reserve_aligned(mz_name, size, socket_id, + RTE_MEMZONE_IOVA_CONTIG, RTE_CACHE_LINE_SIZE); + if (memzone == NULL) + RTE_LOG(ERR, USER1, "Can't allocate memory zone %s\n", + mz_name); + } + return memzone; +} + +static int +comp_perf_allocate_external_mbufs(struct comp_test_data *test_data, + struct cperf_mem_resources *mem) +{ + uint32_t i; + + mem->comp_memzones = rte_zmalloc_socket(NULL, + test_data->total_segs * sizeof(struct rte_memzone *), + 0, rte_socket_id()); + + if (mem->comp_memzones == NULL) { + RTE_LOG(ERR, USER1, + "Memory to hold the compression memzones could not be allocated\n"); + return -1; + } + + mem->decomp_memzones = rte_zmalloc_socket(NULL, + test_data->total_segs * sizeof(struct rte_memzone *), + 0, rte_socket_id()); + + if (mem->decomp_memzones == NULL) { + RTE_LOG(ERR, USER1, + "Memory to hold the decompression memzones could not be allocated\n"); + return -1; + } + + mem->comp_buf_infos = rte_zmalloc_socket(NULL, + test_data->total_segs * sizeof(struct rte_mbuf_ext_shared_info), + 0, rte_socket_id()); + + if (mem->comp_buf_infos == NULL) { + RTE_LOG(ERR, USER1, + "Memory to hold the compression buf infos could not be allocated\n"); + return -1; + } + + mem->decomp_buf_infos = rte_zmalloc_socket(NULL, + test_data->total_segs * sizeof(struct rte_mbuf_ext_shared_info), + 0, rte_socket_id()); + + if (mem->decomp_buf_infos == NULL) { + RTE_LOG(ERR, USER1, + "Memory to hold the decompression buf infos could not be allocated\n"); + return -1; + } + + for (i = 0; i < test_data->total_segs; i++) { + mem->comp_memzones[i] = comp_perf_make_memzone("comp", mem, + i, test_data->out_seg_sz); + if (mem->comp_memzones[i] == NULL) { + RTE_LOG(ERR, USER1, + "Memory to hold the compression memzone could not be allocated\n"); + return -1; + } + + mem->decomp_memzones[i] = comp_perf_make_memzone("decomp", mem, + i, test_data->seg_sz); + if (mem->decomp_memzones[i] == NULL) { + RTE_LOG(ERR, USER1, + "Memory to hold the decompression memzone could not be allocated\n"); + return -1; + } + + mem->comp_buf_infos[i].free_cb = + comp_perf_extbuf_free_cb; + mem->comp_buf_infos[i].fcb_opaque = NULL; + rte_mbuf_ext_refcnt_set(&mem->comp_buf_infos[i], 1); + + mem->decomp_buf_infos[i].free_cb = + comp_perf_extbuf_free_cb; + mem->decomp_buf_infos[i].fcb_opaque = NULL; + rte_mbuf_ext_refcnt_set(&mem->decomp_buf_infos[i], 1); + } + + return 0; } int comp_perf_allocate_memory(struct comp_test_data *test_data, struct cperf_mem_resources *mem) { + uint16_t comp_mbuf_size; + uint16_t decomp_mbuf_size; + test_data->out_seg_sz = find_buf_size(test_data->seg_sz); + /* Number of segments for input and output * (compression and decompression) */ - uint32_t total_segs = DIV_CEIL(test_data->input_data_sz, + test_data->total_segs = DIV_CEIL(test_data->input_data_sz, test_data->seg_sz); + + if (test_data->use_external_mbufs != 0) { + if (comp_perf_allocate_external_mbufs(test_data, mem) < 0) + return -1; + comp_mbuf_size = 0; + decomp_mbuf_size = 0; + } else { + comp_mbuf_size = test_data->out_seg_sz + RTE_PKTMBUF_HEADROOM; + decomp_mbuf_size = test_data->seg_sz + RTE_PKTMBUF_HEADROOM; + } + char pool_name[32] = ""; snprintf(pool_name, sizeof(pool_name), "comp_buf_pool_%u_qp_%u", mem->dev_id, mem->qp_id); mem->comp_buf_pool = rte_pktmbuf_pool_create(pool_name, - total_segs, + test_data->total_segs, 0, 0, - test_data->out_seg_sz + RTE_PKTMBUF_HEADROOM, + comp_mbuf_size, rte_socket_id()); if (mem->comp_buf_pool == NULL) { RTE_LOG(ERR, USER1, "Mbuf mempool could not be created\n"); @@ -109,15 +262,17 @@ comp_perf_allocate_memory(struct comp_test_data *test_data, snprintf(pool_name, sizeof(pool_name), "decomp_buf_pool_%u_qp_%u", mem->dev_id, mem->qp_id); mem->decomp_buf_pool = rte_pktmbuf_pool_create(pool_name, - total_segs, - 0, 0, test_data->seg_sz + RTE_PKTMBUF_HEADROOM, + test_data->total_segs, + 0, 0, + decomp_mbuf_size, rte_socket_id()); if (mem->decomp_buf_pool == NULL) { RTE_LOG(ERR, USER1, "Mbuf mempool could not be created\n"); return -1; } - mem->total_bufs = DIV_CEIL(total_segs, test_data->max_sgl_segs); + mem->total_bufs = DIV_CEIL(test_data->total_segs, + test_data->max_sgl_segs); snprintf(pool_name, sizeof(pool_name), "op_pool_%u_qp_%u", mem->dev_id, mem->qp_id); @@ -134,8 +289,11 @@ comp_perf_allocate_memory(struct comp_test_data *test_data, * if data cannot be compressed */ mem->compressed_data = rte_zmalloc_socket(NULL, - test_data->input_data_sz * EXPANSE_RATIO - + MIN_COMPRESSED_BUF_SIZE, 0, + RTE_MAX( + (size_t) test_data->out_seg_sz * + test_data->total_segs, + (size_t) MIN_COMPRESSED_BUF_SIZE), + 0, rte_socket_id()); if (mem->compressed_data == NULL) { RTE_LOG(ERR, USER1, "Memory to hold the data from the input " @@ -169,6 +327,13 @@ comp_perf_allocate_memory(struct comp_test_data *test_data, " could not be allocated\n"); return -1; } + + buffer_info.total_segments = test_data->total_segs; + buffer_info.segment_sz = test_data->seg_sz; + buffer_info.total_buffs = mem->total_bufs; + buffer_info.segments_per_buff = test_data->max_sgl_segs; + buffer_info.input_data_sz = test_data->input_data_sz; + return 0; } @@ -177,9 +342,12 @@ prepare_bufs(struct comp_test_data *test_data, struct cperf_mem_resources *mem) { uint32_t remaining_data = test_data->input_data_sz; uint8_t *input_data_ptr = test_data->input_data; - size_t data_sz; + size_t data_sz = 0; uint8_t *data_addr; uint32_t i, j; + uint16_t segs_per_mbuf = 0; + uint32_t cmz = 0; + uint32_t dmz = 0; for (i = 0; i < mem->total_bufs; i++) { /* Allocate data in input mbuf and copy data from input file */ @@ -191,6 +359,16 @@ prepare_bufs(struct comp_test_data *test_data, struct cperf_mem_resources *mem) } data_sz = RTE_MIN(remaining_data, test_data->seg_sz); + + if (test_data->use_external_mbufs != 0) { + rte_pktmbuf_attach_extbuf(mem->decomp_bufs[i], + mem->decomp_memzones[dmz]->addr, + mem->decomp_memzones[dmz]->iova, + test_data->seg_sz, + &mem->decomp_buf_infos[dmz]); + dmz++; + } + data_addr = (uint8_t *) rte_pktmbuf_append( mem->decomp_bufs[i], data_sz); if (data_addr == NULL) { @@ -203,7 +381,7 @@ prepare_bufs(struct comp_test_data *test_data, struct cperf_mem_resources *mem) remaining_data -= data_sz; /* Already one segment in the mbuf */ - uint16_t segs_per_mbuf = 1; + segs_per_mbuf = 1; /* Chain mbufs if needed for input mbufs */ while (segs_per_mbuf < test_data->max_sgl_segs @@ -218,6 +396,17 @@ prepare_bufs(struct comp_test_data *test_data, struct cperf_mem_resources *mem) } data_sz = RTE_MIN(remaining_data, test_data->seg_sz); + + if (test_data->use_external_mbufs != 0) { + rte_pktmbuf_attach_extbuf( + next_seg, + mem->decomp_memzones[dmz]->addr, + mem->decomp_memzones[dmz]->iova, + test_data->seg_sz, + &mem->decomp_buf_infos[dmz]); + dmz++; + } + data_addr = (uint8_t *)rte_pktmbuf_append(next_seg, data_sz); @@ -245,6 +434,16 @@ prepare_bufs(struct comp_test_data *test_data, struct cperf_mem_resources *mem) RTE_LOG(ERR, USER1, "Could not allocate mbuf\n"); return -1; } + + if (test_data->use_external_mbufs != 0) { + rte_pktmbuf_attach_extbuf(mem->comp_bufs[i], + mem->comp_memzones[cmz]->addr, + mem->comp_memzones[cmz]->iova, + test_data->out_seg_sz, + &mem->comp_buf_infos[cmz]); + cmz++; + } + data_addr = (uint8_t *) rte_pktmbuf_append( mem->comp_bufs[i], test_data->out_seg_sz); @@ -264,9 +463,18 @@ prepare_bufs(struct comp_test_data *test_data, struct cperf_mem_resources *mem) return -1; } + if (test_data->use_external_mbufs != 0) { + rte_pktmbuf_attach_extbuf( + next_seg, + mem->comp_memzones[cmz]->addr, + mem->comp_memzones[cmz]->iova, + test_data->out_seg_sz, + &mem->comp_buf_infos[cmz]); + cmz++; + } + data_addr = (uint8_t *)rte_pktmbuf_append(next_seg, test_data->out_seg_sz); - if (data_addr == NULL) { RTE_LOG(ERR, USER1, "Could not append data\n"); return -1; @@ -280,5 +488,75 @@ prepare_bufs(struct comp_test_data *test_data, struct cperf_mem_resources *mem) } } + buffer_info.segments_per_last_buff = segs_per_mbuf; + buffer_info.last_segment_sz = data_sz; + return 0; } + +void +print_test_dynamics(void) +{ + uint32_t opt_total_segs = DIV_CEIL(buffer_info.input_data_sz, + MAX_SEG_SIZE); + + if (buffer_info.total_buffs > 1) { + printf("\nWarning: for the current input parameters, number" + " of ops is higher than one, which may result" + " in sub-optimal performance.\n"); + printf("To improve the performance (for the current" + " input data) following parameters are" + " suggested:\n"); + printf(" * Segment size: %d\n", MAX_SEG_SIZE); + printf(" * Number of segments: %u\n", opt_total_segs); + } else if (buffer_info.total_buffs == 1) { + printf("\nInfo: there is only one op with %u segments -" + " the compression ratio is the best.\n", + buffer_info.segments_per_last_buff); + if (buffer_info.segment_sz < MAX_SEG_SIZE) + printf("To reduce compression time, please use" + " bigger segment size: %d.\n", + MAX_SEG_SIZE); + else if (buffer_info.segment_sz == MAX_SEG_SIZE) + printf("Segment size is optimal for the best" + " performance.\n"); + } else + printf("Warning: something wrong happened!!\n"); + + printf("\nFor the current input parameters (segment size = %u," + " maximum segments per SGL = %u):\n", + buffer_info.segment_sz, + buffer_info.segments_per_buff); + printf(" * Total number of buffers: %d\n", + buffer_info.total_segments); + printf(" * %u buffer(s) %u bytes long, last buffer %u" + " byte(s) long\n", + buffer_info.total_segments - 1, + buffer_info.segment_sz, + buffer_info.last_segment_sz); + printf(" * Number of ops: %u\n", buffer_info.total_buffs); + printf(" * Total memory allocation: %u\n", + (buffer_info.total_segments - 1) * buffer_info.segment_sz + + buffer_info.last_segment_sz); + if (buffer_info.total_buffs > 1) + printf(" * %u ops: %u segment(s) in each," + " segment size %u\n", + buffer_info.total_buffs - 1, + buffer_info.segments_per_buff, + buffer_info.segment_sz); + if (buffer_info.segments_per_last_buff > 1) { + printf(" * 1 op %u segments:\n", + buffer_info.segments_per_last_buff); + printf(" o %u segment size %u\n", + buffer_info.segments_per_last_buff - 1, + buffer_info.segment_sz); + printf(" o last segment size %u\n", + buffer_info.last_segment_sz); + } else if (buffer_info.segments_per_last_buff == 1) { + printf(" * 1 op (the last one): %u segment %u" + " byte(s) long\n\n", + buffer_info.segments_per_last_buff, + buffer_info.last_segment_sz); + } + printf("\n"); +}