app/testpmd: fix metering and policing command for RFC4115
[dpdk.git] / app / test / test_dmadev.c
index b0fba1d..2b097e0 100644 (file)
@@ -82,6 +82,98 @@ await_hw(int16_t dev_id, uint16_t vchan)
        }
 }
 
+/* run a series of copy tests just using some different options for enqueues and completions */
+static int
+do_multi_copies(int16_t dev_id, uint16_t vchan,
+               int split_batches,     /* submit 2 x 16 or 1 x 32 burst */
+               int split_completions, /* gather 2 x 16 or 1 x 32 completions */
+               int use_completed_status) /* use completed or completed_status function */
+{
+       struct rte_mbuf *srcs[32], *dsts[32];
+       enum rte_dma_status_code sc[32];
+       unsigned int i, j;
+       bool dma_err = false;
+
+       /* Enqueue burst of copies and hit doorbell */
+       for (i = 0; i < RTE_DIM(srcs); i++) {
+               uint64_t *src_data;
+
+               if (split_batches && i == RTE_DIM(srcs) / 2)
+                       rte_dma_submit(dev_id, vchan);
+
+               srcs[i] = rte_pktmbuf_alloc(pool);
+               dsts[i] = rte_pktmbuf_alloc(pool);
+               if (srcs[i] == NULL || dsts[i] == NULL)
+                       ERR_RETURN("Error allocating buffers\n");
+
+               src_data = rte_pktmbuf_mtod(srcs[i], uint64_t *);
+               for (j = 0; j < COPY_LEN/sizeof(uint64_t); j++)
+                       src_data[j] = rte_rand();
+
+               if (rte_dma_copy(dev_id, vchan, srcs[i]->buf_iova + srcs[i]->data_off,
+                               dsts[i]->buf_iova + dsts[i]->data_off, COPY_LEN, 0) != id_count++)
+                       ERR_RETURN("Error with rte_dma_copy for buffer %u\n", i);
+       }
+       rte_dma_submit(dev_id, vchan);
+
+       await_hw(dev_id, vchan);
+
+       if (split_completions) {
+               /* gather completions in two halves */
+               uint16_t half_len = RTE_DIM(srcs) / 2;
+               int ret = rte_dma_completed(dev_id, vchan, half_len, NULL, &dma_err);
+               if (ret != half_len || dma_err)
+                       ERR_RETURN("Error with rte_dma_completed - first half. ret = %d, expected ret = %u, dma_err = %d\n",
+                                       ret, half_len, dma_err);
+
+               ret = rte_dma_completed(dev_id, vchan, half_len, NULL, &dma_err);
+               if (ret != half_len || dma_err)
+                       ERR_RETURN("Error with rte_dma_completed - second half. ret = %d, expected ret = %u, dma_err = %d\n",
+                                       ret, half_len, dma_err);
+       } else {
+               /* gather all completions in one go, using either
+                * completed or completed_status fns
+                */
+               if (!use_completed_status) {
+                       int n = rte_dma_completed(dev_id, vchan, RTE_DIM(srcs), NULL, &dma_err);
+                       if (n != RTE_DIM(srcs) || dma_err)
+                               ERR_RETURN("Error with rte_dma_completed, %u [expected: %zu], dma_err = %d\n",
+                                               n, RTE_DIM(srcs), dma_err);
+               } else {
+                       int n = rte_dma_completed_status(dev_id, vchan, RTE_DIM(srcs), NULL, sc);
+                       if (n != RTE_DIM(srcs))
+                               ERR_RETURN("Error with rte_dma_completed_status, %u [expected: %zu]\n",
+                                               n, RTE_DIM(srcs));
+
+                       for (j = 0; j < (uint16_t)n; j++)
+                               if (sc[j] != RTE_DMA_STATUS_SUCCESSFUL)
+                                       ERR_RETURN("Error with rte_dma_completed_status, job %u reports failure [code %u]\n",
+                                                       j, sc[j]);
+               }
+       }
+
+       /* check for empty */
+       int ret = use_completed_status ?
+                       rte_dma_completed_status(dev_id, vchan, RTE_DIM(srcs), NULL, sc) :
+                       rte_dma_completed(dev_id, vchan, RTE_DIM(srcs), NULL, &dma_err);
+       if (ret != 0)
+               ERR_RETURN("Error with completion check - ops unexpectedly returned\n");
+
+       for (i = 0; i < RTE_DIM(srcs); i++) {
+               char *src_data, *dst_data;
+
+               src_data = rte_pktmbuf_mtod(srcs[i], char *);
+               dst_data = rte_pktmbuf_mtod(dsts[i], char *);
+               for (j = 0; j < COPY_LEN; j++)
+                       if (src_data[j] != dst_data[j])
+                               ERR_RETURN("Error with copy of packet %u, byte %u\n", i, j);
+
+               rte_pktmbuf_free(srcs[i]);
+               rte_pktmbuf_free(dsts[i]);
+       }
+       return 0;
+}
+
 static int
 test_enqueue_copies(int16_t dev_id, uint16_t vchan)
 {
@@ -176,6 +268,469 @@ test_enqueue_copies(int16_t dev_id, uint16_t vchan)
                rte_pktmbuf_free(dst);
        } while (0);
 
+       /* test doing multiple copies */
+       return do_multi_copies(dev_id, vchan, 0, 0, 0) /* enqueue and complete 1 batch at a time */
+                       /* enqueue 2 batches and then complete both */
+                       || do_multi_copies(dev_id, vchan, 1, 0, 0)
+                       /* enqueue 1 batch, then complete in two halves */
+                       || do_multi_copies(dev_id, vchan, 0, 1, 0)
+                       /* test using completed_status in place of regular completed API */
+                       || do_multi_copies(dev_id, vchan, 0, 0, 1);
+}
+
+/* Failure handling test cases - global macros and variables for those tests*/
+#define COMP_BURST_SZ  16
+#define OPT_FENCE(idx) ((fence && idx == 8) ? RTE_DMA_OP_FLAG_FENCE : 0)
+
+static int
+test_failure_in_full_burst(int16_t dev_id, uint16_t vchan, bool fence,
+               struct rte_mbuf **srcs, struct rte_mbuf **dsts, unsigned int fail_idx)
+{
+       /* Test single full batch statuses with failures */
+       enum rte_dma_status_code status[COMP_BURST_SZ];
+       struct rte_dma_stats baseline, stats;
+       uint16_t invalid_addr_id = 0;
+       uint16_t idx;
+       uint16_t count, status_count;
+       unsigned int i;
+       bool error = false;
+       int err_count = 0;
+
+       rte_dma_stats_get(dev_id, vchan, &baseline); /* get a baseline set of stats */
+       for (i = 0; i < COMP_BURST_SZ; i++) {
+               int id = rte_dma_copy(dev_id, vchan,
+                               (i == fail_idx ? 0 : (srcs[i]->buf_iova + srcs[i]->data_off)),
+                               dsts[i]->buf_iova + dsts[i]->data_off,
+                               COPY_LEN, OPT_FENCE(i));
+               if (id < 0)
+                       ERR_RETURN("Error with rte_dma_copy for buffer %u\n", i);
+               if (i == fail_idx)
+                       invalid_addr_id = id;
+       }
+       rte_dma_submit(dev_id, vchan);
+       rte_dma_stats_get(dev_id, vchan, &stats);
+       if (stats.submitted != baseline.submitted + COMP_BURST_SZ)
+               ERR_RETURN("Submitted stats value not as expected, %"PRIu64" not %"PRIu64"\n",
+                               stats.submitted, baseline.submitted + COMP_BURST_SZ);
+
+       await_hw(dev_id, vchan);
+
+       count = rte_dma_completed(dev_id, vchan, COMP_BURST_SZ, &idx, &error);
+       if (count != fail_idx)
+               ERR_RETURN("Error with rte_dma_completed for failure test. Got returned %u not %u.\n",
+                               count, fail_idx);
+       if (!error)
+               ERR_RETURN("Error, missing expected failed copy, %u. has_error is not set\n",
+                               fail_idx);
+       if (idx != invalid_addr_id - 1)
+               ERR_RETURN("Error, missing expected failed copy, %u. Got last idx %u, not %u\n",
+                               fail_idx, idx, invalid_addr_id - 1);
+
+       /* all checks ok, now verify calling completed() again always returns 0 */
+       for (i = 0; i < 10; i++)
+               if (rte_dma_completed(dev_id, vchan, COMP_BURST_SZ, &idx, &error) != 0
+                               || error == false || idx != (invalid_addr_id - 1))
+                       ERR_RETURN("Error with follow-up completed calls for fail idx %u\n",
+                                       fail_idx);
+
+       status_count = rte_dma_completed_status(dev_id, vchan, COMP_BURST_SZ,
+                       &idx, status);
+       /* some HW may stop on error and be restarted after getting error status for single value
+        * To handle this case, if we get just one error back, wait for more completions and get
+        * status for rest of the burst
+        */
+       if (status_count == 1) {
+               await_hw(dev_id, vchan);
+               status_count += rte_dma_completed_status(dev_id, vchan, COMP_BURST_SZ - 1,
+                                       &idx, &status[1]);
+       }
+       /* check that at this point we have all status values */
+       if (status_count != COMP_BURST_SZ - count)
+               ERR_RETURN("Error with completed_status calls for fail idx %u. Got %u not %u\n",
+                               fail_idx, status_count, COMP_BURST_SZ - count);
+       /* now verify just one failure followed by multiple successful or skipped entries */
+       if (status[0] == RTE_DMA_STATUS_SUCCESSFUL)
+               ERR_RETURN("Error with status returned for fail idx %u. First status was not failure\n",
+                               fail_idx);
+       for (i = 1; i < status_count; i++)
+               /* after a failure in a burst, depending on ordering/fencing,
+                * operations may be successful or skipped because of previous error.
+                */
+               if (status[i] != RTE_DMA_STATUS_SUCCESSFUL
+                               && status[i] != RTE_DMA_STATUS_NOT_ATTEMPTED)
+                       ERR_RETURN("Error with status calls for fail idx %u. Status for job %u (of %u) is not successful\n",
+                                       fail_idx, count + i, COMP_BURST_SZ);
+
+       /* check the completed + errors stats are as expected */
+       rte_dma_stats_get(dev_id, vchan, &stats);
+       if (stats.completed != baseline.completed + COMP_BURST_SZ)
+               ERR_RETURN("Completed stats value not as expected, %"PRIu64" not %"PRIu64"\n",
+                               stats.completed, baseline.completed + COMP_BURST_SZ);
+       for (i = 0; i < status_count; i++)
+               err_count += (status[i] != RTE_DMA_STATUS_SUCCESSFUL);
+       if (stats.errors != baseline.errors + err_count)
+               ERR_RETURN("'Errors' stats value not as expected, %"PRIu64" not %"PRIu64"\n",
+                               stats.errors, baseline.errors + err_count);
+
+       return 0;
+}
+
+static int
+test_individual_status_query_with_failure(int16_t dev_id, uint16_t vchan, bool fence,
+               struct rte_mbuf **srcs, struct rte_mbuf **dsts, unsigned int fail_idx)
+{
+       /* Test gathering batch statuses one at a time */
+       enum rte_dma_status_code status[COMP_BURST_SZ];
+       uint16_t invalid_addr_id = 0;
+       uint16_t idx;
+       uint16_t count = 0, status_count = 0;
+       unsigned int j;
+       bool error = false;
+
+       for (j = 0; j < COMP_BURST_SZ; j++) {
+               int id = rte_dma_copy(dev_id, vchan,
+                               (j == fail_idx ? 0 : (srcs[j]->buf_iova + srcs[j]->data_off)),
+                               dsts[j]->buf_iova + dsts[j]->data_off,
+                               COPY_LEN, OPT_FENCE(j));
+               if (id < 0)
+                       ERR_RETURN("Error with rte_dma_copy for buffer %u\n", j);
+               if (j == fail_idx)
+                       invalid_addr_id = id;
+       }
+       rte_dma_submit(dev_id, vchan);
+       await_hw(dev_id, vchan);
+
+       /* use regular "completed" until we hit error */
+       while (!error) {
+               uint16_t n = rte_dma_completed(dev_id, vchan, 1, &idx, &error);
+               count += n;
+               if (n > 1 || count >= COMP_BURST_SZ)
+                       ERR_RETURN("Error - too many completions got\n");
+               if (n == 0 && !error)
+                       ERR_RETURN("Error, unexpectedly got zero completions after %u completed\n",
+                                       count);
+       }
+       if (idx != invalid_addr_id - 1)
+               ERR_RETURN("Error, last successful index not as expected, got %u, expected %u\n",
+                               idx, invalid_addr_id - 1);
+
+       /* use completed_status until we hit end of burst */
+       while (count + status_count < COMP_BURST_SZ) {
+               uint16_t n = rte_dma_completed_status(dev_id, vchan, 1, &idx,
+                               &status[status_count]);
+               await_hw(dev_id, vchan); /* allow delay to ensure jobs are completed */
+               status_count += n;
+               if (n != 1)
+                       ERR_RETURN("Error: unexpected number of completions received, %u, not 1\n",
+                                       n);
+       }
+
+       /* check for single failure */
+       if (status[0] == RTE_DMA_STATUS_SUCCESSFUL)
+               ERR_RETURN("Error, unexpected successful DMA transaction\n");
+       for (j = 1; j < status_count; j++)
+               if (status[j] != RTE_DMA_STATUS_SUCCESSFUL
+                               && status[j] != RTE_DMA_STATUS_NOT_ATTEMPTED)
+                       ERR_RETURN("Error, unexpected DMA error reported\n");
+
+       return 0;
+}
+
+static int
+test_single_item_status_query_with_failure(int16_t dev_id, uint16_t vchan,
+               struct rte_mbuf **srcs, struct rte_mbuf **dsts, unsigned int fail_idx)
+{
+       /* When error occurs just collect a single error using "completed_status()"
+        * before going to back to completed() calls
+        */
+       enum rte_dma_status_code status;
+       uint16_t invalid_addr_id = 0;
+       uint16_t idx;
+       uint16_t count, status_count, count2;
+       unsigned int j;
+       bool error = false;
+
+       for (j = 0; j < COMP_BURST_SZ; j++) {
+               int id = rte_dma_copy(dev_id, vchan,
+                               (j == fail_idx ? 0 : (srcs[j]->buf_iova + srcs[j]->data_off)),
+                               dsts[j]->buf_iova + dsts[j]->data_off,
+                               COPY_LEN, 0);
+               if (id < 0)
+                       ERR_RETURN("Error with rte_dma_copy for buffer %u\n", j);
+               if (j == fail_idx)
+                       invalid_addr_id = id;
+       }
+       rte_dma_submit(dev_id, vchan);
+       await_hw(dev_id, vchan);
+
+       /* get up to the error point */
+       count = rte_dma_completed(dev_id, vchan, COMP_BURST_SZ, &idx, &error);
+       if (count != fail_idx)
+               ERR_RETURN("Error with rte_dma_completed for failure test. Got returned %u not %u.\n",
+                               count, fail_idx);
+       if (!error)
+               ERR_RETURN("Error, missing expected failed copy, %u. has_error is not set\n",
+                               fail_idx);
+       if (idx != invalid_addr_id - 1)
+               ERR_RETURN("Error, missing expected failed copy, %u. Got last idx %u, not %u\n",
+                               fail_idx, idx, invalid_addr_id - 1);
+
+       /* get the error code */
+       status_count = rte_dma_completed_status(dev_id, vchan, 1, &idx, &status);
+       if (status_count != 1)
+               ERR_RETURN("Error with completed_status calls for fail idx %u. Got %u not %u\n",
+                               fail_idx, status_count, COMP_BURST_SZ - count);
+       if (status == RTE_DMA_STATUS_SUCCESSFUL)
+               ERR_RETURN("Error with status returned for fail idx %u. First status was not failure\n",
+                               fail_idx);
+
+       /* delay in case time needed after err handled to complete other jobs */
+       await_hw(dev_id, vchan);
+
+       /* get the rest of the completions without status */
+       count2 = rte_dma_completed(dev_id, vchan, COMP_BURST_SZ, &idx, &error);
+       if (error == true)
+               ERR_RETURN("Error, got further errors post completed_status() call, for failure case %u.\n",
+                               fail_idx);
+       if (count + status_count + count2 != COMP_BURST_SZ)
+               ERR_RETURN("Error, incorrect number of completions received, got %u not %u\n",
+                               count + status_count + count2, COMP_BURST_SZ);
+
+       return 0;
+}
+
+static int
+test_multi_failure(int16_t dev_id, uint16_t vchan, struct rte_mbuf **srcs, struct rte_mbuf **dsts,
+               const unsigned int *fail, size_t num_fail)
+{
+       /* test having multiple errors in one go */
+       enum rte_dma_status_code status[COMP_BURST_SZ];
+       unsigned int i, j;
+       uint16_t count, err_count = 0;
+       bool error = false;
+
+       /* enqueue and gather completions in one go */
+       for (j = 0; j < COMP_BURST_SZ; j++) {
+               uintptr_t src = srcs[j]->buf_iova + srcs[j]->data_off;
+               /* set up for failure if the current index is anywhere is the fails array */
+               for (i = 0; i < num_fail; i++)
+                       if (j == fail[i])
+                               src = 0;
+
+               int id = rte_dma_copy(dev_id, vchan,
+                               src, dsts[j]->buf_iova + dsts[j]->data_off,
+                               COPY_LEN, 0);
+               if (id < 0)
+                       ERR_RETURN("Error with rte_dma_copy for buffer %u\n", j);
+       }
+       rte_dma_submit(dev_id, vchan);
+       await_hw(dev_id, vchan);
+
+       count = rte_dma_completed_status(dev_id, vchan, COMP_BURST_SZ, NULL, status);
+       while (count < COMP_BURST_SZ) {
+               await_hw(dev_id, vchan);
+
+               uint16_t ret = rte_dma_completed_status(dev_id, vchan, COMP_BURST_SZ - count,
+                               NULL, &status[count]);
+               if (ret == 0)
+                       ERR_RETURN("Error getting all completions for jobs. Got %u of %u\n",
+                                       count, COMP_BURST_SZ);
+               count += ret;
+       }
+       for (i = 0; i < count; i++)
+               if (status[i] != RTE_DMA_STATUS_SUCCESSFUL)
+                       err_count++;
+
+       if (err_count != num_fail)
+               ERR_RETURN("Error: Invalid number of failed completions returned, %u; expected %zu\n",
+                       err_count, num_fail);
+
+       /* enqueue and gather completions in bursts, but getting errors one at a time */
+       for (j = 0; j < COMP_BURST_SZ; j++) {
+               uintptr_t src = srcs[j]->buf_iova + srcs[j]->data_off;
+               /* set up for failure if the current index is anywhere is the fails array */
+               for (i = 0; i < num_fail; i++)
+                       if (j == fail[i])
+                               src = 0;
+
+               int id = rte_dma_copy(dev_id, vchan,
+                               src, dsts[j]->buf_iova + dsts[j]->data_off,
+                               COPY_LEN, 0);
+               if (id < 0)
+                       ERR_RETURN("Error with rte_dma_copy for buffer %u\n", j);
+       }
+       rte_dma_submit(dev_id, vchan);
+       await_hw(dev_id, vchan);
+
+       count = 0;
+       err_count = 0;
+       while (count + err_count < COMP_BURST_SZ) {
+               count += rte_dma_completed(dev_id, vchan, COMP_BURST_SZ, NULL, &error);
+               if (error) {
+                       uint16_t ret = rte_dma_completed_status(dev_id, vchan, 1,
+                                       NULL, status);
+                       if (ret != 1)
+                               ERR_RETURN("Error getting error-status for completions\n");
+                       err_count += ret;
+                       await_hw(dev_id, vchan);
+               }
+       }
+       if (err_count != num_fail)
+               ERR_RETURN("Error: Incorrect number of failed completions received, got %u not %zu\n",
+                               err_count, num_fail);
+
+       return 0;
+}
+
+static int
+test_completion_status(int16_t dev_id, uint16_t vchan, bool fence)
+{
+       const unsigned int fail[] = {0, 7, 14, 15};
+       struct rte_mbuf *srcs[COMP_BURST_SZ], *dsts[COMP_BURST_SZ];
+       unsigned int i;
+
+       for (i = 0; i < COMP_BURST_SZ; i++) {
+               srcs[i] = rte_pktmbuf_alloc(pool);
+               dsts[i] = rte_pktmbuf_alloc(pool);
+       }
+
+       for (i = 0; i < RTE_DIM(fail); i++) {
+               if (test_failure_in_full_burst(dev_id, vchan, fence, srcs, dsts, fail[i]) < 0)
+                       return -1;
+
+               if (test_individual_status_query_with_failure(dev_id, vchan, fence,
+                               srcs, dsts, fail[i]) < 0)
+                       return -1;
+
+               /* test is run the same fenced, or unfenced, but no harm in running it twice */
+               if (test_single_item_status_query_with_failure(dev_id, vchan,
+                               srcs, dsts, fail[i]) < 0)
+                       return -1;
+       }
+
+       if (test_multi_failure(dev_id, vchan, srcs, dsts, fail, RTE_DIM(fail)) < 0)
+               return -1;
+
+       for (i = 0; i < COMP_BURST_SZ; i++) {
+               rte_pktmbuf_free(srcs[i]);
+               rte_pktmbuf_free(dsts[i]);
+       }
+       return 0;
+}
+
+static int
+test_completion_handling(int16_t dev_id, uint16_t vchan)
+{
+       return test_completion_status(dev_id, vchan, false)              /* without fences */
+                       || test_completion_status(dev_id, vchan, true);  /* with fences */
+}
+
+static int
+test_enqueue_fill(int16_t dev_id, uint16_t vchan)
+{
+       const unsigned int lengths[] = {8, 64, 1024, 50, 100, 89};
+       struct rte_mbuf *dst;
+       char *dst_data;
+       uint64_t pattern = 0xfedcba9876543210;
+       unsigned int i, j;
+
+       dst = rte_pktmbuf_alloc(pool);
+       if (dst == NULL)
+               ERR_RETURN("Failed to allocate mbuf\n");
+       dst_data = rte_pktmbuf_mtod(dst, char *);
+
+       for (i = 0; i < RTE_DIM(lengths); i++) {
+               /* reset dst_data */
+               memset(dst_data, 0, rte_pktmbuf_data_len(dst));
+
+               /* perform the fill operation */
+               int id = rte_dma_fill(dev_id, vchan, pattern,
+                               rte_pktmbuf_iova(dst), lengths[i], RTE_DMA_OP_FLAG_SUBMIT);
+               if (id < 0)
+                       ERR_RETURN("Error with rte_dma_fill\n");
+               await_hw(dev_id, vchan);
+
+               if (rte_dma_completed(dev_id, vchan, 1, NULL, NULL) != 1)
+                       ERR_RETURN("Error: fill operation failed (length: %u)\n", lengths[i]);
+               /* check the data from the fill operation is correct */
+               for (j = 0; j < lengths[i]; j++) {
+                       char pat_byte = ((char *)&pattern)[j % 8];
+                       if (dst_data[j] != pat_byte)
+                               ERR_RETURN("Error with fill operation (lengths = %u): got (%x), not (%x)\n",
+                                               lengths[i], dst_data[j], pat_byte);
+               }
+               /* check that the data after the fill operation was not written to */
+               for (; j < rte_pktmbuf_data_len(dst); j++)
+                       if (dst_data[j] != 0)
+                               ERR_RETURN("Error, fill operation wrote too far (lengths = %u): got (%x), not (%x)\n",
+                                               lengths[i], dst_data[j], 0);
+       }
+
+       rte_pktmbuf_free(dst);
+       return 0;
+}
+
+static int
+test_burst_capacity(int16_t dev_id, uint16_t vchan)
+{
+#define CAP_TEST_BURST_SIZE    64
+       const int ring_space = rte_dma_burst_capacity(dev_id, vchan);
+       struct rte_mbuf *src, *dst;
+       int i, j, iter;
+       int cap, ret;
+       bool dma_err;
+
+       src = rte_pktmbuf_alloc(pool);
+       dst = rte_pktmbuf_alloc(pool);
+
+       /* to test capacity, we enqueue elements and check capacity is reduced
+        * by one each time - rebaselining the expected value after each burst
+        * as the capacity is only for a burst. We enqueue multiple bursts to
+        * fill up half the ring, before emptying it again. We do this multiple
+        * times to ensure that we get to test scenarios where we get ring
+        * wrap-around and wrap-around of the ids returned (at UINT16_MAX).
+        */
+       for (iter = 0; iter < 2 * (((int)UINT16_MAX + 1) / ring_space); iter++) {
+               for (i = 0; i < (ring_space / (2 * CAP_TEST_BURST_SIZE)) + 1; i++) {
+                       cap = rte_dma_burst_capacity(dev_id, vchan);
+
+                       for (j = 0; j < CAP_TEST_BURST_SIZE; j++) {
+                               ret = rte_dma_copy(dev_id, vchan, rte_pktmbuf_iova(src),
+                                               rte_pktmbuf_iova(dst), COPY_LEN, 0);
+                               if (ret < 0)
+                                       ERR_RETURN("Error with rte_dmadev_copy\n");
+
+                               if (rte_dma_burst_capacity(dev_id, vchan) != cap - (j + 1))
+                                       ERR_RETURN("Error, ring capacity did not change as expected\n");
+                       }
+                       if (rte_dma_submit(dev_id, vchan) < 0)
+                               ERR_RETURN("Error, failed to submit burst\n");
+
+                       if (cap < rte_dma_burst_capacity(dev_id, vchan))
+                               ERR_RETURN("Error, avail ring capacity has gone up, not down\n");
+               }
+               await_hw(dev_id, vchan);
+
+               for (i = 0; i < (ring_space / (2 * CAP_TEST_BURST_SIZE)) + 1; i++) {
+                       ret = rte_dma_completed(dev_id, vchan,
+                                       CAP_TEST_BURST_SIZE, NULL, &dma_err);
+                       if (ret != CAP_TEST_BURST_SIZE || dma_err) {
+                               enum rte_dma_status_code status;
+
+                               rte_dma_completed_status(dev_id, vchan, 1, NULL, &status);
+                               ERR_RETURN("Error with rte_dmadev_completed, %u [expected: %u], dma_err = %d, i = %u, iter = %u, status = %u\n",
+                                               ret, CAP_TEST_BURST_SIZE, dma_err, i, iter, status);
+                       }
+               }
+               cap = rte_dma_burst_capacity(dev_id, vchan);
+               if (cap != ring_space)
+                       ERR_RETURN("Error, ring capacity has not reset to original value, got %u, expected %u\n",
+                                       cap, ring_space);
+       }
+
+       rte_pktmbuf_free(src);
+       rte_pktmbuf_free(dst);
+
        return 0;
 }
 
@@ -192,11 +747,15 @@ test_dmadev_instance(int16_t dev_id)
                        .nb_desc = TEST_RINGSIZE,
        };
        const int vchan = 0;
+       int ret;
+
+       ret = rte_dma_info_get(dev_id, &info);
+       if (ret != 0)
+               ERR_RETURN("Error with rte_dma_info_get()\n");
 
        printf("\n### Test dmadev instance %u [%s]\n",
-                       dev_id, rte_dma_devices[dev_id].data->dev_name);
+                       dev_id, info.dev_name);
 
-       rte_dma_info_get(dev_id, &info);
        if (info.max_vchans < 1)
                ERR_RETURN("Error, no channels available on device id %u\n", dev_id);
 
@@ -206,8 +765,8 @@ test_dmadev_instance(int16_t dev_id)
        if (rte_dma_vchan_setup(dev_id, vchan, &qconf) < 0)
                ERR_RETURN("Error with queue configuration\n");
 
-       rte_dma_info_get(dev_id, &info);
-       if (info.nb_vchans != 1)
+       ret = rte_dma_info_get(dev_id, &info);
+       if (ret != 0 || info.nb_vchans != 1)
                ERR_RETURN("Error, no configured queues reported on device id %u\n", dev_id);
 
        if (rte_dma_start(dev_id) != 0)
@@ -216,6 +775,9 @@ test_dmadev_instance(int16_t dev_id)
        if (rte_dma_stats_get(dev_id, vchan, &stats) != 0)
                ERR_RETURN("Error with rte_dma_stats_get()\n");
 
+       if (rte_dma_burst_capacity(dev_id, vchan) < 32)
+               ERR_RETURN("Error: Device does not have sufficient burst capacity to run tests");
+
        if (stats.completed != 0 || stats.submitted != 0 || stats.errors != 0)
                ERR_RETURN("Error device stats are not all zero: completed = %"PRIu64", "
                                "submitted = %"PRIu64", errors = %"PRIu64"\n",
@@ -236,6 +798,31 @@ test_dmadev_instance(int16_t dev_id)
        if (runtest("copy", test_enqueue_copies, 640, dev_id, vchan, CHECK_ERRS) < 0)
                goto err;
 
+       /* run some burst capacity tests */
+       if (rte_dma_burst_capacity(dev_id, vchan) < 64)
+               printf("DMA Dev %u: insufficient burst capacity (64 required), skipping tests\n",
+                               dev_id);
+       else if (runtest("burst capacity", test_burst_capacity, 1, dev_id, vchan, CHECK_ERRS) < 0)
+               goto err;
+
+       /* to test error handling we can provide null pointers for source or dest in copies. This
+        * requires VA mode in DPDK, since NULL(0) is a valid physical address.
+        * We also need hardware that can report errors back.
+        */
+       if (rte_eal_iova_mode() != RTE_IOVA_VA)
+               printf("DMA Dev %u: DPDK not in VA mode, skipping error handling tests\n", dev_id);
+       else if ((info.dev_capa & RTE_DMA_CAPA_HANDLES_ERRORS) == 0)
+               printf("DMA Dev %u: device does not report errors, skipping error handling tests\n",
+                               dev_id);
+       else if (runtest("error handling", test_completion_handling, 1,
+                       dev_id, vchan, !CHECK_ERRS) < 0)
+               goto err;
+
+       if ((info.dev_capa & RTE_DMA_CAPA_OPS_FILL) == 0)
+               printf("DMA Dev %u: No device fill support, skipping fill tests\n", dev_id);
+       else if (runtest("fill", test_enqueue_fill, 1, dev_id, vchan, CHECK_ERRS) < 0)
+               goto err;
+
        rte_mempool_free(pool);
        rte_dma_stop(dev_id);
        rte_dma_stats_reset(dev_id, vchan);