]> git.droids-corp.org - dpdk.git/commitdiff
raw/ioat: add separate API for fence call
authorBruce Richardson <bruce.richardson@intel.com>
Thu, 8 Oct 2020 09:51:16 +0000 (10:51 +0100)
committerThomas Monjalon <thomas@monjalon.net>
Thu, 8 Oct 2020 12:33:20 +0000 (14:33 +0200)
Rather than having the fence signalled via a flag on a descriptor - which
requires reading the docs to find out whether the flag needs to go on the
last descriptor before, or the first descriptor after the fence - we can
instead add a separate fence API call. This becomes unambiguous to use,
since the fence call explicitly comes between two other enqueue calls. It
also allows more freedom of implementation in the driver code.

Signed-off-by: Bruce Richardson <bruce.richardson@intel.com>
Reviewed-by: Kevin Laatz <kevin.laatz@intel.com>
Acked-by: Radu Nicolau <radu.nicolau@intel.com>
doc/guides/rawdevs/ioat.rst
doc/guides/rel_notes/release_20_11.rst
drivers/raw/ioat/ioat_rawdev_test.c
drivers/raw/ioat/rte_ioat_rawdev.h
drivers/raw/ioat/rte_ioat_rawdev_fns.h
examples/ioat/ioatfwd.c

index 3db5f5d0972ec20888b5a6da0bc5064d3f0c3a0f..71bca0b28f73588fc9f24f8ab3b3037ae7d56431 100644 (file)
@@ -203,8 +203,7 @@ a burst of copies to the device and start the hardware processing of them:
                                 dsts[i]->buf_iova + dsts[i]->data_off,
                                 length,
                                 (uintptr_t)srcs[i],
-                                (uintptr_t)dsts[i],
-                                0 /* nofence */) != 1) {
+                                (uintptr_t)dsts[i]) != 1) {
                         printf("Error with rte_ioat_enqueue_copy for buffer %u\n",
                                         i);
                         return -1;
index 99014f324192d58060f09f4bc789fbce11be80ce..e6591e8d158b3c5332240dc5606bc65880876a17 100644 (file)
@@ -116,6 +116,10 @@ New Features
     to better reflect the APIs' purposes, and remove the implication that
     they are limited to copy operations only.
     [Note: The old API is still provided but marked as deprecated in the code]
+  * Added a new API ``rte_ioat_fence()`` to add a fence between operations.
+    This API replaces the ``fence`` flag parameter in the ``rte_ioat_enqueue_copies()`` function,
+    and is clearer as there is no ambiguity as to whether the flag should be
+    set on the last operation before the fence or the first operation after it.
 
 * **Updated the pipeline library for alignment with the P4 language.**
 
index 439b46c03cd0f891b69f52c79215a4ba96e540eb..8b665cc9ad57ebcd6c94ebcfb3f329b31b60c275 100644 (file)
@@ -60,8 +60,7 @@ test_enqueue_copies(int dev_id)
                                dst->buf_iova + dst->data_off,
                                length,
                                (uintptr_t)src,
-                               (uintptr_t)dst,
-                               0 /* no fence */) != 1) {
+                               (uintptr_t)dst) != 1) {
                        PRINT_ERR("Error with rte_ioat_enqueue_copy\n");
                        return -1;
                }
@@ -112,8 +111,7 @@ test_enqueue_copies(int dev_id)
                                        dsts[i]->buf_iova + dsts[i]->data_off,
                                        length,
                                        (uintptr_t)srcs[i],
-                                       (uintptr_t)dsts[i],
-                                       0 /* nofence */) != 1) {
+                                       (uintptr_t)dsts[i]) != 1) {
                                PRINT_ERR("Error with rte_ioat_enqueue_copy for buffer %u\n",
                                                i);
                                return -1;
index ae63939515bea038ec993d9511ed90482c2b28a8..21a92901296e7e68fafe4c3ab4ad8ea55ad808e5 100644 (file)
@@ -61,18 +61,33 @@ struct rte_ioat_rawdev_config {
  *   operation has been completed and the user polls for the completion details.
  *   NOTE: If hdls_disable configuration option for the device is set, this
  *   parameter is ignored.
- * @param fence
- *   A flag parameter indicating that hardware should not begin to perform any
- *   subsequently enqueued copy operations until after this operation has
- *   completed
  * @return
  *   Number of operations enqueued, either 0 or 1
  */
 static inline int
 __rte_experimental
 rte_ioat_enqueue_copy(int dev_id, phys_addr_t src, phys_addr_t dst,
-               unsigned int length, uintptr_t src_hdl, uintptr_t dst_hdl,
-               int fence);
+               unsigned int length, uintptr_t src_hdl, uintptr_t dst_hdl);
+
+/**
+ * Add a fence to force ordering between operations
+ *
+ * This adds a fence to a sequence of operations to enforce ordering, such that
+ * all operations enqueued before the fence must be completed before operations
+ * after the fence.
+ * NOTE: Since this fence may be added as a flag to the last operation enqueued,
+ * this API may not function correctly when called immediately after an
+ * "rte_ioat_perform_ops" call i.e. before any new operations are enqueued.
+ *
+ * @param dev_id
+ *   The rawdev device id of the ioat instance
+ * @return
+ *   Number of fences enqueued, either 0 or 1
+ */
+static inline int
+__rte_experimental
+rte_ioat_fence(int dev_id);
+
 
 /**
  * Trigger hardware to begin performing enqueued operations
index b155d79c45c59cc3c74aae1ab9f10f1215db8fc4..466721a23a58e75a4bbf1d7b7e19154024398502 100644 (file)
@@ -47,8 +47,7 @@ struct rte_ioat_rawdev {
  */
 static inline int
 rte_ioat_enqueue_copy(int dev_id, phys_addr_t src, phys_addr_t dst,
-               unsigned int length, uintptr_t src_hdl, uintptr_t dst_hdl,
-               int fence)
+               unsigned int length, uintptr_t src_hdl, uintptr_t dst_hdl)
 {
        struct rte_ioat_rawdev *ioat =
                        (struct rte_ioat_rawdev *)rte_rawdevs[dev_id].dev_private;
@@ -69,7 +68,7 @@ rte_ioat_enqueue_copy(int dev_id, phys_addr_t src, phys_addr_t dst,
        desc = &ioat->desc_ring[write];
        desc->size = length;
        /* set descriptor write-back every 16th descriptor */
-       desc->u.control_raw = (uint32_t)((!!fence << 4) | (!(write & 0xF)) << 3);
+       desc->u.control_raw = (uint32_t)((!(write & 0xF)) << 3);
        desc->src_addr = src;
        desc->dest_addr = dst;
 
@@ -82,6 +81,23 @@ rte_ioat_enqueue_copy(int dev_id, phys_addr_t src, phys_addr_t dst,
        return 1;
 }
 
+/* add fence to last written descriptor */
+static inline int
+rte_ioat_fence(int dev_id)
+{
+       struct rte_ioat_rawdev *ioat =
+                       (struct rte_ioat_rawdev *)rte_rawdevs[dev_id].dev_private;
+       unsigned short write = ioat->next_write;
+       unsigned short mask = ioat->ring_size - 1;
+       struct rte_ioat_generic_hw_desc *desc;
+
+       write = (write - 1) & mask;
+       desc = &ioat->desc_ring[write];
+
+       desc->u.control.fence = 1;
+       return 0;
+}
+
 /*
  * Trigger hardware to begin performing enqueued operations
  */
index 67f75737b6cd54343f5833ec888f21b8e4f14d2e..e6d1d1236e4f7bb1e9e1a166de8ea14557e8382d 100644 (file)
@@ -361,15 +361,11 @@ ioat_enqueue_packets(struct rte_mbuf **pkts,
        for (i = 0; i < nb_rx; i++) {
                /* Perform data copy */
                ret = rte_ioat_enqueue_copy(dev_id,
-                       pkts[i]->buf_iova
-                       - addr_offset,
-                       pkts_copy[i]->buf_iova
-                       - addr_offset,
-                       rte_pktmbuf_data_len(pkts[i])
-                       + addr_offset,
+                       pkts[i]->buf_iova - addr_offset,
+                       pkts_copy[i]->buf_iova - addr_offset,
+                       rte_pktmbuf_data_len(pkts[i]) + addr_offset,
                        (uintptr_t)pkts[i],
-                       (uintptr_t)pkts_copy[i],
-                       0 /* nofence */);
+                       (uintptr_t)pkts_copy[i]);
 
                if (ret != 1)
                        break;