net/txgbe: add build and doc infrastructure
[dpdk.git] / drivers / crypto / ccp / ccp_crypto.c
index 1837c85..db3fb6e 100644 (file)
@@ -1571,7 +1571,7 @@ ccp_perform_hmac(struct rte_crypto_op *op,
                                         ccp_cryptodev_driver_id);
        addr = session->auth.pre_compute;
 
-       src_addr = rte_pktmbuf_mtophys_offset(op->sym->m_src,
+       src_addr = rte_pktmbuf_iova_offset(op->sym->m_src,
                                              op->sym->auth.data.offset);
        append_ptr = (void *)rte_pktmbuf_append(op->sym->m_src,
                                                session->auth.ctx_len);
@@ -1743,7 +1743,7 @@ ccp_perform_sha(struct rte_crypto_op *op,
                                         op->sym->session,
                                        ccp_cryptodev_driver_id);
 
-       src_addr = rte_pktmbuf_mtophys_offset(op->sym->m_src,
+       src_addr = rte_pktmbuf_iova_offset(op->sym->m_src,
                                              op->sym->auth.data.offset);
 
        append_ptr = (void *)rte_pktmbuf_append(op->sym->m_src,
@@ -1832,7 +1832,7 @@ ccp_perform_sha3_hmac(struct rte_crypto_op *op,
                                         op->sym->session,
                                        ccp_cryptodev_driver_id);
 
-       src_addr = rte_pktmbuf_mtophys_offset(op->sym->m_src,
+       src_addr = rte_pktmbuf_iova_offset(op->sym->m_src,
                                              op->sym->auth.data.offset);
        append_ptr = (uint8_t *)rte_pktmbuf_append(op->sym->m_src,
                                                session->auth.ctx_len);
@@ -1972,7 +1972,7 @@ ccp_perform_sha3(struct rte_crypto_op *op,
                                         op->sym->session,
                                        ccp_cryptodev_driver_id);
 
-       src_addr = rte_pktmbuf_mtophys_offset(op->sym->m_src,
+       src_addr = rte_pktmbuf_iova_offset(op->sym->m_src,
                                              op->sym->auth.data.offset);
        append_ptr = (uint8_t *)rte_pktmbuf_append(op->sym->m_src,
                                                session->auth.ctx_len);
@@ -2041,7 +2041,7 @@ ccp_perform_aes_cmac(struct rte_crypto_op *op,
                                        ccp_cryptodev_driver_id);
        key_addr = rte_mem_virt2phy(session->auth.key_ccp);
 
-       src_addr = rte_pktmbuf_mtophys_offset(op->sym->m_src,
+       src_addr = rte_pktmbuf_iova_offset(op->sym->m_src,
                                              op->sym->auth.data.offset);
        append_ptr = (uint8_t *)rte_pktmbuf_append(op->sym->m_src,
                                                session->auth.ctx_len);
@@ -2221,10 +2221,10 @@ ccp_perform_aes(struct rte_crypto_op *op,
 
        desc = &cmd_q->qbase_desc[cmd_q->qidx];
 
-       src_addr = rte_pktmbuf_mtophys_offset(op->sym->m_src,
+       src_addr = rte_pktmbuf_iova_offset(op->sym->m_src,
                                              op->sym->cipher.data.offset);
        if (likely(op->sym->m_dst != NULL))
-               dest_addr = rte_pktmbuf_mtophys_offset(op->sym->m_dst,
+               dest_addr = rte_pktmbuf_iova_offset(op->sym->m_dst,
                                                op->sym->cipher.data.offset);
        else
                dest_addr = src_addr;
@@ -2303,11 +2303,11 @@ ccp_perform_3des(struct rte_crypto_op *op,
                return -ENOTSUP;
        }
 
-       src_addr = rte_pktmbuf_mtophys_offset(op->sym->m_src,
+       src_addr = rte_pktmbuf_iova_offset(op->sym->m_src,
                                              op->sym->cipher.data.offset);
        if (unlikely(op->sym->m_dst != NULL))
                dest_addr =
-                       rte_pktmbuf_mtophys_offset(op->sym->m_dst,
+                       rte_pktmbuf_iova_offset(op->sym->m_dst,
                                                   op->sym->cipher.data.offset);
        else
                dest_addr = src_addr;
@@ -2385,10 +2385,10 @@ ccp_perform_aes_gcm(struct rte_crypto_op *op, struct ccp_queue *cmd_q)
        iv = rte_crypto_op_ctod_offset(op, uint8_t *, session->iv.offset);
        key_addr = session->cipher.key_phys;
 
-       src_addr = rte_pktmbuf_mtophys_offset(op->sym->m_src,
+       src_addr = rte_pktmbuf_iova_offset(op->sym->m_src,
                                              op->sym->aead.data.offset);
        if (unlikely(op->sym->m_dst != NULL))
-               dest_addr = rte_pktmbuf_mtophys_offset(op->sym->m_dst,
+               dest_addr = rte_pktmbuf_iova_offset(op->sym->m_dst,
                                                op->sym->aead.data.offset);
        else
                dest_addr = src_addr;
@@ -2680,7 +2680,9 @@ process_ops_to_enqueue(struct ccp_qp *qp,
                       struct rte_crypto_op **op,
                       struct ccp_queue *cmd_q,
                       uint16_t nb_ops,
-                      int slots_req)
+                      uint16_t total_nb_ops,
+                      int slots_req,
+                      uint16_t b_idx)
 {
        int i, result = 0;
        struct ccp_batch_info *b_info;
@@ -2701,6 +2703,7 @@ process_ops_to_enqueue(struct ccp_qp *qp,
 
        /* populate batch info necessary for dequeue */
        b_info->op_idx = 0;
+       b_info->b_idx = 0;
        b_info->lsb_buf_idx = 0;
        b_info->desccnt = 0;
        b_info->cmd_q = cmd_q;
@@ -2710,7 +2713,7 @@ process_ops_to_enqueue(struct ccp_qp *qp,
 
        b_info->head_offset = (uint32_t)(cmd_q->qbase_phys_addr + cmd_q->qidx *
                                         Q_DESC_SIZE);
-       for (i = 0; i < nb_ops; i++) {
+       for (i = b_idx; i < (nb_ops+b_idx); i++) {
                session = (struct ccp_session *)get_sym_session_private_data(
                                                 op[i]->sym->session,
                                                 ccp_cryptodev_driver_id);
@@ -2762,6 +2765,8 @@ process_ops_to_enqueue(struct ccp_qp *qp,
        }
 
        b_info->opcnt = i;
+       b_info->b_idx = b_idx;
+       b_info->total_nb_ops = total_nb_ops;
        b_info->tail_offset = (uint32_t)(cmd_q->qbase_phys_addr + cmd_q->qidx *
                                         Q_DESC_SIZE);
 
@@ -2776,7 +2781,7 @@ process_ops_to_enqueue(struct ccp_qp *qp,
        rte_ring_enqueue(qp->processed_pkts, (void *)b_info);
 
        EVP_MD_CTX_destroy(auth_ctx);
-       return i;
+       return i-b_idx;
 }
 
 static inline void ccp_auth_dq_prepare(struct rte_crypto_op *op)
@@ -2861,8 +2866,8 @@ ccp_prepare_ops(struct ccp_qp *qp,
        }
        min_ops = RTE_MIN(nb_ops, b_info->opcnt);
 
-       for (i = 0; i < min_ops; i++) {
-               op_d[i] = b_info->op[b_info->op_idx++];
+       for (i =  b_info->b_idx; i < min_ops; i++) {
+               op_d[i] = b_info->op[b_info->b_idx + b_info->op_idx++];
                session = (struct ccp_session *)get_sym_session_private_data(
                                                 op_d[i]->sym->session,
                                                ccp_cryptodev_driver_id);
@@ -2903,7 +2908,8 @@ ccp_prepare_ops(struct ccp_qp *qp,
 int
 process_ops_to_dequeue(struct ccp_qp *qp,
                       struct rte_crypto_op **op,
-                      uint16_t nb_ops)
+                      uint16_t nb_ops,
+                      uint16_t *total_nb_ops)
 {
        struct ccp_batch_info *b_info;
        uint32_t cur_head_offset;
@@ -2918,6 +2924,7 @@ process_ops_to_dequeue(struct ccp_qp *qp,
 
        if (b_info->auth_ctr == b_info->opcnt)
                goto success;
+       *total_nb_ops = b_info->total_nb_ops;
        cur_head_offset = CCP_READ_REG(b_info->cmd_q->reg_base,
                                       CMD_Q_HEAD_LO_BASE);
 
@@ -2927,7 +2934,7 @@ process_ops_to_dequeue(struct ccp_qp *qp,
                        qp->b_info = b_info;
                        return 0;
                }
-       } else {
+       } else if (b_info->tail_offset != b_info->head_offset) {
                if ((cur_head_offset >= b_info->head_offset) ||
                    (cur_head_offset < b_info->tail_offset)) {
                        qp->b_info = b_info;
@@ -2937,6 +2944,7 @@ process_ops_to_dequeue(struct ccp_qp *qp,
 
 
 success:
+       *total_nb_ops = b_info->total_nb_ops;
        nb_ops = ccp_prepare_ops(qp, op, b_info, nb_ops);
        rte_atomic64_add(&b_info->cmd_q->free_slots, b_info->desccnt);
        b_info->desccnt = 0;