cryptodev: move IV parameters to session
[dpdk.git] / drivers / crypto / aesni_gcm / aesni_gcm_pmd.c
index 101ef98..f775829 100644 (file)
@@ -1,7 +1,7 @@
 /*-
  *   BSD LICENSE
  *
- *   Copyright(c) 2016 Intel Corporation. All rights reserved.
+ *   Copyright(c) 2016-2017 Intel Corporation. All rights reserved.
  *
  *   Redistribution and use in source and binary forms, with or without
  *   modification, are permitted provided that the following conditions
@@ -35,6 +35,7 @@
 #include <rte_hexdump.h>
 #include <rte_cryptodev.h>
 #include <rte_cryptodev_pmd.h>
+#include <rte_cryptodev_vdev.h>
 #include <rte_vdev.h>
 #include <rte_malloc.h>
 #include <rte_cpuflags.h>
@@ -103,6 +104,17 @@ aesni_gcm_set_session_parameters(struct aesni_gcm_session *sess,
                return -EINVAL;
        }
 
+       /* Set IV parameters */
+       sess->iv.offset = cipher_xform->cipher.iv.offset;
+       sess->iv.length = cipher_xform->cipher.iv.length;
+
+       /* IV check */
+       if (sess->iv.length != 16 && sess->iv.length != 12 &&
+                       sess->iv.length != 0) {
+               GCM_LOG_ERR("Wrong IV length");
+               return -EINVAL;
+       }
+
        /* Select Crypto operation */
        if (cipher_xform->cipher.op == RTE_CRYPTO_CIPHER_OP_ENCRYPT &&
                        auth_xform->auth.op == RTE_CRYPTO_AUTH_OP_GENERATE)
@@ -138,16 +150,17 @@ aesni_gcm_set_session_parameters(struct aesni_gcm_session *sess,
 
 /** Get gcm session */
 static struct aesni_gcm_session *
-aesni_gcm_get_session(struct aesni_gcm_qp *qp, struct rte_crypto_sym_op *op)
+aesni_gcm_get_session(struct aesni_gcm_qp *qp, struct rte_crypto_op *op)
 {
        struct aesni_gcm_session *sess = NULL;
+       struct rte_crypto_sym_op *sym_op = op->sym;
 
-       if (op->sess_type == RTE_CRYPTO_SYM_OP_WITH_SESSION) {
-               if (unlikely(op->session->dev_type
+       if (op->sess_type == RTE_CRYPTO_OP_WITH_SESSION) {
+               if (unlikely(sym_op->session->dev_type
                                        != RTE_CRYPTODEV_AESNI_GCM_PMD))
                        return sess;
 
-               sess = (struct aesni_gcm_session *)op->session->_private;
+               sess = (struct aesni_gcm_session *)sym_op->session->_private;
        } else  {
                void *_sess;
 
@@ -158,7 +171,7 @@ aesni_gcm_get_session(struct aesni_gcm_qp *qp, struct rte_crypto_sym_op *op)
                        ((struct rte_cryptodev_sym_session *)_sess)->_private;
 
                if (unlikely(aesni_gcm_set_session_parameters(sess,
-                               op->xform) != 0)) {
+                               sym_op->xform) != 0)) {
                        rte_mempool_put(qp->sess_mp, _sess);
                        sess = NULL;
                }
@@ -178,12 +191,14 @@ aesni_gcm_get_session(struct aesni_gcm_qp *qp, struct rte_crypto_sym_op *op)
  *
  */
 static int
-process_gcm_crypto_op(struct rte_crypto_sym_op *op,
+process_gcm_crypto_op(struct rte_crypto_op *op,
                struct aesni_gcm_session *session)
 {
        uint8_t *src, *dst;
-       struct rte_mbuf *m_src = op->m_src;
-       uint32_t offset = op->cipher.data.offset;
+       uint8_t *iv_ptr;
+       struct rte_crypto_sym_op *sym_op = op->sym;
+       struct rte_mbuf *m_src = sym_op->m_src;
+       uint32_t offset = sym_op->cipher.data.offset;
        uint32_t part_len, total_len, data_len;
 
        RTE_ASSERT(m_src != NULL);
@@ -196,46 +211,41 @@ process_gcm_crypto_op(struct rte_crypto_sym_op *op,
        }
 
        data_len = m_src->data_len - offset;
-       part_len = (data_len < op->cipher.data.length) ? data_len :
-                       op->cipher.data.length;
+       part_len = (data_len < sym_op->cipher.data.length) ? data_len :
+                       sym_op->cipher.data.length;
 
        /* Destination buffer is required when segmented source buffer */
-       RTE_ASSERT((part_len == op->cipher.data.length) ||
-                       ((part_len != op->cipher.data.length) &&
-                                       (op->m_dst != NULL)));
+       RTE_ASSERT((part_len == sym_op->cipher.data.length) ||
+                       ((part_len != sym_op->cipher.data.length) &&
+                                       (sym_op->m_dst != NULL)));
        /* Segmented destination buffer is not supported */
-       RTE_ASSERT((op->m_dst == NULL) ||
-                       ((op->m_dst != NULL) &&
-                                       rte_pktmbuf_is_contiguous(op->m_dst)));
+       RTE_ASSERT((sym_op->m_dst == NULL) ||
+                       ((sym_op->m_dst != NULL) &&
+                                       rte_pktmbuf_is_contiguous(sym_op->m_dst)));
 
 
-       dst = op->m_dst ?
-                       rte_pktmbuf_mtod_offset(op->m_dst, uint8_t *,
-                                       op->cipher.data.offset) :
-                       rte_pktmbuf_mtod_offset(op->m_src, uint8_t *,
-                                       op->cipher.data.offset);
+       dst = sym_op->m_dst ?
+                       rte_pktmbuf_mtod_offset(sym_op->m_dst, uint8_t *,
+                                       sym_op->cipher.data.offset) :
+                       rte_pktmbuf_mtod_offset(sym_op->m_src, uint8_t *,
+                                       sym_op->cipher.data.offset);
 
        src = rte_pktmbuf_mtod_offset(m_src, uint8_t *, offset);
 
-       /* sanity checks */
-       if (op->cipher.iv.length != 16 && op->cipher.iv.length != 12 &&
-                       op->cipher.iv.length != 0) {
-               GCM_LOG_ERR("iv");
-               return -1;
-       }
-
+       iv_ptr = rte_crypto_op_ctod_offset(op, uint8_t *,
+                               session->iv.offset);
        /*
         * GCM working in 12B IV mode => 16B pre-counter block we need
         * to set BE LSB to 1, driver expects that 16B is allocated
         */
-       if (op->cipher.iv.length == 12) {
-               uint32_t *iv_padd = (uint32_t *)&op->cipher.iv.data[12];
+       if (session->iv.length == 12) {
+               uint32_t *iv_padd = (uint32_t *)&(iv_ptr[12]);
                *iv_padd = rte_bswap32(1);
        }
 
-       if (op->auth.digest.length != 16 &&
-                       op->auth.digest.length != 12 &&
-                       op->auth.digest.length != 8) {
+       if (sym_op->auth.digest.length != 16 &&
+                       sym_op->auth.digest.length != 12 &&
+                       sym_op->auth.digest.length != 8) {
                GCM_LOG_ERR("digest");
                return -1;
        }
@@ -243,13 +253,13 @@ process_gcm_crypto_op(struct rte_crypto_sym_op *op,
        if (session->op == AESNI_GCM_OP_AUTHENTICATED_ENCRYPTION) {
 
                aesni_gcm_enc[session->key].init(&session->gdata,
-                               op->cipher.iv.data,
-                               op->auth.aad.data,
-                               (uint64_t)op->auth.aad.length);
+                               iv_ptr,
+                               sym_op->auth.aad.data,
+                               (uint64_t)sym_op->auth.aad.length);
 
                aesni_gcm_enc[session->key].update(&session->gdata, dst, src,
                                (uint64_t)part_len);
-               total_len = op->cipher.data.length - part_len;
+               total_len = sym_op->cipher.data.length - part_len;
 
                while (total_len) {
                        dst += part_len;
@@ -268,12 +278,12 @@ process_gcm_crypto_op(struct rte_crypto_sym_op *op,
                }
 
                aesni_gcm_enc[session->key].finalize(&session->gdata,
-                               op->auth.digest.data,
-                               (uint64_t)op->auth.digest.length);
+                               sym_op->auth.digest.data,
+                               (uint64_t)sym_op->auth.digest.length);
        } else { /* session->op == AESNI_GCM_OP_AUTHENTICATED_DECRYPTION */
-               uint8_t *auth_tag = (uint8_t *)rte_pktmbuf_append(op->m_dst ?
-                               op->m_dst : op->m_src,
-                               op->auth.digest.length);
+               uint8_t *auth_tag = (uint8_t *)rte_pktmbuf_append(sym_op->m_dst ?
+                               sym_op->m_dst : sym_op->m_src,
+                               sym_op->auth.digest.length);
 
                if (!auth_tag) {
                        GCM_LOG_ERR("auth_tag");
@@ -281,13 +291,13 @@ process_gcm_crypto_op(struct rte_crypto_sym_op *op,
                }
 
                aesni_gcm_dec[session->key].init(&session->gdata,
-                               op->cipher.iv.data,
-                               op->auth.aad.data,
-                               (uint64_t)op->auth.aad.length);
+                               iv_ptr,
+                               sym_op->auth.aad.data,
+                               (uint64_t)sym_op->auth.aad.length);
 
                aesni_gcm_dec[session->key].update(&session->gdata, dst, src,
                                (uint64_t)part_len);
-               total_len = op->cipher.data.length - part_len;
+               total_len = sym_op->cipher.data.length - part_len;
 
                while (total_len) {
                        dst += part_len;
@@ -307,7 +317,7 @@ process_gcm_crypto_op(struct rte_crypto_sym_op *op,
 
                aesni_gcm_dec[session->key].finalize(&session->gdata,
                                auth_tag,
-                               (uint64_t)op->auth.digest.length);
+                               (uint64_t)sym_op->auth.digest.length);
        }
 
        return 0;
@@ -371,7 +381,7 @@ handle_completed_gcm_crypto_op(struct aesni_gcm_qp *qp,
        post_process_gcm_crypto_op(op);
 
        /* Free session if a session-less crypto op */
-       if (op->sym->sess_type == RTE_CRYPTO_SYM_OP_SESSIONLESS) {
+       if (op->sess_type == RTE_CRYPTO_OP_SESSIONLESS) {
                rte_mempool_put(qp->sess_mp, op->sym->session);
                op->sym->session = NULL;
        }
@@ -392,14 +402,14 @@ aesni_gcm_pmd_dequeue_burst(void *queue_pair,
 
        for (i = 0; i < nb_dequeued; i++) {
 
-               sess = aesni_gcm_get_session(qp, ops[i]->sym);
+               sess = aesni_gcm_get_session(qp, ops[i]);
                if (unlikely(sess == NULL)) {
                        ops[i]->status = RTE_CRYPTO_OP_STATUS_INVALID_ARGS;
                        qp->qp_stats.dequeue_err_count++;
                        break;
                }
 
-               retval = process_gcm_crypto_op(ops[i]->sym, sess);
+               retval = process_gcm_crypto_op(ops[i], sess);
                if (retval < 0) {
                        ops[i]->status = RTE_CRYPTO_OP_STATUS_INVALID_ARGS;
                        qp->qp_stats.dequeue_err_count++;
@@ -449,8 +459,9 @@ aesni_gcm_create(const char *name,
                return -EFAULT;
        }
 
-       dev = rte_cryptodev_pmd_virtual_dev_init(init_params->name,
-                       sizeof(struct aesni_gcm_private), init_params->socket_id);
+       dev = rte_cryptodev_vdev_pmd_init(init_params->name,
+                       sizeof(struct aesni_gcm_private), init_params->socket_id,
+                       vdev);
        if (dev == NULL) {
                GCM_LOG_ERR("failed to create cryptodev vdev");
                goto init_error;
@@ -498,7 +509,7 @@ aesni_gcm_probe(struct rte_vdev_device *vdev)
        if (name == NULL)
                return -EINVAL;
        input_args = rte_vdev_device_args(vdev);
-       rte_cryptodev_parse_vdev_init_params(&init_params, input_args);
+       rte_cryptodev_vdev_parse_init_params(&init_params, input_args);
 
        RTE_LOG(INFO, PMD, "Initialising %s on NUMA node %d\n", name,
                        init_params.socket_id);