git.droids-corp.org
/
dpdk.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
remove experimental tags from all symbol definitions
[dpdk.git]
/
lib
/
librte_vhost
/
vhost_crypto.c
diff --git
a/lib/librte_vhost/vhost_crypto.c
b/lib/librte_vhost/vhost_crypto.c
index
9b4b850
..
684fddc
100644
(file)
--- a/
lib/librte_vhost/vhost_crypto.c
+++ b/
lib/librte_vhost/vhost_crypto.c
@@
-46,116
+46,107
@@
((t)(uintptr_t)vhost_iova_to_vva(r->dev, r->vq, a, l, p))
static int
((t)(uintptr_t)vhost_iova_to_vva(r->dev, r->vq, a, l, p))
static int
-cipher_algo_transform(uint32_t virtio_cipher_algo)
+cipher_algo_transform(uint32_t virtio_cipher_algo,
+ enum rte_crypto_cipher_algorithm *algo)
{
{
- int ret;
-
switch (virtio_cipher_algo) {
case VIRTIO_CRYPTO_CIPHER_AES_CBC:
switch (virtio_cipher_algo) {
case VIRTIO_CRYPTO_CIPHER_AES_CBC:
-
ret
= RTE_CRYPTO_CIPHER_AES_CBC;
+
*algo
= RTE_CRYPTO_CIPHER_AES_CBC;
break;
case VIRTIO_CRYPTO_CIPHER_AES_CTR:
break;
case VIRTIO_CRYPTO_CIPHER_AES_CTR:
-
ret
= RTE_CRYPTO_CIPHER_AES_CTR;
+
*algo
= RTE_CRYPTO_CIPHER_AES_CTR;
break;
case VIRTIO_CRYPTO_CIPHER_DES_ECB:
break;
case VIRTIO_CRYPTO_CIPHER_DES_ECB:
-
ret
= -VIRTIO_CRYPTO_NOTSUPP;
+
*algo
= -VIRTIO_CRYPTO_NOTSUPP;
break;
case VIRTIO_CRYPTO_CIPHER_DES_CBC:
break;
case VIRTIO_CRYPTO_CIPHER_DES_CBC:
-
ret
= RTE_CRYPTO_CIPHER_DES_CBC;
+
*algo
= RTE_CRYPTO_CIPHER_DES_CBC;
break;
case VIRTIO_CRYPTO_CIPHER_3DES_ECB:
break;
case VIRTIO_CRYPTO_CIPHER_3DES_ECB:
-
ret
= RTE_CRYPTO_CIPHER_3DES_ECB;
+
*algo
= RTE_CRYPTO_CIPHER_3DES_ECB;
break;
case VIRTIO_CRYPTO_CIPHER_3DES_CBC:
break;
case VIRTIO_CRYPTO_CIPHER_3DES_CBC:
-
ret
= RTE_CRYPTO_CIPHER_3DES_CBC;
+
*algo
= RTE_CRYPTO_CIPHER_3DES_CBC;
break;
case VIRTIO_CRYPTO_CIPHER_3DES_CTR:
break;
case VIRTIO_CRYPTO_CIPHER_3DES_CTR:
-
ret
= RTE_CRYPTO_CIPHER_3DES_CTR;
+
*algo
= RTE_CRYPTO_CIPHER_3DES_CTR;
break;
case VIRTIO_CRYPTO_CIPHER_KASUMI_F8:
break;
case VIRTIO_CRYPTO_CIPHER_KASUMI_F8:
-
ret
= RTE_CRYPTO_CIPHER_KASUMI_F8;
+
*algo
= RTE_CRYPTO_CIPHER_KASUMI_F8;
break;
case VIRTIO_CRYPTO_CIPHER_SNOW3G_UEA2:
break;
case VIRTIO_CRYPTO_CIPHER_SNOW3G_UEA2:
-
ret
= RTE_CRYPTO_CIPHER_SNOW3G_UEA2;
+
*algo
= RTE_CRYPTO_CIPHER_SNOW3G_UEA2;
break;
case VIRTIO_CRYPTO_CIPHER_AES_F8:
break;
case VIRTIO_CRYPTO_CIPHER_AES_F8:
-
ret
= RTE_CRYPTO_CIPHER_AES_F8;
+
*algo
= RTE_CRYPTO_CIPHER_AES_F8;
break;
case VIRTIO_CRYPTO_CIPHER_AES_XTS:
break;
case VIRTIO_CRYPTO_CIPHER_AES_XTS:
-
ret
= RTE_CRYPTO_CIPHER_AES_XTS;
+
*algo
= RTE_CRYPTO_CIPHER_AES_XTS;
break;
case VIRTIO_CRYPTO_CIPHER_ZUC_EEA3:
break;
case VIRTIO_CRYPTO_CIPHER_ZUC_EEA3:
-
ret
= RTE_CRYPTO_CIPHER_ZUC_EEA3;
+
*algo
= RTE_CRYPTO_CIPHER_ZUC_EEA3;
break;
default:
break;
default:
- ret
=
-VIRTIO_CRYPTO_BADMSG;
+ ret
urn
-VIRTIO_CRYPTO_BADMSG;
break;
}
break;
}
- return
ret
;
+ return
0
;
}
static int
}
static int
-auth_algo_transform(uint32_t virtio_auth_algo)
+auth_algo_transform(uint32_t virtio_auth_algo,
+ enum rte_crypto_auth_algorithm *algo)
{
{
- int ret;
-
switch (virtio_auth_algo) {
switch (virtio_auth_algo) {
-
case VIRTIO_CRYPTO_NO_MAC:
case VIRTIO_CRYPTO_NO_MAC:
-
ret
= RTE_CRYPTO_AUTH_NULL;
+
*algo
= RTE_CRYPTO_AUTH_NULL;
break;
case VIRTIO_CRYPTO_MAC_HMAC_MD5:
break;
case VIRTIO_CRYPTO_MAC_HMAC_MD5:
-
ret
= RTE_CRYPTO_AUTH_MD5_HMAC;
+
*algo
= RTE_CRYPTO_AUTH_MD5_HMAC;
break;
case VIRTIO_CRYPTO_MAC_HMAC_SHA1:
break;
case VIRTIO_CRYPTO_MAC_HMAC_SHA1:
-
ret
= RTE_CRYPTO_AUTH_SHA1_HMAC;
+
*algo
= RTE_CRYPTO_AUTH_SHA1_HMAC;
break;
case VIRTIO_CRYPTO_MAC_HMAC_SHA_224:
break;
case VIRTIO_CRYPTO_MAC_HMAC_SHA_224:
-
ret
= RTE_CRYPTO_AUTH_SHA224_HMAC;
+
*algo
= RTE_CRYPTO_AUTH_SHA224_HMAC;
break;
case VIRTIO_CRYPTO_MAC_HMAC_SHA_256:
break;
case VIRTIO_CRYPTO_MAC_HMAC_SHA_256:
-
ret
= RTE_CRYPTO_AUTH_SHA256_HMAC;
+
*algo
= RTE_CRYPTO_AUTH_SHA256_HMAC;
break;
case VIRTIO_CRYPTO_MAC_HMAC_SHA_384:
break;
case VIRTIO_CRYPTO_MAC_HMAC_SHA_384:
-
ret
= RTE_CRYPTO_AUTH_SHA384_HMAC;
+
*algo
= RTE_CRYPTO_AUTH_SHA384_HMAC;
break;
case VIRTIO_CRYPTO_MAC_HMAC_SHA_512:
break;
case VIRTIO_CRYPTO_MAC_HMAC_SHA_512:
- ret = RTE_CRYPTO_AUTH_SHA512_HMAC;
- break;
- case VIRTIO_CRYPTO_MAC_CMAC_3DES:
- ret = -VIRTIO_CRYPTO_NOTSUPP;
+ *algo = RTE_CRYPTO_AUTH_SHA512_HMAC;
break;
case VIRTIO_CRYPTO_MAC_CMAC_AES:
break;
case VIRTIO_CRYPTO_MAC_CMAC_AES:
-
ret
= RTE_CRYPTO_AUTH_AES_CMAC;
+
*algo
= RTE_CRYPTO_AUTH_AES_CMAC;
break;
case VIRTIO_CRYPTO_MAC_KASUMI_F9:
break;
case VIRTIO_CRYPTO_MAC_KASUMI_F9:
-
ret
= RTE_CRYPTO_AUTH_KASUMI_F9;
+
*algo
= RTE_CRYPTO_AUTH_KASUMI_F9;
break;
case VIRTIO_CRYPTO_MAC_SNOW3G_UIA2:
break;
case VIRTIO_CRYPTO_MAC_SNOW3G_UIA2:
-
ret
= RTE_CRYPTO_AUTH_SNOW3G_UIA2;
+
*algo
= RTE_CRYPTO_AUTH_SNOW3G_UIA2;
break;
case VIRTIO_CRYPTO_MAC_GMAC_AES:
break;
case VIRTIO_CRYPTO_MAC_GMAC_AES:
- ret = RTE_CRYPTO_AUTH_AES_GMAC;
- break;
- case VIRTIO_CRYPTO_MAC_GMAC_TWOFISH:
- ret = -VIRTIO_CRYPTO_NOTSUPP;
+ *algo = RTE_CRYPTO_AUTH_AES_GMAC;
break;
case VIRTIO_CRYPTO_MAC_CBCMAC_AES:
break;
case VIRTIO_CRYPTO_MAC_CBCMAC_AES:
- ret = RTE_CRYPTO_AUTH_AES_CBC_MAC;
- break;
- case VIRTIO_CRYPTO_MAC_CBCMAC_KASUMI_F9:
- ret = -VIRTIO_CRYPTO_NOTSUPP;
+ *algo = RTE_CRYPTO_AUTH_AES_CBC_MAC;
break;
case VIRTIO_CRYPTO_MAC_XCBC_AES:
break;
case VIRTIO_CRYPTO_MAC_XCBC_AES:
-
ret
= RTE_CRYPTO_AUTH_AES_XCBC_MAC;
+
*algo
= RTE_CRYPTO_AUTH_AES_XCBC_MAC;
break;
break;
+ case VIRTIO_CRYPTO_MAC_CMAC_3DES:
+ case VIRTIO_CRYPTO_MAC_GMAC_TWOFISH:
+ case VIRTIO_CRYPTO_MAC_CBCMAC_KASUMI_F9:
+ return -VIRTIO_CRYPTO_NOTSUPP;
default:
default:
- ret = -VIRTIO_CRYPTO_BADMSG;
- break;
+ return -VIRTIO_CRYPTO_BADMSG;
}
}
- return
ret
;
+ return
0
;
}
static int get_iv_len(enum rte_crypto_cipher_algorithm algo)
}
static int get_iv_len(enum rte_crypto_cipher_algorithm algo)
@@
-242,12
+233,11
@@
transform_cipher_param(struct rte_crypto_sym_xform *xform,
{
int ret;
{
int ret;
- ret = cipher_algo_transform(param->cipher_algo);
+ ret = cipher_algo_transform(param->cipher_algo
, &xform->cipher.algo
);
if (unlikely(ret < 0))
return ret;
xform->type = RTE_CRYPTO_SYM_XFORM_CIPHER;
if (unlikely(ret < 0))
return ret;
xform->type = RTE_CRYPTO_SYM_XFORM_CIPHER;
- xform->cipher.algo = (enum rte_crypto_cipher_algorithm)ret;
xform->cipher.key.length = param->cipher_key_len;
if (xform->cipher.key.length > 0)
xform->cipher.key.data = param->cipher_key_buf;
xform->cipher.key.length = param->cipher_key_len;
if (xform->cipher.key.length > 0)
xform->cipher.key.data = param->cipher_key_buf;
@@
-293,11
+283,11
@@
transform_chain_param(struct rte_crypto_sym_xform *xforms,
}
/* cipher */
}
/* cipher */
- ret = cipher_algo_transform(param->cipher_algo);
+ ret = cipher_algo_transform(param->cipher_algo,
+ &xform_cipher->cipher.algo);
if (unlikely(ret < 0))
return ret;
xform_cipher->type = RTE_CRYPTO_SYM_XFORM_CIPHER;
if (unlikely(ret < 0))
return ret;
xform_cipher->type = RTE_CRYPTO_SYM_XFORM_CIPHER;
- xform_cipher->cipher.algo = (enum rte_crypto_cipher_algorithm)ret;
xform_cipher->cipher.key.length = param->cipher_key_len;
xform_cipher->cipher.key.data = param->cipher_key_buf;
ret = get_iv_len(xform_cipher->cipher.algo);
xform_cipher->cipher.key.length = param->cipher_key_len;
xform_cipher->cipher.key.data = param->cipher_key_buf;
ret = get_iv_len(xform_cipher->cipher.algo);
@@
-308,10
+298,9
@@
transform_chain_param(struct rte_crypto_sym_xform *xforms,
/* auth */
xform_auth->type = RTE_CRYPTO_SYM_XFORM_AUTH;
/* auth */
xform_auth->type = RTE_CRYPTO_SYM_XFORM_AUTH;
- ret = auth_algo_transform(param->hash_algo);
+ ret = auth_algo_transform(param->hash_algo
, &xform_auth->auth.algo
);
if (unlikely(ret < 0))
return ret;
if (unlikely(ret < 0))
return ret;
- xform_auth->auth.algo = (enum rte_crypto_auth_algorithm)ret;
xform_auth->auth.digest_length = param->digest_len;
xform_auth->auth.key.length = param->auth_key_len;
xform_auth->auth.key.data = param->auth_key_buf;
xform_auth->auth.digest_length = param->digest_len;
xform_auth->auth.key.length = param->auth_key_len;
xform_auth->auth.key.data = param->auth_key_buf;
@@
-1024,7
+1013,7
@@
prepare_sym_chain_op(struct vhost_crypto *vcrypto, struct rte_crypto_op *op,
}
if (unlikely(copy_data(rte_pktmbuf_mtod(m_src, uint8_t *),
vc_req, &desc, chain->para.src_data_len,
}
if (unlikely(copy_data(rte_pktmbuf_mtod(m_src, uint8_t *),
vc_req, &desc, chain->para.src_data_len,
- nb_descs, vq_size)
) < 0
) {
+ nb_descs, vq_size)
< 0)
) {
ret = VIRTIO_CRYPTO_BADMSG;
goto error_exit;
}
ret = VIRTIO_CRYPTO_BADMSG;
goto error_exit;
}
@@
-1109,7
+1098,7
@@
prepare_sym_chain_op(struct vhost_crypto *vcrypto, struct rte_crypto_op *op,
if (unlikely(copy_data(digest_addr, vc_req, &digest_desc,
chain->para.hash_result_len,
if (unlikely(copy_data(digest_addr, vc_req, &digest_desc,
chain->para.hash_result_len,
- nb_descs, vq_size)
) < 0
) {
+ nb_descs, vq_size)
< 0)
) {
ret = VIRTIO_CRYPTO_BADMSG;
goto error_exit;
}
ret = VIRTIO_CRYPTO_BADMSG;
goto error_exit;
}
@@
-1358,7
+1347,7
@@
vhost_crypto_complete_one_vm_requests(struct rte_crypto_op **ops,
return processed;
}
return processed;
}
-int
__rte_experimental
+int
rte_vhost_crypto_create(int vid, uint8_t cryptodev_id,
struct rte_mempool *sess_pool,
struct rte_mempool *sess_priv_pool,
rte_vhost_crypto_create(int vid, uint8_t cryptodev_id,
struct rte_mempool *sess_pool,
struct rte_mempool *sess_priv_pool,
@@
-1451,7
+1440,7
@@
error_exit:
return ret;
}
return ret;
}
-int
__rte_experimental
+int
rte_vhost_crypto_free(int vid)
{
struct virtio_net *dev = get_device(vid);
rte_vhost_crypto_free(int vid)
{
struct virtio_net *dev = get_device(vid);
@@
-1480,7
+1469,7
@@
rte_vhost_crypto_free(int vid)
return 0;
}
return 0;
}
-int
__rte_experimental
+int
rte_vhost_crypto_set_zero_copy(int vid, enum rte_vhost_crypto_zero_copy option)
{
struct virtio_net *dev = get_device(vid);
rte_vhost_crypto_set_zero_copy(int vid, enum rte_vhost_crypto_zero_copy option)
{
struct virtio_net *dev = get_device(vid);
@@
-1535,7
+1524,7
@@
rte_vhost_crypto_set_zero_copy(int vid, enum rte_vhost_crypto_zero_copy option)
return 0;
}
return 0;
}
-uint16_t
__rte_experimental
+uint16_t
rte_vhost_crypto_fetch_requests(int vid, uint32_t qid,
struct rte_crypto_op **ops, uint16_t nb_ops)
{
rte_vhost_crypto_fetch_requests(int vid, uint32_t qid,
struct rte_crypto_op **ops, uint16_t nb_ops)
{
@@
-1598,7
+1587,7
@@
rte_vhost_crypto_fetch_requests(int vid, uint32_t qid,
op->sym->m_dst->data_off = 0;
if (unlikely(vhost_crypto_process_one_req(vcrypto, vq,
op->sym->m_dst->data_off = 0;
if (unlikely(vhost_crypto_process_one_req(vcrypto, vq,
- op, head, desc_idx)
) < 0
)
+ op, head, desc_idx)
< 0)
)
break;
}
break;
}
@@
-1627,7
+1616,7
@@
rte_vhost_crypto_fetch_requests(int vid, uint32_t qid,
op->sym->m_src->data_off = 0;
if (unlikely(vhost_crypto_process_one_req(vcrypto, vq,
op->sym->m_src->data_off = 0;
if (unlikely(vhost_crypto_process_one_req(vcrypto, vq,
- op, head, desc_idx)
) < 0
)
+ op, head, desc_idx)
< 0)
)
break;
}
break;
}
@@
-1645,7
+1634,7
@@
rte_vhost_crypto_fetch_requests(int vid, uint32_t qid,
return i;
}
return i;
}
-uint16_t
__rte_experimental
+uint16_t
rte_vhost_crypto_finalize_requests(struct rte_crypto_op **ops,
uint16_t nb_ops, int *callfds, uint16_t *nb_callfds)
{
rte_vhost_crypto_finalize_requests(struct rte_crypto_op **ops,
uint16_t nb_ops, int *callfds, uint16_t *nb_callfds)
{