git.droids-corp.org
/
dpdk.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
net/ice/base: fix uninitialized stack variables
[dpdk.git]
/
lib
/
librte_vhost
/
vhost_crypto.c
diff --git
a/lib/librte_vhost/vhost_crypto.c
b/lib/librte_vhost/vhost_crypto.c
index
069b561
..
6891197
100644
(file)
--- a/
lib/librte_vhost/vhost_crypto.c
+++ b/
lib/librte_vhost/vhost_crypto.c
@@
-46,116
+46,107
@@
((t)(uintptr_t)vhost_iova_to_vva(r->dev, r->vq, a, l, p))
static int
((t)(uintptr_t)vhost_iova_to_vva(r->dev, r->vq, a, l, p))
static int
-cipher_algo_transform(uint32_t virtio_cipher_algo)
+cipher_algo_transform(uint32_t virtio_cipher_algo,
+ enum rte_crypto_cipher_algorithm *algo)
{
{
- int ret;
-
switch (virtio_cipher_algo) {
case VIRTIO_CRYPTO_CIPHER_AES_CBC:
switch (virtio_cipher_algo) {
case VIRTIO_CRYPTO_CIPHER_AES_CBC:
-
ret
= RTE_CRYPTO_CIPHER_AES_CBC;
+
*algo
= RTE_CRYPTO_CIPHER_AES_CBC;
break;
case VIRTIO_CRYPTO_CIPHER_AES_CTR:
break;
case VIRTIO_CRYPTO_CIPHER_AES_CTR:
-
ret
= RTE_CRYPTO_CIPHER_AES_CTR;
+
*algo
= RTE_CRYPTO_CIPHER_AES_CTR;
break;
case VIRTIO_CRYPTO_CIPHER_DES_ECB:
break;
case VIRTIO_CRYPTO_CIPHER_DES_ECB:
-
ret
= -VIRTIO_CRYPTO_NOTSUPP;
+
*algo
= -VIRTIO_CRYPTO_NOTSUPP;
break;
case VIRTIO_CRYPTO_CIPHER_DES_CBC:
break;
case VIRTIO_CRYPTO_CIPHER_DES_CBC:
-
ret
= RTE_CRYPTO_CIPHER_DES_CBC;
+
*algo
= RTE_CRYPTO_CIPHER_DES_CBC;
break;
case VIRTIO_CRYPTO_CIPHER_3DES_ECB:
break;
case VIRTIO_CRYPTO_CIPHER_3DES_ECB:
-
ret
= RTE_CRYPTO_CIPHER_3DES_ECB;
+
*algo
= RTE_CRYPTO_CIPHER_3DES_ECB;
break;
case VIRTIO_CRYPTO_CIPHER_3DES_CBC:
break;
case VIRTIO_CRYPTO_CIPHER_3DES_CBC:
-
ret
= RTE_CRYPTO_CIPHER_3DES_CBC;
+
*algo
= RTE_CRYPTO_CIPHER_3DES_CBC;
break;
case VIRTIO_CRYPTO_CIPHER_3DES_CTR:
break;
case VIRTIO_CRYPTO_CIPHER_3DES_CTR:
-
ret
= RTE_CRYPTO_CIPHER_3DES_CTR;
+
*algo
= RTE_CRYPTO_CIPHER_3DES_CTR;
break;
case VIRTIO_CRYPTO_CIPHER_KASUMI_F8:
break;
case VIRTIO_CRYPTO_CIPHER_KASUMI_F8:
-
ret
= RTE_CRYPTO_CIPHER_KASUMI_F8;
+
*algo
= RTE_CRYPTO_CIPHER_KASUMI_F8;
break;
case VIRTIO_CRYPTO_CIPHER_SNOW3G_UEA2:
break;
case VIRTIO_CRYPTO_CIPHER_SNOW3G_UEA2:
-
ret
= RTE_CRYPTO_CIPHER_SNOW3G_UEA2;
+
*algo
= RTE_CRYPTO_CIPHER_SNOW3G_UEA2;
break;
case VIRTIO_CRYPTO_CIPHER_AES_F8:
break;
case VIRTIO_CRYPTO_CIPHER_AES_F8:
-
ret
= RTE_CRYPTO_CIPHER_AES_F8;
+
*algo
= RTE_CRYPTO_CIPHER_AES_F8;
break;
case VIRTIO_CRYPTO_CIPHER_AES_XTS:
break;
case VIRTIO_CRYPTO_CIPHER_AES_XTS:
-
ret
= RTE_CRYPTO_CIPHER_AES_XTS;
+
*algo
= RTE_CRYPTO_CIPHER_AES_XTS;
break;
case VIRTIO_CRYPTO_CIPHER_ZUC_EEA3:
break;
case VIRTIO_CRYPTO_CIPHER_ZUC_EEA3:
-
ret
= RTE_CRYPTO_CIPHER_ZUC_EEA3;
+
*algo
= RTE_CRYPTO_CIPHER_ZUC_EEA3;
break;
default:
break;
default:
- ret
=
-VIRTIO_CRYPTO_BADMSG;
+ ret
urn
-VIRTIO_CRYPTO_BADMSG;
break;
}
break;
}
- return
ret
;
+ return
0
;
}
static int
}
static int
-auth_algo_transform(uint32_t virtio_auth_algo)
+auth_algo_transform(uint32_t virtio_auth_algo,
+ enum rte_crypto_auth_algorithm *algo)
{
{
- int ret;
-
switch (virtio_auth_algo) {
switch (virtio_auth_algo) {
-
case VIRTIO_CRYPTO_NO_MAC:
case VIRTIO_CRYPTO_NO_MAC:
-
ret
= RTE_CRYPTO_AUTH_NULL;
+
*algo
= RTE_CRYPTO_AUTH_NULL;
break;
case VIRTIO_CRYPTO_MAC_HMAC_MD5:
break;
case VIRTIO_CRYPTO_MAC_HMAC_MD5:
-
ret
= RTE_CRYPTO_AUTH_MD5_HMAC;
+
*algo
= RTE_CRYPTO_AUTH_MD5_HMAC;
break;
case VIRTIO_CRYPTO_MAC_HMAC_SHA1:
break;
case VIRTIO_CRYPTO_MAC_HMAC_SHA1:
-
ret
= RTE_CRYPTO_AUTH_SHA1_HMAC;
+
*algo
= RTE_CRYPTO_AUTH_SHA1_HMAC;
break;
case VIRTIO_CRYPTO_MAC_HMAC_SHA_224:
break;
case VIRTIO_CRYPTO_MAC_HMAC_SHA_224:
-
ret
= RTE_CRYPTO_AUTH_SHA224_HMAC;
+
*algo
= RTE_CRYPTO_AUTH_SHA224_HMAC;
break;
case VIRTIO_CRYPTO_MAC_HMAC_SHA_256:
break;
case VIRTIO_CRYPTO_MAC_HMAC_SHA_256:
-
ret
= RTE_CRYPTO_AUTH_SHA256_HMAC;
+
*algo
= RTE_CRYPTO_AUTH_SHA256_HMAC;
break;
case VIRTIO_CRYPTO_MAC_HMAC_SHA_384:
break;
case VIRTIO_CRYPTO_MAC_HMAC_SHA_384:
-
ret
= RTE_CRYPTO_AUTH_SHA384_HMAC;
+
*algo
= RTE_CRYPTO_AUTH_SHA384_HMAC;
break;
case VIRTIO_CRYPTO_MAC_HMAC_SHA_512:
break;
case VIRTIO_CRYPTO_MAC_HMAC_SHA_512:
- ret = RTE_CRYPTO_AUTH_SHA512_HMAC;
- break;
- case VIRTIO_CRYPTO_MAC_CMAC_3DES:
- ret = -VIRTIO_CRYPTO_NOTSUPP;
+ *algo = RTE_CRYPTO_AUTH_SHA512_HMAC;
break;
case VIRTIO_CRYPTO_MAC_CMAC_AES:
break;
case VIRTIO_CRYPTO_MAC_CMAC_AES:
-
ret
= RTE_CRYPTO_AUTH_AES_CMAC;
+
*algo
= RTE_CRYPTO_AUTH_AES_CMAC;
break;
case VIRTIO_CRYPTO_MAC_KASUMI_F9:
break;
case VIRTIO_CRYPTO_MAC_KASUMI_F9:
-
ret
= RTE_CRYPTO_AUTH_KASUMI_F9;
+
*algo
= RTE_CRYPTO_AUTH_KASUMI_F9;
break;
case VIRTIO_CRYPTO_MAC_SNOW3G_UIA2:
break;
case VIRTIO_CRYPTO_MAC_SNOW3G_UIA2:
-
ret
= RTE_CRYPTO_AUTH_SNOW3G_UIA2;
+
*algo
= RTE_CRYPTO_AUTH_SNOW3G_UIA2;
break;
case VIRTIO_CRYPTO_MAC_GMAC_AES:
break;
case VIRTIO_CRYPTO_MAC_GMAC_AES:
- ret = RTE_CRYPTO_AUTH_AES_GMAC;
- break;
- case VIRTIO_CRYPTO_MAC_GMAC_TWOFISH:
- ret = -VIRTIO_CRYPTO_NOTSUPP;
+ *algo = RTE_CRYPTO_AUTH_AES_GMAC;
break;
case VIRTIO_CRYPTO_MAC_CBCMAC_AES:
break;
case VIRTIO_CRYPTO_MAC_CBCMAC_AES:
- ret = RTE_CRYPTO_AUTH_AES_CBC_MAC;
- break;
- case VIRTIO_CRYPTO_MAC_CBCMAC_KASUMI_F9:
- ret = -VIRTIO_CRYPTO_NOTSUPP;
+ *algo = RTE_CRYPTO_AUTH_AES_CBC_MAC;
break;
case VIRTIO_CRYPTO_MAC_XCBC_AES:
break;
case VIRTIO_CRYPTO_MAC_XCBC_AES:
-
ret
= RTE_CRYPTO_AUTH_AES_XCBC_MAC;
+
*algo
= RTE_CRYPTO_AUTH_AES_XCBC_MAC;
break;
break;
+ case VIRTIO_CRYPTO_MAC_CMAC_3DES:
+ case VIRTIO_CRYPTO_MAC_GMAC_TWOFISH:
+ case VIRTIO_CRYPTO_MAC_CBCMAC_KASUMI_F9:
+ return -VIRTIO_CRYPTO_NOTSUPP;
default:
default:
- ret = -VIRTIO_CRYPTO_BADMSG;
- break;
+ return -VIRTIO_CRYPTO_BADMSG;
}
}
- return
ret
;
+ return
0
;
}
static int get_iv_len(enum rte_crypto_cipher_algorithm algo)
}
static int get_iv_len(enum rte_crypto_cipher_algorithm algo)
@@
-242,12
+233,11
@@
transform_cipher_param(struct rte_crypto_sym_xform *xform,
{
int ret;
{
int ret;
- ret = cipher_algo_transform(param->cipher_algo);
+ ret = cipher_algo_transform(param->cipher_algo
, &xform->cipher.algo
);
if (unlikely(ret < 0))
return ret;
xform->type = RTE_CRYPTO_SYM_XFORM_CIPHER;
if (unlikely(ret < 0))
return ret;
xform->type = RTE_CRYPTO_SYM_XFORM_CIPHER;
- xform->cipher.algo = (enum rte_crypto_cipher_algorithm)ret;
xform->cipher.key.length = param->cipher_key_len;
if (xform->cipher.key.length > 0)
xform->cipher.key.data = param->cipher_key_buf;
xform->cipher.key.length = param->cipher_key_len;
if (xform->cipher.key.length > 0)
xform->cipher.key.data = param->cipher_key_buf;
@@
-293,11
+283,11
@@
transform_chain_param(struct rte_crypto_sym_xform *xforms,
}
/* cipher */
}
/* cipher */
- ret = cipher_algo_transform(param->cipher_algo);
+ ret = cipher_algo_transform(param->cipher_algo,
+ &xform_cipher->cipher.algo);
if (unlikely(ret < 0))
return ret;
xform_cipher->type = RTE_CRYPTO_SYM_XFORM_CIPHER;
if (unlikely(ret < 0))
return ret;
xform_cipher->type = RTE_CRYPTO_SYM_XFORM_CIPHER;
- xform_cipher->cipher.algo = (enum rte_crypto_cipher_algorithm)ret;
xform_cipher->cipher.key.length = param->cipher_key_len;
xform_cipher->cipher.key.data = param->cipher_key_buf;
ret = get_iv_len(xform_cipher->cipher.algo);
xform_cipher->cipher.key.length = param->cipher_key_len;
xform_cipher->cipher.key.data = param->cipher_key_buf;
ret = get_iv_len(xform_cipher->cipher.algo);
@@
-308,10
+298,9
@@
transform_chain_param(struct rte_crypto_sym_xform *xforms,
/* auth */
xform_auth->type = RTE_CRYPTO_SYM_XFORM_AUTH;
/* auth */
xform_auth->type = RTE_CRYPTO_SYM_XFORM_AUTH;
- ret = auth_algo_transform(param->hash_algo);
+ ret = auth_algo_transform(param->hash_algo
, &xform_auth->auth.algo
);
if (unlikely(ret < 0))
return ret;
if (unlikely(ret < 0))
return ret;
- xform_auth->auth.algo = (enum rte_crypto_auth_algorithm)ret;
xform_auth->auth.digest_length = param->digest_len;
xform_auth->auth.key.length = param->auth_key_len;
xform_auth->auth.key.data = param->auth_key_buf;
xform_auth->auth.digest_length = param->digest_len;
xform_auth->auth.key.length = param->auth_key_len;
xform_auth->auth.key.data = param->auth_key_buf;
@@
-1358,7
+1347,7
@@
vhost_crypto_complete_one_vm_requests(struct rte_crypto_op **ops,
return processed;
}
return processed;
}
-int
__rte_experimental
+int
rte_vhost_crypto_create(int vid, uint8_t cryptodev_id,
struct rte_mempool *sess_pool,
struct rte_mempool *sess_priv_pool,
rte_vhost_crypto_create(int vid, uint8_t cryptodev_id,
struct rte_mempool *sess_pool,
struct rte_mempool *sess_priv_pool,
@@
-1451,7
+1440,7
@@
error_exit:
return ret;
}
return ret;
}
-int
__rte_experimental
+int
rte_vhost_crypto_free(int vid)
{
struct virtio_net *dev = get_device(vid);
rte_vhost_crypto_free(int vid)
{
struct virtio_net *dev = get_device(vid);
@@
-1480,7
+1469,7
@@
rte_vhost_crypto_free(int vid)
return 0;
}
return 0;
}
-int
__rte_experimental
+int
rte_vhost_crypto_set_zero_copy(int vid, enum rte_vhost_crypto_zero_copy option)
{
struct virtio_net *dev = get_device(vid);
rte_vhost_crypto_set_zero_copy(int vid, enum rte_vhost_crypto_zero_copy option)
{
struct virtio_net *dev = get_device(vid);
@@
-1535,7
+1524,7
@@
rte_vhost_crypto_set_zero_copy(int vid, enum rte_vhost_crypto_zero_copy option)
return 0;
}
return 0;
}
-uint16_t
__rte_experimental
+uint16_t
rte_vhost_crypto_fetch_requests(int vid, uint32_t qid,
struct rte_crypto_op **ops, uint16_t nb_ops)
{
rte_vhost_crypto_fetch_requests(int vid, uint32_t qid,
struct rte_crypto_op **ops, uint16_t nb_ops)
{
@@
-1550,18
+1539,18
@@
rte_vhost_crypto_fetch_requests(int vid, uint32_t qid,
if (unlikely(dev == NULL)) {
VC_LOG_ERR("Invalid vid %i", vid);
if (unlikely(dev == NULL)) {
VC_LOG_ERR("Invalid vid %i", vid);
- return
-EINVAL
;
+ return
0
;
}
if (unlikely(qid >= VHOST_MAX_QUEUE_PAIRS)) {
VC_LOG_ERR("Invalid qid %u", qid);
}
if (unlikely(qid >= VHOST_MAX_QUEUE_PAIRS)) {
VC_LOG_ERR("Invalid qid %u", qid);
- return
-EINVAL
;
+ return
0
;
}
vcrypto = (struct vhost_crypto *)dev->extern_data;
if (unlikely(vcrypto == NULL)) {
VC_LOG_ERR("Cannot find required data, is it initialized?");
}
vcrypto = (struct vhost_crypto *)dev->extern_data;
if (unlikely(vcrypto == NULL)) {
VC_LOG_ERR("Cannot find required data, is it initialized?");
- return
-ENOENT
;
+ return
0
;
}
vq = dev->virtqueue[qid];
}
vq = dev->virtqueue[qid];
@@
-1583,7
+1572,7
@@
rte_vhost_crypto_fetch_requests(int vid, uint32_t qid,
if (unlikely(rte_mempool_get_bulk(vcrypto->mbuf_pool,
(void **)mbufs, count * 2) < 0)) {
VC_LOG_ERR("Insufficient memory");
if (unlikely(rte_mempool_get_bulk(vcrypto->mbuf_pool,
(void **)mbufs, count * 2) < 0)) {
VC_LOG_ERR("Insufficient memory");
- return
-ENOMEM
;
+ return
0
;
}
for (i = 0; i < count; i++) {
}
for (i = 0; i < count; i++) {
@@
-1613,7
+1602,7
@@
rte_vhost_crypto_fetch_requests(int vid, uint32_t qid,
if (unlikely(rte_mempool_get_bulk(vcrypto->mbuf_pool,
(void **)mbufs, count) < 0)) {
VC_LOG_ERR("Insufficient memory");
if (unlikely(rte_mempool_get_bulk(vcrypto->mbuf_pool,
(void **)mbufs, count) < 0)) {
VC_LOG_ERR("Insufficient memory");
- return
-ENOMEM
;
+ return
0
;
}
for (i = 0; i < count; i++) {
}
for (i = 0; i < count; i++) {
@@
-1645,7
+1634,7
@@
rte_vhost_crypto_fetch_requests(int vid, uint32_t qid,
return i;
}
return i;
}
-uint16_t
__rte_experimental
+uint16_t
rte_vhost_crypto_finalize_requests(struct rte_crypto_op **ops,
uint16_t nb_ops, int *callfds, uint16_t *nb_callfds)
{
rte_vhost_crypto_finalize_requests(struct rte_crypto_op **ops,
uint16_t nb_ops, int *callfds, uint16_t *nb_callfds)
{