git.droids-corp.org
/
dpdk.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
doc: announce removal of indirect mbuf check macro
[dpdk.git]
/
drivers
/
crypto
/
dpaa_sec
/
dpaa_sec.c
diff --git
a/drivers/crypto/dpaa_sec/dpaa_sec.c
b/drivers/crypto/dpaa_sec/dpaa_sec.c
index
2dac510
..
06f7e43
100644
(file)
--- a/
drivers/crypto/dpaa_sec/dpaa_sec.c
+++ b/
drivers/crypto/dpaa_sec/dpaa_sec.c
@@
-1,7
+1,7
@@
/* SPDX-License-Identifier: BSD-3-Clause
*
* Copyright (c) 2016 Freescale Semiconductor, Inc. All rights reserved.
/* SPDX-License-Identifier: BSD-3-Clause
*
* Copyright (c) 2016 Freescale Semiconductor, Inc. All rights reserved.
- * Copyright 2017 NXP
+ * Copyright 2017
-2018
NXP
*
*/
*
*/
@@
-39,6
+39,8
@@
enum rta_sec_era rta_sec_era;
enum rta_sec_era rta_sec_era;
+int dpaa_logtype_sec;
+
static uint8_t cryptodev_driver_id;
static __thread struct rte_crypto_op **dpaa_sec_ops;
static uint8_t cryptodev_driver_id;
static __thread struct rte_crypto_op **dpaa_sec_ops;
@@
-53,7
+55,7
@@
dpaa_sec_op_ending(struct dpaa_sec_op_ctx *ctx)
if (!ctx->fd_status) {
ctx->op->status = RTE_CRYPTO_OP_STATUS_SUCCESS;
} else {
if (!ctx->fd_status) {
ctx->op->status = RTE_CRYPTO_OP_STATUS_SUCCESS;
} else {
-
PMD_RX_LOG(ERR,
"SEC return err: 0x%x", ctx->fd_status);
+
DPAA_SEC_DP_WARN(
"SEC return err: 0x%x", ctx->fd_status);
ctx->op->status = RTE_CRYPTO_OP_STATUS_ERROR;
}
ctx->op->status = RTE_CRYPTO_OP_STATUS_ERROR;
}
@@
-69,7
+71,7
@@
dpaa_sec_alloc_ctx(dpaa_sec_session *ses)
retval = rte_mempool_get(ses->ctx_pool, (void **)(&ctx));
if (!ctx || retval) {
retval = rte_mempool_get(ses->ctx_pool, (void **)(&ctx));
if (!ctx || retval) {
-
PMD_TX_LOG(ERR,
"Alloc sec descriptor failed!");
+
DPAA_SEC_DP_WARN(
"Alloc sec descriptor failed!");
return NULL;
}
/*
return NULL;
}
/*
@@
-101,13
+103,6
@@
dpaa_mem_vtop(void *vaddr)
return (size_t)NULL;
}
return (size_t)NULL;
}
-/* virtual address conversin when mempool support is available for ctx */
-static inline phys_addr_t
-dpaa_mem_vtop_ctx(struct dpaa_sec_op_ctx *ctx, void *vaddr)
-{
- return (size_t)vaddr - ctx->vtop_offset;
-}
-
static inline void *
dpaa_mem_ptov(rte_iova_t paddr)
{
static inline void *
dpaa_mem_ptov(rte_iova_t paddr)
{
@@
-119,8
+114,8
@@
ern_sec_fq_handler(struct qman_portal *qm __rte_unused,
struct qman_fq *fq,
const struct qm_mr_entry *msg)
{
struct qman_fq *fq,
const struct qm_mr_entry *msg)
{
-
RTE_LOG_DP(ERR, PMD,
"sec fq %d error, RC = %x, seqnum = %x\n",
- fq->fqid, msg->ern.rc, msg->ern.seqnum);
+
DPAA_SEC_DP_ERR(
"sec fq %d error, RC = %x, seqnum = %x\n",
+
fq->fqid, msg->ern.rc, msg->ern.seqnum);
}
/* initialize the queue with dest chan as caam chan so that
}
/* initialize the queue with dest chan as caam chan so that
@@
-148,11
+143,11
@@
dpaa_sec_init_rx(struct qman_fq *fq_in, rte_iova_t hwdesc,
fq_in->cb.ern = ern_sec_fq_handler;
fq_in->cb.ern = ern_sec_fq_handler;
-
PMD_INIT_LOG(DEBUG,
"in-%x out-%x", fq_in->fqid, fqid_out);
+
DPAA_SEC_DEBUG(
"in-%x out-%x", fq_in->fqid, fqid_out);
ret = qman_init_fq(fq_in, flags, &fq_opts);
if (unlikely(ret != 0))
ret = qman_init_fq(fq_in, flags, &fq_opts);
if (unlikely(ret != 0))
-
PMD_INIT_LOG(ERR,
"qman_init_fq failed %d", ret);
+
DPAA_SEC_ERR(
"qman_init_fq failed %d", ret);
return ret;
}
return ret;
}
@@
-211,7
+206,7
@@
dpaa_sec_init_tx(struct qman_fq *fq)
ret = qman_create_fq(0, flags, fq);
if (unlikely(ret)) {
ret = qman_create_fq(0, flags, fq);
if (unlikely(ret)) {
-
PMD_INIT_LOG(ERR,
"qman_create_fq failed");
+
DPAA_SEC_ERR(
"qman_create_fq failed");
return ret;
}
return ret;
}
@@
-226,7
+221,7
@@
dpaa_sec_init_tx(struct qman_fq *fq)
ret = qman_init_fq(fq, 0, &opts);
if (unlikely(ret)) {
ret = qman_init_fq(fq, 0, &opts);
if (unlikely(ret)) {
-
PMD_INIT_LOG(ERR,
"unable to init caam source fq!");
+
DPAA_SEC_ERR(
"unable to init caam source fq!");
return ret;
}
return ret;
}
@@
-318,7
+313,7
@@
caam_auth_alg(dpaa_sec_session *ses, struct alginfo *alginfo_a)
alginfo_a->algmode = OP_ALG_AAI_HMAC;
break;
default:
alginfo_a->algmode = OP_ALG_AAI_HMAC;
break;
default:
-
PMD_INIT_LOG(ERR,
"unsupported auth alg %u", ses->auth_alg);
+
DPAA_SEC_ERR(
"unsupported auth alg %u", ses->auth_alg);
}
}
}
}
@@
-347,7
+342,7
@@
caam_cipher_alg(dpaa_sec_session *ses, struct alginfo *alginfo_c)
alginfo_c->algmode = OP_ALG_AAI_CTR;
break;
default:
alginfo_c->algmode = OP_ALG_AAI_CTR;
break;
default:
-
PMD_INIT_LOG(ERR,
"unsupported cipher alg %d", ses->cipher_alg);
+
DPAA_SEC_ERR(
"unsupported cipher alg %d", ses->cipher_alg);
}
}
}
}
@@
-360,7
+355,7
@@
caam_aead_alg(dpaa_sec_session *ses, struct alginfo *alginfo)
alginfo->algmode = OP_ALG_AAI_GCM;
break;
default:
alginfo->algmode = OP_ALG_AAI_GCM;
break;
default:
-
PMD_INIT_LOG(ERR,
"unsupported AEAD alg %d", ses->aead_alg);
+
DPAA_SEC_ERR(
"unsupported AEAD alg %d", ses->aead_alg);
}
}
}
}
@@
-384,7
+379,7
@@
dpaa_sec_prep_cdb(dpaa_sec_session *ses)
if (is_cipher_only(ses)) {
caam_cipher_alg(ses, &alginfo_c);
if (alginfo_c.algtype == (unsigned int)DPAA_SEC_ALG_UNSUPPORT) {
if (is_cipher_only(ses)) {
caam_cipher_alg(ses, &alginfo_c);
if (alginfo_c.algtype == (unsigned int)DPAA_SEC_ALG_UNSUPPORT) {
-
PMD_TX_LOG(ERR, "not supported cipher alg\n
");
+
DPAA_SEC_ERR("not supported cipher alg
");
return -ENOTSUP;
}
return -ENOTSUP;
}
@@
-402,7
+397,7
@@
dpaa_sec_prep_cdb(dpaa_sec_session *ses)
} else if (is_auth_only(ses)) {
caam_auth_alg(ses, &alginfo_a);
if (alginfo_a.algtype == (unsigned int)DPAA_SEC_ALG_UNSUPPORT) {
} else if (is_auth_only(ses)) {
caam_auth_alg(ses, &alginfo_a);
if (alginfo_a.algtype == (unsigned int)DPAA_SEC_ALG_UNSUPPORT) {
-
PMD_TX_LOG(ERR, "not supported auth alg\n
");
+
DPAA_SEC_ERR("not supported auth alg
");
return -ENOTSUP;
}
return -ENOTSUP;
}
@@
-418,7
+413,7
@@
dpaa_sec_prep_cdb(dpaa_sec_session *ses)
} else if (is_aead(ses)) {
caam_aead_alg(ses, &alginfo);
if (alginfo.algtype == (unsigned int)DPAA_SEC_ALG_UNSUPPORT) {
} else if (is_aead(ses)) {
caam_aead_alg(ses, &alginfo);
if (alginfo.algtype == (unsigned int)DPAA_SEC_ALG_UNSUPPORT) {
-
PMD_TX_LOG(ERR, "not supported aead alg\n
");
+
DPAA_SEC_ERR("not supported aead alg
");
return -ENOTSUP;
}
alginfo.key = (size_t)ses->aead_key.data;
return -ENOTSUP;
}
alginfo.key = (size_t)ses->aead_key.data;
@@
-441,7
+436,7
@@
dpaa_sec_prep_cdb(dpaa_sec_session *ses)
} else {
caam_cipher_alg(ses, &alginfo_c);
if (alginfo_c.algtype == (unsigned int)DPAA_SEC_ALG_UNSUPPORT) {
} else {
caam_cipher_alg(ses, &alginfo_c);
if (alginfo_c.algtype == (unsigned int)DPAA_SEC_ALG_UNSUPPORT) {
-
PMD_TX_LOG(ERR, "not supported cipher alg\n
");
+
DPAA_SEC_ERR("not supported cipher alg
");
return -ENOTSUP;
}
return -ENOTSUP;
}
@@
-452,7
+447,7
@@
dpaa_sec_prep_cdb(dpaa_sec_session *ses)
caam_auth_alg(ses, &alginfo_a);
if (alginfo_a.algtype == (unsigned int)DPAA_SEC_ALG_UNSUPPORT) {
caam_auth_alg(ses, &alginfo_a);
if (alginfo_a.algtype == (unsigned int)DPAA_SEC_ALG_UNSUPPORT) {
-
PMD_TX_LOG(ERR, "not supported auth alg\n
");
+
DPAA_SEC_ERR("not supported auth alg
");
return -ENOTSUP;
}
return -ENOTSUP;
}
@@
-469,7
+464,7
@@
dpaa_sec_prep_cdb(dpaa_sec_session *ses)
&cdb->sh_desc[2], 2);
if (err < 0) {
&cdb->sh_desc[2], 2);
if (err < 0) {
-
PMD_TX_LOG(ERR,
"Crypto: Incorrect key lengths");
+
DPAA_SEC_ERR(
"Crypto: Incorrect key lengths");
return err;
}
if (cdb->sh_desc[2] & 1)
return err;
}
if (cdb->sh_desc[2] & 1)
@@
-514,7
+509,7
@@
dpaa_sec_prep_cdb(dpaa_sec_session *ses)
}
if (shared_desc_len < 0) {
}
if (shared_desc_len < 0) {
-
PMD_TX_LOG(ERR, "error in preparing command block\n
");
+
DPAA_SEC_ERR("error in preparing command block
");
return shared_desc_len;
}
return shared_desc_len;
}
@@
-573,7
+568,7
@@
dpaa_sec_deq(struct dpaa_sec_qp *qp, struct rte_crypto_op **ops, int nb_ops)
if (!ctx->fd_status) {
op->status = RTE_CRYPTO_OP_STATUS_SUCCESS;
} else {
if (!ctx->fd_status) {
op->status = RTE_CRYPTO_OP_STATUS_SUCCESS;
} else {
-
printf("\nSEC return err:
0x%x", ctx->fd_status);
+
DPAA_SEC_DP_WARN("SEC return err:
0x%x", ctx->fd_status);
op->status = RTE_CRYPTO_OP_STATUS_ERROR;
}
ops[pkts++] = op;
op->status = RTE_CRYPTO_OP_STATUS_ERROR;
}
ops[pkts++] = op;
@@
-604,8
+599,8
@@
build_auth_only_sg(struct rte_crypto_op *op, dpaa_sec_session *ses)
extra_segs = 2;
if ((mbuf->nb_segs + extra_segs) > MAX_SG_ENTRIES) {
extra_segs = 2;
if ((mbuf->nb_segs + extra_segs) > MAX_SG_ENTRIES) {
-
PMD_TX_LOG(ERR, "Auth: Max sec segs supported is %d\n
",
-
MAX_SG_ENTRIES);
+
DPAA_SEC_DP_ERR("Auth: Max sec segs supported is %d
",
+ MAX_SG_ENTRIES);
return NULL;
}
ctx = dpaa_sec_alloc_ctx(ses);
return NULL;
}
ctx = dpaa_sec_alloc_ctx(ses);
@@
-628,7
+623,7
@@
build_auth_only_sg(struct rte_crypto_op *op, dpaa_sec_session *ses)
in_sg->extension = 1;
in_sg->final = 1;
in_sg->length = sym->auth.data.length;
in_sg->extension = 1;
in_sg->final = 1;
in_sg->length = sym->auth.data.length;
- qm_sg_entry_set64(in_sg, dpaa_mem_vtop
_ctx(ctx,
&cf->sg[2]));
+ qm_sg_entry_set64(in_sg, dpaa_mem_vtop
(
&cf->sg[2]));
/* 1st seg */
sg = in_sg + 1;
/* 1st seg */
sg = in_sg + 1;
@@
-652,7
+647,7
@@
build_auth_only_sg(struct rte_crypto_op *op, dpaa_sec_session *ses)
sg++;
rte_memcpy(old_digest, sym->auth.digest.data,
ses->digest_length);
sg++;
rte_memcpy(old_digest, sym->auth.digest.data,
ses->digest_length);
- start_addr = dpaa_mem_vtop
_ctx(ctx,
old_digest);
+ start_addr = dpaa_mem_vtop
(
old_digest);
qm_sg_entry_set64(sg, start_addr);
sg->length = ses->digest_length;
in_sg->length += ses->digest_length;
qm_sg_entry_set64(sg, start_addr);
sg->length = ses->digest_length;
in_sg->length += ses->digest_length;
@@
-706,7
+701,7
@@
build_auth_only(struct rte_crypto_op *op, dpaa_sec_session *ses)
if (is_decode(ses)) {
/* need to extend the input to a compound frame */
sg->extension = 1;
if (is_decode(ses)) {
/* need to extend the input to a compound frame */
sg->extension = 1;
- qm_sg_entry_set64(sg, dpaa_mem_vtop
_ctx(ctx,
&cf->sg[2]));
+ qm_sg_entry_set64(sg, dpaa_mem_vtop
(
&cf->sg[2]));
sg->length = sym->auth.data.length + ses->digest_length;
sg->final = 1;
cpu_to_hw_sg(sg);
sg->length = sym->auth.data.length + ses->digest_length;
sg->final = 1;
cpu_to_hw_sg(sg);
@@
-720,7
+715,7
@@
build_auth_only(struct rte_crypto_op *op, dpaa_sec_session *ses)
cpu_to_hw_sg(sg);
/* let's check digest by hw */
cpu_to_hw_sg(sg);
/* let's check digest by hw */
- start_addr = dpaa_mem_vtop
_ctx(ctx,
old_digest);
+ start_addr = dpaa_mem_vtop
(
old_digest);
sg++;
qm_sg_entry_set64(sg, start_addr);
sg->length = ses->digest_length;
sg++;
qm_sg_entry_set64(sg, start_addr);
sg->length = ses->digest_length;
@@
-757,8
+752,8
@@
build_cipher_only_sg(struct rte_crypto_op *op, dpaa_sec_session *ses)
}
if (req_segs > MAX_SG_ENTRIES) {
}
if (req_segs > MAX_SG_ENTRIES) {
-
PMD_TX_LOG(ERR, "Cipher: Max sec segs supported is %d\n
",
-
MAX_SG_ENTRIES);
+
DPAA_SEC_DP_ERR("Cipher: Max sec segs supported is %d
",
+ MAX_SG_ENTRIES);
return NULL;
}
return NULL;
}
@@
-773,7
+768,7
@@
build_cipher_only_sg(struct rte_crypto_op *op, dpaa_sec_session *ses)
out_sg = &cf->sg[0];
out_sg->extension = 1;
out_sg->length = sym->cipher.data.length;
out_sg = &cf->sg[0];
out_sg->extension = 1;
out_sg->length = sym->cipher.data.length;
- qm_sg_entry_set64(out_sg, dpaa_mem_vtop
_ctx(ctx,
&cf->sg[2]));
+ qm_sg_entry_set64(out_sg, dpaa_mem_vtop
(
&cf->sg[2]));
cpu_to_hw_sg(out_sg);
/* 1st seg */
cpu_to_hw_sg(out_sg);
/* 1st seg */
@@
-802,7
+797,7
@@
build_cipher_only_sg(struct rte_crypto_op *op, dpaa_sec_session *ses)
in_sg->length = sym->cipher.data.length + ses->iv.length;
sg++;
in_sg->length = sym->cipher.data.length + ses->iv.length;
sg++;
- qm_sg_entry_set64(in_sg, dpaa_mem_vtop
_ctx(ctx,
sg));
+ qm_sg_entry_set64(in_sg, dpaa_mem_vtop
(
sg));
cpu_to_hw_sg(in_sg);
/* IV */
cpu_to_hw_sg(in_sg);
/* IV */
@@
-869,7
+864,7
@@
build_cipher_only(struct rte_crypto_op *op, dpaa_sec_session *ses)
sg->extension = 1;
sg->final = 1;
sg->length = sym->cipher.data.length + ses->iv.length;
sg->extension = 1;
sg->final = 1;
sg->length = sym->cipher.data.length + ses->iv.length;
- qm_sg_entry_set64(sg, dpaa_mem_vtop
_ctx(ctx,
&cf->sg[2]));
+ qm_sg_entry_set64(sg, dpaa_mem_vtop
(
&cf->sg[2]));
cpu_to_hw_sg(sg);
sg = &cf->sg[2];
cpu_to_hw_sg(sg);
sg = &cf->sg[2];
@@
-910,7
+905,7
@@
build_cipher_auth_gcm_sg(struct rte_crypto_op *op, dpaa_sec_session *ses)
req_segs++;
if (req_segs > MAX_SG_ENTRIES) {
req_segs++;
if (req_segs > MAX_SG_ENTRIES) {
-
PMD_TX_LOG(ERR, "AEAD: Max sec segs supported is %d\n
",
+
DPAA_SEC_DP_ERR("AEAD: Max sec segs supported is %d
",
MAX_SG_ENTRIES);
return NULL;
}
MAX_SG_ENTRIES);
return NULL;
}
@@
-935,7
+930,7
@@
build_cipher_auth_gcm_sg(struct rte_crypto_op *op, dpaa_sec_session *ses)
/* output sg entries */
sg = &cf->sg[2];
/* output sg entries */
sg = &cf->sg[2];
- qm_sg_entry_set64(out_sg, dpaa_mem_vtop
_ctx(ctx,
sg));
+ qm_sg_entry_set64(out_sg, dpaa_mem_vtop
(
sg));
cpu_to_hw_sg(out_sg);
/* 1st seg */
cpu_to_hw_sg(out_sg);
/* 1st seg */
@@
-979,7
+974,7
@@
build_cipher_auth_gcm_sg(struct rte_crypto_op *op, dpaa_sec_session *ses)
/* input sg entries */
sg++;
/* input sg entries */
sg++;
- qm_sg_entry_set64(in_sg, dpaa_mem_vtop
_ctx(ctx,
sg));
+ qm_sg_entry_set64(in_sg, dpaa_mem_vtop
(
sg));
cpu_to_hw_sg(in_sg);
/* 1st seg IV */
cpu_to_hw_sg(in_sg);
/* 1st seg IV */
@@
-1016,7
+1011,7
@@
build_cipher_auth_gcm_sg(struct rte_crypto_op *op, dpaa_sec_session *ses)
sg++;
memcpy(ctx->digest, sym->aead.digest.data,
ses->digest_length);
sg++;
memcpy(ctx->digest, sym->aead.digest.data,
ses->digest_length);
- qm_sg_entry_set64(sg, dpaa_mem_vtop
_ctx(ctx,
ctx->digest));
+ qm_sg_entry_set64(sg, dpaa_mem_vtop
(
ctx->digest));
sg->length = ses->digest_length;
}
sg->final = 1;
sg->length = ses->digest_length;
}
sg->final = 1;
@@
-1054,7
+1049,7
@@
build_cipher_auth_gcm(struct rte_crypto_op *op, dpaa_sec_session *ses)
/* input */
rte_prefetch0(cf->sg);
sg = &cf->sg[2];
/* input */
rte_prefetch0(cf->sg);
sg = &cf->sg[2];
- qm_sg_entry_set64(&cf->sg[1], dpaa_mem_vtop
_ctx(ctx,
sg));
+ qm_sg_entry_set64(&cf->sg[1], dpaa_mem_vtop
(
sg));
if (is_encode(ses)) {
qm_sg_entry_set64(sg, dpaa_mem_vtop(IV_ptr));
sg->length = ses->iv.length;
if (is_encode(ses)) {
qm_sg_entry_set64(sg, dpaa_mem_vtop(IV_ptr));
sg->length = ses->iv.length;
@@
-1099,7
+1094,7
@@
build_cipher_auth_gcm(struct rte_crypto_op *op, dpaa_sec_session *ses)
ses->digest_length);
sg++;
ses->digest_length);
sg++;
- qm_sg_entry_set64(sg, dpaa_mem_vtop
_ctx(ctx,
ctx->digest));
+ qm_sg_entry_set64(sg, dpaa_mem_vtop
(
ctx->digest));
sg->length = ses->digest_length;
length += sg->length;
sg->final = 1;
sg->length = ses->digest_length;
length += sg->length;
sg->final = 1;
@@
-1113,7
+1108,7
@@
build_cipher_auth_gcm(struct rte_crypto_op *op, dpaa_sec_session *ses)
/* output */
sg++;
/* output */
sg++;
- qm_sg_entry_set64(&cf->sg[0], dpaa_mem_vtop
_ctx(ctx,
sg));
+ qm_sg_entry_set64(&cf->sg[0], dpaa_mem_vtop
(
sg));
qm_sg_entry_set64(sg,
dst_start_addr + sym->aead.data.offset - ses->auth_only_len);
sg->length = sym->aead.data.length + ses->auth_only_len;
qm_sg_entry_set64(sg,
dst_start_addr + sym->aead.data.offset - ses->auth_only_len);
sg->length = sym->aead.data.length + ses->auth_only_len;
@@
-1158,7
+1153,7
@@
build_cipher_auth_sg(struct rte_crypto_op *op, dpaa_sec_session *ses)
}
if (req_segs > MAX_SG_ENTRIES) {
}
if (req_segs > MAX_SG_ENTRIES) {
-
PMD_TX_LOG(ERR, "Cipher-Auth: Max sec segs supported is %d\n
",
+
DPAA_SEC_DP_ERR("Cipher-Auth: Max sec segs supported is %d
",
MAX_SG_ENTRIES);
return NULL;
}
MAX_SG_ENTRIES);
return NULL;
}
@@
-1182,7
+1177,7
@@
build_cipher_auth_sg(struct rte_crypto_op *op, dpaa_sec_session *ses)
/* output sg entries */
sg = &cf->sg[2];
/* output sg entries */
sg = &cf->sg[2];
- qm_sg_entry_set64(out_sg, dpaa_mem_vtop
_ctx(ctx,
sg));
+ qm_sg_entry_set64(out_sg, dpaa_mem_vtop
(
sg));
cpu_to_hw_sg(out_sg);
/* 1st seg */
cpu_to_hw_sg(out_sg);
/* 1st seg */
@@
-1224,7
+1219,7
@@
build_cipher_auth_sg(struct rte_crypto_op *op, dpaa_sec_session *ses)
/* input sg entries */
sg++;
/* input sg entries */
sg++;
- qm_sg_entry_set64(in_sg, dpaa_mem_vtop
_ctx(ctx,
sg));
+ qm_sg_entry_set64(in_sg, dpaa_mem_vtop
(
sg));
cpu_to_hw_sg(in_sg);
/* 1st seg IV */
cpu_to_hw_sg(in_sg);
/* 1st seg IV */
@@
-1254,7
+1249,7
@@
build_cipher_auth_sg(struct rte_crypto_op *op, dpaa_sec_session *ses)
sg++;
memcpy(ctx->digest, sym->auth.digest.data,
ses->digest_length);
sg++;
memcpy(ctx->digest, sym->auth.digest.data,
ses->digest_length);
- qm_sg_entry_set64(sg, dpaa_mem_vtop
_ctx(ctx,
ctx->digest));
+ qm_sg_entry_set64(sg, dpaa_mem_vtop
(
ctx->digest));
sg->length = ses->digest_length;
}
sg->final = 1;
sg->length = ses->digest_length;
}
sg->final = 1;
@@
-1291,7
+1286,7
@@
build_cipher_auth(struct rte_crypto_op *op, dpaa_sec_session *ses)
/* input */
rte_prefetch0(cf->sg);
sg = &cf->sg[2];
/* input */
rte_prefetch0(cf->sg);
sg = &cf->sg[2];
- qm_sg_entry_set64(&cf->sg[1], dpaa_mem_vtop
_ctx(ctx,
sg));
+ qm_sg_entry_set64(&cf->sg[1], dpaa_mem_vtop
(
sg));
if (is_encode(ses)) {
qm_sg_entry_set64(sg, dpaa_mem_vtop(IV_ptr));
sg->length = ses->iv.length;
if (is_encode(ses)) {
qm_sg_entry_set64(sg, dpaa_mem_vtop(IV_ptr));
sg->length = ses->iv.length;
@@
-1321,7
+1316,7
@@
build_cipher_auth(struct rte_crypto_op *op, dpaa_sec_session *ses)
ses->digest_length);
sg++;
ses->digest_length);
sg++;
- qm_sg_entry_set64(sg, dpaa_mem_vtop
_ctx(ctx,
ctx->digest));
+ qm_sg_entry_set64(sg, dpaa_mem_vtop
(
ctx->digest));
sg->length = ses->digest_length;
length += sg->length;
sg->final = 1;
sg->length = ses->digest_length;
length += sg->length;
sg->final = 1;
@@
-1335,7
+1330,7
@@
build_cipher_auth(struct rte_crypto_op *op, dpaa_sec_session *ses)
/* output */
sg++;
/* output */
sg++;
- qm_sg_entry_set64(&cf->sg[0], dpaa_mem_vtop
_ctx(ctx,
sg));
+ qm_sg_entry_set64(&cf->sg[0], dpaa_mem_vtop
(
sg));
qm_sg_entry_set64(sg, dst_start_addr + sym->cipher.data.offset);
sg->length = sym->cipher.data.length;
length = sg->length;
qm_sg_entry_set64(sg, dst_start_addr + sym->cipher.data.offset);
sg->length = sym->cipher.data.length;
length = sg->length;
@@
-1410,7
+1405,6
@@
dpaa_sec_enqueue_burst(void *qp, struct rte_crypto_op **ops,
struct rte_crypto_op *op;
struct dpaa_sec_job *cf;
dpaa_sec_session *ses;
struct rte_crypto_op *op;
struct dpaa_sec_job *cf;
dpaa_sec_session *ses;
- struct dpaa_sec_op_ctx *ctx;
uint32_t auth_only_len;
struct qman_fq *inq[DPAA_SEC_BURST];
uint32_t auth_only_len;
struct qman_fq *inq[DPAA_SEC_BURST];
@@
-1432,15
+1426,15
@@
dpaa_sec_enqueue_burst(void *qp, struct rte_crypto_op **ops,
op->sym->sec_session);
break;
default:
op->sym->sec_session);
break;
default:
- PMD_TX_LOG(ERR,
+ DPAA_SEC_DP_ERR(
"sessionless crypto op not supported");
frames_to_send = loop;
nb_ops = loop;
goto send_pkts;
}
if (unlikely(!ses->qp || ses->qp != qp)) {
"sessionless crypto op not supported");
frames_to_send = loop;
nb_ops = loop;
goto send_pkts;
}
if (unlikely(!ses->qp || ses->qp != qp)) {
-
PMD_INIT_LOG(DEBUG,
"sess->qp - %p qp %p",
-
ses->qp, qp);
+
DPAA_SEC_DP_ERR(
"sess->qp - %p qp %p",
+ ses->qp, qp);
if (dpaa_sec_attach_sess_q(qp, ses)) {
frames_to_send = loop;
nb_ops = loop;
if (dpaa_sec_attach_sess_q(qp, ses)) {
frames_to_send = loop;
nb_ops = loop;
@@
-1463,7
+1457,7
@@
dpaa_sec_enqueue_burst(void *qp, struct rte_crypto_op **ops,
} else if (is_proto_ipsec(ses)) {
cf = build_proto(op, ses);
} else {
} else if (is_proto_ipsec(ses)) {
cf = build_proto(op, ses);
} else {
-
PMD_TX_LOG(ERR, "not supported sec op
");
+
DPAA_SEC_DP_ERR("not supported ops
");
frames_to_send = loop;
nb_ops = loop;
goto send_pkts;
frames_to_send = loop;
nb_ops = loop;
goto send_pkts;
@@
-1479,7
+1473,7
@@
dpaa_sec_enqueue_burst(void *qp, struct rte_crypto_op **ops,
} else if (is_auth_cipher(ses)) {
cf = build_cipher_auth_sg(op, ses);
} else {
} else if (is_auth_cipher(ses)) {
cf = build_cipher_auth_sg(op, ses);
} else {
-
PMD_TX_LOG(ERR, "not supported sec op
");
+
DPAA_SEC_DP_ERR("not supported ops
");
frames_to_send = loop;
nb_ops = loop;
goto send_pkts;
frames_to_send = loop;
nb_ops = loop;
goto send_pkts;
@@
-1495,8
+1489,7
@@
dpaa_sec_enqueue_burst(void *qp, struct rte_crypto_op **ops,
inq[loop] = ses->inq;
fd->opaque_addr = 0;
fd->cmd = 0;
inq[loop] = ses->inq;
fd->opaque_addr = 0;
fd->cmd = 0;
- ctx = container_of(cf, struct dpaa_sec_op_ctx, job);
- qm_fd_addr_set64(fd, dpaa_mem_vtop_ctx(ctx, cf->sg));
+ qm_fd_addr_set64(fd, dpaa_mem_vtop(cf->sg));
fd->_format1 = qm_fd_compound;
fd->length29 = 2 * sizeof(struct qm_sg_entry);
/* Auth_only_len is set as 0 in descriptor and it is
fd->_format1 = qm_fd_compound;
fd->length29 = 2 * sizeof(struct qm_sg_entry);
/* Auth_only_len is set as 0 in descriptor and it is
@@
-1535,7
+1528,7
@@
dpaa_sec_dequeue_burst(void *qp, struct rte_crypto_op **ops,
dpaa_qp->rx_pkts += num_rx;
dpaa_qp->rx_errs += nb_ops - num_rx;
dpaa_qp->rx_pkts += num_rx;
dpaa_qp->rx_errs += nb_ops - num_rx;
-
PMD_RX_LOG(DEBUG,
"SEC Received %d Packets\n", num_rx);
+
DPAA_SEC_DP_DEBUG(
"SEC Received %d Packets\n", num_rx);
return num_rx;
}
return num_rx;
}
@@
-1550,11
+1543,11
@@
dpaa_sec_queue_pair_release(struct rte_cryptodev *dev,
PMD_INIT_FUNC_TRACE();
PMD_INIT_FUNC_TRACE();
-
PMD_INIT_LOG(DEBUG,
"dev =%p, queue =%d", dev, qp_id);
+
DPAA_SEC_DEBUG(
"dev =%p, queue =%d", dev, qp_id);
internals = dev->data->dev_private;
if (qp_id >= internals->max_nb_queue_pairs) {
internals = dev->data->dev_private;
if (qp_id >= internals->max_nb_queue_pairs) {
-
PMD_INIT_LOG(ERR,
"Max supported qpid %d",
+
DPAA_SEC_ERR(
"Max supported qpid %d",
internals->max_nb_queue_pairs);
return -EINVAL;
}
internals->max_nb_queue_pairs);
return -EINVAL;
}
@@
-1576,12
+1569,11
@@
dpaa_sec_queue_pair_setup(struct rte_cryptodev *dev, uint16_t qp_id,
struct dpaa_sec_dev_private *internals;
struct dpaa_sec_qp *qp = NULL;
struct dpaa_sec_dev_private *internals;
struct dpaa_sec_qp *qp = NULL;
- PMD_INIT_LOG(DEBUG, "dev =%p, queue =%d, conf =%p",
- dev, qp_id, qp_conf);
+ DPAA_SEC_DEBUG("dev =%p, queue =%d, conf =%p", dev, qp_id, qp_conf);
internals = dev->data->dev_private;
if (qp_id >= internals->max_nb_queue_pairs) {
internals = dev->data->dev_private;
if (qp_id >= internals->max_nb_queue_pairs) {
-
PMD_INIT_LOG(ERR,
"Max supported qpid %d",
+
DPAA_SEC_ERR(
"Max supported qpid %d",
internals->max_nb_queue_pairs);
return -EINVAL;
}
internals->max_nb_queue_pairs);
return -EINVAL;
}
@@
-1642,7
+1634,7
@@
dpaa_sec_cipher_init(struct rte_cryptodev *dev __rte_unused,
session->cipher_key.data = rte_zmalloc(NULL, xform->cipher.key.length,
RTE_CACHE_LINE_SIZE);
if (session->cipher_key.data == NULL && xform->cipher.key.length > 0) {
session->cipher_key.data = rte_zmalloc(NULL, xform->cipher.key.length,
RTE_CACHE_LINE_SIZE);
if (session->cipher_key.data == NULL && xform->cipher.key.length > 0) {
-
PMD_INIT_LOG(ERR, "No Memory for cipher key\n
");
+
DPAA_SEC_ERR("No Memory for cipher key
");
return -ENOMEM;
}
session->cipher_key.length = xform->cipher.key.length;
return -ENOMEM;
}
session->cipher_key.length = xform->cipher.key.length;
@@
-1664,7
+1656,7
@@
dpaa_sec_auth_init(struct rte_cryptodev *dev __rte_unused,
session->auth_key.data = rte_zmalloc(NULL, xform->auth.key.length,
RTE_CACHE_LINE_SIZE);
if (session->auth_key.data == NULL && xform->auth.key.length > 0) {
session->auth_key.data = rte_zmalloc(NULL, xform->auth.key.length,
RTE_CACHE_LINE_SIZE);
if (session->auth_key.data == NULL && xform->auth.key.length > 0) {
-
PMD_INIT_LOG(ERR, "No Memory for auth key\n
");
+
DPAA_SEC_ERR("No Memory for auth key
");
return -ENOMEM;
}
session->auth_key.length = xform->auth.key.length;
return -ENOMEM;
}
session->auth_key.length = xform->auth.key.length;
@@
-1690,7
+1682,7
@@
dpaa_sec_aead_init(struct rte_cryptodev *dev __rte_unused,
session->aead_key.data = rte_zmalloc(NULL, xform->aead.key.length,
RTE_CACHE_LINE_SIZE);
if (session->aead_key.data == NULL && xform->aead.key.length > 0) {
session->aead_key.data = rte_zmalloc(NULL, xform->aead.key.length,
RTE_CACHE_LINE_SIZE);
if (session->aead_key.data == NULL && xform->aead.key.length > 0) {
-
PMD_INIT_LOG(ERR,
"No Memory for aead key\n");
+
DPAA_SEC_ERR(
"No Memory for aead key\n");
return -ENOMEM;
}
session->aead_key.length = xform->aead.key.length;
return -ENOMEM;
}
session->aead_key.length = xform->aead.key.length;
@@
-1715,7
+1707,7
@@
dpaa_sec_attach_rxq(struct dpaa_sec_dev_private *qi)
return &qi->inq[i];
}
}
return &qi->inq[i];
}
}
-
PMD_DRV_LOG(ERR,
"All ses session in use %x", qi->max_nb_sessions);
+
DPAA_SEC_WARN(
"All ses session in use %x", qi->max_nb_sessions);
return NULL;
}
return NULL;
}
@@
-1744,20
+1736,20
@@
dpaa_sec_attach_sess_q(struct dpaa_sec_qp *qp, dpaa_sec_session *sess)
sess->qp = qp;
ret = dpaa_sec_prep_cdb(sess);
if (ret) {
sess->qp = qp;
ret = dpaa_sec_prep_cdb(sess);
if (ret) {
-
PMD_DRV_LOG(ERR,
"Unable to prepare sec cdb");
+
DPAA_SEC_ERR(
"Unable to prepare sec cdb");
return -1;
}
if (unlikely(!RTE_PER_LCORE(dpaa_io))) {
ret = rte_dpaa_portal_init((void *)0);
if (ret) {
return -1;
}
if (unlikely(!RTE_PER_LCORE(dpaa_io))) {
ret = rte_dpaa_portal_init((void *)0);
if (ret) {
-
PMD_DRV_LOG(ERR,
"Failure in affining portal");
+
DPAA_SEC_ERR(
"Failure in affining portal");
return ret;
}
}
ret = dpaa_sec_init_rx(sess->inq, dpaa_mem_vtop(&sess->cdb),
qman_fq_fqid(&qp->outq));
if (ret)
return ret;
}
}
ret = dpaa_sec_init_rx(sess->inq, dpaa_mem_vtop(&sess->cdb),
qman_fq_fqid(&qp->outq));
if (ret)
-
PMD_DRV_LOG(ERR,
"Unable to init sec queue");
+
DPAA_SEC_ERR(
"Unable to init sec queue");
return ret;
}
return ret;
}
@@
-1800,7
+1792,7
@@
dpaa_sec_set_session_parameters(struct rte_cryptodev *dev,
PMD_INIT_FUNC_TRACE();
if (unlikely(sess == NULL)) {
PMD_INIT_FUNC_TRACE();
if (unlikely(sess == NULL)) {
-
RTE_LOG(ERR, PMD, "invalid session struct\n
");
+
DPAA_SEC_ERR("invalid session struct
");
return -EINVAL;
}
return -EINVAL;
}
@@
-1825,7
+1817,7
@@
dpaa_sec_set_session_parameters(struct rte_cryptodev *dev,
dpaa_sec_cipher_init(dev, xform, session);
dpaa_sec_auth_init(dev, xform->next, session);
} else {
dpaa_sec_cipher_init(dev, xform, session);
dpaa_sec_auth_init(dev, xform->next, session);
} else {
-
PMD_DRV_LOG(ERR,
"Not supported: Auth then Cipher");
+
DPAA_SEC_ERR(
"Not supported: Auth then Cipher");
return -EINVAL;
}
return -EINVAL;
}
@@
-1836,7
+1828,7
@@
dpaa_sec_set_session_parameters(struct rte_cryptodev *dev,
dpaa_sec_auth_init(dev, xform, session);
dpaa_sec_cipher_init(dev, xform->next, session);
} else {
dpaa_sec_auth_init(dev, xform, session);
dpaa_sec_cipher_init(dev, xform->next, session);
} else {
-
PMD_DRV_LOG(ERR,
"Not supported: Auth then Cipher");
+
DPAA_SEC_ERR(
"Not supported: Auth then Cipher");
return -EINVAL;
}
return -EINVAL;
}
@@
-1846,13
+1838,13
@@
dpaa_sec_set_session_parameters(struct rte_cryptodev *dev,
dpaa_sec_aead_init(dev, xform, session);
} else {
dpaa_sec_aead_init(dev, xform, session);
} else {
-
PMD_DRV_LOG(ERR,
"Invalid crypto type");
+
DPAA_SEC_ERR(
"Invalid crypto type");
return -EINVAL;
}
session->ctx_pool = internals->ctx_pool;
session->inq = dpaa_sec_attach_rxq(internals);
if (session->inq == NULL) {
return -EINVAL;
}
session->ctx_pool = internals->ctx_pool;
session->inq = dpaa_sec_attach_rxq(internals);
if (session->inq == NULL) {
-
PMD_DRV_LOG(ERR,
"unable to attach sec queue");
+
DPAA_SEC_ERR(
"unable to attach sec queue");
goto err1;
}
goto err1;
}
@@
-1878,15
+1870,13
@@
dpaa_sec_session_configure(struct rte_cryptodev *dev,
PMD_INIT_FUNC_TRACE();
if (rte_mempool_get(mempool, &sess_private_data)) {
PMD_INIT_FUNC_TRACE();
if (rte_mempool_get(mempool, &sess_private_data)) {
- CDEV_LOG_ERR(
- "Couldn't get object from session mempool");
+ DPAA_SEC_ERR("Couldn't get object from session mempool");
return -ENOMEM;
}
ret = dpaa_sec_set_session_parameters(dev, xform, sess_private_data);
if (ret != 0) {
return -ENOMEM;
}
ret = dpaa_sec_set_session_parameters(dev, xform, sess_private_data);
if (ret != 0) {
- PMD_DRV_LOG(ERR, "DPAA PMD: failed to configure "
- "session parameters");
+ DPAA_SEC_ERR("failed to configure session parameters");
/* Return session to mempool */
rte_mempool_put(mempool, sess_private_data);
/* Return session to mempool */
rte_mempool_put(mempool, sess_private_data);
@@
-1952,7
+1942,7
@@
dpaa_sec_set_ipsec_session(__rte_unused struct rte_cryptodev *dev,
RTE_CACHE_LINE_SIZE);
if (session->cipher_key.data == NULL &&
cipher_xform->key.length > 0) {
RTE_CACHE_LINE_SIZE);
if (session->cipher_key.data == NULL &&
cipher_xform->key.length > 0) {
-
RTE_LOG(ERR, PMD, "No Memory for cipher key\n
");
+
DPAA_SEC_ERR("No Memory for cipher key
");
return -ENOMEM;
}
return -ENOMEM;
}
@@
-1962,7
+1952,7
@@
dpaa_sec_set_ipsec_session(__rte_unused struct rte_cryptodev *dev,
RTE_CACHE_LINE_SIZE);
if (session->auth_key.data == NULL &&
auth_xform->key.length > 0) {
RTE_CACHE_LINE_SIZE);
if (session->auth_key.data == NULL &&
auth_xform->key.length > 0) {
-
RTE_LOG(ERR, PMD, "No Memory for auth key\n
");
+
DPAA_SEC_ERR("No Memory for auth key
");
rte_free(session->cipher_key.data);
return -ENOMEM;
}
rte_free(session->cipher_key.data);
return -ENOMEM;
}
@@
-2007,11
+1997,11
@@
dpaa_sec_set_ipsec_session(__rte_unused struct rte_cryptodev *dev,
case RTE_CRYPTO_AUTH_KASUMI_F9:
case RTE_CRYPTO_AUTH_AES_CBC_MAC:
case RTE_CRYPTO_AUTH_ZUC_EIA3:
case RTE_CRYPTO_AUTH_KASUMI_F9:
case RTE_CRYPTO_AUTH_AES_CBC_MAC:
case RTE_CRYPTO_AUTH_ZUC_EIA3:
-
RTE_LOG(ERR, PMD, "Crypto: Unsupported auth alg %u\n
",
+
DPAA_SEC_ERR("Crypto: Unsupported auth alg %u
",
auth_xform->algo);
goto out;
default:
auth_xform->algo);
goto out;
default:
-
RTE_LOG(ERR, PMD, "Crypto: Undefined Auth specified %u\n
",
+
DPAA_SEC_ERR("Crypto: Undefined Auth specified %u
",
auth_xform->algo);
goto out;
}
auth_xform->algo);
goto out;
}
@@
-2031,11
+2021,11
@@
dpaa_sec_set_ipsec_session(__rte_unused struct rte_cryptodev *dev,
case RTE_CRYPTO_CIPHER_3DES_ECB:
case RTE_CRYPTO_CIPHER_AES_ECB:
case RTE_CRYPTO_CIPHER_KASUMI_F8:
case RTE_CRYPTO_CIPHER_3DES_ECB:
case RTE_CRYPTO_CIPHER_AES_ECB:
case RTE_CRYPTO_CIPHER_KASUMI_F8:
-
RTE_LOG(ERR, PMD, "Crypto: Unsupported Cipher alg %u\n
",
+
DPAA_SEC_ERR("Crypto: Unsupported Cipher alg %u
",
cipher_xform->algo);
goto out;
default:
cipher_xform->algo);
goto out;
default:
-
RTE_LOG(ERR, PMD, "Crypto: Undefined Cipher specified %u\n
",
+
DPAA_SEC_ERR("Crypto: Undefined Cipher specified %u
",
cipher_xform->algo);
goto out;
}
cipher_xform->algo);
goto out;
}
@@
-2080,7
+2070,7
@@
dpaa_sec_set_ipsec_session(__rte_unused struct rte_cryptodev *dev,
session->ctx_pool = internals->ctx_pool;
session->inq = dpaa_sec_attach_rxq(internals);
if (session->inq == NULL) {
session->ctx_pool = internals->ctx_pool;
session->inq = dpaa_sec_attach_rxq(internals);
if (session->inq == NULL) {
-
PMD_DRV_LOG(ERR,
"unable to attach sec queue");
+
DPAA_SEC_ERR(
"unable to attach sec queue");
goto out;
}
goto out;
}
@@
-2104,8
+2094,7
@@
dpaa_sec_security_session_create(void *dev,
int ret;
if (rte_mempool_get(mempool, &sess_private_data)) {
int ret;
if (rte_mempool_get(mempool, &sess_private_data)) {
- CDEV_LOG_ERR(
- "Couldn't get object from session mempool");
+ DPAA_SEC_ERR("Couldn't get object from session mempool");
return -ENOMEM;
}
return -ENOMEM;
}
@@
-2120,9
+2109,7
@@
dpaa_sec_security_session_create(void *dev,
return -EINVAL;
}
if (ret != 0) {
return -EINVAL;
}
if (ret != 0) {
- PMD_DRV_LOG(ERR,
- "DPAA2 PMD: failed to configure session parameters");
-
+ DPAA_SEC_ERR("failed to configure session parameters");
/* Return session to mempool */
rte_mempool_put(mempool, sess_private_data);
return ret;
/* Return session to mempool */
rte_mempool_put(mempool, sess_private_data);
return ret;
@@
-2176,12
+2163,12
@@
dpaa_sec_dev_configure(struct rte_cryptodev *dev,
NULL, NULL, NULL, NULL,
SOCKET_ID_ANY, 0);
if (!internals->ctx_pool) {
NULL, NULL, NULL, NULL,
SOCKET_ID_ANY, 0);
if (!internals->ctx_pool) {
-
RTE_LOG(ERR, PMD,
"%s create failed\n", str);
+
DPAA_SEC_ERR(
"%s create failed\n", str);
return -ENOMEM;
}
} else
return -ENOMEM;
}
} else
-
RTE_LOG(INFO, PMD, "mempool already created for dev_id : %d\n
",
- dev->data->dev_id);
+
DPAA_SEC_INFO("mempool already created for dev_id : %d
",
+
dev->data->dev_id);
return 0;
}
return 0;
}
@@
-2283,8
+2270,8
@@
dpaa_sec_uninit(struct rte_cryptodev *dev)
rte_mempool_free(internals->ctx_pool);
rte_free(internals);
rte_mempool_free(internals->ctx_pool);
rte_free(internals);
-
PMD_INIT_LOG(INFO, "Closing DPAA_SEC device %s on numa socket %u\n
",
- dev->data->name, rte_socket_id());
+
DPAA_SEC_INFO("Closing DPAA_SEC device %s on numa socket %u
",
+
dev->data->name, rte_socket_id());
return 0;
}
return 0;
}
@@
-2321,7
+2308,7
@@
dpaa_sec_dev_init(struct rte_cryptodev *cryptodev)
* RX function
*/
if (rte_eal_process_type() != RTE_PROC_PRIMARY) {
* RX function
*/
if (rte_eal_process_type() != RTE_PROC_PRIMARY) {
-
PMD_INIT_LOG(DEBUG,
"Device already init by primary process");
+
DPAA_SEC_WARN(
"Device already init by primary process");
return 0;
}
return 0;
}
@@
-2340,7
+2327,7
@@
dpaa_sec_dev_init(struct rte_cryptodev *cryptodev)
qp = &internals->qps[i];
ret = dpaa_sec_init_tx(&qp->outq);
if (ret) {
qp = &internals->qps[i];
ret = dpaa_sec_init_tx(&qp->outq);
if (ret) {
-
PMD_INIT_LOG(ERR,
"config tx of queue pair %d", i);
+
DPAA_SEC_ERR(
"config tx of queue pair %d", i);
goto init_error;
}
}
goto init_error;
}
}
@@
-2351,16
+2338,16
@@
dpaa_sec_dev_init(struct rte_cryptodev *cryptodev)
/* create rx qman fq for sessions*/
ret = qman_create_fq(0, flags, &internals->inq[i]);
if (unlikely(ret != 0)) {
/* create rx qman fq for sessions*/
ret = qman_create_fq(0, flags, &internals->inq[i]);
if (unlikely(ret != 0)) {
-
PMD_INIT_LOG(ERR,
"sec qman_create_fq failed");
+
DPAA_SEC_ERR(
"sec qman_create_fq failed");
goto init_error;
}
}
goto init_error;
}
}
-
PMD_INIT_LOG(DEBUG, "driver %s: created
\n", cryptodev->data->name);
+
RTE_LOG(INFO, PMD, "%s cryptodev init
\n", cryptodev->data->name);
return 0;
init_error:
return 0;
init_error:
-
PMD_INIT_LOG(ERR,
"driver %s: create failed\n", cryptodev->data->name);
+
DPAA_SEC_ERR(
"driver %s: create failed\n", cryptodev->data->name);
dpaa_sec_uninit(cryptodev);
return -EFAULT;
dpaa_sec_uninit(cryptodev);
return -EFAULT;
@@
-2461,3
+2448,12
@@
static struct cryptodev_driver dpaa_sec_crypto_drv;
RTE_PMD_REGISTER_DPAA(CRYPTODEV_NAME_DPAA_SEC_PMD, rte_dpaa_sec_driver);
RTE_PMD_REGISTER_CRYPTO_DRIVER(dpaa_sec_crypto_drv, rte_dpaa_sec_driver.driver,
cryptodev_driver_id);
RTE_PMD_REGISTER_DPAA(CRYPTODEV_NAME_DPAA_SEC_PMD, rte_dpaa_sec_driver);
RTE_PMD_REGISTER_CRYPTO_DRIVER(dpaa_sec_crypto_drv, rte_dpaa_sec_driver.driver,
cryptodev_driver_id);
+
+RTE_INIT(dpaa_sec_init_log);
+static void
+dpaa_sec_init_log(void)
+{
+ dpaa_logtype_sec = rte_log_register("pmd.crypto.dpaa");
+ if (dpaa_logtype_sec >= 0)
+ rte_log_set_level(dpaa_logtype_sec, RTE_LOG_NOTICE);
+}