#include "otx2_ethdev_sec.h"
#include "otx2_ipsec_fp.h"
#include "otx2_sec_idev.h"
-
-#define ETH_SEC_MAX_PKT_LEN 1450
-
-#define AH_HDR_LEN 12
-#define AES_GCM_IV_LEN 8
-#define AES_GCM_MAC_LEN 16
-#define AES_CBC_IV_LEN 16
-#define SHA1_HMAC_LEN 12
-
-#define AES_GCM_ROUNDUP_BYTE_LEN 4
-#define AES_CBC_ROUNDUP_BYTE_LEN 16
+#include "otx2_security.h"
struct eth_sec_tag_const {
RTE_STD_C11
sess->partial_len += sizeof(struct rte_esp_hdr);
sess->roundup_len = sizeof(struct rte_esp_tail);
} else if (ipsec->proto == RTE_SECURITY_IPSEC_SA_PROTO_AH) {
- sess->partial_len += AH_HDR_LEN;
+ sess->partial_len += OTX2_SEC_AH_HDR_LEN;
} else {
return -EINVAL;
}
if (xform->type == RTE_CRYPTO_SYM_XFORM_AEAD) {
if (xform->aead.algo == RTE_CRYPTO_AEAD_AES_GCM) {
- sess->partial_len += AES_GCM_IV_LEN;
- sess->partial_len += AES_GCM_MAC_LEN;
- sess->roundup_byte = AES_GCM_ROUNDUP_BYTE_LEN;
+ sess->partial_len += OTX2_SEC_AES_GCM_IV_LEN;
+ sess->partial_len += OTX2_SEC_AES_GCM_MAC_LEN;
+ sess->roundup_byte = OTX2_SEC_AES_GCM_ROUNDUP_BYTE_LEN;
}
return 0;
}
return -EINVAL;
}
if (cipher_xform->cipher.algo == RTE_CRYPTO_CIPHER_AES_CBC) {
- sess->partial_len += AES_CBC_IV_LEN;
- sess->roundup_byte = AES_CBC_ROUNDUP_BYTE_LEN;
+ sess->partial_len += OTX2_SEC_AES_CBC_IV_LEN;
+ sess->roundup_byte = OTX2_SEC_AES_CBC_ROUNDUP_BYTE_LEN;
} else {
return -EINVAL;
}
if (auth_xform->auth.algo == RTE_CRYPTO_AUTH_SHA1_HMAC)
- sess->partial_len += SHA1_HMAC_LEN;
+ sess->partial_len += OTX2_SEC_SHA1_HMAC_LEN;
else
return -EINVAL;
timeout = rte_get_timer_cycles() + 5 * rte_get_timer_hz();
- rte_cio_wmb();
+ rte_io_wmb();
do {
otx2_lmt_mov(qp->lmtline, &inst, 2);
req->ipsec_cfg0.sa_pow2_size =
rte_log2_u32(sizeof(struct otx2_ipsec_fp_in_sa));
- req->ipsec_cfg0.lenm1_max = ETH_SEC_MAX_PKT_LEN - 1;
+ req->ipsec_cfg0.lenm1_max = NIX_MAX_FRS - 1;
req->ipsec_cfg1.sa_idx_w = rte_log2_u32(dev->ipsec_in_max_spi);
req->ipsec_cfg1.sa_idx_max = dev->ipsec_in_max_spi - 1;