X-Git-Url: http://git.droids-corp.org/?a=blobdiff_plain;f=drivers%2Fmempool%2Focteontx%2Focteontx_fpavf.c;h=94dc5cd8152ad56709314af988ebb14f3f7b501d;hb=37d3bde48bf358e53a74c524095cf9fda9c5e1f3;hp=d14ee47eded7fff0c2ff6b4addeb39f57b24a506;hpb=aaf4363e1e9e518c034c7ff9938a2faefde9854d;p=dpdk.git diff --git a/drivers/mempool/octeontx/octeontx_fpavf.c b/drivers/mempool/octeontx/octeontx_fpavf.c index d14ee47ede..94dc5cd815 100644 --- a/drivers/mempool/octeontx/octeontx_fpavf.c +++ b/drivers/mempool/octeontx/octeontx_fpavf.c @@ -46,20 +46,20 @@ struct octeontx_mbox_fpa_cfg { uint64_t aura_cfg; }; -struct __attribute__((__packed__)) gen_req { +struct __rte_packed gen_req { uint32_t value; }; -struct __attribute__((__packed__)) idn_req { +struct __rte_packed idn_req { uint8_t domain_id; }; -struct __attribute__((__packed__)) gen_resp { +struct __rte_packed gen_resp { uint16_t domain_id; uint16_t vfid; }; -struct __attribute__((__packed__)) dcfg_resp { +struct __rte_packed dcfg_resp { uint8_t sso_count; uint8_t ssow_count; uint8_t fpa_count; @@ -105,24 +105,28 @@ struct octeontx_fpadev { static struct octeontx_fpadev fpadev; +RTE_LOG_REGISTER(octeontx_logtype_fpavf, pmd.mempool.octeontx, NOTICE); + /* lock is taken by caller */ static int octeontx_fpa_gpool_alloc(unsigned int object_size) { + uint16_t global_domain = octeontx_get_global_domain(); struct fpavf_res *res = NULL; - uint16_t gpool; unsigned int sz128; + int i; sz128 = FPA_OBJSZ_2_CACHE_LINE(object_size); - for (gpool = 0; gpool < FPA_VF_MAX; gpool++) { + for (i = 0; i < FPA_VF_MAX; i++) { /* Skip VF that is not mapped Or _inuse */ - if ((fpadev.pool[gpool].bar0 == NULL) || - (fpadev.pool[gpool].is_inuse == true)) + if ((fpadev.pool[i].bar0 == NULL) || + (fpadev.pool[i].is_inuse == true) || + (fpadev.pool[i].domain_id != global_domain)) continue; - res = &fpadev.pool[gpool]; + res = &fpadev.pool[i]; RTE_ASSERT(res->domain_id != (uint16_t)~0); RTE_ASSERT(res->vf_id != (uint16_t)~0); @@ -130,15 +134,34 @@ octeontx_fpa_gpool_alloc(unsigned int object_size) if (res->sz128 == 0) { res->sz128 = sz128; + fpavf_log_dbg("gpool %d blk_sz %d\n", res->vf_id, + sz128); - fpavf_log_dbg("gpool %d blk_sz %d\n", gpool, sz128); - return gpool; + return res->vf_id; } } return -ENOSPC; } +static __rte_always_inline struct fpavf_res * +octeontx_get_fpavf(uint16_t gpool) +{ + uint16_t global_domain = octeontx_get_global_domain(); + int i; + + for (i = 0; i < FPA_VF_MAX; i++) { + if (fpadev.pool[i].domain_id != global_domain) + continue; + if (fpadev.pool[i].vf_id != gpool) + continue; + + return &fpadev.pool[i]; + } + + return NULL; +} + /* lock is taken by caller */ static __rte_always_inline uintptr_t octeontx_fpa_gpool2handle(uint16_t gpool) @@ -146,8 +169,10 @@ octeontx_fpa_gpool2handle(uint16_t gpool) struct fpavf_res *res = NULL; RTE_ASSERT(gpool < FPA_VF_MAX); + res = octeontx_get_fpavf(gpool); + if (res == NULL) + return 0; - res = &fpadev.pool[gpool]; return (uintptr_t)res->bar0 | gpool; } @@ -172,7 +197,7 @@ octeontx_fpa_handle_valid(uintptr_t handle) continue; /* validate gpool */ - if (gpool != i) + if (gpool != fpadev.pool[i].vf_id) return false; res = &fpadev.pool[i]; @@ -202,7 +227,10 @@ octeontx_fpapf_pool_setup(unsigned int gpool, unsigned int buf_size, struct octeontx_mbox_fpa_cfg cfg; int ret = -1; - fpa = &fpadev.pool[gpool]; + fpa = octeontx_get_fpavf(gpool); + if (fpa == NULL) + return -EINVAL; + memsz = FPA_ROUND_UP(max_buf_count / fpa->stack_ln_ptr, FPA_LN_SIZE) * FPA_LN_SIZE; @@ -237,7 +265,7 @@ octeontx_fpapf_pool_setup(unsigned int gpool, unsigned int buf_size, cfg.pool_stack_end = phys_addr + memsz; cfg.aura_cfg = (1 << 9); - ret = octeontx_ssovf_mbox_send(&hdr, &cfg, + ret = octeontx_mbox_send(&hdr, &cfg, sizeof(struct octeontx_mbox_fpa_cfg), &resp, sizeof(resp)); if (ret < 0) { @@ -268,7 +296,9 @@ octeontx_fpapf_pool_destroy(unsigned int gpool_index) struct fpavf_res *fpa = NULL; int ret = -1; - fpa = &fpadev.pool[gpool_index]; + fpa = octeontx_get_fpavf(gpool_index); + if (fpa == NULL) + return -EINVAL; hdr.coproc = FPA_COPROC; hdr.msg = FPA_CONFIGSET; @@ -282,7 +312,7 @@ octeontx_fpapf_pool_destroy(unsigned int gpool_index) cfg.pool_stack_end = 0; cfg.aura_cfg = 0; - ret = octeontx_ssovf_mbox_send(&hdr, &cfg, + ret = octeontx_mbox_send(&hdr, &cfg, sizeof(struct octeontx_mbox_fpa_cfg), &resp, sizeof(resp)); if (ret < 0) { @@ -315,15 +345,16 @@ octeontx_fpapf_aura_attach(unsigned int gpool_index) hdr.vfid = gpool_index; hdr.res_code = 0; memset(&cfg, 0x0, sizeof(struct octeontx_mbox_fpa_cfg)); - cfg.aid = gpool_index; /* gpool is guara */ + cfg.aid = 0; - ret = octeontx_ssovf_mbox_send(&hdr, &cfg, + ret = octeontx_mbox_send(&hdr, &cfg, sizeof(struct octeontx_mbox_fpa_cfg), &resp, sizeof(resp)); if (ret < 0) { fpavf_log_err("Could not attach fpa "); fpavf_log_err("aura %d to pool %d. Err=%d. FuncErr=%d\n", - gpool_index, gpool_index, ret, hdr.res_code); + FPA_AURA_IDX(gpool_index), gpool_index, ret, + hdr.res_code); ret = -EACCES; goto err; } @@ -343,14 +374,15 @@ octeontx_fpapf_aura_detach(unsigned int gpool_index) goto err; } - cfg.aid = gpool_index; /* gpool is gaura */ + cfg.aid = 0; hdr.coproc = FPA_COPROC; hdr.msg = FPA_DETACHAURA; hdr.vfid = gpool_index; - ret = octeontx_ssovf_mbox_send(&hdr, &cfg, sizeof(cfg), NULL, 0); + ret = octeontx_mbox_send(&hdr, &cfg, sizeof(cfg), NULL, 0); if (ret < 0) { fpavf_log_err("Couldn't detach FPA aura %d Err=%d FuncErr=%d\n", - gpool_index, ret, hdr.res_code); + FPA_AURA_IDX(gpool_index), ret, + hdr.res_code); ret = -EINVAL; } @@ -358,8 +390,8 @@ err: return ret; } -static int -octeontx_fpavf_pool_setup(uintptr_t handle, unsigned long memsz, +int +octeontx_fpavf_pool_set_range(uintptr_t handle, unsigned long memsz, void *memva, uint16_t gpool) { uint64_t va_end; @@ -394,7 +426,7 @@ octeontx_fpapf_start_count(uint16_t gpool_index) hdr.coproc = FPA_COPROC; hdr.msg = FPA_START_COUNT; hdr.vfid = gpool_index; - ret = octeontx_ssovf_mbox_send(&hdr, NULL, 0, NULL, 0); + ret = octeontx_mbox_send(&hdr, NULL, 0, NULL, 0); if (ret < 0) { fpavf_log_err("Could not start buffer counting for "); fpavf_log_err("FPA pool %d. Err=%d. FuncErr=%d\n", @@ -410,6 +442,7 @@ err: static __rte_always_inline int octeontx_fpavf_free(unsigned int gpool) { + struct fpavf_res *res = octeontx_get_fpavf(gpool); int ret = 0; if (gpool >= FPA_MAX_POOL) { @@ -418,7 +451,8 @@ octeontx_fpavf_free(unsigned int gpool) } /* Pool is free */ - fpadev.pool[gpool].is_inuse = false; + if (res != NULL) + res->is_inuse = false; err: return ret; @@ -427,8 +461,10 @@ err: static __rte_always_inline int octeontx_gpool_free(uint16_t gpool) { - if (fpadev.pool[gpool].sz128 != 0) { - fpadev.pool[gpool].sz128 = 0; + struct fpavf_res *res = octeontx_get_fpavf(gpool); + + if (res && res->sz128 != 0) { + res->sz128 = 0; return 0; } return -EINVAL; @@ -448,8 +484,8 @@ octeontx_fpa_bufpool_block_size(uintptr_t handle) /* get the gpool */ gpool = octeontx_fpa_bufpool_gpool(handle); - res = &fpadev.pool[gpool]; - return FPA_CACHE_LINE_2_OBJSZ(res->sz128); + res = octeontx_get_fpavf(gpool); + return res ? FPA_CACHE_LINE_2_OBJSZ(res->sz128) : 0; } int @@ -457,6 +493,7 @@ octeontx_fpa_bufpool_free_count(uintptr_t handle) { uint64_t cnt, limit, avail; uint8_t gpool; + uint16_t gaura; uintptr_t pool_bar; if (unlikely(!octeontx_fpa_handle_valid(handle))) @@ -464,14 +501,16 @@ octeontx_fpa_bufpool_free_count(uintptr_t handle) /* get the gpool */ gpool = octeontx_fpa_bufpool_gpool(handle); + /* get the aura */ + gaura = octeontx_fpa_bufpool_gaura(handle); /* Get pool bar address from handle */ pool_bar = handle & ~(uint64_t)FPA_GPOOL_MASK; cnt = fpavf_read64((void *)((uintptr_t)pool_bar + - FPA_VF_VHAURA_CNT(gpool))); + FPA_VF_VHAURA_CNT(gaura))); limit = fpavf_read64((void *)((uintptr_t)pool_bar + - FPA_VF_VHAURA_CNT_LIMIT(gpool))); + FPA_VF_VHAURA_CNT_LIMIT(gaura))); avail = fpavf_read64((void *)((uintptr_t)pool_bar + FPA_VF_VHPOOL_AVAILABLE(gpool))); @@ -481,12 +520,10 @@ octeontx_fpa_bufpool_free_count(uintptr_t handle) uintptr_t octeontx_fpa_bufpool_create(unsigned int object_size, unsigned int object_count, - unsigned int buf_offset, char **va_start, - int node_id) + unsigned int buf_offset, int node_id) { unsigned int gpool; - void *memva; - unsigned long memsz; + unsigned int gaura; uintptr_t gpool_handle; uintptr_t pool_bar; int res; @@ -494,9 +531,7 @@ octeontx_fpa_bufpool_create(unsigned int object_size, unsigned int object_count, RTE_SET_USED(node_id); RTE_BUILD_BUG_ON(sizeof(struct rte_mbuf) > OCTEONTX_FPAVF_BUF_OFFSET); - if (unlikely(*va_start == NULL)) - goto error_end; - + octeontx_mbox_init(); object_size = RTE_CACHE_LINE_ROUNDUP(object_size); if (object_size > FPA_MAX_OBJ_SIZE) { errno = EINVAL; @@ -539,32 +574,23 @@ octeontx_fpa_bufpool_create(unsigned int object_size, unsigned int object_count, goto error_pool_destroy; } - /* vf pool setup */ - memsz = object_size * object_count; - memva = *va_start; - res = octeontx_fpavf_pool_setup(pool_bar, memsz, memva, gpool); - if (res < 0) { - errno = res; - goto error_gaura_detach; - } + gaura = FPA_AURA_IDX(gpool); /* Release lock */ rte_spinlock_unlock(&fpadev.lock); /* populate AURA registers */ fpavf_write64(object_count, (void *)((uintptr_t)pool_bar + - FPA_VF_VHAURA_CNT(gpool))); + FPA_VF_VHAURA_CNT(gaura))); fpavf_write64(object_count, (void *)((uintptr_t)pool_bar + - FPA_VF_VHAURA_CNT_LIMIT(gpool))); + FPA_VF_VHAURA_CNT_LIMIT(gaura))); fpavf_write64(object_count + 1, (void *)((uintptr_t)pool_bar + - FPA_VF_VHAURA_CNT_THRESHOLD(gpool))); + FPA_VF_VHAURA_CNT_THRESHOLD(gaura))); octeontx_fpapf_start_count(gpool); return gpool_handle; -error_gaura_detach: - (void) octeontx_fpapf_aura_detach(gpool); error_pool_destroy: octeontx_fpavf_free(gpool); octeontx_fpapf_pool_destroy(gpool); @@ -586,6 +612,7 @@ octeontx_fpa_bufpool_destroy(uintptr_t handle, int node_id) uint64_t sz; uint64_t cnt, avail; uint8_t gpool; + uint16_t gaura; uintptr_t pool_bar; int ret; @@ -599,13 +626,15 @@ octeontx_fpa_bufpool_destroy(uintptr_t handle, int node_id) /* get the pool */ gpool = octeontx_fpa_bufpool_gpool(handle); + /* get the aura */ + gaura = octeontx_fpa_bufpool_gaura(handle); /* Get pool bar address from handle */ pool_bar = handle & ~(uint64_t)FPA_GPOOL_MASK; /* Check for no outstanding buffers */ cnt = fpavf_read64((void *)((uintptr_t)pool_bar + - FPA_VF_VHAURA_CNT(gpool))); + FPA_VF_VHAURA_CNT(gaura))); if (cnt) { fpavf_log_dbg("buffer exist in pool cnt %" PRId64 "\n", cnt); return -EBUSY; @@ -618,9 +647,9 @@ octeontx_fpa_bufpool_destroy(uintptr_t handle, int node_id) /* Prepare to empty the entire POOL */ fpavf_write64(avail, (void *)((uintptr_t)pool_bar + - FPA_VF_VHAURA_CNT_LIMIT(gpool))); + FPA_VF_VHAURA_CNT_LIMIT(gaura))); fpavf_write64(avail + 1, (void *)((uintptr_t)pool_bar + - FPA_VF_VHAURA_CNT_THRESHOLD(gpool))); + FPA_VF_VHAURA_CNT_THRESHOLD(gaura))); /* Empty the pool */ /* Invalidate the POOL */ @@ -632,11 +661,11 @@ octeontx_fpa_bufpool_destroy(uintptr_t handle, int node_id) /* Yank a buffer from the pool */ node = (void *)(uintptr_t) fpavf_read64((void *) - (pool_bar + FPA_VF_VHAURA_OP_ALLOC(gpool))); + (pool_bar + FPA_VF_VHAURA_OP_ALLOC(gaura))); if (node == NULL) { fpavf_log_err("GAURA[%u] missing %" PRIx64 " buf\n", - gpool, avail); + gaura, avail); break; } @@ -670,9 +699,9 @@ octeontx_fpa_bufpool_destroy(uintptr_t handle, int node_id) /* Deactivate the AURA */ fpavf_write64(0, (void *)((uintptr_t)pool_bar + - FPA_VF_VHAURA_CNT_LIMIT(gpool))); + FPA_VF_VHAURA_CNT_LIMIT(gaura))); fpavf_write64(0, (void *)((uintptr_t)pool_bar + - FPA_VF_VHAURA_CNT_THRESHOLD(gpool))); + FPA_VF_VHAURA_CNT_THRESHOLD(gaura))); ret = octeontx_fpapf_aura_detach(gpool); if (ret) { @@ -717,6 +746,7 @@ octeontx_fpavf_identify(void *bar0) uint16_t domain_id; uint16_t vf_id; uint64_t stack_ln_ptr; + static uint16_t vf_idx; val = fpavf_read64((void *)((uintptr_t)bar0 + FPA_VF_VHAURA_CNT_THRESHOLD(0))); @@ -726,23 +756,18 @@ octeontx_fpavf_identify(void *bar0) stack_ln_ptr = fpavf_read64((void *)((uintptr_t)bar0 + FPA_VF_VHPOOL_THRESHOLD(0))); - if (vf_id >= FPA_VF_MAX) { + if (vf_idx >= FPA_VF_MAX) { fpavf_log_err("vf_id(%d) greater than max vf (32)\n", vf_id); - return -1; - } - - if (fpadev.pool[vf_id].is_inuse) { - fpavf_log_err("vf_id %d is_inuse\n", vf_id); - return -1; + return -E2BIG; } - fpadev.pool[vf_id].domain_id = domain_id; - fpadev.pool[vf_id].vf_id = vf_id; - fpadev.pool[vf_id].bar0 = bar0; - fpadev.pool[vf_id].stack_ln_ptr = stack_ln_ptr; + fpadev.pool[vf_idx].domain_id = domain_id; + fpadev.pool[vf_idx].vf_id = vf_id; + fpadev.pool[vf_idx].bar0 = bar0; + fpadev.pool[vf_idx].stack_ln_ptr = stack_ln_ptr; /* SUCCESS */ - return vf_id; + return vf_idx++; } /* FPAVF pcie device aka mempool probe */ @@ -795,7 +820,7 @@ static const struct rte_pci_id pci_fpavf_map[] = { static struct rte_pci_driver pci_fpavf = { .id_table = pci_fpavf_map, - .drv_flags = RTE_PCI_DRV_NEED_MAPPING | RTE_PCI_DRV_IOVA_AS_VA, + .drv_flags = RTE_PCI_DRV_NEED_MAPPING | RTE_PCI_DRV_NEED_IOVA_AS_VA, .probe = fpavf_probe, };