This patch moves the internal symbols to INTERNAL sections
so that any change in them is not reported as ABI breakage.
This patch also removes two symbols, which are not to be exported.
rte_dpaa_mem_ptov - static inline in the headerfile
fman_ccsr_map_fd - local shared variable.
Signed-off-by: Hemant Agrawal <hemant.agrawal@nxp.com>
Acked-by: Ray Kinsella <mdr@ashroe.eu>
file_name_regexp = ^librte_common_dpaax\.
[suppress_file]
file_name_regexp = ^librte_bus_fslmc\.
file_name_regexp = ^librte_common_dpaax\.
[suppress_file]
file_name_regexp = ^librte_bus_fslmc\.
+[suppress_file]
+ file_name_regexp = ^librte_bus_dpaa\.
* the structure provided by the caller can be released or reused after the
* function returns.
*/
* the structure provided by the caller can be released or reused after the
* function returns.
*/
struct bman_pool *bman_new_pool(const struct bman_pool_params *params);
/**
* bman_free_pool - Deallocates a Buffer Pool object
* @pool: the pool object to release
*/
struct bman_pool *bman_new_pool(const struct bman_pool_params *params);
/**
* bman_free_pool - Deallocates a Buffer Pool object
* @pool: the pool object to release
*/
void bman_free_pool(struct bman_pool *pool);
/**
void bman_free_pool(struct bman_pool *pool);
/**
* The returned pointer refers to state within the pool object so must not be
* modified and can no longer be read once the pool object is destroyed.
*/
* The returned pointer refers to state within the pool object so must not be
* modified and can no longer be read once the pool object is destroyed.
*/
const struct bman_pool_params *bman_get_params(const struct bman_pool *pool);
/**
const struct bman_pool_params *bman_get_params(const struct bman_pool *pool);
/**
* @flags: bit-mask of BMAN_RELEASE_FLAG_*** options
*
*/
* @flags: bit-mask of BMAN_RELEASE_FLAG_*** options
*
*/
int bman_release(struct bman_pool *pool, const struct bm_buffer *bufs, u8 num,
u32 flags);
int bman_release(struct bman_pool *pool, const struct bm_buffer *bufs, u8 num,
u32 flags);
* The return value will be the number of buffers obtained from the pool, or a
* negative error code if a h/w error or pool starvation was encountered.
*/
* The return value will be the number of buffers obtained from the pool, or a
* negative error code if a h/w error or pool starvation was encountered.
*/
int bman_acquire(struct bman_pool *pool, struct bm_buffer *bufs, u8 num,
u32 flags);
int bman_acquire(struct bman_pool *pool, struct bm_buffer *bufs, u8 num,
u32 flags);
*
* Return the number of the free buffers
*/
*
* Return the number of the free buffers
*/
u32 bman_query_free_buffers(struct bman_pool *pool);
/**
u32 bman_query_free_buffers(struct bman_pool *pool);
/**
#ifndef __FSL_FMAN_H
#define __FSL_FMAN_H
#ifndef __FSL_FMAN_H
#define __FSL_FMAN_H
+#include <rte_compat.h>
+
#ifdef __cplusplus
extern "C" {
#endif
#ifdef __cplusplus
extern "C" {
#endif
} __rte_packed;
/* Set MAC address for a particular interface */
} __rte_packed;
/* Set MAC address for a particular interface */
int fman_if_add_mac_addr(struct fman_if *p, uint8_t *eth, uint8_t addr_num);
/* Remove a MAC address for a particular interface */
int fman_if_add_mac_addr(struct fman_if *p, uint8_t *eth, uint8_t addr_num);
/* Remove a MAC address for a particular interface */
void fman_if_clear_mac_addr(struct fman_if *p, uint8_t addr_num);
/* Get the FMAN statistics */
void fman_if_clear_mac_addr(struct fman_if *p, uint8_t addr_num);
/* Get the FMAN statistics */
void fman_if_stats_get(struct fman_if *p, struct rte_eth_stats *stats);
/* Reset the FMAN statistics */
void fman_if_stats_get(struct fman_if *p, struct rte_eth_stats *stats);
/* Reset the FMAN statistics */
void fman_if_stats_reset(struct fman_if *p);
/* Get all of the FMAN statistics */
void fman_if_stats_reset(struct fman_if *p);
/* Get all of the FMAN statistics */
void fman_if_stats_get_all(struct fman_if *p, uint64_t *value, int n);
/* Set ignore pause option for a specific interface */
void fman_if_stats_get_all(struct fman_if *p, uint64_t *value, int n);
/* Set ignore pause option for a specific interface */
void fman_if_conf_max_frame_len(struct fman_if *p, unsigned int max_frame_len);
/* Enable/disable Rx promiscuous mode on specified interface */
void fman_if_conf_max_frame_len(struct fman_if *p, unsigned int max_frame_len);
/* Enable/disable Rx promiscuous mode on specified interface */
void fman_if_promiscuous_enable(struct fman_if *p);
void fman_if_promiscuous_enable(struct fman_if *p);
void fman_if_promiscuous_disable(struct fman_if *p);
/* Enable/disable Rx on specific interfaces */
void fman_if_promiscuous_disable(struct fman_if *p);
/* Enable/disable Rx on specific interfaces */
void fman_if_enable_rx(struct fman_if *p);
void fman_if_enable_rx(struct fman_if *p);
void fman_if_disable_rx(struct fman_if *p);
/* Enable/disable loopback on specific interfaces */
void fman_if_disable_rx(struct fman_if *p);
/* Enable/disable loopback on specific interfaces */
void fman_if_loopback_enable(struct fman_if *p);
void fman_if_loopback_enable(struct fman_if *p);
void fman_if_loopback_disable(struct fman_if *p);
/* Set buffer pool on specific interface */
void fman_if_loopback_disable(struct fman_if *p);
/* Set buffer pool on specific interface */
void fman_if_set_bp(struct fman_if *fm_if, unsigned int num, int bpid,
size_t bufsize);
/* Get Flow Control threshold parameters on specific interface */
void fman_if_set_bp(struct fman_if *fm_if, unsigned int num, int bpid,
size_t bufsize);
/* Get Flow Control threshold parameters on specific interface */
int fman_if_get_fc_threshold(struct fman_if *fm_if);
/* Enable and Set Flow Control threshold parameters on specific interface */
int fman_if_get_fc_threshold(struct fman_if *fm_if);
/* Enable and Set Flow Control threshold parameters on specific interface */
int fman_if_set_fc_threshold(struct fman_if *fm_if,
u32 high_water, u32 low_water, u32 bpid);
/* Get Flow Control pause quanta on specific interface */
int fman_if_set_fc_threshold(struct fman_if *fm_if,
u32 high_water, u32 low_water, u32 bpid);
/* Get Flow Control pause quanta on specific interface */
int fman_if_get_fc_quanta(struct fman_if *fm_if);
/* Set Flow Control pause quanta on specific interface */
int fman_if_get_fc_quanta(struct fman_if *fm_if);
/* Set Flow Control pause quanta on specific interface */
int fman_if_set_fc_quanta(struct fman_if *fm_if, u16 pause_quanta);
/* Set default error fqid on specific interface */
int fman_if_set_fc_quanta(struct fman_if *fm_if, u16 pause_quanta);
/* Set default error fqid on specific interface */
int fman_if_get_ic_params(struct fman_if *fm_if, struct fman_if_ic_params *icp);
/* Set IC transfer params */
int fman_if_get_ic_params(struct fman_if *fm_if, struct fman_if_ic_params *icp);
/* Set IC transfer params */
int fman_if_set_ic_params(struct fman_if *fm_if,
const struct fman_if_ic_params *icp);
/* Get interface fd->offset value */
int fman_if_set_ic_params(struct fman_if *fm_if,
const struct fman_if_ic_params *icp);
/* Get interface fd->offset value */
int fman_if_get_fdoff(struct fman_if *fm_if);
/* Set interface fd->offset value */
int fman_if_get_fdoff(struct fman_if *fm_if);
/* Set interface fd->offset value */
void fman_if_set_fdoff(struct fman_if *fm_if, uint32_t fd_offset);
/* Get interface SG enable status value */
void fman_if_set_fdoff(struct fman_if *fm_if, uint32_t fd_offset);
/* Get interface SG enable status value */
int fman_if_get_sg_enable(struct fman_if *fm_if);
/* Set interface SG support mode */
int fman_if_get_sg_enable(struct fman_if *fm_if);
/* Set interface SG support mode */
void fman_if_set_sg(struct fman_if *fm_if, int enable);
/* Get interface Max Frame length (MTU) */
uint16_t fman_if_get_maxfrm(struct fman_if *fm_if);
/* Set interface Max Frame length (MTU) */
void fman_if_set_sg(struct fman_if *fm_if, int enable);
/* Get interface Max Frame length (MTU) */
uint16_t fman_if_get_maxfrm(struct fman_if *fm_if);
/* Set interface Max Frame length (MTU) */
void fman_if_set_maxfrm(struct fman_if *fm_if, uint16_t max_frm);
/* Set interface next invoked action for dequeue operation */
void fman_if_set_dnia(struct fman_if *fm_if, uint32_t nia);
/* discard error packets on rx */
void fman_if_set_maxfrm(struct fman_if *fm_if, uint16_t max_frm);
/* Set interface next invoked action for dequeue operation */
void fman_if_set_dnia(struct fman_if *fm_if, uint32_t nia);
/* discard error packets on rx */
void fman_if_discard_rx_errors(struct fman_if *fm_if);
void fman_if_discard_rx_errors(struct fman_if *fm_if);
void fman_if_set_mcast_filter_table(struct fman_if *p);
void fman_if_set_mcast_filter_table(struct fman_if *p);
void fman_if_reset_mcast_filter_table(struct fman_if *p);
int fman_if_add_hash_mac_addr(struct fman_if *p, uint8_t *eth);
void fman_if_reset_mcast_filter_table(struct fman_if *p);
int fman_if_add_hash_mac_addr(struct fman_if *p, uint8_t *eth);
#define QMAN_CGR_MODE_FRAME 0x00000001
#ifdef CONFIG_FSL_QMAN_FQ_LOOKUP
#define QMAN_CGR_MODE_FRAME 0x00000001
#ifdef CONFIG_FSL_QMAN_FQ_LOOKUP
void qman_set_fq_lookup_table(void **table);
#endif
void qman_set_fq_lookup_table(void **table);
#endif
*/
int qman_get_portal_index(void);
*/
int qman_get_portal_index(void);
u32 qman_portal_dequeue(struct rte_event ev[], unsigned int poll_limit,
void **bufs);
u32 qman_portal_dequeue(struct rte_event ev[], unsigned int poll_limit,
void **bufs);
* processed via qman_poll_***() functions). Returns zero for success, or
* -EINVAL if the current CPU is sharing a portal hosted on another CPU.
*/
* processed via qman_poll_***() functions). Returns zero for success, or
* -EINVAL if the current CPU is sharing a portal hosted on another CPU.
*/
int qman_irqsource_add(u32 bits);
/**
int qman_irqsource_add(u32 bits);
/**
* takes portal (fq specific) as input rather than using the thread affined
* portal.
*/
* takes portal (fq specific) as input rather than using the thread affined
* portal.
*/
int qman_fq_portal_irqsource_add(struct qman_portal *p, u32 bits);
/**
int qman_fq_portal_irqsource_add(struct qman_portal *p, u32 bits);
/**
* instead be processed via qman_poll_***() functions. Returns zero for success,
* or -EINVAL if the current CPU is sharing a portal hosted on another CPU.
*/
* instead be processed via qman_poll_***() functions. Returns zero for success,
* or -EINVAL if the current CPU is sharing a portal hosted on another CPU.
*/
int qman_irqsource_remove(u32 bits);
/**
int qman_irqsource_remove(u32 bits);
/**
* takes portal (fq specific) as input rather than using the thread affined
* portal.
*/
* takes portal (fq specific) as input rather than using the thread affined
* portal.
*/
int qman_fq_portal_irqsource_remove(struct qman_portal *p, u32 bits);
/**
int qman_fq_portal_irqsource_remove(struct qman_portal *p, u32 bits);
/**
*/
u16 qman_affine_channel(int cpu);
*/
u16 qman_affine_channel(int cpu);
unsigned int qman_portal_poll_rx(unsigned int poll_limit,
void **bufs, struct qman_portal *q);
unsigned int qman_portal_poll_rx(unsigned int poll_limit,
void **bufs, struct qman_portal *q);
*
* This function will issue a volatile dequeue command to the QMAN.
*/
*
* This function will issue a volatile dequeue command to the QMAN.
*/
int qman_set_vdq(struct qman_fq *fq, u16 num, uint32_t vdqcr_flags);
/**
int qman_set_vdq(struct qman_fq *fq, u16 num, uint32_t vdqcr_flags);
/**
* is issued. It will keep returning NULL until there is no packet available on
* the DQRR.
*/
* is issued. It will keep returning NULL until there is no packet available on
* the DQRR.
*/
struct qm_dqrr_entry *qman_dequeue(struct qman_fq *fq);
/**
struct qm_dqrr_entry *qman_dequeue(struct qman_fq *fq);
/**
* This will consume the DQRR enrey and make it available for next volatile
* dequeue.
*/
* This will consume the DQRR enrey and make it available for next volatile
* dequeue.
*/
void qman_dqrr_consume(struct qman_fq *fq,
struct qm_dqrr_entry *dq);
void qman_dqrr_consume(struct qman_fq *fq,
struct qm_dqrr_entry *dq);
* this function will return -EINVAL, otherwise the return value is >=0 and
* represents the number of DQRR entries processed.
*/
* this function will return -EINVAL, otherwise the return value is >=0 and
* represents the number of DQRR entries processed.
*/
int qman_poll_dqrr(unsigned int limit);
/**
int qman_poll_dqrr(unsigned int limit);
/**
* (SDQCR). The requested pools are limited to those the portal has dequeue
* access to.
*/
* (SDQCR). The requested pools are limited to those the portal has dequeue
* access to.
*/
void qman_static_dequeue_add(u32 pools, struct qman_portal *qm);
/**
void qman_static_dequeue_add(u32 pools, struct qman_portal *qm);
/**
* function must be called from the same CPU as that which processed the DQRR
* entry in the first place.
*/
* function must be called from the same CPU as that which processed the DQRR
* entry in the first place.
*/
void qman_dca_index(u8 index, int park_request);
/**
void qman_dca_index(u8 index, int park_request);
/**
* a frame queue object based on that, rather than assuming/requiring that it be
* Out of Service.
*/
* a frame queue object based on that, rather than assuming/requiring that it be
* Out of Service.
*/
int qman_create_fq(u32 fqid, u32 flags, struct qman_fq *fq);
/**
int qman_create_fq(u32 fqid, u32 flags, struct qman_fq *fq);
/**
* qman_fq_fqid - Queries the frame queue ID of a FQ object
* @fq: the frame queue object to query
*/
* qman_fq_fqid - Queries the frame queue ID of a FQ object
* @fq: the frame queue object to query
*/
u32 qman_fq_fqid(struct qman_fq *fq);
/**
u32 qman_fq_fqid(struct qman_fq *fq);
/**
* This captures the state, as seen by the driver, at the time the function
* executes.
*/
* This captures the state, as seen by the driver, at the time the function
* executes.
*/
void qman_fq_state(struct qman_fq *fq, enum qman_fq_state *state, u32 *flags);
/**
void qman_fq_state(struct qman_fq *fq, enum qman_fq_state *state, u32 *flags);
/**
* context_a.address fields and will leave the stashing fields provided by the
* user alone, otherwise it will zero out the context_a.stashing fields.
*/
* context_a.address fields and will leave the stashing fields provided by the
* user alone, otherwise it will zero out the context_a.stashing fields.
*/
int qman_init_fq(struct qman_fq *fq, u32 flags, struct qm_mcc_initfq *opts);
/**
int qman_init_fq(struct qman_fq *fq, u32 flags, struct qm_mcc_initfq *opts);
/**
* caller should be prepared to accept the callback as the function is called,
* not only once it has returned.
*/
* caller should be prepared to accept the callback as the function is called,
* not only once it has returned.
*/
int qman_retire_fq(struct qman_fq *fq, u32 *flags);
/**
int qman_retire_fq(struct qman_fq *fq, u32 *flags);
/**
* The frame queue must be retired and empty, and if any order restoration list
* was released as ERNs at the time of retirement, they must all be consumed.
*/
* The frame queue must be retired and empty, and if any order restoration list
* was released as ERNs at the time of retirement, they must all be consumed.
*/
int qman_oos_fq(struct qman_fq *fq);
/**
int qman_oos_fq(struct qman_fq *fq);
/**
* @fq: the frame queue object to be queried
* @np: storage for the queried FQD fields
*/
* @fq: the frame queue object to be queried
* @np: storage for the queried FQD fields
*/
int qman_query_fq_np(struct qman_fq *fq, struct qm_mcr_queryfq_np *np);
/**
int qman_query_fq_np(struct qman_fq *fq, struct qm_mcr_queryfq_np *np);
/**
* @fq: the frame queue object to be queried
* @frm_cnt: number of frames in the queue
*/
* @fq: the frame queue object to be queried
* @frm_cnt: number of frames in the queue
*/
int qman_query_fq_frm_cnt(struct qman_fq *fq, u32 *frm_cnt);
/**
int qman_query_fq_frm_cnt(struct qman_fq *fq, u32 *frm_cnt);
/**
* callback, or by waiting for the QMAN_FQ_STATE_VDQCR bit to disappear from the
* "flags" retrieved from qman_fq_state().
*/
* callback, or by waiting for the QMAN_FQ_STATE_VDQCR bit to disappear from the
* "flags" retrieved from qman_fq_state().
*/
int qman_volatile_dequeue(struct qman_fq *fq, u32 flags, u32 vdqcr);
/**
int qman_volatile_dequeue(struct qman_fq *fq, u32 flags, u32 vdqcr);
/**
* of an already busy hardware resource by throttling many of the to-be-dropped
* enqueues "at the source".
*/
* of an already busy hardware resource by throttling many of the to-be-dropped
* enqueues "at the source".
*/
int qman_enqueue(struct qman_fq *fq, const struct qm_fd *fd, u32 flags);
int qman_enqueue(struct qman_fq *fq, const struct qm_fd *fd, u32 flags);
int qman_enqueue_multi(struct qman_fq *fq, const struct qm_fd *fd, u32 *flags,
int frames_to_send);
int qman_enqueue_multi(struct qman_fq *fq, const struct qm_fd *fd, u32 *flags,
int frames_to_send);
* This API is similar to qman_enqueue_multi(), but it takes fd which needs
* to be processed by different frame queues.
*/
* This API is similar to qman_enqueue_multi(), but it takes fd which needs
* to be processed by different frame queues.
*/
int
qman_enqueue_multi_fq(struct qman_fq *fq[], const struct qm_fd *fd,
u32 *flags, int frames_to_send);
int
qman_enqueue_multi_fq(struct qman_fq *fq[], const struct qm_fd *fd,
u32 *flags, int frames_to_send);
* @fqid: the base FQID of the range to deallocate
* @count: the number of FQIDs in the range
*/
* @fqid: the base FQID of the range to deallocate
* @count: the number of FQIDs in the range
*/
int qman_reserve_fqid_range(u32 fqid, unsigned int count);
static inline int qman_reserve_fqid(u32 fqid)
{
int qman_reserve_fqid_range(u32 fqid, unsigned int count);
static inline int qman_reserve_fqid(u32 fqid)
{
* than requested (though alignment will be as requested). If @partial is zero,
* the return value will either be 'count' or negative.
*/
* than requested (though alignment will be as requested). If @partial is zero,
* the return value will either be 'count' or negative.
*/
int qman_alloc_pool_range(u32 *result, u32 count, u32 align, int partial);
static inline int qman_alloc_pool(u32 *result)
{
int qman_alloc_pool_range(u32 *result, u32 count, u32 align, int partial);
static inline int qman_alloc_pool(u32 *result)
{
* any unspecified parameters) will be used rather than a modify hw hardware
* (which only modifies the specified parameters).
*/
* any unspecified parameters) will be used rather than a modify hw hardware
* (which only modifies the specified parameters).
*/
int qman_create_cgr(struct qman_cgr *cgr, u32 flags,
struct qm_mcc_initcgr *opts);
int qman_create_cgr(struct qman_cgr *cgr, u32 flags,
struct qm_mcc_initcgr *opts);
* is executed. This must be excuted on the same affine portal on which it was
* created.
*/
* is executed. This must be excuted on the same affine portal on which it was
* created.
*/
int qman_delete_cgr(struct qman_cgr *cgr);
/**
int qman_delete_cgr(struct qman_cgr *cgr);
/**
* unspecified parameters) will be used rather than a modify hw hardware (which
* only modifies the specified parameters).
*/
* unspecified parameters) will be used rather than a modify hw hardware (which
* only modifies the specified parameters).
*/
int qman_modify_cgr(struct qman_cgr *cgr, u32 flags,
struct qm_mcc_initcgr *opts);
int qman_modify_cgr(struct qman_cgr *cgr, u32 flags,
struct qm_mcc_initcgr *opts);
* than requested (though alignment will be as requested). If @partial is zero,
* the return value will either be 'count' or negative.
*/
* than requested (though alignment will be as requested). If @partial is zero,
* the return value will either be 'count' or negative.
*/
int qman_alloc_cgrid_range(u32 *result, u32 count, u32 align, int partial);
static inline int qman_alloc_cgrid(u32 *result)
{
int qman_alloc_cgrid_range(u32 *result, u32 count, u32 align, int partial);
static inline int qman_alloc_cgrid(u32 *result)
{
* @id: the base CGR ID of the range to deallocate
* @count: the number of CGR IDs in the range
*/
* @id: the base CGR ID of the range to deallocate
* @count: the number of CGR IDs in the range
*/
void qman_release_cgrid_range(u32 id, unsigned int count);
static inline void qman_release_cgrid(u32 id)
{
void qman_release_cgrid_range(u32 id, unsigned int count);
static inline void qman_release_cgrid(u32 id)
{
int bman_free_raw_portal(struct dpaa_raw_portal *portal);
/* Obtain thread-local UIO file-descriptors */
int bman_free_raw_portal(struct dpaa_raw_portal *portal);
/* Obtain thread-local UIO file-descriptors */
int qman_thread_fd(void);
int bman_thread_fd(void);
int qman_thread_fd(void);
int bman_thread_fd(void);
* processing is complete. As such, it is essential to call this before going
* into another blocking read/select/poll.
*/
* processing is complete. As such, it is essential to call this before going
* into another blocking read/select/poll.
*/
void qman_thread_irq(void);
void qman_thread_irq(void);
void bman_thread_irq(void);
void bman_thread_irq(void);
void qman_fq_portal_thread_irq(struct qman_portal *qp);
void qman_fq_portal_thread_irq(struct qman_portal *qp);
void qman_clear_irq(void);
/* Global setup */
void qman_clear_irq(void);
/* Global setup */
int bman_global_init(void);
/* Direct portal create and destroy */
int bman_global_init(void);
/* Direct portal create and destroy */
struct qman_portal *fsl_qman_fq_portal_create(int *fd);
int fsl_qman_fq_portal_destroy(struct qman_portal *qp);
int fsl_qman_fq_portal_init(struct qman_portal *qp);
struct qman_portal *fsl_qman_fq_portal_create(int *fd);
int fsl_qman_fq_portal_destroy(struct qman_portal *qp);
int fsl_qman_fq_portal_init(struct qman_portal *qp);
* cfg_file: FMC config XML file
* Returns the configuration information in newly allocated memory.
*/
* cfg_file: FMC config XML file
* Returns the configuration information in newly allocated memory.
*/
struct netcfg_info *netcfg_acquire(void);
/* cfg_ptr: configuration information pointer.
* Frees the resources allocated by the configuration layer.
*/
struct netcfg_info *netcfg_acquire(void);
/* cfg_ptr: configuration information pointer.
* Frees the resources allocated by the configuration layer.
*/
void netcfg_release(struct netcfg_info *cfg_ptr);
#ifdef RTE_LIBRTE_DPAA_DEBUG_DRIVER
void netcfg_release(struct netcfg_info *cfg_ptr);
#ifdef RTE_LIBRTE_DPAA_DEBUG_DRIVER
+ local: *;
+};
+
+INTERNAL {
dpaa_logtype_pmd;
dpaa_netcfg;
dpaa_svr_family;
dpaa_logtype_pmd;
dpaa_netcfg;
dpaa_svr_family;
fman_dealloc_bufs_mask_hi;
fman_dealloc_bufs_mask_lo;
fman_if_add_mac_addr;
fman_dealloc_bufs_mask_hi;
fman_dealloc_bufs_mask_lo;
fman_if_add_mac_addr;
qman_volatile_dequeue;
rte_dpaa_driver_register;
rte_dpaa_driver_unregister;
qman_volatile_dequeue;
rte_dpaa_driver_register;
rte_dpaa_driver_unregister;
rte_dpaa_portal_fq_close;
rte_dpaa_portal_fq_init;
rte_dpaa_portal_init;
rte_dpaa_portal_fq_close;
rte_dpaa_portal_fq_init;
rte_dpaa_portal_init;
* A pointer to a rte_dpaa_driver structure describing the driver
* to be registered.
*/
* A pointer to a rte_dpaa_driver structure describing the driver
* to be registered.
*/
void rte_dpaa_driver_register(struct rte_dpaa_driver *driver);
/**
void rte_dpaa_driver_register(struct rte_dpaa_driver *driver);
/**
* A pointer to a rte_dpaa_driver structure describing the driver
* to be unregistered.
*/
* A pointer to a rte_dpaa_driver structure describing the driver
* to be unregistered.
*/
void rte_dpaa_driver_unregister(struct rte_dpaa_driver *driver);
/**
void rte_dpaa_driver_unregister(struct rte_dpaa_driver *driver);
/**
* @return
* 0 in case of success, error otherwise
*/
* @return
* 0 in case of success, error otherwise
*/
int rte_dpaa_portal_init(void *arg);
int rte_dpaa_portal_init(void *arg);
int rte_dpaa_portal_fq_init(void *arg, struct qman_fq *fq);
int rte_dpaa_portal_fq_init(void *arg, struct qman_fq *fq);
int rte_dpaa_portal_fq_close(struct qman_fq *fq);
/**
int rte_dpaa_portal_fq_close(struct qman_fq *fq);
/**