f-stack/dpdk/lib/librte_eal/common/eal_private.h

742 lines
18 KiB
C
Raw Normal View History

2019-06-25 11:12:58 +00:00
/* SPDX-License-Identifier: BSD-3-Clause
* Copyright(c) 2010-2018 Intel Corporation
2017-04-21 10:43:26 +00:00
*/
#ifndef _EAL_PRIVATE_H_
#define _EAL_PRIVATE_H_
#include <stdbool.h>
#include <stdint.h>
2017-04-21 10:43:26 +00:00
#include <stdio.h>
2019-06-25 11:12:58 +00:00
#include <rte_dev.h>
2020-06-18 16:55:50 +00:00
#include <rte_lcore.h>
2021-02-05 08:48:47 +00:00
#include <rte_memory.h>
#include "eal_internal_cfg.h"
2020-06-18 16:55:50 +00:00
/**
* Structure storing internal configuration (per-lcore)
*/
struct lcore_config {
pthread_t thread_id; /**< pthread identifier */
2021-02-05 08:48:47 +00:00
int pipe_main2worker[2]; /**< communication pipe with main */
int pipe_worker2main[2]; /**< communication pipe with main */
2020-06-18 16:55:50 +00:00
lcore_function_t * volatile f; /**< function to call */
void * volatile arg; /**< argument of function */
volatile int ret; /**< return value of function */
volatile enum rte_lcore_state_t state; /**< lcore state */
unsigned int socket_id; /**< physical socket id for this lcore */
unsigned int core_id; /**< core number on socket for this lcore */
int core_index; /**< relative index, starting from 0 */
uint8_t core_role; /**< role of core eg: OFF, RTE, SERVICE */
rte_cpuset_t cpuset; /**< cpu set which the lcore affinity to */
};
extern struct lcore_config lcore_config[RTE_MAX_LCORE];
/**
* The global RTE configuration structure.
*/
struct rte_config {
2021-02-05 08:48:47 +00:00
uint32_t main_lcore; /**< Id of the main lcore */
2020-06-18 16:55:50 +00:00
uint32_t lcore_count; /**< Number of available logical cores. */
uint32_t numa_node_count; /**< Number of detected NUMA nodes. */
uint32_t numa_nodes[RTE_MAX_NUMA_NODES]; /**< List of detected NUMA nodes. */
uint32_t service_lcore_count;/**< Number of available service cores. */
enum rte_lcore_role_t lcore_role[RTE_MAX_LCORE]; /**< State of cores. */
/** Primary or secondary configuration */
enum rte_proc_type_t process_type;
/** PA or VA mapping mode */
enum rte_iova_mode iova_mode;
/**
* Pointer to memory configuration, which may be shared across multiple
* DPDK instances
*/
struct rte_mem_config *mem_config;
2021-02-05 08:48:47 +00:00
} __rte_packed;
2020-06-18 16:55:50 +00:00
/**
* Get the global configuration structure.
*
* @return
* A pointer to the global configuration structure.
*/
struct rte_config *rte_eal_get_configuration(void);
2019-06-25 11:12:58 +00:00
2017-04-21 10:43:26 +00:00
/**
* Initialize the memzone subsystem (private to eal).
*
* @return
* - 0 on success
* - Negative on error
*/
int rte_eal_memzone_init(void);
/**
* Common log initialization function (private to eal). Determines
* where log data is written when no call to rte_openlog_stream is
* in effect.
2017-04-21 10:43:26 +00:00
*
* @param default_log
* The default log stream to be used.
* @return
* - 0 on success
* - Negative on error
*/
void eal_log_set_default(FILE *default_log);
2017-04-21 10:43:26 +00:00
/**
* Fill configuration with number of physical and logical processors
*
* This function is private to EAL.
*
* Parse /proc/cpuinfo to get the number of physical and logical
* processors on the machine.
*
* @return
* 0 on success, negative on error
*/
int rte_eal_cpu_init(void);
2019-06-25 11:12:58 +00:00
/**
* Create memseg lists
*
* This function is private to EAL.
*
* Preallocate virtual memory.
*
* @return
* 0 on success, negative on error
*/
int rte_eal_memseg_init(void);
2017-04-21 10:43:26 +00:00
/**
* Map memory
*
* This function is private to EAL.
*
* Fill configuration structure with these infos, and return 0 on success.
*
* @return
* 0 on success, negative on error
*/
int rte_eal_memory_init(void);
/**
* Configure timers
*
* This function is private to EAL.
*
* Mmap memory areas used by HPET (high precision event timer) that will
* provide our time reference, and configure the TSC frequency also for it
* to be used as a reference.
*
* @return
* 0 on success, negative on error
*/
int rte_eal_timer_init(void);
/**
* Init the default log stream
*
* This function is private to EAL.
*
* @return
* 0 on success, negative on error
*/
int rte_eal_log_init(const char *id, int facility);
2019-06-25 11:12:58 +00:00
/**
* Save the log regexp for later
*/
int rte_log_save_regexp(const char *type, int priority);
int rte_log_save_pattern(const char *pattern, int priority);
2017-04-21 10:43:26 +00:00
/**
* Init tail queues for non-EAL library structures. This is to allow
* the rings, mempools, etc. lists to be shared among multiple processes
*
* This function is private to EAL
*
* @return
* 0 on success, negative on error
*/
int rte_eal_tailqs_init(void);
/**
* Init interrupt handling.
*
* This function is private to EAL.
*
* @return
* 0 on success, negative on error
*/
int rte_eal_intr_init(void);
/**
* Init alarm mechanism. This is to allow a callback be called after
* specific time.
*
* This function is private to EAL.
*
* @return
* 0 on success, negative on error
*/
int rte_eal_alarm_init(void);
/**
* Function is to check if the kernel module(like, vfio, vfio_iommu_type1,
* etc.) loaded.
*
* @param module_name
* The module's name which need to be checked
*
* @return
* -1 means some error happens(NULL pointer or open failure)
* 0 means the module not loaded
* 1 means the module loaded
*/
int rte_eal_check_module(const char *module_name);
2021-02-05 08:48:47 +00:00
/**
* Memory reservation flags.
*/
enum eal_mem_reserve_flags {
/**
* Reserve hugepages. May be unsupported by some platforms.
*/
EAL_RESERVE_HUGEPAGES = 1 << 0,
/**
* Force reserving memory at the requested address.
* This can be a destructive action depending on the implementation.
*
* @see RTE_MAP_FORCE_ADDRESS for description of possible consequences
* (although implementations are not required to use it).
*/
EAL_RESERVE_FORCE_ADDRESS = 1 << 1
};
2019-06-25 11:12:58 +00:00
/**
* Get virtual area of specified size from the OS.
*
* This function is private to the EAL.
*
* @param requested_addr
* Address where to request address space.
* @param size
* Size of requested area.
* @param page_sz
* Page size on which to align requested virtual area.
* @param flags
* EAL_VIRTUAL_AREA_* flags.
2021-02-05 08:48:47 +00:00
* @param reserve_flags
* Extra flags passed directly to eal_mem_reserve().
2019-06-25 11:12:58 +00:00
*
* @return
* Virtual area address if successful.
* NULL if unsuccessful.
*/
#define EAL_VIRTUAL_AREA_ADDR_IS_HINT (1 << 0)
/**< don't fail if cannot get exact requested address. */
#define EAL_VIRTUAL_AREA_ALLOW_SHRINK (1 << 1)
/**< try getting smaller sized (decrement by page size) virtual areas if cannot
* get area of requested size.
*/
#define EAL_VIRTUAL_AREA_UNMAP (1 << 2)
/**< immediately unmap reserved virtual area. */
void *
eal_get_virtual_area(void *requested_addr, size_t *size,
2021-02-05 08:48:47 +00:00
size_t page_sz, int flags, int reserve_flags);
/**
* Initialize a memory segment list and create its backing storage.
*
* @param msl
* Memory segment list to be filled.
* @param name
* Name for the backing storage.
* @param page_sz
* Size of segment pages in the MSL.
* @param n_segs
* Number of segments.
* @param socket_id
* Socket ID. Must not be SOCKET_ID_ANY.
* @param heap
* Mark MSL as pointing to a heap.
* @return
* 0 on success, (-1) on failure and rte_errno is set.
*/
int
eal_memseg_list_init_named(struct rte_memseg_list *msl, const char *name,
uint64_t page_sz, int n_segs, int socket_id, bool heap);
/**
* Initialize memory segment list and create its backing storage
* with a name corresponding to MSL parameters.
*
* @param type_msl_idx
* Index of the MSL among other MSLs of the same socket and page size.
*
* @see eal_memseg_list_init_named for remaining parameters description.
*/
int
eal_memseg_list_init(struct rte_memseg_list *msl, uint64_t page_sz,
int n_segs, int socket_id, int type_msl_idx, bool heap);
/**
* Reserve VA space for a memory segment list
* previously initialized with eal_memseg_list_init().
*
* @param msl
* Initialized memory segment list with page size defined.
* @param reserve_flags
* Extra memory reservation flags. Can be 0 if unnecessary.
* @return
* 0 on success, (-1) on failure and rte_errno is set.
*/
int
eal_memseg_list_alloc(struct rte_memseg_list *msl, int reserve_flags);
/**
* Populate MSL, each segment is one page long.
*
* @param msl
* Initialized memory segment list with page size defined.
* @param addr
* Starting address of list segments.
* @param n_segs
* Number of segments to populate.
*/
void
eal_memseg_list_populate(struct rte_memseg_list *msl, void *addr, int n_segs);
/**
* Distribute available memory between MSLs.
*
* @return
* 0 on success, (-1) on failure.
*/
int
eal_dynmem_memseg_lists_init(void);
/**
* Preallocate hugepages for dynamic allocation.
*
* @return
* 0 on success, (-1) on failure.
*/
int
eal_dynmem_hugepage_init(void);
/**
* Given the list of hugepage sizes and the number of pages thereof,
* calculate the best number of pages of each size to fulfill the request
* for RAM on each NUMA node.
*
* @param memory
* Amounts of memory requested for each NUMA node of RTE_MAX_NUMA_NODES.
* @param hp_info
* Information about hugepages of different size.
* @param hp_used
* Receives information about used hugepages of each size.
* @param num_hp_info
* Number of elements in hp_info and hp_used.
* @return
* 0 on success, (-1) on failure.
*/
int
eal_dynmem_calc_num_pages_per_socket(
uint64_t *memory, struct hugepage_info *hp_info,
struct hugepage_info *hp_used, unsigned int num_hp_info);
2019-06-25 11:12:58 +00:00
2017-04-21 10:43:26 +00:00
/**
* Get cpu core_id.
*
* This function is private to the EAL.
*/
unsigned eal_cpu_core_id(unsigned lcore_id);
/**
* Check if cpu is present.
*
* This function is private to the EAL.
*/
int eal_cpu_detected(unsigned lcore_id);
/**
* Set TSC frequency from precise value or estimation
*
* This function is private to the EAL.
*/
void set_tsc_freq(void);
/**
* Get precise TSC frequency from system
*
* This function is private to the EAL.
*/
uint64_t get_tsc_freq(void);
/**
* Get TSC frequency if the architecture supports.
*
* This function is private to the EAL.
*
* @return
* The number of TSC cycles in one second.
* Returns zero if the architecture support is not available.
*/
uint64_t get_tsc_freq_arch(void);
2021-02-05 08:48:47 +00:00
/**
* Allocate a free lcore to associate to a non-EAL thread.
*
* @return
* - the id of a lcore with role ROLE_NON_EAL on success.
* - RTE_MAX_LCORE if none was available or initializing was refused (see
* rte_lcore_callback_register).
*/
unsigned int eal_lcore_non_eal_allocate(void);
/**
* Release the lcore used by a non-EAL thread.
* Counterpart of eal_lcore_non_eal_allocate().
*
* @param lcore_id
* The lcore with role ROLE_NON_EAL to release.
*/
void eal_lcore_non_eal_release(unsigned int lcore_id);
2017-04-21 10:43:26 +00:00
/**
* Prepare physical memory mapping
* i.e. hugepages on Linux and
* contigmem on BSD.
*
* This function is private to the EAL.
*/
int rte_eal_hugepage_init(void);
/**
* Creates memory mapping in secondary process
* i.e. hugepages on Linux and
* contigmem on BSD.
*
* This function is private to the EAL.
*/
int rte_eal_hugepage_attach(void);
/**
* Find a bus capable of identifying a device.
*
* @param str
* A device identifier (PCI address, virtual PMD name, ...).
*
* @return
* A valid bus handle if found.
* NULL if no bus is able to parse this device.
*/
struct rte_bus *rte_bus_find_by_device_name(const char *str);
2019-06-25 11:12:58 +00:00
/**
* Create the unix channel for primary/secondary communication.
*
* @return
* 0 on success;
* (<0) on failure.
*/
int rte_mp_channel_init(void);
2019-06-26 10:17:41 +00:00
/**
* Primary/secondary communication cleanup.
*/
void rte_mp_channel_cleanup(void);
2019-06-25 11:12:58 +00:00
/**
* @internal
* Parse a device string and store its information in an
* rte_devargs structure.
*
* A device description is split by layers of abstraction of the device:
* bus, class and driver. Each layer will offer a set of properties that
* can be applied either to configure or recognize a device.
*
* This function will parse those properties and prepare the rte_devargs
* to be given to each layers for processing.
*
* Note: if the "data" field of the devargs points to devstr,
* then no dynamic allocation is performed and the rte_devargs
* can be safely discarded.
*
* Otherwise ``data`` will hold a workable copy of devstr, that will be
* used by layers descriptors within rte_devargs. In this case,
* any rte_devargs should be cleaned-up before being freed.
*
* @param da
* rte_devargs structure to fill.
*
* @param devstr
* Device string.
*
* @return
* 0 on success.
* Negative errno values on error (rte_errno is set).
*/
int
rte_devargs_layers_parse(struct rte_devargs *devargs,
const char *devstr);
/*
* probe a device at local process.
*
* @param devargs
* Device arguments including bus, class and driver properties.
* @param new_dev
* new device be probed as output.
* @return
* 0 on success, negative on error.
*/
int local_dev_probe(const char *devargs, struct rte_device **new_dev);
/**
* Hotplug remove a given device from a specific bus at local process.
*
* @param dev
* Data structure of the device to remove.
* @return
* 0 on success, negative on error.
*/
int local_dev_remove(struct rte_device *dev);
/**
* Iterate over all buses to find the corresponding bus to handle the sigbus
* error.
* @param failure_addr
* Pointer of the fault address of the sigbus error.
*
* @return
* 0 success to handle the sigbus.
* -1 failed to handle the sigbus
* 1 no bus can handler the sigbus
*/
int rte_bus_sigbus_handler(const void *failure_addr);
/**
* @internal
* Register the sigbus handler.
*
* @return
* - On success, zero.
* - On failure, a negative value.
*/
int
dev_sigbus_handler_register(void);
/**
* @internal
* Unregister the sigbus handler.
*
* @return
* - On success, zero.
* - On failure, a negative value.
*/
int
dev_sigbus_handler_unregister(void);
/**
2021-02-05 08:48:47 +00:00
* Get OS-specific EAL mapping base address.
*/
uint64_t
eal_get_baseaddr(void);
void *
eal_malloc_no_trace(const char *type, size_t size, unsigned int align);
void eal_free_no_trace(void *addr);
/** Options for eal_file_open(). */
enum eal_open_flags {
/** Open file for reading. */
EAL_OPEN_READONLY = 0x00,
/** Open file for reading and writing. */
EAL_OPEN_READWRITE = 0x02,
/**
* Create the file if it doesn't exist.
* New files are only accessible to the owner (0600 equivalent).
*/
EAL_OPEN_CREATE = 0x04
};
/**
* Open or create a file.
2019-06-25 11:12:58 +00:00
*
2021-02-05 08:48:47 +00:00
* @param path
* Path to the file.
* @param flags
* A combination of eal_open_flags controlling operation and FD behavior.
* @return
* Open file descriptor on success, (-1) on failure and rte_errno is set.
*/
int
eal_file_open(const char *path, int flags);
/** File locking operation. */
enum eal_flock_op {
EAL_FLOCK_SHARED, /**< Acquire a shared lock. */
EAL_FLOCK_EXCLUSIVE, /**< Acquire an exclusive lock. */
EAL_FLOCK_UNLOCK /**< Release a previously taken lock. */
};
/** Behavior on file locking conflict. */
enum eal_flock_mode {
EAL_FLOCK_WAIT, /**< Wait until the file gets unlocked to lock it. */
EAL_FLOCK_RETURN /**< Return immediately if the file is locked. */
};
/**
* Lock or unlock the file.
2019-06-25 11:12:58 +00:00
*
2021-02-05 08:48:47 +00:00
* On failure @code rte_errno @endcode is set to the error code
* specified by POSIX flock(3) description.
*
* @param fd
* Opened file descriptor.
* @param op
* Operation to perform.
* @param mode
* Behavior on conflict.
2019-06-25 11:12:58 +00:00
* @return
2021-02-05 08:48:47 +00:00
* 0 on success, (-1) on failure.
*/
int
eal_file_lock(int fd, enum eal_flock_op op, enum eal_flock_mode mode);
/**
* Truncate or extend the file to the specified size.
*
* On failure @code rte_errno @endcode is set to the error code
* specified by POSIX ftruncate(3) description.
*
* @param fd
* Opened file descriptor.
* @param size
* Desired file size.
2019-06-25 11:12:58 +00:00
* @return
2021-02-05 08:48:47 +00:00
* 0 on success, (-1) on failure.
2019-06-25 11:12:58 +00:00
*/
int
2021-02-05 08:48:47 +00:00
eal_file_truncate(int fd, ssize_t size);
2019-06-25 11:12:58 +00:00
/**
2021-02-05 08:48:47 +00:00
* Reserve a region of virtual memory.
*
* Use eal_mem_free() to free reserved memory.
*
* @param requested_addr
* A desired reservation address which must be page-aligned.
* The system might not respect it.
* NULL means the address will be chosen by the system.
* @param size
* Reservation size. Must be a multiple of system page size.
* @param flags
* Reservation options, a combination of eal_mem_reserve_flags.
* @returns
* Starting address of the reserved area on success, NULL on failure.
* Callers must not access this memory until remapping it.
2019-06-25 11:12:58 +00:00
*/
2021-02-05 08:48:47 +00:00
void *
eal_mem_reserve(void *requested_addr, size_t size, int flags);
2019-06-25 11:12:58 +00:00
2020-06-18 16:55:50 +00:00
/**
2021-02-05 08:48:47 +00:00
* Free memory obtained by eal_mem_reserve() and possibly allocated.
*
* If *virt* and *size* describe a part of the reserved region,
* only this part of the region is freed (accurately up to the system
* page size). If *virt* points to allocated memory, *size* must match
* the one specified on allocation. The behavior is undefined
* if the memory pointed by *virt* is obtained from another source
* than listed above.
*
* @param virt
* A virtual address in a region previously reserved.
* @param size
* Number of bytes to unreserve.
2020-06-18 16:55:50 +00:00
*/
void
2021-02-05 08:48:47 +00:00
eal_mem_free(void *virt, size_t size);
2020-06-18 16:55:50 +00:00
/**
2021-02-05 08:48:47 +00:00
* Configure memory region inclusion into dumps.
*
* @param virt
* Starting address of the region.
* @param size
* Size of the region.
* @param dump
* True to include memory into dumps, false to exclude.
* @return
* 0 on success, (-1) on failure and rte_errno is set.
2020-06-18 16:55:50 +00:00
*/
2021-02-05 08:48:47 +00:00
int
eal_mem_set_dump(void *virt, size_t size, bool dump);
/**
* Sets the runtime directory of DPDK
*
* @param run_dir
* The new runtime directory path of DPDK
* @param size
* The size of the new runtime directory path in bytes.
* @return
* 0 on success, (-1) on failure.
*/
int
eal_set_runtime_dir(char *run_dir, size_t size);
/**
* Get the internal configuration structure.
*
* @return
* A pointer to the internal configuration structure.
*/
struct internal_config *
eal_get_internal_configuration(void);
/**
* Get the current value of the rte_application_usage pointer
*
* @return
* Pointer to the current value of rte_application_usage .
*/
rte_usage_hook_t
eal_get_application_usage_hook(void);
/**
* Instruct primary process that a secondary process wants to attach.
*/
bool __rte_mp_enable(void);
/**
* Init per-lcore info in current thread.
*
* @param lcore_id
* identifier of lcore.
* @param cpuset
* CPU affinity for this thread.
*/
void __rte_thread_init(unsigned int lcore_id, rte_cpuset_t *cpuset);
/**
* Uninitialize per-lcore info for current thread.
*/
void __rte_thread_uninit(void);
2020-06-18 16:55:50 +00:00
2017-04-21 10:43:26 +00:00
#endif /* _EAL_PRIVATE_H_ */