f-stack/dpdk/drivers/common/qat/qat_qp.h

219 lines
5.9 KiB
C
Raw Normal View History

2019-06-25 11:12:58 +00:00
/* SPDX-License-Identifier: BSD-3-Clause
2023-09-13 12:21:49 +00:00
* Copyright(c) 2018-2022 Intel Corporation
2019-06-25 11:12:58 +00:00
*/
#ifndef _QAT_QP_H_
#define _QAT_QP_H_
#include "qat_common.h"
#include "adf_transport_access_macros.h"
#define QAT_CSR_HEAD_WRITE_THRESH 32U
/* number of requests to accumulate before writing head CSR */
2021-01-28 17:08:59 +00:00
#define QAT_QP_MIN_INFL_THRESHOLD 256
2019-06-25 11:12:58 +00:00
2022-09-06 04:00:10 +00:00
struct qat_pci_device;
2019-06-25 11:12:58 +00:00
/**
* Structure associated with each queue.
*/
struct qat_queue {
char memz_name[RTE_MEMZONE_NAMESIZE];
void *base_addr; /* Base address */
rte_iova_t base_phys_addr; /* Queue physical address */
uint32_t head; /* Shadow copy of the head */
uint32_t tail; /* Shadow copy of the tail */
uint32_t modulo_mask;
uint32_t msg_size;
uint32_t queue_size;
2021-02-05 08:48:47 +00:00
uint8_t trailz;
2019-06-25 11:12:58 +00:00
uint8_t hw_bundle_number;
uint8_t hw_queue_number;
/* HW queue aka ring offset on bundle */
uint32_t csr_head; /* last written head value */
uint32_t csr_tail; /* last written tail value */
uint16_t nb_processed_responses;
/* number of responses processed since last CSR head write */
};
2023-09-13 12:21:49 +00:00
/**
* Type define qat_op_build_request_t function pointer, passed in as argument
* in enqueue op burst, where a build request assigned base on the type of
* crypto op.
*
* @param in_op
* An input op pointer
* @param out_msg
* out_meg pointer
* @param op_cookie
* op cookie pointer
* @param opaque
* an opaque data may be used to store context may be useful between
* 2 enqueue operations.
* @param dev_gen
* qat device gen id
* @return
* - 0 if the crypto request is build successfully,
* - EINVAL if error
**/
typedef int (*qat_op_build_request_t)(void *in_op, uint8_t *out_msg,
void *op_cookie, uint64_t *opaque, enum qat_device_gen dev_gen);
/**
* Type define qat_op_dequeue_t function pointer, passed in as argument
* in dequeue op burst, where a dequeue op assigned base on the type of
* crypto op.
*
* @param op
* An input op pointer
* @param resp
* qat response msg pointer
* @param op_cookie
* op cookie pointer
* @param dequeue_err_count
* dequeue error counter
* @return
* - 0 if dequeue OP is successful
* - EINVAL if error
**/
typedef int (*qat_op_dequeue_t)(void **op, uint8_t *resp, void *op_cookie,
uint64_t *dequeue_err_count __rte_unused);
#define QAT_BUILD_REQUEST_MAX_OPAQUE_SIZE 2
2019-06-25 11:12:58 +00:00
struct qat_qp {
void *mmap_bar_addr;
struct qat_queue tx_q;
struct qat_queue rx_q;
struct qat_common_stats stats;
struct rte_mempool *op_cookie_pool;
void **op_cookies;
uint32_t nb_descriptors;
2023-09-13 12:21:49 +00:00
uint64_t opaque[QAT_BUILD_REQUEST_MAX_OPAQUE_SIZE];
2019-06-25 11:12:58 +00:00
enum qat_device_gen qat_dev_gen;
enum qat_service_type service_type;
struct qat_pci_device *qat_dev;
/**< qat device this qp is on */
2021-01-28 17:08:59 +00:00
uint32_t enqueued;
uint32_t dequeued __rte_aligned(4);
uint16_t max_inflights;
uint16_t min_enq_burst_threshold;
2019-06-25 11:12:58 +00:00
} __rte_cache_aligned;
2022-09-06 04:00:10 +00:00
/**
* Structure with data needed for creation of queue pair.
*/
struct qat_qp_hw_data {
enum qat_service_type service_type;
uint8_t hw_bundle_num;
uint8_t tx_ring_num;
uint8_t rx_ring_num;
uint16_t tx_msg_size;
uint16_t rx_msg_size;
};
/**
* Structure with data needed for creation of queue pair.
*/
struct qat_qp_config {
const struct qat_qp_hw_data *hw;
uint32_t nb_descriptors;
uint32_t cookie_size;
int socket_id;
const char *service_str;
};
2019-06-25 11:12:58 +00:00
uint16_t
2023-09-13 12:21:49 +00:00
qat_enqueue_op_burst(void *qp, qat_op_build_request_t op_build_request,
void **ops, uint16_t nb_ops);
2019-06-25 11:12:58 +00:00
uint16_t
2023-09-13 12:21:49 +00:00
qat_dequeue_op_burst(void *qp, void **ops,
qat_op_dequeue_t qat_dequeue_process_response, uint16_t nb_ops);
2019-06-25 11:12:58 +00:00
int
2022-09-06 04:00:10 +00:00
qat_qp_release(enum qat_device_gen qat_dev_gen, struct qat_qp **qp_addr);
2019-06-25 11:12:58 +00:00
int
qat_qp_setup(struct qat_pci_device *qat_dev,
struct qat_qp **qp_addr, uint16_t queue_pair_id,
struct qat_qp_config *qat_qp_conf);
int
2022-09-06 04:00:10 +00:00
qat_qps_per_service(struct qat_pci_device *qat_dev,
enum qat_service_type service);
const struct qat_qp_hw_data *
qat_qp_get_hw_data(struct qat_pci_device *qat_dev,
enum qat_service_type service, uint16_t qp_id);
2019-06-25 11:12:58 +00:00
2021-01-28 17:08:59 +00:00
int
qat_cq_get_fw_version(struct qat_qp *qp);
2019-06-25 11:12:58 +00:00
/* Needed for weak function*/
int
2019-06-26 10:17:41 +00:00
qat_comp_process_response(void **op __rte_unused, uint8_t *resp __rte_unused,
2020-06-18 16:55:50 +00:00
void *op_cookie __rte_unused,
uint64_t *dequeue_err_count __rte_unused);
2022-09-06 04:00:10 +00:00
int
qat_read_qp_config(struct qat_pci_device *qat_dev);
/**
* Function prototypes for GENx specific queue pair operations.
**/
typedef int (*qat_qp_rings_per_service_t)
(struct qat_pci_device *, enum qat_service_type);
typedef void (*qat_qp_build_ring_base_t)(void *, struct qat_queue *);
typedef void (*qat_qp_adf_arb_enable_t)(const struct qat_queue *, void *,
rte_spinlock_t *);
typedef void (*qat_qp_adf_arb_disable_t)(const struct qat_queue *, void *,
rte_spinlock_t *);
typedef void (*qat_qp_adf_configure_queues_t)(struct qat_qp *);
typedef void (*qat_qp_csr_write_tail_t)(struct qat_qp *qp, struct qat_queue *q);
typedef void (*qat_qp_csr_write_head_t)(struct qat_qp *qp, struct qat_queue *q,
uint32_t new_head);
typedef void (*qat_qp_csr_setup_t)(struct qat_pci_device*, void *,
struct qat_qp *);
typedef const struct qat_qp_hw_data * (*qat_qp_get_hw_data_t)(
struct qat_pci_device *dev, enum qat_service_type service_type,
uint16_t qp_id);
struct qat_qp_hw_spec_funcs {
qat_qp_rings_per_service_t qat_qp_rings_per_service;
qat_qp_build_ring_base_t qat_qp_build_ring_base;
qat_qp_adf_arb_enable_t qat_qp_adf_arb_enable;
qat_qp_adf_arb_disable_t qat_qp_adf_arb_disable;
qat_qp_adf_configure_queues_t qat_qp_adf_configure_queues;
qat_qp_csr_write_tail_t qat_qp_csr_write_tail;
qat_qp_csr_write_head_t qat_qp_csr_write_head;
qat_qp_csr_setup_t qat_qp_csr_setup;
qat_qp_get_hw_data_t qat_qp_get_hw_data;
};
2023-09-11 06:58:14 +00:00
extern struct qat_qp_hw_spec_funcs*
qat_qp_hw_spec[];
static inline void
txq_write_tail(enum qat_device_gen qat_dev_gen,
struct qat_qp *qp, struct qat_queue *q)
{
struct qat_qp_hw_spec_funcs *ops =
qat_qp_hw_spec[qat_dev_gen];
/*
* Pointer check should be done during
* initialization
*/
ops->qat_qp_csr_write_tail(qp, q);
}
2019-06-25 11:12:58 +00:00
#endif /* _QAT_QP_H_ */