| /f-stack/dpdk/drivers/crypto/nitrox/ |
| H A D | nitrox_qp.c | 35 qp->cmdq.mz = mz; in nitrox_setup_cmdq() 38 qp->cmdq.ring = mz->addr; in nitrox_setup_cmdq() 39 qp->cmdq.instr_size = instr_size; in nitrox_setup_cmdq() 40 setup_nps_pkt_input_ring(bar_addr, qp->qno, qp->count, mz->iova); in nitrox_setup_cmdq() 49 size_t ridq_size = qp->count * sizeof(*qp->ridq); in nitrox_setup_ridq() 54 if (!qp->ridq) { in nitrox_setup_ridq() 86 qp->count = count; in nitrox_qp_setup() 87 qp->head = qp->tail = 0; in nitrox_qp_setup() 100 nitrox_release_cmdq(qp, bar_addr); in nitrox_qp_setup() 107 rte_free(qp->ridq); in nitrox_release_ridq() [all …]
|
| H A D | nitrox_qp.h | 42 RTE_ASSERT(qp->count >= pending_count); in nitrox_qp_free_count() 43 return (qp->count - pending_count); in nitrox_qp_free_count() 61 uint32_t tail = qp->tail % qp->count; in nitrox_qp_get_softreq() 64 return qp->ridq[tail].sr; in nitrox_qp_get_softreq() 82 uint32_t head = qp->head % qp->count; in nitrox_qp_enqueue() 84 qp->head++; in nitrox_qp_enqueue() 85 memcpy(&qp->cmdq.ring[head * qp->cmdq.instr_size], in nitrox_qp_enqueue() 86 instr, qp->cmdq.instr_size); in nitrox_qp_enqueue() 87 qp->ridq[head].sr = sr; in nitrox_qp_enqueue() 89 rte_atomic16_inc(&qp->pending_count); in nitrox_qp_enqueue() [all …]
|
| /f-stack/dpdk/drivers/crypto/bcmfs/ |
| H A D | bcmfs_qp.c | 176 qp->ops->stopq(qp); in bcmfs_qp_release() 178 bcmfs_queue_delete(&qp->tx_q, qp->qpair_id); in bcmfs_qp_release() 179 bcmfs_queue_delete(&qp->cmpl_q, qp->qpair_id); in bcmfs_qp_release() 188 rte_free(qp); in bcmfs_qp_release() 235 rc = bcmfs_queue_create(&qp->tx_q, qp_conf, qp->qpair_id, in bcmfs_qp_setup() 243 rc = bcmfs_queue_create(&qp->cmpl_q, qp_conf, qp->qpair_id, in bcmfs_qp_setup() 263 qp->ctx_bmp = rte_bitmap_init(nb_descriptors, qp->ctx_bmp_mem, in bcmfs_qp_setup() 285 qp->ops->startq(qp); in bcmfs_qp_setup() 287 *qp_addr = qp; in bcmfs_qp_setup() 300 rte_free(qp); in bcmfs_qp_setup() [all …]
|
| /f-stack/dpdk/drivers/compress/isal/ |
| H A D | isal_compress_pmd_ops.c | 161 memset(&qp->qp_stats, 0, sizeof(qp->qp_stats)); in isal_comp_pmd_stats_reset() 171 if (qp == NULL) in isal_comp_pmd_qp_release() 174 if (qp->stream) in isal_comp_pmd_qp_release() 180 rte_free(qp); in isal_comp_pmd_qp_release() 198 qp->name); in isal_comp_pmd_qp_create_processed_pkts_ring() 205 qp->name); in isal_comp_pmd_qp_create_processed_pkts_ring() 218 unsigned int n = snprintf(qp->name, sizeof(qp->name), in isal_comp_pmd_qp_set_unique_name() 274 qp->id = qp_id; in isal_comp_pmd_qp_setup() 294 memset(&qp->qp_stats, 0, sizeof(qp->qp_stats)); in isal_comp_pmd_qp_setup() 298 if (qp->stream) in isal_comp_pmd_qp_setup() [all …]
|
| H A D | isal_compress_pmd.c | 261 if (unlikely(!qp->stream->next_in || !qp->stream->next_out)) { in chained_mbuf_compression() 287 qp->stream->next_in = in chained_mbuf_compression() 289 qp->stream->avail_in = in chained_mbuf_compression() 303 qp->stream->next_out = in chained_mbuf_compression() 361 if (qp->state->avail_in == 0 in chained_mbuf_decompression() 368 qp->state->next_in = in chained_mbuf_decompression() 370 qp->state->avail_in = in chained_mbuf_decompression() 393 qp->state->next_out = in chained_mbuf_decompression() 429 qp->stream->flush = NO_FLUSH; in process_isal_deflate() 480 if (unlikely(!qp->stream->next_in || !qp->stream->next_out)) { in process_isal_deflate() [all …]
|
| /f-stack/freebsd/crypto/ccp/ |
| H A D | ccp_hardware.c | 335 bus_dmamem_free(qp->ring_desc_tag, qp->desc_ring, qp->ring_desc_map); in ccp_hw_detach_queue() 858 desc = &qp->desc_ring[qp->cq_tail]; in ccp_passthrough() 1048 desc = &qp->desc_ring[qp->cq_tail]; in ccp_sha_single_desc() 1424 desc = &qp->desc_ring[qp->cq_tail]; in ccp_do_xts() 1461 qp->cq_tail = (qp->cq_tail + 1) % in ccp_do_xts() 1665 qp->cq_tail = (qp->cq_tail + 1) % in ccp_do_blkcipher() 1754 qp->cq_tail = (qp->cq_tail + 1) % in ccp_do_ghash_aad() 1769 desc = &qp->desc_ring[qp->cq_tail]; in ccp_do_gctr() 1792 qp->cq_tail = (qp->cq_tail + 1) % in ccp_do_gctr() 1805 desc = &qp->desc_ring[qp->cq_tail]; in ccp_do_ghash_final() [all …]
|
| H A D | ccp.c | 139 struct ccp_queue *qp; in ccp_initialize_queues() local 145 qp->cq_softc = sc; in ccp_initialize_queues() 146 qp->cq_qindex = i; in ccp_initialize_queues() 541 qp = NULL; in ccp_process() 657 sc = qp->cq_softc; in ccp_queue_reserve_space() 668 msleep(&qp->cq_tail, &qp->cq_lock, 0, "ccpqfull", 0); in ccp_queue_reserve_space() 678 qp->cq_acq_tail = qp->cq_tail; in ccp_queue_acquire_reserve() 690 if (qp->cq_tail != qp->cq_acq_tail) { in ccp_queue_release() 705 for (i = qp->cq_acq_tail; i != qp->cq_tail; in ccp_queue_abort() 707 memset(&qp->desc_ring[i], 0, sizeof(qp->desc_ring[i])); in ccp_queue_abort() [all …]
|
| /f-stack/dpdk/drivers/crypto/kasumi/ |
| H A D | rte_kasumi_pmd_ops.c | 114 memset(&qp->qp_stats, 0, sizeof(qp->qp_stats)); in kasumi_pmd_stats_reset() 144 rte_free(qp); in kasumi_pmd_qp_release() 155 unsigned n = snprintf(qp->name, sizeof(qp->name), in kasumi_pmd_qp_set_unique_name() 177 qp->name); in kasumi_pmd_qp_create_processed_ops_ring() 183 qp->name); in kasumi_pmd_qp_create_processed_ops_ring() 205 qp = rte_zmalloc_socket("KASUMI PMD Queue Pair", sizeof(*qp), in kasumi_pmd_qp_setup() 207 if (qp == NULL) in kasumi_pmd_qp_setup() 210 qp->id = qp_id; in kasumi_pmd_qp_setup() 216 qp->processed_ops = kasumi_pmd_qp_create_processed_ops_ring(qp, in kasumi_pmd_qp_setup() 225 memset(&qp->qp_stats, 0, sizeof(qp->qp_stats)); in kasumi_pmd_qp_setup() [all …]
|
| H A D | rte_kasumi_pmd.c | 148 if (rte_mempool_get(qp->sess_mp_priv, in kasumi_get_session() 156 rte_mempool_put(qp->sess_mp, _sess); in kasumi_get_session() 262 dst = qp->temp_digest; in process_kasumi_hash_op() 263 IMB_KASUMI_F9_1_BUFFER(qp->mgr, in process_kasumi_hash_op() 274 IMB_KASUMI_F9_1_BUFFER(qp->mgr, in process_kasumi_hash_op() 404 struct kasumi_qp *qp = queue_pair; in kasumi_pmd_enqueue_burst() local 444 qp, burst_size, &enqueued_ops); in kasumi_pmd_enqueue_burst() 455 qp, &enqueued_ops); in kasumi_pmd_enqueue_burst() 489 qp, burst_size, &enqueued_ops); in kasumi_pmd_enqueue_burst() 505 qp, burst_size, &enqueued_ops); in kasumi_pmd_enqueue_burst() [all …]
|
| /f-stack/dpdk/drivers/compress/zlib/ |
| H A D | zlib_pmd_ops.c | 109 memset(&qp->qp_stats, 0, sizeof(qp->qp_stats)); in zlib_pmd_stats_reset() 133 rte_free(qp); in zlib_pmd_qp_release() 144 unsigned int n = snprintf(qp->name, sizeof(qp->name), in zlib_pmd_qp_set_unique_name() 189 qp = rte_zmalloc_socket("ZLIB PMD Queue Pair", sizeof(*qp), in zlib_pmd_qp_setup() 191 if (qp == NULL) in zlib_pmd_qp_setup() 194 qp->id = qp_id; in zlib_pmd_qp_setup() 200 qp->processed_pkts = zlib_pmd_qp_create_processed_pkts_ring(qp, in zlib_pmd_qp_setup() 205 memset(&qp->qp_stats, 0, sizeof(qp->qp_stats)); in zlib_pmd_qp_setup() 209 if (qp) { in zlib_pmd_qp_setup() 210 rte_free(qp); in zlib_pmd_qp_setup() [all …]
|
| /f-stack/dpdk/drivers/crypto/zuc/ |
| H A D | rte_zuc_pmd_ops.c | 118 memset(&qp->qp_stats, 0, sizeof(qp->qp_stats)); in zuc_pmd_stats_reset() 161 unsigned n = snprintf(qp->name, sizeof(qp->name), in zuc_pmd_qp_set_unique_name() 183 qp->name); in zuc_pmd_qp_create_processed_ops_ring() 189 qp->name); in zuc_pmd_qp_create_processed_ops_ring() 211 qp = rte_zmalloc_socket("ZUC PMD Queue Pair", sizeof(*qp), in zuc_pmd_qp_setup() 213 if (qp == NULL) in zuc_pmd_qp_setup() 216 qp->id = qp_id; in zuc_pmd_qp_setup() 222 qp->processed_ops = zuc_pmd_qp_create_processed_ops_ring(qp, in zuc_pmd_qp_setup() 231 memset(&qp->qp_stats, 0, sizeof(qp->qp_stats)); in zuc_pmd_qp_setup() 236 if (qp) in zuc_pmd_qp_setup() [all …]
|
| /f-stack/dpdk/drivers/crypto/snow3g/ |
| H A D | rte_snow3g_pmd_ops.c | 118 memset(&qp->qp_stats, 0, sizeof(qp->qp_stats)); in snow3g_pmd_stats_reset() 161 unsigned n = snprintf(qp->name, sizeof(qp->name), in snow3g_pmd_qp_set_unique_name() 183 qp->name); in snow3g_pmd_qp_create_processed_ops_ring() 189 qp->name); in snow3g_pmd_qp_create_processed_ops_ring() 211 qp = rte_zmalloc_socket("SNOW 3G PMD Queue Pair", sizeof(*qp), in snow3g_pmd_qp_setup() 213 if (qp == NULL) in snow3g_pmd_qp_setup() 216 qp->id = qp_id; in snow3g_pmd_qp_setup() 222 qp->processed_ops = snow3g_pmd_qp_create_processed_ops_ring(qp, in snow3g_pmd_qp_setup() 231 memset(&qp->qp_stats, 0, sizeof(qp->qp_stats)); in snow3g_pmd_qp_setup() 236 if (qp) in snow3g_pmd_qp_setup() [all …]
|
| /f-stack/dpdk/drivers/crypto/aesni_gcm/ |
| H A D | aesni_gcm_pmd_ops.c | 128 memset(&qp->qp_stats, 0, sizeof(qp->qp_stats)); in aesni_gcm_pmd_stats_reset() 158 if (qp->processed_pkts) in aesni_gcm_pmd_qp_release() 170 struct aesni_gcm_qp *qp) in aesni_gcm_pmd_qp_set_unique_name() argument 172 unsigned n = snprintf(qp->name, sizeof(qp->name), in aesni_gcm_pmd_qp_set_unique_name() 219 qp = rte_zmalloc_socket("AES-NI PMD Queue Pair", sizeof(*qp), in aesni_gcm_pmd_qp_setup() 221 if (qp == NULL) in aesni_gcm_pmd_qp_setup() 224 qp->id = qp_id; in aesni_gcm_pmd_qp_setup() 232 qp->processed_pkts = aesni_gcm_pmd_qp_create_processed_pkts_ring(qp, in aesni_gcm_pmd_qp_setup() 240 memset(&qp->qp_stats, 0, sizeof(qp->qp_stats)); in aesni_gcm_pmd_qp_setup() 245 if (qp) in aesni_gcm_pmd_qp_setup() [all …]
|
| /f-stack/dpdk/drivers/crypto/null/ |
| H A D | null_crypto_pmd_ops.c | 109 memset(&qp->qp_stats, 0, sizeof(qp->qp_stats)); in null_crypto_pmd_stats_reset() 138 if (qp->processed_pkts) in null_crypto_pmd_qp_release() 152 unsigned n = snprintf(qp->name, sizeof(qp->name), in null_crypto_pmd_qp_set_unique_name() 156 if (n >= sizeof(qp->name)) in null_crypto_pmd_qp_set_unique_name() 210 qp = rte_zmalloc_socket("Null Crypto PMD Queue Pair", sizeof(*qp), in null_crypto_pmd_qp_setup() 212 if (qp == NULL) { in null_crypto_pmd_qp_setup() 217 qp->id = qp_id; in null_crypto_pmd_qp_setup() 228 qp->processed_pkts = null_crypto_pmd_qp_create_processed_pkts_ring(qp, in null_crypto_pmd_qp_setup() 239 memset(&qp->qp_stats, 0, sizeof(qp->qp_stats)); in null_crypto_pmd_qp_setup() 244 if (qp) in null_crypto_pmd_qp_setup() [all …]
|
| /f-stack/dpdk/drivers/crypto/armv8/ |
| H A D | rte_armv8_pmd_ops.c | 138 memset(&qp->stats, 0, sizeof(qp->stats)); in armv8_crypto_pmd_stats_reset() 180 n = snprintf(qp->name, sizeof(qp->name), "armv8_crypto_pmd_%u_qp_%u", in armv8_crypto_pmd_qp_set_unique_name() 202 qp->name); in armv8_crypto_pmd_qp_create_processed_ops_ring() 208 qp->name); in armv8_crypto_pmd_qp_create_processed_ops_ring() 230 qp = rte_zmalloc_socket("ARMv8 PMD Queue Pair", sizeof(*qp), in armv8_crypto_pmd_qp_setup() 232 if (qp == NULL) in armv8_crypto_pmd_qp_setup() 235 qp->id = qp_id; in armv8_crypto_pmd_qp_setup() 241 qp->processed_ops = armv8_crypto_pmd_qp_create_processed_ops_ring(qp, in armv8_crypto_pmd_qp_setup() 249 memset(&qp->stats, 0, sizeof(qp->stats)); in armv8_crypto_pmd_qp_setup() 254 if (qp) in armv8_crypto_pmd_qp_setup() [all …]
|
| /f-stack/dpdk/drivers/common/qat/ |
| H A D | qat_qp.c | 231 rte_free(qp); in qat_qp_setup() 236 qp->enqueued = qp->dequeued = 0; in qat_qp_setup() 245 qp->max_inflights = ADF_MAX_INFLIGHTS(qp->tx_q.queue_size, in qat_qp_setup() 263 adf_queue_arb_enable(&qp->tx_q, qp->mmap_bar_addr, in qat_qp_setup() 287 if (rte_mempool_get(qp->op_cookie_pool, &qp->op_cookies[i])) { in qat_qp_setup() 301 *qp_addr = qp; in qat_qp_setup() 308 rte_free(qp); in qat_qp_setup() 326 if ((qp->enqueued - qp->dequeued) == 0) { in qat_qp_release() 333 adf_queue_arb_disable(&(qp->tx_q), qp->mmap_bar_addr, in qat_qp_release() 337 rte_mempool_put(qp->op_cookie_pool, qp->op_cookies[i]); in qat_qp_release() [all …]
|
| /f-stack/dpdk/drivers/common/octeontx2/ |
| H A D | otx2_sec_idev.c | 45 cfg->tx_cpt[i].qp = NULL; in otx2_sec_idev_cfg_init() 58 if (qp == NULL || port_id >= OTX2_MAX_INLINE_PORTS) in otx2_sec_idev_tx_cpt_qp_add() 68 if (cfg->tx_cpt[i].qp == NULL) { in otx2_sec_idev_tx_cpt_qp_add() 69 cfg->tx_cpt[i].qp = qp; in otx2_sec_idev_tx_cpt_qp_add() 89 if (qp == NULL) in otx2_sec_idev_tx_cpt_qp_remove() 98 if (cfg->tx_cpt[i].qp != qp) in otx2_sec_idev_tx_cpt_qp_remove() 105 cfg->tx_cpt[i].qp = NULL; in otx2_sec_idev_tx_cpt_qp_remove() 140 if (cfg->tx_cpt[index].qp != NULL) in otx2_sec_idev_tx_cpt_qp_get() 150 *qp = cfg->tx_cpt[index].qp; in otx2_sec_idev_tx_cpt_qp_get() 169 if (qp == NULL) in otx2_sec_idev_tx_cpt_qp_put() [all …]
|
| /f-stack/dpdk/drivers/crypto/bcmfs/hw/ |
| H A D | bcmfs5_rm.c | 389 struct bcmfs_queue *txq = &qp->tx_q; in bcmfs5_enqueue_single_request_qp() 418 qp->qpair_id); in bcmfs5_enqueue_single_request_qp() 428 qp->nb_pending_requests++; in bcmfs5_enqueue_single_request_qp() 470 budget = qp->nb_pending_requests; in bcmfs5_dequeue_qp() 508 context = qp->ctx_pool[reqid]; in bcmfs5_dequeue_qp() 513 qp->ctx_pool[reqid] = 0; in bcmfs5_dequeue_qp() 526 qp->nb_pending_requests -= count; in bcmfs5_dequeue_qp() 532 bcmfs5_start_qp(struct bcmfs_qp *qp) in bcmfs5_start_qp() argument 585 qp->qpair_id); in bcmfs5_start_qp() 599 qp->qpair_id); in bcmfs5_start_qp() [all …]
|
| H A D | bcmfs4_rm.c | 458 struct bcmfs_queue *txq = &qp->tx_q; in bcmfs4_enqueue_single_request_qp() 495 qp->qpair_id); in bcmfs4_enqueue_single_request_qp() 505 qp->nb_pending_requests++; in bcmfs4_enqueue_single_request_qp() 539 budget = qp->nb_pending_requests; in bcmfs4_dequeue_qp() 577 context = qp->ctx_pool[reqid]; in bcmfs4_dequeue_qp() 582 qp->ctx_pool[reqid] = 0; in bcmfs4_dequeue_qp() 595 qp->nb_pending_requests -= count; in bcmfs4_dequeue_qp() 601 bcmfs4_start_qp(struct bcmfs_qp *qp) in bcmfs4_start_qp() argument 654 qp->qpair_id); in bcmfs4_start_qp() 668 qp->qpair_id); in bcmfs4_start_qp() [all …]
|
| /f-stack/dpdk/drivers/compress/octeontx/ |
| H A D | otx_zip_pmd.c | 312 memset(&qp->qp_stats, 0, sizeof(qp->qp_stats)); in zip_pmd_stats_reset() 343 rte_free(qp); in zip_pmd_qp_release() 400 qp = rte_zmalloc_socket(name, sizeof(*qp), in zip_pmd_qp_setup() 402 if (qp == NULL) in zip_pmd_qp_setup() 405 qp->name = name; in zip_pmd_qp_setup() 408 qp->processed_pkts = zip_pmd_qp_create_processed_pkts_ring(qp, in zip_pmd_qp_setup() 413 qp->id = qp_id; in zip_pmd_qp_setup() 414 qp->vf = vf; in zip_pmd_qp_setup() 422 memset(&qp->qp_stats, 0, sizeof(qp->qp_stats)); in zip_pmd_qp_setup() 428 if (qp) in zip_pmd_qp_setup() [all …]
|
| /f-stack/dpdk/drivers/crypto/ccp/ |
| H A D | ccp_pmd_ops.c | 610 memset(&qp->qp_stats, 0, sizeof(qp->qp_stats)); in ccp_pmd_stats_reset() 641 rte_free(qp); in ccp_pmd_qp_release() 651 unsigned int n = snprintf(qp->name, sizeof(qp->name), in ccp_pmd_qp_set_unique_name() 672 qp->name); in ccp_pmd_qp_create_batch_info_ring() 677 qp->name); in ccp_pmd_qp_create_batch_info_ring() 712 qp->dev = dev; in ccp_pmd_qp_setup() 722 qp->processed_pkts = ccp_pmd_qp_create_batch_info_ring(qp, in ccp_pmd_qp_setup() 734 qp->name, in ccp_pmd_qp_setup() 742 memset(&qp->qp_stats, 0, sizeof(qp->qp_stats)); in ccp_pmd_qp_setup() 747 if (qp) in ccp_pmd_qp_setup() [all …]
|
| /f-stack/dpdk/drivers/regex/mlx5/ |
| H A D | mlx5_regex_fastpath.c | 339 if (!qp->metadata) { in setup_buffers() 355 if (!qp->outputs) { in setup_buffers() 386 qp->jobs = rte_calloc(__func__, qp->nb_desc, sizeof(*qp->jobs), 64); in mlx5_regexdev_setup_fastpath() 387 if (!qp->jobs) in mlx5_regexdev_setup_fastpath() 391 rte_free(qp->jobs); in mlx5_regexdev_setup_fastpath() 394 setup_sqs(qp); in mlx5_regexdev_setup_fastpath() 401 if (qp->metadata) { in free_buffers() 405 if (qp->outputs) { in free_buffers() 416 if (qp) { in mlx5_regexdev_teardown_fastpath() 417 free_buffers(qp); in mlx5_regexdev_teardown_fastpath() [all …]
|
| H A D | mlx5_regex_control.c | 235 attr.cqn = qp->cq.obj->id; in regex_ctrl_create_sq() 273 (void)qp; in regex_ctrl_create_sq() 337 struct mlx5_regex_qp *qp; in mlx5_regex_qp_setup() local 343 qp = &priv->qps[qp_ind]; in mlx5_regex_qp_setup() 346 qp->nb_desc = 1 << qp->cq.log_nb_desc; in mlx5_regex_qp_setup() 348 qp->nb_obj = regex_ctrl_get_nb_obj(qp->nb_desc); in mlx5_regex_qp_setup() 350 qp->nb_obj = 1; in mlx5_regex_qp_setup() 351 qp->sqs = rte_malloc(NULL, in mlx5_regex_qp_setup() 353 if (!qp->sqs) { in mlx5_regex_qp_setup() 358 log_desc = rte_log2_u32(qp->nb_desc / qp->nb_obj); in mlx5_regex_qp_setup() [all …]
|
| /f-stack/dpdk/drivers/event/octeontx2/ |
| H A D | otx2_evdev_crypto_adptr.c | 32 struct otx2_cpt_qp *qp; in otx2_ca_qp_add() local 35 qp = cdev->data->queue_pairs[queue_pair_id]; in otx2_ca_qp_add() 37 qp->ca_enable = 1; in otx2_ca_qp_add() 38 rte_memcpy(&qp->ev, event, sizeof(struct rte_event)); in otx2_ca_qp_add() 40 ret = otx2_cpt_af_reg_read(cdev, OTX2_CPT_AF_LF_CTL2(qp->id), in otx2_ca_qp_add() 46 ret = otx2_cpt_af_reg_write(cdev, OTX2_CPT_AF_LF_CTL2(qp->id), in otx2_ca_qp_add() 62 struct otx2_cpt_qp *qp; in otx2_ca_qp_del() local 67 qp = cdev->data->queue_pairs[queue_pair_id]; in otx2_ca_qp_del() 68 qp->ca_enable = 0; in otx2_ca_qp_del() 69 memset(&qp->ev, 0, sizeof(struct rte_event)); in otx2_ca_qp_del() [all …]
|
| /f-stack/dpdk/drivers/crypto/octeontx2/ |
| H A D | otx2_cryptodev_ops.c | 181 if (qp == NULL) { in otx2_cpt_qp_create() 243 qp->id = qp_id; in otx2_cpt_qp_create() 252 qp->lf_nq_reg = qp->base + OTX2_CPT_LF_NQ(0); in otx2_cpt_qp_create() 275 return qp; in otx2_cpt_qp_create() 282 rte_free(qp); in otx2_cpt_qp_create() 304 qp->id); in otx2_cpt_qp_destroy() 312 rte_free(qp); in otx2_cpt_qp_destroy() 458 qp->ev.flow_id; in otx2_ca_enqueue_req() 460 req->qp = qp; in otx2_ca_enqueue_req() 1195 if (qp == NULL) { in otx2_cpt_queue_pair_setup() [all …]
|