| /f-stack/dpdk/drivers/crypto/null/ |
| H A D | null_crypto_pmd_ops.c | 87 int qp_id; in null_crypto_pmd_stats_get() local 89 for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { in null_crypto_pmd_stats_get() 104 int qp_id; in null_crypto_pmd_stats_reset() local 106 for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { in null_crypto_pmd_stats_reset() 141 rte_free(dev->data->queue_pairs[qp_id]); in null_crypto_pmd_qp_release() 142 dev->data->queue_pairs[qp_id] = NULL; in null_crypto_pmd_qp_release() 198 if (qp_id >= internals->max_nb_qpairs) { in null_crypto_pmd_qp_setup() 201 qp_id, internals->max_nb_qpairs); in null_crypto_pmd_qp_setup() 207 null_crypto_pmd_qp_release(dev, qp_id); in null_crypto_pmd_qp_setup() 217 qp->id = qp_id; in null_crypto_pmd_qp_setup() [all …]
|
| /f-stack/dpdk/drivers/crypto/armv8/ |
| H A D | rte_armv8_pmd_ops.c | 116 int qp_id; in armv8_crypto_pmd_stats_get() local 118 for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { in armv8_crypto_pmd_stats_get() 133 int qp_id; in armv8_crypto_pmd_stats_reset() local 135 for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { in armv8_crypto_pmd_stats_reset() 165 if (dev->data->queue_pairs[qp_id] != NULL) { in armv8_crypto_pmd_qp_release() 166 rte_free(dev->data->queue_pairs[qp_id]); in armv8_crypto_pmd_qp_release() 167 dev->data->queue_pairs[qp_id] = NULL; in armv8_crypto_pmd_qp_release() 226 if (dev->data->queue_pairs[qp_id] != NULL) in armv8_crypto_pmd_qp_setup() 227 armv8_crypto_pmd_qp_release(dev, qp_id); in armv8_crypto_pmd_qp_setup() 235 qp->id = qp_id; in armv8_crypto_pmd_qp_setup() [all …]
|
| /f-stack/dpdk/drivers/crypto/zuc/ |
| H A D | rte_zuc_pmd_ops.c | 96 int qp_id; in zuc_pmd_stats_get() local 98 for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { in zuc_pmd_stats_get() 113 int qp_id; in zuc_pmd_stats_reset() local 115 for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { in zuc_pmd_stats_reset() 144 if (dev->data->queue_pairs[qp_id] != NULL) { in zuc_pmd_qp_release() 150 rte_free(dev->data->queue_pairs[qp_id]); in zuc_pmd_qp_release() 151 dev->data->queue_pairs[qp_id] = NULL; in zuc_pmd_qp_release() 207 if (dev->data->queue_pairs[qp_id] != NULL) in zuc_pmd_qp_setup() 208 zuc_pmd_qp_release(dev, qp_id); in zuc_pmd_qp_setup() 216 qp->id = qp_id; in zuc_pmd_qp_setup() [all …]
|
| /f-stack/dpdk/drivers/crypto/snow3g/ |
| H A D | rte_snow3g_pmd_ops.c | 96 int qp_id; in snow3g_pmd_stats_get() local 98 for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { in snow3g_pmd_stats_get() 113 int qp_id; in snow3g_pmd_stats_reset() local 115 for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { in snow3g_pmd_stats_reset() 144 if (dev->data->queue_pairs[qp_id] != NULL) { in snow3g_pmd_qp_release() 150 rte_free(dev->data->queue_pairs[qp_id]); in snow3g_pmd_qp_release() 151 dev->data->queue_pairs[qp_id] = NULL; in snow3g_pmd_qp_release() 207 if (dev->data->queue_pairs[qp_id] != NULL) in snow3g_pmd_qp_setup() 208 snow3g_pmd_qp_release(dev, qp_id); in snow3g_pmd_qp_setup() 216 qp->id = qp_id; in snow3g_pmd_qp_setup() [all …]
|
| /f-stack/dpdk/drivers/crypto/aesni_gcm/ |
| H A D | aesni_gcm_pmd_ops.c | 106 int qp_id; in aesni_gcm_pmd_stats_get() local 108 for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { in aesni_gcm_pmd_stats_get() 123 int qp_id; in aesni_gcm_pmd_stats_reset() local 125 for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { in aesni_gcm_pmd_stats_reset() 155 if (dev->data->queue_pairs[qp_id] != NULL) { in aesni_gcm_pmd_qp_release() 161 rte_free(dev->data->queue_pairs[qp_id]); in aesni_gcm_pmd_qp_release() 162 dev->data->queue_pairs[qp_id] = NULL; in aesni_gcm_pmd_qp_release() 215 if (dev->data->queue_pairs[qp_id] != NULL) in aesni_gcm_pmd_qp_setup() 216 aesni_gcm_pmd_qp_release(dev, qp_id); in aesni_gcm_pmd_qp_setup() 224 qp->id = qp_id; in aesni_gcm_pmd_qp_setup() [all …]
|
| /f-stack/dpdk/drivers/compress/zlib/ |
| H A D | zlib_pmd_ops.c | 87 int qp_id; in zlib_pmd_stats_get() local 89 for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { in zlib_pmd_stats_get() 90 struct zlib_qp *qp = dev->data->queue_pairs[qp_id]; in zlib_pmd_stats_get() 104 int qp_id; in zlib_pmd_stats_reset() local 106 for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { in zlib_pmd_stats_reset() 129 struct zlib_qp *qp = dev->data->queue_pairs[qp_id]; in zlib_pmd_qp_release() 134 dev->data->queue_pairs[qp_id] = NULL; in zlib_pmd_qp_release() 185 if (dev->data->queue_pairs[qp_id] != NULL) in zlib_pmd_qp_setup() 186 zlib_pmd_qp_release(dev, qp_id); in zlib_pmd_qp_setup() 194 qp->id = qp_id; in zlib_pmd_qp_setup() [all …]
|
| /f-stack/dpdk/drivers/crypto/kasumi/ |
| H A D | rte_kasumi_pmd_ops.c | 92 int qp_id; in kasumi_pmd_stats_get() local 94 for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { in kasumi_pmd_stats_get() 95 struct kasumi_qp *qp = dev->data->queue_pairs[qp_id]; in kasumi_pmd_stats_get() 109 int qp_id; in kasumi_pmd_stats_reset() local 111 for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { in kasumi_pmd_stats_reset() 140 struct kasumi_qp *qp = dev->data->queue_pairs[qp_id]; in kasumi_pmd_qp_release() 145 dev->data->queue_pairs[qp_id] = NULL; in kasumi_pmd_qp_release() 201 if (dev->data->queue_pairs[qp_id] != NULL) in kasumi_pmd_qp_setup() 202 kasumi_pmd_qp_release(dev, qp_id); in kasumi_pmd_qp_setup() 210 qp->id = qp_id; in kasumi_pmd_qp_setup() [all …]
|
| /f-stack/dpdk/drivers/crypto/mvsam/ |
| H A D | rte_mrvl_pmd_ops.c | 533 int qp_id; in mrvl_crypto_pmd_stats_get() local 535 for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { in mrvl_crypto_pmd_stats_get() 554 int qp_id; in mrvl_crypto_pmd_stats_reset() local 556 for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { in mrvl_crypto_pmd_stats_reset() 601 dev->data->queue_pairs[qp_id] = NULL; in mrvl_crypto_pmd_qp_release() 616 int qp_id; in mrvl_crypto_pmd_close() local 618 for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) in mrvl_crypto_pmd_close() 619 mrvl_crypto_pmd_qp_release(dev, qp_id); in mrvl_crypto_pmd_close() 650 mrvl_crypto_pmd_qp_release(dev, qp_id); in mrvl_crypto_pmd_qp_setup() 682 qp_id % num, qp_id / num); in mrvl_crypto_pmd_qp_setup() [all …]
|
| /f-stack/dpdk/drivers/compress/isal/ |
| H A D | isal_compress_pmd_ops.c | 118 uint16_t qp_id; in isal_comp_pmd_stats_get() local 120 for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { in isal_comp_pmd_stats_get() 121 struct isal_comp_qp *qp = dev->data->queue_pairs[qp_id]; in isal_comp_pmd_stats_get() 157 uint16_t qp_id; in isal_comp_pmd_stats_reset() local 159 for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { in isal_comp_pmd_stats_reset() 169 struct isal_comp_qp *qp = dev->data->queue_pairs[qp_id]; in isal_comp_pmd_qp_release() 181 dev->data->queue_pairs[qp_id] = NULL; in isal_comp_pmd_qp_release() 237 if (dev->data->queue_pairs[qp_id] != NULL) in isal_comp_pmd_qp_setup() 238 isal_comp_pmd_qp_release(dev, qp_id); in isal_comp_pmd_qp_setup() 274 qp->id = qp_id; in isal_comp_pmd_qp_setup() [all …]
|
| /f-stack/dpdk/drivers/crypto/ccp/ |
| H A D | ccp_pmd_ops.c | 588 int qp_id; in ccp_pmd_stats_get() local 590 for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { in ccp_pmd_stats_get() 605 int qp_id; in ccp_pmd_stats_reset() local 607 for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { in ccp_pmd_stats_reset() 642 dev->data->queue_pairs[qp_id] = NULL; in ccp_pmd_qp_release() 694 if (qp_id >= internals->max_nb_qpairs) { in ccp_pmd_qp_setup() 696 qp_id, internals->max_nb_qpairs); in ccp_pmd_qp_setup() 702 ccp_pmd_qp_release(dev, qp_id); in ccp_pmd_qp_setup() 713 qp->id = qp_id; in ccp_pmd_qp_setup() 714 dev->data->queue_pairs[qp_id] = qp; in ccp_pmd_qp_setup() [all …]
|
| /f-stack/dpdk/drivers/crypto/nitrox/ |
| H A D | nitrox_sym.c | 72 uint16_t qp_id); 140 int qp_id; in nitrox_sym_dev_stats_get() local 142 for (qp_id = 0; qp_id < cdev->data->nb_queue_pairs; qp_id++) { in nitrox_sym_dev_stats_get() 158 int qp_id; in nitrox_sym_dev_stats_reset() local 160 for (qp_id = 0; qp_id < cdev->data->nb_queue_pairs; qp_id++) { in nitrox_sym_dev_stats_reset() 181 if (qp_id >= ndev->nr_queues) { in nitrox_sym_dev_qp_setup() 183 qp_id, ndev->nr_queues); in nitrox_sym_dev_qp_setup() 201 qp->qno = qp_id; in nitrox_sym_dev_qp_setup() 213 cdev->data->queue_pairs[qp_id] = qp; in nitrox_sym_dev_qp_setup() 233 if (qp_id >= ndev->nr_queues) { in nitrox_sym_dev_qp_release() [all …]
|
| /f-stack/dpdk/drivers/compress/octeontx/ |
| H A D | otx_zip_pmd.c | 291 int qp_id; in zip_pmd_stats_get() local 293 for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { in zip_pmd_stats_get() 294 struct zipvf_qp *qp = dev->data->queue_pairs[qp_id]; in zip_pmd_stats_get() 308 int qp_id; in zip_pmd_stats_reset() local 310 for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { in zip_pmd_stats_reset() 335 struct zipvf_qp *qp = dev->data->queue_pairs[qp_id]; in zip_pmd_qp_release() 344 dev->data->queue_pairs[qp_id] = NULL; in zip_pmd_qp_release() 389 if (dev->data->queue_pairs[qp_id] != NULL) { in zip_pmd_qp_setup() 397 dev->data->dev_id, qp_id); in zip_pmd_qp_setup() 413 qp->id = qp_id; in zip_pmd_qp_setup() [all …]
|
| /f-stack/dpdk/lib/librte_cryptodev/ |
| H A D | rte_cryptodev_trace_fp.h | 16 RTE_TRACE_POINT_ARGS(uint8_t dev_id, uint16_t qp_id, void **ops, 19 rte_trace_point_emit_u16(qp_id); 26 RTE_TRACE_POINT_ARGS(uint8_t dev_id, uint16_t qp_id, void **ops, 29 rte_trace_point_emit_u16(qp_id);
|
| /f-stack/dpdk/drivers/crypto/bcmfs/ |
| H A D | bcmfs_sym_pmd.c | 151 uint32_t nobjs, uint16_t qp_id, in bcmfs_sym_req_pool_create() argument 158 "bcm_sym", qp_id); in bcmfs_sym_req_pool_create() 167 qp_id, rte_errno); in bcmfs_sym_req_pool_create() 182 (struct bcmfs_qp **)&cdev->data->queue_pairs[qp_id]; in bcmfs_sym_qp_setup() 189 ret = bcmfs_sym_qp_release(cdev, qp_id); in bcmfs_sym_qp_setup() 194 if (qp_id >= fsdev->max_hw_qps) { in bcmfs_sym_qp_setup() 205 ret = bcmfs_qp_setup(qp_addr, qp_id, &bcmfs_qp_conf); in bcmfs_sym_qp_setup() 212 qp_id, socket_id); in bcmfs_sym_qp_setup() 217 bcmfs_private->fsdev->qps_in_use[qp_id] = *qp_addr; in bcmfs_sym_qp_setup() 219 cdev->data->queue_pairs[qp_id] = qp; in bcmfs_sym_qp_setup() [all …]
|
| /f-stack/dpdk/lib/librte_regexdev/ |
| H A D | rte_regexdev.h | 758 typedef void (*regexdev_stop_flush_t)(uint8_t dev_id, uint16_t qp_id, 1464 rte_regexdev_enqueue_burst(uint8_t dev_id, uint16_t qp_id, in rte_regexdev_enqueue_burst() argument 1471 if (qp_id >= dev->data->dev_conf.nb_queue_pairs) { in rte_regexdev_enqueue_burst() 1472 RTE_REGEXDEV_LOG(ERR, "Invalid queue %d\n", qp_id); in rte_regexdev_enqueue_burst() 1476 return (*dev->enqueue)(dev, qp_id, ops, nb_ops); in rte_regexdev_enqueue_burst() 1523 rte_regexdev_dequeue_burst(uint8_t dev_id, uint16_t qp_id, in rte_regexdev_dequeue_burst() argument 1530 if (qp_id >= dev->data->dev_conf.nb_queue_pairs) { in rte_regexdev_dequeue_burst() 1531 RTE_REGEXDEV_LOG(ERR, "Invalid queue %d\n", qp_id); in rte_regexdev_dequeue_burst() 1535 return (*dev->dequeue)(dev, qp_id, ops, nb_ops); in rte_regexdev_dequeue_burst()
|
| /f-stack/dpdk/app/test-crypto-perf/ |
| H A D | cperf_test_throughput.c | 16 uint16_t qp_id; member 62 uint8_t dev_id, uint16_t qp_id, in cperf_throughput_test_constructor() argument 74 ctx->qp_id = qp_id; in cperf_throughput_test_constructor() 89 if (cperf_alloc_common_memory(options, test_vector, dev_id, qp_id, 0, in cperf_throughput_test_constructor() 134 rte_cryptodev_enqueue_burst(ctx->dev_id, ctx->qp_id, NULL, 0); in cperf_throughput_test_runner() 208 ops_enqd = rte_cryptodev_enqueue_burst(ctx->dev_id, ctx->qp_id, in cperf_throughput_test_runner() 222 ops_deqd = rte_cryptodev_dequeue_burst(ctx->dev_id, ctx->qp_id, in cperf_throughput_test_runner() 246 rte_cryptodev_enqueue_burst(ctx->dev_id, ctx->qp_id, NULL, 0); in cperf_throughput_test_runner() 249 ops_deqd = rte_cryptodev_dequeue_burst(ctx->dev_id, ctx->qp_id, in cperf_throughput_test_runner()
|
| H A D | cperf_test_latency.c | 22 uint16_t qp_id; member 63 uint8_t dev_id, uint16_t qp_id, in cperf_latency_test_constructor() argument 76 ctx->qp_id = qp_id; in cperf_latency_test_constructor() 92 if (cperf_alloc_common_memory(options, test_vector, dev_id, qp_id, in cperf_latency_test_constructor() 158 rte_cryptodev_enqueue_burst(ctx->dev_id, ctx->qp_id, NULL, 0); in cperf_latency_test_runner() 219 ops_enqd = rte_cryptodev_enqueue_burst(ctx->dev_id, ctx->qp_id, in cperf_latency_test_runner() 223 ops_deqd = rte_cryptodev_dequeue_burst(ctx->dev_id, ctx->qp_id, in cperf_latency_test_runner() 268 rte_cryptodev_enqueue_burst(ctx->dev_id, ctx->qp_id, NULL, 0); in cperf_latency_test_runner() 271 ops_deqd = rte_cryptodev_dequeue_burst(ctx->dev_id, ctx->qp_id, in cperf_latency_test_runner()
|
| H A D | cperf_test_verify.c | 16 uint16_t qp_id; member 55 uint8_t dev_id, uint16_t qp_id, in cperf_verify_test_constructor() argument 67 ctx->qp_id = qp_id; in cperf_verify_test_constructor() 82 if (cperf_alloc_common_memory(options, test_vector, dev_id, qp_id, 0, in cperf_verify_test_constructor() 323 ops_enqd = rte_cryptodev_enqueue_burst(ctx->dev_id, ctx->qp_id, in cperf_verify_test_runner() 337 ops_deqd = rte_cryptodev_dequeue_burst(ctx->dev_id, ctx->qp_id, in cperf_verify_test_runner() 365 rte_cryptodev_enqueue_burst(ctx->dev_id, ctx->qp_id, NULL, 0); in cperf_verify_test_runner() 368 ops_deqd = rte_cryptodev_dequeue_burst(ctx->dev_id, ctx->qp_id, in cperf_verify_test_runner()
|
| /f-stack/dpdk/drivers/crypto/aesni_mb/ |
| H A D | rte_aesni_mb_pmd_ops.c | 775 int qp_id; in aesni_mb_pmd_stats_get() local 777 for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { in aesni_mb_pmd_stats_get() 792 int qp_id; in aesni_mb_pmd_stats_reset() local 794 for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { in aesni_mb_pmd_stats_reset() 823 struct aesni_mb_qp *qp = dev->data->queue_pairs[qp_id]; in aesni_mb_pmd_qp_release() 833 dev->data->queue_pairs[qp_id] = NULL; in aesni_mb_pmd_qp_release() 894 if (dev->data->queue_pairs[qp_id] != NULL) in aesni_mb_pmd_qp_setup() 895 aesni_mb_pmd_qp_release(dev, qp_id); in aesni_mb_pmd_qp_setup() 903 qp->id = qp_id; in aesni_mb_pmd_qp_setup() 904 dev->data->queue_pairs[qp_id] = qp; in aesni_mb_pmd_qp_setup() [all …]
|
| /f-stack/dpdk/drivers/crypto/scheduler/ |
| H A D | scheduler_pmd_ops.c | 84 update_order_ring(struct rte_cryptodev *dev, uint16_t qp_id) in update_order_ring() argument 87 struct scheduler_qp_ctx *qp_ctx = dev->data->queue_pairs[qp_id]; in update_order_ring() 104 dev->data->dev_id, qp_id) < 0) { in update_order_ring() 385 dev->data->queue_pairs[qp_id] = NULL; in scheduler_pmd_qp_release() 403 dev->data->dev_id, qp_id) < 0) { in scheduler_pmd_qp_setup() 409 if (dev->data->queue_pairs[qp_id] != NULL) in scheduler_pmd_qp_setup() 410 scheduler_pmd_qp_release(dev, qp_id); in scheduler_pmd_qp_setup() 420 ret = rte_cryptodev_queue_pair_setup(worker_id, qp_id, in scheduler_pmd_qp_setup() 435 dev->data->queue_pairs[qp_id] = qp_ctx; in scheduler_pmd_qp_setup() 443 scheduler_pmd_qp_release(dev, qp_id); in scheduler_pmd_qp_setup() [all …]
|
| /f-stack/dpdk/drivers/regex/octeontx2/ |
| H A D | otx2_regexdev.c | 108 qp_id); in ree_qp_create() 134 qp->id = qp_id; in ree_qp_create() 135 qp->base = OTX2_REE_LF_BAR2(vf, qp_id); in ree_qp_create() 179 struct otx2_ree_qp *qp = data->queue_pairs[qp_id]; in ree_queue_pair_release() 182 ree_func_trace("Queue=%d", qp_id); in ree_queue_pair_release() 193 data->queue_pairs[qp_id] = NULL; in ree_queue_pair_release() 682 ree_func_trace("Queue=%d", qp_id); in otx2_ree_queue_pair_setup() 684 if (data->queue_pairs[qp_id] != NULL) in otx2_ree_queue_pair_setup() 685 ree_queue_pair_release(dev, qp_id); in otx2_ree_queue_pair_setup() 698 qp = ree_qp_create(dev, qp_id); in otx2_ree_queue_pair_setup() [all …]
|
| /f-stack/dpdk/app/test-compress-perf/ |
| H A D | comp_perf.h | 14 uint16_t qp_id, 30 uint16_t qp_id __rte_unused, 41 uint16_t qp_id __rte_unused,
|
| /f-stack/dpdk/drivers/crypto/openssl/ |
| H A D | rte_openssl_pmd_ops.c | 611 int qp_id; in openssl_pmd_stats_get() local 613 for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { in openssl_pmd_stats_get() 628 int qp_id; in openssl_pmd_stats_reset() local 630 for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { in openssl_pmd_stats_reset() 659 if (dev->data->queue_pairs[qp_id] != NULL) { in openssl_pmd_qp_release() 665 rte_free(dev->data->queue_pairs[qp_id]); in openssl_pmd_qp_release() 666 dev->data->queue_pairs[qp_id] = NULL; in openssl_pmd_qp_release() 723 if (dev->data->queue_pairs[qp_id] != NULL) in openssl_pmd_qp_setup() 724 openssl_pmd_qp_release(dev, qp_id); in openssl_pmd_qp_setup() 732 qp->id = qp_id; in openssl_pmd_qp_setup() [all …]
|
| /f-stack/dpdk/drivers/crypto/qat/ |
| H A D | qat_asym_pmd.c | 119 static int qat_asym_qp_setup(struct rte_cryptodev *dev, uint16_t qp_id, in qat_asym_qp_setup() argument 129 (struct qat_qp **)&(dev->data->queue_pairs[qp_id]); in qat_asym_qp_setup() 134 const struct qat_qp_hw_data *qp_hw_data = asym_hw_qps + qp_id; in qat_asym_qp_setup() 138 ret = qat_asym_qp_release(dev, qp_id); in qat_asym_qp_setup() 142 if (qp_id >= qat_qps_per_service(asym_hw_qps, QAT_SERVICE_ASYMMETRIC)) { in qat_asym_qp_setup() 143 QAT_LOG(ERR, "qp_id %u invalid for this device", qp_id); in qat_asym_qp_setup() 153 ret = qat_qp_setup(qat_private->qat_dev, qp_addr, qp_id, &qat_qp_conf); in qat_asym_qp_setup() 158 qat_private->qat_dev->qps_in_use[QAT_SERVICE_ASYMMETRIC][qp_id] in qat_asym_qp_setup()
|
| H A D | qat_sym_pmd.c | 155 static int qat_sym_qp_setup(struct rte_cryptodev *dev, uint16_t qp_id, in qat_sym_qp_setup() argument 165 (struct qat_qp **)&(dev->data->queue_pairs[qp_id]); in qat_sym_qp_setup() 170 const struct qat_qp_hw_data *qp_hw_data = sym_hw_qps + qp_id; in qat_sym_qp_setup() 174 ret = qat_sym_qp_release(dev, qp_id); in qat_sym_qp_setup() 178 if (qp_id >= qat_qps_per_service(sym_hw_qps, QAT_SERVICE_SYMMETRIC)) { in qat_sym_qp_setup() 179 QAT_LOG(ERR, "qp_id %u invalid for this device", qp_id); in qat_sym_qp_setup() 189 ret = qat_qp_setup(qat_private->qat_dev, qp_addr, qp_id, &qat_qp_conf); in qat_sym_qp_setup() 194 qat_private->qat_dev->qps_in_use[QAT_SERVICE_SYMMETRIC][qp_id] in qat_sym_qp_setup() 222 qat_sym_qp_release(dev, qp_id); in qat_sym_qp_setup()
|