| /dpdk/drivers/event/dsw/ |
| H A D | dsw_event.c | 106 port->dequeued += num; in dsw_port_dequeue_stats() 118 if (dequeued > 0 && port->busy_start == 0) in dsw_port_load_record() 1342 port->in_buffer_start += dequeued; in dsw_port_dequeue_burst() 1343 port->in_buffer_len -= dequeued; in dsw_port_dequeue_burst() 1348 return dequeued; in dsw_port_dequeue_burst() 1360 uint16_t dequeued; in dsw_event_dequeue_burst() local 1371 source_port->pending_releases = dequeued; in dsw_event_dequeue_burst() 1375 dsw_port_note_op(source_port, dequeued); in dsw_event_dequeue_burst() 1377 if (dequeued > 0) { in dsw_event_dequeue_burst() 1379 dequeued); in dsw_event_dequeue_burst() [all …]
|
| H A D | dsw_xstats.c | 78 DSW_GEN_PORT_ACCESS_FN(dequeued) in DSW_GEN_PORT_ACCESS_FN() argument 106 uint64_t dequeued = in dsw_xstats_port_get_event_proc_latency() local 107 dsw->ports[port_id].dequeued; in dsw_xstats_port_get_event_proc_latency() 109 return dequeued > 0 ? total_busy_cycles / dequeued : 0; in dsw_xstats_port_get_event_proc_latency()
|
| H A D | dsw_evdev.h | 206 uint64_t dequeued; member
|
| /dpdk/app/test-bbdev/ |
| H A D | test_bbdev_perf.c | 3982 for (i = 0, dequeued = 0; dequeued < num_to_process; ++i) { in latency_test_dec() 4033 dequeued += deq; in latency_test_dec() 4055 for (i = 0, dequeued = 0; dequeued < num_to_process; ++i) { in latency_test_ldpc_dec() 4124 dequeued += deq; in latency_test_ldpc_dec() 4141 for (i = 0, dequeued = 0; dequeued < num_to_process; ++i) { in latency_test_enc() 4190 dequeued += deq; in latency_test_enc() 4208 for (i = 0, dequeued = 0; dequeued < num_to_process; ++i) { in latency_test_ldpc_enc() 4384 for (i = 0, dequeued = 0; dequeued < num_to_process; ++i) { in offload_latency_test_dec() 4469 for (i = 0, dequeued = 0; dequeued < num_to_process; ++i) { in offload_latency_test_ldpc_dec() 4562 for (i = 0, dequeued = 0; dequeued < num_to_process; ++i) { in offload_latency_test_enc() [all …]
|
| /dpdk/app/test/ |
| H A D | test_pie.c | 76 uint32_t *dequeued; /**< Test operations dequeued */ member 728 uint32_t *dequeued, in enqueue_dequeue_perf() argument 767 (*dequeued)++; in enqueue_dequeue_perf() 830 tcfg->tvar->dequeued, in perf_test() 860 .dequeued = pt_dequeued 898 *tcfg->tvar->dequeued = 0; in perf_test2() 908 tcfg->tvar->dequeued, in perf_test2() 914 total, *tcfg->tvar->dequeued, in perf_test2() 915 ((double)(*tcfg->tvar->dequeued) / (double)total) * 100.0, in perf_test2()
|
| /dpdk/drivers/common/qat/ |
| H A D | qat_qp.c | 108 qp->enqueued = qp->dequeued = 0; in qat_qp_setup() 315 if ((qp->enqueued - qp->dequeued) == 0) { in qat_qp_release() 582 tmp_qp->enqueued - tmp_qp->dequeued; in qat_enqueue_op_burst() 672 tmp_qp->enqueued - tmp_qp->dequeued; in qat_enqueue_comp_op_burst() 855 tmp_qp->dequeued += fw_resp_counter; in qat_dequeue_op_burst()
|
| H A D | qat_qp.h | 98 uint32_t dequeued __rte_aligned(4);
|
| /dpdk/doc/guides/sample_app_ug/ |
| H A D | bbdev_app.rst | 19 operation -> dequeued from the downlink baseband device -> enqueued for uplink 20 baseband operation -> dequeued from the baseband device -> then the received
|
| H A D | server_node_efd.rst | 184 After initialization, packets are dequeued from the shared ring 193 :start-after: Packets dequeued from the shared ring. 8<
|
| H A D | dist_app.rst | 132 statistics include the number of packets enqueued and dequeued at each stage
|
| H A D | l2_forward_crypto.rst | 329 After this, the operations are dequeued from the device, and the transformed mbuf
|
| /dpdk/doc/guides/eventdevs/ |
| H A D | dlb2.rst | 63 event is dequeued, its sched_type will be ``RTE_SCHED_TYPE_ORDERED``. Else if 123 will be set to ``RTE_SCHED_TYPE_ATOMIC`` when it is dequeued from a directed 190 nb_events_limit (without any events being dequeued), since the other ports are 207 events are dequeued and released. 215 until it is last dequeued. If the event goes through multiple event queues, it 236 "loops", in which an event in dequeued from queue A and later enqueued back to
|
| H A D | opdl.rst | 53 Packets dequeued from this queue do not need to be re-enqueued (as is the
|
| /dpdk/examples/bbdev_app/ |
| H A D | main.c | 125 unsigned int dequeued; member 564 lcore_id, stats_border, lstats->dequeued); in print_lcore_stats() 800 lcore_stats->dequeued += nb_deq; in run_encoding() 896 lcore_stats->dequeued += nb_deq; in run_decoding()
|
| /dpdk/doc/guides/tools/ |
| H A D | testbbdev.rst | 103 - Compares the dequeued data buffer with a expected values in the test 105 - Fails if any dequeued value does not match the data in the TV 124 appearance of a dequeued result
|
| /dpdk/doc/guides/prog_guide/ |
| H A D | kernel_nic_interface.rst | 344 If an mbuf is dequeued, it will be converted to a sk_buff and sent to the net stack via netif_rx(). 345 The dequeued mbuf must be freed, so the same pointer is sent back in the free_q FIFO, 358 The mbuf is dequeued (without waiting due the cache) and filled with data from sk_buff.
|
| H A D | compressdev.rst | 414 processed i.e. dequeued with status = RTE_COMP_OP_STATUS_SUCCESS. 419 processed i.e. dequeued with status = RTE_COMP_OP_STATUS_SUCCESS. 564 (having flush = RTE_COMP_FLUSH_FINAL) is successfully processed i.e. dequeued 570 (having flush = RTE_COMP_FLUSH_FINAL) is successfully processed i.e. dequeued
|
| H A D | event_crypto_adapter.rst | 44 In this mode, events dequeued from the adapter will be treated as new events. 69 dequeued from the adapter will be treated as forwarded events. The application
|
| H A D | vhost_lib.rst | 113 packets enqueued/dequeued by async APIs are processed through the async 214 After the ``ops`` are dequeued from Cryptodev, finalizes the jobs and
|
| H A D | cryptodev_lib.rst | 762 - ``1``: the operation(s) is/are dequeued successfully. 763 - ``0``: the operation(s) is/are completed but is not actually dequeued (hence 766 operations (e.g. all operations in a descriptor) are dequeued. The crypto 775 when the user wants to abandon partially dequeued data and try dequeuing again
|
| H A D | writing_efficient_code.rst | 114 /* Process as many elements as can be dequeued. */
|
| /dpdk/doc/guides/nics/ |
| H A D | memif.rst | 127 last 1024 will belong to S2C ring. In case of zero-copy, buffers are dequeued and 207 Buffers are dequeued and enqueued as needed. Offset descriptor field is calculated at tx.
|
| /dpdk/examples/l2fwd-crypto/ |
| H A D | main.c | 246 uint64_t dequeued; member 314 crypto_statistics[cdevid].dequeued, in print_stats() 318 total_packets_dequeued += crypto_statistics[cdevid].dequeued; in print_stats() 962 crypto_statistics[cparams->dev_id].dequeued += in l2fwd_main_loop()
|
| /dpdk/drivers/crypto/qat/dev/ |
| H A D | qat_sym_pmd_gen1.c | 921 inflight = qp->enqueued - qp->dequeued; in qat_sym_dp_dequeue_burst_gen1() 1055 qp->dequeued += n; in qat_sym_dp_dequeue_done_gen1()
|
| H A D | qat_crypto_pmd_gens.h | 14 RTE_MIN((q->max_inflights - q->enqueued + q->dequeued - c), n)
|