| /f-stack/dpdk/drivers/event/dsw/ |
| H A D | dsw_event.c | 103 port->dequeued += num; in dsw_port_dequeue_stats() 115 if (dequeued > 0 && port->busy_start == 0) in dsw_port_load_record() 1335 port->in_buffer_start += dequeued; in dsw_port_dequeue_burst() 1336 port->in_buffer_len -= dequeued; in dsw_port_dequeue_burst() 1341 return dequeued; in dsw_port_dequeue_burst() 1353 uint16_t dequeued; in dsw_event_dequeue_burst() local 1364 source_port->pending_releases = dequeued; in dsw_event_dequeue_burst() 1368 dsw_port_note_op(source_port, dequeued); in dsw_event_dequeue_burst() 1370 if (dequeued > 0) { in dsw_event_dequeue_burst() 1372 dequeued); in dsw_event_dequeue_burst() [all …]
|
| H A D | dsw_xstats.c | 78 DSW_GEN_PORT_ACCESS_FN(dequeued) in DSW_GEN_PORT_ACCESS_FN() argument 106 uint64_t dequeued = in dsw_xstats_port_get_event_proc_latency() local 107 dsw->ports[port_id].dequeued; in dsw_xstats_port_get_event_proc_latency() 109 return dequeued > 0 ? total_busy_cycles / dequeued : 0; in dsw_xstats_port_get_event_proc_latency()
|
| H A D | dsw_evdev.h | 204 uint64_t dequeued; member
|
| /f-stack/dpdk/app/test-bbdev/ |
| H A D | test_bbdev_perf.c | 3953 for (i = 0, dequeued = 0; dequeued < num_to_process; ++i) { in latency_test_dec() 4004 dequeued += deq; in latency_test_dec() 4026 for (i = 0, dequeued = 0; dequeued < num_to_process; ++i) { in latency_test_ldpc_dec() 4095 dequeued += deq; in latency_test_ldpc_dec() 4112 for (i = 0, dequeued = 0; dequeued < num_to_process; ++i) { in latency_test_enc() 4161 dequeued += deq; in latency_test_enc() 4179 for (i = 0, dequeued = 0; dequeued < num_to_process; ++i) { in latency_test_ldpc_enc() 4355 for (i = 0, dequeued = 0; dequeued < num_to_process; ++i) { in offload_latency_test_dec() 4440 for (i = 0, dequeued = 0; dequeued < num_to_process; ++i) { in offload_latency_test_ldpc_dec() 4533 for (i = 0, dequeued = 0; dequeued < num_to_process; ++i) { in offload_latency_test_enc() [all …]
|
| /f-stack/dpdk/doc/guides/eventdevs/ |
| H A D | dlb.rst | 97 event is dequeued, its sched_type will be ``RTE_SCHED_TYPE_ORDERED``. Else if 157 will be set to ``RTE_SCHED_TYPE_ATOMIC`` when it is dequeued from a directed 166 the enqueued event will not be in the dequeued event. If this field is 226 nb_events_limit (without any events being dequeued), since the other ports are 243 events are dequeued and released. 251 until it is last dequeued. If the event goes through multiple event queues, it 272 "loops", in which an event in dequeued from queue A and later enqueued back to 325 dequeued event before returning from rte_event_dequeue_burst(). This frees the
|
| H A D | dlb2.rst | 62 event is dequeued, its sched_type will be ``RTE_SCHED_TYPE_ORDERED``. Else if 97 event is dequeued, its sched_type will be ``RTE_SCHED_TYPE_ORDERED``. Else if 157 will be set to ``RTE_SCHED_TYPE_ATOMIC`` when it is dequeued from a directed 217 nb_events_limit (without any events being dequeued), since the other ports are 234 events are dequeued and released. 242 until it is last dequeued. If the event goes through multiple event queues, it 263 "loops", in which an event in dequeued from queue A and later enqueued back to 321 dequeued event before returning from rte_event_dequeue_burst(). This frees the
|
| H A D | opdl.rst | 53 Packets dequeued from this queue do not need to be re-enqueued (as is the
|
| /f-stack/dpdk/drivers/common/qat/ |
| H A D | qat_qp.h | 75 uint32_t dequeued __rte_aligned(4);
|
| H A D | qat_qp.c | 236 qp->enqueued = qp->dequeued = 0; in qat_qp_setup() 326 if ((qp->enqueued - qp->dequeued) == 0) { in qat_qp_release() 610 tmp_qp->enqueued - tmp_qp->dequeued; in qat_enqueue_op_burst() 711 tmp_qp->enqueued - tmp_qp->dequeued; in qat_enqueue_comp_op_burst() 901 tmp_qp->dequeued += fw_resp_counter; in qat_dequeue_op_burst()
|
| /f-stack/dpdk/doc/guides/sample_app_ug/ |
| H A D | bbdev_app.rst | 19 operation -> dequeued from the downlink baseband device -> enqueued for uplink 20 baseband operation -> dequeued from the baseband device -> then the received
|
| H A D | l2_forward_crypto.rst | 462 After this, the operations are dequeued from the device, and the transformed mbuf 474 crypto_statistics[cparams->dev_id].dequeued +=
|
| H A D | dist_app.rst | 132 statistics include the number of packets enqueued and dequeued at each stage
|
| H A D | server_node_efd.rst | 339 After initialization, packets are dequeued from the shared ring
|
| /f-stack/dpdk/examples/bbdev_app/ |
| H A D | main.c | 127 unsigned int dequeued; member 571 lcore_id, stats_border, lstats->dequeued); in print_lcore_stats() 807 lcore_stats->dequeued += nb_deq; in run_encoding() 903 lcore_stats->dequeued += nb_deq; in run_decoding()
|
| /f-stack/dpdk/doc/guides/prog_guide/ |
| H A D | kernel_nic_interface.rst | 283 If an mbuf is dequeued, it will be converted to a sk_buff and sent to the net stack via netif_rx(). 284 The dequeued mbuf must be freed, so the same pointer is sent back in the free_q FIFO, 297 The mbuf is dequeued (without waiting due the cache) and filled with data from sk_buff.
|
| H A D | event_crypto_adapter.rst | 44 In this mode, events dequeued from the adapter will be treated as new events. 68 In this mode, events dequeued from the adapter will be treated as
|
| H A D | compressdev.rst | 414 processed i.e. dequeued with status = RTE_COMP_OP_STATUS_SUCCESS. 419 processed i.e. dequeued with status = RTE_COMP_OP_STATUS_SUCCESS. 564 (having flush = RTE_COMP_FLUSH_FINAL) is successfully processed i.e. dequeued 570 (having flush = RTE_COMP_FLUSH_FINAL) is successfully processed i.e. dequeued
|
| H A D | vhost_lib.rst | 113 the async capability. Only packets enqueued/dequeued by async APIs are 202 After the ``ops`` are dequeued from Cryptodev, finalizes the jobs and
|
| H A D | cryptodev_lib.rst | 718 - ``1``: the operation(s) is/are dequeued successfully. 719 - ``0``: the operation(s) is/are completed but is not actually dequeued (hence 722 operations (e.g. all operations in a descriptor) are dequeued. The crypto 731 when the user wants to abandon partially dequeued data and try dequeuing again
|
| H A D | writing_efficient_code.rst | 114 /* Process as many elements as can be dequeued. */
|
| H A D | ring_lib.rst | 183 The dequeued object (obj1) is copied in the pointer given by the user.
|
| /f-stack/dpdk/doc/guides/tools/ |
| H A D | testbbdev.rst | 103 - Compares the dequeued data buffer with a expected values in the test 105 - Fails if any dequeued value does not match the data in the TV 124 appearance of a dequeued result
|
| /f-stack/dpdk/drivers/crypto/qat/ |
| H A D | qat_sym_hw_dp.c | 102 RTE_MIN((q->max_inflights - q->enqueued + q->dequeued - c), n) 728 inflight = qp->enqueued - qp->dequeued; in qat_sym_dp_dequeue_burst() 856 qp->dequeued += n; in qat_sym_dp_update_head()
|
| /f-stack/dpdk/doc/guides/nics/ |
| H A D | memif.rst | 127 last 1024 will belong to M2S ring. In case of zero-copy, buffers are dequeued and 207 Buffers are dequeued and enqueued as needed. Offset descriptor field is calculated at tx.
|
| /f-stack/dpdk/examples/l2fwd-crypto/ |
| H A D | main.c | 246 uint64_t dequeued; member 316 crypto_statistics[cdevid].dequeued, in print_stats() 320 total_packets_dequeued += crypto_statistics[cdevid].dequeued; in print_stats() 934 crypto_statistics[cparams->dev_id].dequeued += in l2fwd_main_loop()
|