| /dpdk/drivers/crypto/mlx5/ |
| H A D | mlx5_crypto.c | 317 uint32_t *remain) in mlx5_crypto_klm_set() argument 322 if (data_len > *remain) in mlx5_crypto_klm_set() 323 data_len = *remain; in mlx5_crypto_klm_set() 324 *remain -= data_len; in mlx5_crypto_klm_set() 442 if (remain < nb_ops) in mlx5_crypto_enqueue_burst() 443 nb_ops = remain; in mlx5_crypto_enqueue_burst() 445 remain = nb_ops; in mlx5_crypto_enqueue_burst() 446 if (unlikely(remain == 0)) in mlx5_crypto_enqueue_burst() 455 if (remain != nb_ops) { in mlx5_crypto_enqueue_burst() 456 qp->stats.enqueued_count -= remain; in mlx5_crypto_enqueue_burst() [all …]
|
| /dpdk/drivers/compress/mlx5/ |
| H A D | mlx5_compress.c | 478 uint16_t remain = qp->entries_n - (qp->pi - qp->ci); in mlx5_compress_enqueue_burst() local 482 if (remain < nb_ops) in mlx5_compress_enqueue_burst() 483 nb_ops = remain; in mlx5_compress_enqueue_burst() 485 remain = nb_ops; in mlx5_compress_enqueue_burst() 486 if (unlikely(remain == 0)) in mlx5_compress_enqueue_burst() 517 nb_ops -= remain; in mlx5_compress_enqueue_burst() 526 } while (--remain); in mlx5_compress_enqueue_burst()
|
| /dpdk/examples/vhost/ |
| H A D | virtio_net.c | 95 uint64_t remain = dev->hdr_len; in enqueue_pkt() local 99 while (remain) { in enqueue_pkt() 100 len = remain; in enqueue_pkt() 110 remain -= len; in enqueue_pkt()
|
| /dpdk/drivers/common/sfc_efx/base/ |
| H A D | ef10_rx.c | 647 uint32_t index, remain, batch; in ef10_rx_scale_tbl_set() local 652 for (remain = nentries; remain > 0; remain -= batch) { in ef10_rx_scale_tbl_set() 653 if (batch > remain) in ef10_rx_scale_tbl_set() 654 batch = remain; in ef10_rx_scale_tbl_set()
|
| /dpdk/lib/vhost/ |
| H A D | vhost_crypto.c | 575 uint64_t remain, addr, dlen, len; in copy_data() local 592 remain = to_copy - dlen; in copy_data() 595 while (remain) { in copy_data() 596 len = remain; in copy_data() 606 remain -= len; in copy_data() 628 remain = to_copy - dlen; in copy_data() 631 while (remain) { in copy_data() 632 len = remain; in copy_data() 642 remain -= len; in copy_data()
|
| H A D | virtio_net.c | 903 uint64_t remain = dev->vhost_hlen; in copy_vnet_hdr_to_desc() local 907 while (remain) { in copy_vnet_hdr_to_desc() 908 len = RTE_MIN(remain, in copy_vnet_hdr_to_desc() 920 remain -= len; in copy_vnet_hdr_to_desc() 2454 uint64_t remain = sizeof(struct virtio_net_hdr); in copy_vnet_hdr_from_desc() local 2458 while (remain) { in copy_vnet_hdr_from_desc() 2459 len = RTE_MIN(remain, buf_vec->buf_len); in copy_vnet_hdr_from_desc() 2464 remain -= len; in copy_vnet_hdr_from_desc()
|
| H A D | vhost.c | 257 uint64_t len, remain = desc_len; in vhost_alloc_copy_ind_table() local 265 while (remain) { in vhost_alloc_copy_ind_table() 266 len = remain; in vhost_alloc_copy_ind_table() 276 remain -= len; in vhost_alloc_copy_ind_table()
|
| /dpdk/doc/guides/tools/ |
| H A D | dumpcap.rst | 29 packet capture framework, other applications remain untouched.
|
| H A D | pdump.rst | 16 framework, other applications remain untouched. So, if the ``dpdk-pdump``
|
| /dpdk/app/test/ |
| H A D | test_mbuf.c | 1276 int remain; in test_mbuf_linearize() local 1295 remain = pkt_len; in test_mbuf_linearize() 1298 for (seg = 0; remain > 0; seg++) { in test_mbuf_linearize() 1310 data_len = remain; in test_mbuf_linearize() 1329 remain -= data_len; in test_mbuf_linearize()
|
| /dpdk/doc/guides/platform/ |
| H A D | dpaa.rst | 65 configuration will remain in the hardware till it is re-configured. This
|
| /dpdk/doc/guides/howto/ |
| H A D | packet_capture_framework.rst | 55 other applications remain untouched. So, if the ``dpdk-dumpcap`` tool has to
|
| /dpdk/doc/guides/nics/ |
| H A D | ark.rst | 53 remain constant over a broad and extendable feature set. 130 be offloaded or remain in host software.
|
| H A D | mlx4.rst | 41 which remain functional, although they stop receiving unicast packets as 176 Unlike most other PMDs, these modules must remain loaded and bound to
|
| H A D | thunderx.rst | 204 They are at the beginning of VF list (the remain ones are secondary VF's).
|
| H A D | ixgbe.rst | 207 remain consistent. Running a DPDK application without polling the statistics will
|
| /dpdk/doc/guides/cryptodevs/ |
| H A D | mlx5.rst | 75 All values not specified will remain 0.
|
| /dpdk/doc/guides/rel_notes/ |
| H A D | deprecation.rst | 80 struct will remain.
|
| /dpdk/doc/guides/sample_app_ug/ |
| H A D | ipv4_multicast.rst | 210 The input packet's metadata must remain intact, so for N destination ports,
|
| H A D | link_status_intr.rst | 257 To ensure that no packets remain in the tables, each lcore does a draining of the TX queue in its m…
|
| H A D | l2_forward_real_virtual.rst | 309 To ensure that no packets remain in the tables, each lcore does a draining of TX queue in its main …
|
| H A D | l2_forward_job_stats.rst | 344 To ensure that no packets remain in the tables, the flush job exists. The l2fwd_flush_job()
|
| H A D | l2_forward_event.rst | 391 To ensure that no packets remain in the tables, each lcore does a draining of TX
|
| /dpdk/doc/guides/prog_guide/ |
| H A D | power_man.rst | 79 remain at the highest frequency until all Hyper-Threads on that core request a scale down.
|
| H A D | rte_flow.rst | 48 In order to remain as hardware-agnostic as possible, by default all rules 104 before stopping the device to ensure no rules remain. 108 However, ``rte_eth_dev_configure()`` may fail if any rules remain, 376 items, as long as those remain stacked properly. The resulting matching 3049 before stopping the device to ensure no indirect actions remain. 3054 However, ``rte_eth_dev_configure()`` may fail if any indirect actions remain, 3456 The returned value is guaranteed to remain valid only as long as no
|