| /dpdk/drivers/baseband/la12xx/ |
| H A D | bbdev_la12xx.h | 49 #define join_32_bits(upper, lower) \ argument 50 ((size_t)(((uint64_t)(upper) << 32) | (uint32_t)(lower)))
|
| /dpdk/drivers/net/igc/ |
| H A D | igc_txrx.c | 333 rxm->vlan_tci = rte_le_to_cpu_16(rxd->wb.upper.vlan); in rx_desc_get_pkt_info() 376 staterr = rte_cpu_to_le_32(rxdp->wb.upper.status_error); in igc_recv_pkts() 410 rte_le_to_cpu_16(rxd.wb.upper.length)); in igc_recv_pkts() 454 data_len = rte_le_to_cpu_16(rxd.wb.upper.length) - rxq->crc_len; in igc_recv_pkts() 524 staterr = rte_cpu_to_le_32(rxdp->wb.upper.status_error); in igc_recv_scattered_pkts() 554 rte_le_to_cpu_16(rxd.wb.upper.length)); in igc_recv_scattered_pkts() 601 data_len = rte_le_to_cpu_16(rxd.wb.upper.length); in igc_recv_scattered_pkts() 743 if (unlikely(!(rxdp->wb.upper.status_error & in eth_igc_rx_queue_count() 752 (rxdp->wb.upper.status_error & IGC_RXD_STAT_DD)) { in eth_igc_rx_queue_count() 776 status = &rxq->rx_ring[desc].wb.upper.status_error; in eth_igc_rx_descriptor_status()
|
| /dpdk/drivers/net/e1000/base/ |
| H A D | e1000_base.h | 112 } upper; member
|
| H A D | e1000_hw.h | 395 } upper; member 430 } upper; member 453 } upper; member 503 } upper; member
|
| H A D | e1000_vf.h | 82 } upper; member
|
| H A D | e1000_82575.h | 48 } upper; member
|
| /dpdk/drivers/net/igc/base/ |
| H A D | igc_base.h | 112 } upper; member
|
| H A D | igc_hw.h | 407 } upper; member 442 } upper; member 465 } upper; member 515 } upper; member
|
| H A D | igc_82575.h | 48 } upper; member
|
| /dpdk/drivers/net/e1000/ |
| H A D | igb_rxtx.c | 841 staterr = rxdp->wb.upper.status_error; in eth_igb_recv_pkts() 876 (unsigned) rte_le_to_cpu_16(rxd.wb.upper.length)); in eth_igb_recv_pkts() 945 rxm->vlan_tci = rte_be_to_cpu_16(rxd.wb.upper.vlan); in eth_igb_recv_pkts() 947 rxm->vlan_tci = rte_le_to_cpu_16(rxd.wb.upper.vlan); in eth_igb_recv_pkts() 1036 staterr = rxdp->wb.upper.status_error; in eth_igb_recv_scattered_pkts() 1067 (unsigned) rte_le_to_cpu_16(rxd.wb.upper.length)); in eth_igb_recv_scattered_pkts() 1110 data_len = rte_le_to_cpu_16(rxd.wb.upper.length); in eth_igb_recv_scattered_pkts() 1186 rte_be_to_cpu_16(rxd.wb.upper.vlan); in eth_igb_recv_scattered_pkts() 1189 rte_le_to_cpu_16(rxd.wb.upper.vlan); in eth_igb_recv_scattered_pkts() 1788 (rxdp->wb.upper.status_error & E1000_RXD_STAT_DD)) { in eth_igb_rx_queue_count() [all …]
|
| H A D | em_rxtx.c | 313 if (! (txr[desc_to_clean_to].upper.fields.status & E1000_TXD_STAT_DD)) in em_xmit_cleanup() 343 txr[desc_to_clean_to].upper.fields.status = 0; in em_xmit_cleanup() 566 txd->upper.data = rte_cpu_to_le_32(popts_spec); in eth_em_xmit_pkts() 1137 .upper.fields = {.status = E1000_TXD_STAT_DD}, in em_reset_tx_queue() 1558 status = &txq->tx_ring[desc].upper.fields.status; in eth_em_tx_descriptor_status() 2033 tx_desc->upper.data = 0; in e1000_flush_tx_ring()
|
| /dpdk/drivers/net/ixgbe/ |
| H A D | ixgbe_rxtx.c | 1392 pmc->addr = &rxdp->wb.upper.status_error; in ixgbe_get_monitor_addr() 1548 status = rxdp->wb.upper.status_error; in ixgbe_rx_scan_hw_ring() 1579 pkt_len = rte_le_to_cpu_16(rxdp[j].wb.upper.length) - in ixgbe_rx_scan_hw_ring() 1823 staterr = rxdp->wb.upper.status_error; in ixgbe_recv_pkts() 1858 (unsigned) rte_le_to_cpu_16(rxd.wb.upper.length)); in ixgbe_recv_pkts() 1920 rxm->vlan_tci = rte_le_to_cpu_16(rxd.wb.upper.vlan); in ixgbe_recv_pkts() 2015 head->vlan_tci = rte_le_to_cpu_16(desc->wb.upper.vlan); in ixgbe_fill_cluster_head_buf() 2130 rte_le_to_cpu_16(rxd.wb.upper.length)); in ixgbe_recv_pkts_lro() 2204 data_len = rte_le_to_cpu_16(rxd.wb.upper.length); in ixgbe_recv_pkts_lro() 3272 (rxdp->wb.upper.status_error & in ixgbe_dev_rx_queue_count() [all …]
|
| H A D | ixgbe_rxtx_vec_neon.c | 327 if (!(rxdp->wb.upper.status_error & in _recv_raw_pkts_vec()
|
| H A D | ixgbe_rxtx_vec_sse.c | 397 if (!(rxdp->wb.upper.status_error & in _recv_raw_pkts_vec()
|
| /dpdk/lib/bpf/ |
| H A D | bpf_jit_arm64.c | 332 uint16_t upper = val >> 16; in emit_mov_imm32() local 338 if (upper) in emit_mov_imm32() 339 mov_imm(ctx, is64, rd, A64_MOVK, upper, 16); in emit_mov_imm32() 341 if (upper == 0xffff) { in emit_mov_imm32() 344 mov_imm(ctx, is64, rd, A64_MOVN, ~upper, 16); in emit_mov_imm32()
|
| /dpdk/usertools/ |
| H A D | dpdk-hugepages.py | 30 match = re.match(r'(\d+)([' + BINARY_PREFIX + r']?)$', arg.upper())
|
| /dpdk/doc/guides/linux_gsg/ |
| H A D | linux_eal_parameters.rst | 75 Place a per-socket upper limit on memory use (non-legacy memory mode only).
|
| /dpdk/doc/guides/prog_guide/ |
| H A D | qos_framework.rst | 183 …| | | | event), pipe TC upper limit is capped … 480 …sed on time (for example, subport and pipe traffic shaping, traffic class upper limit enforcement,… 619 … | or on demand, based on the bucket_rate. Credits cannot exceed the upper | 658 …| 4 | tb_size | Bytes | Bucket size, i.e. upper limit for the tb_credits. … 749 The upper limit for the traffic classes at the subport and 762 …| 1 | tc_time | Bytes | Time of the next update (upper limit refill) for the TCs of … 778 …| 4 | tc_credits | Bytes | Current upper limit for the number of credits that can be co… 990 To ease implementation, it is also assumed that the upper limit for subport best effort TC is set t… 991 and that the upper limit for pipe best effort TC is set to 100% of pipe rate for all subport member… 998 The watermark is computed at the subport level at the beginning of each traffic class upper limit e…
|
| H A D | rte_flow.rst | 236 - ``last``: upper bound for an inclusive range with corresponding fields in 599 | ``last`` | ``id`` | upper range value | 637 | ``last`` | ``index`` | upper range value | 671 | ``last`` | ``id`` | upper range value | 701 | ``last`` | ``id`` | upper range value | 725 | ``last`` | ``data`` | upper range value | 755 | ``last`` | ``data`` | upper range value | 787 | ``last`` | ``num`` | upper range value | 1157 | ``last`` | ``threshold`` | upper range value | 1547 | ``last`` | ``port_id`` | upper range value |
|
| H A D | poll_mode_drv.rst | 75 PMDs must help global policy-oriented decisions to be enforced at the upper application level. 76 Conversely, NIC PMD functions should not impede the benefits expected by upper-level global policie… 341 It is the responsibility of the upper-level application to enforce this rule.
|
| H A D | generic_segmentation_offload_lib.rst | 11 Much like TSO, GSO gains performance by enabling upper layer applications to
|
| /dpdk/doc/guides/eventdevs/ |
| H A D | cnxk.rst | 65 upper limit for in-flight events.
|
| /dpdk/drivers/net/dpaa2/ |
| H A D | dpaa2_recycle.c | 157 uint32_t upper; member
|
| /dpdk/doc/guides/rel_notes/ |
| H A D | release_21_02.rst | 196 the upper-case component name, e.g. ``EAL``, ``ETHDEV``, ``VIRTIO``, 197 and ``CLASS`` is the upper-case driver class, e.g. ``NET``, ``CRYPTO``.
|
| /dpdk/doc/guides/nics/ |
| H A D | ixgbe.rst | 192 the info about the header length from the upper layer, because parsing the
|