| /f-stack/tools/traffic/ |
| H A D | traffic.c | 12 int traffic_status(struct ff_traffic_args *traffic) in traffic_status() argument 43 *traffic = retmsg->traffic; in traffic_status() 55 struct ff_traffic_args traffic = {0, 0, 0, 0}, otr; in main() local 64 #define DIFF(member) (traffic.member - otr.member) in main() 103 if (traffic_status(&traffic)) { in main() 109 printf("%lu,%lu,%lu,%lu\n", traffic.rx_packets, traffic.rx_bytes, in main() 110 traffic.tx_packets, traffic.tx_bytes); in main() 131 "total", traffic.rx_packets, traffic.rx_bytes, in main() 132 traffic.tx_packets, traffic.tx_bytes); in main() 140 if (traffic_status(&traffic)) { in main() [all …]
|
| H A D | Makefile | 7 PROG=traffic
|
| /f-stack/dpdk/doc/guides/prog_guide/ |
| H A D | switch_representation.rst | 23 layer 2 (L2) traffic (such as OVS) need to steer traffic themselves 465 | | | traffic : | | | traffic 474 | | | traffic : | ^ | | traffic 527 | | | traffic | v 541 | | <-- traffic | 562 - Matches **F** in `traffic steering`_. 569 - Targets **F** in `traffic steering`_. 606 - Matches **A** in `traffic steering`_. 615 - Targets **A** in `traffic steering`_. 778 traffic. [all …]
|
| H A D | qos_framework.rst | 133 Each traffic class is the representation of a different traffic type with specific loss rate, 391 …These fields are: port, subport, traffic class and queue within traffic class, and are typically s… 442 …ength and the available credits (of current pipe, pipe traffic class, subport and subport traffic … 479 which requires credit updates based on time (for example, subport and pipe traffic shaping, traffic… 550 subport S traffic class TC, pipe P, pipe P traffic class TC. 746 The traffic classes at the pipe and subport levels are not traffic shaped, 748 The upper limit for the traffic classes at the subport and 931 allocated for the same traffic class at the parent subport level. 934 traffic class is solely the result of pipe and 986 which is typically used for best effort traffic, [all …]
|
| H A D | power_man.rst | 128 to improved handling of bursts of traffic. 131 frequency (including turbo) to do best effort for intensive traffic. This gives 132 us more flexible and balanced traffic awareness over the standard l3fwd-power 153 * MED: the frequency is used to process modest traffic workload. 155 * HIGH: the frequency is used to process busy traffic workload. 163 In this phase, the user must ensure that no traffic can enter the 167 displayed, and normal mode resumes, and traffic can be allowed into
|
| /f-stack/dpdk/drivers/event/dlb/ |
| H A D | dlb_xstats.c | 94 val += port->stats.traffic.rx_ok; in dlb_device_traffic_stat_get() 97 val += port->stats.traffic.rx_drop; in dlb_device_traffic_stat_get() 106 val += port->stats.traffic.tx_ok; in dlb_device_traffic_stat_get() 109 val += port->stats.traffic.total_polls; in dlb_device_traffic_stat_get() 112 val += port->stats.traffic.zero_polls; in dlb_device_traffic_stat_get() 1097 p->stats.traffic.rx_ok); in dlb_eventdev_dump() 1100 p->stats.traffic.rx_drop); in dlb_eventdev_dump() 1103 p->stats.traffic.rx_interrupt_wait); in dlb_eventdev_dump() 1109 p->stats.traffic.tx_ok); in dlb_eventdev_dump() 1112 p->stats.traffic.total_polls); in dlb_eventdev_dump() [all …]
|
| /f-stack/dpdk/drivers/event/dlb2/ |
| H A D | dlb2_xstats.c | 106 val += port->stats.traffic.rx_ok; in dlb2_device_traffic_stat_get() 109 val += port->stats.traffic.rx_drop; in dlb2_device_traffic_stat_get() 118 val += port->stats.traffic.tx_ok; in dlb2_device_traffic_stat_get() 121 val += port->stats.traffic.total_polls; in dlb2_device_traffic_stat_get() 124 val += port->stats.traffic.zero_polls; in dlb2_device_traffic_stat_get() 1115 p->stats.traffic.rx_ok); in dlb2_eventdev_dump() 1118 p->stats.traffic.rx_drop); in dlb2_eventdev_dump() 1121 p->stats.traffic.rx_interrupt_wait); in dlb2_eventdev_dump() 1127 p->stats.traffic.tx_ok); in dlb2_eventdev_dump() 1130 p->stats.traffic.total_polls); in dlb2_eventdev_dump() [all …]
|
| /f-stack/dpdk/examples/ipsec-secgw/ |
| H A D | ipsec-secgw.c | 725 traffic->ipsec.saptr, traffic->ipsec.num); in process_pkts_inbound() 784 traffic->ipsec.res, traffic->ipsec.num, in process_pkts_outbound() 800 traffic->ipsec.saptr, traffic->ipsec.num); in process_pkts_outbound() 813 free_pkts(traffic->ip4.pkts, traffic->ip4.num); in process_pkts_inbound_nosp() 818 free_pkts(traffic->ip6.pkts, traffic->ip6.num); in process_pkts_inbound_nosp() 840 traffic->ipsec.saptr, traffic->ipsec.num); in process_pkts_inbound_nosp() 859 traffic->ipsec.pkts[n] = traffic->ip4.pkts[i]; in process_pkts_outbound_nosp() 864 traffic->ipsec.pkts[n] = traffic->ip6.pkts[i]; in process_pkts_outbound_nosp() 875 traffic->ipsec.res, traffic->ipsec.num, in process_pkts_outbound_nosp() 884 traffic->ip4.pkts[i] = traffic->ipsec.pkts[i]; in process_pkts_outbound_nosp() [all …]
|
| /f-stack/dpdk/doc/guides/howto/ |
| H A D | flow_bifurcation.rst | 8 to split traffic between Linux user space and kernel space. Since it is a 12 movement during the traffic split. This can yield better performance with 15 The Flow Bifurcation splits the incoming data traffic to user space 17 the Linux kernel stack). It can direct some traffic, for example data plane 18 traffic, to DPDK, while directing some other traffic, for example control 19 plane traffic, to the traditional Linux networking stack. 26 with physical functions (PF). The network adapter will direct traffic to a 35 In this way the Linux networking stack can receive specific traffic through 36 the kernel driver while a DPDK application can receive specific traffic
|
| /f-stack/freebsd/contrib/ngatm/netnatm/msg/ |
| H A D | traffic.c | 73 const struct uni_ie_traffic *traffic, in uni_classify_traffic() argument 192 ft = traffic->h.present & fmask; in uni_classify_traffic() 193 bt = traffic->h.present & bmask; in uni_classify_traffic() 194 be = traffic->h.present & UNI_TRAFFIC_BEST_P; in uni_classify_traffic() 195 ftag = (traffic->h.present & UNI_TRAFFIC_MOPT_P) && traffic->t.ftag; in uni_classify_traffic() 196 btag = (traffic->h.present & UNI_TRAFFIC_MOPT_P) && traffic->t.btag; in uni_classify_traffic()
|
| H A D | ie.def | 46 element traffic 0x59 itu 30 47 element traffic 0x59 net 67 element mintraffic 0x81 itu 20 file=traffic 69 element atraffic 0x82 itu 30 file=traffic 83 element mdcr 0xf0 net 13 file=traffic
|
| H A D | uni_msg.h | 40 struct uni_ie_traffic traffic; member 79 struct uni_ie_traffic traffic; member 272 struct uni_ie_traffic traffic; member 283 struct uni_ie_traffic traffic; member
|
| /f-stack/dpdk/doc/guides/sample_app_ug/ |
| H A D | packet_ordering.rst | 15 * RX core (main core) receives traffic from the NIC ports and feeds Worker 16 cores with traffic through SW queues. 22 * TX Core (worker core) receives traffic from Worker cores through software queues, 54 When setting more than 1 port, traffic would be forwarded in pairs. 55 For example, if we enable 4 ports, traffic from port 0 to 1 and from 1 to 0, 59 of traffic, which should help evaluate reordering performance impact.
|
| H A D | vmdq_dcb_forwarding.rst | 8 The application performs L2 forwarding using VMDQ and DCB to divide the incoming traffic into queue… 9 The traffic splitting is performed in hardware by the VMDQ and DCB features of the Intel® 82599 and… 15 uses VMDQ and DCB for traffic partitioning. 17 The VMDQ and DCB filters work on MAC and VLAN traffic to divide the traffic into input queues on th… 19 VMDQ filters split the traffic into 16 or 32 groups based on the Destination MAC and VLAN ID. 22 All traffic is read from a single incoming port (port 0) and output on port 1, without any processi… 23 With Intel® 82599 NIC, for example, the traffic is split into 128 queues on input, where each threa… 27 The Intel® 82599 10 Gigabit Ethernet Controller NIC also supports the splitting of traffic into 16 … 146 (also referred to as traffic classes) within each pool. With Intel® 82599 NIC, 173 * given above, and the number of traffic classes available for use. */
|
| H A D | bbdev_app.rst | 93 To allow the bbdev sample app to do the loopback, an influx of traffic is required. 94 This can be done by using DPDK Pktgen to burst traffic on two ethernet ports, and 95 it will print the transmitted along with the looped-back traffic on Rx ports. 96 Executing the command below will generate traffic on the two allowed ethernet
|
| H A D | ipsec_secgw.rst | 44 The Path for IPsec Inbound traffic is: 405 * The traffic direction 411 * *in*: inbound traffic 412 * *out*: outbound traffic 424 * *bypass*: the specified traffic traffic is bypassed 509 * The traffic direction 515 * *in*: inbound traffic 516 * *out*: outbound traffic 791 * The traffic output port id 856 * The traffic input port id [all …]
|
| H A D | vmdq_forwarding.rst | 8 The application performs L2 forwarding using VMDq to divide the incoming traffic into queues. 9 The traffic splitting is performed in hardware by the VMDq feature of the Intel® 82599 and X710/XL7… 15 uses VMDq for traffic partitioning. 20 All traffic is read from a single incoming port and output on another port, without any processing … 21 With Intel® 82599 NIC, for example, the traffic is split into 128 queues on input, where each threa… 25 The Intel® 82599 10 Gigabit Ethernet Controller NIC also supports the splitting of traffic into 16 …
|
| H A D | qos_metering.rst | 74 To simplify debugging (for example, by using the traffic generator RX side MAC address based packet… 77 The traffic meter parameters are configured in the application source code with following default v… 93 Assuming the input traffic is generated at line rate and all packets are 64 bytes Ethernet frames (… 94 and green, the expected output traffic should be marked as shown in the following table:
|
| H A D | test_pipeline.rst | 15 * Core A ("RX core") receives traffic from the NIC ports and feeds core B with traffic through SW… 19 Core B receives traffic from core A through software queues, 23 * Core C ("TX core") receives traffic from core B through software queues and sends it to the NIC… 229 the same input traffic can be used to hit all table entries with uniform distribution, 231 The profile for input traffic is TCP/IPv4 packets with:
|
| /f-stack/dpdk/doc/guides/nics/ |
| H A D | softnic.rst | 31 such as: memory pools, SW queues, traffic manager, action profiles, pipelines, 73 #. ``tm_n_queues``: number of traffic manager's scheduler queues. The traffic manager 76 #. ``tm_qsize0``: size of scheduler queue 0 (traffic class 0) of the pipes/subscribers. 79 #. ``tm_qsize1``: size of scheduler queue 1 (traffic class 1) of the pipes/subscribers. 82 #. ``tm_qsize2``: size of scheduler queue 2 (traffic class 2) of the pipes/subscribers. 85 #. ``tm_qsize3``: size of scheduler queue 3 (traffic class 3) of the pipes/subscribers. 274 SoftNIC PMD implements ethdev traffic management APIs ``rte_tm.h`` that 275 allow building and committing traffic manager hierarchy, configuring hierarchy 277 library. Furthermore, APIs for run-time update to the traffic manager hierarchy 280 SoftNIC PMD also implements ethdev traffic metering and policing APIs [all …]
|
| /f-stack/dpdk/examples/qos_sched/ |
| H A D | profile.cfg | 8 ; - Each of the 13 traffic classes has rate set to 100% of port rate 11 ; - Each of the 13 traffic classes has rate set to 100% of pipe rate 12 ; - Within lowest priority traffic class (best-effort), the byte-level 13 ; WRR weights for the 4 queues of best effort traffic class are set 78 ; RED params per traffic class and color (Green / Yellow / Red)
|
| /f-stack/dpdk/lib/librte_power/ |
| H A D | channel_commands.h | 72 struct traffic { struct 89 struct traffic traffic_policy; argument
|
| /f-stack/freebsd/contrib/ngatm/netnatm/sig/ |
| H A D | sig_unimsgcpy.c | 70 if(IE_ISGOOD(src->traffic)) in copy_msg_connect() 71 dst->traffic = src->traffic; in copy_msg_connect() 146 if(IE_ISGOOD(src->traffic)) in copy_msg_setup() 147 dst->traffic = src->traffic; in copy_msg_setup() 522 if(IE_ISGOOD(src->traffic)) in copy_msg_modify_req() 523 dst->traffic = src->traffic; in copy_msg_modify_req() 544 if(IE_ISGOOD(src->traffic)) in copy_msg_modify_ack() 545 dst->traffic = src->traffic; in copy_msg_modify_ack()
|
| /f-stack/tools/ |
| H A D | Makefile | 1 …il libmemstat libxo libnetgraph sysctl ifconfig route top netstat ngctl ipfw arp traffic knictl ndp 23 ln -sf ${PREFIX_BIN}/f-stack/traffic ${PREFIX_BIN}/ff_traffic
|
| /f-stack/dpdk/doc/guides/vdpadevs/ |
| H A D | mlx5.rst | 111 automatically adjusts its delays to the coming traffic rate. 117 arms the CQ in order to get an interrupt event in the next traffic burst. 133 A nonzero value defines the traffic off time, in seconds, that moves the 134 driver to no-traffic mode. In this mode the timer events are stopped and 135 interrupts are configured to the device in order to notify traffic for the
|