1*2bfe3f2eSlogwang /*-
2*2bfe3f2eSlogwang  *   BSD LICENSE
3*2bfe3f2eSlogwang  *
4*2bfe3f2eSlogwang  *   Copyright(c) 2016-2017 Intel Corporation. All rights reserved.
5*2bfe3f2eSlogwang  *   All rights reserved.
6*2bfe3f2eSlogwang  *
7*2bfe3f2eSlogwang  *   Redistribution and use in source and binary forms, with or without
8*2bfe3f2eSlogwang  *   modification, are permitted provided that the following conditions
9*2bfe3f2eSlogwang  *   are met:
10*2bfe3f2eSlogwang  *
11*2bfe3f2eSlogwang  *     * Redistributions of source code must retain the above copyright
12*2bfe3f2eSlogwang  *       notice, this list of conditions and the following disclaimer.
13*2bfe3f2eSlogwang  *     * Redistributions in binary form must reproduce the above copyright
14*2bfe3f2eSlogwang  *       notice, this list of conditions and the following disclaimer in
15*2bfe3f2eSlogwang  *       the documentation and/or other materials provided with the
16*2bfe3f2eSlogwang  *       distribution.
17*2bfe3f2eSlogwang  *     * Neither the name of Intel Corporation nor the names of its
18*2bfe3f2eSlogwang  *       contributors may be used to endorse or promote products derived
19*2bfe3f2eSlogwang  *       from this software without specific prior written permission.
20*2bfe3f2eSlogwang  *
21*2bfe3f2eSlogwang  *   THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
22*2bfe3f2eSlogwang  *   "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
23*2bfe3f2eSlogwang  *   LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
24*2bfe3f2eSlogwang  *   A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
25*2bfe3f2eSlogwang  *   OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
26*2bfe3f2eSlogwang  *   SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
27*2bfe3f2eSlogwang  *   LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
28*2bfe3f2eSlogwang  *   DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
29*2bfe3f2eSlogwang  *   THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
30*2bfe3f2eSlogwang  *   (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
31*2bfe3f2eSlogwang  *   OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
32*2bfe3f2eSlogwang  */
33*2bfe3f2eSlogwang 
34*2bfe3f2eSlogwang #include <stdio.h>
35*2bfe3f2eSlogwang #include <stdlib.h>
36*2bfe3f2eSlogwang #include <string.h>
37*2bfe3f2eSlogwang #include <unistd.h>
38*2bfe3f2eSlogwang #include <stdint.h>
39*2bfe3f2eSlogwang #include <stdarg.h>
40*2bfe3f2eSlogwang #include <inttypes.h>
41*2bfe3f2eSlogwang #include <sys/queue.h>
42*2bfe3f2eSlogwang #include <errno.h>
43*2bfe3f2eSlogwang #include <netinet/ip.h>
44*2bfe3f2eSlogwang 
45*2bfe3f2eSlogwang #include <rte_common.h>
46*2bfe3f2eSlogwang #include <rte_memory.h>
47*2bfe3f2eSlogwang #include <rte_eal.h>
48*2bfe3f2eSlogwang #include <rte_launch.h>
49*2bfe3f2eSlogwang #include <rte_per_lcore.h>
50*2bfe3f2eSlogwang #include <rte_lcore.h>
51*2bfe3f2eSlogwang #include <rte_branch_prediction.h>
52*2bfe3f2eSlogwang #include <rte_atomic.h>
53*2bfe3f2eSlogwang #include <rte_ring.h>
54*2bfe3f2eSlogwang #include <rte_log.h>
55*2bfe3f2eSlogwang #include <rte_debug.h>
56*2bfe3f2eSlogwang #include <rte_mempool.h>
57*2bfe3f2eSlogwang #include <rte_memcpy.h>
58*2bfe3f2eSlogwang #include <rte_mbuf.h>
59*2bfe3f2eSlogwang #include <rte_ether.h>
60*2bfe3f2eSlogwang #include <rte_interrupts.h>
61*2bfe3f2eSlogwang #include <rte_ethdev.h>
62*2bfe3f2eSlogwang #include <rte_byteorder.h>
63*2bfe3f2eSlogwang #include <rte_malloc.h>
64*2bfe3f2eSlogwang #include <rte_string_fns.h>
65*2bfe3f2eSlogwang #include <rte_efd.h>
66*2bfe3f2eSlogwang #include <rte_ip.h>
67*2bfe3f2eSlogwang 
68*2bfe3f2eSlogwang #include "common.h"
69*2bfe3f2eSlogwang #include "args.h"
70*2bfe3f2eSlogwang #include "init.h"
71*2bfe3f2eSlogwang 
72*2bfe3f2eSlogwang /*
73*2bfe3f2eSlogwang  * When doing reads from the NIC or the node queues,
74*2bfe3f2eSlogwang  * use this batch size
75*2bfe3f2eSlogwang  */
76*2bfe3f2eSlogwang #define PACKET_READ_SIZE 32
77*2bfe3f2eSlogwang 
78*2bfe3f2eSlogwang /*
79*2bfe3f2eSlogwang  * Local buffers to put packets in, used to send packets in bursts to the
80*2bfe3f2eSlogwang  * nodes
81*2bfe3f2eSlogwang  */
82*2bfe3f2eSlogwang struct node_rx_buf {
83*2bfe3f2eSlogwang 	struct rte_mbuf *buffer[PACKET_READ_SIZE];
84*2bfe3f2eSlogwang 	uint16_t count;
85*2bfe3f2eSlogwang };
86*2bfe3f2eSlogwang 
87*2bfe3f2eSlogwang struct efd_stats {
88*2bfe3f2eSlogwang 	uint64_t distributed;
89*2bfe3f2eSlogwang 	uint64_t drop;
90*2bfe3f2eSlogwang } flow_dist_stats;
91*2bfe3f2eSlogwang 
92*2bfe3f2eSlogwang /* One buffer per node rx queue - dynamically allocate array */
93*2bfe3f2eSlogwang static struct node_rx_buf *cl_rx_buf;
94*2bfe3f2eSlogwang 
95*2bfe3f2eSlogwang static const char *
96*2bfe3f2eSlogwang get_printable_mac_addr(uint16_t port)
97*2bfe3f2eSlogwang {
98*2bfe3f2eSlogwang 	static const char err_address[] = "00:00:00:00:00:00";
99*2bfe3f2eSlogwang 	static char addresses[RTE_MAX_ETHPORTS][sizeof(err_address)];
100*2bfe3f2eSlogwang 	struct ether_addr mac;
101*2bfe3f2eSlogwang 
102*2bfe3f2eSlogwang 	if (unlikely(port >= RTE_MAX_ETHPORTS))
103*2bfe3f2eSlogwang 		return err_address;
104*2bfe3f2eSlogwang 	if (unlikely(addresses[port][0] == '\0')) {
105*2bfe3f2eSlogwang 		rte_eth_macaddr_get(port, &mac);
106*2bfe3f2eSlogwang 		snprintf(addresses[port], sizeof(addresses[port]),
107*2bfe3f2eSlogwang 				"%02x:%02x:%02x:%02x:%02x:%02x\n",
108*2bfe3f2eSlogwang 				mac.addr_bytes[0], mac.addr_bytes[1],
109*2bfe3f2eSlogwang 				mac.addr_bytes[2], mac.addr_bytes[3],
110*2bfe3f2eSlogwang 				mac.addr_bytes[4], mac.addr_bytes[5]);
111*2bfe3f2eSlogwang 	}
112*2bfe3f2eSlogwang 	return addresses[port];
113*2bfe3f2eSlogwang }
114*2bfe3f2eSlogwang 
115*2bfe3f2eSlogwang /*
116*2bfe3f2eSlogwang  * This function displays the recorded statistics for each port
117*2bfe3f2eSlogwang  * and for each node. It uses ANSI terminal codes to clear
118*2bfe3f2eSlogwang  * screen when called. It is called from a single non-master
119*2bfe3f2eSlogwang  * thread in the server process, when the process is run with more
120*2bfe3f2eSlogwang  * than one lcore enabled.
121*2bfe3f2eSlogwang  */
122*2bfe3f2eSlogwang static void
123*2bfe3f2eSlogwang do_stats_display(void)
124*2bfe3f2eSlogwang {
125*2bfe3f2eSlogwang 	unsigned int i, j;
126*2bfe3f2eSlogwang 	const char clr[] = {27, '[', '2', 'J', '\0'};
127*2bfe3f2eSlogwang 	const char topLeft[] = {27, '[', '1', ';', '1', 'H', '\0'};
128*2bfe3f2eSlogwang 	uint64_t port_tx[RTE_MAX_ETHPORTS], port_tx_drop[RTE_MAX_ETHPORTS];
129*2bfe3f2eSlogwang 	uint64_t node_tx[MAX_NODES], node_tx_drop[MAX_NODES];
130*2bfe3f2eSlogwang 
131*2bfe3f2eSlogwang 	/* to get TX stats, we need to do some summing calculations */
132*2bfe3f2eSlogwang 	memset(port_tx, 0, sizeof(port_tx));
133*2bfe3f2eSlogwang 	memset(port_tx_drop, 0, sizeof(port_tx_drop));
134*2bfe3f2eSlogwang 	memset(node_tx, 0, sizeof(node_tx));
135*2bfe3f2eSlogwang 	memset(node_tx_drop, 0, sizeof(node_tx_drop));
136*2bfe3f2eSlogwang 
137*2bfe3f2eSlogwang 	for (i = 0; i < num_nodes; i++) {
138*2bfe3f2eSlogwang 		const struct tx_stats *tx = &info->tx_stats[i];
139*2bfe3f2eSlogwang 
140*2bfe3f2eSlogwang 		for (j = 0; j < info->num_ports; j++) {
141*2bfe3f2eSlogwang 			const uint64_t tx_val = tx->tx[info->id[j]];
142*2bfe3f2eSlogwang 			const uint64_t drop_val = tx->tx_drop[info->id[j]];
143*2bfe3f2eSlogwang 
144*2bfe3f2eSlogwang 			port_tx[j] += tx_val;
145*2bfe3f2eSlogwang 			port_tx_drop[j] += drop_val;
146*2bfe3f2eSlogwang 			node_tx[i] += tx_val;
147*2bfe3f2eSlogwang 			node_tx_drop[i] += drop_val;
148*2bfe3f2eSlogwang 		}
149*2bfe3f2eSlogwang 	}
150*2bfe3f2eSlogwang 
151*2bfe3f2eSlogwang 	/* Clear screen and move to top left */
152*2bfe3f2eSlogwang 	printf("%s%s", clr, topLeft);
153*2bfe3f2eSlogwang 
154*2bfe3f2eSlogwang 	printf("PORTS\n");
155*2bfe3f2eSlogwang 	printf("-----\n");
156*2bfe3f2eSlogwang 	for (i = 0; i < info->num_ports; i++)
157*2bfe3f2eSlogwang 		printf("Port %u: '%s'\t", (unsigned int)info->id[i],
158*2bfe3f2eSlogwang 				get_printable_mac_addr(info->id[i]));
159*2bfe3f2eSlogwang 	printf("\n\n");
160*2bfe3f2eSlogwang 	for (i = 0; i < info->num_ports; i++) {
161*2bfe3f2eSlogwang 		printf("Port %u - rx: %9"PRIu64"\t"
162*2bfe3f2eSlogwang 				"tx: %9"PRIu64"\n",
163*2bfe3f2eSlogwang 				(unsigned int)info->id[i], info->rx_stats.rx[i],
164*2bfe3f2eSlogwang 				port_tx[i]);
165*2bfe3f2eSlogwang 	}
166*2bfe3f2eSlogwang 
167*2bfe3f2eSlogwang 	printf("\nSERVER\n");
168*2bfe3f2eSlogwang 	printf("-----\n");
169*2bfe3f2eSlogwang 	printf("distributed: %9"PRIu64", drop: %9"PRIu64"\n",
170*2bfe3f2eSlogwang 			flow_dist_stats.distributed, flow_dist_stats.drop);
171*2bfe3f2eSlogwang 
172*2bfe3f2eSlogwang 	printf("\nNODES\n");
173*2bfe3f2eSlogwang 	printf("-------\n");
174*2bfe3f2eSlogwang 	for (i = 0; i < num_nodes; i++) {
175*2bfe3f2eSlogwang 		const unsigned long long rx = nodes[i].stats.rx;
176*2bfe3f2eSlogwang 		const unsigned long long rx_drop = nodes[i].stats.rx_drop;
177*2bfe3f2eSlogwang 		const struct filter_stats *filter = &info->filter_stats[i];
178*2bfe3f2eSlogwang 
179*2bfe3f2eSlogwang 		printf("Node %2u - rx: %9llu, rx_drop: %9llu\n"
180*2bfe3f2eSlogwang 				"            tx: %9"PRIu64", tx_drop: %9"PRIu64"\n"
181*2bfe3f2eSlogwang 				"            filter_passed: %9"PRIu64", "
182*2bfe3f2eSlogwang 				"filter_drop: %9"PRIu64"\n",
183*2bfe3f2eSlogwang 				i, rx, rx_drop, node_tx[i], node_tx_drop[i],
184*2bfe3f2eSlogwang 				filter->passed, filter->drop);
185*2bfe3f2eSlogwang 	}
186*2bfe3f2eSlogwang 
187*2bfe3f2eSlogwang 	printf("\n");
188*2bfe3f2eSlogwang }
189*2bfe3f2eSlogwang 
190*2bfe3f2eSlogwang /*
191*2bfe3f2eSlogwang  * The function called from each non-master lcore used by the process.
192*2bfe3f2eSlogwang  * The test_and_set function is used to randomly pick a single lcore on which
193*2bfe3f2eSlogwang  * the code to display the statistics will run. Otherwise, the code just
194*2bfe3f2eSlogwang  * repeatedly sleeps.
195*2bfe3f2eSlogwang  */
196*2bfe3f2eSlogwang static int
197*2bfe3f2eSlogwang sleep_lcore(__attribute__((unused)) void *dummy)
198*2bfe3f2eSlogwang {
199*2bfe3f2eSlogwang 	/* Used to pick a display thread - static, so zero-initialised */
200*2bfe3f2eSlogwang 	static rte_atomic32_t display_stats;
201*2bfe3f2eSlogwang 
202*2bfe3f2eSlogwang 	/* Only one core should display stats */
203*2bfe3f2eSlogwang 	if (rte_atomic32_test_and_set(&display_stats)) {
204*2bfe3f2eSlogwang 		const unsigned int sleeptime = 1;
205*2bfe3f2eSlogwang 
206*2bfe3f2eSlogwang 		printf("Core %u displaying statistics\n", rte_lcore_id());
207*2bfe3f2eSlogwang 
208*2bfe3f2eSlogwang 		/* Longer initial pause so above printf is seen */
209*2bfe3f2eSlogwang 		sleep(sleeptime * 3);
210*2bfe3f2eSlogwang 
211*2bfe3f2eSlogwang 		/* Loop forever: sleep always returns 0 or <= param */
212*2bfe3f2eSlogwang 		while (sleep(sleeptime) <= sleeptime)
213*2bfe3f2eSlogwang 			do_stats_display();
214*2bfe3f2eSlogwang 	}
215*2bfe3f2eSlogwang 	return 0;
216*2bfe3f2eSlogwang }
217*2bfe3f2eSlogwang 
218*2bfe3f2eSlogwang /*
219*2bfe3f2eSlogwang  * Function to set all the node statistic values to zero.
220*2bfe3f2eSlogwang  * Called at program startup.
221*2bfe3f2eSlogwang  */
222*2bfe3f2eSlogwang static void
223*2bfe3f2eSlogwang clear_stats(void)
224*2bfe3f2eSlogwang {
225*2bfe3f2eSlogwang 	unsigned int i;
226*2bfe3f2eSlogwang 
227*2bfe3f2eSlogwang 	for (i = 0; i < num_nodes; i++)
228*2bfe3f2eSlogwang 		nodes[i].stats.rx = nodes[i].stats.rx_drop = 0;
229*2bfe3f2eSlogwang }
230*2bfe3f2eSlogwang 
231*2bfe3f2eSlogwang /*
232*2bfe3f2eSlogwang  * send a burst of traffic to a node, assuming there are packets
233*2bfe3f2eSlogwang  * available to be sent to this node
234*2bfe3f2eSlogwang  */
235*2bfe3f2eSlogwang static void
236*2bfe3f2eSlogwang flush_rx_queue(uint16_t node)
237*2bfe3f2eSlogwang {
238*2bfe3f2eSlogwang 	uint16_t j;
239*2bfe3f2eSlogwang 	struct node *cl;
240*2bfe3f2eSlogwang 
241*2bfe3f2eSlogwang 	if (cl_rx_buf[node].count == 0)
242*2bfe3f2eSlogwang 		return;
243*2bfe3f2eSlogwang 
244*2bfe3f2eSlogwang 	cl = &nodes[node];
245*2bfe3f2eSlogwang 	if (rte_ring_enqueue_bulk(cl->rx_q, (void **)cl_rx_buf[node].buffer,
246*2bfe3f2eSlogwang 			cl_rx_buf[node].count, NULL) != cl_rx_buf[node].count){
247*2bfe3f2eSlogwang 		for (j = 0; j < cl_rx_buf[node].count; j++)
248*2bfe3f2eSlogwang 			rte_pktmbuf_free(cl_rx_buf[node].buffer[j]);
249*2bfe3f2eSlogwang 		cl->stats.rx_drop += cl_rx_buf[node].count;
250*2bfe3f2eSlogwang 	} else
251*2bfe3f2eSlogwang 		cl->stats.rx += cl_rx_buf[node].count;
252*2bfe3f2eSlogwang 
253*2bfe3f2eSlogwang 	cl_rx_buf[node].count = 0;
254*2bfe3f2eSlogwang }
255*2bfe3f2eSlogwang 
256*2bfe3f2eSlogwang /*
257*2bfe3f2eSlogwang  * marks a packet down to be sent to a particular node process
258*2bfe3f2eSlogwang  */
259*2bfe3f2eSlogwang static inline void
260*2bfe3f2eSlogwang enqueue_rx_packet(uint8_t node, struct rte_mbuf *buf)
261*2bfe3f2eSlogwang {
262*2bfe3f2eSlogwang 	cl_rx_buf[node].buffer[cl_rx_buf[node].count++] = buf;
263*2bfe3f2eSlogwang }
264*2bfe3f2eSlogwang 
265*2bfe3f2eSlogwang /*
266*2bfe3f2eSlogwang  * This function takes a group of packets and routes them
267*2bfe3f2eSlogwang  * individually to the node process. Very simply round-robins the packets
268*2bfe3f2eSlogwang  * without checking any of the packet contents.
269*2bfe3f2eSlogwang  */
270*2bfe3f2eSlogwang static void
271*2bfe3f2eSlogwang process_packets(uint32_t port_num __rte_unused, struct rte_mbuf *pkts[],
272*2bfe3f2eSlogwang 		uint16_t rx_count, unsigned int socket_id)
273*2bfe3f2eSlogwang {
274*2bfe3f2eSlogwang 	uint16_t i;
275*2bfe3f2eSlogwang 	uint8_t node;
276*2bfe3f2eSlogwang 	efd_value_t data[RTE_EFD_BURST_MAX];
277*2bfe3f2eSlogwang 	const void *key_ptrs[RTE_EFD_BURST_MAX];
278*2bfe3f2eSlogwang 
279*2bfe3f2eSlogwang 	struct ipv4_hdr *ipv4_hdr;
280*2bfe3f2eSlogwang 	uint32_t ipv4_dst_ip[RTE_EFD_BURST_MAX];
281*2bfe3f2eSlogwang 
282*2bfe3f2eSlogwang 	for (i = 0; i < rx_count; i++) {
283*2bfe3f2eSlogwang 		/* Handle IPv4 header.*/
284*2bfe3f2eSlogwang 		ipv4_hdr = rte_pktmbuf_mtod_offset(pkts[i], struct ipv4_hdr *,
285*2bfe3f2eSlogwang 				sizeof(struct ether_hdr));
286*2bfe3f2eSlogwang 		ipv4_dst_ip[i] = ipv4_hdr->dst_addr;
287*2bfe3f2eSlogwang 		key_ptrs[i] = (void *)&ipv4_dst_ip[i];
288*2bfe3f2eSlogwang 	}
289*2bfe3f2eSlogwang 
290*2bfe3f2eSlogwang 	rte_efd_lookup_bulk(efd_table, socket_id, rx_count,
291*2bfe3f2eSlogwang 				(const void **) key_ptrs, data);
292*2bfe3f2eSlogwang 	for (i = 0; i < rx_count; i++) {
293*2bfe3f2eSlogwang 		node = (uint8_t) ((uintptr_t)data[i]);
294*2bfe3f2eSlogwang 
295*2bfe3f2eSlogwang 		if (node >= num_nodes) {
296*2bfe3f2eSlogwang 			/*
297*2bfe3f2eSlogwang 			 * Node is out of range, which means that
298*2bfe3f2eSlogwang 			 * flow has not been inserted
299*2bfe3f2eSlogwang 			 */
300*2bfe3f2eSlogwang 			flow_dist_stats.drop++;
301*2bfe3f2eSlogwang 			rte_pktmbuf_free(pkts[i]);
302*2bfe3f2eSlogwang 		} else {
303*2bfe3f2eSlogwang 			flow_dist_stats.distributed++;
304*2bfe3f2eSlogwang 			enqueue_rx_packet(node, pkts[i]);
305*2bfe3f2eSlogwang 		}
306*2bfe3f2eSlogwang 	}
307*2bfe3f2eSlogwang 
308*2bfe3f2eSlogwang 	for (i = 0; i < num_nodes; i++)
309*2bfe3f2eSlogwang 		flush_rx_queue(i);
310*2bfe3f2eSlogwang }
311*2bfe3f2eSlogwang 
312*2bfe3f2eSlogwang /*
313*2bfe3f2eSlogwang  * Function called by the master lcore of the DPDK process.
314*2bfe3f2eSlogwang  */
315*2bfe3f2eSlogwang static void
316*2bfe3f2eSlogwang do_packet_forwarding(void)
317*2bfe3f2eSlogwang {
318*2bfe3f2eSlogwang 	unsigned int port_num = 0; /* indexes the port[] array */
319*2bfe3f2eSlogwang 	unsigned int socket_id = rte_socket_id();
320*2bfe3f2eSlogwang 
321*2bfe3f2eSlogwang 	for (;;) {
322*2bfe3f2eSlogwang 		struct rte_mbuf *buf[PACKET_READ_SIZE];
323*2bfe3f2eSlogwang 		uint16_t rx_count;
324*2bfe3f2eSlogwang 
325*2bfe3f2eSlogwang 		/* read a port */
326*2bfe3f2eSlogwang 		rx_count = rte_eth_rx_burst(info->id[port_num], 0,
327*2bfe3f2eSlogwang 				buf, PACKET_READ_SIZE);
328*2bfe3f2eSlogwang 		info->rx_stats.rx[port_num] += rx_count;
329*2bfe3f2eSlogwang 
330*2bfe3f2eSlogwang 		/* Now process the NIC packets read */
331*2bfe3f2eSlogwang 		if (likely(rx_count > 0))
332*2bfe3f2eSlogwang 			process_packets(port_num, buf, rx_count, socket_id);
333*2bfe3f2eSlogwang 
334*2bfe3f2eSlogwang 		/* move to next port */
335*2bfe3f2eSlogwang 		if (++port_num == info->num_ports)
336*2bfe3f2eSlogwang 			port_num = 0;
337*2bfe3f2eSlogwang 	}
338*2bfe3f2eSlogwang }
339*2bfe3f2eSlogwang 
340*2bfe3f2eSlogwang int
341*2bfe3f2eSlogwang main(int argc, char *argv[])
342*2bfe3f2eSlogwang {
343*2bfe3f2eSlogwang 	/* initialise the system */
344*2bfe3f2eSlogwang 	if (init(argc, argv) < 0)
345*2bfe3f2eSlogwang 		return -1;
346*2bfe3f2eSlogwang 	RTE_LOG(INFO, APP, "Finished Process Init.\n");
347*2bfe3f2eSlogwang 
348*2bfe3f2eSlogwang 	cl_rx_buf = calloc(num_nodes, sizeof(cl_rx_buf[0]));
349*2bfe3f2eSlogwang 
350*2bfe3f2eSlogwang 	/* clear statistics */
351*2bfe3f2eSlogwang 	clear_stats();
352*2bfe3f2eSlogwang 
353*2bfe3f2eSlogwang 	/* put all other cores to sleep bar master */
354*2bfe3f2eSlogwang 	rte_eal_mp_remote_launch(sleep_lcore, NULL, SKIP_MASTER);
355*2bfe3f2eSlogwang 
356*2bfe3f2eSlogwang 	do_packet_forwarding();
357*2bfe3f2eSlogwang 	return 0;
358*2bfe3f2eSlogwang }
359