xref: /f-stack/dpdk/drivers/net/axgbe/axgbe_dev.c (revision 2d9fd380)
1d30ea906Sjfb8856606 /*   SPDX-License-Identifier: BSD-3-Clause
2d30ea906Sjfb8856606  *   Copyright(c) 2018 Advanced Micro Devices, Inc. All rights reserved.
3d30ea906Sjfb8856606  *   Copyright(c) 2018 Synopsys, Inc. All rights reserved.
4d30ea906Sjfb8856606  */
5d30ea906Sjfb8856606 
6d30ea906Sjfb8856606 #include "axgbe_ethdev.h"
7d30ea906Sjfb8856606 #include "axgbe_common.h"
8d30ea906Sjfb8856606 #include "axgbe_phy.h"
9d30ea906Sjfb8856606 #include "axgbe_rxtx.h"
10d30ea906Sjfb8856606 
axgbe_get_max_frame(struct axgbe_port * pdata)11d30ea906Sjfb8856606 static inline unsigned int axgbe_get_max_frame(struct axgbe_port *pdata)
12d30ea906Sjfb8856606 {
134418919fSjohnjiang 	return pdata->eth_dev->data->mtu + RTE_ETHER_HDR_LEN +
144418919fSjohnjiang 		RTE_ETHER_CRC_LEN + VLAN_HLEN;
15d30ea906Sjfb8856606 }
16d30ea906Sjfb8856606 
17d30ea906Sjfb8856606 /* query busy bit */
mdio_complete(struct axgbe_port * pdata)18d30ea906Sjfb8856606 static int mdio_complete(struct axgbe_port *pdata)
19d30ea906Sjfb8856606 {
20d30ea906Sjfb8856606 	if (!AXGMAC_IOREAD_BITS(pdata, MAC_MDIOSCCDR, BUSY))
21d30ea906Sjfb8856606 		return 1;
22d30ea906Sjfb8856606 
23d30ea906Sjfb8856606 	return 0;
24d30ea906Sjfb8856606 }
25d30ea906Sjfb8856606 
axgbe_write_ext_mii_regs(struct axgbe_port * pdata,int addr,int reg,u16 val)26d30ea906Sjfb8856606 static int axgbe_write_ext_mii_regs(struct axgbe_port *pdata, int addr,
27d30ea906Sjfb8856606 				    int reg, u16 val)
28d30ea906Sjfb8856606 {
29d30ea906Sjfb8856606 	unsigned int mdio_sca, mdio_sccd;
30d30ea906Sjfb8856606 	uint64_t timeout;
31d30ea906Sjfb8856606 
32d30ea906Sjfb8856606 	mdio_sca = 0;
33d30ea906Sjfb8856606 	AXGMAC_SET_BITS(mdio_sca, MAC_MDIOSCAR, REG, reg);
34d30ea906Sjfb8856606 	AXGMAC_SET_BITS(mdio_sca, MAC_MDIOSCAR, DA, addr);
35d30ea906Sjfb8856606 	AXGMAC_IOWRITE(pdata, MAC_MDIOSCAR, mdio_sca);
36d30ea906Sjfb8856606 
37d30ea906Sjfb8856606 	mdio_sccd = 0;
38d30ea906Sjfb8856606 	AXGMAC_SET_BITS(mdio_sccd, MAC_MDIOSCCDR, DATA, val);
39d30ea906Sjfb8856606 	AXGMAC_SET_BITS(mdio_sccd, MAC_MDIOSCCDR, CMD, 1);
40d30ea906Sjfb8856606 	AXGMAC_SET_BITS(mdio_sccd, MAC_MDIOSCCDR, BUSY, 1);
41d30ea906Sjfb8856606 	AXGMAC_IOWRITE(pdata, MAC_MDIOSCCDR, mdio_sccd);
42d30ea906Sjfb8856606 
43d30ea906Sjfb8856606 	timeout = rte_get_timer_cycles() + rte_get_timer_hz();
44d30ea906Sjfb8856606 	while (time_before(rte_get_timer_cycles(), timeout)) {
45d30ea906Sjfb8856606 		rte_delay_us(100);
46d30ea906Sjfb8856606 		if (mdio_complete(pdata))
47d30ea906Sjfb8856606 			return 0;
48d30ea906Sjfb8856606 	}
49d30ea906Sjfb8856606 
50d30ea906Sjfb8856606 	PMD_DRV_LOG(ERR, "Mdio write operation timed out\n");
51d30ea906Sjfb8856606 	return -ETIMEDOUT;
52d30ea906Sjfb8856606 }
53d30ea906Sjfb8856606 
axgbe_read_ext_mii_regs(struct axgbe_port * pdata,int addr,int reg)54d30ea906Sjfb8856606 static int axgbe_read_ext_mii_regs(struct axgbe_port *pdata, int addr,
55d30ea906Sjfb8856606 				   int reg)
56d30ea906Sjfb8856606 {
57d30ea906Sjfb8856606 	unsigned int mdio_sca, mdio_sccd;
58d30ea906Sjfb8856606 	uint64_t timeout;
59d30ea906Sjfb8856606 
60d30ea906Sjfb8856606 	mdio_sca = 0;
61d30ea906Sjfb8856606 	AXGMAC_SET_BITS(mdio_sca, MAC_MDIOSCAR, REG, reg);
62d30ea906Sjfb8856606 	AXGMAC_SET_BITS(mdio_sca, MAC_MDIOSCAR, DA, addr);
63d30ea906Sjfb8856606 	AXGMAC_IOWRITE(pdata, MAC_MDIOSCAR, mdio_sca);
64d30ea906Sjfb8856606 
65d30ea906Sjfb8856606 	mdio_sccd = 0;
66d30ea906Sjfb8856606 	AXGMAC_SET_BITS(mdio_sccd, MAC_MDIOSCCDR, CMD, 3);
67d30ea906Sjfb8856606 	AXGMAC_SET_BITS(mdio_sccd, MAC_MDIOSCCDR, BUSY, 1);
68d30ea906Sjfb8856606 	AXGMAC_IOWRITE(pdata, MAC_MDIOSCCDR, mdio_sccd);
69d30ea906Sjfb8856606 
70d30ea906Sjfb8856606 	timeout = rte_get_timer_cycles() + rte_get_timer_hz();
71d30ea906Sjfb8856606 
72d30ea906Sjfb8856606 	while (time_before(rte_get_timer_cycles(), timeout)) {
73d30ea906Sjfb8856606 		rte_delay_us(100);
74d30ea906Sjfb8856606 		if (mdio_complete(pdata))
75d30ea906Sjfb8856606 			goto success;
76d30ea906Sjfb8856606 	}
77d30ea906Sjfb8856606 
78d30ea906Sjfb8856606 	PMD_DRV_LOG(ERR, "Mdio read operation timed out\n");
79d30ea906Sjfb8856606 	return -ETIMEDOUT;
80d30ea906Sjfb8856606 
81d30ea906Sjfb8856606 success:
82d30ea906Sjfb8856606 	return AXGMAC_IOREAD_BITS(pdata, MAC_MDIOSCCDR, DATA);
83d30ea906Sjfb8856606 }
84d30ea906Sjfb8856606 
axgbe_set_ext_mii_mode(struct axgbe_port * pdata,unsigned int port,enum axgbe_mdio_mode mode)85d30ea906Sjfb8856606 static int axgbe_set_ext_mii_mode(struct axgbe_port *pdata, unsigned int port,
86d30ea906Sjfb8856606 				  enum axgbe_mdio_mode mode)
87d30ea906Sjfb8856606 {
88d30ea906Sjfb8856606 	unsigned int reg_val = 0;
89d30ea906Sjfb8856606 
90d30ea906Sjfb8856606 	switch (mode) {
91d30ea906Sjfb8856606 	case AXGBE_MDIO_MODE_CL22:
92d30ea906Sjfb8856606 		if (port > AXGMAC_MAX_C22_PORT)
93d30ea906Sjfb8856606 			return -EINVAL;
94d30ea906Sjfb8856606 		reg_val |= (1 << port);
95d30ea906Sjfb8856606 		break;
96d30ea906Sjfb8856606 	case AXGBE_MDIO_MODE_CL45:
97d30ea906Sjfb8856606 		break;
98d30ea906Sjfb8856606 	default:
99d30ea906Sjfb8856606 		return -EINVAL;
100d30ea906Sjfb8856606 	}
101d30ea906Sjfb8856606 	AXGMAC_IOWRITE(pdata, MAC_MDIOCL22R, reg_val);
102d30ea906Sjfb8856606 
103d30ea906Sjfb8856606 	return 0;
104d30ea906Sjfb8856606 }
105d30ea906Sjfb8856606 
axgbe_read_mmd_regs_v2(struct axgbe_port * pdata,int prtad __rte_unused,int mmd_reg)106d30ea906Sjfb8856606 static int axgbe_read_mmd_regs_v2(struct axgbe_port *pdata,
107d30ea906Sjfb8856606 				  int prtad __rte_unused, int mmd_reg)
108d30ea906Sjfb8856606 {
109d30ea906Sjfb8856606 	unsigned int mmd_address, index, offset;
110d30ea906Sjfb8856606 	int mmd_data;
111d30ea906Sjfb8856606 
112d30ea906Sjfb8856606 	if (mmd_reg & MII_ADDR_C45)
113d30ea906Sjfb8856606 		mmd_address = mmd_reg & ~MII_ADDR_C45;
114d30ea906Sjfb8856606 	else
115d30ea906Sjfb8856606 		mmd_address = (pdata->mdio_mmd << 16) | (mmd_reg & 0xffff);
116d30ea906Sjfb8856606 
117d30ea906Sjfb8856606 	/* The PCS registers are accessed using mmio. The underlying
118d30ea906Sjfb8856606 	 * management interface uses indirect addressing to access the MMD
119d30ea906Sjfb8856606 	 * register sets. This requires accessing of the PCS register in two
120d30ea906Sjfb8856606 	 * phases, an address phase and a data phase.
121d30ea906Sjfb8856606 	 *
122d30ea906Sjfb8856606 	 * The mmio interface is based on 16-bit offsets and values. All
123d30ea906Sjfb8856606 	 * register offsets must therefore be adjusted by left shifting the
124d30ea906Sjfb8856606 	 * offset 1 bit and reading 16 bits of data.
125d30ea906Sjfb8856606 	 */
126d30ea906Sjfb8856606 	mmd_address <<= 1;
127d30ea906Sjfb8856606 	index = mmd_address & ~pdata->xpcs_window_mask;
128d30ea906Sjfb8856606 	offset = pdata->xpcs_window + (mmd_address & pdata->xpcs_window_mask);
129d30ea906Sjfb8856606 
130d30ea906Sjfb8856606 	pthread_mutex_lock(&pdata->xpcs_mutex);
131d30ea906Sjfb8856606 
132d30ea906Sjfb8856606 	XPCS32_IOWRITE(pdata, pdata->xpcs_window_sel_reg, index);
133d30ea906Sjfb8856606 	mmd_data = XPCS16_IOREAD(pdata, offset);
134d30ea906Sjfb8856606 
135d30ea906Sjfb8856606 	pthread_mutex_unlock(&pdata->xpcs_mutex);
136d30ea906Sjfb8856606 
137d30ea906Sjfb8856606 	return mmd_data;
138d30ea906Sjfb8856606 }
139d30ea906Sjfb8856606 
axgbe_write_mmd_regs_v2(struct axgbe_port * pdata,int prtad __rte_unused,int mmd_reg,int mmd_data)140d30ea906Sjfb8856606 static void axgbe_write_mmd_regs_v2(struct axgbe_port *pdata,
141d30ea906Sjfb8856606 				    int prtad __rte_unused,
142d30ea906Sjfb8856606 				    int mmd_reg, int mmd_data)
143d30ea906Sjfb8856606 {
144d30ea906Sjfb8856606 	unsigned int mmd_address, index, offset;
145d30ea906Sjfb8856606 
146d30ea906Sjfb8856606 	if (mmd_reg & MII_ADDR_C45)
147d30ea906Sjfb8856606 		mmd_address = mmd_reg & ~MII_ADDR_C45;
148d30ea906Sjfb8856606 	else
149d30ea906Sjfb8856606 		mmd_address = (pdata->mdio_mmd << 16) | (mmd_reg & 0xffff);
150d30ea906Sjfb8856606 
151d30ea906Sjfb8856606 	/* The PCS registers are accessed using mmio. The underlying
152d30ea906Sjfb8856606 	 * management interface uses indirect addressing to access the MMD
153d30ea906Sjfb8856606 	 * register sets. This requires accessing of the PCS register in two
154d30ea906Sjfb8856606 	 * phases, an address phase and a data phase.
155d30ea906Sjfb8856606 	 *
156d30ea906Sjfb8856606 	 * The mmio interface is based on 16-bit offsets and values. All
157d30ea906Sjfb8856606 	 * register offsets must therefore be adjusted by left shifting the
158d30ea906Sjfb8856606 	 * offset 1 bit and writing 16 bits of data.
159d30ea906Sjfb8856606 	 */
160d30ea906Sjfb8856606 	mmd_address <<= 1;
161d30ea906Sjfb8856606 	index = mmd_address & ~pdata->xpcs_window_mask;
162d30ea906Sjfb8856606 	offset = pdata->xpcs_window + (mmd_address & pdata->xpcs_window_mask);
163d30ea906Sjfb8856606 
164d30ea906Sjfb8856606 	pthread_mutex_lock(&pdata->xpcs_mutex);
165d30ea906Sjfb8856606 
166d30ea906Sjfb8856606 	XPCS32_IOWRITE(pdata, pdata->xpcs_window_sel_reg, index);
167d30ea906Sjfb8856606 	XPCS16_IOWRITE(pdata, offset, mmd_data);
168d30ea906Sjfb8856606 
169d30ea906Sjfb8856606 	pthread_mutex_unlock(&pdata->xpcs_mutex);
170d30ea906Sjfb8856606 }
171d30ea906Sjfb8856606 
axgbe_read_mmd_regs(struct axgbe_port * pdata,int prtad,int mmd_reg)172d30ea906Sjfb8856606 static int axgbe_read_mmd_regs(struct axgbe_port *pdata, int prtad,
173d30ea906Sjfb8856606 			       int mmd_reg)
174d30ea906Sjfb8856606 {
175d30ea906Sjfb8856606 	switch (pdata->vdata->xpcs_access) {
176d30ea906Sjfb8856606 	case AXGBE_XPCS_ACCESS_V1:
177d30ea906Sjfb8856606 		PMD_DRV_LOG(ERR, "PHY_Version 1 is not supported\n");
178d30ea906Sjfb8856606 		return -1;
179d30ea906Sjfb8856606 	case AXGBE_XPCS_ACCESS_V2:
180d30ea906Sjfb8856606 	default:
181d30ea906Sjfb8856606 		return axgbe_read_mmd_regs_v2(pdata, prtad, mmd_reg);
182d30ea906Sjfb8856606 	}
183d30ea906Sjfb8856606 }
184d30ea906Sjfb8856606 
axgbe_write_mmd_regs(struct axgbe_port * pdata,int prtad,int mmd_reg,int mmd_data)185d30ea906Sjfb8856606 static void axgbe_write_mmd_regs(struct axgbe_port *pdata, int prtad,
186d30ea906Sjfb8856606 				 int mmd_reg, int mmd_data)
187d30ea906Sjfb8856606 {
188d30ea906Sjfb8856606 	switch (pdata->vdata->xpcs_access) {
189d30ea906Sjfb8856606 	case AXGBE_XPCS_ACCESS_V1:
190d30ea906Sjfb8856606 		PMD_DRV_LOG(ERR, "PHY_Version 1 is not supported\n");
191d30ea906Sjfb8856606 		return;
192d30ea906Sjfb8856606 	case AXGBE_XPCS_ACCESS_V2:
193d30ea906Sjfb8856606 	default:
194d30ea906Sjfb8856606 		return axgbe_write_mmd_regs_v2(pdata, prtad, mmd_reg, mmd_data);
195d30ea906Sjfb8856606 	}
196d30ea906Sjfb8856606 }
197d30ea906Sjfb8856606 
axgbe_set_speed(struct axgbe_port * pdata,int speed)198d30ea906Sjfb8856606 static int axgbe_set_speed(struct axgbe_port *pdata, int speed)
199d30ea906Sjfb8856606 {
200d30ea906Sjfb8856606 	unsigned int ss;
201d30ea906Sjfb8856606 
202d30ea906Sjfb8856606 	switch (speed) {
203d30ea906Sjfb8856606 	case SPEED_1000:
204d30ea906Sjfb8856606 		ss = 0x03;
205d30ea906Sjfb8856606 		break;
206d30ea906Sjfb8856606 	case SPEED_2500:
207d30ea906Sjfb8856606 		ss = 0x02;
208d30ea906Sjfb8856606 		break;
209d30ea906Sjfb8856606 	case SPEED_10000:
210d30ea906Sjfb8856606 		ss = 0x00;
211d30ea906Sjfb8856606 		break;
212d30ea906Sjfb8856606 	default:
213d30ea906Sjfb8856606 		return -EINVAL;
214d30ea906Sjfb8856606 	}
215d30ea906Sjfb8856606 
216d30ea906Sjfb8856606 	if (AXGMAC_IOREAD_BITS(pdata, MAC_TCR, SS) != ss)
217d30ea906Sjfb8856606 		AXGMAC_IOWRITE_BITS(pdata, MAC_TCR, SS, ss);
218d30ea906Sjfb8856606 
219d30ea906Sjfb8856606 	return 0;
220d30ea906Sjfb8856606 }
221d30ea906Sjfb8856606 
axgbe_disable_tx_flow_control(struct axgbe_port * pdata)222d30ea906Sjfb8856606 static int axgbe_disable_tx_flow_control(struct axgbe_port *pdata)
223d30ea906Sjfb8856606 {
224d30ea906Sjfb8856606 	unsigned int max_q_count, q_count;
225d30ea906Sjfb8856606 	unsigned int reg, reg_val;
226d30ea906Sjfb8856606 	unsigned int i;
227d30ea906Sjfb8856606 
228d30ea906Sjfb8856606 	/* Clear MTL flow control */
229d30ea906Sjfb8856606 	for (i = 0; i < pdata->rx_q_count; i++)
230d30ea906Sjfb8856606 		AXGMAC_MTL_IOWRITE_BITS(pdata, i, MTL_Q_RQOMR, EHFC, 0);
231d30ea906Sjfb8856606 
232d30ea906Sjfb8856606 	/* Clear MAC flow control */
233d30ea906Sjfb8856606 	max_q_count = AXGMAC_MAX_FLOW_CONTROL_QUEUES;
234d30ea906Sjfb8856606 	q_count = RTE_MIN(pdata->tx_q_count,
235d30ea906Sjfb8856606 			max_q_count);
236d30ea906Sjfb8856606 	reg = MAC_Q0TFCR;
237d30ea906Sjfb8856606 	for (i = 0; i < q_count; i++) {
238d30ea906Sjfb8856606 		reg_val = AXGMAC_IOREAD(pdata, reg);
239d30ea906Sjfb8856606 		AXGMAC_SET_BITS(reg_val, MAC_Q0TFCR, TFE, 0);
240d30ea906Sjfb8856606 		AXGMAC_IOWRITE(pdata, reg, reg_val);
241d30ea906Sjfb8856606 
242d30ea906Sjfb8856606 		reg += MAC_QTFCR_INC;
243d30ea906Sjfb8856606 	}
244d30ea906Sjfb8856606 
245d30ea906Sjfb8856606 	return 0;
246d30ea906Sjfb8856606 }
247d30ea906Sjfb8856606 
axgbe_enable_tx_flow_control(struct axgbe_port * pdata)248d30ea906Sjfb8856606 static int axgbe_enable_tx_flow_control(struct axgbe_port *pdata)
249d30ea906Sjfb8856606 {
250d30ea906Sjfb8856606 	unsigned int max_q_count, q_count;
251d30ea906Sjfb8856606 	unsigned int reg, reg_val;
252d30ea906Sjfb8856606 	unsigned int i;
253d30ea906Sjfb8856606 
254d30ea906Sjfb8856606 	/* Set MTL flow control */
255d30ea906Sjfb8856606 	for (i = 0; i < pdata->rx_q_count; i++) {
256d30ea906Sjfb8856606 		unsigned int ehfc = 0;
257d30ea906Sjfb8856606 
258d30ea906Sjfb8856606 		/* Flow control thresholds are established */
259d30ea906Sjfb8856606 		if (pdata->rx_rfd[i])
260d30ea906Sjfb8856606 			ehfc = 1;
261d30ea906Sjfb8856606 
262d30ea906Sjfb8856606 		AXGMAC_MTL_IOWRITE_BITS(pdata, i, MTL_Q_RQOMR, EHFC, ehfc);
263*2d9fd380Sjfb8856606 
264*2d9fd380Sjfb8856606 		PMD_DRV_LOG(DEBUG, "flow control %s for RXq%u\n",
265*2d9fd380Sjfb8856606 			    ehfc ? "enabled" : "disabled", i);
266d30ea906Sjfb8856606 	}
267d30ea906Sjfb8856606 
268d30ea906Sjfb8856606 	/* Set MAC flow control */
269d30ea906Sjfb8856606 	max_q_count = AXGMAC_MAX_FLOW_CONTROL_QUEUES;
270d30ea906Sjfb8856606 	q_count = RTE_MIN(pdata->tx_q_count,
271d30ea906Sjfb8856606 			max_q_count);
272d30ea906Sjfb8856606 	reg = MAC_Q0TFCR;
273d30ea906Sjfb8856606 	for (i = 0; i < q_count; i++) {
274d30ea906Sjfb8856606 		reg_val = AXGMAC_IOREAD(pdata, reg);
275d30ea906Sjfb8856606 
276d30ea906Sjfb8856606 		/* Enable transmit flow control */
277d30ea906Sjfb8856606 		AXGMAC_SET_BITS(reg_val, MAC_Q0TFCR, TFE, 1);
278d30ea906Sjfb8856606 		/* Set pause time */
279d30ea906Sjfb8856606 		AXGMAC_SET_BITS(reg_val, MAC_Q0TFCR, PT, 0xffff);
280d30ea906Sjfb8856606 
281d30ea906Sjfb8856606 		AXGMAC_IOWRITE(pdata, reg, reg_val);
282d30ea906Sjfb8856606 
283d30ea906Sjfb8856606 		reg += MAC_QTFCR_INC;
284d30ea906Sjfb8856606 	}
285d30ea906Sjfb8856606 
286d30ea906Sjfb8856606 	return 0;
287d30ea906Sjfb8856606 }
288d30ea906Sjfb8856606 
axgbe_disable_rx_flow_control(struct axgbe_port * pdata)289d30ea906Sjfb8856606 static int axgbe_disable_rx_flow_control(struct axgbe_port *pdata)
290d30ea906Sjfb8856606 {
291d30ea906Sjfb8856606 	AXGMAC_IOWRITE_BITS(pdata, MAC_RFCR, RFE, 0);
292d30ea906Sjfb8856606 
293d30ea906Sjfb8856606 	return 0;
294d30ea906Sjfb8856606 }
295d30ea906Sjfb8856606 
axgbe_enable_rx_flow_control(struct axgbe_port * pdata)296d30ea906Sjfb8856606 static int axgbe_enable_rx_flow_control(struct axgbe_port *pdata)
297d30ea906Sjfb8856606 {
298d30ea906Sjfb8856606 	AXGMAC_IOWRITE_BITS(pdata, MAC_RFCR, RFE, 1);
299d30ea906Sjfb8856606 
300d30ea906Sjfb8856606 	return 0;
301d30ea906Sjfb8856606 }
302d30ea906Sjfb8856606 
axgbe_config_tx_flow_control(struct axgbe_port * pdata)303d30ea906Sjfb8856606 static int axgbe_config_tx_flow_control(struct axgbe_port *pdata)
304d30ea906Sjfb8856606 {
305d30ea906Sjfb8856606 	if (pdata->tx_pause)
306d30ea906Sjfb8856606 		axgbe_enable_tx_flow_control(pdata);
307d30ea906Sjfb8856606 	else
308d30ea906Sjfb8856606 		axgbe_disable_tx_flow_control(pdata);
309d30ea906Sjfb8856606 
310d30ea906Sjfb8856606 	return 0;
311d30ea906Sjfb8856606 }
312d30ea906Sjfb8856606 
axgbe_config_rx_flow_control(struct axgbe_port * pdata)313d30ea906Sjfb8856606 static int axgbe_config_rx_flow_control(struct axgbe_port *pdata)
314d30ea906Sjfb8856606 {
315d30ea906Sjfb8856606 	if (pdata->rx_pause)
316d30ea906Sjfb8856606 		axgbe_enable_rx_flow_control(pdata);
317d30ea906Sjfb8856606 	else
318d30ea906Sjfb8856606 		axgbe_disable_rx_flow_control(pdata);
319d30ea906Sjfb8856606 
320d30ea906Sjfb8856606 	return 0;
321d30ea906Sjfb8856606 }
322d30ea906Sjfb8856606 
axgbe_config_flow_control(struct axgbe_port * pdata)323d30ea906Sjfb8856606 static void axgbe_config_flow_control(struct axgbe_port *pdata)
324d30ea906Sjfb8856606 {
325d30ea906Sjfb8856606 	axgbe_config_tx_flow_control(pdata);
326d30ea906Sjfb8856606 	axgbe_config_rx_flow_control(pdata);
327d30ea906Sjfb8856606 
328d30ea906Sjfb8856606 	AXGMAC_IOWRITE_BITS(pdata, MAC_RFCR, PFCE, 0);
329d30ea906Sjfb8856606 }
330d30ea906Sjfb8856606 
axgbe_queue_flow_control_threshold(struct axgbe_port * pdata,unsigned int queue,unsigned int q_fifo_size)331d30ea906Sjfb8856606 static void axgbe_queue_flow_control_threshold(struct axgbe_port *pdata,
332d30ea906Sjfb8856606 					       unsigned int queue,
333d30ea906Sjfb8856606 					       unsigned int q_fifo_size)
334d30ea906Sjfb8856606 {
335d30ea906Sjfb8856606 	unsigned int frame_fifo_size;
336d30ea906Sjfb8856606 	unsigned int rfa, rfd;
337d30ea906Sjfb8856606 
338d30ea906Sjfb8856606 	frame_fifo_size = AXGMAC_FLOW_CONTROL_ALIGN(axgbe_get_max_frame(pdata));
339d30ea906Sjfb8856606 
340d30ea906Sjfb8856606 	/* This path deals with just maximum frame sizes which are
341d30ea906Sjfb8856606 	 * limited to a jumbo frame of 9,000 (plus headers, etc.)
342d30ea906Sjfb8856606 	 * so we can never exceed the maximum allowable RFA/RFD
343d30ea906Sjfb8856606 	 * values.
344d30ea906Sjfb8856606 	 */
345d30ea906Sjfb8856606 	if (q_fifo_size <= 2048) {
346d30ea906Sjfb8856606 		/* rx_rfd to zero to signal no flow control */
347d30ea906Sjfb8856606 		pdata->rx_rfa[queue] = 0;
348d30ea906Sjfb8856606 		pdata->rx_rfd[queue] = 0;
349d30ea906Sjfb8856606 		return;
350d30ea906Sjfb8856606 	}
351d30ea906Sjfb8856606 
352d30ea906Sjfb8856606 	if (q_fifo_size <= 4096) {
353d30ea906Sjfb8856606 		/* Between 2048 and 4096 */
354d30ea906Sjfb8856606 		pdata->rx_rfa[queue] = 0;	/* Full - 1024 bytes */
355d30ea906Sjfb8856606 		pdata->rx_rfd[queue] = 1;	/* Full - 1536 bytes */
356d30ea906Sjfb8856606 		return;
357d30ea906Sjfb8856606 	}
358d30ea906Sjfb8856606 
359d30ea906Sjfb8856606 	if (q_fifo_size <= frame_fifo_size) {
360d30ea906Sjfb8856606 		/* Between 4096 and max-frame */
361d30ea906Sjfb8856606 		pdata->rx_rfa[queue] = 2;	/* Full - 2048 bytes */
362d30ea906Sjfb8856606 		pdata->rx_rfd[queue] = 5;	/* Full - 3584 bytes */
363d30ea906Sjfb8856606 		return;
364d30ea906Sjfb8856606 	}
365d30ea906Sjfb8856606 
366d30ea906Sjfb8856606 	if (q_fifo_size <= (frame_fifo_size * 3)) {
367d30ea906Sjfb8856606 		/* Between max-frame and 3 max-frames,
368d30ea906Sjfb8856606 		 * trigger if we get just over a frame of data and
369d30ea906Sjfb8856606 		 * resume when we have just under half a frame left.
370d30ea906Sjfb8856606 		 */
371d30ea906Sjfb8856606 		rfa = q_fifo_size - frame_fifo_size;
372d30ea906Sjfb8856606 		rfd = rfa + (frame_fifo_size / 2);
373d30ea906Sjfb8856606 	} else {
374d30ea906Sjfb8856606 		/* Above 3 max-frames - trigger when just over
375d30ea906Sjfb8856606 		 * 2 frames of space available
376d30ea906Sjfb8856606 		 */
377d30ea906Sjfb8856606 		rfa = frame_fifo_size * 2;
378d30ea906Sjfb8856606 		rfa += AXGMAC_FLOW_CONTROL_UNIT;
379d30ea906Sjfb8856606 		rfd = rfa + frame_fifo_size;
380d30ea906Sjfb8856606 	}
381d30ea906Sjfb8856606 
382d30ea906Sjfb8856606 	pdata->rx_rfa[queue] = AXGMAC_FLOW_CONTROL_VALUE(rfa);
383d30ea906Sjfb8856606 	pdata->rx_rfd[queue] = AXGMAC_FLOW_CONTROL_VALUE(rfd);
384d30ea906Sjfb8856606 }
385d30ea906Sjfb8856606 
axgbe_calculate_flow_control_threshold(struct axgbe_port * pdata)386d30ea906Sjfb8856606 static void axgbe_calculate_flow_control_threshold(struct axgbe_port *pdata)
387d30ea906Sjfb8856606 {
388d30ea906Sjfb8856606 	unsigned int q_fifo_size;
389d30ea906Sjfb8856606 	unsigned int i;
390d30ea906Sjfb8856606 
391d30ea906Sjfb8856606 	for (i = 0; i < pdata->rx_q_count; i++) {
392d30ea906Sjfb8856606 		q_fifo_size = (pdata->fifo + 1) * AXGMAC_FIFO_UNIT;
393d30ea906Sjfb8856606 
394d30ea906Sjfb8856606 		axgbe_queue_flow_control_threshold(pdata, i, q_fifo_size);
395d30ea906Sjfb8856606 	}
396d30ea906Sjfb8856606 }
397d30ea906Sjfb8856606 
axgbe_config_flow_control_threshold(struct axgbe_port * pdata)398d30ea906Sjfb8856606 static void axgbe_config_flow_control_threshold(struct axgbe_port *pdata)
399d30ea906Sjfb8856606 {
400d30ea906Sjfb8856606 	unsigned int i;
401d30ea906Sjfb8856606 
402d30ea906Sjfb8856606 	for (i = 0; i < pdata->rx_q_count; i++) {
403d30ea906Sjfb8856606 		AXGMAC_MTL_IOWRITE_BITS(pdata, i, MTL_Q_RQFCR, RFA,
404d30ea906Sjfb8856606 					pdata->rx_rfa[i]);
405d30ea906Sjfb8856606 		AXGMAC_MTL_IOWRITE_BITS(pdata, i, MTL_Q_RQFCR, RFD,
406d30ea906Sjfb8856606 					pdata->rx_rfd[i]);
407d30ea906Sjfb8856606 	}
408d30ea906Sjfb8856606 }
409d30ea906Sjfb8856606 
__axgbe_exit(struct axgbe_port * pdata)410d30ea906Sjfb8856606 static int __axgbe_exit(struct axgbe_port *pdata)
411d30ea906Sjfb8856606 {
412d30ea906Sjfb8856606 	unsigned int count = 2000;
413d30ea906Sjfb8856606 
414d30ea906Sjfb8856606 	/* Issue a software reset */
415d30ea906Sjfb8856606 	AXGMAC_IOWRITE_BITS(pdata, DMA_MR, SWR, 1);
416d30ea906Sjfb8856606 	rte_delay_us(10);
417d30ea906Sjfb8856606 
418d30ea906Sjfb8856606 	/* Poll Until Poll Condition */
419d30ea906Sjfb8856606 	while (--count && AXGMAC_IOREAD_BITS(pdata, DMA_MR, SWR))
420d30ea906Sjfb8856606 		rte_delay_us(500);
421d30ea906Sjfb8856606 
422d30ea906Sjfb8856606 	if (!count)
423d30ea906Sjfb8856606 		return -EBUSY;
424d30ea906Sjfb8856606 
425d30ea906Sjfb8856606 	return 0;
426d30ea906Sjfb8856606 }
427d30ea906Sjfb8856606 
axgbe_exit(struct axgbe_port * pdata)428d30ea906Sjfb8856606 static int axgbe_exit(struct axgbe_port *pdata)
429d30ea906Sjfb8856606 {
430d30ea906Sjfb8856606 	int ret;
431d30ea906Sjfb8856606 
432d30ea906Sjfb8856606 	/* To guard against possible incorrectly generated interrupts,
433d30ea906Sjfb8856606 	 * issue the software reset twice.
434d30ea906Sjfb8856606 	 */
435d30ea906Sjfb8856606 	ret = __axgbe_exit(pdata);
436d30ea906Sjfb8856606 	if (ret)
437d30ea906Sjfb8856606 		return ret;
438d30ea906Sjfb8856606 
439d30ea906Sjfb8856606 	return __axgbe_exit(pdata);
440d30ea906Sjfb8856606 }
441d30ea906Sjfb8856606 
axgbe_flush_tx_queues(struct axgbe_port * pdata)442d30ea906Sjfb8856606 static int axgbe_flush_tx_queues(struct axgbe_port *pdata)
443d30ea906Sjfb8856606 {
444d30ea906Sjfb8856606 	unsigned int i, count;
445d30ea906Sjfb8856606 
446d30ea906Sjfb8856606 	if (AXGMAC_GET_BITS(pdata->hw_feat.version, MAC_VR, SNPSVER) < 0x21)
447d30ea906Sjfb8856606 		return 0;
448d30ea906Sjfb8856606 
449d30ea906Sjfb8856606 	for (i = 0; i < pdata->tx_q_count; i++)
450d30ea906Sjfb8856606 		AXGMAC_MTL_IOWRITE_BITS(pdata, i, MTL_Q_TQOMR, FTQ, 1);
451d30ea906Sjfb8856606 
452d30ea906Sjfb8856606 	/* Poll Until Poll Condition */
453d30ea906Sjfb8856606 	for (i = 0; i < pdata->tx_q_count; i++) {
454d30ea906Sjfb8856606 		count = 2000;
455d30ea906Sjfb8856606 		while (--count && AXGMAC_MTL_IOREAD_BITS(pdata, i,
456d30ea906Sjfb8856606 							 MTL_Q_TQOMR, FTQ))
457d30ea906Sjfb8856606 			rte_delay_us(500);
458d30ea906Sjfb8856606 
459d30ea906Sjfb8856606 		if (!count)
460d30ea906Sjfb8856606 			return -EBUSY;
461d30ea906Sjfb8856606 	}
462d30ea906Sjfb8856606 
463d30ea906Sjfb8856606 	return 0;
464d30ea906Sjfb8856606 }
465d30ea906Sjfb8856606 
axgbe_config_dma_bus(struct axgbe_port * pdata)466d30ea906Sjfb8856606 static void axgbe_config_dma_bus(struct axgbe_port *pdata)
467d30ea906Sjfb8856606 {
468d30ea906Sjfb8856606 	/* Set enhanced addressing mode */
469d30ea906Sjfb8856606 	AXGMAC_IOWRITE_BITS(pdata, DMA_SBMR, EAME, 1);
470d30ea906Sjfb8856606 
471d30ea906Sjfb8856606 	/* Out standing read/write requests*/
472d30ea906Sjfb8856606 	AXGMAC_IOWRITE_BITS(pdata, DMA_SBMR, RD_OSR, 0x3f);
473d30ea906Sjfb8856606 	AXGMAC_IOWRITE_BITS(pdata, DMA_SBMR, WR_OSR, 0x3f);
474d30ea906Sjfb8856606 
475d30ea906Sjfb8856606 	/* Set the System Bus mode */
476d30ea906Sjfb8856606 	AXGMAC_IOWRITE_BITS(pdata, DMA_SBMR, UNDEF, 1);
477d30ea906Sjfb8856606 	AXGMAC_IOWRITE_BITS(pdata, DMA_SBMR, BLEN_32, 1);
478d30ea906Sjfb8856606 	AXGMAC_IOWRITE_BITS(pdata, DMA_SBMR, AAL, 1);
479d30ea906Sjfb8856606 }
480d30ea906Sjfb8856606 
axgbe_config_dma_cache(struct axgbe_port * pdata)481d30ea906Sjfb8856606 static void axgbe_config_dma_cache(struct axgbe_port *pdata)
482d30ea906Sjfb8856606 {
483d30ea906Sjfb8856606 	unsigned int arcache, awcache, arwcache;
484d30ea906Sjfb8856606 
485d30ea906Sjfb8856606 	arcache = 0;
486d30ea906Sjfb8856606 	AXGMAC_SET_BITS(arcache, DMA_AXIARCR, DRC, 0x3);
487d30ea906Sjfb8856606 	AXGMAC_IOWRITE(pdata, DMA_AXIARCR, arcache);
488d30ea906Sjfb8856606 
489d30ea906Sjfb8856606 	awcache = 0;
490d30ea906Sjfb8856606 	AXGMAC_SET_BITS(awcache, DMA_AXIAWCR, DWC, 0x3);
491d30ea906Sjfb8856606 	AXGMAC_SET_BITS(awcache, DMA_AXIAWCR, RPC, 0x3);
492d30ea906Sjfb8856606 	AXGMAC_SET_BITS(awcache, DMA_AXIAWCR, RPD, 0x1);
493d30ea906Sjfb8856606 	AXGMAC_SET_BITS(awcache, DMA_AXIAWCR, RHC, 0x3);
494d30ea906Sjfb8856606 	AXGMAC_SET_BITS(awcache, DMA_AXIAWCR, RHD, 0x1);
495d30ea906Sjfb8856606 	AXGMAC_SET_BITS(awcache, DMA_AXIAWCR, RDC, 0x3);
496d30ea906Sjfb8856606 	AXGMAC_SET_BITS(awcache, DMA_AXIAWCR, RDD, 0x1);
497d30ea906Sjfb8856606 	AXGMAC_IOWRITE(pdata, DMA_AXIAWCR, awcache);
498d30ea906Sjfb8856606 
499d30ea906Sjfb8856606 	arwcache = 0;
500d30ea906Sjfb8856606 	AXGMAC_SET_BITS(arwcache, DMA_AXIAWRCR, TDWD, 0x1);
501d30ea906Sjfb8856606 	AXGMAC_SET_BITS(arwcache, DMA_AXIAWRCR, TDWC, 0x3);
502d30ea906Sjfb8856606 	AXGMAC_SET_BITS(arwcache, DMA_AXIAWRCR, RDRC, 0x3);
503d30ea906Sjfb8856606 	AXGMAC_IOWRITE(pdata, DMA_AXIAWRCR, arwcache);
504d30ea906Sjfb8856606 }
505d30ea906Sjfb8856606 
axgbe_config_edma_control(struct axgbe_port * pdata)506d30ea906Sjfb8856606 static void axgbe_config_edma_control(struct axgbe_port *pdata)
507d30ea906Sjfb8856606 {
508d30ea906Sjfb8856606 	AXGMAC_IOWRITE(pdata, EDMA_TX_CONTROL, 0x5);
509d30ea906Sjfb8856606 	AXGMAC_IOWRITE(pdata, EDMA_RX_CONTROL, 0x5);
510d30ea906Sjfb8856606 }
511d30ea906Sjfb8856606 
axgbe_config_osp_mode(struct axgbe_port * pdata)512d30ea906Sjfb8856606 static int axgbe_config_osp_mode(struct axgbe_port *pdata)
513d30ea906Sjfb8856606 {
514d30ea906Sjfb8856606 	/* Force DMA to operate on second packet before closing descriptors
515d30ea906Sjfb8856606 	 *  of first packet
516d30ea906Sjfb8856606 	 */
517d30ea906Sjfb8856606 	struct axgbe_tx_queue *txq;
518d30ea906Sjfb8856606 	unsigned int i;
519d30ea906Sjfb8856606 
520d30ea906Sjfb8856606 	for (i = 0; i < pdata->eth_dev->data->nb_tx_queues; i++) {
521d30ea906Sjfb8856606 		txq = pdata->eth_dev->data->tx_queues[i];
522d30ea906Sjfb8856606 		AXGMAC_DMA_IOWRITE_BITS(txq, DMA_CH_TCR, OSP,
523d30ea906Sjfb8856606 					pdata->tx_osp_mode);
524d30ea906Sjfb8856606 	}
525d30ea906Sjfb8856606 
526d30ea906Sjfb8856606 	return 0;
527d30ea906Sjfb8856606 }
528d30ea906Sjfb8856606 
axgbe_config_pblx8(struct axgbe_port * pdata)529d30ea906Sjfb8856606 static int axgbe_config_pblx8(struct axgbe_port *pdata)
530d30ea906Sjfb8856606 {
531d30ea906Sjfb8856606 	struct axgbe_tx_queue *txq;
532d30ea906Sjfb8856606 	unsigned int i;
533d30ea906Sjfb8856606 
534d30ea906Sjfb8856606 	for (i = 0; i < pdata->eth_dev->data->nb_tx_queues; i++) {
535d30ea906Sjfb8856606 		txq = pdata->eth_dev->data->tx_queues[i];
536d30ea906Sjfb8856606 		AXGMAC_DMA_IOWRITE_BITS(txq, DMA_CH_CR, PBLX8,
537d30ea906Sjfb8856606 					pdata->pblx8);
538d30ea906Sjfb8856606 	}
539d30ea906Sjfb8856606 	return 0;
540d30ea906Sjfb8856606 }
541d30ea906Sjfb8856606 
axgbe_config_tx_pbl_val(struct axgbe_port * pdata)542d30ea906Sjfb8856606 static int axgbe_config_tx_pbl_val(struct axgbe_port *pdata)
543d30ea906Sjfb8856606 {
544d30ea906Sjfb8856606 	struct axgbe_tx_queue *txq;
545d30ea906Sjfb8856606 	unsigned int i;
546d30ea906Sjfb8856606 
547d30ea906Sjfb8856606 	for (i = 0; i < pdata->eth_dev->data->nb_tx_queues; i++) {
548d30ea906Sjfb8856606 		txq = pdata->eth_dev->data->tx_queues[i];
549d30ea906Sjfb8856606 		AXGMAC_DMA_IOWRITE_BITS(txq, DMA_CH_TCR, PBL,
550d30ea906Sjfb8856606 				pdata->tx_pbl);
551d30ea906Sjfb8856606 	}
552d30ea906Sjfb8856606 
553d30ea906Sjfb8856606 	return 0;
554d30ea906Sjfb8856606 }
555d30ea906Sjfb8856606 
axgbe_config_rx_pbl_val(struct axgbe_port * pdata)556d30ea906Sjfb8856606 static int axgbe_config_rx_pbl_val(struct axgbe_port *pdata)
557d30ea906Sjfb8856606 {
558d30ea906Sjfb8856606 	struct axgbe_rx_queue *rxq;
559d30ea906Sjfb8856606 	unsigned int i;
560d30ea906Sjfb8856606 
561d30ea906Sjfb8856606 	for (i = 0; i < pdata->eth_dev->data->nb_rx_queues; i++) {
562d30ea906Sjfb8856606 		rxq = pdata->eth_dev->data->rx_queues[i];
563d30ea906Sjfb8856606 		AXGMAC_DMA_IOWRITE_BITS(rxq, DMA_CH_RCR, PBL,
564d30ea906Sjfb8856606 				pdata->rx_pbl);
565d30ea906Sjfb8856606 	}
566d30ea906Sjfb8856606 
567d30ea906Sjfb8856606 	return 0;
568d30ea906Sjfb8856606 }
569d30ea906Sjfb8856606 
axgbe_config_rx_buffer_size(struct axgbe_port * pdata)570d30ea906Sjfb8856606 static void axgbe_config_rx_buffer_size(struct axgbe_port *pdata)
571d30ea906Sjfb8856606 {
572d30ea906Sjfb8856606 	struct axgbe_rx_queue *rxq;
573d30ea906Sjfb8856606 	unsigned int i;
574d30ea906Sjfb8856606 
575d30ea906Sjfb8856606 	for (i = 0; i < pdata->eth_dev->data->nb_rx_queues; i++) {
576d30ea906Sjfb8856606 		rxq = pdata->eth_dev->data->rx_queues[i];
577d30ea906Sjfb8856606 
578d30ea906Sjfb8856606 		rxq->buf_size = rte_pktmbuf_data_room_size(rxq->mb_pool) -
579d30ea906Sjfb8856606 			RTE_PKTMBUF_HEADROOM;
580d30ea906Sjfb8856606 		rxq->buf_size = (rxq->buf_size + AXGBE_RX_BUF_ALIGN - 1) &
581d30ea906Sjfb8856606 			~(AXGBE_RX_BUF_ALIGN - 1);
582d30ea906Sjfb8856606 
583d30ea906Sjfb8856606 		if (rxq->buf_size > pdata->rx_buf_size)
584d30ea906Sjfb8856606 			pdata->rx_buf_size = rxq->buf_size;
585d30ea906Sjfb8856606 
586d30ea906Sjfb8856606 		AXGMAC_DMA_IOWRITE_BITS(rxq, DMA_CH_RCR, RBSZ,
587d30ea906Sjfb8856606 					rxq->buf_size);
588d30ea906Sjfb8856606 	}
589d30ea906Sjfb8856606 }
590d30ea906Sjfb8856606 
axgbe_write_rss_reg(struct axgbe_port * pdata,unsigned int type,unsigned int index,unsigned int val)591d30ea906Sjfb8856606 static int axgbe_write_rss_reg(struct axgbe_port *pdata, unsigned int type,
592d30ea906Sjfb8856606 			       unsigned int index, unsigned int val)
593d30ea906Sjfb8856606 {
594d30ea906Sjfb8856606 	unsigned int wait;
595d30ea906Sjfb8856606 
596d30ea906Sjfb8856606 	if (AXGMAC_IOREAD_BITS(pdata, MAC_RSSAR, OB))
597d30ea906Sjfb8856606 		return -EBUSY;
598d30ea906Sjfb8856606 
599d30ea906Sjfb8856606 	AXGMAC_IOWRITE(pdata, MAC_RSSDR, val);
600d30ea906Sjfb8856606 
601d30ea906Sjfb8856606 	AXGMAC_IOWRITE_BITS(pdata, MAC_RSSAR, RSSIA, index);
602d30ea906Sjfb8856606 	AXGMAC_IOWRITE_BITS(pdata, MAC_RSSAR, ADDRT, type);
603d30ea906Sjfb8856606 	AXGMAC_IOWRITE_BITS(pdata, MAC_RSSAR, CT, 0);
604d30ea906Sjfb8856606 	AXGMAC_IOWRITE_BITS(pdata, MAC_RSSAR, OB, 1);
605d30ea906Sjfb8856606 
606d30ea906Sjfb8856606 	wait = 1000;
607d30ea906Sjfb8856606 	while (wait--) {
608d30ea906Sjfb8856606 		if (!AXGMAC_IOREAD_BITS(pdata, MAC_RSSAR, OB))
609d30ea906Sjfb8856606 			return 0;
610d30ea906Sjfb8856606 
611d30ea906Sjfb8856606 		rte_delay_us(1500);
612d30ea906Sjfb8856606 	}
613d30ea906Sjfb8856606 
614d30ea906Sjfb8856606 	return -EBUSY;
615d30ea906Sjfb8856606 }
616d30ea906Sjfb8856606 
axgbe_write_rss_hash_key(struct axgbe_port * pdata)617*2d9fd380Sjfb8856606 int axgbe_write_rss_hash_key(struct axgbe_port *pdata)
618d30ea906Sjfb8856606 {
619d30ea906Sjfb8856606 	struct rte_eth_rss_conf *rss_conf;
620d30ea906Sjfb8856606 	unsigned int key_regs = sizeof(pdata->rss_key) / sizeof(u32);
621d30ea906Sjfb8856606 	unsigned int *key;
622d30ea906Sjfb8856606 	int ret;
623d30ea906Sjfb8856606 
624d30ea906Sjfb8856606 	rss_conf = &pdata->eth_dev->data->dev_conf.rx_adv_conf.rss_conf;
625d30ea906Sjfb8856606 
626d30ea906Sjfb8856606 	if (!rss_conf->rss_key)
627d30ea906Sjfb8856606 		key = (unsigned int *)&pdata->rss_key;
628d30ea906Sjfb8856606 	else
629d30ea906Sjfb8856606 		key = (unsigned int *)&rss_conf->rss_key;
630d30ea906Sjfb8856606 
631d30ea906Sjfb8856606 	while (key_regs--) {
632d30ea906Sjfb8856606 		ret = axgbe_write_rss_reg(pdata, AXGBE_RSS_HASH_KEY_TYPE,
633d30ea906Sjfb8856606 					  key_regs, *key++);
634d30ea906Sjfb8856606 		if (ret)
635d30ea906Sjfb8856606 			return ret;
636d30ea906Sjfb8856606 	}
637d30ea906Sjfb8856606 
638d30ea906Sjfb8856606 	return 0;
639d30ea906Sjfb8856606 }
640d30ea906Sjfb8856606 
axgbe_write_rss_lookup_table(struct axgbe_port * pdata)641*2d9fd380Sjfb8856606 int axgbe_write_rss_lookup_table(struct axgbe_port *pdata)
642d30ea906Sjfb8856606 {
643d30ea906Sjfb8856606 	unsigned int i;
644d30ea906Sjfb8856606 	int ret;
645d30ea906Sjfb8856606 
646d30ea906Sjfb8856606 	for (i = 0; i < ARRAY_SIZE(pdata->rss_table); i++) {
647d30ea906Sjfb8856606 		ret = axgbe_write_rss_reg(pdata,
648d30ea906Sjfb8856606 					  AXGBE_RSS_LOOKUP_TABLE_TYPE, i,
649d30ea906Sjfb8856606 					  pdata->rss_table[i]);
650d30ea906Sjfb8856606 		if (ret)
651d30ea906Sjfb8856606 			return ret;
652d30ea906Sjfb8856606 	}
653d30ea906Sjfb8856606 
654d30ea906Sjfb8856606 	return 0;
655d30ea906Sjfb8856606 }
656d30ea906Sjfb8856606 
axgbe_enable_rss(struct axgbe_port * pdata)657d30ea906Sjfb8856606 static int axgbe_enable_rss(struct axgbe_port *pdata)
658d30ea906Sjfb8856606 {
659d30ea906Sjfb8856606 	int ret;
660d30ea906Sjfb8856606 
661d30ea906Sjfb8856606 	/* Program the hash key */
662d30ea906Sjfb8856606 	ret = axgbe_write_rss_hash_key(pdata);
663d30ea906Sjfb8856606 	if (ret)
664d30ea906Sjfb8856606 		return ret;
665d30ea906Sjfb8856606 
666d30ea906Sjfb8856606 	/* Program the lookup table */
667d30ea906Sjfb8856606 	ret = axgbe_write_rss_lookup_table(pdata);
668d30ea906Sjfb8856606 	if (ret)
669d30ea906Sjfb8856606 		return ret;
670d30ea906Sjfb8856606 
671d30ea906Sjfb8856606 	/* Set the RSS options */
672d30ea906Sjfb8856606 	AXGMAC_IOWRITE(pdata, MAC_RSSCR, pdata->rss_options);
673d30ea906Sjfb8856606 
674d30ea906Sjfb8856606 	/* Enable RSS */
675d30ea906Sjfb8856606 	AXGMAC_IOWRITE_BITS(pdata, MAC_RSSCR, RSSE, 1);
676d30ea906Sjfb8856606 
677d30ea906Sjfb8856606 	return 0;
678d30ea906Sjfb8856606 }
679d30ea906Sjfb8856606 
axgbe_rss_options(struct axgbe_port * pdata)680d30ea906Sjfb8856606 static void axgbe_rss_options(struct axgbe_port *pdata)
681d30ea906Sjfb8856606 {
682d30ea906Sjfb8856606 	struct rte_eth_rss_conf *rss_conf;
683d30ea906Sjfb8856606 	uint64_t rss_hf;
684d30ea906Sjfb8856606 
685d30ea906Sjfb8856606 	rss_conf = &pdata->eth_dev->data->dev_conf.rx_adv_conf.rss_conf;
686*2d9fd380Sjfb8856606 	pdata->rss_hf = rss_conf->rss_hf;
687d30ea906Sjfb8856606 	rss_hf = rss_conf->rss_hf;
688d30ea906Sjfb8856606 
689d30ea906Sjfb8856606 	if (rss_hf & (ETH_RSS_IPV4 | ETH_RSS_IPV6))
690d30ea906Sjfb8856606 		AXGMAC_SET_BITS(pdata->rss_options, MAC_RSSCR, IP2TE, 1);
691d30ea906Sjfb8856606 	if (rss_hf & (ETH_RSS_NONFRAG_IPV4_TCP | ETH_RSS_NONFRAG_IPV6_TCP))
692d30ea906Sjfb8856606 		AXGMAC_SET_BITS(pdata->rss_options, MAC_RSSCR, TCP4TE, 1);
693d30ea906Sjfb8856606 	if (rss_hf & (ETH_RSS_NONFRAG_IPV4_UDP | ETH_RSS_NONFRAG_IPV6_UDP))
694d30ea906Sjfb8856606 		AXGMAC_SET_BITS(pdata->rss_options, MAC_RSSCR, UDP4TE, 1);
695d30ea906Sjfb8856606 }
696d30ea906Sjfb8856606 
axgbe_config_rss(struct axgbe_port * pdata)697d30ea906Sjfb8856606 static int axgbe_config_rss(struct axgbe_port *pdata)
698d30ea906Sjfb8856606 {
699d30ea906Sjfb8856606 	uint32_t i;
700d30ea906Sjfb8856606 
701d30ea906Sjfb8856606 	if (pdata->rss_enable) {
702d30ea906Sjfb8856606 		/* Initialize RSS hash key and lookup table */
703d30ea906Sjfb8856606 		uint32_t *key = (uint32_t *)pdata->rss_key;
704d30ea906Sjfb8856606 
705d30ea906Sjfb8856606 		for (i = 0; i < sizeof(pdata->rss_key) / 4; i++)
706d30ea906Sjfb8856606 			*key++ = (uint32_t)rte_rand();
707d30ea906Sjfb8856606 		for (i = 0; i < AXGBE_RSS_MAX_TABLE_SIZE; i++)
708d30ea906Sjfb8856606 			AXGMAC_SET_BITS(pdata->rss_table[i], MAC_RSSDR, DMCH,
709d30ea906Sjfb8856606 					i % pdata->eth_dev->data->nb_rx_queues);
710d30ea906Sjfb8856606 		axgbe_rss_options(pdata);
711d30ea906Sjfb8856606 		if (axgbe_enable_rss(pdata)) {
712d30ea906Sjfb8856606 			PMD_DRV_LOG(ERR, "Error in enabling RSS support\n");
713d30ea906Sjfb8856606 			return -1;
714d30ea906Sjfb8856606 		}
715d30ea906Sjfb8856606 	} else {
716d30ea906Sjfb8856606 		AXGMAC_IOWRITE_BITS(pdata, MAC_RSSCR, RSSE, 0);
717d30ea906Sjfb8856606 	}
718d30ea906Sjfb8856606 
719d30ea906Sjfb8856606 	return 0;
720d30ea906Sjfb8856606 }
721d30ea906Sjfb8856606 
axgbe_enable_dma_interrupts(struct axgbe_port * pdata)722d30ea906Sjfb8856606 static void axgbe_enable_dma_interrupts(struct axgbe_port *pdata)
723d30ea906Sjfb8856606 {
724d30ea906Sjfb8856606 	struct axgbe_tx_queue *txq;
725d30ea906Sjfb8856606 	unsigned int dma_ch_isr, dma_ch_ier;
726d30ea906Sjfb8856606 	unsigned int i;
727d30ea906Sjfb8856606 
728d30ea906Sjfb8856606 	for (i = 0; i < pdata->eth_dev->data->nb_tx_queues; i++) {
729d30ea906Sjfb8856606 		txq = pdata->eth_dev->data->tx_queues[i];
730d30ea906Sjfb8856606 
731d30ea906Sjfb8856606 		/* Clear all the interrupts which are set */
732d30ea906Sjfb8856606 		dma_ch_isr = AXGMAC_DMA_IOREAD(txq, DMA_CH_SR);
733d30ea906Sjfb8856606 		AXGMAC_DMA_IOWRITE(txq, DMA_CH_SR, dma_ch_isr);
734d30ea906Sjfb8856606 
735d30ea906Sjfb8856606 		/* Clear all interrupt enable bits */
736d30ea906Sjfb8856606 		dma_ch_ier = 0;
737d30ea906Sjfb8856606 
738d30ea906Sjfb8856606 		/* Enable following interrupts
739d30ea906Sjfb8856606 		 *   NIE  - Normal Interrupt Summary Enable
740d30ea906Sjfb8856606 		 *   AIE  - Abnormal Interrupt Summary Enable
741d30ea906Sjfb8856606 		 *   FBEE - Fatal Bus Error Enable
742d30ea906Sjfb8856606 		 */
743d30ea906Sjfb8856606 		AXGMAC_SET_BITS(dma_ch_ier, DMA_CH_IER, NIE, 0);
744d30ea906Sjfb8856606 		AXGMAC_SET_BITS(dma_ch_ier, DMA_CH_IER, AIE, 1);
745d30ea906Sjfb8856606 		AXGMAC_SET_BITS(dma_ch_ier, DMA_CH_IER, FBEE, 1);
746d30ea906Sjfb8856606 
747d30ea906Sjfb8856606 		/* Enable following Rx interrupts
748d30ea906Sjfb8856606 		 *   RBUE - Receive Buffer Unavailable Enable
749d30ea906Sjfb8856606 		 *   RIE  - Receive Interrupt Enable (unless using
750d30ea906Sjfb8856606 		 *          per channel interrupts in edge triggered
751d30ea906Sjfb8856606 		 *          mode)
752d30ea906Sjfb8856606 		 */
753d30ea906Sjfb8856606 		AXGMAC_SET_BITS(dma_ch_ier, DMA_CH_IER, RBUE, 0);
754d30ea906Sjfb8856606 
755d30ea906Sjfb8856606 		AXGMAC_DMA_IOWRITE(txq, DMA_CH_IER, dma_ch_ier);
756d30ea906Sjfb8856606 	}
757d30ea906Sjfb8856606 }
758d30ea906Sjfb8856606 
wrapper_tx_desc_init(struct axgbe_port * pdata)759d30ea906Sjfb8856606 static void wrapper_tx_desc_init(struct axgbe_port *pdata)
760d30ea906Sjfb8856606 {
761d30ea906Sjfb8856606 	struct axgbe_tx_queue *txq;
762d30ea906Sjfb8856606 	unsigned int i;
763d30ea906Sjfb8856606 
764d30ea906Sjfb8856606 	for (i = 0; i < pdata->eth_dev->data->nb_tx_queues; i++) {
765d30ea906Sjfb8856606 		txq = pdata->eth_dev->data->tx_queues[i];
766d30ea906Sjfb8856606 		txq->cur = 0;
767d30ea906Sjfb8856606 		txq->dirty = 0;
768d30ea906Sjfb8856606 		/* Update the total number of Tx descriptors */
769d30ea906Sjfb8856606 		AXGMAC_DMA_IOWRITE(txq, DMA_CH_TDRLR, txq->nb_desc - 1);
770d30ea906Sjfb8856606 		/* Update the starting address of descriptor ring */
771d30ea906Sjfb8856606 		AXGMAC_DMA_IOWRITE(txq, DMA_CH_TDLR_HI,
772d30ea906Sjfb8856606 					high32_value(txq->ring_phys_addr));
773d30ea906Sjfb8856606 		AXGMAC_DMA_IOWRITE(txq, DMA_CH_TDLR_LO,
774d30ea906Sjfb8856606 					low32_value(txq->ring_phys_addr));
775d30ea906Sjfb8856606 	}
776d30ea906Sjfb8856606 }
777d30ea906Sjfb8856606 
wrapper_rx_desc_init(struct axgbe_port * pdata)778d30ea906Sjfb8856606 static int wrapper_rx_desc_init(struct axgbe_port *pdata)
779d30ea906Sjfb8856606 {
780d30ea906Sjfb8856606 	struct axgbe_rx_queue *rxq;
781d30ea906Sjfb8856606 	struct rte_mbuf *mbuf;
782d30ea906Sjfb8856606 	volatile union axgbe_rx_desc *desc;
783d30ea906Sjfb8856606 	unsigned int i, j;
784d30ea906Sjfb8856606 
785d30ea906Sjfb8856606 	for (i = 0; i < pdata->eth_dev->data->nb_rx_queues; i++) {
786d30ea906Sjfb8856606 		rxq = pdata->eth_dev->data->rx_queues[i];
787d30ea906Sjfb8856606 
788d30ea906Sjfb8856606 		/* Initialize software ring entries */
789d30ea906Sjfb8856606 		rxq->mbuf_alloc = 0;
790d30ea906Sjfb8856606 		rxq->cur = 0;
791d30ea906Sjfb8856606 		rxq->dirty = 0;
792d30ea906Sjfb8856606 		desc = AXGBE_GET_DESC_PT(rxq, 0);
793d30ea906Sjfb8856606 
794d30ea906Sjfb8856606 		for (j = 0; j < rxq->nb_desc; j++) {
795d30ea906Sjfb8856606 			mbuf = rte_mbuf_raw_alloc(rxq->mb_pool);
796d30ea906Sjfb8856606 			if (mbuf == NULL) {
797d30ea906Sjfb8856606 				PMD_DRV_LOG(ERR, "RX mbuf alloc failed queue_id = %u, idx = %d\n",
798d30ea906Sjfb8856606 					    (unsigned int)rxq->queue_id, j);
799d30ea906Sjfb8856606 				axgbe_dev_rx_queue_release(rxq);
800d30ea906Sjfb8856606 				return -ENOMEM;
801d30ea906Sjfb8856606 			}
802d30ea906Sjfb8856606 			rxq->sw_ring[j] = mbuf;
803d30ea906Sjfb8856606 			/* Mbuf populate */
804d30ea906Sjfb8856606 			mbuf->next = NULL;
805d30ea906Sjfb8856606 			mbuf->data_off = RTE_PKTMBUF_HEADROOM;
806d30ea906Sjfb8856606 			mbuf->nb_segs = 1;
807d30ea906Sjfb8856606 			mbuf->port = rxq->port_id;
808d30ea906Sjfb8856606 			desc->read.baddr =
809d30ea906Sjfb8856606 				rte_cpu_to_le_64(
810d30ea906Sjfb8856606 					rte_mbuf_data_iova_default(mbuf));
811d30ea906Sjfb8856606 			rte_wmb();
812d30ea906Sjfb8856606 			AXGMAC_SET_BITS_LE(desc->read.desc3,
813d30ea906Sjfb8856606 						RX_NORMAL_DESC3, OWN, 1);
814d30ea906Sjfb8856606 			rte_wmb();
815d30ea906Sjfb8856606 			rxq->mbuf_alloc++;
816d30ea906Sjfb8856606 			desc++;
817d30ea906Sjfb8856606 		}
818d30ea906Sjfb8856606 		/* Update the total number of Rx descriptors */
819d30ea906Sjfb8856606 		AXGMAC_DMA_IOWRITE(rxq, DMA_CH_RDRLR,
820d30ea906Sjfb8856606 					rxq->nb_desc - 1);
821d30ea906Sjfb8856606 		/* Update the starting address of descriptor ring */
822d30ea906Sjfb8856606 		AXGMAC_DMA_IOWRITE(rxq, DMA_CH_RDLR_HI,
823d30ea906Sjfb8856606 					high32_value(rxq->ring_phys_addr));
824d30ea906Sjfb8856606 		AXGMAC_DMA_IOWRITE(rxq, DMA_CH_RDLR_LO,
825d30ea906Sjfb8856606 					low32_value(rxq->ring_phys_addr));
826d30ea906Sjfb8856606 		/* Update the Rx Descriptor Tail Pointer */
827d30ea906Sjfb8856606 		AXGMAC_DMA_IOWRITE(rxq, DMA_CH_RDTR_LO,
828d30ea906Sjfb8856606 				   low32_value(rxq->ring_phys_addr +
829d30ea906Sjfb8856606 				   (rxq->nb_desc - 1) *
830d30ea906Sjfb8856606 				   sizeof(union axgbe_rx_desc)));
831d30ea906Sjfb8856606 	}
832d30ea906Sjfb8856606 	return 0;
833d30ea906Sjfb8856606 }
834d30ea906Sjfb8856606 
axgbe_config_mtl_mode(struct axgbe_port * pdata)835d30ea906Sjfb8856606 static void axgbe_config_mtl_mode(struct axgbe_port *pdata)
836d30ea906Sjfb8856606 {
837d30ea906Sjfb8856606 	unsigned int i;
838d30ea906Sjfb8856606 
839d30ea906Sjfb8856606 	/* Set Tx to weighted round robin scheduling algorithm */
840d30ea906Sjfb8856606 	AXGMAC_IOWRITE_BITS(pdata, MTL_OMR, ETSALG, MTL_ETSALG_WRR);
841d30ea906Sjfb8856606 
842d30ea906Sjfb8856606 	/* Set Tx traffic classes to use WRR algorithm with equal weights */
843d30ea906Sjfb8856606 	for (i = 0; i < pdata->hw_feat.tc_cnt; i++) {
844d30ea906Sjfb8856606 		AXGMAC_MTL_IOWRITE_BITS(pdata, i, MTL_TC_ETSCR, TSA,
845d30ea906Sjfb8856606 				MTL_TSA_ETS);
846d30ea906Sjfb8856606 		AXGMAC_MTL_IOWRITE_BITS(pdata, i, MTL_TC_QWR, QW, 1);
847d30ea906Sjfb8856606 	}
848d30ea906Sjfb8856606 
849d30ea906Sjfb8856606 	/* Set Rx to strict priority algorithm */
850d30ea906Sjfb8856606 	AXGMAC_IOWRITE_BITS(pdata, MTL_OMR, RAA, MTL_RAA_SP);
851d30ea906Sjfb8856606 }
852d30ea906Sjfb8856606 
axgbe_config_tsf_mode(struct axgbe_port * pdata,unsigned int val)853d30ea906Sjfb8856606 static int axgbe_config_tsf_mode(struct axgbe_port *pdata, unsigned int val)
854d30ea906Sjfb8856606 {
855d30ea906Sjfb8856606 	unsigned int i;
856d30ea906Sjfb8856606 
857d30ea906Sjfb8856606 	for (i = 0; i < pdata->tx_q_count; i++)
858d30ea906Sjfb8856606 		AXGMAC_MTL_IOWRITE_BITS(pdata, i, MTL_Q_TQOMR, TSF, val);
859d30ea906Sjfb8856606 
860d30ea906Sjfb8856606 	return 0;
861d30ea906Sjfb8856606 }
862d30ea906Sjfb8856606 
axgbe_config_rsf_mode(struct axgbe_port * pdata,unsigned int val)863d30ea906Sjfb8856606 static int axgbe_config_rsf_mode(struct axgbe_port *pdata, unsigned int val)
864d30ea906Sjfb8856606 {
865d30ea906Sjfb8856606 	unsigned int i;
866d30ea906Sjfb8856606 
867d30ea906Sjfb8856606 	for (i = 0; i < pdata->rx_q_count; i++)
868d30ea906Sjfb8856606 		AXGMAC_MTL_IOWRITE_BITS(pdata, i, MTL_Q_RQOMR, RSF, val);
869d30ea906Sjfb8856606 
870d30ea906Sjfb8856606 	return 0;
871d30ea906Sjfb8856606 }
872d30ea906Sjfb8856606 
axgbe_config_tx_threshold(struct axgbe_port * pdata,unsigned int val)873d30ea906Sjfb8856606 static int axgbe_config_tx_threshold(struct axgbe_port *pdata,
874d30ea906Sjfb8856606 				     unsigned int val)
875d30ea906Sjfb8856606 {
876d30ea906Sjfb8856606 	unsigned int i;
877d30ea906Sjfb8856606 
878d30ea906Sjfb8856606 	for (i = 0; i < pdata->tx_q_count; i++)
879d30ea906Sjfb8856606 		AXGMAC_MTL_IOWRITE_BITS(pdata, i, MTL_Q_TQOMR, TTC, val);
880d30ea906Sjfb8856606 
881d30ea906Sjfb8856606 	return 0;
882d30ea906Sjfb8856606 }
883d30ea906Sjfb8856606 
axgbe_config_rx_threshold(struct axgbe_port * pdata,unsigned int val)884d30ea906Sjfb8856606 static int axgbe_config_rx_threshold(struct axgbe_port *pdata,
885d30ea906Sjfb8856606 				     unsigned int val)
886d30ea906Sjfb8856606 {
887d30ea906Sjfb8856606 	unsigned int i;
888d30ea906Sjfb8856606 
889d30ea906Sjfb8856606 	for (i = 0; i < pdata->rx_q_count; i++)
890d30ea906Sjfb8856606 		AXGMAC_MTL_IOWRITE_BITS(pdata, i, MTL_Q_RQOMR, RTC, val);
891d30ea906Sjfb8856606 
892d30ea906Sjfb8856606 	return 0;
893d30ea906Sjfb8856606 }
894d30ea906Sjfb8856606 
895d30ea906Sjfb8856606 /*Distrubting fifo size  */
axgbe_config_rx_fifo_size(struct axgbe_port * pdata)896d30ea906Sjfb8856606 static void axgbe_config_rx_fifo_size(struct axgbe_port *pdata)
897d30ea906Sjfb8856606 {
898d30ea906Sjfb8856606 	unsigned int fifo_size;
899d30ea906Sjfb8856606 	unsigned int q_fifo_size;
900d30ea906Sjfb8856606 	unsigned int p_fifo, i;
901d30ea906Sjfb8856606 
902d30ea906Sjfb8856606 	fifo_size = RTE_MIN(pdata->rx_max_fifo_size,
903d30ea906Sjfb8856606 			  pdata->hw_feat.rx_fifo_size);
904d30ea906Sjfb8856606 	q_fifo_size = fifo_size / pdata->rx_q_count;
905d30ea906Sjfb8856606 
906d30ea906Sjfb8856606 	/* Calculate the fifo setting by dividing the queue's fifo size
907d30ea906Sjfb8856606 	 * by the fifo allocation increment (with 0 representing the
908d30ea906Sjfb8856606 	 * base allocation increment so decrement the result
909d30ea906Sjfb8856606 	 * by 1).
910d30ea906Sjfb8856606 	 */
911d30ea906Sjfb8856606 	p_fifo = q_fifo_size / AXGMAC_FIFO_UNIT;
912d30ea906Sjfb8856606 	if (p_fifo)
913d30ea906Sjfb8856606 		p_fifo--;
914d30ea906Sjfb8856606 
915d30ea906Sjfb8856606 	for (i = 0; i < pdata->rx_q_count; i++)
916d30ea906Sjfb8856606 		AXGMAC_MTL_IOWRITE_BITS(pdata, i, MTL_Q_RQOMR, RQS, p_fifo);
917d30ea906Sjfb8856606 	pdata->fifo = p_fifo;
918d30ea906Sjfb8856606 
919d30ea906Sjfb8856606 	/*Calculate and config Flow control threshold*/
920d30ea906Sjfb8856606 	axgbe_calculate_flow_control_threshold(pdata);
921d30ea906Sjfb8856606 	axgbe_config_flow_control_threshold(pdata);
922*2d9fd380Sjfb8856606 
923*2d9fd380Sjfb8856606 	PMD_DRV_LOG(DEBUG, "%d Rx hardware queues, %d byte fifo per queue\n",
924*2d9fd380Sjfb8856606 		    pdata->rx_q_count, q_fifo_size);
925d30ea906Sjfb8856606 }
926d30ea906Sjfb8856606 
axgbe_config_tx_fifo_size(struct axgbe_port * pdata)927d30ea906Sjfb8856606 static void axgbe_config_tx_fifo_size(struct axgbe_port *pdata)
928d30ea906Sjfb8856606 {
929d30ea906Sjfb8856606 	unsigned int fifo_size;
930d30ea906Sjfb8856606 	unsigned int q_fifo_size;
931d30ea906Sjfb8856606 	unsigned int p_fifo, i;
932d30ea906Sjfb8856606 
933d30ea906Sjfb8856606 	fifo_size = RTE_MIN(pdata->tx_max_fifo_size,
934d30ea906Sjfb8856606 				pdata->hw_feat.tx_fifo_size);
935d30ea906Sjfb8856606 	q_fifo_size = fifo_size / pdata->tx_q_count;
936d30ea906Sjfb8856606 
937d30ea906Sjfb8856606 	/* Calculate the fifo setting by dividing the queue's fifo size
938d30ea906Sjfb8856606 	 * by the fifo allocation increment (with 0 representing the
939d30ea906Sjfb8856606 	 * base allocation increment so decrement the result
940d30ea906Sjfb8856606 	 * by 1).
941d30ea906Sjfb8856606 	 */
942d30ea906Sjfb8856606 	p_fifo = q_fifo_size / AXGMAC_FIFO_UNIT;
943d30ea906Sjfb8856606 	if (p_fifo)
944d30ea906Sjfb8856606 		p_fifo--;
945d30ea906Sjfb8856606 
946d30ea906Sjfb8856606 	for (i = 0; i < pdata->tx_q_count; i++)
947d30ea906Sjfb8856606 		AXGMAC_MTL_IOWRITE_BITS(pdata, i, MTL_Q_TQOMR, TQS, p_fifo);
948*2d9fd380Sjfb8856606 
949*2d9fd380Sjfb8856606 	PMD_DRV_LOG(DEBUG, "%d Tx hardware queues, %d byte fifo per queue\n",
950*2d9fd380Sjfb8856606 		    pdata->tx_q_count, q_fifo_size);
951d30ea906Sjfb8856606 }
952d30ea906Sjfb8856606 
axgbe_config_queue_mapping(struct axgbe_port * pdata)953d30ea906Sjfb8856606 static void axgbe_config_queue_mapping(struct axgbe_port *pdata)
954d30ea906Sjfb8856606 {
955d30ea906Sjfb8856606 	unsigned int qptc, qptc_extra, queue;
956d30ea906Sjfb8856606 	unsigned int i, j, reg, reg_val;
957d30ea906Sjfb8856606 
958d30ea906Sjfb8856606 	/* Map the MTL Tx Queues to Traffic Classes
959d30ea906Sjfb8856606 	 *   Note: Tx Queues >= Traffic Classes
960d30ea906Sjfb8856606 	 */
961d30ea906Sjfb8856606 	qptc = pdata->tx_q_count / pdata->hw_feat.tc_cnt;
962d30ea906Sjfb8856606 	qptc_extra = pdata->tx_q_count % pdata->hw_feat.tc_cnt;
963d30ea906Sjfb8856606 
964d30ea906Sjfb8856606 	for (i = 0, queue = 0; i < pdata->hw_feat.tc_cnt; i++) {
965*2d9fd380Sjfb8856606 		for (j = 0; j < qptc; j++) {
966*2d9fd380Sjfb8856606 			PMD_DRV_LOG(DEBUG, "TXq%u mapped to TC%u\n", queue, i);
967d30ea906Sjfb8856606 			AXGMAC_MTL_IOWRITE_BITS(pdata, queue, MTL_Q_TQOMR,
968d30ea906Sjfb8856606 						Q2TCMAP, i);
969*2d9fd380Sjfb8856606 		}
970*2d9fd380Sjfb8856606 		if (i < qptc_extra) {
971*2d9fd380Sjfb8856606 			PMD_DRV_LOG(DEBUG, "TXq%u mapped to TC%u\n", queue, i);
972d30ea906Sjfb8856606 			AXGMAC_MTL_IOWRITE_BITS(pdata, queue, MTL_Q_TQOMR,
973d30ea906Sjfb8856606 						Q2TCMAP, i);
974d30ea906Sjfb8856606 		}
975*2d9fd380Sjfb8856606 	}
976d30ea906Sjfb8856606 
977d30ea906Sjfb8856606 	if (pdata->rss_enable) {
978d30ea906Sjfb8856606 		/* Select dynamic mapping of MTL Rx queue to DMA Rx channel */
979d30ea906Sjfb8856606 		reg = MTL_RQDCM0R;
980d30ea906Sjfb8856606 		reg_val = 0;
981d30ea906Sjfb8856606 		for (i = 0; i < pdata->rx_q_count;) {
982d30ea906Sjfb8856606 			reg_val |= (0x80 << ((i++ % MTL_RQDCM_Q_PER_REG) << 3));
983d30ea906Sjfb8856606 
984d30ea906Sjfb8856606 			if ((i % MTL_RQDCM_Q_PER_REG) &&
985d30ea906Sjfb8856606 			    (i != pdata->rx_q_count))
986d30ea906Sjfb8856606 				continue;
987d30ea906Sjfb8856606 
988d30ea906Sjfb8856606 			AXGMAC_IOWRITE(pdata, reg, reg_val);
989d30ea906Sjfb8856606 
990d30ea906Sjfb8856606 			reg += MTL_RQDCM_INC;
991d30ea906Sjfb8856606 			reg_val = 0;
992d30ea906Sjfb8856606 		}
993d30ea906Sjfb8856606 	}
994d30ea906Sjfb8856606 }
995d30ea906Sjfb8856606 
axgbe_enable_mtl_interrupts(struct axgbe_port * pdata)996d30ea906Sjfb8856606 static void axgbe_enable_mtl_interrupts(struct axgbe_port *pdata)
997d30ea906Sjfb8856606 {
998d30ea906Sjfb8856606 	unsigned int mtl_q_isr;
999d30ea906Sjfb8856606 	unsigned int q_count, i;
1000d30ea906Sjfb8856606 
1001d30ea906Sjfb8856606 	q_count = RTE_MAX(pdata->hw_feat.tx_q_cnt, pdata->hw_feat.rx_q_cnt);
1002d30ea906Sjfb8856606 	for (i = 0; i < q_count; i++) {
1003d30ea906Sjfb8856606 		/* Clear all the interrupts which are set */
1004d30ea906Sjfb8856606 		mtl_q_isr = AXGMAC_MTL_IOREAD(pdata, i, MTL_Q_ISR);
1005d30ea906Sjfb8856606 		AXGMAC_MTL_IOWRITE(pdata, i, MTL_Q_ISR, mtl_q_isr);
1006d30ea906Sjfb8856606 
1007d30ea906Sjfb8856606 		/* No MTL interrupts to be enabled */
1008d30ea906Sjfb8856606 		AXGMAC_MTL_IOWRITE(pdata, i, MTL_Q_IER, 0);
1009d30ea906Sjfb8856606 	}
1010d30ea906Sjfb8856606 }
1011d30ea906Sjfb8856606 
bitrev32(uint32_t x)1012*2d9fd380Sjfb8856606 static uint32_t bitrev32(uint32_t x)
1013*2d9fd380Sjfb8856606 {
1014*2d9fd380Sjfb8856606 	x = (x >> 16) | (x << 16);
1015*2d9fd380Sjfb8856606 	x = (((x & 0xff00ff00) >> 8) | ((x & 0x00ff00ff) << 8));
1016*2d9fd380Sjfb8856606 	x = (((x & 0xf0f0f0f0) >> 4) | ((x & 0x0f0f0f0f) << 4));
1017*2d9fd380Sjfb8856606 	x = (((x & 0xcccccccc) >> 2) | ((x & 0x33333333) << 2));
1018*2d9fd380Sjfb8856606 	x = (((x & 0xaaaaaaaa) >> 1) | ((x & 0x55555555) << 1));
1019*2d9fd380Sjfb8856606 	return x;
1020*2d9fd380Sjfb8856606 }
1021*2d9fd380Sjfb8856606 
crc32_le(uint32_t crc,uint8_t * p,uint32_t len)1022*2d9fd380Sjfb8856606 static uint32_t crc32_le(uint32_t crc, uint8_t *p, uint32_t len)
1023*2d9fd380Sjfb8856606 {
1024*2d9fd380Sjfb8856606 	int i;
1025*2d9fd380Sjfb8856606 	while (len--) {
1026*2d9fd380Sjfb8856606 		crc ^= *p++;
1027*2d9fd380Sjfb8856606 		for (i = 0; i < 8; i++)
1028*2d9fd380Sjfb8856606 			crc = (crc >> 1) ^ ((crc & 1) ? 0xedb88320 : 0);
1029*2d9fd380Sjfb8856606 	}
1030*2d9fd380Sjfb8856606 	return crc;
1031*2d9fd380Sjfb8856606 }
1032*2d9fd380Sjfb8856606 
axgbe_set_mac_hash_table(struct axgbe_port * pdata,u8 * addr,bool add)1033*2d9fd380Sjfb8856606 void axgbe_set_mac_hash_table(struct axgbe_port *pdata, u8 *addr, bool add)
1034*2d9fd380Sjfb8856606 {
1035*2d9fd380Sjfb8856606 	uint32_t crc, htable_index, htable_bitmask;
1036*2d9fd380Sjfb8856606 
1037*2d9fd380Sjfb8856606 	crc = bitrev32(~crc32_le(~0, addr, RTE_ETHER_ADDR_LEN));
1038*2d9fd380Sjfb8856606 	crc >>= pdata->hash_table_shift;
1039*2d9fd380Sjfb8856606 	htable_index = crc >> 5;
1040*2d9fd380Sjfb8856606 	htable_bitmask = 1 << (crc & 0x1f);
1041*2d9fd380Sjfb8856606 
1042*2d9fd380Sjfb8856606 	if (add) {
1043*2d9fd380Sjfb8856606 		pdata->uc_hash_table[htable_index] |= htable_bitmask;
1044*2d9fd380Sjfb8856606 		pdata->uc_hash_mac_addr++;
1045*2d9fd380Sjfb8856606 	} else {
1046*2d9fd380Sjfb8856606 		pdata->uc_hash_table[htable_index] &= ~htable_bitmask;
1047*2d9fd380Sjfb8856606 		pdata->uc_hash_mac_addr--;
1048*2d9fd380Sjfb8856606 	}
1049*2d9fd380Sjfb8856606 	PMD_DRV_LOG(DEBUG, "%s MAC hash table Bit %d at Index %#x\n",
1050*2d9fd380Sjfb8856606 		    add ? "set" : "clear", (crc & 0x1f), htable_index);
1051*2d9fd380Sjfb8856606 
1052*2d9fd380Sjfb8856606 	AXGMAC_IOWRITE(pdata, MAC_HTR(htable_index),
1053*2d9fd380Sjfb8856606 		       pdata->uc_hash_table[htable_index]);
1054*2d9fd380Sjfb8856606 }
1055*2d9fd380Sjfb8856606 
axgbe_set_mac_addn_addr(struct axgbe_port * pdata,u8 * addr,uint32_t index)1056*2d9fd380Sjfb8856606 void axgbe_set_mac_addn_addr(struct axgbe_port *pdata, u8 *addr, uint32_t index)
1057*2d9fd380Sjfb8856606 {
1058*2d9fd380Sjfb8856606 	unsigned int mac_addr_hi, mac_addr_lo;
1059*2d9fd380Sjfb8856606 	u8 *mac_addr;
1060*2d9fd380Sjfb8856606 
1061*2d9fd380Sjfb8856606 	mac_addr_lo = 0;
1062*2d9fd380Sjfb8856606 	mac_addr_hi = 0;
1063*2d9fd380Sjfb8856606 
1064*2d9fd380Sjfb8856606 	if (addr) {
1065*2d9fd380Sjfb8856606 		mac_addr = (u8 *)&mac_addr_lo;
1066*2d9fd380Sjfb8856606 		mac_addr[0] = addr[0];
1067*2d9fd380Sjfb8856606 		mac_addr[1] = addr[1];
1068*2d9fd380Sjfb8856606 		mac_addr[2] = addr[2];
1069*2d9fd380Sjfb8856606 		mac_addr[3] = addr[3];
1070*2d9fd380Sjfb8856606 		mac_addr = (u8 *)&mac_addr_hi;
1071*2d9fd380Sjfb8856606 		mac_addr[0] = addr[4];
1072*2d9fd380Sjfb8856606 		mac_addr[1] = addr[5];
1073*2d9fd380Sjfb8856606 
1074*2d9fd380Sjfb8856606 		/*Address Enable: Use this Addr for Perfect Filtering */
1075*2d9fd380Sjfb8856606 		AXGMAC_SET_BITS(mac_addr_hi, MAC_MACA1HR, AE, 1);
1076*2d9fd380Sjfb8856606 	}
1077*2d9fd380Sjfb8856606 
1078*2d9fd380Sjfb8856606 	PMD_DRV_LOG(DEBUG, "%s mac address at %#x\n",
1079*2d9fd380Sjfb8856606 		    addr ? "set" : "clear", index);
1080*2d9fd380Sjfb8856606 
1081*2d9fd380Sjfb8856606 	AXGMAC_IOWRITE(pdata, MAC_MACAHR(index), mac_addr_hi);
1082*2d9fd380Sjfb8856606 	AXGMAC_IOWRITE(pdata, MAC_MACALR(index), mac_addr_lo);
1083*2d9fd380Sjfb8856606 }
1084*2d9fd380Sjfb8856606 
axgbe_set_mac_address(struct axgbe_port * pdata,u8 * addr)1085d30ea906Sjfb8856606 static int axgbe_set_mac_address(struct axgbe_port *pdata, u8 *addr)
1086d30ea906Sjfb8856606 {
1087d30ea906Sjfb8856606 	unsigned int mac_addr_hi, mac_addr_lo;
1088d30ea906Sjfb8856606 
1089d30ea906Sjfb8856606 	mac_addr_hi = (addr[5] <<  8) | (addr[4] <<  0);
1090d30ea906Sjfb8856606 	mac_addr_lo = (addr[3] << 24) | (addr[2] << 16) |
1091d30ea906Sjfb8856606 		(addr[1] <<  8) | (addr[0] <<  0);
1092d30ea906Sjfb8856606 
1093d30ea906Sjfb8856606 	AXGMAC_IOWRITE(pdata, MAC_MACA0HR, mac_addr_hi);
1094d30ea906Sjfb8856606 	AXGMAC_IOWRITE(pdata, MAC_MACA0LR, mac_addr_lo);
1095d30ea906Sjfb8856606 
1096d30ea906Sjfb8856606 	return 0;
1097d30ea906Sjfb8856606 }
1098d30ea906Sjfb8856606 
axgbe_config_mac_hash_table(struct axgbe_port * pdata)1099*2d9fd380Sjfb8856606 static void axgbe_config_mac_hash_table(struct axgbe_port *pdata)
1100*2d9fd380Sjfb8856606 {
1101*2d9fd380Sjfb8856606 	struct axgbe_hw_features *hw_feat = &pdata->hw_feat;
1102*2d9fd380Sjfb8856606 
1103*2d9fd380Sjfb8856606 	pdata->hash_table_shift = 0;
1104*2d9fd380Sjfb8856606 	pdata->hash_table_count = 0;
1105*2d9fd380Sjfb8856606 	pdata->uc_hash_mac_addr = 0;
1106*2d9fd380Sjfb8856606 	memset(pdata->uc_hash_table, 0, sizeof(pdata->uc_hash_table));
1107*2d9fd380Sjfb8856606 
1108*2d9fd380Sjfb8856606 	if (hw_feat->hash_table_size) {
1109*2d9fd380Sjfb8856606 		pdata->hash_table_shift = 26 - (hw_feat->hash_table_size >> 7);
1110*2d9fd380Sjfb8856606 		pdata->hash_table_count = hw_feat->hash_table_size / 32;
1111*2d9fd380Sjfb8856606 	}
1112*2d9fd380Sjfb8856606 }
1113*2d9fd380Sjfb8856606 
axgbe_config_mac_address(struct axgbe_port * pdata)1114d30ea906Sjfb8856606 static void axgbe_config_mac_address(struct axgbe_port *pdata)
1115d30ea906Sjfb8856606 {
1116d30ea906Sjfb8856606 	axgbe_set_mac_address(pdata, pdata->mac_addr.addr_bytes);
1117d30ea906Sjfb8856606 }
1118d30ea906Sjfb8856606 
axgbe_config_jumbo_enable(struct axgbe_port * pdata)1119d30ea906Sjfb8856606 static void axgbe_config_jumbo_enable(struct axgbe_port *pdata)
1120d30ea906Sjfb8856606 {
1121d30ea906Sjfb8856606 	unsigned int val;
1122d30ea906Sjfb8856606 
1123d30ea906Sjfb8856606 	val = (pdata->rx_buf_size > AXGMAC_STD_PACKET_MTU) ? 1 : 0;
1124d30ea906Sjfb8856606 
1125d30ea906Sjfb8856606 	AXGMAC_IOWRITE_BITS(pdata, MAC_RCR, JE, val);
1126d30ea906Sjfb8856606 }
1127d30ea906Sjfb8856606 
axgbe_config_mac_speed(struct axgbe_port * pdata)1128d30ea906Sjfb8856606 static void axgbe_config_mac_speed(struct axgbe_port *pdata)
1129d30ea906Sjfb8856606 {
1130d30ea906Sjfb8856606 	axgbe_set_speed(pdata, pdata->phy_speed);
1131d30ea906Sjfb8856606 }
1132d30ea906Sjfb8856606 
axgbe_config_checksum_offload(struct axgbe_port * pdata)1133d30ea906Sjfb8856606 static void axgbe_config_checksum_offload(struct axgbe_port *pdata)
1134d30ea906Sjfb8856606 {
1135d30ea906Sjfb8856606 	if (pdata->rx_csum_enable)
1136d30ea906Sjfb8856606 		AXGMAC_IOWRITE_BITS(pdata, MAC_RCR, IPC, 1);
1137d30ea906Sjfb8856606 	else
1138d30ea906Sjfb8856606 		AXGMAC_IOWRITE_BITS(pdata, MAC_RCR, IPC, 0);
1139d30ea906Sjfb8856606 }
1140d30ea906Sjfb8856606 
axgbe_config_mmc(struct axgbe_port * pdata)1141*2d9fd380Sjfb8856606 static void axgbe_config_mmc(struct axgbe_port *pdata)
1142*2d9fd380Sjfb8856606 {
1143*2d9fd380Sjfb8856606 	struct axgbe_mmc_stats *stats = &pdata->mmc_stats;
1144*2d9fd380Sjfb8856606 
1145*2d9fd380Sjfb8856606 	/* Reset stats */
1146*2d9fd380Sjfb8856606 	memset(stats, 0, sizeof(*stats));
1147*2d9fd380Sjfb8856606 
1148*2d9fd380Sjfb8856606 	/* Set counters to reset on read */
1149*2d9fd380Sjfb8856606 	AXGMAC_IOWRITE_BITS(pdata, MMC_CR, ROR, 1);
1150*2d9fd380Sjfb8856606 
1151*2d9fd380Sjfb8856606 	/* Reset the counters */
1152*2d9fd380Sjfb8856606 	AXGMAC_IOWRITE_BITS(pdata, MMC_CR, CR, 1);
1153*2d9fd380Sjfb8856606 }
1154*2d9fd380Sjfb8856606 
axgbe_init(struct axgbe_port * pdata)1155d30ea906Sjfb8856606 static int axgbe_init(struct axgbe_port *pdata)
1156d30ea906Sjfb8856606 {
1157d30ea906Sjfb8856606 	int ret;
1158d30ea906Sjfb8856606 
1159d30ea906Sjfb8856606 	/* Flush Tx queues */
1160d30ea906Sjfb8856606 	ret = axgbe_flush_tx_queues(pdata);
1161d30ea906Sjfb8856606 	if (ret)
1162d30ea906Sjfb8856606 		return ret;
1163d30ea906Sjfb8856606 	/* Initialize DMA related features */
1164d30ea906Sjfb8856606 	axgbe_config_dma_bus(pdata);
1165d30ea906Sjfb8856606 	axgbe_config_dma_cache(pdata);
1166d30ea906Sjfb8856606 	axgbe_config_edma_control(pdata);
1167d30ea906Sjfb8856606 	axgbe_config_osp_mode(pdata);
1168d30ea906Sjfb8856606 	axgbe_config_pblx8(pdata);
1169d30ea906Sjfb8856606 	axgbe_config_tx_pbl_val(pdata);
1170d30ea906Sjfb8856606 	axgbe_config_rx_pbl_val(pdata);
1171d30ea906Sjfb8856606 	axgbe_config_rx_buffer_size(pdata);
1172d30ea906Sjfb8856606 	axgbe_config_rss(pdata);
1173d30ea906Sjfb8856606 	wrapper_tx_desc_init(pdata);
1174d30ea906Sjfb8856606 	ret = wrapper_rx_desc_init(pdata);
1175d30ea906Sjfb8856606 	if (ret)
1176d30ea906Sjfb8856606 		return ret;
1177d30ea906Sjfb8856606 	axgbe_enable_dma_interrupts(pdata);
1178d30ea906Sjfb8856606 
1179d30ea906Sjfb8856606 	/* Initialize MTL related features */
1180d30ea906Sjfb8856606 	axgbe_config_mtl_mode(pdata);
1181d30ea906Sjfb8856606 	axgbe_config_queue_mapping(pdata);
1182d30ea906Sjfb8856606 	axgbe_config_tsf_mode(pdata, pdata->tx_sf_mode);
1183d30ea906Sjfb8856606 	axgbe_config_rsf_mode(pdata, pdata->rx_sf_mode);
1184d30ea906Sjfb8856606 	axgbe_config_tx_threshold(pdata, pdata->tx_threshold);
1185d30ea906Sjfb8856606 	axgbe_config_rx_threshold(pdata, pdata->rx_threshold);
1186d30ea906Sjfb8856606 	axgbe_config_tx_fifo_size(pdata);
1187d30ea906Sjfb8856606 	axgbe_config_rx_fifo_size(pdata);
1188d30ea906Sjfb8856606 
1189d30ea906Sjfb8856606 	axgbe_enable_mtl_interrupts(pdata);
1190d30ea906Sjfb8856606 
1191d30ea906Sjfb8856606 	/* Initialize MAC related features */
1192*2d9fd380Sjfb8856606 	axgbe_config_mac_hash_table(pdata);
1193d30ea906Sjfb8856606 	axgbe_config_mac_address(pdata);
1194d30ea906Sjfb8856606 	axgbe_config_jumbo_enable(pdata);
1195d30ea906Sjfb8856606 	axgbe_config_flow_control(pdata);
1196d30ea906Sjfb8856606 	axgbe_config_mac_speed(pdata);
1197d30ea906Sjfb8856606 	axgbe_config_checksum_offload(pdata);
1198*2d9fd380Sjfb8856606 	axgbe_config_mmc(pdata);
1199d30ea906Sjfb8856606 
1200d30ea906Sjfb8856606 	return 0;
1201d30ea906Sjfb8856606 }
1202d30ea906Sjfb8856606 
axgbe_init_function_ptrs_dev(struct axgbe_hw_if * hw_if)1203d30ea906Sjfb8856606 void axgbe_init_function_ptrs_dev(struct axgbe_hw_if *hw_if)
1204d30ea906Sjfb8856606 {
1205d30ea906Sjfb8856606 	hw_if->exit = axgbe_exit;
1206d30ea906Sjfb8856606 	hw_if->config_flow_control = axgbe_config_flow_control;
1207d30ea906Sjfb8856606 
1208d30ea906Sjfb8856606 	hw_if->init = axgbe_init;
1209d30ea906Sjfb8856606 
1210d30ea906Sjfb8856606 	hw_if->read_mmd_regs = axgbe_read_mmd_regs;
1211d30ea906Sjfb8856606 	hw_if->write_mmd_regs = axgbe_write_mmd_regs;
1212d30ea906Sjfb8856606 
1213d30ea906Sjfb8856606 	hw_if->set_speed = axgbe_set_speed;
1214d30ea906Sjfb8856606 
1215d30ea906Sjfb8856606 	hw_if->set_ext_mii_mode = axgbe_set_ext_mii_mode;
1216d30ea906Sjfb8856606 	hw_if->read_ext_mii_regs = axgbe_read_ext_mii_regs;
1217d30ea906Sjfb8856606 	hw_if->write_ext_mii_regs = axgbe_write_ext_mii_regs;
1218d30ea906Sjfb8856606 	/* For FLOW ctrl */
1219d30ea906Sjfb8856606 	hw_if->config_tx_flow_control = axgbe_config_tx_flow_control;
1220d30ea906Sjfb8856606 	hw_if->config_rx_flow_control = axgbe_config_rx_flow_control;
1221d30ea906Sjfb8856606 }
1222