xref: /dpdk/drivers/event/cnxk/cnxk_eventdev.h (revision 68c05095)
1 /* SPDX-License-Identifier: BSD-3-Clause
2  * Copyright(C) 2021 Marvell.
3  */
4 
5 #ifndef __CNXK_EVENTDEV_H__
6 #define __CNXK_EVENTDEV_H__
7 
8 #include <string.h>
9 
10 #include <cryptodev_pmd.h>
11 #include <rte_devargs.h>
12 #include <rte_ethdev.h>
13 #include <rte_event_eth_rx_adapter.h>
14 #include <rte_event_eth_tx_adapter.h>
15 #include <rte_kvargs.h>
16 #include <rte_mbuf_pool_ops.h>
17 #include <rte_pci.h>
18 
19 #include <eventdev_pmd_pci.h>
20 
21 #include "roc_api.h"
22 
23 #include "cnxk_tim_evdev.h"
24 
25 #define CNXK_SSO_XAE_CNT   "xae_cnt"
26 #define CNXK_SSO_GGRP_QOS  "qos"
27 #define CNXK_SSO_FORCE_BP  "force_rx_bp"
28 #define CN9K_SSO_SINGLE_WS "single_ws"
29 #define CN10K_SSO_GW_MODE  "gw_mode"
30 
31 #define NSEC2USEC(__ns)		((__ns) / 1E3)
32 #define USEC2NSEC(__us)		((__us)*1E3)
33 #define NSEC2TICK(__ns, __freq) (((__ns) * (__freq)) / 1E9)
34 
35 #define CNXK_SSO_MAX_HWGRP     (RTE_EVENT_MAX_QUEUES_PER_DEV + 1)
36 #define CNXK_SSO_FC_NAME       "cnxk_evdev_xaq_fc"
37 #define CNXK_SSO_MZ_NAME       "cnxk_evdev_mz"
38 #define CNXK_SSO_XAQ_CACHE_CNT (0x7)
39 #define CNXK_SSO_XAQ_SLACK     (8)
40 #define CNXK_SSO_WQE_SG_PTR    (9)
41 #define CNXK_SSO_PRIORITY_CNT  (0x8)
42 #define CNXK_SSO_WEIGHT_MAX    (0x3f)
43 #define CNXK_SSO_WEIGHT_MIN    (0x3)
44 #define CNXK_SSO_WEIGHT_CNT    (CNXK_SSO_WEIGHT_MAX - CNXK_SSO_WEIGHT_MIN + 1)
45 #define CNXK_SSO_AFFINITY_CNT  (0x10)
46 
47 #define CNXK_TT_FROM_TAG(x)	    (((x) >> 32) & SSO_TT_EMPTY)
48 #define CNXK_TT_FROM_EVENT(x)	    (((x) >> 38) & SSO_TT_EMPTY)
49 #define CNXK_EVENT_TYPE_FROM_TAG(x) (((x) >> 28) & 0xf)
50 #define CNXK_SUB_EVENT_FROM_TAG(x)  (((x) >> 20) & 0xff)
51 #define CNXK_CLR_SUB_EVENT(x)	    (~(0xffull << 20) & x)
52 #define CNXK_GRP_FROM_TAG(x)	    (((x) >> 36) & 0x3ff)
53 #define CNXK_SWTAG_PEND(x)	    (BIT_ULL(62) & x)
54 #define CNXK_TAG_IS_HEAD(x)	    (BIT_ULL(35) & x)
55 
56 #define CN9K_SSOW_GET_BASE_ADDR(_GW) ((_GW)-SSOW_LF_GWS_OP_GET_WORK0)
57 
58 #define CN10K_GW_MODE_NONE     0
59 #define CN10K_GW_MODE_PREF     1
60 #define CN10K_GW_MODE_PREF_WFE 2
61 
62 #define CNXK_QOS_NORMALIZE(val, min, max, cnt)                                 \
63 	(min + val / ((max + cnt - 1) / cnt))
64 #define CNXK_SSO_FLUSH_RETRY_MAX 0xfff
65 
66 #define CNXK_VALID_DEV_OR_ERR_RET(dev, drv_name)                               \
67 	do {                                                                   \
68 		if (strncmp(dev->driver->name, drv_name, strlen(drv_name)))    \
69 			return -EINVAL;                                        \
70 	} while (0)
71 
72 typedef void *(*cnxk_sso_init_hws_mem_t)(void *dev, uint8_t port_id);
73 typedef void (*cnxk_sso_hws_setup_t)(void *dev, void *ws, uintptr_t grp_base);
74 typedef void (*cnxk_sso_hws_release_t)(void *dev, void *ws);
75 typedef int (*cnxk_sso_link_t)(void *dev, void *ws, uint16_t *map,
76 			       uint16_t nb_link);
77 typedef int (*cnxk_sso_unlink_t)(void *dev, void *ws, uint16_t *map,
78 				 uint16_t nb_link);
79 typedef void (*cnxk_handle_event_t)(void *arg, struct rte_event ev);
80 typedef void (*cnxk_sso_hws_reset_t)(void *arg, void *ws);
81 typedef int (*cnxk_sso_hws_flush_t)(void *ws, uint8_t queue_id, uintptr_t base,
82 				    cnxk_handle_event_t fn, void *arg);
83 
84 struct cnxk_sso_qos {
85 	uint16_t queue;
86 	uint16_t xaq_prcnt;
87 	uint16_t taq_prcnt;
88 	uint16_t iaq_prcnt;
89 };
90 
91 struct cnxk_sso_mlt_prio {
92 	uint8_t weight;
93 	uint8_t affinity;
94 };
95 
96 struct cnxk_sso_evdev {
97 	struct roc_sso sso;
98 	uint8_t max_event_queues;
99 	uint8_t max_event_ports;
100 	uint8_t is_timeout_deq;
101 	uint8_t nb_event_queues;
102 	uint8_t nb_event_ports;
103 	uint8_t configured;
104 	uint32_t deq_tmo_ns;
105 	uint32_t min_dequeue_timeout_ns;
106 	uint32_t max_dequeue_timeout_ns;
107 	int32_t max_num_events;
108 	uint64_t xaq_lmt;
109 	rte_iova_t fc_iova;
110 	uint64_t rx_offloads;
111 	uint64_t tx_offloads;
112 	uint64_t adptr_xae_cnt;
113 	uint16_t rx_adptr_pool_cnt;
114 	uint64_t *rx_adptr_pools;
115 	uint64_t *tx_adptr_data;
116 	size_t tx_adptr_data_sz;
117 	uint16_t max_port_id;
118 	uint16_t max_queue_id[RTE_MAX_ETHPORTS];
119 	uint8_t tx_adptr_configured;
120 	uint16_t tim_adptr_ring_cnt;
121 	uint16_t *timer_adptr_rings;
122 	uint64_t *timer_adptr_sz;
123 	uint16_t vec_pool_cnt;
124 	uint64_t *vec_pools;
125 	struct cnxk_sso_mlt_prio mlt_prio[RTE_EVENT_MAX_QUEUES_PER_DEV];
126 	/* Dev args */
127 	uint32_t xae_cnt;
128 	uint8_t qos_queue_cnt;
129 	struct cnxk_sso_qos *qos_parse_data;
130 	uint8_t force_ena_bp;
131 	/* CN9K */
132 	uint8_t dual_ws;
133 	/* CN10K */
134 	uint8_t gw_mode;
135 	/* Crypto adapter */
136 	uint8_t is_ca_internal_port;
137 } __rte_cache_aligned;
138 
139 struct cn10k_sso_hws {
140 	uint64_t base;
141 	uint64_t gw_rdata;
142 	/* PTP timestamp */
143 	struct cnxk_timesync_info *tstamp;
144 	void *lookup_mem;
145 	uint32_t gw_wdata;
146 	uint8_t swtag_req;
147 	uint8_t hws_id;
148 	/* Add Work Fastpath data */
149 	uint64_t xaq_lmt __rte_cache_aligned;
150 	uint64_t *fc_mem;
151 	uintptr_t grp_base;
152 	/* Tx Fastpath data */
153 	uintptr_t lmt_base __rte_cache_aligned;
154 	uint64_t lso_tun_fmt;
155 	uint8_t tx_adptr_data[];
156 } __rte_cache_aligned;
157 
158 /* Event port a.k.a GWS */
159 struct cn9k_sso_hws {
160 	uint64_t base;
161 	uint64_t gw_wdata;
162 	/* PTP timestamp */
163 	struct cnxk_timesync_info *tstamp;
164 	void *lookup_mem;
165 	uint8_t swtag_req;
166 	uint8_t hws_id;
167 	/* Add Work Fastpath data */
168 	uint64_t xaq_lmt __rte_cache_aligned;
169 	uint64_t *fc_mem;
170 	uintptr_t grp_base;
171 	/* Tx Fastpath data */
172 	uint64_t lso_tun_fmt __rte_cache_aligned;
173 	uint8_t tx_adptr_data[];
174 } __rte_cache_aligned;
175 
176 struct cn9k_sso_hws_dual {
177 	uint64_t base[2]; /* Ping and Pong */
178 	uint64_t gw_wdata;
179 	/* PTP timestamp */
180 	struct cnxk_timesync_info *tstamp;
181 	void *lookup_mem;
182 	uint8_t swtag_req;
183 	uint8_t vws; /* Ping pong bit */
184 	uint8_t hws_id;
185 	/* Add Work Fastpath data */
186 	uint64_t xaq_lmt __rte_cache_aligned;
187 	uint64_t *fc_mem;
188 	uintptr_t grp_base;
189 	/* Tx Fastpath data */
190 	uint64_t lso_tun_fmt __rte_cache_aligned;
191 	uint8_t tx_adptr_data[];
192 } __rte_cache_aligned;
193 
194 struct cnxk_sso_hws_cookie {
195 	const struct rte_eventdev *event_dev;
196 	bool configured;
197 } __rte_cache_aligned;
198 
199 static inline int
parse_kvargs_flag(const char * key,const char * value,void * opaque)200 parse_kvargs_flag(const char *key, const char *value, void *opaque)
201 {
202 	RTE_SET_USED(key);
203 
204 	*(uint8_t *)opaque = !!atoi(value);
205 	return 0;
206 }
207 
208 static inline int
parse_kvargs_value(const char * key,const char * value,void * opaque)209 parse_kvargs_value(const char *key, const char *value, void *opaque)
210 {
211 	RTE_SET_USED(key);
212 
213 	*(uint32_t *)opaque = (uint32_t)atoi(value);
214 	return 0;
215 }
216 
217 static inline struct cnxk_sso_evdev *
cnxk_sso_pmd_priv(const struct rte_eventdev * event_dev)218 cnxk_sso_pmd_priv(const struct rte_eventdev *event_dev)
219 {
220 	return event_dev->data->dev_private;
221 }
222 
223 static inline struct cnxk_sso_hws_cookie *
cnxk_sso_hws_get_cookie(void * ws)224 cnxk_sso_hws_get_cookie(void *ws)
225 {
226 	return RTE_PTR_SUB(ws, sizeof(struct cnxk_sso_hws_cookie));
227 }
228 
229 /* Configuration functions */
230 int cnxk_sso_xae_reconfigure(struct rte_eventdev *event_dev);
231 int cnxk_sso_xaq_allocate(struct cnxk_sso_evdev *dev);
232 void cnxk_sso_updt_xae_cnt(struct cnxk_sso_evdev *dev, void *data,
233 			   uint32_t event_type);
234 
235 /* Common ops API. */
236 int cnxk_sso_init(struct rte_eventdev *event_dev);
237 int cnxk_sso_fini(struct rte_eventdev *event_dev);
238 int cnxk_sso_remove(struct rte_pci_device *pci_dev);
239 void cnxk_sso_info_get(struct cnxk_sso_evdev *dev,
240 		       struct rte_event_dev_info *dev_info);
241 int cnxk_sso_dev_validate(const struct rte_eventdev *event_dev);
242 int cnxk_setup_event_ports(const struct rte_eventdev *event_dev,
243 			   cnxk_sso_init_hws_mem_t init_hws_mem,
244 			   cnxk_sso_hws_setup_t hws_setup);
245 void cnxk_sso_restore_links(const struct rte_eventdev *event_dev,
246 			    cnxk_sso_link_t link_fn);
247 void cnxk_sso_queue_def_conf(struct rte_eventdev *event_dev, uint8_t queue_id,
248 			     struct rte_event_queue_conf *queue_conf);
249 int cnxk_sso_queue_setup(struct rte_eventdev *event_dev, uint8_t queue_id,
250 			 const struct rte_event_queue_conf *queue_conf);
251 void cnxk_sso_queue_release(struct rte_eventdev *event_dev, uint8_t queue_id);
252 int cnxk_sso_queue_attribute_get(struct rte_eventdev *event_dev,
253 				 uint8_t queue_id, uint32_t attr_id,
254 				 uint32_t *attr_value);
255 int cnxk_sso_queue_attribute_set(struct rte_eventdev *event_dev,
256 				 uint8_t queue_id, uint32_t attr_id,
257 				 uint64_t attr_value);
258 void cnxk_sso_port_def_conf(struct rte_eventdev *event_dev, uint8_t port_id,
259 			    struct rte_event_port_conf *port_conf);
260 int cnxk_sso_port_setup(struct rte_eventdev *event_dev, uint8_t port_id,
261 			cnxk_sso_hws_setup_t hws_setup_fn);
262 int cnxk_sso_timeout_ticks(struct rte_eventdev *event_dev, uint64_t ns,
263 			   uint64_t *tmo_ticks);
264 int cnxk_sso_start(struct rte_eventdev *event_dev,
265 		   cnxk_sso_hws_reset_t reset_fn,
266 		   cnxk_sso_hws_flush_t flush_fn);
267 void cnxk_sso_stop(struct rte_eventdev *event_dev,
268 		   cnxk_sso_hws_reset_t reset_fn,
269 		   cnxk_sso_hws_flush_t flush_fn);
270 int cnxk_sso_close(struct rte_eventdev *event_dev, cnxk_sso_unlink_t unlink_fn);
271 int cnxk_sso_selftest(const char *dev_name);
272 void cnxk_sso_dump(struct rte_eventdev *event_dev, FILE *f);
273 
274 /* Stats API. */
275 int cnxk_sso_xstats_get_names(const struct rte_eventdev *event_dev,
276 			      enum rte_event_dev_xstats_mode mode,
277 			      uint8_t queue_port_id,
278 			      struct rte_event_dev_xstats_name *xstats_names,
279 			      unsigned int *ids, unsigned int size);
280 int cnxk_sso_xstats_get(const struct rte_eventdev *event_dev,
281 			enum rte_event_dev_xstats_mode mode,
282 			uint8_t queue_port_id, const unsigned int ids[],
283 			uint64_t values[], unsigned int n);
284 int cnxk_sso_xstats_reset(struct rte_eventdev *event_dev,
285 			  enum rte_event_dev_xstats_mode mode,
286 			  int16_t queue_port_id, const uint32_t ids[],
287 			  uint32_t n);
288 
289 /* Crypto adapter APIs. */
290 int cnxk_crypto_adapter_qp_add(const struct rte_eventdev *event_dev,
291 			       const struct rte_cryptodev *cdev,
292 			       int32_t queue_pair_id);
293 int cnxk_crypto_adapter_qp_del(const struct rte_cryptodev *cdev,
294 			       int32_t queue_pair_id);
295 
296 /* CN9K */
297 void cn9k_sso_set_rsrc(void *arg);
298 
299 /* Common adapter ops */
300 int cnxk_sso_rx_adapter_queue_add(
301 	const struct rte_eventdev *event_dev, const struct rte_eth_dev *eth_dev,
302 	int32_t rx_queue_id,
303 	const struct rte_event_eth_rx_adapter_queue_conf *queue_conf);
304 int cnxk_sso_rx_adapter_queue_del(const struct rte_eventdev *event_dev,
305 				  const struct rte_eth_dev *eth_dev,
306 				  int32_t rx_queue_id);
307 int cnxk_sso_rx_adapter_start(const struct rte_eventdev *event_dev,
308 			      const struct rte_eth_dev *eth_dev);
309 int cnxk_sso_rx_adapter_stop(const struct rte_eventdev *event_dev,
310 			     const struct rte_eth_dev *eth_dev);
311 int cnxk_sso_tx_adapter_queue_add(const struct rte_eventdev *event_dev,
312 				  const struct rte_eth_dev *eth_dev,
313 				  int32_t tx_queue_id);
314 int cnxk_sso_tx_adapter_queue_del(const struct rte_eventdev *event_dev,
315 				  const struct rte_eth_dev *eth_dev,
316 				  int32_t tx_queue_id);
317 
318 #endif /* __CNXK_EVENTDEV_H__ */
319