1 /* SPDX-License-Identifier: BSD-3-Clause
2 * Copyright(c) 2017 Intel Corporation
3 */
4
5 #include <stdio.h>
6 #include <errno.h>
7 #include <stdint.h>
8 #include <string.h>
9 #include <unistd.h>
10 #include <stdarg.h>
11 #include <inttypes.h>
12 #include <rte_byteorder.h>
13 #include <rte_common.h>
14
15 #include <rte_debug.h>
16 #include <rte_atomic.h>
17 #include <rte_eal.h>
18 #include <rte_ether.h>
19 #include <rte_ethdev_driver.h>
20 #include <rte_ethdev_pci.h>
21 #include <rte_dev.h>
22
23 #include "iavf.h"
24 #include "iavf_rxtx.h"
25
26 #define MAX_TRY_TIMES 200
27 #define ASQ_DELAY_MS 10
28
29 static uint32_t
iavf_convert_link_speed(enum virtchnl_link_speed virt_link_speed)30 iavf_convert_link_speed(enum virtchnl_link_speed virt_link_speed)
31 {
32 uint32_t speed;
33
34 switch (virt_link_speed) {
35 case VIRTCHNL_LINK_SPEED_100MB:
36 speed = 100;
37 break;
38 case VIRTCHNL_LINK_SPEED_1GB:
39 speed = 1000;
40 break;
41 case VIRTCHNL_LINK_SPEED_10GB:
42 speed = 10000;
43 break;
44 case VIRTCHNL_LINK_SPEED_40GB:
45 speed = 40000;
46 break;
47 case VIRTCHNL_LINK_SPEED_20GB:
48 speed = 20000;
49 break;
50 case VIRTCHNL_LINK_SPEED_25GB:
51 speed = 25000;
52 break;
53 case VIRTCHNL_LINK_SPEED_2_5GB:
54 speed = 2500;
55 break;
56 case VIRTCHNL_LINK_SPEED_5GB:
57 speed = 5000;
58 break;
59 default:
60 speed = 0;
61 break;
62 }
63
64 return speed;
65 }
66
67 /* Read data in admin queue to get msg from pf driver */
68 static enum iavf_aq_result
iavf_read_msg_from_pf(struct iavf_adapter * adapter,uint16_t buf_len,uint8_t * buf)69 iavf_read_msg_from_pf(struct iavf_adapter *adapter, uint16_t buf_len,
70 uint8_t *buf)
71 {
72 struct iavf_hw *hw = IAVF_DEV_PRIVATE_TO_HW(adapter);
73 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
74 struct rte_eth_dev *dev = adapter->eth_dev;
75 struct iavf_arq_event_info event;
76 enum iavf_aq_result result = IAVF_MSG_NON;
77 enum virtchnl_ops opcode;
78 int ret;
79
80 event.buf_len = buf_len;
81 event.msg_buf = buf;
82 ret = iavf_clean_arq_element(hw, &event, NULL);
83 /* Can't read any msg from adminQ */
84 if (ret) {
85 PMD_DRV_LOG(DEBUG, "Can't read msg from AQ");
86 if (ret != IAVF_ERR_ADMIN_QUEUE_NO_WORK)
87 result = IAVF_MSG_ERR;
88 return result;
89 }
90
91 opcode = (enum virtchnl_ops)rte_le_to_cpu_32(event.desc.cookie_high);
92 vf->cmd_retval = (enum virtchnl_status_code)rte_le_to_cpu_32(
93 event.desc.cookie_low);
94
95 PMD_DRV_LOG(DEBUG, "AQ from pf carries opcode %u, retval %d",
96 opcode, vf->cmd_retval);
97
98 if (opcode == VIRTCHNL_OP_EVENT) {
99 struct virtchnl_pf_event *vpe =
100 (struct virtchnl_pf_event *)event.msg_buf;
101
102 result = IAVF_MSG_SYS;
103 switch (vpe->event) {
104 case VIRTCHNL_EVENT_LINK_CHANGE:
105 vf->link_up =
106 vpe->event_data.link_event.link_status;
107 if (vf->vf_res->vf_cap_flags &
108 VIRTCHNL_VF_CAP_ADV_LINK_SPEED) {
109 vf->link_speed =
110 vpe->event_data.link_event_adv.link_speed;
111 } else {
112 enum virtchnl_link_speed speed;
113 speed = vpe->event_data.link_event.link_speed;
114 vf->link_speed = iavf_convert_link_speed(speed);
115 }
116 iavf_dev_link_update(dev, 0);
117 PMD_DRV_LOG(INFO, "Link status update:%s",
118 vf->link_up ? "up" : "down");
119 break;
120 case VIRTCHNL_EVENT_RESET_IMPENDING:
121 vf->vf_reset = true;
122 PMD_DRV_LOG(INFO, "VF is resetting");
123 break;
124 case VIRTCHNL_EVENT_PF_DRIVER_CLOSE:
125 vf->dev_closed = true;
126 PMD_DRV_LOG(INFO, "PF driver closed");
127 break;
128 default:
129 PMD_DRV_LOG(ERR, "%s: Unknown event %d from pf",
130 __func__, vpe->event);
131 }
132 } else {
133 /* async reply msg on command issued by vf previously */
134 result = IAVF_MSG_CMD;
135 if (opcode != vf->pend_cmd) {
136 PMD_DRV_LOG(WARNING, "command mismatch, expect %u, get %u",
137 vf->pend_cmd, opcode);
138 result = IAVF_MSG_ERR;
139 }
140 }
141
142 return result;
143 }
144
145 static int
iavf_execute_vf_cmd(struct iavf_adapter * adapter,struct iavf_cmd_info * args)146 iavf_execute_vf_cmd(struct iavf_adapter *adapter, struct iavf_cmd_info *args)
147 {
148 struct iavf_hw *hw = IAVF_DEV_PRIVATE_TO_HW(adapter);
149 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
150 enum iavf_aq_result result;
151 enum iavf_status ret;
152 int err = 0;
153 int i = 0;
154
155 if (vf->vf_reset)
156 return -EIO;
157
158 if (_atomic_set_cmd(vf, args->ops))
159 return -1;
160
161 ret = iavf_aq_send_msg_to_pf(hw, args->ops, IAVF_SUCCESS,
162 args->in_args, args->in_args_size, NULL);
163 if (ret) {
164 PMD_DRV_LOG(ERR, "fail to send cmd %d", args->ops);
165 _clear_cmd(vf);
166 return err;
167 }
168
169 switch (args->ops) {
170 case VIRTCHNL_OP_RESET_VF:
171 /*no need to wait for response */
172 _clear_cmd(vf);
173 break;
174 case VIRTCHNL_OP_VERSION:
175 case VIRTCHNL_OP_GET_VF_RESOURCES:
176 case VIRTCHNL_OP_GET_SUPPORTED_RXDIDS:
177 /* for init virtchnl ops, need to poll the response */
178 do {
179 result = iavf_read_msg_from_pf(adapter, args->out_size,
180 args->out_buffer);
181 if (result == IAVF_MSG_CMD)
182 break;
183 rte_delay_ms(ASQ_DELAY_MS);
184 } while (i++ < MAX_TRY_TIMES);
185 if (i >= MAX_TRY_TIMES ||
186 vf->cmd_retval != VIRTCHNL_STATUS_SUCCESS) {
187 err = -1;
188 PMD_DRV_LOG(ERR, "No response or return failure (%d)"
189 " for cmd %d", vf->cmd_retval, args->ops);
190 }
191 _clear_cmd(vf);
192 break;
193 case VIRTCHNL_OP_REQUEST_QUEUES:
194 /*
195 * ignore async reply, only wait for system message,
196 * vf_reset = true if get VIRTCHNL_EVENT_RESET_IMPENDING,
197 * if not, means request queues failed.
198 */
199 do {
200 result = iavf_read_msg_from_pf(adapter, args->out_size,
201 args->out_buffer);
202 if (result == IAVF_MSG_SYS && vf->vf_reset) {
203 break;
204 } else if (result == IAVF_MSG_CMD ||
205 result == IAVF_MSG_ERR) {
206 err = -1;
207 break;
208 }
209 rte_delay_ms(ASQ_DELAY_MS);
210 /* If don't read msg or read sys event, continue */
211 } while (i++ < MAX_TRY_TIMES);
212 if (i >= MAX_TRY_TIMES ||
213 vf->cmd_retval != VIRTCHNL_STATUS_SUCCESS) {
214 err = -1;
215 PMD_DRV_LOG(ERR, "No response or return failure (%d)"
216 " for cmd %d", vf->cmd_retval, args->ops);
217 }
218 _clear_cmd(vf);
219 break;
220 default:
221 /* For other virtchnl ops in running time,
222 * wait for the cmd done flag.
223 */
224 do {
225 if (vf->pend_cmd == VIRTCHNL_OP_UNKNOWN)
226 break;
227 rte_delay_ms(ASQ_DELAY_MS);
228 /* If don't read msg or read sys event, continue */
229 } while (i++ < MAX_TRY_TIMES);
230 /* If there's no response is received, clear command */
231 if (i >= MAX_TRY_TIMES ||
232 vf->cmd_retval != VIRTCHNL_STATUS_SUCCESS) {
233 err = -1;
234 PMD_DRV_LOG(ERR, "No response or return failure (%d)"
235 " for cmd %d", vf->cmd_retval, args->ops);
236 _clear_cmd(vf);
237 }
238 break;
239 }
240
241 return err;
242 }
243
244 static void
iavf_handle_pf_event_msg(struct rte_eth_dev * dev,uint8_t * msg,uint16_t msglen)245 iavf_handle_pf_event_msg(struct rte_eth_dev *dev, uint8_t *msg,
246 uint16_t msglen)
247 {
248 struct virtchnl_pf_event *pf_msg =
249 (struct virtchnl_pf_event *)msg;
250 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(dev->data->dev_private);
251
252 if (msglen < sizeof(struct virtchnl_pf_event)) {
253 PMD_DRV_LOG(DEBUG, "Error event");
254 return;
255 }
256 switch (pf_msg->event) {
257 case VIRTCHNL_EVENT_RESET_IMPENDING:
258 PMD_DRV_LOG(DEBUG, "VIRTCHNL_EVENT_RESET_IMPENDING event");
259 vf->vf_reset = true;
260 rte_eth_dev_callback_process(dev, RTE_ETH_EVENT_INTR_RESET,
261 NULL);
262 break;
263 case VIRTCHNL_EVENT_LINK_CHANGE:
264 PMD_DRV_LOG(DEBUG, "VIRTCHNL_EVENT_LINK_CHANGE event");
265 vf->link_up = pf_msg->event_data.link_event.link_status;
266 if (vf->vf_res->vf_cap_flags & VIRTCHNL_VF_CAP_ADV_LINK_SPEED) {
267 vf->link_speed =
268 pf_msg->event_data.link_event_adv.link_speed;
269 } else {
270 enum virtchnl_link_speed speed;
271 speed = pf_msg->event_data.link_event.link_speed;
272 vf->link_speed = iavf_convert_link_speed(speed);
273 }
274 iavf_dev_link_update(dev, 0);
275 rte_eth_dev_callback_process(dev, RTE_ETH_EVENT_INTR_LSC, NULL);
276 break;
277 case VIRTCHNL_EVENT_PF_DRIVER_CLOSE:
278 PMD_DRV_LOG(DEBUG, "VIRTCHNL_EVENT_PF_DRIVER_CLOSE event");
279 break;
280 default:
281 PMD_DRV_LOG(ERR, " unknown event received %u", pf_msg->event);
282 break;
283 }
284 }
285
286 void
iavf_handle_virtchnl_msg(struct rte_eth_dev * dev)287 iavf_handle_virtchnl_msg(struct rte_eth_dev *dev)
288 {
289 struct iavf_hw *hw = IAVF_DEV_PRIVATE_TO_HW(dev->data->dev_private);
290 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(dev->data->dev_private);
291 struct iavf_arq_event_info info;
292 uint16_t pending, aq_opc;
293 enum virtchnl_ops msg_opc;
294 enum iavf_status msg_ret;
295 int ret;
296
297 info.buf_len = IAVF_AQ_BUF_SZ;
298 if (!vf->aq_resp) {
299 PMD_DRV_LOG(ERR, "Buffer for adminq resp should not be NULL");
300 return;
301 }
302 info.msg_buf = vf->aq_resp;
303
304 pending = 1;
305 while (pending) {
306 ret = iavf_clean_arq_element(hw, &info, &pending);
307
308 if (ret != IAVF_SUCCESS) {
309 PMD_DRV_LOG(INFO, "Failed to read msg from AdminQ,"
310 "ret: %d", ret);
311 break;
312 }
313 aq_opc = rte_le_to_cpu_16(info.desc.opcode);
314 /* For the message sent from pf to vf, opcode is stored in
315 * cookie_high of struct iavf_aq_desc, while return error code
316 * are stored in cookie_low, Which is done by PF driver.
317 */
318 msg_opc = (enum virtchnl_ops)rte_le_to_cpu_32(
319 info.desc.cookie_high);
320 msg_ret = (enum iavf_status)rte_le_to_cpu_32(
321 info.desc.cookie_low);
322 switch (aq_opc) {
323 case iavf_aqc_opc_send_msg_to_vf:
324 if (msg_opc == VIRTCHNL_OP_EVENT) {
325 iavf_handle_pf_event_msg(dev, info.msg_buf,
326 info.msg_len);
327 } else {
328 /* read message and it's expected one */
329 if (msg_opc == vf->pend_cmd)
330 _notify_cmd(vf, msg_ret);
331 else
332 PMD_DRV_LOG(ERR, "command mismatch,"
333 "expect %u, get %u",
334 vf->pend_cmd, msg_opc);
335 PMD_DRV_LOG(DEBUG,
336 "adminq response is received,"
337 " opcode = %d", msg_opc);
338 }
339 break;
340 default:
341 PMD_DRV_LOG(DEBUG, "Request %u is not supported yet",
342 aq_opc);
343 break;
344 }
345 }
346 }
347
348 int
iavf_enable_vlan_strip(struct iavf_adapter * adapter)349 iavf_enable_vlan_strip(struct iavf_adapter *adapter)
350 {
351 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
352 struct iavf_cmd_info args;
353 int ret;
354
355 memset(&args, 0, sizeof(args));
356 args.ops = VIRTCHNL_OP_ENABLE_VLAN_STRIPPING;
357 args.in_args = NULL;
358 args.in_args_size = 0;
359 args.out_buffer = vf->aq_resp;
360 args.out_size = IAVF_AQ_BUF_SZ;
361 ret = iavf_execute_vf_cmd(adapter, &args);
362 if (ret)
363 PMD_DRV_LOG(ERR, "Failed to execute command of"
364 " OP_ENABLE_VLAN_STRIPPING");
365
366 return ret;
367 }
368
369 int
iavf_disable_vlan_strip(struct iavf_adapter * adapter)370 iavf_disable_vlan_strip(struct iavf_adapter *adapter)
371 {
372 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
373 struct iavf_cmd_info args;
374 int ret;
375
376 memset(&args, 0, sizeof(args));
377 args.ops = VIRTCHNL_OP_DISABLE_VLAN_STRIPPING;
378 args.in_args = NULL;
379 args.in_args_size = 0;
380 args.out_buffer = vf->aq_resp;
381 args.out_size = IAVF_AQ_BUF_SZ;
382 ret = iavf_execute_vf_cmd(adapter, &args);
383 if (ret)
384 PMD_DRV_LOG(ERR, "Failed to execute command of"
385 " OP_DISABLE_VLAN_STRIPPING");
386
387 return ret;
388 }
389
390 #define VIRTCHNL_VERSION_MAJOR_START 1
391 #define VIRTCHNL_VERSION_MINOR_START 1
392
393 /* Check API version with sync wait until version read from admin queue */
394 int
iavf_check_api_version(struct iavf_adapter * adapter)395 iavf_check_api_version(struct iavf_adapter *adapter)
396 {
397 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
398 struct virtchnl_version_info version, *pver;
399 struct iavf_cmd_info args;
400 int err;
401
402 version.major = VIRTCHNL_VERSION_MAJOR;
403 version.minor = VIRTCHNL_VERSION_MINOR;
404
405 args.ops = VIRTCHNL_OP_VERSION;
406 args.in_args = (uint8_t *)&version;
407 args.in_args_size = sizeof(version);
408 args.out_buffer = vf->aq_resp;
409 args.out_size = IAVF_AQ_BUF_SZ;
410
411 err = iavf_execute_vf_cmd(adapter, &args);
412 if (err) {
413 PMD_INIT_LOG(ERR, "Fail to execute command of OP_VERSION");
414 return err;
415 }
416
417 pver = (struct virtchnl_version_info *)args.out_buffer;
418 vf->virtchnl_version = *pver;
419
420 if (vf->virtchnl_version.major < VIRTCHNL_VERSION_MAJOR_START ||
421 (vf->virtchnl_version.major == VIRTCHNL_VERSION_MAJOR_START &&
422 vf->virtchnl_version.minor < VIRTCHNL_VERSION_MINOR_START)) {
423 PMD_INIT_LOG(ERR, "VIRTCHNL API version should not be lower"
424 " than (%u.%u) to support Adapative VF",
425 VIRTCHNL_VERSION_MAJOR_START,
426 VIRTCHNL_VERSION_MAJOR_START);
427 return -1;
428 } else if (vf->virtchnl_version.major > VIRTCHNL_VERSION_MAJOR ||
429 (vf->virtchnl_version.major == VIRTCHNL_VERSION_MAJOR &&
430 vf->virtchnl_version.minor > VIRTCHNL_VERSION_MINOR)) {
431 PMD_INIT_LOG(ERR, "PF/VF API version mismatch:(%u.%u)-(%u.%u)",
432 vf->virtchnl_version.major,
433 vf->virtchnl_version.minor,
434 VIRTCHNL_VERSION_MAJOR,
435 VIRTCHNL_VERSION_MINOR);
436 return -1;
437 }
438
439 PMD_DRV_LOG(DEBUG, "Peer is supported PF host");
440 return 0;
441 }
442
443 int
iavf_get_vf_resource(struct iavf_adapter * adapter)444 iavf_get_vf_resource(struct iavf_adapter *adapter)
445 {
446 struct iavf_hw *hw = IAVF_DEV_PRIVATE_TO_HW(adapter);
447 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
448 struct iavf_cmd_info args;
449 uint32_t caps, len;
450 int err, i;
451
452 args.ops = VIRTCHNL_OP_GET_VF_RESOURCES;
453 args.out_buffer = vf->aq_resp;
454 args.out_size = IAVF_AQ_BUF_SZ;
455
456 caps = IAVF_BASIC_OFFLOAD_CAPS | VIRTCHNL_VF_CAP_ADV_LINK_SPEED |
457 VIRTCHNL_VF_OFFLOAD_RX_FLEX_DESC |
458 VIRTCHNL_VF_OFFLOAD_FDIR_PF |
459 VIRTCHNL_VF_OFFLOAD_ADV_RSS_PF |
460 VIRTCHNL_VF_OFFLOAD_REQ_QUEUES |
461 VIRTCHNL_VF_LARGE_NUM_QPAIRS;
462
463 args.in_args = (uint8_t *)∩︀
464 args.in_args_size = sizeof(caps);
465
466 err = iavf_execute_vf_cmd(adapter, &args);
467
468 if (err) {
469 PMD_DRV_LOG(ERR,
470 "Failed to execute command of OP_GET_VF_RESOURCE");
471 return -1;
472 }
473
474 len = sizeof(struct virtchnl_vf_resource) +
475 IAVF_MAX_VF_VSI * sizeof(struct virtchnl_vsi_resource);
476
477 rte_memcpy(vf->vf_res, args.out_buffer,
478 RTE_MIN(args.out_size, len));
479 /* parse VF config message back from PF*/
480 iavf_vf_parse_hw_config(hw, vf->vf_res);
481 for (i = 0; i < vf->vf_res->num_vsis; i++) {
482 if (vf->vf_res->vsi_res[i].vsi_type == VIRTCHNL_VSI_SRIOV)
483 vf->vsi_res = &vf->vf_res->vsi_res[i];
484 }
485
486 if (!vf->vsi_res) {
487 PMD_INIT_LOG(ERR, "no LAN VSI found");
488 return -1;
489 }
490
491 vf->vsi.vsi_id = vf->vsi_res->vsi_id;
492 vf->vsi.nb_qps = vf->vsi_res->num_queue_pairs;
493 vf->vsi.adapter = adapter;
494
495 return 0;
496 }
497
498 int
iavf_get_supported_rxdid(struct iavf_adapter * adapter)499 iavf_get_supported_rxdid(struct iavf_adapter *adapter)
500 {
501 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
502 struct iavf_cmd_info args;
503 int ret;
504
505 args.ops = VIRTCHNL_OP_GET_SUPPORTED_RXDIDS;
506 args.in_args = NULL;
507 args.in_args_size = 0;
508 args.out_buffer = vf->aq_resp;
509 args.out_size = IAVF_AQ_BUF_SZ;
510
511 ret = iavf_execute_vf_cmd(adapter, &args);
512 if (ret) {
513 PMD_DRV_LOG(ERR,
514 "Failed to execute command of OP_GET_SUPPORTED_RXDIDS");
515 return ret;
516 }
517
518 vf->supported_rxdid =
519 ((struct virtchnl_supported_rxdids *)args.out_buffer)->supported_rxdids;
520
521 return 0;
522 }
523
524 int
iavf_enable_queues(struct iavf_adapter * adapter)525 iavf_enable_queues(struct iavf_adapter *adapter)
526 {
527 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
528 struct virtchnl_queue_select queue_select;
529 struct iavf_cmd_info args;
530 int err;
531
532 memset(&queue_select, 0, sizeof(queue_select));
533 queue_select.vsi_id = vf->vsi_res->vsi_id;
534
535 queue_select.rx_queues = BIT(adapter->eth_dev->data->nb_rx_queues) - 1;
536 queue_select.tx_queues = BIT(adapter->eth_dev->data->nb_tx_queues) - 1;
537
538 args.ops = VIRTCHNL_OP_ENABLE_QUEUES;
539 args.in_args = (u8 *)&queue_select;
540 args.in_args_size = sizeof(queue_select);
541 args.out_buffer = vf->aq_resp;
542 args.out_size = IAVF_AQ_BUF_SZ;
543 err = iavf_execute_vf_cmd(adapter, &args);
544 if (err) {
545 PMD_DRV_LOG(ERR,
546 "Failed to execute command of OP_ENABLE_QUEUES");
547 return err;
548 }
549 return 0;
550 }
551
552 int
iavf_disable_queues(struct iavf_adapter * adapter)553 iavf_disable_queues(struct iavf_adapter *adapter)
554 {
555 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
556 struct virtchnl_queue_select queue_select;
557 struct iavf_cmd_info args;
558 int err;
559
560 memset(&queue_select, 0, sizeof(queue_select));
561 queue_select.vsi_id = vf->vsi_res->vsi_id;
562
563 queue_select.rx_queues = BIT(adapter->eth_dev->data->nb_rx_queues) - 1;
564 queue_select.tx_queues = BIT(adapter->eth_dev->data->nb_tx_queues) - 1;
565
566 args.ops = VIRTCHNL_OP_DISABLE_QUEUES;
567 args.in_args = (u8 *)&queue_select;
568 args.in_args_size = sizeof(queue_select);
569 args.out_buffer = vf->aq_resp;
570 args.out_size = IAVF_AQ_BUF_SZ;
571 err = iavf_execute_vf_cmd(adapter, &args);
572 if (err) {
573 PMD_DRV_LOG(ERR,
574 "Failed to execute command of OP_DISABLE_QUEUES");
575 return err;
576 }
577 return 0;
578 }
579
580 int
iavf_switch_queue(struct iavf_adapter * adapter,uint16_t qid,bool rx,bool on)581 iavf_switch_queue(struct iavf_adapter *adapter, uint16_t qid,
582 bool rx, bool on)
583 {
584 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
585 struct virtchnl_queue_select queue_select;
586 struct iavf_cmd_info args;
587 int err;
588
589 memset(&queue_select, 0, sizeof(queue_select));
590 queue_select.vsi_id = vf->vsi_res->vsi_id;
591 if (rx)
592 queue_select.rx_queues |= 1 << qid;
593 else
594 queue_select.tx_queues |= 1 << qid;
595
596 if (on)
597 args.ops = VIRTCHNL_OP_ENABLE_QUEUES;
598 else
599 args.ops = VIRTCHNL_OP_DISABLE_QUEUES;
600 args.in_args = (u8 *)&queue_select;
601 args.in_args_size = sizeof(queue_select);
602 args.out_buffer = vf->aq_resp;
603 args.out_size = IAVF_AQ_BUF_SZ;
604 err = iavf_execute_vf_cmd(adapter, &args);
605 if (err)
606 PMD_DRV_LOG(ERR, "Failed to execute command of %s",
607 on ? "OP_ENABLE_QUEUES" : "OP_DISABLE_QUEUES");
608 return err;
609 }
610
611 int
iavf_enable_queues_lv(struct iavf_adapter * adapter)612 iavf_enable_queues_lv(struct iavf_adapter *adapter)
613 {
614 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
615 struct virtchnl_del_ena_dis_queues *queue_select;
616 struct virtchnl_queue_chunk *queue_chunk;
617 struct iavf_cmd_info args;
618 int err, len;
619
620 len = sizeof(struct virtchnl_del_ena_dis_queues) +
621 sizeof(struct virtchnl_queue_chunk) *
622 (IAVF_RXTX_QUEUE_CHUNKS_NUM - 1);
623 queue_select = rte_zmalloc("queue_select", len, 0);
624 if (!queue_select)
625 return -ENOMEM;
626
627 queue_chunk = queue_select->chunks.chunks;
628 queue_select->chunks.num_chunks = IAVF_RXTX_QUEUE_CHUNKS_NUM;
629 queue_select->vport_id = vf->vsi_res->vsi_id;
630
631 queue_chunk[VIRTCHNL_QUEUE_TYPE_TX].type = VIRTCHNL_QUEUE_TYPE_TX;
632 queue_chunk[VIRTCHNL_QUEUE_TYPE_TX].start_queue_id = 0;
633 queue_chunk[VIRTCHNL_QUEUE_TYPE_TX].num_queues =
634 adapter->eth_dev->data->nb_tx_queues;
635
636 queue_chunk[VIRTCHNL_QUEUE_TYPE_RX].type = VIRTCHNL_QUEUE_TYPE_RX;
637 queue_chunk[VIRTCHNL_QUEUE_TYPE_RX].start_queue_id = 0;
638 queue_chunk[VIRTCHNL_QUEUE_TYPE_RX].num_queues =
639 adapter->eth_dev->data->nb_rx_queues;
640
641 args.ops = VIRTCHNL_OP_ENABLE_QUEUES_V2;
642 args.in_args = (u8 *)queue_select;
643 args.in_args_size = len;
644 args.out_buffer = vf->aq_resp;
645 args.out_size = IAVF_AQ_BUF_SZ;
646 err = iavf_execute_vf_cmd(adapter, &args);
647 if (err) {
648 PMD_DRV_LOG(ERR,
649 "Failed to execute command of OP_ENABLE_QUEUES_V2");
650 return err;
651 }
652 return 0;
653 }
654
655 int
iavf_disable_queues_lv(struct iavf_adapter * adapter)656 iavf_disable_queues_lv(struct iavf_adapter *adapter)
657 {
658 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
659 struct virtchnl_del_ena_dis_queues *queue_select;
660 struct virtchnl_queue_chunk *queue_chunk;
661 struct iavf_cmd_info args;
662 int err, len;
663
664 len = sizeof(struct virtchnl_del_ena_dis_queues) +
665 sizeof(struct virtchnl_queue_chunk) *
666 (IAVF_RXTX_QUEUE_CHUNKS_NUM - 1);
667 queue_select = rte_zmalloc("queue_select", len, 0);
668 if (!queue_select)
669 return -ENOMEM;
670
671 queue_chunk = queue_select->chunks.chunks;
672 queue_select->chunks.num_chunks = IAVF_RXTX_QUEUE_CHUNKS_NUM;
673 queue_select->vport_id = vf->vsi_res->vsi_id;
674
675 queue_chunk[VIRTCHNL_QUEUE_TYPE_TX].type = VIRTCHNL_QUEUE_TYPE_TX;
676 queue_chunk[VIRTCHNL_QUEUE_TYPE_TX].start_queue_id = 0;
677 queue_chunk[VIRTCHNL_QUEUE_TYPE_TX].num_queues =
678 adapter->eth_dev->data->nb_tx_queues;
679
680 queue_chunk[VIRTCHNL_QUEUE_TYPE_RX].type = VIRTCHNL_QUEUE_TYPE_RX;
681 queue_chunk[VIRTCHNL_QUEUE_TYPE_RX].start_queue_id = 0;
682 queue_chunk[VIRTCHNL_QUEUE_TYPE_RX].num_queues =
683 adapter->eth_dev->data->nb_rx_queues;
684
685 args.ops = VIRTCHNL_OP_DISABLE_QUEUES_V2;
686 args.in_args = (u8 *)queue_select;
687 args.in_args_size = len;
688 args.out_buffer = vf->aq_resp;
689 args.out_size = IAVF_AQ_BUF_SZ;
690 err = iavf_execute_vf_cmd(adapter, &args);
691 if (err) {
692 PMD_DRV_LOG(ERR,
693 "Failed to execute command of OP_DISABLE_QUEUES_V2");
694 return err;
695 }
696 return 0;
697 }
698
699 int
iavf_switch_queue_lv(struct iavf_adapter * adapter,uint16_t qid,bool rx,bool on)700 iavf_switch_queue_lv(struct iavf_adapter *adapter, uint16_t qid,
701 bool rx, bool on)
702 {
703 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
704 struct virtchnl_del_ena_dis_queues *queue_select;
705 struct virtchnl_queue_chunk *queue_chunk;
706 struct iavf_cmd_info args;
707 int err, len;
708
709 len = sizeof(struct virtchnl_del_ena_dis_queues);
710 queue_select = rte_zmalloc("queue_select", len, 0);
711 if (!queue_select)
712 return -ENOMEM;
713
714 queue_chunk = queue_select->chunks.chunks;
715 queue_select->chunks.num_chunks = 1;
716 queue_select->vport_id = vf->vsi_res->vsi_id;
717
718 if (rx) {
719 queue_chunk->type = VIRTCHNL_QUEUE_TYPE_RX;
720 queue_chunk->start_queue_id = qid;
721 queue_chunk->num_queues = 1;
722 } else {
723 queue_chunk->type = VIRTCHNL_QUEUE_TYPE_TX;
724 queue_chunk->start_queue_id = qid;
725 queue_chunk->num_queues = 1;
726 }
727
728 if (on)
729 args.ops = VIRTCHNL_OP_ENABLE_QUEUES_V2;
730 else
731 args.ops = VIRTCHNL_OP_DISABLE_QUEUES_V2;
732 args.in_args = (u8 *)queue_select;
733 args.in_args_size = len;
734 args.out_buffer = vf->aq_resp;
735 args.out_size = IAVF_AQ_BUF_SZ;
736 err = iavf_execute_vf_cmd(adapter, &args);
737 if (err)
738 PMD_DRV_LOG(ERR, "Failed to execute command of %s",
739 on ? "OP_ENABLE_QUEUES_V2" : "OP_DISABLE_QUEUES_V2");
740 return err;
741 }
742
743 int
iavf_configure_rss_lut(struct iavf_adapter * adapter)744 iavf_configure_rss_lut(struct iavf_adapter *adapter)
745 {
746 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
747 struct virtchnl_rss_lut *rss_lut;
748 struct iavf_cmd_info args;
749 int len, err = 0;
750
751 len = sizeof(*rss_lut) + vf->vf_res->rss_lut_size - 1;
752 rss_lut = rte_zmalloc("rss_lut", len, 0);
753 if (!rss_lut)
754 return -ENOMEM;
755
756 rss_lut->vsi_id = vf->vsi_res->vsi_id;
757 rss_lut->lut_entries = vf->vf_res->rss_lut_size;
758 rte_memcpy(rss_lut->lut, vf->rss_lut, vf->vf_res->rss_lut_size);
759
760 args.ops = VIRTCHNL_OP_CONFIG_RSS_LUT;
761 args.in_args = (u8 *)rss_lut;
762 args.in_args_size = len;
763 args.out_buffer = vf->aq_resp;
764 args.out_size = IAVF_AQ_BUF_SZ;
765
766 err = iavf_execute_vf_cmd(adapter, &args);
767 if (err)
768 PMD_DRV_LOG(ERR,
769 "Failed to execute command of OP_CONFIG_RSS_LUT");
770
771 rte_free(rss_lut);
772 return err;
773 }
774
775 int
iavf_configure_rss_key(struct iavf_adapter * adapter)776 iavf_configure_rss_key(struct iavf_adapter *adapter)
777 {
778 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
779 struct virtchnl_rss_key *rss_key;
780 struct iavf_cmd_info args;
781 int len, err = 0;
782
783 len = sizeof(*rss_key) + vf->vf_res->rss_key_size - 1;
784 rss_key = rte_zmalloc("rss_key", len, 0);
785 if (!rss_key)
786 return -ENOMEM;
787
788 rss_key->vsi_id = vf->vsi_res->vsi_id;
789 rss_key->key_len = vf->vf_res->rss_key_size;
790 rte_memcpy(rss_key->key, vf->rss_key, vf->vf_res->rss_key_size);
791
792 args.ops = VIRTCHNL_OP_CONFIG_RSS_KEY;
793 args.in_args = (u8 *)rss_key;
794 args.in_args_size = len;
795 args.out_buffer = vf->aq_resp;
796 args.out_size = IAVF_AQ_BUF_SZ;
797
798 err = iavf_execute_vf_cmd(adapter, &args);
799 if (err)
800 PMD_DRV_LOG(ERR,
801 "Failed to execute command of OP_CONFIG_RSS_KEY");
802
803 rte_free(rss_key);
804 return err;
805 }
806
807 int
iavf_configure_queues(struct iavf_adapter * adapter,uint16_t num_queue_pairs,uint16_t index)808 iavf_configure_queues(struct iavf_adapter *adapter,
809 uint16_t num_queue_pairs, uint16_t index)
810 {
811 struct iavf_rx_queue **rxq =
812 (struct iavf_rx_queue **)adapter->eth_dev->data->rx_queues;
813 struct iavf_tx_queue **txq =
814 (struct iavf_tx_queue **)adapter->eth_dev->data->tx_queues;
815 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
816 struct virtchnl_vsi_queue_config_info *vc_config;
817 struct virtchnl_queue_pair_info *vc_qp;
818 struct iavf_cmd_info args;
819 uint16_t i, size;
820 int err;
821
822 size = sizeof(*vc_config) +
823 sizeof(vc_config->qpair[0]) * num_queue_pairs;
824 vc_config = rte_zmalloc("cfg_queue", size, 0);
825 if (!vc_config)
826 return -ENOMEM;
827
828 vc_config->vsi_id = vf->vsi_res->vsi_id;
829 vc_config->num_queue_pairs = num_queue_pairs;
830
831 for (i = index, vc_qp = vc_config->qpair;
832 i < index + num_queue_pairs;
833 i++, vc_qp++) {
834 vc_qp->txq.vsi_id = vf->vsi_res->vsi_id;
835 vc_qp->txq.queue_id = i;
836
837 /* Virtchnnl configure tx queues by pairs */
838 if (i < adapter->eth_dev->data->nb_tx_queues) {
839 vc_qp->txq.ring_len = txq[i]->nb_tx_desc;
840 vc_qp->txq.dma_ring_addr = txq[i]->tx_ring_phys_addr;
841 }
842
843 vc_qp->rxq.vsi_id = vf->vsi_res->vsi_id;
844 vc_qp->rxq.queue_id = i;
845 vc_qp->rxq.max_pkt_size = vf->max_pkt_len;
846
847 if (i >= adapter->eth_dev->data->nb_rx_queues)
848 continue;
849
850 /* Virtchnnl configure rx queues by pairs */
851 vc_qp->rxq.ring_len = rxq[i]->nb_rx_desc;
852 vc_qp->rxq.dma_ring_addr = rxq[i]->rx_ring_phys_addr;
853 vc_qp->rxq.databuffer_size = rxq[i]->rx_buf_len;
854
855 #ifndef RTE_LIBRTE_IAVF_16BYTE_RX_DESC
856 if (vf->vf_res->vf_cap_flags &
857 VIRTCHNL_VF_OFFLOAD_RX_FLEX_DESC &&
858 vf->supported_rxdid & BIT(rxq[i]->rxdid)) {
859 vc_qp->rxq.rxdid = rxq[i]->rxdid;
860 PMD_DRV_LOG(NOTICE, "request RXDID[%d] in Queue[%d]",
861 vc_qp->rxq.rxdid, i);
862 } else {
863 PMD_DRV_LOG(NOTICE, "RXDID[%d] is not supported, "
864 "request default RXDID[%d] in Queue[%d]",
865 rxq[i]->rxdid, IAVF_RXDID_LEGACY_1, i);
866 vc_qp->rxq.rxdid = IAVF_RXDID_LEGACY_1;
867 }
868 #else
869 if (vf->vf_res->vf_cap_flags &
870 VIRTCHNL_VF_OFFLOAD_RX_FLEX_DESC &&
871 vf->supported_rxdid & BIT(IAVF_RXDID_LEGACY_0)) {
872 vc_qp->rxq.rxdid = IAVF_RXDID_LEGACY_0;
873 PMD_DRV_LOG(NOTICE, "request RXDID[%d] in Queue[%d]",
874 vc_qp->rxq.rxdid, i);
875 } else {
876 PMD_DRV_LOG(ERR, "RXDID[%d] is not supported",
877 IAVF_RXDID_LEGACY_0);
878 return -1;
879 }
880 #endif
881 }
882
883 memset(&args, 0, sizeof(args));
884 args.ops = VIRTCHNL_OP_CONFIG_VSI_QUEUES;
885 args.in_args = (uint8_t *)vc_config;
886 args.in_args_size = size;
887 args.out_buffer = vf->aq_resp;
888 args.out_size = IAVF_AQ_BUF_SZ;
889
890 err = iavf_execute_vf_cmd(adapter, &args);
891 if (err)
892 PMD_DRV_LOG(ERR, "Failed to execute command of"
893 " VIRTCHNL_OP_CONFIG_VSI_QUEUES");
894
895 rte_free(vc_config);
896 return err;
897 }
898
899 int
iavf_config_irq_map(struct iavf_adapter * adapter)900 iavf_config_irq_map(struct iavf_adapter *adapter)
901 {
902 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
903 struct virtchnl_irq_map_info *map_info;
904 struct virtchnl_vector_map *vecmap;
905 struct iavf_cmd_info args;
906 int len, i, err;
907
908 len = sizeof(struct virtchnl_irq_map_info) +
909 sizeof(struct virtchnl_vector_map) * vf->nb_msix;
910
911 map_info = rte_zmalloc("map_info", len, 0);
912 if (!map_info)
913 return -ENOMEM;
914
915 map_info->num_vectors = vf->nb_msix;
916 for (i = 0; i < adapter->eth_dev->data->nb_rx_queues; i++) {
917 vecmap =
918 &map_info->vecmap[vf->qv_map[i].vector_id - vf->msix_base];
919 vecmap->vsi_id = vf->vsi_res->vsi_id;
920 vecmap->rxitr_idx = IAVF_ITR_INDEX_DEFAULT;
921 vecmap->vector_id = vf->qv_map[i].vector_id;
922 vecmap->txq_map = 0;
923 vecmap->rxq_map |= 1 << vf->qv_map[i].queue_id;
924 }
925
926 args.ops = VIRTCHNL_OP_CONFIG_IRQ_MAP;
927 args.in_args = (u8 *)map_info;
928 args.in_args_size = len;
929 args.out_buffer = vf->aq_resp;
930 args.out_size = IAVF_AQ_BUF_SZ;
931 err = iavf_execute_vf_cmd(adapter, &args);
932 if (err)
933 PMD_DRV_LOG(ERR, "fail to execute command OP_CONFIG_IRQ_MAP");
934
935 rte_free(map_info);
936 return err;
937 }
938
939 int
iavf_config_irq_map_lv(struct iavf_adapter * adapter,uint16_t num,uint16_t index)940 iavf_config_irq_map_lv(struct iavf_adapter *adapter, uint16_t num,
941 uint16_t index)
942 {
943 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
944 struct virtchnl_queue_vector_maps *map_info;
945 struct virtchnl_queue_vector *qv_maps;
946 struct iavf_cmd_info args;
947 int len, i, err;
948 int count = 0;
949
950 len = sizeof(struct virtchnl_queue_vector_maps) +
951 sizeof(struct virtchnl_queue_vector) * (num - 1);
952
953 map_info = rte_zmalloc("map_info", len, 0);
954 if (!map_info)
955 return -ENOMEM;
956
957 map_info->vport_id = vf->vsi_res->vsi_id;
958 map_info->num_qv_maps = num;
959 for (i = index; i < index + map_info->num_qv_maps; i++) {
960 qv_maps = &map_info->qv_maps[count++];
961 qv_maps->itr_idx = VIRTCHNL_ITR_IDX_0;
962 qv_maps->queue_type = VIRTCHNL_QUEUE_TYPE_RX;
963 qv_maps->queue_id = vf->qv_map[i].queue_id;
964 qv_maps->vector_id = vf->qv_map[i].vector_id;
965 }
966
967 args.ops = VIRTCHNL_OP_MAP_QUEUE_VECTOR;
968 args.in_args = (u8 *)map_info;
969 args.in_args_size = len;
970 args.out_buffer = vf->aq_resp;
971 args.out_size = IAVF_AQ_BUF_SZ;
972 err = iavf_execute_vf_cmd(adapter, &args);
973 if (err)
974 PMD_DRV_LOG(ERR, "fail to execute command OP_MAP_QUEUE_VECTOR");
975
976 rte_free(map_info);
977 return err;
978 }
979
980 void
iavf_add_del_all_mac_addr(struct iavf_adapter * adapter,bool add)981 iavf_add_del_all_mac_addr(struct iavf_adapter *adapter, bool add)
982 {
983 struct virtchnl_ether_addr_list *list;
984 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
985 struct rte_ether_addr *addr;
986 struct iavf_cmd_info args;
987 int len, err, i, j;
988 int next_begin = 0;
989 int begin = 0;
990
991 do {
992 j = 0;
993 len = sizeof(struct virtchnl_ether_addr_list);
994 for (i = begin; i < IAVF_NUM_MACADDR_MAX; i++, next_begin++) {
995 addr = &adapter->eth_dev->data->mac_addrs[i];
996 if (rte_is_zero_ether_addr(addr))
997 continue;
998 len += sizeof(struct virtchnl_ether_addr);
999 if (len >= IAVF_AQ_BUF_SZ) {
1000 next_begin = i + 1;
1001 break;
1002 }
1003 }
1004
1005 list = rte_zmalloc("iavf_del_mac_buffer", len, 0);
1006 if (!list) {
1007 PMD_DRV_LOG(ERR, "fail to allocate memory");
1008 return;
1009 }
1010
1011 for (i = begin; i < next_begin; i++) {
1012 addr = &adapter->eth_dev->data->mac_addrs[i];
1013 if (rte_is_zero_ether_addr(addr))
1014 continue;
1015 rte_memcpy(list->list[j].addr, addr->addr_bytes,
1016 sizeof(addr->addr_bytes));
1017 PMD_DRV_LOG(DEBUG, "add/rm mac:%x:%x:%x:%x:%x:%x",
1018 addr->addr_bytes[0], addr->addr_bytes[1],
1019 addr->addr_bytes[2], addr->addr_bytes[3],
1020 addr->addr_bytes[4], addr->addr_bytes[5]);
1021 j++;
1022 }
1023 list->vsi_id = vf->vsi_res->vsi_id;
1024 list->num_elements = j;
1025 args.ops = add ? VIRTCHNL_OP_ADD_ETH_ADDR :
1026 VIRTCHNL_OP_DEL_ETH_ADDR;
1027 args.in_args = (uint8_t *)list;
1028 args.in_args_size = len;
1029 args.out_buffer = vf->aq_resp;
1030 args.out_size = IAVF_AQ_BUF_SZ;
1031 err = iavf_execute_vf_cmd(adapter, &args);
1032 if (err)
1033 PMD_DRV_LOG(ERR, "fail to execute command %s",
1034 add ? "OP_ADD_ETHER_ADDRESS" :
1035 "OP_DEL_ETHER_ADDRESS");
1036 rte_free(list);
1037 begin = next_begin;
1038 } while (begin < IAVF_NUM_MACADDR_MAX);
1039 }
1040
1041 int
iavf_query_stats(struct iavf_adapter * adapter,struct virtchnl_eth_stats ** pstats)1042 iavf_query_stats(struct iavf_adapter *adapter,
1043 struct virtchnl_eth_stats **pstats)
1044 {
1045 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
1046 struct virtchnl_queue_select q_stats;
1047 struct iavf_cmd_info args;
1048 int err;
1049
1050 memset(&q_stats, 0, sizeof(q_stats));
1051 q_stats.vsi_id = vf->vsi_res->vsi_id;
1052 args.ops = VIRTCHNL_OP_GET_STATS;
1053 args.in_args = (uint8_t *)&q_stats;
1054 args.in_args_size = sizeof(q_stats);
1055 args.out_buffer = vf->aq_resp;
1056 args.out_size = IAVF_AQ_BUF_SZ;
1057
1058 err = iavf_execute_vf_cmd(adapter, &args);
1059 if (err) {
1060 PMD_DRV_LOG(ERR, "fail to execute command OP_GET_STATS");
1061 *pstats = NULL;
1062 return err;
1063 }
1064 *pstats = (struct virtchnl_eth_stats *)args.out_buffer;
1065 return 0;
1066 }
1067
1068 int
iavf_config_promisc(struct iavf_adapter * adapter,bool enable_unicast,bool enable_multicast)1069 iavf_config_promisc(struct iavf_adapter *adapter,
1070 bool enable_unicast,
1071 bool enable_multicast)
1072 {
1073 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
1074 struct virtchnl_promisc_info promisc;
1075 struct iavf_cmd_info args;
1076 int err;
1077
1078 promisc.flags = 0;
1079 promisc.vsi_id = vf->vsi_res->vsi_id;
1080
1081 if (enable_unicast)
1082 promisc.flags |= FLAG_VF_UNICAST_PROMISC;
1083
1084 if (enable_multicast)
1085 promisc.flags |= FLAG_VF_MULTICAST_PROMISC;
1086
1087 args.ops = VIRTCHNL_OP_CONFIG_PROMISCUOUS_MODE;
1088 args.in_args = (uint8_t *)&promisc;
1089 args.in_args_size = sizeof(promisc);
1090 args.out_buffer = vf->aq_resp;
1091 args.out_size = IAVF_AQ_BUF_SZ;
1092
1093 err = iavf_execute_vf_cmd(adapter, &args);
1094
1095 if (err) {
1096 PMD_DRV_LOG(ERR,
1097 "fail to execute command CONFIG_PROMISCUOUS_MODE");
1098
1099 if (err == IAVF_NOT_SUPPORTED)
1100 return -ENOTSUP;
1101
1102 return -EAGAIN;
1103 }
1104
1105 vf->promisc_unicast_enabled = enable_unicast;
1106 vf->promisc_multicast_enabled = enable_multicast;
1107 return 0;
1108 }
1109
1110 int
iavf_add_del_eth_addr(struct iavf_adapter * adapter,struct rte_ether_addr * addr,bool add)1111 iavf_add_del_eth_addr(struct iavf_adapter *adapter, struct rte_ether_addr *addr,
1112 bool add)
1113 {
1114 struct virtchnl_ether_addr_list *list;
1115 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
1116 uint8_t cmd_buffer[sizeof(struct virtchnl_ether_addr_list) +
1117 sizeof(struct virtchnl_ether_addr)];
1118 struct iavf_cmd_info args;
1119 int err;
1120
1121 list = (struct virtchnl_ether_addr_list *)cmd_buffer;
1122 list->vsi_id = vf->vsi_res->vsi_id;
1123 list->num_elements = 1;
1124 rte_memcpy(list->list[0].addr, addr->addr_bytes,
1125 sizeof(addr->addr_bytes));
1126
1127 args.ops = add ? VIRTCHNL_OP_ADD_ETH_ADDR : VIRTCHNL_OP_DEL_ETH_ADDR;
1128 args.in_args = cmd_buffer;
1129 args.in_args_size = sizeof(cmd_buffer);
1130 args.out_buffer = vf->aq_resp;
1131 args.out_size = IAVF_AQ_BUF_SZ;
1132 err = iavf_execute_vf_cmd(adapter, &args);
1133 if (err)
1134 PMD_DRV_LOG(ERR, "fail to execute command %s",
1135 add ? "OP_ADD_ETH_ADDR" : "OP_DEL_ETH_ADDR");
1136 return err;
1137 }
1138
1139 int
iavf_add_del_vlan(struct iavf_adapter * adapter,uint16_t vlanid,bool add)1140 iavf_add_del_vlan(struct iavf_adapter *adapter, uint16_t vlanid, bool add)
1141 {
1142 struct virtchnl_vlan_filter_list *vlan_list;
1143 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
1144 uint8_t cmd_buffer[sizeof(struct virtchnl_vlan_filter_list) +
1145 sizeof(uint16_t)];
1146 struct iavf_cmd_info args;
1147 int err;
1148
1149 vlan_list = (struct virtchnl_vlan_filter_list *)cmd_buffer;
1150 vlan_list->vsi_id = vf->vsi_res->vsi_id;
1151 vlan_list->num_elements = 1;
1152 vlan_list->vlan_id[0] = vlanid;
1153
1154 args.ops = add ? VIRTCHNL_OP_ADD_VLAN : VIRTCHNL_OP_DEL_VLAN;
1155 args.in_args = cmd_buffer;
1156 args.in_args_size = sizeof(cmd_buffer);
1157 args.out_buffer = vf->aq_resp;
1158 args.out_size = IAVF_AQ_BUF_SZ;
1159 err = iavf_execute_vf_cmd(adapter, &args);
1160 if (err)
1161 PMD_DRV_LOG(ERR, "fail to execute command %s",
1162 add ? "OP_ADD_VLAN" : "OP_DEL_VLAN");
1163
1164 return err;
1165 }
1166
1167 int
iavf_fdir_add(struct iavf_adapter * adapter,struct iavf_fdir_conf * filter)1168 iavf_fdir_add(struct iavf_adapter *adapter,
1169 struct iavf_fdir_conf *filter)
1170 {
1171 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
1172 struct virtchnl_fdir_add *fdir_ret;
1173
1174 struct iavf_cmd_info args;
1175 int err;
1176
1177 filter->add_fltr.vsi_id = vf->vsi_res->vsi_id;
1178 filter->add_fltr.validate_only = 0;
1179
1180 args.ops = VIRTCHNL_OP_ADD_FDIR_FILTER;
1181 args.in_args = (uint8_t *)(&filter->add_fltr);
1182 args.in_args_size = sizeof(*(&filter->add_fltr));
1183 args.out_buffer = vf->aq_resp;
1184 args.out_size = IAVF_AQ_BUF_SZ;
1185
1186 err = iavf_execute_vf_cmd(adapter, &args);
1187 if (err) {
1188 PMD_DRV_LOG(ERR, "fail to execute command OP_ADD_FDIR_FILTER");
1189 return err;
1190 }
1191
1192 fdir_ret = (struct virtchnl_fdir_add *)args.out_buffer;
1193 filter->flow_id = fdir_ret->flow_id;
1194
1195 if (fdir_ret->status == VIRTCHNL_FDIR_SUCCESS) {
1196 PMD_DRV_LOG(INFO,
1197 "Succeed in adding rule request by PF");
1198 } else if (fdir_ret->status == VIRTCHNL_FDIR_FAILURE_RULE_NORESOURCE) {
1199 PMD_DRV_LOG(ERR,
1200 "Failed to add rule request due to no hw resource");
1201 return -1;
1202 } else if (fdir_ret->status == VIRTCHNL_FDIR_FAILURE_RULE_EXIST) {
1203 PMD_DRV_LOG(ERR,
1204 "Failed to add rule request due to the rule is already existed");
1205 return -1;
1206 } else if (fdir_ret->status == VIRTCHNL_FDIR_FAILURE_RULE_CONFLICT) {
1207 PMD_DRV_LOG(ERR,
1208 "Failed to add rule request due to the rule is conflict with existing rule");
1209 return -1;
1210 } else if (fdir_ret->status == VIRTCHNL_FDIR_FAILURE_RULE_INVALID) {
1211 PMD_DRV_LOG(ERR,
1212 "Failed to add rule request due to the hw doesn't support");
1213 return -1;
1214 } else if (fdir_ret->status == VIRTCHNL_FDIR_FAILURE_RULE_TIMEOUT) {
1215 PMD_DRV_LOG(ERR,
1216 "Failed to add rule request due to time out for programming");
1217 return -1;
1218 } else {
1219 PMD_DRV_LOG(ERR,
1220 "Failed to add rule request due to other reasons");
1221 return -1;
1222 }
1223
1224 return 0;
1225 };
1226
1227 int
iavf_fdir_del(struct iavf_adapter * adapter,struct iavf_fdir_conf * filter)1228 iavf_fdir_del(struct iavf_adapter *adapter,
1229 struct iavf_fdir_conf *filter)
1230 {
1231 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
1232 struct virtchnl_fdir_del *fdir_ret;
1233
1234 struct iavf_cmd_info args;
1235 int err;
1236
1237 filter->del_fltr.vsi_id = vf->vsi_res->vsi_id;
1238 filter->del_fltr.flow_id = filter->flow_id;
1239
1240 args.ops = VIRTCHNL_OP_DEL_FDIR_FILTER;
1241 args.in_args = (uint8_t *)(&filter->del_fltr);
1242 args.in_args_size = sizeof(filter->del_fltr);
1243 args.out_buffer = vf->aq_resp;
1244 args.out_size = IAVF_AQ_BUF_SZ;
1245
1246 err = iavf_execute_vf_cmd(adapter, &args);
1247 if (err) {
1248 PMD_DRV_LOG(ERR, "fail to execute command OP_DEL_FDIR_FILTER");
1249 return err;
1250 }
1251
1252 fdir_ret = (struct virtchnl_fdir_del *)args.out_buffer;
1253
1254 if (fdir_ret->status == VIRTCHNL_FDIR_SUCCESS) {
1255 PMD_DRV_LOG(INFO,
1256 "Succeed in deleting rule request by PF");
1257 } else if (fdir_ret->status == VIRTCHNL_FDIR_FAILURE_RULE_NONEXIST) {
1258 PMD_DRV_LOG(ERR,
1259 "Failed to delete rule request due to this rule doesn't exist");
1260 return -1;
1261 } else if (fdir_ret->status == VIRTCHNL_FDIR_FAILURE_RULE_TIMEOUT) {
1262 PMD_DRV_LOG(ERR,
1263 "Failed to delete rule request due to time out for programming");
1264 return -1;
1265 } else {
1266 PMD_DRV_LOG(ERR,
1267 "Failed to delete rule request due to other reasons");
1268 return -1;
1269 }
1270
1271 return 0;
1272 };
1273
1274 int
iavf_fdir_check(struct iavf_adapter * adapter,struct iavf_fdir_conf * filter)1275 iavf_fdir_check(struct iavf_adapter *adapter,
1276 struct iavf_fdir_conf *filter)
1277 {
1278 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
1279 struct virtchnl_fdir_add *fdir_ret;
1280
1281 struct iavf_cmd_info args;
1282 int err;
1283
1284 filter->add_fltr.vsi_id = vf->vsi_res->vsi_id;
1285 filter->add_fltr.validate_only = 1;
1286
1287 args.ops = VIRTCHNL_OP_ADD_FDIR_FILTER;
1288 args.in_args = (uint8_t *)(&filter->add_fltr);
1289 args.in_args_size = sizeof(*(&filter->add_fltr));
1290 args.out_buffer = vf->aq_resp;
1291 args.out_size = IAVF_AQ_BUF_SZ;
1292
1293 err = iavf_execute_vf_cmd(adapter, &args);
1294 if (err) {
1295 PMD_DRV_LOG(ERR, "fail to check flow direcotor rule");
1296 return err;
1297 }
1298
1299 fdir_ret = (struct virtchnl_fdir_add *)args.out_buffer;
1300
1301 if (fdir_ret->status == VIRTCHNL_FDIR_SUCCESS) {
1302 PMD_DRV_LOG(INFO,
1303 "Succeed in checking rule request by PF");
1304 } else if (fdir_ret->status == VIRTCHNL_FDIR_FAILURE_RULE_INVALID) {
1305 PMD_DRV_LOG(ERR,
1306 "Failed to check rule request due to parameters validation"
1307 " or HW doesn't support");
1308 return -1;
1309 } else {
1310 PMD_DRV_LOG(ERR,
1311 "Failed to check rule request due to other reasons");
1312 return -1;
1313 }
1314
1315 return 0;
1316 }
1317
1318 int
iavf_add_del_rss_cfg(struct iavf_adapter * adapter,struct virtchnl_rss_cfg * rss_cfg,bool add)1319 iavf_add_del_rss_cfg(struct iavf_adapter *adapter,
1320 struct virtchnl_rss_cfg *rss_cfg, bool add)
1321 {
1322 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
1323 struct iavf_cmd_info args;
1324 int err;
1325
1326 memset(&args, 0, sizeof(args));
1327 args.ops = add ? VIRTCHNL_OP_ADD_RSS_CFG :
1328 VIRTCHNL_OP_DEL_RSS_CFG;
1329 args.in_args = (u8 *)rss_cfg;
1330 args.in_args_size = sizeof(*rss_cfg);
1331 args.out_buffer = vf->aq_resp;
1332 args.out_size = IAVF_AQ_BUF_SZ;
1333
1334 err = iavf_execute_vf_cmd(adapter, &args);
1335 if (err)
1336 PMD_DRV_LOG(ERR,
1337 "Failed to execute command of %s",
1338 add ? "OP_ADD_RSS_CFG" :
1339 "OP_DEL_RSS_INPUT_CFG");
1340
1341 return err;
1342 }
1343
1344 int
iavf_add_del_mc_addr_list(struct iavf_adapter * adapter,struct rte_ether_addr * mc_addrs,uint32_t mc_addrs_num,bool add)1345 iavf_add_del_mc_addr_list(struct iavf_adapter *adapter,
1346 struct rte_ether_addr *mc_addrs,
1347 uint32_t mc_addrs_num, bool add)
1348 {
1349 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
1350 uint8_t cmd_buffer[sizeof(struct virtchnl_ether_addr_list) +
1351 (IAVF_NUM_MACADDR_MAX * sizeof(struct virtchnl_ether_addr))];
1352 struct virtchnl_ether_addr_list *list;
1353 struct iavf_cmd_info args;
1354 uint32_t i;
1355 int err;
1356
1357 if (mc_addrs == NULL || mc_addrs_num == 0)
1358 return 0;
1359
1360 list = (struct virtchnl_ether_addr_list *)cmd_buffer;
1361 list->vsi_id = vf->vsi_res->vsi_id;
1362 list->num_elements = mc_addrs_num;
1363
1364 for (i = 0; i < mc_addrs_num; i++) {
1365 if (!IAVF_IS_MULTICAST(mc_addrs[i].addr_bytes)) {
1366 PMD_DRV_LOG(ERR, "Invalid mac:%x:%x:%x:%x:%x:%x",
1367 mc_addrs[i].addr_bytes[0],
1368 mc_addrs[i].addr_bytes[1],
1369 mc_addrs[i].addr_bytes[2],
1370 mc_addrs[i].addr_bytes[3],
1371 mc_addrs[i].addr_bytes[4],
1372 mc_addrs[i].addr_bytes[5]);
1373 return -EINVAL;
1374 }
1375
1376 memcpy(list->list[i].addr, mc_addrs[i].addr_bytes,
1377 sizeof(list->list[i].addr));
1378 }
1379
1380 args.ops = add ? VIRTCHNL_OP_ADD_ETH_ADDR : VIRTCHNL_OP_DEL_ETH_ADDR;
1381 args.in_args = cmd_buffer;
1382 args.in_args_size = sizeof(struct virtchnl_ether_addr_list) +
1383 i * sizeof(struct virtchnl_ether_addr);
1384 args.out_buffer = vf->aq_resp;
1385 args.out_size = IAVF_AQ_BUF_SZ;
1386 err = iavf_execute_vf_cmd(adapter, &args);
1387
1388 if (err) {
1389 PMD_DRV_LOG(ERR, "fail to execute command %s",
1390 add ? "OP_ADD_ETH_ADDR" : "OP_DEL_ETH_ADDR");
1391 return err;
1392 }
1393
1394 return 0;
1395 }
1396
1397 int
iavf_request_queues(struct iavf_adapter * adapter,uint16_t num)1398 iavf_request_queues(struct iavf_adapter *adapter, uint16_t num)
1399 {
1400 struct rte_eth_dev *dev = adapter->eth_dev;
1401 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
1402 struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(dev);
1403 struct virtchnl_vf_res_request vfres;
1404 struct iavf_cmd_info args;
1405 uint16_t num_queue_pairs;
1406 int err;
1407
1408 if (!(vf->vf_res->vf_cap_flags &
1409 VIRTCHNL_VF_OFFLOAD_REQ_QUEUES)) {
1410 PMD_DRV_LOG(ERR, "request queues not supported");
1411 return -1;
1412 }
1413
1414 if (num == 0) {
1415 PMD_DRV_LOG(ERR, "queue number cannot be zero");
1416 return -1;
1417 }
1418 vfres.num_queue_pairs = num;
1419
1420 args.ops = VIRTCHNL_OP_REQUEST_QUEUES;
1421 args.in_args = (u8 *)&vfres;
1422 args.in_args_size = sizeof(vfres);
1423 args.out_buffer = vf->aq_resp;
1424 args.out_size = IAVF_AQ_BUF_SZ;
1425
1426 /*
1427 * disable interrupt to avoid the admin queue message to be read
1428 * before iavf_read_msg_from_pf.
1429 */
1430 rte_intr_disable(&pci_dev->intr_handle);
1431 err = iavf_execute_vf_cmd(adapter, &args);
1432 rte_intr_enable(&pci_dev->intr_handle);
1433 if (err) {
1434 PMD_DRV_LOG(ERR, "fail to execute command OP_REQUEST_QUEUES");
1435 return err;
1436 }
1437
1438 /* request queues succeeded, vf is resetting */
1439 if (vf->vf_reset) {
1440 PMD_DRV_LOG(INFO, "vf is resetting");
1441 return 0;
1442 }
1443
1444 /* request additional queues failed, return available number */
1445 num_queue_pairs =
1446 ((struct virtchnl_vf_res_request *)args.out_buffer)->num_queue_pairs;
1447 PMD_DRV_LOG(ERR, "request queues failed, only %u queues "
1448 "available", num_queue_pairs);
1449
1450 return -1;
1451 }
1452
1453 int
iavf_get_max_rss_queue_region(struct iavf_adapter * adapter)1454 iavf_get_max_rss_queue_region(struct iavf_adapter *adapter)
1455 {
1456 struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
1457 struct iavf_cmd_info args;
1458 uint16_t qregion_width;
1459 int err;
1460
1461 args.ops = VIRTCHNL_OP_GET_MAX_RSS_QREGION;
1462 args.in_args = NULL;
1463 args.in_args_size = 0;
1464 args.out_buffer = vf->aq_resp;
1465 args.out_size = IAVF_AQ_BUF_SZ;
1466
1467 err = iavf_execute_vf_cmd(adapter, &args);
1468 if (err) {
1469 PMD_DRV_LOG(ERR, "Failed to execute command of VIRTCHNL_OP_GET_MAX_RSS_QREGION");
1470 return err;
1471 }
1472
1473 qregion_width =
1474 ((struct virtchnl_max_rss_qregion *)args.out_buffer)->qregion_width;
1475
1476 vf->max_rss_qregion = (uint16_t)(1 << qregion_width);
1477
1478 return 0;
1479 }
1480