Lines Matching refs:dev

61 			(rte_event_devices[i].dev ? (strncmp(  in rte_event_dev_get_dev_id()
62 rte_event_devices[i].dev->driver->name, name, in rte_event_dev_get_dev_id()
74 struct rte_eventdev *dev; in rte_event_dev_socket_id() local
77 dev = &rte_eventdevs[dev_id]; in rte_event_dev_socket_id()
79 return dev->data->socket_id; in rte_event_dev_socket_id()
85 struct rte_eventdev *dev; in rte_event_dev_info_get() local
88 dev = &rte_eventdevs[dev_id]; in rte_event_dev_info_get()
95 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->dev_infos_get, -ENOTSUP); in rte_event_dev_info_get()
96 (*dev->dev_ops->dev_infos_get)(dev, dev_info); in rte_event_dev_info_get()
98 dev_info->dequeue_timeout_ns = dev->data->dev_conf.dequeue_timeout_ns; in rte_event_dev_info_get()
100 dev_info->dev = dev->dev; in rte_event_dev_info_get()
108 struct rte_eventdev *dev; in rte_event_eth_rx_adapter_caps_get() local
113 dev = &rte_eventdevs[dev_id]; in rte_event_eth_rx_adapter_caps_get()
118 if (dev->dev_ops->eth_rx_adapter_caps_get == NULL) in rte_event_eth_rx_adapter_caps_get()
123 return dev->dev_ops->eth_rx_adapter_caps_get ? in rte_event_eth_rx_adapter_caps_get()
124 (*dev->dev_ops->eth_rx_adapter_caps_get)(dev, in rte_event_eth_rx_adapter_caps_get()
133 struct rte_eventdev *dev; in rte_event_timer_adapter_caps_get() local
138 dev = &rte_eventdevs[dev_id]; in rte_event_timer_adapter_caps_get()
144 return dev->dev_ops->timer_adapter_caps_get ? in rte_event_timer_adapter_caps_get()
145 (*dev->dev_ops->timer_adapter_caps_get)(dev, in rte_event_timer_adapter_caps_get()
156 struct rte_eventdev *dev; in rte_event_crypto_adapter_caps_get() local
163 dev = &rte_eventdevs[dev_id]; in rte_event_crypto_adapter_caps_get()
169 if (dev->dev_ops->crypto_adapter_caps_get == NULL) in rte_event_crypto_adapter_caps_get()
174 return dev->dev_ops->crypto_adapter_caps_get ? in rte_event_crypto_adapter_caps_get()
175 (*dev->dev_ops->crypto_adapter_caps_get) in rte_event_crypto_adapter_caps_get()
176 (dev, cdev, caps) : 0; in rte_event_crypto_adapter_caps_get()
183 struct rte_eventdev *dev; in rte_event_eth_tx_adapter_caps_get() local
189 dev = &rte_eventdevs[dev_id]; in rte_event_eth_tx_adapter_caps_get()
195 if (dev->dev_ops->eth_tx_adapter_caps_get == NULL) in rte_event_eth_tx_adapter_caps_get()
200 return dev->dev_ops->eth_tx_adapter_caps_get ? in rte_event_eth_tx_adapter_caps_get()
201 (*dev->dev_ops->eth_tx_adapter_caps_get)(dev, in rte_event_eth_tx_adapter_caps_get()
208 event_dev_queue_config(struct rte_eventdev *dev, uint8_t nb_queues) in event_dev_queue_config() argument
210 uint8_t old_nb_queues = dev->data->nb_queues; in event_dev_queue_config()
215 dev->data->dev_id); in event_dev_queue_config()
218 queues_cfg = dev->data->queues_cfg; in event_dev_queue_config()
219 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->queue_release, -ENOTSUP); in event_dev_queue_config()
222 (*dev->dev_ops->queue_release)(dev, i); in event_dev_queue_config()
232 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->queue_release, -ENOTSUP); in event_dev_queue_config()
235 (*dev->dev_ops->queue_release)(dev, i); in event_dev_queue_config()
238 dev->data->nb_queues = nb_queues; in event_dev_queue_config()
245 event_dev_port_config(struct rte_eventdev *dev, uint8_t nb_ports) in event_dev_port_config() argument
247 uint8_t old_nb_ports = dev->data->nb_ports; in event_dev_port_config()
254 dev->data->dev_id); in event_dev_port_config()
257 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->port_release, -ENOTSUP); in event_dev_port_config()
259 ports = dev->data->ports; in event_dev_port_config()
260 ports_cfg = dev->data->ports_cfg; in event_dev_port_config()
261 links_map = dev->data->links_map; in event_dev_port_config()
264 (*dev->dev_ops->port_release)(ports[i]); in event_dev_port_config()
282 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->port_release, -ENOTSUP); in event_dev_port_config()
284 ports = dev->data->ports; in event_dev_port_config()
286 (*dev->dev_ops->port_release)(ports[i]); in event_dev_port_config()
291 dev->data->nb_ports = nb_ports; in event_dev_port_config()
300 struct rte_eventdev *dev; in rte_event_dev_configure() local
304 dev = &rte_eventdevs[dev_id]; in rte_event_dev_configure()
306 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->dev_infos_get, -ENOTSUP); in rte_event_dev_configure()
307 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->dev_configure, -ENOTSUP); in rte_event_dev_configure()
309 if (dev->data->dev_started) { in rte_event_dev_configure()
318 (*dev->dev_ops->dev_infos_get)(dev, &info); in rte_event_dev_configure()
450 memcpy(&dev->data->dev_conf, dev_conf, sizeof(dev->data->dev_conf)); in rte_event_dev_configure()
453 diag = event_dev_queue_config(dev, dev_conf->nb_event_queues); in rte_event_dev_configure()
461 diag = event_dev_port_config(dev, dev_conf->nb_event_ports); in rte_event_dev_configure()
463 event_dev_queue_config(dev, 0); in rte_event_dev_configure()
472 diag = (*dev->dev_ops->dev_configure)(dev); in rte_event_dev_configure()
476 event_dev_queue_config(dev, 0); in rte_event_dev_configure()
477 event_dev_port_config(dev, 0); in rte_event_dev_configure()
480 dev->data->event_dev_cap = info.event_dev_cap; in rte_event_dev_configure()
486 is_valid_queue(struct rte_eventdev *dev, uint8_t queue_id) in is_valid_queue() argument
488 if (queue_id < dev->data->nb_queues && queue_id < in is_valid_queue()
499 struct rte_eventdev *dev; in rte_event_queue_default_conf_get() local
502 dev = &rte_eventdevs[dev_id]; in rte_event_queue_default_conf_get()
507 if (!is_valid_queue(dev, queue_id)) { in rte_event_queue_default_conf_get()
512 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->queue_def_conf, -ENOTSUP); in rte_event_queue_default_conf_get()
514 (*dev->dev_ops->queue_def_conf)(dev, queue_id, queue_conf); in rte_event_queue_default_conf_get()
555 struct rte_eventdev *dev; in rte_event_queue_setup() local
559 dev = &rte_eventdevs[dev_id]; in rte_event_queue_setup()
561 if (!is_valid_queue(dev, queue_id)) { in rte_event_queue_setup()
570 dev->data->dev_conf.nb_event_queue_flows) { in rte_event_queue_setup()
574 dev->data->dev_conf.nb_event_queue_flows); in rte_event_queue_setup()
583 dev->data->dev_conf.nb_event_queue_flows) { in rte_event_queue_setup()
587 dev->data->dev_conf.nb_event_queue_flows); in rte_event_queue_setup()
592 if (dev->data->dev_started) { in rte_event_queue_setup()
598 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->queue_setup, -ENOTSUP); in rte_event_queue_setup()
601 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->queue_def_conf, in rte_event_queue_setup()
603 (*dev->dev_ops->queue_def_conf)(dev, queue_id, &def_conf); in rte_event_queue_setup()
607 dev->data->queues_cfg[queue_id] = *queue_conf; in rte_event_queue_setup()
609 return (*dev->dev_ops->queue_setup)(dev, queue_id, queue_conf); in rte_event_queue_setup()
613 is_valid_port(struct rte_eventdev *dev, uint8_t port_id) in is_valid_port() argument
615 if (port_id < dev->data->nb_ports) in is_valid_port()
625 struct rte_eventdev *dev; in rte_event_port_default_conf_get() local
628 dev = &rte_eventdevs[dev_id]; in rte_event_port_default_conf_get()
633 if (!is_valid_port(dev, port_id)) { in rte_event_port_default_conf_get()
638 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->port_def_conf, -ENOTSUP); in rte_event_port_default_conf_get()
640 (*dev->dev_ops->port_def_conf)(dev, port_id, port_conf); in rte_event_port_default_conf_get()
648 struct rte_eventdev *dev; in rte_event_port_setup() local
653 dev = &rte_eventdevs[dev_id]; in rte_event_port_setup()
655 if (!is_valid_port(dev, port_id)) { in rte_event_port_setup()
663 dev->data->dev_conf.nb_events_limit)) { in rte_event_port_setup()
667 dev->data->dev_conf.nb_events_limit); in rte_event_port_setup()
674 dev->data->dev_conf.nb_event_port_dequeue_depth)) { in rte_event_port_setup()
678 dev->data->dev_conf.nb_event_port_dequeue_depth); in rte_event_port_setup()
685 dev->data->dev_conf.nb_event_port_enqueue_depth)) { in rte_event_port_setup()
689 dev->data->dev_conf.nb_event_port_enqueue_depth); in rte_event_port_setup()
695 !(dev->data->event_dev_cap & in rte_event_port_setup()
703 if (dev->data->dev_started) { in rte_event_port_setup()
709 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->port_setup, -ENOTSUP); in rte_event_port_setup()
712 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->port_def_conf, in rte_event_port_setup()
714 (*dev->dev_ops->port_def_conf)(dev, port_id, &def_conf); in rte_event_port_setup()
718 dev->data->ports_cfg[port_id] = *port_conf; in rte_event_port_setup()
720 diag = (*dev->dev_ops->port_setup)(dev, port_id, port_conf); in rte_event_port_setup()
737 struct rte_eventdev *dev; in rte_event_port_quiesce() local
740 dev = &rte_eventdevs[dev_id]; in rte_event_port_quiesce()
742 if (!is_valid_port(dev, port_id)) { in rte_event_port_quiesce()
747 if (dev->dev_ops->port_quiesce) in rte_event_port_quiesce()
748 (*dev->dev_ops->port_quiesce)(dev, dev->data->ports[port_id], in rte_event_port_quiesce()
756 struct rte_eventdev *dev; in rte_event_dev_attr_get() local
761 dev = &rte_eventdevs[dev_id]; in rte_event_dev_attr_get()
765 *attr_value = dev->data->nb_ports; in rte_event_dev_attr_get()
768 *attr_value = dev->data->nb_queues; in rte_event_dev_attr_get()
771 *attr_value = dev->data->dev_started; in rte_event_dev_attr_get()
784 struct rte_eventdev *dev; in rte_event_port_attr_get() local
790 dev = &rte_eventdevs[dev_id]; in rte_event_port_attr_get()
791 if (!is_valid_port(dev, port_id)) { in rte_event_port_attr_get()
798 *attr_value = dev->data->ports_cfg[port_id].enqueue_depth; in rte_event_port_attr_get()
801 *attr_value = dev->data->ports_cfg[port_id].dequeue_depth; in rte_event_port_attr_get()
804 *attr_value = dev->data->ports_cfg[port_id].new_event_threshold; in rte_event_port_attr_get()
810 config = dev->data->ports_cfg[port_id].event_port_cfg; in rte_event_port_attr_get()
825 struct rte_eventdev *dev; in rte_event_queue_attr_get() local
831 dev = &rte_eventdevs[dev_id]; in rte_event_queue_attr_get()
832 if (!is_valid_queue(dev, queue_id)) { in rte_event_queue_attr_get()
837 conf = &dev->data->queues_cfg[queue_id]; in rte_event_queue_attr_get()
842 if (dev->data->event_dev_cap & RTE_EVENT_DEV_CAP_QUEUE_QOS) in rte_event_queue_attr_get()
862 if (dev->dev_ops->queue_attr_get) in rte_event_queue_attr_get()
863 return (*dev->dev_ops->queue_attr_get)( in rte_event_queue_attr_get()
864 dev, queue_id, attr_id, attr_value); in rte_event_queue_attr_get()
868 if (dev->dev_ops->queue_attr_get) in rte_event_queue_attr_get()
869 return (*dev->dev_ops->queue_attr_get)( in rte_event_queue_attr_get()
870 dev, queue_id, attr_id, attr_value); in rte_event_queue_attr_get()
882 struct rte_eventdev *dev; in rte_event_queue_attr_set() local
885 dev = &rte_eventdevs[dev_id]; in rte_event_queue_attr_set()
886 if (!is_valid_queue(dev, queue_id)) { in rte_event_queue_attr_set()
891 if (!(dev->data->event_dev_cap & in rte_event_queue_attr_set()
899 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->queue_attr_set, -ENOTSUP); in rte_event_queue_attr_set()
900 return (*dev->dev_ops->queue_attr_set)(dev, queue_id, attr_id, in rte_event_queue_attr_set()
909 struct rte_eventdev *dev; in rte_event_port_link() local
916 dev = &rte_eventdevs[dev_id]; in rte_event_port_link()
918 if (*dev->dev_ops->port_link == NULL) { in rte_event_port_link()
924 if (!is_valid_port(dev, port_id)) { in rte_event_port_link()
931 for (i = 0; i < dev->data->nb_queues; i++) in rte_event_port_link()
935 nb_links = dev->data->nb_queues; in rte_event_port_link()
946 if (queues[i] >= dev->data->nb_queues) { in rte_event_port_link()
951 diag = (*dev->dev_ops->port_link)(dev, dev->data->ports[port_id], in rte_event_port_link()
956 links_map = dev->data->links_map; in rte_event_port_link()
970 struct rte_eventdev *dev; in rte_event_port_unlink() local
976 dev = &rte_eventdevs[dev_id]; in rte_event_port_unlink()
978 if (*dev->dev_ops->port_unlink == NULL) { in rte_event_port_unlink()
984 if (!is_valid_port(dev, port_id)) { in rte_event_port_unlink()
990 links_map = dev->data->links_map; in rte_event_port_unlink()
996 for (i = 0; i < dev->data->nb_queues; i++) { in rte_event_port_unlink()
1014 if (queues[i] >= dev->data->nb_queues) { in rte_event_port_unlink()
1019 diag = (*dev->dev_ops->port_unlink)(dev, dev->data->ports[port_id], in rte_event_port_unlink()
1035 struct rte_eventdev *dev; in rte_event_port_unlinks_in_progress() local
1038 dev = &rte_eventdevs[dev_id]; in rte_event_port_unlinks_in_progress()
1039 if (!is_valid_port(dev, port_id)) { in rte_event_port_unlinks_in_progress()
1048 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->port_unlinks_in_progress, 0); in rte_event_port_unlinks_in_progress()
1050 return (*dev->dev_ops->port_unlinks_in_progress)(dev, in rte_event_port_unlinks_in_progress()
1051 dev->data->ports[port_id]); in rte_event_port_unlinks_in_progress()
1058 struct rte_eventdev *dev; in rte_event_port_links_get() local
1063 dev = &rte_eventdevs[dev_id]; in rte_event_port_links_get()
1064 if (!is_valid_port(dev, port_id)) { in rte_event_port_links_get()
1069 links_map = dev->data->links_map; in rte_event_port_links_get()
1072 for (i = 0; i < dev->data->nb_queues; i++) { in rte_event_port_links_get()
1086 struct rte_eventdev *dev; in rte_event_dequeue_timeout_ticks() local
1089 dev = &rte_eventdevs[dev_id]; in rte_event_dequeue_timeout_ticks()
1090 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->timeout_ticks, -ENOTSUP); in rte_event_dequeue_timeout_ticks()
1095 return (*dev->dev_ops->timeout_ticks)(dev, ns, timeout_ticks); in rte_event_dequeue_timeout_ticks()
1101 struct rte_eventdev *dev; in rte_event_dev_service_id_get() local
1104 dev = &rte_eventdevs[dev_id]; in rte_event_dev_service_id_get()
1109 if (dev->data->service_inited) in rte_event_dev_service_id_get()
1110 *service_id = dev->data->service_id; in rte_event_dev_service_id_get()
1112 return dev->data->service_inited ? 0 : -ESRCH; in rte_event_dev_service_id_get()
1118 struct rte_eventdev *dev; in rte_event_dev_dump() local
1121 dev = &rte_eventdevs[dev_id]; in rte_event_dev_dump()
1122 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->dump, -ENOTSUP); in rte_event_dev_dump()
1126 (*dev->dev_ops->dump)(dev, f); in rte_event_dev_dump()
1135 struct rte_eventdev *dev = &rte_eventdevs[dev_id]; in xstats_get_count() local
1136 if (dev->dev_ops->xstats_get_names != NULL) in xstats_get_count()
1137 return (*dev->dev_ops->xstats_get_names)(dev, mode, in xstats_get_count()
1157 const struct rte_eventdev *dev = &rte_eventdevs[dev_id]; in rte_event_dev_xstats_names_get() local
1159 if (dev->dev_ops->xstats_get_names != NULL) in rte_event_dev_xstats_names_get()
1160 return (*dev->dev_ops->xstats_get_names)(dev, mode, in rte_event_dev_xstats_names_get()
1173 const struct rte_eventdev *dev = &rte_eventdevs[dev_id]; in rte_event_dev_xstats_get() local
1176 if (dev->dev_ops->xstats_get != NULL) in rte_event_dev_xstats_get()
1177 return (*dev->dev_ops->xstats_get)(dev, mode, queue_port_id, in rte_event_dev_xstats_get()
1187 const struct rte_eventdev *dev = &rte_eventdevs[dev_id]; in rte_event_dev_xstats_by_name_get() local
1196 if (dev->dev_ops->xstats_get_by_name != NULL) in rte_event_dev_xstats_by_name_get()
1197 return (*dev->dev_ops->xstats_get_by_name)(dev, name, id); in rte_event_dev_xstats_by_name_get()
1206 struct rte_eventdev *dev = &rte_eventdevs[dev_id]; in rte_event_dev_xstats_reset() local
1208 if (dev->dev_ops->xstats_reset != NULL) in rte_event_dev_xstats_reset()
1209 return (*dev->dev_ops->xstats_reset)(dev, mode, queue_port_id, in rte_event_dev_xstats_reset()
1224 struct rte_eventdev *dev = &rte_eventdevs[dev_id]; in rte_event_dev_selftest() local
1226 if (dev->dev_ops->dev_selftest != NULL) { in rte_event_dev_selftest()
1231 return (*dev->dev_ops->dev_selftest)(); in rte_event_dev_selftest()
1281 struct rte_eventdev *dev; in rte_event_dev_start() local
1287 dev = &rte_eventdevs[dev_id]; in rte_event_dev_start()
1288 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->dev_start, -ENOTSUP); in rte_event_dev_start()
1290 if (dev->data->dev_started != 0) { in rte_event_dev_start()
1296 diag = (*dev->dev_ops->dev_start)(dev); in rte_event_dev_start()
1299 dev->data->dev_started = 1; in rte_event_dev_start()
1303 event_dev_fp_ops_set(rte_event_fp_ops + dev_id, dev); in rte_event_dev_start()
1312 struct rte_eventdev *dev; in rte_event_dev_stop_flush_callback_register() local
1317 dev = &rte_eventdevs[dev_id]; in rte_event_dev_stop_flush_callback_register()
1319 dev->dev_ops->dev_stop_flush = callback; in rte_event_dev_stop_flush_callback_register()
1320 dev->data->dev_stop_flush_arg = userdata; in rte_event_dev_stop_flush_callback_register()
1328 struct rte_eventdev *dev; in rte_event_dev_stop() local
1333 dev = &rte_eventdevs[dev_id]; in rte_event_dev_stop()
1334 RTE_FUNC_PTR_OR_RET(*dev->dev_ops->dev_stop); in rte_event_dev_stop()
1336 if (dev->data->dev_started == 0) { in rte_event_dev_stop()
1342 dev->data->dev_started = 0; in rte_event_dev_stop()
1343 (*dev->dev_ops->dev_stop)(dev); in rte_event_dev_stop()
1351 struct rte_eventdev *dev; in rte_event_dev_close() local
1354 dev = &rte_eventdevs[dev_id]; in rte_event_dev_close()
1355 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->dev_close, -ENOTSUP); in rte_event_dev_close()
1358 if (dev->data->dev_started == 1) { in rte_event_dev_close()
1366 return (*dev->dev_ops->dev_close)(dev); in rte_event_dev_close()
1540 struct rte_eventdev *dev; in handle_port_list() local
1552 dev = &rte_eventdevs[dev_id]; in handle_port_list()
1555 for (i = 0; i < dev->data->nb_ports; i++) in handle_port_list()
1568 struct rte_eventdev *dev; in handle_queue_list() local
1580 dev = &rte_eventdevs[dev_id]; in handle_queue_list()
1583 for (i = 0; i < dev->data->nb_queues; i++) in handle_queue_list()