Lines Matching refs:dev

69 			(rte_event_devices[i].dev ? (strncmp(  in rte_event_dev_get_dev_id()
70 rte_event_devices[i].dev->driver->name, name, in rte_event_dev_get_dev_id()
82 struct rte_eventdev *dev; in rte_event_dev_socket_id() local
85 dev = &rte_eventdevs[dev_id]; in rte_event_dev_socket_id()
87 return dev->data->socket_id; in rte_event_dev_socket_id()
93 struct rte_eventdev *dev; in rte_event_dev_info_get() local
96 dev = &rte_eventdevs[dev_id]; in rte_event_dev_info_get()
103 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->dev_infos_get, -ENOTSUP); in rte_event_dev_info_get()
104 (*dev->dev_ops->dev_infos_get)(dev, dev_info); in rte_event_dev_info_get()
106 dev_info->dequeue_timeout_ns = dev->data->dev_conf.dequeue_timeout_ns; in rte_event_dev_info_get()
108 dev_info->dev = dev->dev; in rte_event_dev_info_get()
116 struct rte_eventdev *dev; in rte_event_eth_rx_adapter_caps_get() local
121 dev = &rte_eventdevs[dev_id]; in rte_event_eth_rx_adapter_caps_get()
127 return dev->dev_ops->eth_rx_adapter_caps_get ? in rte_event_eth_rx_adapter_caps_get()
128 (*dev->dev_ops->eth_rx_adapter_caps_get)(dev, in rte_event_eth_rx_adapter_caps_get()
137 struct rte_eventdev *dev; in rte_event_timer_adapter_caps_get() local
142 dev = &rte_eventdevs[dev_id]; in rte_event_timer_adapter_caps_get()
148 return dev->dev_ops->timer_adapter_caps_get ? in rte_event_timer_adapter_caps_get()
149 (*dev->dev_ops->timer_adapter_caps_get)(dev, in rte_event_timer_adapter_caps_get()
160 struct rte_eventdev *dev; in rte_event_crypto_adapter_caps_get() local
167 dev = &rte_eventdevs[dev_id]; in rte_event_crypto_adapter_caps_get()
174 return dev->dev_ops->crypto_adapter_caps_get ? in rte_event_crypto_adapter_caps_get()
175 (*dev->dev_ops->crypto_adapter_caps_get) in rte_event_crypto_adapter_caps_get()
176 (dev, cdev, caps) : -ENOTSUP; in rte_event_crypto_adapter_caps_get()
183 struct rte_eventdev *dev; in rte_event_eth_tx_adapter_caps_get() local
189 dev = &rte_eventdevs[dev_id]; in rte_event_eth_tx_adapter_caps_get()
197 return dev->dev_ops->eth_tx_adapter_caps_get ? in rte_event_eth_tx_adapter_caps_get()
198 (*dev->dev_ops->eth_tx_adapter_caps_get)(dev, in rte_event_eth_tx_adapter_caps_get()
205 rte_event_dev_queue_config(struct rte_eventdev *dev, uint8_t nb_queues) in rte_event_dev_queue_config() argument
207 uint8_t old_nb_queues = dev->data->nb_queues; in rte_event_dev_queue_config()
212 dev->data->dev_id); in rte_event_dev_queue_config()
215 if (dev->data->queues_cfg == NULL && nb_queues != 0) { in rte_event_dev_queue_config()
217 dev->data->queues_cfg = rte_zmalloc_socket( in rte_event_dev_queue_config()
219 sizeof(dev->data->queues_cfg[0]) * nb_queues, in rte_event_dev_queue_config()
220 RTE_CACHE_LINE_SIZE, dev->data->socket_id); in rte_event_dev_queue_config()
221 if (dev->data->queues_cfg == NULL) { in rte_event_dev_queue_config()
222 dev->data->nb_queues = 0; in rte_event_dev_queue_config()
228 } else if (dev->data->queues_cfg != NULL && nb_queues != 0) { in rte_event_dev_queue_config()
229 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->queue_release, -ENOTSUP); in rte_event_dev_queue_config()
232 (*dev->dev_ops->queue_release)(dev, i); in rte_event_dev_queue_config()
235 queues_cfg = dev->data->queues_cfg; in rte_event_dev_queue_config()
244 dev->data->queues_cfg = queues_cfg; in rte_event_dev_queue_config()
252 } else if (dev->data->queues_cfg != NULL && nb_queues == 0) { in rte_event_dev_queue_config()
253 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->queue_release, -ENOTSUP); in rte_event_dev_queue_config()
256 (*dev->dev_ops->queue_release)(dev, i); in rte_event_dev_queue_config()
259 dev->data->nb_queues = nb_queues; in rte_event_dev_queue_config()
266 rte_event_dev_port_config(struct rte_eventdev *dev, uint8_t nb_ports) in rte_event_dev_port_config() argument
268 uint8_t old_nb_ports = dev->data->nb_ports; in rte_event_dev_port_config()
275 dev->data->dev_id); in rte_event_dev_port_config()
278 if (dev->data->ports == NULL && nb_ports != 0) { in rte_event_dev_port_config()
279 dev->data->ports = rte_zmalloc_socket("eventdev->data->ports", in rte_event_dev_port_config()
280 sizeof(dev->data->ports[0]) * nb_ports, in rte_event_dev_port_config()
281 RTE_CACHE_LINE_SIZE, dev->data->socket_id); in rte_event_dev_port_config()
282 if (dev->data->ports == NULL) { in rte_event_dev_port_config()
283 dev->data->nb_ports = 0; in rte_event_dev_port_config()
290 dev->data->ports_cfg = in rte_event_dev_port_config()
292 sizeof(dev->data->ports_cfg[0]) * nb_ports, in rte_event_dev_port_config()
293 RTE_CACHE_LINE_SIZE, dev->data->socket_id); in rte_event_dev_port_config()
294 if (dev->data->ports_cfg == NULL) { in rte_event_dev_port_config()
295 dev->data->nb_ports = 0; in rte_event_dev_port_config()
302 dev->data->links_map = in rte_event_dev_port_config()
304 sizeof(dev->data->links_map[0]) * nb_ports * in rte_event_dev_port_config()
306 RTE_CACHE_LINE_SIZE, dev->data->socket_id); in rte_event_dev_port_config()
307 if (dev->data->links_map == NULL) { in rte_event_dev_port_config()
308 dev->data->nb_ports = 0; in rte_event_dev_port_config()
314 dev->data->links_map[i] = in rte_event_dev_port_config()
316 } else if (dev->data->ports != NULL && nb_ports != 0) {/* re-config */ in rte_event_dev_port_config()
317 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->port_release, -ENOTSUP); in rte_event_dev_port_config()
319 ports = dev->data->ports; in rte_event_dev_port_config()
320 ports_cfg = dev->data->ports_cfg; in rte_event_dev_port_config()
321 links_map = dev->data->links_map; in rte_event_dev_port_config()
324 (*dev->dev_ops->port_release)(ports[i]); in rte_event_dev_port_config()
347 sizeof(dev->data->links_map[0]) * nb_ports * in rte_event_dev_port_config()
351 dev->data->nb_ports = 0; in rte_event_dev_port_config()
373 dev->data->ports = ports; in rte_event_dev_port_config()
374 dev->data->ports_cfg = ports_cfg; in rte_event_dev_port_config()
375 dev->data->links_map = links_map; in rte_event_dev_port_config()
376 } else if (dev->data->ports != NULL && nb_ports == 0) { in rte_event_dev_port_config()
377 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->port_release, -ENOTSUP); in rte_event_dev_port_config()
379 ports = dev->data->ports; in rte_event_dev_port_config()
381 (*dev->dev_ops->port_release)(ports[i]); in rte_event_dev_port_config()
384 dev->data->nb_ports = nb_ports; in rte_event_dev_port_config()
392 struct rte_eventdev *dev; in rte_event_dev_configure() local
397 dev = &rte_eventdevs[dev_id]; in rte_event_dev_configure()
399 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->dev_infos_get, -ENOTSUP); in rte_event_dev_configure()
400 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->dev_configure, -ENOTSUP); in rte_event_dev_configure()
402 if (dev->data->dev_started) { in rte_event_dev_configure()
411 (*dev->dev_ops->dev_infos_get)(dev, &info); in rte_event_dev_configure()
543 memcpy(&dev->data->dev_conf, dev_conf, sizeof(dev->data->dev_conf)); in rte_event_dev_configure()
546 diag = rte_event_dev_queue_config(dev, dev_conf->nb_event_queues); in rte_event_dev_configure()
554 diag = rte_event_dev_port_config(dev, dev_conf->nb_event_ports); in rte_event_dev_configure()
556 rte_event_dev_queue_config(dev, 0); in rte_event_dev_configure()
563 diag = (*dev->dev_ops->dev_configure)(dev); in rte_event_dev_configure()
566 rte_event_dev_queue_config(dev, 0); in rte_event_dev_configure()
567 rte_event_dev_port_config(dev, 0); in rte_event_dev_configure()
570 dev->data->event_dev_cap = info.event_dev_cap; in rte_event_dev_configure()
576 is_valid_queue(struct rte_eventdev *dev, uint8_t queue_id) in is_valid_queue() argument
578 if (queue_id < dev->data->nb_queues && queue_id < in is_valid_queue()
589 struct rte_eventdev *dev; in rte_event_queue_default_conf_get() local
592 dev = &rte_eventdevs[dev_id]; in rte_event_queue_default_conf_get()
597 if (!is_valid_queue(dev, queue_id)) { in rte_event_queue_default_conf_get()
602 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->queue_def_conf, -ENOTSUP); in rte_event_queue_default_conf_get()
604 (*dev->dev_ops->queue_def_conf)(dev, queue_id, queue_conf); in rte_event_queue_default_conf_get()
645 struct rte_eventdev *dev; in rte_event_queue_setup() local
649 dev = &rte_eventdevs[dev_id]; in rte_event_queue_setup()
651 if (!is_valid_queue(dev, queue_id)) { in rte_event_queue_setup()
660 dev->data->dev_conf.nb_event_queue_flows) { in rte_event_queue_setup()
664 dev->data->dev_conf.nb_event_queue_flows); in rte_event_queue_setup()
673 dev->data->dev_conf.nb_event_queue_flows) { in rte_event_queue_setup()
677 dev->data->dev_conf.nb_event_queue_flows); in rte_event_queue_setup()
682 if (dev->data->dev_started) { in rte_event_queue_setup()
688 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->queue_setup, -ENOTSUP); in rte_event_queue_setup()
691 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->queue_def_conf, in rte_event_queue_setup()
693 (*dev->dev_ops->queue_def_conf)(dev, queue_id, &def_conf); in rte_event_queue_setup()
697 dev->data->queues_cfg[queue_id] = *queue_conf; in rte_event_queue_setup()
699 return (*dev->dev_ops->queue_setup)(dev, queue_id, queue_conf); in rte_event_queue_setup()
703 is_valid_port(struct rte_eventdev *dev, uint8_t port_id) in is_valid_port() argument
705 if (port_id < dev->data->nb_ports) in is_valid_port()
715 struct rte_eventdev *dev; in rte_event_port_default_conf_get() local
718 dev = &rte_eventdevs[dev_id]; in rte_event_port_default_conf_get()
723 if (!is_valid_port(dev, port_id)) { in rte_event_port_default_conf_get()
728 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->port_def_conf, -ENOTSUP); in rte_event_port_default_conf_get()
730 (*dev->dev_ops->port_def_conf)(dev, port_id, port_conf); in rte_event_port_default_conf_get()
738 struct rte_eventdev *dev; in rte_event_port_setup() local
743 dev = &rte_eventdevs[dev_id]; in rte_event_port_setup()
745 if (!is_valid_port(dev, port_id)) { in rte_event_port_setup()
753 dev->data->dev_conf.nb_events_limit)) { in rte_event_port_setup()
757 dev->data->dev_conf.nb_events_limit); in rte_event_port_setup()
764 dev->data->dev_conf.nb_event_port_dequeue_depth)) { in rte_event_port_setup()
768 dev->data->dev_conf.nb_event_port_dequeue_depth); in rte_event_port_setup()
775 dev->data->dev_conf.nb_event_port_enqueue_depth)) { in rte_event_port_setup()
779 dev->data->dev_conf.nb_event_port_enqueue_depth); in rte_event_port_setup()
785 !(dev->data->event_dev_cap & in rte_event_port_setup()
793 if (dev->data->dev_started) { in rte_event_port_setup()
799 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->port_setup, -ENOTSUP); in rte_event_port_setup()
802 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->port_def_conf, in rte_event_port_setup()
804 (*dev->dev_ops->port_def_conf)(dev, port_id, &def_conf); in rte_event_port_setup()
808 dev->data->ports_cfg[port_id] = *port_conf; in rte_event_port_setup()
810 diag = (*dev->dev_ops->port_setup)(dev, port_id, port_conf); in rte_event_port_setup()
827 struct rte_eventdev *dev; in rte_event_dev_attr_get() local
832 dev = &rte_eventdevs[dev_id]; in rte_event_dev_attr_get()
836 *attr_value = dev->data->nb_ports; in rte_event_dev_attr_get()
839 *attr_value = dev->data->nb_queues; in rte_event_dev_attr_get()
842 *attr_value = dev->data->dev_started; in rte_event_dev_attr_get()
855 struct rte_eventdev *dev; in rte_event_port_attr_get() local
861 dev = &rte_eventdevs[dev_id]; in rte_event_port_attr_get()
862 if (!is_valid_port(dev, port_id)) { in rte_event_port_attr_get()
869 *attr_value = dev->data->ports_cfg[port_id].enqueue_depth; in rte_event_port_attr_get()
872 *attr_value = dev->data->ports_cfg[port_id].dequeue_depth; in rte_event_port_attr_get()
875 *attr_value = dev->data->ports_cfg[port_id].new_event_threshold; in rte_event_port_attr_get()
881 config = dev->data->ports_cfg[port_id].event_port_cfg; in rte_event_port_attr_get()
896 struct rte_eventdev *dev; in rte_event_queue_attr_get() local
902 dev = &rte_eventdevs[dev_id]; in rte_event_queue_attr_get()
903 if (!is_valid_queue(dev, queue_id)) { in rte_event_queue_attr_get()
908 conf = &dev->data->queues_cfg[queue_id]; in rte_event_queue_attr_get()
913 if (dev->data->event_dev_cap & RTE_EVENT_DEV_CAP_QUEUE_QOS) in rte_event_queue_attr_get()
942 struct rte_eventdev *dev; in rte_event_port_link() local
949 dev = &rte_eventdevs[dev_id]; in rte_event_port_link()
951 if (*dev->dev_ops->port_link == NULL) { in rte_event_port_link()
957 if (!is_valid_port(dev, port_id)) { in rte_event_port_link()
964 for (i = 0; i < dev->data->nb_queues; i++) in rte_event_port_link()
968 nb_links = dev->data->nb_queues; in rte_event_port_link()
979 if (queues[i] >= dev->data->nb_queues) { in rte_event_port_link()
984 diag = (*dev->dev_ops->port_link)(dev, dev->data->ports[port_id], in rte_event_port_link()
989 links_map = dev->data->links_map; in rte_event_port_link()
1003 struct rte_eventdev *dev; in rte_event_port_unlink() local
1009 dev = &rte_eventdevs[dev_id]; in rte_event_port_unlink()
1011 if (*dev->dev_ops->port_unlink == NULL) { in rte_event_port_unlink()
1017 if (!is_valid_port(dev, port_id)) { in rte_event_port_unlink()
1023 links_map = dev->data->links_map; in rte_event_port_unlink()
1029 for (i = 0; i < dev->data->nb_queues; i++) { in rte_event_port_unlink()
1047 if (queues[i] >= dev->data->nb_queues) { in rte_event_port_unlink()
1052 diag = (*dev->dev_ops->port_unlink)(dev, dev->data->ports[port_id], in rte_event_port_unlink()
1068 struct rte_eventdev *dev; in rte_event_port_unlinks_in_progress() local
1071 dev = &rte_eventdevs[dev_id]; in rte_event_port_unlinks_in_progress()
1072 if (!is_valid_port(dev, port_id)) { in rte_event_port_unlinks_in_progress()
1081 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->port_unlinks_in_progress, 0); in rte_event_port_unlinks_in_progress()
1083 return (*dev->dev_ops->port_unlinks_in_progress)(dev, in rte_event_port_unlinks_in_progress()
1084 dev->data->ports[port_id]); in rte_event_port_unlinks_in_progress()
1091 struct rte_eventdev *dev; in rte_event_port_links_get() local
1096 dev = &rte_eventdevs[dev_id]; in rte_event_port_links_get()
1097 if (!is_valid_port(dev, port_id)) { in rte_event_port_links_get()
1102 links_map = dev->data->links_map; in rte_event_port_links_get()
1105 for (i = 0; i < dev->data->nb_queues; i++) { in rte_event_port_links_get()
1119 struct rte_eventdev *dev; in rte_event_dequeue_timeout_ticks() local
1122 dev = &rte_eventdevs[dev_id]; in rte_event_dequeue_timeout_ticks()
1123 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->timeout_ticks, -ENOTSUP); in rte_event_dequeue_timeout_ticks()
1128 return (*dev->dev_ops->timeout_ticks)(dev, ns, timeout_ticks); in rte_event_dequeue_timeout_ticks()
1134 struct rte_eventdev *dev; in rte_event_dev_service_id_get() local
1137 dev = &rte_eventdevs[dev_id]; in rte_event_dev_service_id_get()
1142 if (dev->data->service_inited) in rte_event_dev_service_id_get()
1143 *service_id = dev->data->service_id; in rte_event_dev_service_id_get()
1145 return dev->data->service_inited ? 0 : -ESRCH; in rte_event_dev_service_id_get()
1151 struct rte_eventdev *dev; in rte_event_dev_dump() local
1154 dev = &rte_eventdevs[dev_id]; in rte_event_dev_dump()
1155 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->dump, -ENOTSUP); in rte_event_dev_dump()
1159 (*dev->dev_ops->dump)(dev, f); in rte_event_dev_dump()
1168 struct rte_eventdev *dev = &rte_eventdevs[dev_id]; in xstats_get_count() local
1169 if (dev->dev_ops->xstats_get_names != NULL) in xstats_get_count()
1170 return (*dev->dev_ops->xstats_get_names)(dev, mode, in xstats_get_count()
1190 const struct rte_eventdev *dev = &rte_eventdevs[dev_id]; in rte_event_dev_xstats_names_get() local
1192 if (dev->dev_ops->xstats_get_names != NULL) in rte_event_dev_xstats_names_get()
1193 return (*dev->dev_ops->xstats_get_names)(dev, mode, in rte_event_dev_xstats_names_get()
1206 const struct rte_eventdev *dev = &rte_eventdevs[dev_id]; in rte_event_dev_xstats_get() local
1209 if (dev->dev_ops->xstats_get != NULL) in rte_event_dev_xstats_get()
1210 return (*dev->dev_ops->xstats_get)(dev, mode, queue_port_id, in rte_event_dev_xstats_get()
1220 const struct rte_eventdev *dev = &rte_eventdevs[dev_id]; in rte_event_dev_xstats_by_name_get() local
1229 if (dev->dev_ops->xstats_get_by_name != NULL) in rte_event_dev_xstats_by_name_get()
1230 return (*dev->dev_ops->xstats_get_by_name)(dev, name, id); in rte_event_dev_xstats_by_name_get()
1239 struct rte_eventdev *dev = &rte_eventdevs[dev_id]; in rte_event_dev_xstats_reset() local
1241 if (dev->dev_ops->xstats_reset != NULL) in rte_event_dev_xstats_reset()
1242 return (*dev->dev_ops->xstats_reset)(dev, mode, queue_port_id, in rte_event_dev_xstats_reset()
1257 struct rte_eventdev *dev = &rte_eventdevs[dev_id]; in rte_event_dev_selftest() local
1259 if (dev->dev_ops->dev_selftest != NULL) { in rte_event_dev_selftest()
1264 return (*dev->dev_ops->dev_selftest)(); in rte_event_dev_selftest()
1272 struct rte_eventdev *dev; in rte_event_dev_start() local
1278 dev = &rte_eventdevs[dev_id]; in rte_event_dev_start()
1279 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->dev_start, -ENOTSUP); in rte_event_dev_start()
1281 if (dev->data->dev_started != 0) { in rte_event_dev_start()
1287 diag = (*dev->dev_ops->dev_start)(dev); in rte_event_dev_start()
1290 dev->data->dev_started = 1; in rte_event_dev_start()
1301 struct rte_eventdev *dev; in rte_event_dev_stop_flush_callback_register() local
1306 dev = &rte_eventdevs[dev_id]; in rte_event_dev_stop_flush_callback_register()
1308 dev->dev_ops->dev_stop_flush = callback; in rte_event_dev_stop_flush_callback_register()
1309 dev->data->dev_stop_flush_arg = userdata; in rte_event_dev_stop_flush_callback_register()
1317 struct rte_eventdev *dev; in rte_event_dev_stop() local
1322 dev = &rte_eventdevs[dev_id]; in rte_event_dev_stop()
1323 RTE_FUNC_PTR_OR_RET(*dev->dev_ops->dev_stop); in rte_event_dev_stop()
1325 if (dev->data->dev_started == 0) { in rte_event_dev_stop()
1331 dev->data->dev_started = 0; in rte_event_dev_stop()
1332 (*dev->dev_ops->dev_stop)(dev); in rte_event_dev_stop()
1339 struct rte_eventdev *dev; in rte_event_dev_close() local
1342 dev = &rte_eventdevs[dev_id]; in rte_event_dev_close()
1343 RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->dev_close, -ENOTSUP); in rte_event_dev_close()
1346 if (dev->data->dev_started == 1) { in rte_event_dev_close()
1353 return (*dev->dev_ops->dev_close)(dev); in rte_event_dev_close()
1523 struct rte_eventdev *dev; in handle_port_list() local
1535 dev = &rte_eventdevs[dev_id]; in handle_port_list()
1538 for (i = 0; i < dev->data->nb_ports; i++) in handle_port_list()
1551 struct rte_eventdev *dev; in handle_queue_list() local
1563 dev = &rte_eventdevs[dev_id]; in handle_queue_list()
1566 for (i = 0; i < dev->data->nb_queues; i++) in handle_queue_list()