| /f-stack/dpdk/examples/performance-thread/common/ |
| H A D | lthread_sched.c | 264 return (int)rte_atomic16_read(&num_schedulers); in lthread_num_schedulers_set() 272 return (int)rte_atomic16_read(&active_schedulers); in lthread_active_schedulers() 302 while (rte_atomic16_read(&active_schedulers) < in lthread_scheduler_shutdown_all() 303 rte_atomic16_read(&num_schedulers)) in lthread_scheduler_shutdown_all() 425 while (rte_atomic16_read(&active_schedulers) < in _lthread_schedulers_sync_start() 426 rte_atomic16_read(&num_schedulers)) in _lthread_schedulers_sync_start() 444 while (rte_atomic16_read(&active_schedulers) > 0) in _lthread_schedulers_sync_stop()
|
| /f-stack/dpdk/drivers/common/octeontx2/ |
| H A D | otx2_common.c | 118 if (idev != NULL && rte_atomic16_read(&idev->npa_refcnt)) in otx2_npa_lf_obj_get() 139 return rte_atomic16_read(&idev->npa_refcnt); in otx2_npa_lf_active() 160 cnt = rte_atomic16_read(&idev->npa_refcnt); in otx2_npa_lf_obj_ref() 166 cnt = rte_atomic16_read(&idev->npa_refcnt); in otx2_npa_lf_obj_ref()
|
| H A D | otx2_sec_idev.c | 102 if (rte_atomic16_read(&cfg->tx_cpt[i].ref_cnt)) { in otx2_sec_idev_tx_cpt_qp_remove()
|
| /f-stack/dpdk/drivers/crypto/nitrox/ |
| H A D | nitrox_qp.h | 40 uint16_t pending_count = rte_atomic16_read(&qp->pending_count); in nitrox_qp_free_count() 49 return (rte_atomic16_read(&qp->pending_count) == 0); in nitrox_qp_is_empty() 55 return rte_atomic16_read(&qp->pending_count); in nitrox_qp_used_count()
|
| /f-stack/dpdk/app/test/ |
| H A D | test_timer.c | 227 while (rte_atomic16_read(&lcore_state[i]) != WORKER_RUNNING) in main_start_workers() 238 while (rte_atomic16_read(&lcore_state[i]) != WORKER_FINISHED) in main_wait_for_workers() 248 while (rte_atomic16_read(&lcore_state[lcore_id]) != WORKER_RUN_SIGNAL) in worker_wait_to_start()
|
| H A D | test_atomic.c | 464 if (rte_atomic16_read(&a16) != 1UL << 10) { in test_atomic()
|
| /f-stack/dpdk/app/test-bbdev/ |
| H A D | test_bbdev_perf.c | 2600 burst_sz = rte_atomic16_read(&tp->burst_sz); in dequeue_event_callback() 2606 rte_atomic16_read(&tp->nb_dequeued)], in dequeue_event_callback() 2611 rte_atomic16_read(&tp->nb_dequeued)], in dequeue_event_callback() 2616 rte_atomic16_read(&tp->nb_dequeued)], in dequeue_event_callback() 2621 rte_atomic16_read(&tp->nb_dequeued)], in dequeue_event_callback() 2795 while (rte_atomic16_read(&tp->nb_dequeued) != in throughput_intr_lcore_ldpc_dec() 2885 while (rte_atomic16_read(&tp->nb_dequeued) != in throughput_intr_lcore_dec() 2974 while (rte_atomic16_read(&tp->nb_dequeued) != in throughput_intr_lcore_enc() 3066 while (rte_atomic16_read(&tp->nb_dequeued) != in throughput_intr_lcore_ldpc_enc() 3902 while ((rte_atomic16_read(&tp->nb_dequeued) < in throughput_test() [all …]
|
| /f-stack/dpdk/drivers/net/hns3/ |
| H A D | hns3_mbx.c | 86 if (rte_atomic16_read(&hw->reset.disable_cmd)) { in hns3_get_mbx_resp() 370 if (rte_atomic16_read(&hw->reset.disable_cmd)) in hns3_dev_handle_mbx_msg()
|
| H A D | hns3_ethdev_vf.c | 901 if (rte_atomic16_read(&hw->reset.resetting)) { in hns3vf_dev_mtu_set() 1441 if (rte_atomic16_read(&hw->reset.resetting)) in hns3vf_request_link_info() 1474 if (rte_atomic16_read(&hw->reset.resetting)) { in hns3vf_vlan_filter_set() 1513 if (rte_atomic16_read(&hw->reset.resetting)) { in hns3vf_vlan_offload_set() 1895 if (rte_atomic16_read(&hw->reset.disable_cmd) == 0) { in hns3vf_do_stop() 1960 if (rte_atomic16_read(&hw->reset.resetting) == 0) { in hns3vf_dev_stop() 2191 if (rte_atomic16_read(&hw->reset.resetting)) in hns3vf_dev_start() 2405 if (rte_atomic16_read(&hw->reset.disable_cmd) == 0) in hns3vf_stop_service() 2577 if (rte_atomic16_read(&hns->hw.reset.schedule) == SCHEDULE_DEFERRED) { in hns3vf_reset_service() 2812 if (rte_atomic16_read(&hns->hw.reset.schedule) == SCHEDULE_PENDING) { in hns3vf_dev_init()
|
| H A D | hns3_cmd.c | 316 if (rte_atomic16_read(&hw->reset.disable_cmd)) { in hns3_cmd_poll_reply() 363 if (rte_atomic16_read(&hw->reset.disable_cmd)) in hns3_cmd_send()
|
| H A D | hns3_intr.c | 1790 if (rte_atomic16_read(&hns->hw.reset.schedule) == SCHEDULE_REQUESTED) in hns3_schedule_reset() 1792 if (rte_atomic16_read(&hns->hw.reset.schedule) == SCHEDULE_DEFERRED) in hns3_schedule_reset() 1811 if (rte_atomic16_read(&hns->hw.reset.schedule) != SCHEDULE_NONE) in hns3_schedule_delayed_reset() 1986 if (rte_atomic16_read(&hw->reset.disable_cmd)) in hns3_reset_err_handle()
|
| H A D | hns3_ethdev.c | 1011 if (rte_atomic16_read(&hw->reset.resetting) == 0) { in hns3_init_vlan_config() 1035 if (rte_atomic16_read(&hw->reset.resetting) == 0) { in hns3_init_vlan_config() 4866 if (rte_atomic16_read(&hw->reset.resetting)) in hns3_dev_start() 4945 if (rte_atomic16_read(&hw->reset.disable_cmd) == 0) { in hns3_do_stop() 5012 if (rte_atomic16_read(&hw->reset.resetting) == 0) { in hns3_dev_stop() 5545 if (rte_atomic16_read(&hw->reset.disable_cmd) == 0) in hns3_stop_service() 5659 if (rte_atomic16_read(&hns->hw.reset.schedule) == SCHEDULE_DEFERRED) { in hns3_reset_service() 6174 if (rte_atomic16_read(&hns->hw.reset.schedule) == SCHEDULE_PENDING) { in hns3_dev_init()
|
| H A D | hns3_dcb.c | 636 if (rte_atomic16_read(&hw->reset.resetting) == 0) { in hns3_set_rss_size() 1565 if (map_changed || rte_atomic16_read(&hw->reset.resetting)) { in hns3_dcb_configure()
|
| H A D | hns3_rxtx.c | 3747 rte_atomic16_read(&hns->hw.reset.resetting) == 0) { in hns3_set_rxtx_function()
|
| /f-stack/dpdk/examples/bbdev_app/ |
| H A D | main.c | 331 !rte_atomic16_read(&global_exit_flag); count++) { in check_port_link_status() 685 while (!rte_atomic16_read(&global_exit_flag)) { in stats_loop() 931 while (!rte_atomic16_read(&global_exit_flag)) { in processing_loop()
|
| /f-stack/dpdk/drivers/event/dsw/ |
| H A D | dsw_xstats.c | 129 load = rte_atomic16_read(&dsw->ports[port_id].load); in DSW_GEN_PORT_ACCESS_FN()
|
| H A D | dsw_event.c | 155 old_load = rte_atomic16_read(&port->load); in dsw_port_load_update() 370 int16_t measured_load = rte_atomic16_read(&dsw->ports[i].load); in dsw_retrieve_port_loads() 853 source_port_load = rte_atomic16_read(&source_port->load); in dsw_port_consider_emigration()
|
| /f-stack/dpdk/lib/librte_eal/include/generic/ |
| H A D | rte_atomic.h | 219 rte_atomic16_read(const rte_atomic16_t *v) in rte_atomic16_read() function
|
| /f-stack/dpdk/examples/l2fwd-jobstats/ |
| H A D | main.c | 518 rte_atomic16_read(&qconf->stats_read_pending); in l2fwd_main_loop()
|
| /f-stack/dpdk/examples/performance-thread/l3fwd-thread/ |
| H A D | main.c | 1913 while (rte_atomic16_read(&rx_counter) < n_rx_thread) in cpu_load_collector() 1916 while (rte_atomic16_read(&tx_counter) < n_tx_thread) in cpu_load_collector() 2246 while (rte_atomic16_read(&rx_counter) < n_rx_thread) in lthread_spawner()
|
| /f-stack/dpdk/doc/guides/sample_app_ug/ |
| H A D | l2_forward_job_stats.rst | 362 rte_atomic16_read(&qconf->stats_read_pending);
|