| /f-stack/dpdk/lib/librte_ring/ |
| H A D | rte_ring_rts_c11_mem.h | 34 ot.raw = __atomic_load_n(&ht->tail.raw, __ATOMIC_ACQUIRE); in __rte_ring_rts_update_tail() 45 0, __ATOMIC_RELEASE, __ATOMIC_ACQUIRE) == 0); in __rte_ring_rts_update_tail() 62 h->raw = __atomic_load_n(&ht->head.raw, __ATOMIC_ACQUIRE); in __rte_ring_rts_head_wait() 79 oh.raw = __atomic_load_n(&r->rts_prod.head.raw, __ATOMIC_ACQUIRE); in __rte_ring_rts_move_prod_head() 118 0, __ATOMIC_ACQUIRE, __ATOMIC_ACQUIRE) == 0); in __rte_ring_rts_move_prod_head() 135 oh.raw = __atomic_load_n(&r->rts_cons.head.raw, __ATOMIC_ACQUIRE); in __rte_ring_rts_move_cons_head() 173 0, __ATOMIC_ACQUIRE, __ATOMIC_ACQUIRE) == 0); in __rte_ring_rts_move_cons_head()
|
| H A D | rte_ring_hts_c11_mem.h | 47 p->raw = __atomic_load_n(&ht->ht.raw, __ATOMIC_ACQUIRE); in __rte_ring_hts_head_wait() 64 op.raw = __atomic_load_n(&r->hts_prod.ht.raw, __ATOMIC_ACQUIRE); in __rte_ring_hts_move_prod_head() 103 0, __ATOMIC_ACQUIRE, __ATOMIC_ACQUIRE) == 0); in __rte_ring_hts_move_prod_head() 120 op.raw = __atomic_load_n(&r->hts_cons.ht.raw, __ATOMIC_ACQUIRE); in __rte_ring_hts_move_cons_head() 158 0, __ATOMIC_ACQUIRE, __ATOMIC_ACQUIRE) == 0); in __rte_ring_hts_move_cons_head()
|
| H A D | rte_ring_c11_mem.h | 70 __atomic_thread_fence(__ATOMIC_ACQUIRE); in __rte_ring_move_prod_head() 76 __ATOMIC_ACQUIRE); in __rte_ring_move_prod_head() 146 __atomic_thread_fence(__ATOMIC_ACQUIRE); in __rte_ring_move_cons_head() 152 __ATOMIC_ACQUIRE); in __rte_ring_move_cons_head()
|
| /f-stack/dpdk/lib/librte_eal/include/generic/ |
| H A D | rte_rwlock.h | 74 __ATOMIC_ACQUIRE, __ATOMIC_RELAXED); in rte_rwlock_read_lock() 104 __ATOMIC_ACQUIRE, __ATOMIC_RELAXED); in rte_rwlock_read_trylock() 143 __ATOMIC_ACQUIRE, __ATOMIC_RELAXED) == 0) in rte_rwlock_write_trylock() 169 __ATOMIC_ACQUIRE, __ATOMIC_RELAXED); in rte_rwlock_write_lock()
|
| H A D | rte_pause.h | 89 assert(memorder == __ATOMIC_ACQUIRE || memorder == __ATOMIC_RELAXED); in rte_wait_until_equal_16() 99 assert(memorder == __ATOMIC_ACQUIRE || memorder == __ATOMIC_RELAXED); in rte_wait_until_equal_32() 109 assert(memorder == __ATOMIC_ACQUIRE || memorder == __ATOMIC_RELAXED); in rte_wait_until_equal_64()
|
| H A D | rte_spinlock.h | 67 __ATOMIC_ACQUIRE, __ATOMIC_RELAXED)) { in rte_spinlock_lock() 110 __ATOMIC_ACQUIRE, __ATOMIC_RELAXED); in rte_spinlock_trylock() 124 return __atomic_load_n(&sl->locked, __ATOMIC_ACQUIRE); in rte_spinlock_is_locked()
|
| H A D | rte_ticketlock.h | 67 rte_wait_until_equal_16(&tl->s.current, me, __ATOMIC_ACQUIRE); in rte_ticketlock_lock() 100 new.tickets, 0, __ATOMIC_ACQUIRE, __ATOMIC_RELAXED)) in rte_ticketlock_trylock() 119 tic.tickets = __atomic_load_n(&tl->tickets, __ATOMIC_ACQUIRE); in rte_ticketlock_is_locked()
|
| H A D | rte_mcslock.h | 86 while (__atomic_load_n(&me->locked, __ATOMIC_ACQUIRE)) in rte_mcslock_lock() 115 __atomic_thread_fence(__ATOMIC_ACQUIRE); in rte_mcslock_unlock()
|
| /f-stack/dpdk/lib/librte_eal/ppc/include/ |
| H A D | rte_atomic.h | 53 return __atomic_compare_exchange(dst, &exp, &src, 0, __ATOMIC_ACQUIRE, in rte_atomic16_cmpset() 54 __ATOMIC_ACQUIRE) ? 1 : 0; in rte_atomic16_cmpset() 65 __atomic_add_fetch(&v->cnt, 1, __ATOMIC_ACQUIRE); in rte_atomic16_inc() 71 __atomic_sub_fetch(&v->cnt, 1, __ATOMIC_ACQUIRE); in rte_atomic16_dec() 76 return __atomic_add_fetch(&v->cnt, 1, __ATOMIC_ACQUIRE) == 0; in rte_atomic16_inc_and_test() 81 return __atomic_sub_fetch(&v->cnt, 1, __ATOMIC_ACQUIRE) == 0; in rte_atomic16_dec_and_test()
|
| /f-stack/dpdk/lib/librte_eal/arm/include/ |
| H A D | rte_pause_64.h | 40 assert(memorder == __ATOMIC_ACQUIRE || memorder == __ATOMIC_RELAXED); in rte_wait_until_equal_16() 78 assert(memorder == __ATOMIC_ACQUIRE || memorder == __ATOMIC_RELAXED); in rte_wait_until_equal_32() 116 assert(memorder == __ATOMIC_ACQUIRE || memorder == __ATOMIC_RELAXED); in rte_wait_until_equal_64()
|
| H A D | rte_atomic_64.h | 105 success == __ATOMIC_ACQUIRE || in rte_atomic128_cmp_exchange() 117 else if (success == __ATOMIC_ACQUIRE) in rte_atomic128_cmp_exchange() 128 int ldx_mo = __HAS_ACQ(success) ? __ATOMIC_ACQUIRE : __ATOMIC_RELAXED; in rte_atomic128_cmp_exchange()
|
| /f-stack/dpdk/lib/librte_rcu/ |
| H A D | rte_rcu_qsbr.h | 406 1, __ATOMIC_ACQUIRE); in rte_rcu_qsbr_lock() 504 t = __atomic_load_n(&v->token, __ATOMIC_ACQUIRE); in rte_rcu_qsbr_quiescent() 537 bmap = __atomic_load_n(reg_thread_id, __ATOMIC_ACQUIRE); in __rte_rcu_qsbr_check_selective() 547 __ATOMIC_ACQUIRE); in __rte_rcu_qsbr_check_selective() 566 __ATOMIC_ACQUIRE); in __rte_rcu_qsbr_check_selective() 609 c = __atomic_load_n(&cnt->cnt, __ATOMIC_ACQUIRE); in __rte_rcu_qsbr_check_all()
|
| H A D | rte_rcu_qsbr.c | 235 __ATOMIC_ACQUIRE); in rte_rcu_qsbr_dump() 248 __atomic_load_n(&v->token, __ATOMIC_ACQUIRE)); in rte_rcu_qsbr_dump() 251 __atomic_load_n(&v->acked_token, __ATOMIC_ACQUIRE)); in rte_rcu_qsbr_dump() 256 __ATOMIC_ACQUIRE); in rte_rcu_qsbr_dump()
|
| /f-stack/dpdk/drivers/event/octeontx2/ |
| H A D | otx2_tim_worker.h | 27 return __atomic_load_n(&bktp->chunk_remainder, __ATOMIC_ACQUIRE); in tim_bkt_get_rem() 67 __ATOMIC_ACQUIRE); in tim_bkt_fetch_sema_lock() 81 return __atomic_fetch_add(&bktp->w1, v, __ATOMIC_ACQUIRE); in tim_bkt_inc_lock() 264 __ATOMIC_ACQUIRE); in tim_add_entry_sp() 349 __ATOMIC_ACQUIRE); in tim_add_entry_mp() 377 __ATOMIC_ACQUIRE) < 0) in tim_add_entry_mp() 401 lock_sema = __atomic_load_n(&bkt->w1, __ATOMIC_ACQUIRE); in tim_add_entry_mp() 489 __ATOMIC_ACQUIRE); in tim_add_entry_brst()
|
| /f-stack/dpdk/lib/librte_eal/common/ |
| H A D | rte_service.c | 347 if (__atomic_load_n(&s->comp_runstate, __ATOMIC_ACQUIRE) == in rte_service_runstate_get() 349 __atomic_load_n(&s->app_runstate, __ATOMIC_ACQUIRE) == in rte_service_runstate_get() 392 if (__atomic_load_n(&s->comp_runstate, __ATOMIC_ACQUIRE) != in service_run() 394 __atomic_load_n(&s->app_runstate, __ATOMIC_ACQUIRE) != in service_run() 467 while (__atomic_load_n(&cs->runstate, __ATOMIC_ACQUIRE) == in service_runner_func() 499 __ATOMIC_ACQUIRE); in rte_service_lcore_may_be_active() 711 if (__atomic_load_n(&cs->runstate, __ATOMIC_ACQUIRE) != in rte_service_lcore_del() 735 if (__atomic_load_n(&cs->runstate, __ATOMIC_ACQUIRE) == in rte_service_lcore_start() 762 if (__atomic_load_n(&lcore_states[lcore].runstate, __ATOMIC_ACQUIRE) == in rte_service_lcore_stop()
|
| /f-stack/dpdk/app/test/ |
| H A D | test_lcores.c | 45 while (__atomic_load_n(t->registered_count, __ATOMIC_ACQUIRE) != 0) in thread_loop() 86 while (__atomic_load_n(®istered_count, __ATOMIC_ACQUIRE) != in test_non_eal_lcores() 102 while (__atomic_load_n(®istered_count, __ATOMIC_ACQUIRE) != in test_non_eal_lcores() 268 while (__atomic_load_n(®istered_count, __ATOMIC_ACQUIRE) != in test_non_eal_lcores_callback() 291 while (__atomic_load_n(®istered_count, __ATOMIC_ACQUIRE) != in test_non_eal_lcores_callback()
|
| H A D | test_distributor.c | 378 false, __ATOMIC_ACQ_REL, __ATOMIC_ACQUIRE); in handle_work_for_shutdown_test() 380 zero_id = __atomic_load_n(&zero_idx, __ATOMIC_ACQUIRE); in handle_work_for_shutdown_test() 392 false, __ATOMIC_ACQ_REL, __ATOMIC_ACQUIRE); in handle_work_for_shutdown_test() 394 zero_id = __atomic_load_n(&zero_idx, __ATOMIC_ACQUIRE); in handle_work_for_shutdown_test() 480 while (!__atomic_load_n(&zero_sleep, __ATOMIC_ACQUIRE)) in sanity_test_with_worker_shutdown() 484 while (__atomic_load_n(&zero_sleep, __ATOMIC_ACQUIRE)) in sanity_test_with_worker_shutdown() 549 while (!__atomic_load_n(&zero_sleep, __ATOMIC_ACQUIRE)) in test_flush_with_worker_shutdown() 554 while (__atomic_load_n(&zero_sleep, __ATOMIC_ACQUIRE)) in test_flush_with_worker_shutdown()
|
| /f-stack/dpdk/lib/librte_distributor/ |
| H A D | rte_distributor.c | 54 while (unlikely(__atomic_load_n(retptr64, __ATOMIC_ACQUIRE) in EAL_REGISTER_TAILQ() 106 if (__atomic_load_n(&(buf->bufptr64[0]), __ATOMIC_ACQUIRE) in rte_distributor_poll_pkt() 191 __atomic_thread_fence(__ATOMIC_ACQUIRE); in rte_distributor_return_pkt() 301 if (!(__atomic_load_n(&(buf->bufptr64[0]), __ATOMIC_ACQUIRE) in handle_worker_shutdown() 352 if (__atomic_load_n(&(buf->retptr64[0]), __ATOMIC_ACQUIRE) in handle_returns() 408 while (!(__atomic_load_n(&(d->bufs[wkr].bufptr64[0]), __ATOMIC_ACQUIRE) in release() 468 __ATOMIC_ACQUIRE) & RTE_DISTRIB_GET_BUF) { in rte_distributor_process() 603 __ATOMIC_ACQUIRE) & RTE_DISTRIB_GET_BUF)) { in rte_distributor_process()
|
| H A D | rte_distributor_single.c | 50 if (__atomic_load_n(&buf->bufptr64, __ATOMIC_ACQUIRE) in rte_distributor_poll_pkt_single() 173 __ATOMIC_ACQUIRE); in process_returns() 225 __ATOMIC_ACQUIRE); in rte_distributor_process_single() 303 __ATOMIC_ACQUIRE) & RTE_DISTRIB_GET_BUF)) { in rte_distributor_process_single()
|
| /f-stack/dpdk/lib/librte_stack/ |
| H A D | rte_stack_lf_c11.h | 93 1, __ATOMIC_ACQUIRE, in __rte_stack_lf_pop_elems() 113 __atomic_thread_fence(__ATOMIC_ACQUIRE); in __rte_stack_lf_pop_elems()
|
| /f-stack/dpdk/drivers/event/opdl/ |
| H A D | opdl_ring.c | 154 __ATOMIC_ACQUIRE) - in available() 156 __ATOMIC_ACQUIRE); in available() 172 uint32_t min_seq = __atomic_load_n(&s->deps[0]->tail, __ATOMIC_ACQUIRE); in update_available_seq() 184 __ATOMIC_ACQUIRE); in update_available_seq() 192 __ATOMIC_ACQUIRE); in update_available_seq() 385 if (__atomic_load_n(&s->shared.tail, __ATOMIC_ACQUIRE) == in opdl_stage_disclaim_multithread_n() 424 *old_head = __atomic_load_n(&s->shared.head, __ATOMIC_ACQUIRE); in move_head_atomically() 448 __ATOMIC_ACQUIRE); /* memory order on fail */ in move_head_atomically() 476 rte_wait_until_equal_32(&s->shared.tail, old_head, __ATOMIC_ACQUIRE); in opdl_ring_input_multithread() 530 __ATOMIC_ACQUIRE); in opdl_stage_claim_singlethread() [all …]
|
| /f-stack/freebsd/sys/ |
| H A D | stdatomic.h | 107 #ifndef __ATOMIC_ACQUIRE 108 #define __ATOMIC_ACQUIRE 2 macro 130 memory_order_acquire = __ATOMIC_ACQUIRE,
|
| /f-stack/dpdk/lib/librte_hash/ |
| H A D | rte_cuckoo_hash.c | 1291 __ATOMIC_ACQUIRE); in search_one_bucket_lf() 1300 __ATOMIC_ACQUIRE); in search_one_bucket_lf() 1375 __ATOMIC_ACQUIRE); in __rte_hash_lookup_with_hash_lf() 1406 __ATOMIC_ACQUIRE); in __rte_hash_lookup_with_hash_lf() 1651 __ATOMIC_ACQUIRE); in search_and_remove() 2074 __ATOMIC_ACQUIRE); in __bulk_lookup_lf() 2119 __ATOMIC_ACQUIRE); in __bulk_lookup_lf() 2135 __ATOMIC_ACQUIRE); in __bulk_lookup_lf() 2151 __ATOMIC_ACQUIRE); in __bulk_lookup_lf() 2168 __ATOMIC_ACQUIRE); in __bulk_lookup_lf() [all …]
|
| /f-stack/dpdk/drivers/net/mlx5/ |
| H A D | mlx5_utils.c | 138 prev_gen_cnt = __atomic_load_n(&h->gen_cnt, __ATOMIC_ACQUIRE); in mlx5_hlist_register() 146 prev_gen_cnt != __atomic_load_n(&h->gen_cnt, __ATOMIC_ACQUIRE)) { in mlx5_hlist_register() 303 prev_gen_cnt = __atomic_load_n(&list->gen_cnt, __ATOMIC_ACQUIRE); in mlx5_cache_register() 311 if (prev_gen_cnt != __atomic_load_n(&list->gen_cnt, __ATOMIC_ACQUIRE)) { in mlx5_cache_register() 326 __atomic_add_fetch(&list->count, 1, __ATOMIC_ACQUIRE); in mlx5_cache_register() 346 __atomic_add_fetch(&list->gen_cnt, 1, __ATOMIC_ACQUIRE); in mlx5_cache_unregister() 347 __atomic_sub_fetch(&list->count, 1, __ATOMIC_ACQUIRE); in mlx5_cache_unregister()
|
| /f-stack/dpdk/lib/librte_kni/ |
| H A D | rte_kni_fifo.h | 20 __atomic_load_n((src), __ATOMIC_ACQUIRE); \
|