Home
last modified time | relevance | path

Searched refs:__ATOMIC_ACQUIRE (Results 1 – 25 of 46) sorted by relevance

12

/f-stack/dpdk/lib/librte_ring/
H A Drte_ring_rts_c11_mem.h34 ot.raw = __atomic_load_n(&ht->tail.raw, __ATOMIC_ACQUIRE); in __rte_ring_rts_update_tail()
45 0, __ATOMIC_RELEASE, __ATOMIC_ACQUIRE) == 0); in __rte_ring_rts_update_tail()
62 h->raw = __atomic_load_n(&ht->head.raw, __ATOMIC_ACQUIRE); in __rte_ring_rts_head_wait()
79 oh.raw = __atomic_load_n(&r->rts_prod.head.raw, __ATOMIC_ACQUIRE); in __rte_ring_rts_move_prod_head()
118 0, __ATOMIC_ACQUIRE, __ATOMIC_ACQUIRE) == 0); in __rte_ring_rts_move_prod_head()
135 oh.raw = __atomic_load_n(&r->rts_cons.head.raw, __ATOMIC_ACQUIRE); in __rte_ring_rts_move_cons_head()
173 0, __ATOMIC_ACQUIRE, __ATOMIC_ACQUIRE) == 0); in __rte_ring_rts_move_cons_head()
H A Drte_ring_hts_c11_mem.h47 p->raw = __atomic_load_n(&ht->ht.raw, __ATOMIC_ACQUIRE); in __rte_ring_hts_head_wait()
64 op.raw = __atomic_load_n(&r->hts_prod.ht.raw, __ATOMIC_ACQUIRE); in __rte_ring_hts_move_prod_head()
103 0, __ATOMIC_ACQUIRE, __ATOMIC_ACQUIRE) == 0); in __rte_ring_hts_move_prod_head()
120 op.raw = __atomic_load_n(&r->hts_cons.ht.raw, __ATOMIC_ACQUIRE); in __rte_ring_hts_move_cons_head()
158 0, __ATOMIC_ACQUIRE, __ATOMIC_ACQUIRE) == 0); in __rte_ring_hts_move_cons_head()
H A Drte_ring_c11_mem.h70 __atomic_thread_fence(__ATOMIC_ACQUIRE); in __rte_ring_move_prod_head()
76 __ATOMIC_ACQUIRE); in __rte_ring_move_prod_head()
146 __atomic_thread_fence(__ATOMIC_ACQUIRE); in __rte_ring_move_cons_head()
152 __ATOMIC_ACQUIRE); in __rte_ring_move_cons_head()
/f-stack/dpdk/lib/librte_eal/include/generic/
H A Drte_rwlock.h74 __ATOMIC_ACQUIRE, __ATOMIC_RELAXED); in rte_rwlock_read_lock()
104 __ATOMIC_ACQUIRE, __ATOMIC_RELAXED); in rte_rwlock_read_trylock()
143 __ATOMIC_ACQUIRE, __ATOMIC_RELAXED) == 0) in rte_rwlock_write_trylock()
169 __ATOMIC_ACQUIRE, __ATOMIC_RELAXED); in rte_rwlock_write_lock()
H A Drte_pause.h89 assert(memorder == __ATOMIC_ACQUIRE || memorder == __ATOMIC_RELAXED); in rte_wait_until_equal_16()
99 assert(memorder == __ATOMIC_ACQUIRE || memorder == __ATOMIC_RELAXED); in rte_wait_until_equal_32()
109 assert(memorder == __ATOMIC_ACQUIRE || memorder == __ATOMIC_RELAXED); in rte_wait_until_equal_64()
H A Drte_spinlock.h67 __ATOMIC_ACQUIRE, __ATOMIC_RELAXED)) { in rte_spinlock_lock()
110 __ATOMIC_ACQUIRE, __ATOMIC_RELAXED); in rte_spinlock_trylock()
124 return __atomic_load_n(&sl->locked, __ATOMIC_ACQUIRE); in rte_spinlock_is_locked()
H A Drte_ticketlock.h67 rte_wait_until_equal_16(&tl->s.current, me, __ATOMIC_ACQUIRE); in rte_ticketlock_lock()
100 new.tickets, 0, __ATOMIC_ACQUIRE, __ATOMIC_RELAXED)) in rte_ticketlock_trylock()
119 tic.tickets = __atomic_load_n(&tl->tickets, __ATOMIC_ACQUIRE); in rte_ticketlock_is_locked()
H A Drte_mcslock.h86 while (__atomic_load_n(&me->locked, __ATOMIC_ACQUIRE)) in rte_mcslock_lock()
115 __atomic_thread_fence(__ATOMIC_ACQUIRE); in rte_mcslock_unlock()
/f-stack/dpdk/lib/librte_eal/ppc/include/
H A Drte_atomic.h53 return __atomic_compare_exchange(dst, &exp, &src, 0, __ATOMIC_ACQUIRE, in rte_atomic16_cmpset()
54 __ATOMIC_ACQUIRE) ? 1 : 0; in rte_atomic16_cmpset()
65 __atomic_add_fetch(&v->cnt, 1, __ATOMIC_ACQUIRE); in rte_atomic16_inc()
71 __atomic_sub_fetch(&v->cnt, 1, __ATOMIC_ACQUIRE); in rte_atomic16_dec()
76 return __atomic_add_fetch(&v->cnt, 1, __ATOMIC_ACQUIRE) == 0; in rte_atomic16_inc_and_test()
81 return __atomic_sub_fetch(&v->cnt, 1, __ATOMIC_ACQUIRE) == 0; in rte_atomic16_dec_and_test()
/f-stack/dpdk/lib/librte_eal/arm/include/
H A Drte_pause_64.h40 assert(memorder == __ATOMIC_ACQUIRE || memorder == __ATOMIC_RELAXED); in rte_wait_until_equal_16()
78 assert(memorder == __ATOMIC_ACQUIRE || memorder == __ATOMIC_RELAXED); in rte_wait_until_equal_32()
116 assert(memorder == __ATOMIC_ACQUIRE || memorder == __ATOMIC_RELAXED); in rte_wait_until_equal_64()
H A Drte_atomic_64.h105 success == __ATOMIC_ACQUIRE || in rte_atomic128_cmp_exchange()
117 else if (success == __ATOMIC_ACQUIRE) in rte_atomic128_cmp_exchange()
128 int ldx_mo = __HAS_ACQ(success) ? __ATOMIC_ACQUIRE : __ATOMIC_RELAXED; in rte_atomic128_cmp_exchange()
/f-stack/dpdk/lib/librte_rcu/
H A Drte_rcu_qsbr.h406 1, __ATOMIC_ACQUIRE); in rte_rcu_qsbr_lock()
504 t = __atomic_load_n(&v->token, __ATOMIC_ACQUIRE); in rte_rcu_qsbr_quiescent()
537 bmap = __atomic_load_n(reg_thread_id, __ATOMIC_ACQUIRE); in __rte_rcu_qsbr_check_selective()
547 __ATOMIC_ACQUIRE); in __rte_rcu_qsbr_check_selective()
566 __ATOMIC_ACQUIRE); in __rte_rcu_qsbr_check_selective()
609 c = __atomic_load_n(&cnt->cnt, __ATOMIC_ACQUIRE); in __rte_rcu_qsbr_check_all()
H A Drte_rcu_qsbr.c235 __ATOMIC_ACQUIRE); in rte_rcu_qsbr_dump()
248 __atomic_load_n(&v->token, __ATOMIC_ACQUIRE)); in rte_rcu_qsbr_dump()
251 __atomic_load_n(&v->acked_token, __ATOMIC_ACQUIRE)); in rte_rcu_qsbr_dump()
256 __ATOMIC_ACQUIRE); in rte_rcu_qsbr_dump()
/f-stack/dpdk/drivers/event/octeontx2/
H A Dotx2_tim_worker.h27 return __atomic_load_n(&bktp->chunk_remainder, __ATOMIC_ACQUIRE); in tim_bkt_get_rem()
67 __ATOMIC_ACQUIRE); in tim_bkt_fetch_sema_lock()
81 return __atomic_fetch_add(&bktp->w1, v, __ATOMIC_ACQUIRE); in tim_bkt_inc_lock()
264 __ATOMIC_ACQUIRE); in tim_add_entry_sp()
349 __ATOMIC_ACQUIRE); in tim_add_entry_mp()
377 __ATOMIC_ACQUIRE) < 0) in tim_add_entry_mp()
401 lock_sema = __atomic_load_n(&bkt->w1, __ATOMIC_ACQUIRE); in tim_add_entry_mp()
489 __ATOMIC_ACQUIRE); in tim_add_entry_brst()
/f-stack/dpdk/lib/librte_eal/common/
H A Drte_service.c347 if (__atomic_load_n(&s->comp_runstate, __ATOMIC_ACQUIRE) == in rte_service_runstate_get()
349 __atomic_load_n(&s->app_runstate, __ATOMIC_ACQUIRE) == in rte_service_runstate_get()
392 if (__atomic_load_n(&s->comp_runstate, __ATOMIC_ACQUIRE) != in service_run()
394 __atomic_load_n(&s->app_runstate, __ATOMIC_ACQUIRE) != in service_run()
467 while (__atomic_load_n(&cs->runstate, __ATOMIC_ACQUIRE) == in service_runner_func()
499 __ATOMIC_ACQUIRE); in rte_service_lcore_may_be_active()
711 if (__atomic_load_n(&cs->runstate, __ATOMIC_ACQUIRE) != in rte_service_lcore_del()
735 if (__atomic_load_n(&cs->runstate, __ATOMIC_ACQUIRE) == in rte_service_lcore_start()
762 if (__atomic_load_n(&lcore_states[lcore].runstate, __ATOMIC_ACQUIRE) == in rte_service_lcore_stop()
/f-stack/dpdk/app/test/
H A Dtest_lcores.c45 while (__atomic_load_n(t->registered_count, __ATOMIC_ACQUIRE) != 0) in thread_loop()
86 while (__atomic_load_n(&registered_count, __ATOMIC_ACQUIRE) != in test_non_eal_lcores()
102 while (__atomic_load_n(&registered_count, __ATOMIC_ACQUIRE) != in test_non_eal_lcores()
268 while (__atomic_load_n(&registered_count, __ATOMIC_ACQUIRE) != in test_non_eal_lcores_callback()
291 while (__atomic_load_n(&registered_count, __ATOMIC_ACQUIRE) != in test_non_eal_lcores_callback()
H A Dtest_distributor.c378 false, __ATOMIC_ACQ_REL, __ATOMIC_ACQUIRE); in handle_work_for_shutdown_test()
380 zero_id = __atomic_load_n(&zero_idx, __ATOMIC_ACQUIRE); in handle_work_for_shutdown_test()
392 false, __ATOMIC_ACQ_REL, __ATOMIC_ACQUIRE); in handle_work_for_shutdown_test()
394 zero_id = __atomic_load_n(&zero_idx, __ATOMIC_ACQUIRE); in handle_work_for_shutdown_test()
480 while (!__atomic_load_n(&zero_sleep, __ATOMIC_ACQUIRE)) in sanity_test_with_worker_shutdown()
484 while (__atomic_load_n(&zero_sleep, __ATOMIC_ACQUIRE)) in sanity_test_with_worker_shutdown()
549 while (!__atomic_load_n(&zero_sleep, __ATOMIC_ACQUIRE)) in test_flush_with_worker_shutdown()
554 while (__atomic_load_n(&zero_sleep, __ATOMIC_ACQUIRE)) in test_flush_with_worker_shutdown()
/f-stack/dpdk/lib/librte_distributor/
H A Drte_distributor.c54 while (unlikely(__atomic_load_n(retptr64, __ATOMIC_ACQUIRE) in EAL_REGISTER_TAILQ()
106 if (__atomic_load_n(&(buf->bufptr64[0]), __ATOMIC_ACQUIRE) in rte_distributor_poll_pkt()
191 __atomic_thread_fence(__ATOMIC_ACQUIRE); in rte_distributor_return_pkt()
301 if (!(__atomic_load_n(&(buf->bufptr64[0]), __ATOMIC_ACQUIRE) in handle_worker_shutdown()
352 if (__atomic_load_n(&(buf->retptr64[0]), __ATOMIC_ACQUIRE) in handle_returns()
408 while (!(__atomic_load_n(&(d->bufs[wkr].bufptr64[0]), __ATOMIC_ACQUIRE) in release()
468 __ATOMIC_ACQUIRE) & RTE_DISTRIB_GET_BUF) { in rte_distributor_process()
603 __ATOMIC_ACQUIRE) & RTE_DISTRIB_GET_BUF)) { in rte_distributor_process()
H A Drte_distributor_single.c50 if (__atomic_load_n(&buf->bufptr64, __ATOMIC_ACQUIRE) in rte_distributor_poll_pkt_single()
173 __ATOMIC_ACQUIRE); in process_returns()
225 __ATOMIC_ACQUIRE); in rte_distributor_process_single()
303 __ATOMIC_ACQUIRE) & RTE_DISTRIB_GET_BUF)) { in rte_distributor_process_single()
/f-stack/dpdk/lib/librte_stack/
H A Drte_stack_lf_c11.h93 1, __ATOMIC_ACQUIRE, in __rte_stack_lf_pop_elems()
113 __atomic_thread_fence(__ATOMIC_ACQUIRE); in __rte_stack_lf_pop_elems()
/f-stack/dpdk/drivers/event/opdl/
H A Dopdl_ring.c154 __ATOMIC_ACQUIRE) - in available()
156 __ATOMIC_ACQUIRE); in available()
172 uint32_t min_seq = __atomic_load_n(&s->deps[0]->tail, __ATOMIC_ACQUIRE); in update_available_seq()
184 __ATOMIC_ACQUIRE); in update_available_seq()
192 __ATOMIC_ACQUIRE); in update_available_seq()
385 if (__atomic_load_n(&s->shared.tail, __ATOMIC_ACQUIRE) == in opdl_stage_disclaim_multithread_n()
424 *old_head = __atomic_load_n(&s->shared.head, __ATOMIC_ACQUIRE); in move_head_atomically()
448 __ATOMIC_ACQUIRE); /* memory order on fail */ in move_head_atomically()
476 rte_wait_until_equal_32(&s->shared.tail, old_head, __ATOMIC_ACQUIRE); in opdl_ring_input_multithread()
530 __ATOMIC_ACQUIRE); in opdl_stage_claim_singlethread()
[all …]
/f-stack/freebsd/sys/
H A Dstdatomic.h107 #ifndef __ATOMIC_ACQUIRE
108 #define __ATOMIC_ACQUIRE 2 macro
130 memory_order_acquire = __ATOMIC_ACQUIRE,
/f-stack/dpdk/lib/librte_hash/
H A Drte_cuckoo_hash.c1291 __ATOMIC_ACQUIRE); in search_one_bucket_lf()
1300 __ATOMIC_ACQUIRE); in search_one_bucket_lf()
1375 __ATOMIC_ACQUIRE); in __rte_hash_lookup_with_hash_lf()
1406 __ATOMIC_ACQUIRE); in __rte_hash_lookup_with_hash_lf()
1651 __ATOMIC_ACQUIRE); in search_and_remove()
2074 __ATOMIC_ACQUIRE); in __bulk_lookup_lf()
2119 __ATOMIC_ACQUIRE); in __bulk_lookup_lf()
2135 __ATOMIC_ACQUIRE); in __bulk_lookup_lf()
2151 __ATOMIC_ACQUIRE); in __bulk_lookup_lf()
2168 __ATOMIC_ACQUIRE); in __bulk_lookup_lf()
[all …]
/f-stack/dpdk/drivers/net/mlx5/
H A Dmlx5_utils.c138 prev_gen_cnt = __atomic_load_n(&h->gen_cnt, __ATOMIC_ACQUIRE); in mlx5_hlist_register()
146 prev_gen_cnt != __atomic_load_n(&h->gen_cnt, __ATOMIC_ACQUIRE)) { in mlx5_hlist_register()
303 prev_gen_cnt = __atomic_load_n(&list->gen_cnt, __ATOMIC_ACQUIRE); in mlx5_cache_register()
311 if (prev_gen_cnt != __atomic_load_n(&list->gen_cnt, __ATOMIC_ACQUIRE)) { in mlx5_cache_register()
326 __atomic_add_fetch(&list->count, 1, __ATOMIC_ACQUIRE); in mlx5_cache_register()
346 __atomic_add_fetch(&list->gen_cnt, 1, __ATOMIC_ACQUIRE); in mlx5_cache_unregister()
347 __atomic_sub_fetch(&list->count, 1, __ATOMIC_ACQUIRE); in mlx5_cache_unregister()
/f-stack/dpdk/lib/librte_kni/
H A Drte_kni_fifo.h20 __atomic_load_n((src), __ATOMIC_ACQUIRE); \

12