Searched refs:spinlock (Results 1 – 12 of 12) sorted by relevance
139 #define ENA_SPINLOCK_INIT(spinlock) rte_spinlock_init(&(spinlock)) argument140 #define ENA_SPINLOCK_LOCK(spinlock, flags) \ argument141 ({(void)flags; rte_spinlock_lock(&(spinlock)); })142 #define ENA_SPINLOCK_UNLOCK(spinlock, flags) \ argument143 ({(void)flags; rte_spinlock_unlock(&(spinlock)); })144 #define ENA_SPINLOCK_DESTROY(spinlock) ((void)(spinlock)) argument
175 rte_spinlock_t spinlock; member188 rte_spinlock_init(&sp->spinlock); in iavf_init_spinlock_d()194 rte_spinlock_lock(&sp->spinlock); in iavf_acquire_spinlock_d()200 rte_spinlock_unlock(&sp->spinlock); in iavf_release_spinlock_d()
211 rte_spinlock_t spinlock; member217 rte_spinlock_init(&sp->spinlock); in ice_init_lock()223 rte_spinlock_lock(&sp->spinlock); in ice_acquire_lock()229 rte_spinlock_unlock(&sp->spinlock); in ice_release_lock()
37 index, and a spinlock. Accesses to the stack are made multi-thread safe by the38 spinlock.
213 For example, the memory operations before the spinlock lock are allowed to217 On the other hand, the memory operations after the spinlock unlock are allowed
245 fail, as this API tries to acquire the spinlock of vhost
125 * Remove explicit spinlock in some applications where lcores are not mapped to Tx queues with 1:1 …
215 rte_spinlock_t spinlock; member
79 [spinlock] (@ref rte_spinlock.h),
617 * A pointer to the spinlock.
301 and spinlock (a.k.a. lock elision). The methods are implemented for x86
4613 rte_spinlock_init(&sp->spinlock); in i40e_init_spinlock_d()4619 rte_spinlock_lock(&sp->spinlock); in i40e_acquire_spinlock_d()4625 rte_spinlock_unlock(&sp->spinlock); in i40e_release_spinlock_d()