Lines Matching refs:rcu_data

16 static bool rcu_rdp_is_offloaded(struct rcu_data *rdp)  in rcu_rdp_is_offloaded()
33 rdp == this_cpu_ptr(&rcu_data)) || in rcu_rdp_is_offloaded()
162 static void rcu_preempt_ctxt_queue(struct rcu_node *rnp, struct rcu_data *rdp) in rcu_preempt_ctxt_queue()
301 if (__this_cpu_read(rcu_data.cpu_no_qs.b.norm)) { in rcu_qs()
303 __this_cpu_read(rcu_data.gp_seq), in rcu_qs()
305 __this_cpu_write(rcu_data.cpu_no_qs.b.norm, false); in rcu_qs()
327 struct rcu_data *rdp = this_cpu_ptr(&rcu_data); in rcu_note_context_switch()
485 struct rcu_data *rdp; in rcu_preempt_deferred_qs_irqrestore()
495 rdp = this_cpu_ptr(&rcu_data); in rcu_preempt_deferred_qs_irqrestore()
600 return (__this_cpu_read(rcu_data.cpu_no_qs.b.exp) || in rcu_preempt_need_deferred_qs()
627 struct rcu_data *rdp; in rcu_preempt_deferred_qs_handler()
629 rdp = container_of(iwp, struct rcu_data, defer_qs_iw); in rcu_preempt_deferred_qs_handler()
653 struct rcu_data *rdp = this_cpu_ptr(&rcu_data); in rcu_read_unlock_special()
753 __this_cpu_read(rcu_data.core_needs_qs) && in rcu_flavor_sched_clock_irq()
754 __this_cpu_read(rcu_data.cpu_no_qs.b.norm) && in rcu_flavor_sched_clock_irq()
795 struct rcu_data *rdp; in dump_blkd_tasks()
817 rdp = per_cpu_ptr(&rcu_data, cpu); in dump_blkd_tasks()
834 struct rcu_data *rdp; in rcu_read_unlock_strict()
847 rdp = this_cpu_ptr(&rcu_data); in rcu_read_unlock_strict()
872 if (!__this_cpu_read(rcu_data.cpu_no_qs.s)) in rcu_qs()
875 __this_cpu_read(rcu_data.gp_seq), TPS("cpuqs")); in rcu_qs()
876 __this_cpu_write(rcu_data.cpu_no_qs.b.norm, false); in rcu_qs()
877 if (__this_cpu_read(rcu_data.cpu_no_qs.b.exp)) in rcu_qs()
878 rcu_report_exp_rdp(this_cpu_ptr(&rcu_data)); in rcu_qs()
892 if (!raw_cpu_read(rcu_data.rcu_urgent_qs)) in rcu_all_qs()
896 if (!smp_load_acquire(this_cpu_ptr(&rcu_data.rcu_urgent_qs))) { in rcu_all_qs()
900 this_cpu_write(rcu_data.rcu_urgent_qs, false); in rcu_all_qs()
901 if (unlikely(raw_cpu_read(rcu_data.rcu_need_heavy_qs))) { in rcu_all_qs()
919 if (!smp_load_acquire(this_cpu_ptr(&rcu_data.rcu_urgent_qs))) in rcu_note_context_switch()
921 this_cpu_write(rcu_data.rcu_urgent_qs, false); in rcu_note_context_switch()
922 if (unlikely(raw_cpu_read(rcu_data.rcu_need_heavy_qs))) in rcu_note_context_switch()
965 struct rcu_data *rdp = this_cpu_ptr(&rcu_data); in rcu_preempt_deferred_qs()
1031 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); in rcu_cpu_kthread_setup()
1042 static bool rcu_is_callbacks_nocb_kthread(struct rcu_data *rdp) in rcu_is_callbacks_nocb_kthread()
1055 static bool rcu_is_callbacks_kthread(struct rcu_data *rdp) in rcu_is_callbacks_kthread()