// heavyweight readers executing explicit memory barriers.
unsigned long n_heavy_reader_attempts;
unsigned long n_heavy_reader_updates;
+unsigned long n_heavy_reader_ofl_updates;
void call_rcu_tasks_trace(struct rcu_head *rhp, rcu_callback_t func);
DEFINE_RCU_TASKS(rcu_tasks_trace, rcu_tasks_wait_gp, call_rcu_tasks_trace,
!rcu_dynticks_zero_in_eqs(cpu, &t->trc_reader_nesting))
return false; // No quiescent state, do it the hard way.
n_heavy_reader_updates++;
+ if (ofl)
+ n_heavy_reader_ofl_updates++;
in_qs = true;
} else {
in_qs = likely(!t->trc_reader_nesting);
{
char buf[64];
- sprintf(buf, "N%d h:%lu/%lu", atomic_read(&trc_n_readers_need_end),
+ sprintf(buf, "N%d h:%lu/%lu/%lu", atomic_read(&trc_n_readers_need_end),
+ data_race(n_heavy_reader_ofl_updates),
data_race(n_heavy_reader_updates),
data_race(n_heavy_reader_attempts));
show_rcu_tasks_generic_gp_kthread(&rcu_tasks_trace, buf);