linux_dsm_epyc7002/kernel/rcu
Paul E. McKenney d6ada2cf2f rcu: Rework synchronize_sched_expedited() counter handling
Now that synchronize_sched_expedited() have a mutex, it can use simpler
work-already-done detection scheme.  This commit simplifies this scheme
by using something similar to the sequence-locking counter scheme.
A counter is incremented before and after each grace period, so that
the counter is odd in the midst of the grace period and even otherwise.
So if the counter has advanced to the second even number that is
greater than or equal to the snapshot, the required grace period has
already happened.

Signed-off-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
2015-07-17 14:58:47 -07:00
..
Makefile rcu: Make SRCU optional by using CONFIG_SRCU 2015-01-06 11:04:29 -08:00
rcu.h tiny_rcu: Directly force QS when call_rcu_[bh|sched]() on idle_task 2015-01-06 11:01:12 -08:00
rcutorture.c rcutorture: Test SRCU cleanup code path 2015-05-27 12:59:58 -07:00
srcu.c rcu: Convert ACCESS_ONCE() to READ_ONCE() and WRITE_ONCE() 2015-05-27 12:56:15 -07:00
tiny_plugin.h rcu: Convert ACCESS_ONCE() to READ_ONCE() and WRITE_ONCE() 2015-05-27 12:56:15 -07:00
tiny.c This patch series contains several clean ups and even a new trace clock 2015-06-26 14:02:43 -07:00
tree_plugin.h rcu: Remove CONFIG_RCU_CPU_STALL_INFO 2015-07-17 14:58:44 -07:00
tree_trace.c rcu: Rework synchronize_sched_expedited() counter handling 2015-07-17 14:58:47 -07:00
tree.c rcu: Rework synchronize_sched_expedited() counter handling 2015-07-17 14:58:47 -07:00
tree.h rcu: Rework synchronize_sched_expedited() counter handling 2015-07-17 14:58:47 -07:00
update.c rcu: Convert ACCESS_ONCE() to READ_ONCE() and WRITE_ONCE() 2015-05-27 12:56:15 -07:00