A Tour Through TREE_RCU’s Expedited Grace Periods¶
Introduction¶
This document describes RCU’s expedited grace periods. Unlike RCU’s normal grace periods, which accept long latencies to attain high efficiency and minimal disturbance, expedited grace periods accept lower efficiency and significant disturbance to attain shorter latencies.
There are two flavors of RCU (RCU-preempt and RCU-sched), with an earlier third RCU-bh flavor having been implemented in terms of the other two. Each of the two implementations is covered in its own section.
Expedited Grace Period Design¶
The expedited RCU grace periods cannot be accused of being subtle,
given that they for all intents and purposes hammer every CPU that
has not yet provided a quiescent state for the current expedited
grace period.
The one saving grace is that the hammer has grown a bit smaller
over time: The old call to try_stop_cpus()
has been
replaced with a set of calls to smp_call_function_single()
,
each of which results in an IPI to the target CPU.
The corresponding handler function checks the CPU’s state, motivating
a faster quiescent state where possible, and triggering a report
of that quiescent state.
As always for RCU, once everything has spent some time in a quiescent
state, the expedited grace period has completed.
The details of the smp_call_function_single()
handler’s
operation depend on the RCU flavor, as described in the following
sections.
RCU-preempt Expedited Grace Periods¶
CONFIG_PREEMPTION=y
kernels implement RCU-preempt.
The overall flow of the handling of a given CPU by an RCU-preempt
expedited grace period is shown in the following diagram:
The solid arrows denote direct action, for example, a function call. The dotted arrows denote indirect action, for example, an IPI or a state that is reached after some time.
If a given CPU is offline or idle, synchronize_rcu_expedited()
will ignore it because idle and offline CPUs are already residing
in quiescent states.
Otherwise, the expedited grace period will use
smp_call_function_single()
to send the CPU an IPI, which
is handled by rcu_exp_handler()
.
However, because this is preemptible RCU, rcu_exp_handler()
can check to see if the CPU is currently running in an RCU read-side
critical section.
If not, the handler can immediately report a quiescent state.
Otherwise, it sets flags so that the outermost rcu_read_unlock()
invocation will provide the needed quiescent-state report.
This flag-setting avoids the previous forced preemption of all
CPUs that might have RCU read-side critical sections.
In addition, this flag-setting is done so as to avoid increasing
the overhead of the common-case fastpath through the scheduler.
Again because this is preemptible RCU, an RCU read-side critical section
can be preempted.
When that happens, RCU will enqueue the task, which will the continue to
block the current expedited grace period until it resumes and finds its
outermost rcu_read_unlock()
.
The CPU will report a quiescent state just after enqueuing the task because
the CPU is no longer blocking the grace period.
It is instead the preempted task doing the blocking.
The list of blocked tasks is managed by rcu_preempt_ctxt_queue()
,
which is called from rcu_preempt_note_context_switch()
, which
in turn is called from rcu_note_context_switch()
, which in
turn is called from the scheduler.
Quick Quiz: |
Why not just have the expedited grace period check the state of all the CPUs? After all, that would avoid all those real-time-unfriendly IPIs. |
Answer: |
Because we want the RCU read-side critical sections to run fast,
which means no memory barriers. Therefore, it is not possible to
safely check the state from some other CPU. And even if it was
possible to safely check the state, it would still be necessary to
IPI the CPU to safely interact with the upcoming
One way to prevent your real-time application from getting hit with
these IPIs is to build your kernel with |
Please note that this is just the overall flow: Additional complications can arise due to races with CPUs going idle or offline, among other things.
RCU-sched Expedited Grace Periods¶
CONFIG_PREEMPTION=n
kernels implement RCU-sched. The overall flow of
the handling of a given CPU by an RCU-sched expedited grace period is
shown in the following diagram:
As with RCU-preempt, RCU-sched’s synchronize_rcu_expedited()
ignores
offline and idle CPUs, again because they are in remotely detectable
quiescent states. However, because the rcu_read_lock_sched()
and
rcu_read_unlock_sched()
leave no trace of their invocation, in
general it is not possible to tell whether or not the current CPU is in
an RCU read-side critical section. The best that RCU-sched’s
rcu_exp_handler()
can do is to check for idle, on the off-chance
that the CPU went idle while the IPI was in flight. If the CPU is idle,
then rcu_exp_handler()
reports the quiescent state.
Otherwise, the handler forces a future context switch by setting the NEED_RESCHED flag of the current task’s thread flag and the CPU preempt counter. At the time of the context switch, the CPU reports the quiescent state. Should the CPU go offline first, it will report the quiescent state at that time.
Expedited Grace Period and CPU Hotplug¶
The expedited nature of expedited grace periods require a much tighter interaction with CPU hotplug operations than is required for normal grace periods. In addition, attempting to IPI offline CPUs will result in splats, but failing to IPI online CPUs can result in too-short grace periods. Neither option is acceptable in production kernels.
The interaction between expedited grace periods and CPU hotplug operations is carried out at several levels:
The number of CPUs that have ever been online is tracked by the
rcu_state
structure’s->ncpus
field. Thercu_state
structure’s->ncpus_snap
field tracks the number of CPUs that have ever been online at the beginning of an RCU expedited grace period. Note that this number never decreases, at least in the absence of a time machine.The identities of the CPUs that have ever been online is tracked by the
rcu_node
structure’s->expmaskinitnext
field. Thercu_node
structure’s->expmaskinit
field tracks the identities of the CPUs that were online at least once at the beginning of the most recent RCU expedited grace period. Thercu_state
structure’s->ncpus
and->ncpus_snap
fields are used to detect when new CPUs have come online for the first time, that is, when thercu_node
structure’s->expmaskinitnext
field has changed since the beginning of the last RCU expedited grace period, which triggers an update of eachrcu_node
structure’s->expmaskinit
field from its->expmaskinitnext
field.Each
rcu_node
structure’s->expmaskinit
field is used to initialize that structure’s->expmask
at the beginning of each RCU expedited grace period. This means that only those CPUs that have been online at least once will be considered for a given grace period.Any CPU that goes offline will clear its bit in its leaf
rcu_node
structure’s->qsmaskinitnext
field, so any CPU with that bit clear can safely be ignored. However, it is possible for a CPU coming online or going offline to have this bit set for some time whilecpu_online
returnsfalse
.For each non-idle CPU that RCU believes is currently online, the grace period invokes
smp_call_function_single()
. If this succeeds, the CPU was fully online. Failure indicates that the CPU is in the process of coming online or going offline, in which case it is necessary to wait for a short time period and try again. The purpose of this wait (or series of waits, as the case may be) is to permit a concurrent CPU-hotplug operation to complete.In the case of RCU-sched, one of the last acts of an outgoing CPU is to invoke
rcu_report_dead()
, which reports a quiescent state for that CPU. However, this is likely paranoia-induced redundancy.
Quick Quiz: |
Why all the dancing around with multiple counters and masks tracking CPUs that were once online? Why not just have a single set of masks tracking the currently online CPUs and be done with it? |
Answer: |
Maintaining single set of masks tracking the online CPUs sounds easier, at least until you try working out all the race conditions between grace-period initialization and CPU-hotplug operations. For example, suppose initialization is progressing down the tree while a CPU-offline operation is progressing up the tree. This situation can result in bits set at the top of the tree that have no counterparts at the bottom of the tree. Those bits will never be cleared, which will result in grace-period hangs. In short, that way lies madness, to say nothing of a great many bugs, hangs, and deadlocks. In contrast, the current multi-mask multi-counter scheme ensures that grace-period initialization will always see consistent masks up and down the tree, which brings significant simplifications over the single-mask method. This is an instance of deferring work in order to avoid
synchronization.
Lazily recording CPU-hotplug events at the beginning of the next
grace period greatly simplifies maintenance of the CPU-tracking
bitmasks in the |
Expedited Grace Period Refinements¶
Idle-CPU Checks¶
Each expedited grace period checks for idle CPUs when initially forming
the mask of CPUs to be IPIed and again just before IPIing a CPU (both
checks are carried out by sync_rcu_exp_select_cpus()
). If the CPU is
idle at any time between those two times, the CPU will not be IPIed.
Instead, the task pushing the grace period forward will include the idle
CPUs in the mask passed to rcu_report_exp_cpu_mult()
.
For RCU-sched, there is an additional check: If the IPI has interrupted
the idle loop, then rcu_exp_handler()
invokes
rcu_report_exp_rdp()
to report the corresponding quiescent state.
For RCU-preempt, there is no specific check for idle in the IPI handler
(rcu_exp_handler()
), but because RCU read-side critical sections are
not permitted within the idle loop, if rcu_exp_handler()
sees that
the CPU is within RCU read-side critical section, the CPU cannot
possibly be idle. Otherwise, rcu_exp_handler()
invokes
rcu_report_exp_rdp()
to report the corresponding quiescent state,
regardless of whether or not that quiescent state was due to the CPU
being idle.
In summary, RCU expedited grace periods check for idle when building the bitmask of CPUs that must be IPIed, just before sending each IPI, and (either explicitly or implicitly) within the IPI handler.
Batching via Sequence Counter¶
If each grace-period request was carried out separately, expedited grace periods would have abysmal scalability and problematic high-load characteristics. Because each grace-period operation can serve an unlimited number of updates, it is important to batch requests, so that a single expedited grace-period operation will cover all requests in the corresponding batch.
This batching is controlled by a sequence counter named
->expedited_sequence
in the rcu_state
structure. This counter
has an odd value when there is an expedited grace period in progress and
an even value otherwise, so that dividing the counter value by two gives
the number of completed grace periods. During any given update request,
the counter must transition from even to odd and then back to even, thus
indicating that a grace period has elapsed. Therefore, if the initial
value of the counter is s
, the updater must wait until the counter
reaches at least the value (s+3)&~0x1
. This counter is managed by
the following access functions:
rcu_exp_gp_seq_start()
, which marks the start of an expedited grace period.rcu_exp_gp_seq_end()
, which marks the end of an expedited grace period.rcu_exp_gp_seq_snap()
, which obtains a snapshot of the counter.rcu_exp_gp_seq_done()
, which returnstrue
if a full expedited grace period has elapsed since the corresponding call torcu_exp_gp_seq_snap()
.
Again, only one request in a given batch need actually carry out a grace-period operation, which means there must be an efficient way to identify which of many concurrent reqeusts will initiate the grace period, and that there be an efficient way for the remaining requests to wait for that grace period to complete. However, that is the topic of the next section.
Funnel Locking and Wait/Wakeup¶
The natural way to sort out which of a batch of updaters will initiate
the expedited grace period is to use the rcu_node
combining tree, as
implemented by the exp_funnel_lock()
function. The first updater
corresponding to a given grace period arriving at a given rcu_node
structure records its desired grace-period sequence number in the
->exp_seq_rq
field and moves up to the next level in the tree.
Otherwise, if the ->exp_seq_rq
field already contains the sequence
number for the desired grace period or some later one, the updater
blocks on one of four wait queues in the ->exp_wq[]
array, using the
second-from-bottom and third-from bottom bits as an index. An
->exp_lock
field in the rcu_node
structure synchronizes access
to these fields.
An empty rcu_node
tree is shown in the following diagram, with the
white cells representing the ->exp_seq_rq
field and the red cells
representing the elements of the ->exp_wq[]
array.
The next diagram shows the situation after the arrival of Task A and
Task B at the leftmost and rightmost leaf rcu_node
structures,
respectively. The current value of the rcu_state
structure’s
->expedited_sequence
field is zero, so adding three and clearing the
bottom bit results in the value two, which both tasks record in the
->exp_seq_rq
field of their respective rcu_node
structures:
Each of Tasks A and B will move up to the root rcu_node
structure.
Suppose that Task A wins, recording its desired grace-period sequence
number and resulting in the state shown below:
Task A now advances to initiate a new grace period, while Task B moves
up to the root rcu_node
structure, and, seeing that its desired
sequence number is already recorded, blocks on ->exp_wq[1]
.
Quick Quiz: |
Why |
Answer: |
No.
Recall that the bottom bit of the desired sequence number indicates
whether or not a grace period is currently in progress. It is
therefore necessary to shift the sequence number right one bit
position to obtain the number of the grace period. This results in
|
If Tasks C and D also arrive at this point, they will compute the same
desired grace-period sequence number, and see that both leaf
rcu_node
structures already have that value recorded. They will
therefore block on their respective rcu_node
structures’
->exp_wq[1]
fields, as shown below:
Task A now acquires the rcu_state
structure’s ->exp_mutex
and
initiates the grace period, which increments ->expedited_sequence
.
Therefore, if Tasks E and F arrive, they will compute a desired sequence
number of 4 and will record this value as shown below:
Tasks E and F will propagate up the rcu_node
combining tree, with
Task F blocking on the root rcu_node
structure and Task E wait for
Task A to finish so that it can start the next grace period. The
resulting state is as shown below:
Once the grace period completes, Task A starts waking up the tasks
waiting for this grace period to complete, increments the
->expedited_sequence
, acquires the ->exp_wake_mutex
and then
releases the ->exp_mutex
. This results in the following state:
Task E can then acquire ->exp_mutex
and increment
->expedited_sequence
to the value three. If new tasks G and H arrive
and moves up the combining tree at the same time, the state will be as
follows:
Note that three of the root rcu_node
structure’s waitqueues are now
occupied. However, at some point, Task A will wake up the tasks blocked
on the ->exp_wq
waitqueues, resulting in the following state:
Execution will continue with Tasks E and H completing their grace periods and carrying out their wakeups.
Quick Quiz: |
What happens if Task A takes so long to do its wakeups that Task E’s grace period completes? |
Answer: |
Then Task E will block on the |
Use of Workqueues¶
In earlier implementations, the task requesting the expedited grace period also drove it to completion. This straightforward approach had the disadvantage of needing to account for POSIX signals sent to user tasks, so more recent implemementations use the Linux kernel’s workqueues (see Concurrency Managed Workqueue (cmwq)).
The requesting task still does counter snapshotting and funnel-lock
processing, but the task reaching the top of the funnel lock does a
schedule_work()
(from _synchronize_rcu_expedited()
so that a
workqueue kthread does the actual grace-period processing. Because
workqueue kthreads do not accept POSIX signals, grace-period-wait
processing need not allow for POSIX signals. In addition, this approach
allows wakeups for the previous expedited grace period to be overlapped
with processing for the next expedited grace period. Because there are
only four sets of waitqueues, it is necessary to ensure that the
previous grace period’s wakeups complete before the next grace period’s
wakeups start. This is handled by having the ->exp_mutex
guard
expedited grace-period processing and the ->exp_wake_mutex
guard
wakeups. The key point is that the ->exp_mutex
is not released until
the first wakeup is complete, which means that the ->exp_wake_mutex
has already been acquired at that point. This approach ensures that the
previous grace period’s wakeups can be carried out while the current
grace period is in process, but that these wakeups will complete before
the next grace period starts. This means that only three waitqueues are
required, guaranteeing that the four that are provided are sufficient.
Stall Warnings¶
Expediting grace periods does nothing to speed things up when RCU readers take too long, and therefore expedited grace periods check for stalls just as normal grace periods do.
Quick Quiz: |
But why not just let the normal grace-period machinery detect the stalls, given that a given reader must block both normal and expedited grace periods? |
Answer: |
Because it is quite possible that at a given time there is no normal grace period in progress, in which case the normal grace period cannot emit a stall warning. |
The synchronize_sched_expedited_wait()
function loops waiting for
the expedited grace period to end, but with a timeout set to the current
RCU CPU stall-warning time. If this time is exceeded, any CPUs or
rcu_node
structures blocking the current grace period are printed.
Each stall warning results in another pass through the loop, but the
second and subsequent passes use longer stall times.
Mid-boot operation¶
The use of workqueues has the advantage that the expedited grace-period code need not worry about POSIX signals. Unfortunately, it has the corresponding disadvantage that workqueues cannot be used until they are initialized, which does not happen until some time after the scheduler spawns the first task. Given that there are parts of the kernel that really do want to execute grace periods during this mid-boot “dead zone”, expedited grace periods must do something else during thie time.
What they do is to fall back to the old practice of requiring that the requesting task drive the expedited grace period, as was the case before the use of workqueues. However, the requesting task is only required to drive the grace period during the mid-boot dead zone. Before mid-boot, a synchronous grace period is a no-op. Some time after mid-boot, workqueues are used.
Non-expedited non-SRCU synchronous grace periods must also operate normally during mid-boot. This is handled by causing non-expedited grace periods to take the expedited code path during mid-boot.
The current code assumes that there are no POSIX signals during the mid-boot dead zone. However, if an overwhelming need for POSIX signals somehow arises, appropriate adjustments can be made to the expedited stall-warning code. One such adjustment would reinstate the pre-workqueue stall-warning checks, but only during the mid-boot dead zone.
With this refinement, synchronous grace periods can now be used from task context pretty much any time during the life of the kernel. That is, aside from some points in the suspend, hibernate, or shutdown code path.
Summary¶
Expedited grace periods use a sequence-number approach to promote
batching, so that a single grace-period operation can serve numerous
requests. A funnel lock is used to efficiently identify the one task out
of a concurrent group that will request the grace period. All members of
the group will block on waitqueues provided in the rcu_node
structure. The actual grace-period processing is carried out by a
workqueue.
CPU-hotplug operations are noted lazily in order to prevent the need for tight synchronization between expedited grace periods and CPU-hotplug operations. The dyntick-idle counters are used to avoid sending IPIs to idle CPUs, at least in the common case. RCU-preempt and RCU-sched use different IPI handlers and different code to respond to the state changes carried out by those handlers, but otherwise use common code.
Quiescent states are tracked using the rcu_node
tree, and once all
necessary quiescent states have been reported, all tasks waiting on this
expedited grace period are awakened. A pair of mutexes are used to allow
one grace period’s wakeups to proceed concurrently with the next grace
period’s processing.
This combination of mechanisms allows expedited grace periods to run reasonably efficiently. However, for non-time-critical tasks, normal grace periods should be used instead because their longer duration permits much higher degrees of batching, and thus much lower per-request overheads.