Commit 0c09ab96 authored by Thomas Gleixner's avatar Thomas Gleixner
Browse files

cpu/hotplug: Cache number of online CPUs

Re-evaluating the bitmap wheight of the online cpus bitmap in every
invocation of num_online_cpus() over and over is a pretty useless
exercise. Especially when num_online_cpus() is used in code paths
like the IPI delivery of x86 or the membarrier code.

Cache the number of online CPUs in the core and just return the cached
variable. The accessor function provides only a snapshot when used without
protection against concurrent CPU hotplug.

The storage needs to use an atomic_t because the kexec and reboot code
(ab)use set_cpu_online() in their 'shutdown' handlers without any form of
serialization as pointed out by Mathieu. Regular CPU hotplug usage is
properly serialized.
Signed-off-by: default avatarThomas Gleixner <>
Reviewed-by: default avatarMathieu Desnoyers <>
parent b9fa6442
......@@ -10,6 +10,7 @@
#include <linux/kernel.h>
#include <linux/threads.h>
#include <linux/bitmap.h>
#include <linux/atomic.h>
#include <linux/bug.h>
/* Don't assign or return these: may not be this big! */
......@@ -95,8 +96,21 @@ extern struct cpumask __cpu_active_mask;
#define cpu_present_mask ((const struct cpumask *)&__cpu_present_mask)
#define cpu_active_mask ((const struct cpumask *)&__cpu_active_mask)
extern atomic_t __num_online_cpus;
#if NR_CPUS > 1
#define num_online_cpus() cpumask_weight(cpu_online_mask)
* num_online_cpus() - Read the number of online CPUs
* Despite the fact that __num_online_cpus is of type atomic_t, this
* interface gives only a momentary snapshot and is not protected against
* concurrent CPU hotplug operations unless invoked from a cpuhp_lock held
* region.
static inline unsigned int num_online_cpus(void)
return atomic_read(&__num_online_cpus);
#define num_possible_cpus() cpumask_weight(cpu_possible_mask)
#define num_present_cpus() cpumask_weight(cpu_present_mask)
#define num_active_cpus() cpumask_weight(cpu_active_mask)
......@@ -821,14 +835,7 @@ set_cpu_present(unsigned int cpu, bool present)
cpumask_clear_cpu(cpu, &__cpu_present_mask);
static inline void
set_cpu_online(unsigned int cpu, bool online)
if (online)
cpumask_set_cpu(cpu, &__cpu_online_mask);
cpumask_clear_cpu(cpu, &__cpu_online_mask);
void set_cpu_online(unsigned int cpu, bool online);
static inline void
set_cpu_active(unsigned int cpu, bool active)
......@@ -2298,6 +2298,9 @@ EXPORT_SYMBOL(__cpu_present_mask);
struct cpumask __cpu_active_mask __read_mostly;
atomic_t __num_online_cpus __read_mostly;
void init_cpu_present(const struct cpumask *src)
cpumask_copy(&__cpu_present_mask, src);
......@@ -2313,6 +2316,27 @@ void init_cpu_online(const struct cpumask *src)
cpumask_copy(&__cpu_online_mask, src);
void set_cpu_online(unsigned int cpu, bool online)
* atomic_inc/dec() is required to handle the horrid abuse of this
* function by the reboot and kexec code which invoke it from
* IPI/NMI broadcasts when shutting down CPUs. Invocation from
* regular CPU hotplug is properly serialized.
* Note, that the fact that __num_online_cpus is of type atomic_t
* does not protect readers which are not serialized against
* concurrent hotplug operations.
if (online) {
if (!cpumask_test_and_set_cpu(cpu, &__cpu_online_mask))
} else {
if (cpumask_test_and_clear_cpu(cpu, &__cpu_online_mask))
* Activate the first processor.
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment