License cleanup: add SPDX GPL-2.0 license identifier to files with no license
Many source files in the tree are missing licensing information, which
makes it harder for compliance tools to determine the correct license.
By default all files without license information are under the default
license of the kernel, which is GPL version 2.
Update the files which contain no license information with the 'GPL-2.0'
SPDX license identifier. The SPDX identifier is a legally binding
shorthand, which can be used instead of the full boiler plate text.
This patch is based on work done by Thomas Gleixner and Kate Stewart and
Philippe Ombredanne.
How this work was done:
Patches were generated and checked against linux-4.14-rc6 for a subset of
the use cases:
- file had no licensing information it it.
- file was a */uapi/* one with no licensing information in it,
- file was a */uapi/* one with existing licensing information,
Further patches will be generated in subsequent months to fix up cases
where non-standard license headers were used, and references to license
had to be inferred by heuristics based on keywords.
The analysis to determine which SPDX License Identifier to be applied to
a file was done in a spreadsheet of side by side results from of the
output of two independent scanners (ScanCode & Windriver) producing SPDX
tag:value files created by Philippe Ombredanne. Philippe prepared the
base worksheet, and did an initial spot review of a few 1000 files.
The 4.13 kernel was the starting point of the analysis with 60,537 files
assessed. Kate Stewart did a file by file comparison of the scanner
results in the spreadsheet to determine which SPDX license identifier(s)
to be applied to the file. She confirmed any determination that was not
immediately clear with lawyers working with the Linux Foundation.
Criteria used to select files for SPDX license identifier tagging was:
- Files considered eligible had to be source code files.
- Make and config files were included as candidates if they contained >5
lines of source
- File already had some variant of a license header in it (even if <5
lines).
All documentation files were explicitly excluded.
The following heuristics were used to determine which SPDX license
identifiers to apply.
- when both scanners couldn't find any license traces, file was
considered to have no license information in it, and the top level
COPYING file license applied.
For non */uapi/* files that summary was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 11139
and resulted in the first patch in this series.
If that file was a */uapi/* path one, it was "GPL-2.0 WITH
Linux-syscall-note" otherwise it was "GPL-2.0". Results of that was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 WITH Linux-syscall-note 930
and resulted in the second patch in this series.
- if a file had some form of licensing information in it, and was one
of the */uapi/* ones, it was denoted with the Linux-syscall-note if
any GPL family license was found in the file or had no licensing in
it (per prior point). Results summary:
SPDX license identifier # files
---------------------------------------------------|------
GPL-2.0 WITH Linux-syscall-note 270
GPL-2.0+ WITH Linux-syscall-note 169
((GPL-2.0 WITH Linux-syscall-note) OR BSD-2-Clause) 21
((GPL-2.0 WITH Linux-syscall-note) OR BSD-3-Clause) 17
LGPL-2.1+ WITH Linux-syscall-note 15
GPL-1.0+ WITH Linux-syscall-note 14
((GPL-2.0+ WITH Linux-syscall-note) OR BSD-3-Clause) 5
LGPL-2.0+ WITH Linux-syscall-note 4
LGPL-2.1 WITH Linux-syscall-note 3
((GPL-2.0 WITH Linux-syscall-note) OR MIT) 3
((GPL-2.0 WITH Linux-syscall-note) AND MIT) 1
and that resulted in the third patch in this series.
- when the two scanners agreed on the detected license(s), that became
the concluded license(s).
- when there was disagreement between the two scanners (one detected a
license but the other didn't, or they both detected different
licenses) a manual inspection of the file occurred.
- In most cases a manual inspection of the information in the file
resulted in a clear resolution of the license that should apply (and
which scanner probably needed to revisit its heuristics).
- When it was not immediately clear, the license identifier was
confirmed with lawyers working with the Linux Foundation.
- If there was any question as to the appropriate license identifier,
the file was flagged for further research and to be revisited later
in time.
In total, over 70 hours of logged manual review was done on the
spreadsheet to determine the SPDX license identifiers to apply to the
source files by Kate, Philippe, Thomas and, in some cases, confirmation
by lawyers working with the Linux Foundation.
Kate also obtained a third independent scan of the 4.13 code base from
FOSSology, and compared selected files where the other two scanners
disagreed against that SPDX file, to see if there was new insights. The
Windriver scanner is based on an older version of FOSSology in part, so
they are related.
Thomas did random spot checks in about 500 files from the spreadsheets
for the uapi headers and agreed with SPDX license identifier in the
files he inspected. For the non-uapi files Thomas did random spot checks
in about 15000 files.
In initial set of patches against 4.14-rc6, 3 files were found to have
copy/paste license identifier errors, and have been fixed to reflect the
correct identifier.
Additionally Philippe spent 10 hours this week doing a detailed manual
inspection and review of the 12,461 patched files from the initial patch
version early this week with:
- a full scancode scan run, collecting the matched texts, detected
license ids and scores
- reviewing anything where there was a license detected (about 500+
files) to ensure that the applied SPDX license was correct
- reviewing anything where there was no detection but the patch license
was not GPL-2.0 WITH Linux-syscall-note to ensure that the applied
SPDX license was correct
This produced a worksheet with 20 files needing minor correction. This
worksheet was then exported into 3 different .csv files for the
different types of files to be modified.
These .csv files were then reviewed by Greg. Thomas wrote a script to
parse the csv files and add the proper SPDX tag to the file, in the
format that the file expected. This script was further refined by Greg
based on the output to detect more types of files automatically and to
distinguish between header and source .c files (which need different
comment types.) Finally Greg ran the script using the .csv files to
generate the patches.
Reviewed-by: Kate Stewart <kstewart@linuxfoundation.org>
Reviewed-by: Philippe Ombredanne <pombredanne@nexb.com>
Reviewed-by: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
2017-11-01 21:07:57 +07:00
|
|
|
// SPDX-License-Identifier: GPL-2.0
|
2005-10-30 08:16:54 +07:00
|
|
|
/*
|
2011-12-22 05:48:43 +07:00
|
|
|
* Memory subsystem support
|
2005-10-30 08:16:54 +07:00
|
|
|
*
|
|
|
|
* Written by Matt Tolentino <matthew.e.tolentino@intel.com>
|
|
|
|
* Dave Hansen <haveblue@us.ibm.com>
|
|
|
|
*
|
|
|
|
* This file provides the necessary infrastructure to represent
|
|
|
|
* a SPARSEMEM-memory-model system's physical memory in /sysfs.
|
|
|
|
* All arch-independent code that assumes MEMORY_HOTPLUG requires
|
|
|
|
* SPARSEMEM should be contained here, or in mm/memory_hotplug.c.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <linux/module.h>
|
|
|
|
#include <linux/init.h>
|
|
|
|
#include <linux/topology.h>
|
2006-01-12 03:17:46 +07:00
|
|
|
#include <linux/capability.h>
|
2005-10-30 08:16:54 +07:00
|
|
|
#include <linux/device.h>
|
|
|
|
#include <linux/memory.h>
|
|
|
|
#include <linux/memory_hotplug.h>
|
|
|
|
#include <linux/mm.h>
|
2008-10-19 10:27:12 +07:00
|
|
|
#include <linux/stat.h>
|
include cleanup: Update gfp.h and slab.h includes to prepare for breaking implicit slab.h inclusion from percpu.h
percpu.h is included by sched.h and module.h and thus ends up being
included when building most .c files. percpu.h includes slab.h which
in turn includes gfp.h making everything defined by the two files
universally available and complicating inclusion dependencies.
percpu.h -> slab.h dependency is about to be removed. Prepare for
this change by updating users of gfp and slab facilities include those
headers directly instead of assuming availability. As this conversion
needs to touch large number of source files, the following script is
used as the basis of conversion.
http://userweb.kernel.org/~tj/misc/slabh-sweep.py
The script does the followings.
* Scan files for gfp and slab usages and update includes such that
only the necessary includes are there. ie. if only gfp is used,
gfp.h, if slab is used, slab.h.
* When the script inserts a new include, it looks at the include
blocks and try to put the new include such that its order conforms
to its surrounding. It's put in the include block which contains
core kernel includes, in the same order that the rest are ordered -
alphabetical, Christmas tree, rev-Xmas-tree or at the end if there
doesn't seem to be any matching order.
* If the script can't find a place to put a new include (mostly
because the file doesn't have fitting include block), it prints out
an error message indicating which .h file needs to be added to the
file.
The conversion was done in the following steps.
1. The initial automatic conversion of all .c files updated slightly
over 4000 files, deleting around 700 includes and adding ~480 gfp.h
and ~3000 slab.h inclusions. The script emitted errors for ~400
files.
2. Each error was manually checked. Some didn't need the inclusion,
some needed manual addition while adding it to implementation .h or
embedding .c file was more appropriate for others. This step added
inclusions to around 150 files.
3. The script was run again and the output was compared to the edits
from #2 to make sure no file was left behind.
4. Several build tests were done and a couple of problems were fixed.
e.g. lib/decompress_*.c used malloc/free() wrappers around slab
APIs requiring slab.h to be added manually.
5. The script was run on all .h files but without automatically
editing them as sprinkling gfp.h and slab.h inclusions around .h
files could easily lead to inclusion dependency hell. Most gfp.h
inclusion directives were ignored as stuff from gfp.h was usually
wildly available and often used in preprocessor macros. Each
slab.h inclusion directive was examined and added manually as
necessary.
6. percpu.h was updated not to include slab.h.
7. Build test were done on the following configurations and failures
were fixed. CONFIG_GCOV_KERNEL was turned off for all tests (as my
distributed build env didn't work with gcov compiles) and a few
more options had to be turned off depending on archs to make things
build (like ipr on powerpc/64 which failed due to missing writeq).
* x86 and x86_64 UP and SMP allmodconfig and a custom test config.
* powerpc and powerpc64 SMP allmodconfig
* sparc and sparc64 SMP allmodconfig
* ia64 SMP allmodconfig
* s390 SMP allmodconfig
* alpha SMP allmodconfig
* um on x86_64 SMP allmodconfig
8. percpu.h modifications were reverted so that it could be applied as
a separate patch and serve as bisection point.
Given the fact that I had only a couple of failures from tests on step
6, I'm fairly confident about the coverage of this conversion patch.
If there is a breakage, it's likely to be something in one of the arch
headers which should be easily discoverable easily on most builds of
the specific arch.
Signed-off-by: Tejun Heo <tj@kernel.org>
Guess-its-ok-by: Christoph Lameter <cl@linux-foundation.org>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Lee Schermerhorn <Lee.Schermerhorn@hp.com>
2010-03-24 15:04:11 +07:00
|
|
|
#include <linux/slab.h>
|
2008-10-19 10:27:12 +07:00
|
|
|
|
2011-07-27 06:09:06 +07:00
|
|
|
#include <linux/atomic.h>
|
2016-12-25 02:46:01 +07:00
|
|
|
#include <linux/uaccess.h>
|
2005-10-30 08:16:54 +07:00
|
|
|
|
|
|
|
#define MEMORY_CLASS_NAME "memory"
|
2011-01-20 23:43:34 +07:00
|
|
|
|
2020-04-07 10:07:24 +07:00
|
|
|
static const char *const online_type_to_str[] = {
|
|
|
|
[MMOP_OFFLINE] = "offline",
|
|
|
|
[MMOP_ONLINE] = "online",
|
|
|
|
[MMOP_ONLINE_KERNEL] = "online_kernel",
|
|
|
|
[MMOP_ONLINE_MOVABLE] = "online_movable",
|
|
|
|
};
|
|
|
|
|
mm/memory_hotplug: allow to specify a default online_type
For now, distributions implement advanced udev rules to essentially
- Don't online any hotplugged memory (s390x)
- Online all memory to ZONE_NORMAL (e.g., most virt environments like
hyperv)
- Online all memory to ZONE_MOVABLE in case the zone imbalance is taken
care of (e.g., bare metal, special virt environments)
In summary: All memory is usually onlined the same way, however, the
kernel always has to ask user space to come up with the same answer.
E.g., Hyper-V always waits for a memory block to get onlined before
continuing, otherwise it might end up adding memory faster than
onlining it, which can result in strange OOM situations. This waiting
slows down adding of a bigger amount of memory.
Let's allow to specify a default online_type, not just "online" and
"offline". This allows distributions to configure the default online_type
when booting up and be done with it.
We can now specify "offline", "online", "online_movable" and
"online_kernel" via
- "memhp_default_state=" on the kernel cmdline
- /sys/devices/system/memory/auto_online_blocks
just like we are able to specify for a single memory block via
/sys/devices/system/memory/memoryX/state
Signed-off-by: David Hildenbrand <david@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Reviewed-by: Wei Yang <richard.weiyang@gmail.com>
Reviewed-by: Baoquan He <bhe@redhat.com>
Acked-by: Michal Hocko <mhocko@suse.com>
Acked-by: Pankaj Gupta <pankaj.gupta.linux@gmail.com>
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: Oscar Salvador <osalvador@suse.de>
Cc: "Rafael J. Wysocki" <rafael@kernel.org>
Cc: Wei Yang <richard.weiyang@gmail.com>
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
Cc: Eduardo Habkost <ehabkost@redhat.com>
Cc: Haiyang Zhang <haiyangz@microsoft.com>
Cc: Igor Mammedov <imammedo@redhat.com>
Cc: "K. Y. Srinivasan" <kys@microsoft.com>
Cc: Michael Ellerman <mpe@ellerman.id.au>
Cc: Paul Mackerras <paulus@samba.org>
Cc: Stephen Hemminger <sthemmin@microsoft.com>
Cc: Vitaly Kuznetsov <vkuznets@redhat.com>
Cc: Wei Liu <wei.liu@kernel.org>
Cc: Yumei Huang <yuhuang@redhat.com>
Link: http://lkml.kernel.org/r/20200317104942.11178-9-david@redhat.com
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2020-04-07 10:07:44 +07:00
|
|
|
int memhp_online_type_from_str(const char *str)
|
2020-04-07 10:07:24 +07:00
|
|
|
{
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < ARRAY_SIZE(online_type_to_str); i++) {
|
|
|
|
if (sysfs_streq(str, online_type_to_str[i]))
|
|
|
|
return i;
|
|
|
|
}
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
2013-08-28 13:38:27 +07:00
|
|
|
#define to_memory_block(dev) container_of(dev, struct memory_block, dev)
|
|
|
|
|
2011-01-20 23:43:34 +07:00
|
|
|
static int sections_per_block;
|
|
|
|
|
2019-07-19 05:57:40 +07:00
|
|
|
static inline unsigned long base_memory_block_id(unsigned long section_nr)
|
2011-01-20 23:43:34 +07:00
|
|
|
{
|
|
|
|
return section_nr / sections_per_block;
|
|
|
|
}
|
2005-10-30 08:16:54 +07:00
|
|
|
|
2019-07-19 05:57:40 +07:00
|
|
|
static inline unsigned long pfn_to_block_id(unsigned long pfn)
|
2019-07-19 05:56:56 +07:00
|
|
|
{
|
|
|
|
return base_memory_block_id(pfn_to_section_nr(pfn));
|
|
|
|
}
|
|
|
|
|
2019-07-19 05:57:50 +07:00
|
|
|
static inline unsigned long phys_to_block_id(unsigned long phys)
|
|
|
|
{
|
|
|
|
return pfn_to_block_id(PFN_DOWN(phys));
|
|
|
|
}
|
|
|
|
|
2013-05-08 19:18:37 +07:00
|
|
|
static int memory_subsys_online(struct device *dev);
|
|
|
|
static int memory_subsys_offline(struct device *dev);
|
|
|
|
|
2011-12-22 05:48:43 +07:00
|
|
|
static struct bus_type memory_subsys = {
|
2007-12-20 08:09:39 +07:00
|
|
|
.name = MEMORY_CLASS_NAME,
|
2011-12-22 05:48:43 +07:00
|
|
|
.dev_name = MEMORY_CLASS_NAME,
|
2013-05-08 19:18:37 +07:00
|
|
|
.online = memory_subsys_online,
|
|
|
|
.offline = memory_subsys_offline,
|
2005-10-30 08:16:54 +07:00
|
|
|
};
|
|
|
|
|
[PATCH] Notifier chain update: API changes
The kernel's implementation of notifier chains is unsafe. There is no
protection against entries being added to or removed from a chain while the
chain is in use. The issues were discussed in this thread:
http://marc.theaimsgroup.com/?l=linux-kernel&m=113018709002036&w=2
We noticed that notifier chains in the kernel fall into two basic usage
classes:
"Blocking" chains are always called from a process context
and the callout routines are allowed to sleep;
"Atomic" chains can be called from an atomic context and
the callout routines are not allowed to sleep.
We decided to codify this distinction and make it part of the API. Therefore
this set of patches introduces three new, parallel APIs: one for blocking
notifiers, one for atomic notifiers, and one for "raw" notifiers (which is
really just the old API under a new name). New kinds of data structures are
used for the heads of the chains, and new routines are defined for
registration, unregistration, and calling a chain. The three APIs are
explained in include/linux/notifier.h and their implementation is in
kernel/sys.c.
With atomic and blocking chains, the implementation guarantees that the chain
links will not be corrupted and that chain callers will not get messed up by
entries being added or removed. For raw chains the implementation provides no
guarantees at all; users of this API must provide their own protections. (The
idea was that situations may come up where the assumptions of the atomic and
blocking APIs are not appropriate, so it should be possible for users to
handle these things in their own way.)
There are some limitations, which should not be too hard to live with. For
atomic/blocking chains, registration and unregistration must always be done in
a process context since the chain is protected by a mutex/rwsem. Also, a
callout routine for a non-raw chain must not try to register or unregister
entries on its own chain. (This did happen in a couple of places and the code
had to be changed to avoid it.)
Since atomic chains may be called from within an NMI handler, they cannot use
spinlocks for synchronization. Instead we use RCU. The overhead falls almost
entirely in the unregister routine, which is okay since unregistration is much
less frequent that calling a chain.
Here is the list of chains that we adjusted and their classifications. None
of them use the raw API, so for the moment it is only a placeholder.
ATOMIC CHAINS
-------------
arch/i386/kernel/traps.c: i386die_chain
arch/ia64/kernel/traps.c: ia64die_chain
arch/powerpc/kernel/traps.c: powerpc_die_chain
arch/sparc64/kernel/traps.c: sparc64die_chain
arch/x86_64/kernel/traps.c: die_chain
drivers/char/ipmi/ipmi_si_intf.c: xaction_notifier_list
kernel/panic.c: panic_notifier_list
kernel/profile.c: task_free_notifier
net/bluetooth/hci_core.c: hci_notifier
net/ipv4/netfilter/ip_conntrack_core.c: ip_conntrack_chain
net/ipv4/netfilter/ip_conntrack_core.c: ip_conntrack_expect_chain
net/ipv6/addrconf.c: inet6addr_chain
net/netfilter/nf_conntrack_core.c: nf_conntrack_chain
net/netfilter/nf_conntrack_core.c: nf_conntrack_expect_chain
net/netlink/af_netlink.c: netlink_chain
BLOCKING CHAINS
---------------
arch/powerpc/platforms/pseries/reconfig.c: pSeries_reconfig_chain
arch/s390/kernel/process.c: idle_chain
arch/x86_64/kernel/process.c idle_notifier
drivers/base/memory.c: memory_chain
drivers/cpufreq/cpufreq.c cpufreq_policy_notifier_list
drivers/cpufreq/cpufreq.c cpufreq_transition_notifier_list
drivers/macintosh/adb.c: adb_client_list
drivers/macintosh/via-pmu.c sleep_notifier_list
drivers/macintosh/via-pmu68k.c sleep_notifier_list
drivers/macintosh/windfarm_core.c wf_client_list
drivers/usb/core/notify.c usb_notifier_list
drivers/video/fbmem.c fb_notifier_list
kernel/cpu.c cpu_chain
kernel/module.c module_notify_list
kernel/profile.c munmap_notifier
kernel/profile.c task_exit_notifier
kernel/sys.c reboot_notifier_list
net/core/dev.c netdev_chain
net/decnet/dn_dev.c: dnaddr_chain
net/ipv4/devinet.c: inetaddr_chain
It's possible that some of these classifications are wrong. If they are,
please let us know or submit a patch to fix them. Note that any chain that
gets called very frequently should be atomic, because the rwsem read-locking
used for blocking chains is very likely to incur cache misses on SMP systems.
(However, if the chain's callout routines may sleep then the chain cannot be
atomic.)
The patch set was written by Alan Stern and Chandra Seetharaman, incorporating
material written by Keith Owens and suggestions from Paul McKenney and Andrew
Morton.
[jes@sgi.com: restructure the notifier chain initialization macros]
Signed-off-by: Alan Stern <stern@rowland.harvard.edu>
Signed-off-by: Chandra Seetharaman <sekharan@us.ibm.com>
Signed-off-by: Jes Sorensen <jes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-03-27 16:16:30 +07:00
|
|
|
static BLOCKING_NOTIFIER_HEAD(memory_chain);
|
2005-10-30 08:16:54 +07:00
|
|
|
|
2006-01-06 15:10:35 +07:00
|
|
|
int register_memory_notifier(struct notifier_block *nb)
|
2005-10-30 08:16:54 +07:00
|
|
|
{
|
2015-03-08 17:48:35 +07:00
|
|
|
return blocking_notifier_chain_register(&memory_chain, nb);
|
2005-10-30 08:16:54 +07:00
|
|
|
}
|
2008-05-07 19:43:01 +07:00
|
|
|
EXPORT_SYMBOL(register_memory_notifier);
|
2005-10-30 08:16:54 +07:00
|
|
|
|
2006-01-06 15:10:35 +07:00
|
|
|
void unregister_memory_notifier(struct notifier_block *nb)
|
2005-10-30 08:16:54 +07:00
|
|
|
{
|
2015-03-08 17:48:35 +07:00
|
|
|
blocking_notifier_chain_unregister(&memory_chain, nb);
|
2005-10-30 08:16:54 +07:00
|
|
|
}
|
2008-05-07 19:43:01 +07:00
|
|
|
EXPORT_SYMBOL(unregister_memory_notifier);
|
2005-10-30 08:16:54 +07:00
|
|
|
|
2012-12-12 07:00:44 +07:00
|
|
|
static void memory_block_release(struct device *dev)
|
|
|
|
{
|
2013-08-28 13:38:27 +07:00
|
|
|
struct memory_block *mem = to_memory_block(dev);
|
2012-12-12 07:00:44 +07:00
|
|
|
|
|
|
|
kfree(mem);
|
|
|
|
}
|
|
|
|
|
2011-01-20 23:43:34 +07:00
|
|
|
unsigned long __weak memory_block_size_bytes(void)
|
|
|
|
{
|
|
|
|
return MIN_MEMORY_BLOCK_SIZE;
|
|
|
|
}
|
device-dax: "Hotplug" persistent memory for use like normal RAM
This is intended for use with NVDIMMs that are physically persistent
(physically like flash) so that they can be used as a cost-effective
RAM replacement. Intel Optane DC persistent memory is one
implementation of this kind of NVDIMM.
Currently, a persistent memory region is "owned" by a device driver,
either the "Direct DAX" or "Filesystem DAX" drivers. These drivers
allow applications to explicitly use persistent memory, generally
by being modified to use special, new libraries. (DIMM-based
persistent memory hardware/software is described in great detail
here: Documentation/nvdimm/nvdimm.txt).
However, this limits persistent memory use to applications which
*have* been modified. To make it more broadly usable, this driver
"hotplugs" memory into the kernel, to be managed and used just like
normal RAM would be.
To make this work, management software must remove the device from
being controlled by the "Device DAX" infrastructure:
echo dax0.0 > /sys/bus/dax/drivers/device_dax/unbind
and then tell the new driver that it can bind to the device:
echo dax0.0 > /sys/bus/dax/drivers/kmem/new_id
After this, there will be a number of new memory sections visible
in sysfs that can be onlined, or that may get onlined by existing
udev-initiated memory hotplug rules.
This rebinding procedure is currently a one-way trip. Once memory
is bound to "kmem", it's there permanently and can not be
unbound and assigned back to device_dax.
The kmem driver will never bind to a dax device unless the device
is *explicitly* bound to the driver. There are two reasons for
this: One, since it is a one-way trip, it can not be undone if
bound incorrectly. Two, the kmem driver destroys data on the
device. Think of if you had good data on a pmem device. It
would be catastrophic if you compile-in "kmem", but leave out
the "device_dax" driver. kmem would take over the device and
write volatile data all over your good data.
This inherits any existing NUMA information for the newly-added
memory from the persistent memory device that came from the
firmware. On Intel platforms, the firmware has guarantees that
require each socket's persistent memory to be in a separate
memory-only NUMA node. That means that this patch is not expected
to create NUMA nodes, but will simply hotplug memory into existing
nodes.
Because NUMA nodes are created, the existing NUMA APIs and tools
are sufficient to create policies for applications or memory areas
to have affinity for or an aversion to using this memory.
There is currently some metadata at the beginning of pmem regions.
The section-size memory hotplug restrictions, plus this small
reserved area can cause the "loss" of a section or two of capacity.
This should be fixable in follow-on patches. But, as a first step,
losing 256MB of memory (worst case) out of hundreds of gigabytes
is a good tradeoff vs. the required code to fix this up precisely.
This calculation is also the reason we export
memory_block_size_bytes().
Signed-off-by: Dave Hansen <dave.hansen@linux.intel.com>
Reviewed-by: Dan Williams <dan.j.williams@intel.com>
Reviewed-by: Keith Busch <keith.busch@intel.com>
Cc: Dave Jiang <dave.jiang@intel.com>
Cc: Ross Zwisler <zwisler@kernel.org>
Cc: Vishal Verma <vishal.l.verma@intel.com>
Cc: Tom Lendacky <thomas.lendacky@amd.com>
Cc: Andrew Morton <akpm@linux-foundation.org>
Cc: Michal Hocko <mhocko@suse.com>
Cc: linux-nvdimm@lists.01.org
Cc: linux-kernel@vger.kernel.org
Cc: linux-mm@kvack.org
Cc: Huang Ying <ying.huang@intel.com>
Cc: Fengguang Wu <fengguang.wu@intel.com>
Cc: Borislav Petkov <bp@suse.de>
Cc: Bjorn Helgaas <bhelgaas@google.com>
Cc: Yaowei Bai <baiyaowei@cmss.chinamobile.com>
Cc: Takashi Iwai <tiwai@suse.de>
Cc: Jerome Glisse <jglisse@redhat.com>
Reviewed-by: Vishal Verma <vishal.l.verma@intel.com>
Signed-off-by: Dan Williams <dan.j.williams@intel.com>
2019-02-26 01:57:40 +07:00
|
|
|
EXPORT_SYMBOL_GPL(memory_block_size_bytes);
|
2011-01-20 23:43:34 +07:00
|
|
|
|
2005-10-30 08:16:54 +07:00
|
|
|
/*
|
2019-09-24 05:35:43 +07:00
|
|
|
* Show the first physical section index (number) of this memory block.
|
2005-10-30 08:16:54 +07:00
|
|
|
*/
|
2018-12-03 18:16:11 +07:00
|
|
|
static ssize_t phys_index_show(struct device *dev,
|
|
|
|
struct device_attribute *attr, char *buf)
|
2005-10-30 08:16:54 +07:00
|
|
|
{
|
2013-08-28 13:38:27 +07:00
|
|
|
struct memory_block *mem = to_memory_block(dev);
|
2011-01-20 23:44:29 +07:00
|
|
|
unsigned long phys_index;
|
|
|
|
|
|
|
|
phys_index = mem->start_section_nr / sections_per_block;
|
|
|
|
return sprintf(buf, "%08lx\n", phys_index);
|
|
|
|
}
|
|
|
|
|
2008-07-24 11:28:19 +07:00
|
|
|
/*
|
drivers/base/memory.c: indicate all memory blocks as removable
We see multiple issues with the implementation/interface to compute
whether a memory block can be offlined (exposed via
/sys/devices/system/memory/memoryX/removable) and would like to simplify
it (remove the implementation).
1. It runs basically lockless. While this might be good for performance,
we see possible races with memory offlining that will require at
least some sort of locking to fix.
2. Nowadays, more false positives are possible. No arch-specific checks
are performed that validate if memory offlining will not be denied
right away (and such check will require locking). For example, arm64
won't allow to offline any memory block that was added during boot -
which will imply a very high error rate. Other archs have other
constraints.
3. The interface is inherently racy. E.g., if a memory block is detected
to be removable (and was not a false positive at that time), there is
still no guarantee that offlining will actually succeed. So any
caller already has to deal with false positives.
4. It is unclear which performance benefit this interface actually
provides. The introducing commit 5c755e9fd813 ("memory-hotplug: add
sysfs removable attribute for hotplug memory remove") mentioned
"A user-level agent must be able to identify which sections
of memory are likely to be removable before attempting the
potentially expensive operation."
However, no actual performance comparison was included.
Known users:
- lsmem: Will group memory blocks based on the "removable" property. [1]
- chmem: Indirect user. It has a RANGE mode where one can specify
removable ranges identified via lsmem to be offlined. However,
it also has a "SIZE" mode, which allows a sysadmin to skip the
manual "identify removable blocks" step. [2]
- powerpc-utils: Uses the "removable" attribute to skip some memory
blocks right away when trying to find some to offline+remove.
However, with ballooning enabled, it already skips this
information completely (because it once resulted in many false
negatives). Therefore, the implementation can deal with false
positives properly already. [3]
According to Nathan Fontenot, DLPAR on powerpc is nowadays no longer
driven from userspace via the drmgr command (powerpc-utils). Nowadays
it's managed in the kernel - including onlining/offlining of memory
blocks - triggered by drmgr writing to /sys/kernel/dlpar. So the
affected legacy userspace handling is only active on old kernels. Only
very old versions of drmgr on a new kernel (unlikely) might execute
slower - totally acceptable.
With CONFIG_MEMORY_HOTREMOVE, always indicating "removable" should not
break any user space tool. We implement a very bad heuristic now.
Without CONFIG_MEMORY_HOTREMOVE we cannot offline anything, so report
"not removable" as before.
Original discussion can be found in [4] ("[PATCH RFC v1] mm:
is_mem_section_removable() overhaul").
Other users of is_mem_section_removable() will be removed next, so that
we can remove is_mem_section_removable() completely.
[1] http://man7.org/linux/man-pages/man1/lsmem.1.html
[2] http://man7.org/linux/man-pages/man8/chmem.8.html
[3] https://github.com/ibm-power-utilities/powerpc-utils
[4] https://lkml.kernel.org/r/20200117105759.27905-1-david@redhat.com
Also, this patch probably fixes a crash reported by Steve.
http://lkml.kernel.org/r/CAPcyv4jpdaNvJ67SkjyUJLBnBnXXQv686BiVW042g03FUmWLXw@mail.gmail.com
Reported-by: "Scargall, Steve" <steve.scargall@intel.com>
Suggested-by: Michal Hocko <mhocko@kernel.org>
Signed-off-by: David Hildenbrand <david@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Reviewed-by: Nathan Fontenot <ndfont@gmail.com>
Acked-by: Michal Hocko <mhocko@suse.com>
Cc: Dan Williams <dan.j.williams@intel.com>
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: "Rafael J. Wysocki" <rafael@kernel.org>
Cc: Badari Pulavarty <pbadari@us.ibm.com>
Cc: Robert Jennings <rcj@linux.vnet.ibm.com>
Cc: Heiko Carstens <heiko.carstens@de.ibm.com>
Cc: Karel Zak <kzak@redhat.com>
Cc: <stable@vger.kernel.org>
Link: http://lkml.kernel.org/r/20200128093542.6908-1-david@redhat.com
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2020-03-29 09:17:19 +07:00
|
|
|
* Legacy interface that we cannot remove. Always indicate "removable"
|
|
|
|
* with CONFIG_MEMORY_HOTREMOVE - bad heuristic.
|
2008-07-24 11:28:19 +07:00
|
|
|
*/
|
2018-12-03 18:16:11 +07:00
|
|
|
static ssize_t removable_show(struct device *dev, struct device_attribute *attr,
|
|
|
|
char *buf)
|
2008-07-24 11:28:19 +07:00
|
|
|
{
|
drivers/base/memory.c: indicate all memory blocks as removable
We see multiple issues with the implementation/interface to compute
whether a memory block can be offlined (exposed via
/sys/devices/system/memory/memoryX/removable) and would like to simplify
it (remove the implementation).
1. It runs basically lockless. While this might be good for performance,
we see possible races with memory offlining that will require at
least some sort of locking to fix.
2. Nowadays, more false positives are possible. No arch-specific checks
are performed that validate if memory offlining will not be denied
right away (and such check will require locking). For example, arm64
won't allow to offline any memory block that was added during boot -
which will imply a very high error rate. Other archs have other
constraints.
3. The interface is inherently racy. E.g., if a memory block is detected
to be removable (and was not a false positive at that time), there is
still no guarantee that offlining will actually succeed. So any
caller already has to deal with false positives.
4. It is unclear which performance benefit this interface actually
provides. The introducing commit 5c755e9fd813 ("memory-hotplug: add
sysfs removable attribute for hotplug memory remove") mentioned
"A user-level agent must be able to identify which sections
of memory are likely to be removable before attempting the
potentially expensive operation."
However, no actual performance comparison was included.
Known users:
- lsmem: Will group memory blocks based on the "removable" property. [1]
- chmem: Indirect user. It has a RANGE mode where one can specify
removable ranges identified via lsmem to be offlined. However,
it also has a "SIZE" mode, which allows a sysadmin to skip the
manual "identify removable blocks" step. [2]
- powerpc-utils: Uses the "removable" attribute to skip some memory
blocks right away when trying to find some to offline+remove.
However, with ballooning enabled, it already skips this
information completely (because it once resulted in many false
negatives). Therefore, the implementation can deal with false
positives properly already. [3]
According to Nathan Fontenot, DLPAR on powerpc is nowadays no longer
driven from userspace via the drmgr command (powerpc-utils). Nowadays
it's managed in the kernel - including onlining/offlining of memory
blocks - triggered by drmgr writing to /sys/kernel/dlpar. So the
affected legacy userspace handling is only active on old kernels. Only
very old versions of drmgr on a new kernel (unlikely) might execute
slower - totally acceptable.
With CONFIG_MEMORY_HOTREMOVE, always indicating "removable" should not
break any user space tool. We implement a very bad heuristic now.
Without CONFIG_MEMORY_HOTREMOVE we cannot offline anything, so report
"not removable" as before.
Original discussion can be found in [4] ("[PATCH RFC v1] mm:
is_mem_section_removable() overhaul").
Other users of is_mem_section_removable() will be removed next, so that
we can remove is_mem_section_removable() completely.
[1] http://man7.org/linux/man-pages/man1/lsmem.1.html
[2] http://man7.org/linux/man-pages/man8/chmem.8.html
[3] https://github.com/ibm-power-utilities/powerpc-utils
[4] https://lkml.kernel.org/r/20200117105759.27905-1-david@redhat.com
Also, this patch probably fixes a crash reported by Steve.
http://lkml.kernel.org/r/CAPcyv4jpdaNvJ67SkjyUJLBnBnXXQv686BiVW042g03FUmWLXw@mail.gmail.com
Reported-by: "Scargall, Steve" <steve.scargall@intel.com>
Suggested-by: Michal Hocko <mhocko@kernel.org>
Signed-off-by: David Hildenbrand <david@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Reviewed-by: Nathan Fontenot <ndfont@gmail.com>
Acked-by: Michal Hocko <mhocko@suse.com>
Cc: Dan Williams <dan.j.williams@intel.com>
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: "Rafael J. Wysocki" <rafael@kernel.org>
Cc: Badari Pulavarty <pbadari@us.ibm.com>
Cc: Robert Jennings <rcj@linux.vnet.ibm.com>
Cc: Heiko Carstens <heiko.carstens@de.ibm.com>
Cc: Karel Zak <kzak@redhat.com>
Cc: <stable@vger.kernel.org>
Link: http://lkml.kernel.org/r/20200128093542.6908-1-david@redhat.com
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2020-03-29 09:17:19 +07:00
|
|
|
return sprintf(buf, "%d\n", (int)IS_ENABLED(CONFIG_MEMORY_HOTREMOVE));
|
2008-07-24 11:28:19 +07:00
|
|
|
}
|
|
|
|
|
2005-10-30 08:16:54 +07:00
|
|
|
/*
|
|
|
|
* online, offline, going offline, etc.
|
|
|
|
*/
|
2018-12-03 18:16:11 +07:00
|
|
|
static ssize_t state_show(struct device *dev, struct device_attribute *attr,
|
|
|
|
char *buf)
|
2005-10-30 08:16:54 +07:00
|
|
|
{
|
2013-08-28 13:38:27 +07:00
|
|
|
struct memory_block *mem = to_memory_block(dev);
|
2005-10-30 08:16:54 +07:00
|
|
|
ssize_t len = 0;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* We can probably put these states in a nice little array
|
|
|
|
* so that they're not open-coded
|
|
|
|
*/
|
|
|
|
switch (mem->state) {
|
2015-03-08 17:29:04 +07:00
|
|
|
case MEM_ONLINE:
|
|
|
|
len = sprintf(buf, "online\n");
|
|
|
|
break;
|
|
|
|
case MEM_OFFLINE:
|
|
|
|
len = sprintf(buf, "offline\n");
|
|
|
|
break;
|
|
|
|
case MEM_GOING_OFFLINE:
|
|
|
|
len = sprintf(buf, "going-offline\n");
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
len = sprintf(buf, "ERROR-UNKNOWN-%ld\n",
|
|
|
|
mem->state);
|
|
|
|
WARN_ON(1);
|
|
|
|
break;
|
2005-10-30 08:16:54 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
return len;
|
|
|
|
}
|
|
|
|
|
2007-10-22 06:41:36 +07:00
|
|
|
int memory_notify(unsigned long val, void *v)
|
2005-10-30 08:16:54 +07:00
|
|
|
{
|
[PATCH] Notifier chain update: API changes
The kernel's implementation of notifier chains is unsafe. There is no
protection against entries being added to or removed from a chain while the
chain is in use. The issues were discussed in this thread:
http://marc.theaimsgroup.com/?l=linux-kernel&m=113018709002036&w=2
We noticed that notifier chains in the kernel fall into two basic usage
classes:
"Blocking" chains are always called from a process context
and the callout routines are allowed to sleep;
"Atomic" chains can be called from an atomic context and
the callout routines are not allowed to sleep.
We decided to codify this distinction and make it part of the API. Therefore
this set of patches introduces three new, parallel APIs: one for blocking
notifiers, one for atomic notifiers, and one for "raw" notifiers (which is
really just the old API under a new name). New kinds of data structures are
used for the heads of the chains, and new routines are defined for
registration, unregistration, and calling a chain. The three APIs are
explained in include/linux/notifier.h and their implementation is in
kernel/sys.c.
With atomic and blocking chains, the implementation guarantees that the chain
links will not be corrupted and that chain callers will not get messed up by
entries being added or removed. For raw chains the implementation provides no
guarantees at all; users of this API must provide their own protections. (The
idea was that situations may come up where the assumptions of the atomic and
blocking APIs are not appropriate, so it should be possible for users to
handle these things in their own way.)
There are some limitations, which should not be too hard to live with. For
atomic/blocking chains, registration and unregistration must always be done in
a process context since the chain is protected by a mutex/rwsem. Also, a
callout routine for a non-raw chain must not try to register or unregister
entries on its own chain. (This did happen in a couple of places and the code
had to be changed to avoid it.)
Since atomic chains may be called from within an NMI handler, they cannot use
spinlocks for synchronization. Instead we use RCU. The overhead falls almost
entirely in the unregister routine, which is okay since unregistration is much
less frequent that calling a chain.
Here is the list of chains that we adjusted and their classifications. None
of them use the raw API, so for the moment it is only a placeholder.
ATOMIC CHAINS
-------------
arch/i386/kernel/traps.c: i386die_chain
arch/ia64/kernel/traps.c: ia64die_chain
arch/powerpc/kernel/traps.c: powerpc_die_chain
arch/sparc64/kernel/traps.c: sparc64die_chain
arch/x86_64/kernel/traps.c: die_chain
drivers/char/ipmi/ipmi_si_intf.c: xaction_notifier_list
kernel/panic.c: panic_notifier_list
kernel/profile.c: task_free_notifier
net/bluetooth/hci_core.c: hci_notifier
net/ipv4/netfilter/ip_conntrack_core.c: ip_conntrack_chain
net/ipv4/netfilter/ip_conntrack_core.c: ip_conntrack_expect_chain
net/ipv6/addrconf.c: inet6addr_chain
net/netfilter/nf_conntrack_core.c: nf_conntrack_chain
net/netfilter/nf_conntrack_core.c: nf_conntrack_expect_chain
net/netlink/af_netlink.c: netlink_chain
BLOCKING CHAINS
---------------
arch/powerpc/platforms/pseries/reconfig.c: pSeries_reconfig_chain
arch/s390/kernel/process.c: idle_chain
arch/x86_64/kernel/process.c idle_notifier
drivers/base/memory.c: memory_chain
drivers/cpufreq/cpufreq.c cpufreq_policy_notifier_list
drivers/cpufreq/cpufreq.c cpufreq_transition_notifier_list
drivers/macintosh/adb.c: adb_client_list
drivers/macintosh/via-pmu.c sleep_notifier_list
drivers/macintosh/via-pmu68k.c sleep_notifier_list
drivers/macintosh/windfarm_core.c wf_client_list
drivers/usb/core/notify.c usb_notifier_list
drivers/video/fbmem.c fb_notifier_list
kernel/cpu.c cpu_chain
kernel/module.c module_notify_list
kernel/profile.c munmap_notifier
kernel/profile.c task_exit_notifier
kernel/sys.c reboot_notifier_list
net/core/dev.c netdev_chain
net/decnet/dn_dev.c: dnaddr_chain
net/ipv4/devinet.c: inetaddr_chain
It's possible that some of these classifications are wrong. If they are,
please let us know or submit a patch to fix them. Note that any chain that
gets called very frequently should be atomic, because the rwsem read-locking
used for blocking chains is very likely to incur cache misses on SMP systems.
(However, if the chain's callout routines may sleep then the chain cannot be
atomic.)
The patch set was written by Alan Stern and Chandra Seetharaman, incorporating
material written by Keith Owens and suggestions from Paul McKenney and Andrew
Morton.
[jes@sgi.com: restructure the notifier chain initialization macros]
Signed-off-by: Alan Stern <stern@rowland.harvard.edu>
Signed-off-by: Chandra Seetharaman <sekharan@us.ibm.com>
Signed-off-by: Jes Sorensen <jes@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-03-27 16:16:30 +07:00
|
|
|
return blocking_notifier_call_chain(&memory_chain, val, v);
|
2005-10-30 08:16:54 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* MEMORY_HOTPLUG depends on SPARSEMEM in mm/Kconfig, so it is
|
|
|
|
* OK to have direct references to sparsemem variables in here.
|
|
|
|
*/
|
|
|
|
static int
|
2019-05-14 07:19:35 +07:00
|
|
|
memory_block_action(unsigned long start_section_nr, unsigned long action,
|
2020-01-31 13:14:54 +07:00
|
|
|
int online_type, int nid)
|
2005-10-30 08:16:54 +07:00
|
|
|
{
|
2012-10-09 06:33:58 +07:00
|
|
|
unsigned long start_pfn;
|
2011-05-11 14:25:14 +07:00
|
|
|
unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block;
|
2005-10-30 08:16:54 +07:00
|
|
|
int ret;
|
|
|
|
|
2019-05-14 07:19:35 +07:00
|
|
|
start_pfn = section_nr_to_pfn(start_section_nr);
|
2011-10-19 04:00:57 +07:00
|
|
|
|
2005-10-30 08:16:54 +07:00
|
|
|
switch (action) {
|
2015-03-08 17:29:04 +07:00
|
|
|
case MEM_ONLINE:
|
2020-01-31 13:14:54 +07:00
|
|
|
ret = online_pages(start_pfn, nr_pages, online_type, nid);
|
2015-03-08 17:29:04 +07:00
|
|
|
break;
|
|
|
|
case MEM_OFFLINE:
|
|
|
|
ret = offline_pages(start_pfn, nr_pages);
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
WARN(1, KERN_WARNING "%s(%ld, %ld) unknown action: "
|
2019-05-14 07:19:35 +07:00
|
|
|
"%ld\n", __func__, start_section_nr, action, action);
|
2015-03-08 17:29:04 +07:00
|
|
|
ret = -EINVAL;
|
2005-10-30 08:16:54 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2017-02-25 06:00:02 +07:00
|
|
|
static int memory_block_change_state(struct memory_block *mem,
|
2013-08-21 04:05:05 +07:00
|
|
|
unsigned long to_state, unsigned long from_state_req)
|
2005-10-30 08:16:54 +07:00
|
|
|
{
|
2011-10-19 04:00:57 +07:00
|
|
|
int ret = 0;
|
2011-01-20 23:43:34 +07:00
|
|
|
|
2013-05-08 19:18:37 +07:00
|
|
|
if (mem->state != from_state_req)
|
|
|
|
return -EINVAL;
|
2005-10-30 08:16:54 +07:00
|
|
|
|
2011-01-20 23:43:34 +07:00
|
|
|
if (to_state == MEM_OFFLINE)
|
|
|
|
mem->state = MEM_GOING_OFFLINE;
|
|
|
|
|
2013-08-21 04:05:05 +07:00
|
|
|
ret = memory_block_action(mem->start_section_nr, to_state,
|
2020-01-31 13:14:54 +07:00
|
|
|
mem->online_type, mem->nid);
|
2013-08-21 04:05:05 +07:00
|
|
|
|
2013-05-23 15:38:55 +07:00
|
|
|
mem->state = ret ? from_state_req : to_state;
|
2013-08-21 04:05:05 +07:00
|
|
|
|
2013-05-08 19:18:37 +07:00
|
|
|
return ret;
|
|
|
|
}
|
2011-01-20 23:43:34 +07:00
|
|
|
|
2013-08-21 04:05:05 +07:00
|
|
|
/* The device lock serializes operations on memory_subsys_[online|offline] */
|
2013-05-08 19:18:37 +07:00
|
|
|
static int memory_subsys_online(struct device *dev)
|
|
|
|
{
|
2013-08-28 13:38:27 +07:00
|
|
|
struct memory_block *mem = to_memory_block(dev);
|
2013-05-08 19:18:37 +07:00
|
|
|
int ret;
|
2005-10-30 08:16:54 +07:00
|
|
|
|
2013-08-21 04:05:05 +07:00
|
|
|
if (mem->state == MEM_ONLINE)
|
|
|
|
return 0;
|
2013-05-08 19:18:37 +07:00
|
|
|
|
2013-08-21 04:05:05 +07:00
|
|
|
/*
|
2020-04-07 10:07:20 +07:00
|
|
|
* When called via device_online() without configuring the online_type,
|
|
|
|
* we want to default to MMOP_ONLINE.
|
2013-08-21 04:05:05 +07:00
|
|
|
*/
|
2020-04-07 10:07:20 +07:00
|
|
|
if (mem->online_type == MMOP_OFFLINE)
|
drivers/base/memory: rename MMOP_ONLINE_KEEP to MMOP_ONLINE
Patch series "mm/memory_hotplug: allow to specify a default online_type", v3.
Distributions nowadays use udev rules ([1] [2]) to specify if and how to
online hotplugged memory. The rules seem to get more complex with many
special cases. Due to the various special cases,
CONFIG_MEMORY_HOTPLUG_DEFAULT_ONLINE cannot be used. All memory hotplug
is handled via udev rules.
Every time we hotplug memory, the udev rule will come to the same
conclusion. Especially Hyper-V (but also soon virtio-mem) add a lot of
memory in separate memory blocks and wait for memory to get onlined by
user space before continuing to add more memory blocks (to not add memory
faster than it is getting onlined). This of course slows down the whole
memory hotplug process.
To make the job of distributions easier and to avoid udev rules that get
more and more complicated, let's extend the mechanism provided by
- /sys/devices/system/memory/auto_online_blocks
- "memhp_default_state=" on the kernel cmdline
to be able to specify also "online_movable" as well as "online_kernel"
=== Example /usr/libexec/config-memhotplug ===
#!/bin/bash
VIRT=`systemd-detect-virt --vm`
ARCH=`uname -p`
sense_virtio_mem() {
if [ -d "/sys/bus/virtio/drivers/virtio_mem/" ]; then
DEVICES=`find /sys/bus/virtio/drivers/virtio_mem/ -maxdepth 1 -type l | wc -l`
if [ $DEVICES != "0" ]; then
return 0
fi
fi
return 1
}
if [ ! -e "/sys/devices/system/memory/auto_online_blocks" ]; then
echo "Memory hotplug configuration support missing in the kernel"
exit 1
fi
if grep "memhp_default_state=" /proc/cmdline > /dev/null; then
echo "Memory hotplug configuration overridden in kernel cmdline (memhp_default_state=)"
exit 1
fi
if [ $VIRT == "microsoft" ]; then
echo "Detected Hyper-V on $ARCH"
# Hyper-V wants all memory in ZONE_NORMAL
ONLINE_TYPE="online_kernel"
elif sense_virtio_mem; then
echo "Detected virtio-mem on $ARCH"
# virtio-mem wants all memory in ZONE_NORMAL
ONLINE_TYPE="online_kernel"
elif [ $ARCH == "s390x" ] || [ $ARCH == "s390" ]; then
echo "Detected $ARCH"
# standby memory should not be onlined automatically
ONLINE_TYPE="offline"
elif [ $ARCH == "ppc64" ] || [ $ARCH == "ppc64le" ]; then
echo "Detected" $ARCH
# PPC64 onlines all hotplugged memory right from the kernel
ONLINE_TYPE="offline"
elif [ $VIRT == "none" ]; then
echo "Detected bare-metal on $ARCH"
# Bare metal users expect hotplugged memory to be unpluggable. We assume
# that ZONE imbalances on such enterpise servers cannot happen and is
# properly documented
ONLINE_TYPE="online_movable"
else
# TODO: Hypervisors that want to unplug DIMMs and can guarantee that ZONE
# imbalances won't happen
echo "Detected $VIRT on $ARCH"
# Usually, ballooning is used in virtual environments, so memory should go to
# ZONE_NORMAL. However, sometimes "movable_node" is relevant.
ONLINE_TYPE="online"
fi
echo "Selected online_type:" $ONLINE_TYPE
# Configure what to do with memory that will be hotplugged in the future
echo $ONLINE_TYPE 2>/dev/null > /sys/devices/system/memory/auto_online_blocks
if [ $? != "0" ]; then
echo "Memory hotplug cannot be configured (e.g., old kernel or missing permissions)"
# A backup udev rule should handle old kernels if necessary
exit 1
fi
# Process all already pluggedd blocks (e.g., DIMMs, but also Hyper-V or virtio-mem)
if [ $ONLINE_TYPE != "offline" ]; then
for MEMORY in /sys/devices/system/memory/memory*; do
STATE=`cat $MEMORY/state`
if [ $STATE == "offline" ]; then
echo $ONLINE_TYPE > $MEMORY/state
fi
done
fi
=== Example /usr/lib/systemd/system/config-memhotplug.service ===
[Unit]
Description=Configure memory hotplug behavior
DefaultDependencies=no
Conflicts=shutdown.target
Before=sysinit.target shutdown.target
After=systemd-modules-load.service
ConditionPathExists=|/sys/devices/system/memory/auto_online_blocks
[Service]
ExecStart=/usr/libexec/config-memhotplug
Type=oneshot
TimeoutSec=0
RemainAfterExit=yes
[Install]
WantedBy=sysinit.target
=== Example modification to the 40-redhat.rules [2] ===
: diff --git a/40-redhat.rules b/40-redhat.rules-new
: index 2c690e5..168fd03 100644
: --- a/40-redhat.rules
: +++ b/40-redhat.rules-new
: @@ -6,6 +6,9 @@ SUBSYSTEM=="cpu", ACTION=="add", TEST=="online", ATTR{online}=="0", ATTR{online}
: # Memory hotadd request
: SUBSYSTEM!="memory", GOTO="memory_hotplug_end"
: ACTION!="add", GOTO="memory_hotplug_end"
: +# memory hotplug behavior configured
: +PROGRAM=="grep online /sys/devices/system/memory/auto_online_blocks", GOTO="memory_hotplug_end"
: +
: PROGRAM="/bin/uname -p", RESULT=="s390*", GOTO="memory_hotplug_end"
:
: ENV{.state}="online"
===
[1] https://github.com/lnykryn/systemd-rhel/pull/281
[2] https://github.com/lnykryn/systemd-rhel/blob/staging/rules/40-redhat.rules
This patch (of 8):
The name is misleading and it's not really clear what is "kept". Let's
just name it like the online_type name we expose to user space ("online").
Add some documentation to the types.
Signed-off-by: David Hildenbrand <david@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Reviewed-by: Wei Yang <richard.weiyang@gmail.com>
Reviewed-by: Baoquan He <bhe@redhat.com>
Acked-by: Pankaj Gupta <pankaj.gupta.linux@gmail.com>
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Oscar Salvador <osalvador@suse.de>
Cc: "Rafael J. Wysocki" <rafael@kernel.org>
Cc: Wei Yang <richard.weiyang@gmail.com>
Cc: Vitaly Kuznetsov <vkuznets@redhat.com>
Cc: Yumei Huang <yuhuang@redhat.com>
Cc: Igor Mammedov <imammedo@redhat.com>
Cc: Eduardo Habkost <ehabkost@redhat.com>
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
Cc: Haiyang Zhang <haiyangz@microsoft.com>
Cc: K. Y. Srinivasan <kys@microsoft.com>
Cc: Michael Ellerman <mpe@ellerman.id.au> (powerpc)
Cc: Paul Mackerras <paulus@samba.org>
Cc: Stephen Hemminger <sthemmin@microsoft.com>
Cc: Wei Liu <wei.liu@kernel.org>
Link: http://lkml.kernel.org/r/20200319131221.14044-1-david@redhat.com
Link: http://lkml.kernel.org/r/20200317104942.11178-2-david@redhat.com
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2020-04-07 10:07:16 +07:00
|
|
|
mem->online_type = MMOP_ONLINE;
|
2013-05-08 19:18:37 +07:00
|
|
|
|
2013-08-21 04:05:05 +07:00
|
|
|
ret = memory_block_change_state(mem, MEM_ONLINE, MEM_OFFLINE);
|
2020-04-07 10:07:20 +07:00
|
|
|
mem->online_type = MMOP_OFFLINE;
|
2013-05-08 19:18:37 +07:00
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int memory_subsys_offline(struct device *dev)
|
2012-10-09 06:34:01 +07:00
|
|
|
{
|
2013-08-28 13:38:27 +07:00
|
|
|
struct memory_block *mem = to_memory_block(dev);
|
2012-10-09 06:34:01 +07:00
|
|
|
|
2013-08-21 04:05:05 +07:00
|
|
|
if (mem->state == MEM_OFFLINE)
|
|
|
|
return 0;
|
2012-10-09 06:34:01 +07:00
|
|
|
|
2013-08-21 04:05:05 +07:00
|
|
|
return memory_block_change_state(mem, MEM_OFFLINE, MEM_ONLINE);
|
2012-10-09 06:34:01 +07:00
|
|
|
}
|
2013-05-08 19:18:37 +07:00
|
|
|
|
2018-12-03 18:16:11 +07:00
|
|
|
static ssize_t state_store(struct device *dev, struct device_attribute *attr,
|
|
|
|
const char *buf, size_t count)
|
2005-10-30 08:16:54 +07:00
|
|
|
{
|
2020-04-07 10:07:24 +07:00
|
|
|
const int online_type = memhp_online_type_from_str(buf);
|
2013-08-28 13:38:27 +07:00
|
|
|
struct memory_block *mem = to_memory_block(dev);
|
2020-04-07 10:07:24 +07:00
|
|
|
int ret;
|
|
|
|
|
|
|
|
if (online_type < 0)
|
|
|
|
return -EINVAL;
|
2005-10-30 08:16:54 +07:00
|
|
|
|
driver core / ACPI: Avoid device hot remove locking issues
device_hotplug_lock is held around the acpi_bus_trim() call in
acpi_scan_hot_remove() which generally removes devices (it removes
ACPI device objects at least, but it may also remove "physical"
device objects through .detach() callbacks of ACPI scan handlers).
Thus, potentially, device sysfs attributes are removed under that
lock and to remove those attributes it is necessary to hold the
s_active references of their directory entries for writing.
On the other hand, the execution of a .show() or .store() callback
from a sysfs attribute is carried out with that attribute's s_active
reference held for reading. Consequently, if any device sysfs
attribute that may be removed from within acpi_scan_hot_remove()
through acpi_bus_trim() has a .store() or .show() callback which
acquires device_hotplug_lock, the execution of that callback may
deadlock with the removal of the attribute. [Unfortunately, the
"online" device attribute of CPUs and memory blocks is one of them.]
To avoid such deadlocks, make all of the sysfs attribute callbacks
that need to lock device hotplug, for example store_online(), use
a special function, lock_device_hotplug_sysfs(), to lock device
hotplug and return the result of that function immediately if it is
not zero. This will cause the s_active reference of the directory
entry in question to be released and the syscall to be restarted
if device_hotplug_lock cannot be acquired.
[show_online() actually doesn't need to lock device hotplug, but
it is useful to serialize it with respect to device_offline() and
device_online() for the same device (in case user space attempts to
run them concurrently) which can be done with the help of
device_lock().]
Reported-by: Yasuaki Ishimatsu <isimatu.yasuaki@jp.fujitsu.com>
Reported-and-tested-by: Gu Zheng <guz.fnst@cn.fujitsu.com>
Suggested-by: Tejun Heo <tj@kernel.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
Acked-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Acked-by: Toshi Kani <toshi.kani@hp.com>
2013-08-29 02:41:01 +07:00
|
|
|
ret = lock_device_hotplug_sysfs();
|
|
|
|
if (ret)
|
|
|
|
return ret;
|
2013-05-08 19:18:37 +07:00
|
|
|
|
2013-08-21 04:05:05 +07:00
|
|
|
switch (online_type) {
|
2014-08-07 06:05:13 +07:00
|
|
|
case MMOP_ONLINE_KERNEL:
|
|
|
|
case MMOP_ONLINE_MOVABLE:
|
drivers/base/memory: rename MMOP_ONLINE_KEEP to MMOP_ONLINE
Patch series "mm/memory_hotplug: allow to specify a default online_type", v3.
Distributions nowadays use udev rules ([1] [2]) to specify if and how to
online hotplugged memory. The rules seem to get more complex with many
special cases. Due to the various special cases,
CONFIG_MEMORY_HOTPLUG_DEFAULT_ONLINE cannot be used. All memory hotplug
is handled via udev rules.
Every time we hotplug memory, the udev rule will come to the same
conclusion. Especially Hyper-V (but also soon virtio-mem) add a lot of
memory in separate memory blocks and wait for memory to get onlined by
user space before continuing to add more memory blocks (to not add memory
faster than it is getting onlined). This of course slows down the whole
memory hotplug process.
To make the job of distributions easier and to avoid udev rules that get
more and more complicated, let's extend the mechanism provided by
- /sys/devices/system/memory/auto_online_blocks
- "memhp_default_state=" on the kernel cmdline
to be able to specify also "online_movable" as well as "online_kernel"
=== Example /usr/libexec/config-memhotplug ===
#!/bin/bash
VIRT=`systemd-detect-virt --vm`
ARCH=`uname -p`
sense_virtio_mem() {
if [ -d "/sys/bus/virtio/drivers/virtio_mem/" ]; then
DEVICES=`find /sys/bus/virtio/drivers/virtio_mem/ -maxdepth 1 -type l | wc -l`
if [ $DEVICES != "0" ]; then
return 0
fi
fi
return 1
}
if [ ! -e "/sys/devices/system/memory/auto_online_blocks" ]; then
echo "Memory hotplug configuration support missing in the kernel"
exit 1
fi
if grep "memhp_default_state=" /proc/cmdline > /dev/null; then
echo "Memory hotplug configuration overridden in kernel cmdline (memhp_default_state=)"
exit 1
fi
if [ $VIRT == "microsoft" ]; then
echo "Detected Hyper-V on $ARCH"
# Hyper-V wants all memory in ZONE_NORMAL
ONLINE_TYPE="online_kernel"
elif sense_virtio_mem; then
echo "Detected virtio-mem on $ARCH"
# virtio-mem wants all memory in ZONE_NORMAL
ONLINE_TYPE="online_kernel"
elif [ $ARCH == "s390x" ] || [ $ARCH == "s390" ]; then
echo "Detected $ARCH"
# standby memory should not be onlined automatically
ONLINE_TYPE="offline"
elif [ $ARCH == "ppc64" ] || [ $ARCH == "ppc64le" ]; then
echo "Detected" $ARCH
# PPC64 onlines all hotplugged memory right from the kernel
ONLINE_TYPE="offline"
elif [ $VIRT == "none" ]; then
echo "Detected bare-metal on $ARCH"
# Bare metal users expect hotplugged memory to be unpluggable. We assume
# that ZONE imbalances on such enterpise servers cannot happen and is
# properly documented
ONLINE_TYPE="online_movable"
else
# TODO: Hypervisors that want to unplug DIMMs and can guarantee that ZONE
# imbalances won't happen
echo "Detected $VIRT on $ARCH"
# Usually, ballooning is used in virtual environments, so memory should go to
# ZONE_NORMAL. However, sometimes "movable_node" is relevant.
ONLINE_TYPE="online"
fi
echo "Selected online_type:" $ONLINE_TYPE
# Configure what to do with memory that will be hotplugged in the future
echo $ONLINE_TYPE 2>/dev/null > /sys/devices/system/memory/auto_online_blocks
if [ $? != "0" ]; then
echo "Memory hotplug cannot be configured (e.g., old kernel or missing permissions)"
# A backup udev rule should handle old kernels if necessary
exit 1
fi
# Process all already pluggedd blocks (e.g., DIMMs, but also Hyper-V or virtio-mem)
if [ $ONLINE_TYPE != "offline" ]; then
for MEMORY in /sys/devices/system/memory/memory*; do
STATE=`cat $MEMORY/state`
if [ $STATE == "offline" ]; then
echo $ONLINE_TYPE > $MEMORY/state
fi
done
fi
=== Example /usr/lib/systemd/system/config-memhotplug.service ===
[Unit]
Description=Configure memory hotplug behavior
DefaultDependencies=no
Conflicts=shutdown.target
Before=sysinit.target shutdown.target
After=systemd-modules-load.service
ConditionPathExists=|/sys/devices/system/memory/auto_online_blocks
[Service]
ExecStart=/usr/libexec/config-memhotplug
Type=oneshot
TimeoutSec=0
RemainAfterExit=yes
[Install]
WantedBy=sysinit.target
=== Example modification to the 40-redhat.rules [2] ===
: diff --git a/40-redhat.rules b/40-redhat.rules-new
: index 2c690e5..168fd03 100644
: --- a/40-redhat.rules
: +++ b/40-redhat.rules-new
: @@ -6,6 +6,9 @@ SUBSYSTEM=="cpu", ACTION=="add", TEST=="online", ATTR{online}=="0", ATTR{online}
: # Memory hotadd request
: SUBSYSTEM!="memory", GOTO="memory_hotplug_end"
: ACTION!="add", GOTO="memory_hotplug_end"
: +# memory hotplug behavior configured
: +PROGRAM=="grep online /sys/devices/system/memory/auto_online_blocks", GOTO="memory_hotplug_end"
: +
: PROGRAM="/bin/uname -p", RESULT=="s390*", GOTO="memory_hotplug_end"
:
: ENV{.state}="online"
===
[1] https://github.com/lnykryn/systemd-rhel/pull/281
[2] https://github.com/lnykryn/systemd-rhel/blob/staging/rules/40-redhat.rules
This patch (of 8):
The name is misleading and it's not really clear what is "kept". Let's
just name it like the online_type name we expose to user space ("online").
Add some documentation to the types.
Signed-off-by: David Hildenbrand <david@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Reviewed-by: Wei Yang <richard.weiyang@gmail.com>
Reviewed-by: Baoquan He <bhe@redhat.com>
Acked-by: Pankaj Gupta <pankaj.gupta.linux@gmail.com>
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Oscar Salvador <osalvador@suse.de>
Cc: "Rafael J. Wysocki" <rafael@kernel.org>
Cc: Wei Yang <richard.weiyang@gmail.com>
Cc: Vitaly Kuznetsov <vkuznets@redhat.com>
Cc: Yumei Huang <yuhuang@redhat.com>
Cc: Igor Mammedov <imammedo@redhat.com>
Cc: Eduardo Habkost <ehabkost@redhat.com>
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
Cc: Haiyang Zhang <haiyangz@microsoft.com>
Cc: K. Y. Srinivasan <kys@microsoft.com>
Cc: Michael Ellerman <mpe@ellerman.id.au> (powerpc)
Cc: Paul Mackerras <paulus@samba.org>
Cc: Stephen Hemminger <sthemmin@microsoft.com>
Cc: Wei Liu <wei.liu@kernel.org>
Link: http://lkml.kernel.org/r/20200319131221.14044-1-david@redhat.com
Link: http://lkml.kernel.org/r/20200317104942.11178-2-david@redhat.com
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2020-04-07 10:07:16 +07:00
|
|
|
case MMOP_ONLINE:
|
mm/memory_hotplug: fix online/offline_pages called w.o. mem_hotplug_lock
There seem to be some problems as result of 30467e0b3be ("mm, hotplug:
fix concurrent memory hot-add deadlock"), which tried to fix a possible
lock inversion reported and discussed in [1] due to the two locks
a) device_lock()
b) mem_hotplug_lock
While add_memory() first takes b), followed by a) during
bus_probe_device(), onlining of memory from user space first took a),
followed by b), exposing a possible deadlock.
In [1], and it was decided to not make use of device_hotplug_lock, but
rather to enforce a locking order.
The problems I spotted related to this:
1. Memory block device attributes: While .state first calls
mem_hotplug_begin() and the calls device_online() - which takes
device_lock() - .online does no longer call mem_hotplug_begin(), so
effectively calls online_pages() without mem_hotplug_lock.
2. device_online() should be called under device_hotplug_lock, however
onlining memory during add_memory() does not take care of that.
In addition, I think there is also something wrong about the locking in
3. arch/powerpc/platforms/powernv/memtrace.c calls offline_pages()
without locks. This was introduced after 30467e0b3be. And skimming over
the code, I assume it could need some more care in regards to locking
(e.g. device_online() called without device_hotplug_lock. This will
be addressed in the following patches.
Now that we hold the device_hotplug_lock when
- adding memory (e.g. via add_memory()/add_memory_resource())
- removing memory (e.g. via remove_memory())
- device_online()/device_offline()
We can move mem_hotplug_lock usage back into
online_pages()/offline_pages().
Why is mem_hotplug_lock still needed? Essentially to make
get_online_mems()/put_online_mems() be very fast (relying on
device_hotplug_lock would be very slow), and to serialize against
addition of memory that does not create memory block devices (hmm).
[1] http://driverdev.linuxdriverproject.org/pipermail/ driverdev-devel/
2015-February/065324.html
This patch is partly based on a patch by Vitaly Kuznetsov.
Link: http://lkml.kernel.org/r/20180925091457.28651-4-david@redhat.com
Signed-off-by: David Hildenbrand <david@redhat.com>
Reviewed-by: Pavel Tatashin <pavel.tatashin@microsoft.com>
Reviewed-by: Rashmica Gupta <rashmica.g@gmail.com>
Reviewed-by: Oscar Salvador <osalvador@suse.de>
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
Cc: Paul Mackerras <paulus@samba.org>
Cc: Michael Ellerman <mpe@ellerman.id.au>
Cc: "Rafael J. Wysocki" <rjw@rjwysocki.net>
Cc: Len Brown <lenb@kernel.org>
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: "K. Y. Srinivasan" <kys@microsoft.com>
Cc: Haiyang Zhang <haiyangz@microsoft.com>
Cc: Stephen Hemminger <sthemmin@microsoft.com>
Cc: Martin Schwidefsky <schwidefsky@de.ibm.com>
Cc: Heiko Carstens <heiko.carstens@de.ibm.com>
Cc: Boris Ostrovsky <boris.ostrovsky@oracle.com>
Cc: Juergen Gross <jgross@suse.com>
Cc: Rashmica Gupta <rashmica.g@gmail.com>
Cc: Michael Neuling <mikey@neuling.org>
Cc: Balbir Singh <bsingharora@gmail.com>
Cc: Kate Stewart <kstewart@linuxfoundation.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: Philippe Ombredanne <pombredanne@nexb.com>
Cc: Pavel Tatashin <pavel.tatashin@microsoft.com>
Cc: Vlastimil Babka <vbabka@suse.cz>
Cc: Dan Williams <dan.j.williams@intel.com>
Cc: Oscar Salvador <osalvador@suse.de>
Cc: YASUAKI ISHIMATSU <yasu.isimatu@gmail.com>
Cc: Mathieu Malaterre <malat@debian.org>
Cc: John Allen <jallen@linux.vnet.ibm.com>
Cc: Jonathan Corbet <corbet@lwn.net>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Nathan Fontenot <nfont@linux.vnet.ibm.com>
Cc: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2018-10-31 05:10:29 +07:00
|
|
|
/* mem->online_type is protected by device_hotplug_lock */
|
2013-08-21 04:05:05 +07:00
|
|
|
mem->online_type = online_type;
|
|
|
|
ret = device_online(&mem->dev);
|
|
|
|
break;
|
2014-08-07 06:05:13 +07:00
|
|
|
case MMOP_OFFLINE:
|
2013-08-21 04:05:05 +07:00
|
|
|
ret = device_offline(&mem->dev);
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
ret = -EINVAL; /* should never happen */
|
2013-05-08 19:18:37 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
unlock_device_hotplug();
|
2011-01-20 23:43:34 +07:00
|
|
|
|
2016-10-08 07:00:15 +07:00
|
|
|
if (ret < 0)
|
2005-10-30 08:16:54 +07:00
|
|
|
return ret;
|
2016-10-08 07:00:15 +07:00
|
|
|
if (ret)
|
|
|
|
return -EINVAL;
|
|
|
|
|
2005-10-30 08:16:54 +07:00
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* phys_device is a bad name for this. What I really want
|
|
|
|
* is a way to differentiate between memory ranges that
|
|
|
|
* are part of physical devices that constitute
|
|
|
|
* a complete removable unit or fru.
|
|
|
|
* i.e. do these ranges belong to the same physical device,
|
|
|
|
* s.t. if I offline all of these sections I can then
|
|
|
|
* remove the physical device?
|
|
|
|
*/
|
2018-12-03 18:16:11 +07:00
|
|
|
static ssize_t phys_device_show(struct device *dev,
|
2011-12-22 05:48:43 +07:00
|
|
|
struct device_attribute *attr, char *buf)
|
2005-10-30 08:16:54 +07:00
|
|
|
{
|
2013-08-28 13:38:27 +07:00
|
|
|
struct memory_block *mem = to_memory_block(dev);
|
2005-10-30 08:16:54 +07:00
|
|
|
return sprintf(buf, "%d\n", mem->phys_device);
|
|
|
|
}
|
|
|
|
|
2014-10-10 05:26:31 +07:00
|
|
|
#ifdef CONFIG_MEMORY_HOTREMOVE
|
2017-09-07 06:19:37 +07:00
|
|
|
static void print_allowed_zone(char *buf, int nid, unsigned long start_pfn,
|
|
|
|
unsigned long nr_pages, int online_type,
|
|
|
|
struct zone *default_zone)
|
|
|
|
{
|
|
|
|
struct zone *zone;
|
|
|
|
|
|
|
|
zone = zone_for_pfn_range(online_type, nid, start_pfn, nr_pages);
|
|
|
|
if (zone != default_zone) {
|
|
|
|
strcat(buf, " ");
|
|
|
|
strcat(buf, zone->name);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-12-03 18:16:11 +07:00
|
|
|
static ssize_t valid_zones_show(struct device *dev,
|
2014-10-10 05:26:31 +07:00
|
|
|
struct device_attribute *attr, char *buf)
|
|
|
|
{
|
|
|
|
struct memory_block *mem = to_memory_block(dev);
|
2017-07-07 05:38:11 +07:00
|
|
|
unsigned long start_pfn = section_nr_to_pfn(mem->start_section_nr);
|
2014-10-10 05:26:31 +07:00
|
|
|
unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block;
|
2017-09-07 06:19:37 +07:00
|
|
|
struct zone *default_zone;
|
2017-07-07 05:38:11 +07:00
|
|
|
int nid;
|
2014-10-10 05:26:31 +07:00
|
|
|
|
2017-07-07 05:38:11 +07:00
|
|
|
/*
|
|
|
|
* Check the existing zone. Make sure that we do that only on the
|
|
|
|
* online nodes otherwise the page_zone is not reliable
|
|
|
|
*/
|
|
|
|
if (mem->state == MEM_ONLINE) {
|
2018-09-05 05:46:09 +07:00
|
|
|
/*
|
|
|
|
* The block contains more than one zone can not be offlined.
|
|
|
|
* This can happen e.g. for ZONE_DMA and ZONE_DMA32
|
|
|
|
*/
|
2020-02-04 08:34:26 +07:00
|
|
|
default_zone = test_pages_in_a_zone(start_pfn,
|
|
|
|
start_pfn + nr_pages);
|
|
|
|
if (!default_zone)
|
2018-09-05 05:46:09 +07:00
|
|
|
return sprintf(buf, "none\n");
|
2020-02-04 08:34:26 +07:00
|
|
|
strcat(buf, default_zone->name);
|
2017-07-07 05:38:11 +07:00
|
|
|
goto out;
|
2014-10-10 05:26:31 +07:00
|
|
|
}
|
|
|
|
|
2018-09-05 05:46:09 +07:00
|
|
|
nid = mem->nid;
|
drivers/base/memory: rename MMOP_ONLINE_KEEP to MMOP_ONLINE
Patch series "mm/memory_hotplug: allow to specify a default online_type", v3.
Distributions nowadays use udev rules ([1] [2]) to specify if and how to
online hotplugged memory. The rules seem to get more complex with many
special cases. Due to the various special cases,
CONFIG_MEMORY_HOTPLUG_DEFAULT_ONLINE cannot be used. All memory hotplug
is handled via udev rules.
Every time we hotplug memory, the udev rule will come to the same
conclusion. Especially Hyper-V (but also soon virtio-mem) add a lot of
memory in separate memory blocks and wait for memory to get onlined by
user space before continuing to add more memory blocks (to not add memory
faster than it is getting onlined). This of course slows down the whole
memory hotplug process.
To make the job of distributions easier and to avoid udev rules that get
more and more complicated, let's extend the mechanism provided by
- /sys/devices/system/memory/auto_online_blocks
- "memhp_default_state=" on the kernel cmdline
to be able to specify also "online_movable" as well as "online_kernel"
=== Example /usr/libexec/config-memhotplug ===
#!/bin/bash
VIRT=`systemd-detect-virt --vm`
ARCH=`uname -p`
sense_virtio_mem() {
if [ -d "/sys/bus/virtio/drivers/virtio_mem/" ]; then
DEVICES=`find /sys/bus/virtio/drivers/virtio_mem/ -maxdepth 1 -type l | wc -l`
if [ $DEVICES != "0" ]; then
return 0
fi
fi
return 1
}
if [ ! -e "/sys/devices/system/memory/auto_online_blocks" ]; then
echo "Memory hotplug configuration support missing in the kernel"
exit 1
fi
if grep "memhp_default_state=" /proc/cmdline > /dev/null; then
echo "Memory hotplug configuration overridden in kernel cmdline (memhp_default_state=)"
exit 1
fi
if [ $VIRT == "microsoft" ]; then
echo "Detected Hyper-V on $ARCH"
# Hyper-V wants all memory in ZONE_NORMAL
ONLINE_TYPE="online_kernel"
elif sense_virtio_mem; then
echo "Detected virtio-mem on $ARCH"
# virtio-mem wants all memory in ZONE_NORMAL
ONLINE_TYPE="online_kernel"
elif [ $ARCH == "s390x" ] || [ $ARCH == "s390" ]; then
echo "Detected $ARCH"
# standby memory should not be onlined automatically
ONLINE_TYPE="offline"
elif [ $ARCH == "ppc64" ] || [ $ARCH == "ppc64le" ]; then
echo "Detected" $ARCH
# PPC64 onlines all hotplugged memory right from the kernel
ONLINE_TYPE="offline"
elif [ $VIRT == "none" ]; then
echo "Detected bare-metal on $ARCH"
# Bare metal users expect hotplugged memory to be unpluggable. We assume
# that ZONE imbalances on such enterpise servers cannot happen and is
# properly documented
ONLINE_TYPE="online_movable"
else
# TODO: Hypervisors that want to unplug DIMMs and can guarantee that ZONE
# imbalances won't happen
echo "Detected $VIRT on $ARCH"
# Usually, ballooning is used in virtual environments, so memory should go to
# ZONE_NORMAL. However, sometimes "movable_node" is relevant.
ONLINE_TYPE="online"
fi
echo "Selected online_type:" $ONLINE_TYPE
# Configure what to do with memory that will be hotplugged in the future
echo $ONLINE_TYPE 2>/dev/null > /sys/devices/system/memory/auto_online_blocks
if [ $? != "0" ]; then
echo "Memory hotplug cannot be configured (e.g., old kernel or missing permissions)"
# A backup udev rule should handle old kernels if necessary
exit 1
fi
# Process all already pluggedd blocks (e.g., DIMMs, but also Hyper-V or virtio-mem)
if [ $ONLINE_TYPE != "offline" ]; then
for MEMORY in /sys/devices/system/memory/memory*; do
STATE=`cat $MEMORY/state`
if [ $STATE == "offline" ]; then
echo $ONLINE_TYPE > $MEMORY/state
fi
done
fi
=== Example /usr/lib/systemd/system/config-memhotplug.service ===
[Unit]
Description=Configure memory hotplug behavior
DefaultDependencies=no
Conflicts=shutdown.target
Before=sysinit.target shutdown.target
After=systemd-modules-load.service
ConditionPathExists=|/sys/devices/system/memory/auto_online_blocks
[Service]
ExecStart=/usr/libexec/config-memhotplug
Type=oneshot
TimeoutSec=0
RemainAfterExit=yes
[Install]
WantedBy=sysinit.target
=== Example modification to the 40-redhat.rules [2] ===
: diff --git a/40-redhat.rules b/40-redhat.rules-new
: index 2c690e5..168fd03 100644
: --- a/40-redhat.rules
: +++ b/40-redhat.rules-new
: @@ -6,6 +6,9 @@ SUBSYSTEM=="cpu", ACTION=="add", TEST=="online", ATTR{online}=="0", ATTR{online}
: # Memory hotadd request
: SUBSYSTEM!="memory", GOTO="memory_hotplug_end"
: ACTION!="add", GOTO="memory_hotplug_end"
: +# memory hotplug behavior configured
: +PROGRAM=="grep online /sys/devices/system/memory/auto_online_blocks", GOTO="memory_hotplug_end"
: +
: PROGRAM="/bin/uname -p", RESULT=="s390*", GOTO="memory_hotplug_end"
:
: ENV{.state}="online"
===
[1] https://github.com/lnykryn/systemd-rhel/pull/281
[2] https://github.com/lnykryn/systemd-rhel/blob/staging/rules/40-redhat.rules
This patch (of 8):
The name is misleading and it's not really clear what is "kept". Let's
just name it like the online_type name we expose to user space ("online").
Add some documentation to the types.
Signed-off-by: David Hildenbrand <david@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Reviewed-by: Wei Yang <richard.weiyang@gmail.com>
Reviewed-by: Baoquan He <bhe@redhat.com>
Acked-by: Pankaj Gupta <pankaj.gupta.linux@gmail.com>
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Oscar Salvador <osalvador@suse.de>
Cc: "Rafael J. Wysocki" <rafael@kernel.org>
Cc: Wei Yang <richard.weiyang@gmail.com>
Cc: Vitaly Kuznetsov <vkuznets@redhat.com>
Cc: Yumei Huang <yuhuang@redhat.com>
Cc: Igor Mammedov <imammedo@redhat.com>
Cc: Eduardo Habkost <ehabkost@redhat.com>
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
Cc: Haiyang Zhang <haiyangz@microsoft.com>
Cc: K. Y. Srinivasan <kys@microsoft.com>
Cc: Michael Ellerman <mpe@ellerman.id.au> (powerpc)
Cc: Paul Mackerras <paulus@samba.org>
Cc: Stephen Hemminger <sthemmin@microsoft.com>
Cc: Wei Liu <wei.liu@kernel.org>
Link: http://lkml.kernel.org/r/20200319131221.14044-1-david@redhat.com
Link: http://lkml.kernel.org/r/20200317104942.11178-2-david@redhat.com
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2020-04-07 10:07:16 +07:00
|
|
|
default_zone = zone_for_pfn_range(MMOP_ONLINE, nid, start_pfn,
|
|
|
|
nr_pages);
|
2017-09-07 06:19:37 +07:00
|
|
|
strcat(buf, default_zone->name);
|
2014-10-10 05:26:31 +07:00
|
|
|
|
2017-09-07 06:19:37 +07:00
|
|
|
print_allowed_zone(buf, nid, start_pfn, nr_pages, MMOP_ONLINE_KERNEL,
|
|
|
|
default_zone);
|
|
|
|
print_allowed_zone(buf, nid, start_pfn, nr_pages, MMOP_ONLINE_MOVABLE,
|
|
|
|
default_zone);
|
2017-07-07 05:38:11 +07:00
|
|
|
out:
|
2016-07-27 05:22:27 +07:00
|
|
|
strcat(buf, "\n");
|
|
|
|
|
|
|
|
return strlen(buf);
|
2014-10-10 05:26:31 +07:00
|
|
|
}
|
2018-12-03 18:16:11 +07:00
|
|
|
static DEVICE_ATTR_RO(valid_zones);
|
2014-10-10 05:26:31 +07:00
|
|
|
#endif
|
|
|
|
|
2018-12-03 18:16:11 +07:00
|
|
|
static DEVICE_ATTR_RO(phys_index);
|
|
|
|
static DEVICE_ATTR_RW(state);
|
|
|
|
static DEVICE_ATTR_RO(phys_device);
|
|
|
|
static DEVICE_ATTR_RO(removable);
|
2005-10-30 08:16:54 +07:00
|
|
|
|
|
|
|
/*
|
2019-09-24 05:35:43 +07:00
|
|
|
* Show the memory block size (shared by all memory blocks).
|
2005-10-30 08:16:54 +07:00
|
|
|
*/
|
2018-12-03 18:16:11 +07:00
|
|
|
static ssize_t block_size_bytes_show(struct device *dev,
|
|
|
|
struct device_attribute *attr, char *buf)
|
2005-10-30 08:16:54 +07:00
|
|
|
{
|
2019-09-24 05:35:46 +07:00
|
|
|
return sprintf(buf, "%lx\n", memory_block_size_bytes());
|
2005-10-30 08:16:54 +07:00
|
|
|
}
|
|
|
|
|
2018-12-03 18:16:11 +07:00
|
|
|
static DEVICE_ATTR_RO(block_size_bytes);
|
2005-10-30 08:16:54 +07:00
|
|
|
|
2016-03-16 04:56:48 +07:00
|
|
|
/*
|
|
|
|
* Memory auto online policy.
|
|
|
|
*/
|
|
|
|
|
2018-12-03 18:16:11 +07:00
|
|
|
static ssize_t auto_online_blocks_show(struct device *dev,
|
|
|
|
struct device_attribute *attr, char *buf)
|
2016-03-16 04:56:48 +07:00
|
|
|
{
|
2020-04-07 10:07:40 +07:00
|
|
|
return sprintf(buf, "%s\n",
|
|
|
|
online_type_to_str[memhp_default_online_type]);
|
2016-03-16 04:56:48 +07:00
|
|
|
}
|
|
|
|
|
2018-12-03 18:16:11 +07:00
|
|
|
static ssize_t auto_online_blocks_store(struct device *dev,
|
|
|
|
struct device_attribute *attr,
|
|
|
|
const char *buf, size_t count)
|
2016-03-16 04:56:48 +07:00
|
|
|
{
|
mm/memory_hotplug: allow to specify a default online_type
For now, distributions implement advanced udev rules to essentially
- Don't online any hotplugged memory (s390x)
- Online all memory to ZONE_NORMAL (e.g., most virt environments like
hyperv)
- Online all memory to ZONE_MOVABLE in case the zone imbalance is taken
care of (e.g., bare metal, special virt environments)
In summary: All memory is usually onlined the same way, however, the
kernel always has to ask user space to come up with the same answer.
E.g., Hyper-V always waits for a memory block to get onlined before
continuing, otherwise it might end up adding memory faster than
onlining it, which can result in strange OOM situations. This waiting
slows down adding of a bigger amount of memory.
Let's allow to specify a default online_type, not just "online" and
"offline". This allows distributions to configure the default online_type
when booting up and be done with it.
We can now specify "offline", "online", "online_movable" and
"online_kernel" via
- "memhp_default_state=" on the kernel cmdline
- /sys/devices/system/memory/auto_online_blocks
just like we are able to specify for a single memory block via
/sys/devices/system/memory/memoryX/state
Signed-off-by: David Hildenbrand <david@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Reviewed-by: Wei Yang <richard.weiyang@gmail.com>
Reviewed-by: Baoquan He <bhe@redhat.com>
Acked-by: Michal Hocko <mhocko@suse.com>
Acked-by: Pankaj Gupta <pankaj.gupta.linux@gmail.com>
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: Oscar Salvador <osalvador@suse.de>
Cc: "Rafael J. Wysocki" <rafael@kernel.org>
Cc: Wei Yang <richard.weiyang@gmail.com>
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
Cc: Eduardo Habkost <ehabkost@redhat.com>
Cc: Haiyang Zhang <haiyangz@microsoft.com>
Cc: Igor Mammedov <imammedo@redhat.com>
Cc: "K. Y. Srinivasan" <kys@microsoft.com>
Cc: Michael Ellerman <mpe@ellerman.id.au>
Cc: Paul Mackerras <paulus@samba.org>
Cc: Stephen Hemminger <sthemmin@microsoft.com>
Cc: Vitaly Kuznetsov <vkuznets@redhat.com>
Cc: Wei Liu <wei.liu@kernel.org>
Cc: Yumei Huang <yuhuang@redhat.com>
Link: http://lkml.kernel.org/r/20200317104942.11178-9-david@redhat.com
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2020-04-07 10:07:44 +07:00
|
|
|
const int online_type = memhp_online_type_from_str(buf);
|
|
|
|
|
|
|
|
if (online_type < 0)
|
2016-03-16 04:56:48 +07:00
|
|
|
return -EINVAL;
|
|
|
|
|
mm/memory_hotplug: allow to specify a default online_type
For now, distributions implement advanced udev rules to essentially
- Don't online any hotplugged memory (s390x)
- Online all memory to ZONE_NORMAL (e.g., most virt environments like
hyperv)
- Online all memory to ZONE_MOVABLE in case the zone imbalance is taken
care of (e.g., bare metal, special virt environments)
In summary: All memory is usually onlined the same way, however, the
kernel always has to ask user space to come up with the same answer.
E.g., Hyper-V always waits for a memory block to get onlined before
continuing, otherwise it might end up adding memory faster than
onlining it, which can result in strange OOM situations. This waiting
slows down adding of a bigger amount of memory.
Let's allow to specify a default online_type, not just "online" and
"offline". This allows distributions to configure the default online_type
when booting up and be done with it.
We can now specify "offline", "online", "online_movable" and
"online_kernel" via
- "memhp_default_state=" on the kernel cmdline
- /sys/devices/system/memory/auto_online_blocks
just like we are able to specify for a single memory block via
/sys/devices/system/memory/memoryX/state
Signed-off-by: David Hildenbrand <david@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Reviewed-by: Wei Yang <richard.weiyang@gmail.com>
Reviewed-by: Baoquan He <bhe@redhat.com>
Acked-by: Michal Hocko <mhocko@suse.com>
Acked-by: Pankaj Gupta <pankaj.gupta.linux@gmail.com>
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: Oscar Salvador <osalvador@suse.de>
Cc: "Rafael J. Wysocki" <rafael@kernel.org>
Cc: Wei Yang <richard.weiyang@gmail.com>
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
Cc: Eduardo Habkost <ehabkost@redhat.com>
Cc: Haiyang Zhang <haiyangz@microsoft.com>
Cc: Igor Mammedov <imammedo@redhat.com>
Cc: "K. Y. Srinivasan" <kys@microsoft.com>
Cc: Michael Ellerman <mpe@ellerman.id.au>
Cc: Paul Mackerras <paulus@samba.org>
Cc: Stephen Hemminger <sthemmin@microsoft.com>
Cc: Vitaly Kuznetsov <vkuznets@redhat.com>
Cc: Wei Liu <wei.liu@kernel.org>
Cc: Yumei Huang <yuhuang@redhat.com>
Link: http://lkml.kernel.org/r/20200317104942.11178-9-david@redhat.com
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2020-04-07 10:07:44 +07:00
|
|
|
memhp_default_online_type = online_type;
|
2016-03-16 04:56:48 +07:00
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
2018-12-03 18:16:11 +07:00
|
|
|
static DEVICE_ATTR_RW(auto_online_blocks);
|
2016-03-16 04:56:48 +07:00
|
|
|
|
2005-10-30 08:16:54 +07:00
|
|
|
/*
|
|
|
|
* Some architectures will have custom drivers to do this, and
|
|
|
|
* will not need to do it from userspace. The fake hot-add code
|
|
|
|
* as well as ppc64 will do all of their discovery in userspace
|
|
|
|
* and will require this interface.
|
|
|
|
*/
|
|
|
|
#ifdef CONFIG_ARCH_MEMORY_PROBE
|
2018-12-03 18:16:11 +07:00
|
|
|
static ssize_t probe_store(struct device *dev, struct device_attribute *attr,
|
|
|
|
const char *buf, size_t count)
|
2005-10-30 08:16:54 +07:00
|
|
|
{
|
|
|
|
u64 phys_addr;
|
2016-01-15 06:22:16 +07:00
|
|
|
int nid, ret;
|
2011-09-15 03:26:15 +07:00
|
|
|
unsigned long pages_per_block = PAGES_PER_SECTION * sections_per_block;
|
2005-10-30 08:16:54 +07:00
|
|
|
|
2014-08-07 06:06:06 +07:00
|
|
|
ret = kstrtoull(buf, 0, &phys_addr);
|
|
|
|
if (ret)
|
|
|
|
return ret;
|
2005-10-30 08:16:54 +07:00
|
|
|
|
2011-09-15 03:26:15 +07:00
|
|
|
if (phys_addr & ((pages_per_block << PAGE_SHIFT) - 1))
|
|
|
|
return -EINVAL;
|
|
|
|
|
2018-10-31 05:10:24 +07:00
|
|
|
ret = lock_device_hotplug_sysfs();
|
|
|
|
if (ret)
|
2019-04-19 07:50:16 +07:00
|
|
|
return ret;
|
2018-10-31 05:10:24 +07:00
|
|
|
|
2016-01-15 06:22:16 +07:00
|
|
|
nid = memory_add_physaddr_to_nid(phys_addr);
|
2018-10-31 05:10:24 +07:00
|
|
|
ret = __add_memory(nid, phys_addr,
|
|
|
|
MIN_MEMORY_BLOCK_SIZE * sections_per_block);
|
2011-01-31 23:55:23 +07:00
|
|
|
|
2016-01-15 06:22:16 +07:00
|
|
|
if (ret)
|
|
|
|
goto out;
|
2005-10-30 08:16:54 +07:00
|
|
|
|
2011-03-24 13:16:18 +07:00
|
|
|
ret = count;
|
|
|
|
out:
|
2018-10-31 05:10:24 +07:00
|
|
|
unlock_device_hotplug();
|
2011-03-24 13:16:18 +07:00
|
|
|
return ret;
|
2005-10-30 08:16:54 +07:00
|
|
|
}
|
|
|
|
|
2018-12-03 18:16:11 +07:00
|
|
|
static DEVICE_ATTR_WO(probe);
|
2005-10-30 08:16:54 +07:00
|
|
|
#endif
|
|
|
|
|
2009-12-16 18:20:00 +07:00
|
|
|
#ifdef CONFIG_MEMORY_FAILURE
|
|
|
|
/*
|
|
|
|
* Support for offlining pages of memory
|
|
|
|
*/
|
|
|
|
|
|
|
|
/* Soft offline a page */
|
2018-12-03 18:16:11 +07:00
|
|
|
static ssize_t soft_offline_page_store(struct device *dev,
|
|
|
|
struct device_attribute *attr,
|
|
|
|
const char *buf, size_t count)
|
2009-12-16 18:20:00 +07:00
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
u64 pfn;
|
|
|
|
if (!capable(CAP_SYS_ADMIN))
|
|
|
|
return -EPERM;
|
2013-07-26 11:10:22 +07:00
|
|
|
if (kstrtoull(buf, 0, &pfn) < 0)
|
2009-12-16 18:20:00 +07:00
|
|
|
return -EINVAL;
|
|
|
|
pfn >>= PAGE_SHIFT;
|
2019-12-01 08:53:38 +07:00
|
|
|
ret = soft_offline_page(pfn, 0);
|
2009-12-16 18:20:00 +07:00
|
|
|
return ret == 0 ? count : ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Forcibly offline a page, including killing processes. */
|
2018-12-03 18:16:11 +07:00
|
|
|
static ssize_t hard_offline_page_store(struct device *dev,
|
|
|
|
struct device_attribute *attr,
|
|
|
|
const char *buf, size_t count)
|
2009-12-16 18:20:00 +07:00
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
u64 pfn;
|
|
|
|
if (!capable(CAP_SYS_ADMIN))
|
|
|
|
return -EPERM;
|
2013-07-26 11:10:22 +07:00
|
|
|
if (kstrtoull(buf, 0, &pfn) < 0)
|
2009-12-16 18:20:00 +07:00
|
|
|
return -EINVAL;
|
|
|
|
pfn >>= PAGE_SHIFT;
|
2017-07-10 06:14:01 +07:00
|
|
|
ret = memory_failure(pfn, 0);
|
2009-12-16 18:20:00 +07:00
|
|
|
return ret ? ret : count;
|
|
|
|
}
|
|
|
|
|
2018-12-03 18:16:11 +07:00
|
|
|
static DEVICE_ATTR_WO(soft_offline_page);
|
|
|
|
static DEVICE_ATTR_WO(hard_offline_page);
|
2009-12-16 18:20:00 +07:00
|
|
|
#endif
|
|
|
|
|
2005-10-30 08:16:54 +07:00
|
|
|
/*
|
|
|
|
* Note that phys_device is optional. It is here to allow for
|
|
|
|
* differentiation between which *physical* devices each
|
|
|
|
* section belongs to...
|
|
|
|
*/
|
2010-03-15 11:35:03 +07:00
|
|
|
int __weak arch_get_memory_phys_device(unsigned long start_pfn)
|
|
|
|
{
|
|
|
|
return 0;
|
|
|
|
}
|
2005-10-30 08:16:54 +07:00
|
|
|
|
2019-07-19 05:57:53 +07:00
|
|
|
/* A reference for the returned memory block device is acquired. */
|
|
|
|
static struct memory_block *find_memory_block_by_id(unsigned long block_id)
|
2005-10-30 08:16:54 +07:00
|
|
|
{
|
2011-12-22 05:48:43 +07:00
|
|
|
struct device *dev;
|
2005-10-30 08:16:54 +07:00
|
|
|
|
2019-07-19 05:57:53 +07:00
|
|
|
dev = subsys_find_device_by_id(&memory_subsys, block_id, NULL);
|
|
|
|
return dev ? to_memory_block(dev) : NULL;
|
2019-07-19 05:56:56 +07:00
|
|
|
}
|
|
|
|
|
2010-09-30 02:00:55 +07:00
|
|
|
/*
|
|
|
|
* For now, we have a linear search to go find the appropriate
|
|
|
|
* memory_block corresponding to a particular phys_index. If
|
|
|
|
* this gets to be a real problem, we can always use a radix
|
|
|
|
* tree or something here.
|
|
|
|
*
|
2011-12-22 05:48:43 +07:00
|
|
|
* This could be made generic for all device subsystems.
|
2010-09-30 02:00:55 +07:00
|
|
|
*/
|
|
|
|
struct memory_block *find_memory_block(struct mem_section *section)
|
|
|
|
{
|
2019-07-19 05:57:53 +07:00
|
|
|
unsigned long block_id = base_memory_block_id(__section_nr(section));
|
|
|
|
|
|
|
|
return find_memory_block_by_id(block_id);
|
2010-09-30 02:00:55 +07:00
|
|
|
}
|
|
|
|
|
2013-06-05 02:42:28 +07:00
|
|
|
static struct attribute *memory_memblk_attrs[] = {
|
|
|
|
&dev_attr_phys_index.attr,
|
|
|
|
&dev_attr_state.attr,
|
|
|
|
&dev_attr_phys_device.attr,
|
|
|
|
&dev_attr_removable.attr,
|
2014-10-10 05:26:31 +07:00
|
|
|
#ifdef CONFIG_MEMORY_HOTREMOVE
|
|
|
|
&dev_attr_valid_zones.attr,
|
|
|
|
#endif
|
2013-06-05 02:42:28 +07:00
|
|
|
NULL
|
|
|
|
};
|
|
|
|
|
|
|
|
static struct attribute_group memory_memblk_attr_group = {
|
|
|
|
.attrs = memory_memblk_attrs,
|
|
|
|
};
|
|
|
|
|
|
|
|
static const struct attribute_group *memory_memblk_attr_groups[] = {
|
|
|
|
&memory_memblk_attr_group,
|
|
|
|
NULL,
|
|
|
|
};
|
|
|
|
|
|
|
|
/*
|
|
|
|
* register_memory - Setup a sysfs device for a memory block
|
|
|
|
*/
|
|
|
|
static
|
|
|
|
int register_memory(struct memory_block *memory)
|
|
|
|
{
|
2018-04-26 22:42:09 +07:00
|
|
|
int ret;
|
|
|
|
|
2013-06-05 02:42:28 +07:00
|
|
|
memory->dev.bus = &memory_subsys;
|
|
|
|
memory->dev.id = memory->start_section_nr / sections_per_block;
|
|
|
|
memory->dev.release = memory_block_release;
|
|
|
|
memory->dev.groups = memory_memblk_attr_groups;
|
Power management and ACPI updates for 3.11-rc1
- Hotplug changes allowing device hot-removal operations to fail
gracefully (instead of crashing the kernel) if they cannot be
carried out completely. From Rafael J Wysocki and Toshi Kani.
- Freezer update from Colin Cross and Mandeep Singh Baines targeted
at making the freezing of tasks a bit less heavy weight operation.
- cpufreq resume fix from Srivatsa S Bhat for a regression introduced
during the 3.10 cycle causing some cpufreq sysfs attributes to
return wrong values to user space after resume.
- New freqdomain_cpus sysfs attribute for the acpi-cpufreq driver to
provide information previously available via related_cpus from
Lan Tianyu.
- cpufreq fixes and cleanups from Viresh Kumar, Jacob Shin,
Heiko Stübner, Xiaoguang Chen, Ezequiel Garcia, Arnd Bergmann, and
Tang Yuantian.
- Fix for an ACPICA regression causing suspend/resume issues to
appear on some systems introduced during the 3.4 development cycle
from Lv Zheng.
- ACPICA fixes and cleanups from Bob Moore, Tomasz Nowicki, Lv Zheng,
Chao Guan, and Zhang Rui.
- New cupidle driver for Xilinx Zynq processors from Michal Simek.
- cpuidle fixes and cleanups from Daniel Lezcano.
- Changes to make suspend/resume work correctly in Xen guests from
Konrad Rzeszutek Wilk.
- ACPI device power management fixes and cleanups from Fengguang Wu
and Rafael J Wysocki.
- ACPI documentation updates from Lv Zheng, Aaron Lu and Hanjun Guo.
- Fix for the IA-64 issue that was the reason for reverting commit
9f29ab1 and updates of the ACPI scan code from Rafael J Wysocki.
- Mechanism for adding CMOS RTC address space handlers from Lan Tianyu
(to allow some EC-related breakage to be fixed on some systems).
- Spec-compliant implementation of acpi_os_get_timer() from
Mika Westerberg.
- Modification of do_acpi_find_child() to execute _STA in order to
to avoid situations in which a pointer to a disabled device object
is returned instead of an enabled one with the same _ADR value.
From Jeff Wu.
- Intel BayTrail PCH (Platform Controller Hub) support for the ACPI
Intel Low-Power Subsystems (LPSS) driver and modificaions of that
driver to work around a couple of known BIOS issues from
Mika Westerberg and Heikki Krogerus.
- EC driver fix from Vasiliy Kulikov to make it use get_user() and
put_user() instead of dereferencing user space pointers blindly.
- Assorted ACPI code cleanups from Bjorn Helgaas, Nicholas Mazzuca and
Toshi Kani.
- Modification of the "runtime idle" helper routine to take the return
values of the callbacks executed by it into account and to call
rpm_suspend() if they return 0, which allows some code bloat
reduction to be done, from Rafael J Wysocki and Alan Stern.
- New trace points for PM QoS from Sahara <keun-o.park@windriver.com>.
- PM QoS documentation update from Lan Tianyu.
- Assorted core PM code cleanups and changes from Bernie Thompson,
Bjorn Helgaas, Julius Werner, and Shuah Khan.
- New devfreq driver for the Exynos5-bus device from Abhilash Kesavan.
- Minor devfreq cleanups, fixes and MAINTAINERS update from
MyungJoo Ham, Abhilash Kesavan, Paul Bolle, Rajagopal Venkat, and
Wei Yongjun.
- OMAP Adaptive Voltage Scaling (AVS) SmartReflex voltage control
driver updates from Andrii Tseglytskyi and Nishanth Menon.
/
-----BEGIN PGP SIGNATURE-----
Version: GnuPG v2.0.19 (GNU/Linux)
iQIcBAABAgAGBQJR0ZNOAAoJEKhOf7ml8uNsDLYP/0EU4rmvw0TWTITfp6RS1KDE
9GwBn96ZR4Q5bJd9gBCTPSqhHOYMqxWEUp99sn/M2wehG1pk/jw5LO56+2IhM3UZ
g1HDcJ7te2nVT/iXsKiAGTVhU9Rk0aYwoVSknwk27qpIBGxW9w/s5tLX8pY3Q3Zq
wL/7aTPjyL+PFFFEaxgH7qLqsl3DhbtYW5AriUBTkXout/tJ4eO1b7MNBncLDh8X
VQ/0DNCKE95VEJfkO4rk9RKUyVp9GDn0i+HXCD/FS4IA5oYzePdVdNDmXf7g+swe
CGlTZq8pB+oBpDiHl4lxzbNrKQjRNbGnDUkoRcWqn0nAw56xK+vmYnWJhW99gQ/I
fKnvxeLca5po1aiqmC4VSJxZIatFZqLrZAI4dzoCLWY+bGeTnCKmj0/F8ytFnZA2
8IuLLs7/dFOaHXV/pKmpg6FAlFa9CPxoqRFoyqb4M0GjEarADyalXUWsPtG+6xCp
R/p0CISpwk+guKZR/qPhL7M654S7SHrPwd2DPF0KgGsvk+G2GhoB8EzvD8BVp98Z
9siCGCdgKQfJQVI6R0k9aFmn/4gRQIAgyPhkhv9tqULUUkiaXki+/t8kPfnb8O/d
zep+CA57E2G8MYLkDJfpFeKS7GpPD6TIdgFdGmOUC0Y6sl9iTdiw4yTx8O2JM37z
rHBZfYGkJBrbGRu+Q1gs
=VBBq
-----END PGP SIGNATURE-----
Merge tag 'pm+acpi-3.11-rc1' of git://git.kernel.org/pub/scm/linux/kernel/git/rafael/linux-pm
Pull power management and ACPI updates from Rafael Wysocki:
"This time the total number of ACPI commits is slightly greater than
the number of cpufreq commits, but Viresh Kumar (who works on cpufreq)
remains the most active patch submitter.
To me, the most significant change is the addition of offline/online
device operations to the driver core (with the Greg's blessing) and
the related modifications of the ACPI core hotplug code. Next are the
freezer updates from Colin Cross that should make the freezing of
tasks a bit less heavy weight.
We also have a couple of regression fixes, a number of fixes for
issues that have not been identified as regressions, two new drivers
and a bunch of cleanups all over.
Highlights:
- Hotplug changes to support graceful hot-removal failures.
It sometimes is necessary to fail device hot-removal operations
gracefully if they cannot be carried out completely. For example,
if memory from a memory module being hot-removed has been allocated
for the kernel's own use and cannot be moved elsewhere, it's
desirable to fail the hot-removal operation in a graceful way
rather than to crash the kernel, but currenty a success or a kernel
crash are the only possible outcomes of an attempted memory
hot-removal. Needless to say, that is not a very attractive
alternative and it had to be addressed.
However, in order to make it work for memory, I first had to make
it work for CPUs and for this purpose I needed to modify the ACPI
processor driver. It's been split into two parts, a resident one
handling the low-level initialization/cleanup and a modular one
playing the actual driver's role (but it binds to the CPU system
device objects rather than to the ACPI device objects representing
processors). That's been sort of like a live brain surgery on a
patient who's riding a bike.
So this is a little scary, but since we found and fixed a couple of
regressions it caused to happen during the early linux-next testing
(a month ago), nobody has complained.
As a bonus we remove some duplicated ACPI hotplug code, because the
ACPI-based CPU hotplug is now going to use the common ACPI hotplug
code.
- Lighter weight freezing of tasks.
These changes from Colin Cross and Mandeep Singh Baines are
targeted at making the freezing of tasks a bit less heavy weight
operation. They reduce the number of tasks woken up every time
during the freezing, by using the observation that the freezer
simply doesn't need to wake up some of them and wait for them all
to call refrigerator(). The time needed for the freezer to decide
to report a failure is reduced too.
Also reintroduced is the check causing a lockdep warining to
trigger when try_to_freeze() is called with locks held (which is
generally unsafe and shouldn't happen).
- cpufreq updates
First off, a commit from Srivatsa S Bhat fixes a resume regression
introduced during the 3.10 cycle causing some cpufreq sysfs
attributes to return wrong values to user space after resume. The
fix is kind of fresh, but also it's pretty obvious once Srivatsa
has identified the root cause.
Second, we have a new freqdomain_cpus sysfs attribute for the
acpi-cpufreq driver to provide information previously available via
related_cpus. From Lan Tianyu.
Finally, we fix a number of issues, mostly related to the
CPUFREQ_POSTCHANGE notifier and cpufreq Kconfig options and clean
up some code. The majority of changes from Viresh Kumar with bits
from Jacob Shin, Heiko Stübner, Xiaoguang Chen, Ezequiel Garcia,
Arnd Bergmann, and Tang Yuantian.
- ACPICA update
A usual bunch of updates from the ACPICA upstream.
During the 3.4 cycle we introduced support for ACPI 5 extended
sleep registers, but they are only supposed to be used if the
HW-reduced mode bit is set in the FADT flags and the code attempted
to use them without checking that bit. That caused suspend/resume
regressions to happen on some systems. Fix from Lv Zheng causes
those registers to be used only if the HW-reduced mode bit is set.
Apart from this some other ACPICA bugs are fixed and code cleanups
are made by Bob Moore, Tomasz Nowicki, Lv Zheng, Chao Guan, and
Zhang Rui.
- cpuidle updates
New driver for Xilinx Zynq processors is added by Michal Simek.
Multidriver support simplification, addition of some missing
kerneldoc comments and Kconfig-related fixes come from Daniel
Lezcano.
- ACPI power management updates
Changes to make suspend/resume work correctly in Xen guests from
Konrad Rzeszutek Wilk, sparse warning fix from Fengguang Wu and
cleanups and fixes of the ACPI device power state selection
routine.
- ACPI documentation updates
Some previously missing pieces of ACPI documentation are added by
Lv Zheng and Aaron Lu (hopefully, that will help people to
uderstand how the ACPI subsystem works) and one outdated doc is
updated by Hanjun Guo.
- Assorted ACPI updates
We finally nailed down the IA-64 issue that was the reason for
reverting commit 9f29ab11ddbf ("ACPI / scan: do not match drivers
against objects having scan handlers"), so we can fix it and move
the ACPI scan handler check added to the ACPI video driver back to
the core.
A mechanism for adding CMOS RTC address space handlers is
introduced by Lan Tianyu to allow some EC-related breakage to be
fixed on some systems.
A spec-compliant implementation of acpi_os_get_timer() is added by
Mika Westerberg.
The evaluation of _STA is added to do_acpi_find_child() to avoid
situations in which a pointer to a disabled device object is
returned instead of an enabled one with the same _ADR value. From
Jeff Wu.
Intel BayTrail PCH (Platform Controller Hub) support is added to
the ACPI driver for Intel Low-Power Subsystems (LPSS) and that
driver is modified to work around a couple of known BIOS issues.
Changes from Mika Westerberg and Heikki Krogerus.
The EC driver is fixed by Vasiliy Kulikov to use get_user() and
put_user() instead of dereferencing user space pointers blindly.
Code cleanups are made by Bjorn Helgaas, Nicholas Mazzuca and Toshi
Kani.
- Assorted power management updates
The "runtime idle" helper routine is changed to take the return
values of the callbacks executed by it into account and to call
rpm_suspend() if they return 0, which allows us to reduce the
overall code bloat a bit (by dropping some code that's not
necessary any more after that modification).
The runtime PM documentation is updated by Alan Stern (to reflect
the "runtime idle" behavior change).
New trace points for PM QoS are added by Sahara
(<keun-o.park@windriver.com>).
PM QoS documentation is updated by Lan Tianyu.
Code cleanups are made and minor issues are addressed by Bernie
Thompson, Bjorn Helgaas, Julius Werner, and Shuah Khan.
- devfreq updates
New driver for the Exynos5-bus device from Abhilash Kesavan.
Minor cleanups, fixes and MAINTAINERS update from MyungJoo Ham,
Abhilash Kesavan, Paul Bolle, Rajagopal Venkat, and Wei Yongjun.
- OMAP power management updates
Adaptive Voltage Scaling (AVS) SmartReflex voltage control driver
updates from Andrii Tseglytskyi and Nishanth Menon."
* tag 'pm+acpi-3.11-rc1' of git://git.kernel.org/pub/scm/linux/kernel/git/rafael/linux-pm: (162 commits)
cpufreq: Fix cpufreq regression after suspend/resume
ACPI / PM: Fix possible NULL pointer deref in acpi_pm_device_sleep_state()
PM / Sleep: Warn about system time after resume with pm_trace
cpufreq: don't leave stale policy pointer in cdbs->cur_policy
acpi-cpufreq: Add new sysfs attribute freqdomain_cpus
cpufreq: make sure frequency transitions are serialized
ACPI: implement acpi_os_get_timer() according the spec
ACPI / EC: Add HP Folio 13 to ec_dmi_table in order to skip DSDT scan
ACPI: Add CMOS RTC Operation Region handler support
ACPI / processor: Drop unused variable from processor_perflib.c
cpufreq: tegra: call CPUFREQ_POSTCHANGE notfier in error cases
cpufreq: s3c64xx: call CPUFREQ_POSTCHANGE notfier in error cases
cpufreq: omap: call CPUFREQ_POSTCHANGE notfier in error cases
cpufreq: imx6q: call CPUFREQ_POSTCHANGE notfier in error cases
cpufreq: exynos: call CPUFREQ_POSTCHANGE notfier in error cases
cpufreq: dbx500: call CPUFREQ_POSTCHANGE notfier in error cases
cpufreq: davinci: call CPUFREQ_POSTCHANGE notfier in error cases
cpufreq: arm-big-little: call CPUFREQ_POSTCHANGE notfier in error cases
cpufreq: powernow-k8: call CPUFREQ_POSTCHANGE notfier in error cases
cpufreq: pcc: call CPUFREQ_POSTCHANGE notfier in error cases
...
2013-07-04 04:35:40 +07:00
|
|
|
memory->dev.offline = memory->state == MEM_OFFLINE;
|
2013-06-05 02:42:28 +07:00
|
|
|
|
2018-04-26 22:42:09 +07:00
|
|
|
ret = device_register(&memory->dev);
|
|
|
|
if (ret)
|
|
|
|
put_device(&memory->dev);
|
|
|
|
|
|
|
|
return ret;
|
2013-06-05 02:42:28 +07:00
|
|
|
}
|
|
|
|
|
2019-07-19 05:57:40 +07:00
|
|
|
static int init_memory_block(struct memory_block **memory,
|
|
|
|
unsigned long block_id, unsigned long state)
|
2010-10-20 00:44:20 +07:00
|
|
|
{
|
2011-01-20 23:43:34 +07:00
|
|
|
struct memory_block *mem;
|
2010-10-20 00:44:20 +07:00
|
|
|
unsigned long start_pfn;
|
|
|
|
int ret = 0;
|
|
|
|
|
2019-07-19 05:57:53 +07:00
|
|
|
mem = find_memory_block_by_id(block_id);
|
2019-07-19 05:56:56 +07:00
|
|
|
if (mem) {
|
|
|
|
put_device(&mem->dev);
|
|
|
|
return -EEXIST;
|
|
|
|
}
|
2011-01-20 23:43:34 +07:00
|
|
|
mem = kzalloc(sizeof(*mem), GFP_KERNEL);
|
2010-10-20 00:44:20 +07:00
|
|
|
if (!mem)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
2019-07-19 05:56:46 +07:00
|
|
|
mem->start_section_nr = block_id * sections_per_block;
|
2010-10-20 00:44:20 +07:00
|
|
|
mem->state = state;
|
2011-01-20 23:44:29 +07:00
|
|
|
start_pfn = section_nr_to_pfn(mem->start_section_nr);
|
2010-10-20 00:44:20 +07:00
|
|
|
mem->phys_device = arch_get_memory_phys_device(start_pfn);
|
2019-09-24 05:35:40 +07:00
|
|
|
mem->nid = NUMA_NO_NODE;
|
2010-10-20 00:44:20 +07:00
|
|
|
|
2011-01-20 23:43:34 +07:00
|
|
|
ret = register_memory(mem);
|
|
|
|
|
|
|
|
*memory = mem;
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2019-07-19 05:57:37 +07:00
|
|
|
static int add_memory_block(unsigned long base_section_nr)
|
2011-01-20 23:43:34 +07:00
|
|
|
{
|
2020-04-07 10:06:40 +07:00
|
|
|
int section_count = 0;
|
2013-08-21 00:13:03 +07:00
|
|
|
struct memory_block *mem;
|
2019-07-19 05:57:37 +07:00
|
|
|
unsigned long nr;
|
2011-01-20 23:43:34 +07:00
|
|
|
|
2019-07-19 05:57:37 +07:00
|
|
|
for (nr = base_section_nr; nr < base_section_nr + sections_per_block;
|
|
|
|
nr++)
|
|
|
|
if (present_section_nr(nr))
|
2019-07-19 05:56:46 +07:00
|
|
|
section_count++;
|
2010-10-20 00:44:20 +07:00
|
|
|
|
2013-08-21 00:13:03 +07:00
|
|
|
if (section_count == 0)
|
|
|
|
return 0;
|
2020-04-07 10:06:40 +07:00
|
|
|
return init_memory_block(&mem, base_memory_block_id(base_section_nr),
|
|
|
|
MEM_ONLINE);
|
2010-10-20 00:44:20 +07:00
|
|
|
}
|
|
|
|
|
2019-07-19 05:56:56 +07:00
|
|
|
static void unregister_memory(struct memory_block *memory)
|
|
|
|
{
|
|
|
|
if (WARN_ON_ONCE(memory->dev.bus != &memory_subsys))
|
|
|
|
return;
|
|
|
|
|
|
|
|
/* drop the ref. we got via find_memory_block() */
|
|
|
|
put_device(&memory->dev);
|
|
|
|
device_unregister(&memory->dev);
|
|
|
|
}
|
|
|
|
|
2013-04-30 05:08:22 +07:00
|
|
|
/*
|
2019-07-19 05:56:56 +07:00
|
|
|
* Create memory block devices for the given memory area. Start and size
|
|
|
|
* have to be aligned to memory block granularity. Memory block devices
|
|
|
|
* will be initialized as offline.
|
drivers/base/memory.c: drop the mem_sysfs_mutex
The mem_sysfs_mutex isn't really helpful. Also, it's not really clear
what the mutex protects at all.
The device lists of the memory subsystem are protected separately. We
don't need that mutex when looking up. creating, or removing
independent devices. find_memory_block_by_id() will perform locking on
its own and grab a reference of the returned device.
At the time memory_dev_init() is called, we cannot have concurrent
hot(un)plug operations yet - we're still fairly early during boot. We
don't need any locking.
The creation/removal of memory block devices should be protected on a
higher level - especially using the device hotplug lock to avoid
documented issues (see Documentation/core-api/memory-hotplug.rst) - or
if that is reworked, using similar locking.
Protecting in the context of these functions only doesn't really make
sense. Especially, if we would have a situation where the same memory
blocks are created/deleted at the same time, there is something horribly
going wrong (imagining adding/removing a DIMM at the same time from two
call paths) - after the functions succeeded something else in the
callers would blow up (e.g., create_memory_block_devices() succeeded but
there are no memory block devices anymore).
All relevant call paths (except when adding memory early during boot via
ACPI, which is now documented) hold the device hotplug lock when adding
memory, and when removing memory. Let's document that instead.
Add a simple safety net to create_memory_block_devices() in case we
would actually remove memory blocks while adding them, so we'll never
dereference a NULL pointer. Simplify memory_dev_init() now that the
lock is gone.
Link: http://lkml.kernel.org/r/20190925082621.4927-1-david@redhat.com
Signed-off-by: David Hildenbrand <david@redhat.com>
Reviewed-by: Andrew Morton <akpm@linux-foundation.org>
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: "Rafael J. Wysocki" <rafael@kernel.org>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Oscar Salvador <osalvador@suse.de>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2019-12-01 08:54:14 +07:00
|
|
|
*
|
|
|
|
* Called under device_hotplug_lock.
|
2013-04-30 05:08:22 +07:00
|
|
|
*/
|
2019-07-19 05:56:56 +07:00
|
|
|
int create_memory_block_devices(unsigned long start, unsigned long size)
|
2013-04-30 05:08:22 +07:00
|
|
|
{
|
2019-07-19 05:57:40 +07:00
|
|
|
const unsigned long start_block_id = pfn_to_block_id(PFN_DOWN(start));
|
|
|
|
unsigned long end_block_id = pfn_to_block_id(PFN_DOWN(start + size));
|
2013-08-21 00:13:00 +07:00
|
|
|
struct memory_block *mem;
|
2019-07-19 05:56:56 +07:00
|
|
|
unsigned long block_id;
|
|
|
|
int ret = 0;
|
2013-08-21 00:12:57 +07:00
|
|
|
|
2019-07-19 05:56:56 +07:00
|
|
|
if (WARN_ON_ONCE(!IS_ALIGNED(start, memory_block_size_bytes()) ||
|
|
|
|
!IS_ALIGNED(size, memory_block_size_bytes())))
|
|
|
|
return -EINVAL;
|
2013-08-21 00:12:57 +07:00
|
|
|
|
2019-07-19 05:56:56 +07:00
|
|
|
for (block_id = start_block_id; block_id != end_block_id; block_id++) {
|
2019-07-19 05:56:46 +07:00
|
|
|
ret = init_memory_block(&mem, block_id, MEM_OFFLINE);
|
2013-08-21 00:13:00 +07:00
|
|
|
if (ret)
|
2019-07-19 05:56:56 +07:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (ret) {
|
|
|
|
end_block_id = block_id;
|
|
|
|
for (block_id = start_block_id; block_id != end_block_id;
|
|
|
|
block_id++) {
|
2019-07-19 05:57:53 +07:00
|
|
|
mem = find_memory_block_by_id(block_id);
|
drivers/base/memory.c: drop the mem_sysfs_mutex
The mem_sysfs_mutex isn't really helpful. Also, it's not really clear
what the mutex protects at all.
The device lists of the memory subsystem are protected separately. We
don't need that mutex when looking up. creating, or removing
independent devices. find_memory_block_by_id() will perform locking on
its own and grab a reference of the returned device.
At the time memory_dev_init() is called, we cannot have concurrent
hot(un)plug operations yet - we're still fairly early during boot. We
don't need any locking.
The creation/removal of memory block devices should be protected on a
higher level - especially using the device hotplug lock to avoid
documented issues (see Documentation/core-api/memory-hotplug.rst) - or
if that is reworked, using similar locking.
Protecting in the context of these functions only doesn't really make
sense. Especially, if we would have a situation where the same memory
blocks are created/deleted at the same time, there is something horribly
going wrong (imagining adding/removing a DIMM at the same time from two
call paths) - after the functions succeeded something else in the
callers would blow up (e.g., create_memory_block_devices() succeeded but
there are no memory block devices anymore).
All relevant call paths (except when adding memory early during boot via
ACPI, which is now documented) hold the device hotplug lock when adding
memory, and when removing memory. Let's document that instead.
Add a simple safety net to create_memory_block_devices() in case we
would actually remove memory blocks while adding them, so we'll never
dereference a NULL pointer. Simplify memory_dev_init() now that the
lock is gone.
Link: http://lkml.kernel.org/r/20190925082621.4927-1-david@redhat.com
Signed-off-by: David Hildenbrand <david@redhat.com>
Reviewed-by: Andrew Morton <akpm@linux-foundation.org>
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: "Rafael J. Wysocki" <rafael@kernel.org>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Oscar Salvador <osalvador@suse.de>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2019-12-01 08:54:14 +07:00
|
|
|
if (WARN_ON_ONCE(!mem))
|
|
|
|
continue;
|
2019-07-19 05:56:56 +07:00
|
|
|
unregister_memory(mem);
|
|
|
|
}
|
2013-08-21 00:13:00 +07:00
|
|
|
}
|
2013-08-21 00:12:57 +07:00
|
|
|
return ret;
|
2013-04-30 05:08:22 +07:00
|
|
|
}
|
|
|
|
|
2019-07-19 05:57:06 +07:00
|
|
|
/*
|
|
|
|
* Remove memory block devices for the given memory area. Start and size
|
|
|
|
* have to be aligned to memory block granularity. Memory block devices
|
|
|
|
* have to be offline.
|
drivers/base/memory.c: drop the mem_sysfs_mutex
The mem_sysfs_mutex isn't really helpful. Also, it's not really clear
what the mutex protects at all.
The device lists of the memory subsystem are protected separately. We
don't need that mutex when looking up. creating, or removing
independent devices. find_memory_block_by_id() will perform locking on
its own and grab a reference of the returned device.
At the time memory_dev_init() is called, we cannot have concurrent
hot(un)plug operations yet - we're still fairly early during boot. We
don't need any locking.
The creation/removal of memory block devices should be protected on a
higher level - especially using the device hotplug lock to avoid
documented issues (see Documentation/core-api/memory-hotplug.rst) - or
if that is reworked, using similar locking.
Protecting in the context of these functions only doesn't really make
sense. Especially, if we would have a situation where the same memory
blocks are created/deleted at the same time, there is something horribly
going wrong (imagining adding/removing a DIMM at the same time from two
call paths) - after the functions succeeded something else in the
callers would blow up (e.g., create_memory_block_devices() succeeded but
there are no memory block devices anymore).
All relevant call paths (except when adding memory early during boot via
ACPI, which is now documented) hold the device hotplug lock when adding
memory, and when removing memory. Let's document that instead.
Add a simple safety net to create_memory_block_devices() in case we
would actually remove memory blocks while adding them, so we'll never
dereference a NULL pointer. Simplify memory_dev_init() now that the
lock is gone.
Link: http://lkml.kernel.org/r/20190925082621.4927-1-david@redhat.com
Signed-off-by: David Hildenbrand <david@redhat.com>
Reviewed-by: Andrew Morton <akpm@linux-foundation.org>
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: "Rafael J. Wysocki" <rafael@kernel.org>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Oscar Salvador <osalvador@suse.de>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2019-12-01 08:54:14 +07:00
|
|
|
*
|
|
|
|
* Called under device_hotplug_lock.
|
2019-07-19 05:57:06 +07:00
|
|
|
*/
|
|
|
|
void remove_memory_block_devices(unsigned long start, unsigned long size)
|
2005-10-30 08:16:54 +07:00
|
|
|
{
|
2019-07-19 05:57:40 +07:00
|
|
|
const unsigned long start_block_id = pfn_to_block_id(PFN_DOWN(start));
|
|
|
|
const unsigned long end_block_id = pfn_to_block_id(PFN_DOWN(start + size));
|
2005-10-30 08:16:54 +07:00
|
|
|
struct memory_block *mem;
|
2019-07-19 05:57:40 +07:00
|
|
|
unsigned long block_id;
|
2005-10-30 08:16:54 +07:00
|
|
|
|
2019-07-19 05:57:06 +07:00
|
|
|
if (WARN_ON_ONCE(!IS_ALIGNED(start, memory_block_size_bytes()) ||
|
|
|
|
!IS_ALIGNED(size, memory_block_size_bytes())))
|
2019-05-14 07:21:37 +07:00
|
|
|
return;
|
|
|
|
|
2019-07-19 05:57:06 +07:00
|
|
|
for (block_id = start_block_id; block_id != end_block_id; block_id++) {
|
2019-07-19 05:57:53 +07:00
|
|
|
mem = find_memory_block_by_id(block_id);
|
2019-07-19 05:57:06 +07:00
|
|
|
if (WARN_ON_ONCE(!mem))
|
|
|
|
continue;
|
|
|
|
unregister_memory_block_under_nodes(mem);
|
2011-01-20 23:43:34 +07:00
|
|
|
unregister_memory(mem);
|
2019-07-19 05:57:06 +07:00
|
|
|
}
|
2005-10-30 08:16:54 +07:00
|
|
|
}
|
|
|
|
|
2013-02-23 07:32:52 +07:00
|
|
|
/* return true if the memory block is offlined, otherwise, return false */
|
|
|
|
bool is_memblock_offlined(struct memory_block *mem)
|
|
|
|
{
|
|
|
|
return mem->state == MEM_OFFLINE;
|
|
|
|
}
|
|
|
|
|
2013-06-05 02:42:28 +07:00
|
|
|
static struct attribute *memory_root_attrs[] = {
|
|
|
|
#ifdef CONFIG_ARCH_MEMORY_PROBE
|
|
|
|
&dev_attr_probe.attr,
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#ifdef CONFIG_MEMORY_FAILURE
|
|
|
|
&dev_attr_soft_offline_page.attr,
|
|
|
|
&dev_attr_hard_offline_page.attr,
|
|
|
|
#endif
|
|
|
|
|
|
|
|
&dev_attr_block_size_bytes.attr,
|
2016-03-16 04:56:48 +07:00
|
|
|
&dev_attr_auto_online_blocks.attr,
|
2013-06-05 02:42:28 +07:00
|
|
|
NULL
|
|
|
|
};
|
|
|
|
|
|
|
|
static struct attribute_group memory_root_attr_group = {
|
|
|
|
.attrs = memory_root_attrs,
|
|
|
|
};
|
|
|
|
|
|
|
|
static const struct attribute_group *memory_root_attr_groups[] = {
|
|
|
|
&memory_root_attr_group,
|
|
|
|
NULL,
|
|
|
|
};
|
|
|
|
|
2005-10-30 08:16:54 +07:00
|
|
|
/*
|
drivers/base/memory.c: drop the mem_sysfs_mutex
The mem_sysfs_mutex isn't really helpful. Also, it's not really clear
what the mutex protects at all.
The device lists of the memory subsystem are protected separately. We
don't need that mutex when looking up. creating, or removing
independent devices. find_memory_block_by_id() will perform locking on
its own and grab a reference of the returned device.
At the time memory_dev_init() is called, we cannot have concurrent
hot(un)plug operations yet - we're still fairly early during boot. We
don't need any locking.
The creation/removal of memory block devices should be protected on a
higher level - especially using the device hotplug lock to avoid
documented issues (see Documentation/core-api/memory-hotplug.rst) - or
if that is reworked, using similar locking.
Protecting in the context of these functions only doesn't really make
sense. Especially, if we would have a situation where the same memory
blocks are created/deleted at the same time, there is something horribly
going wrong (imagining adding/removing a DIMM at the same time from two
call paths) - after the functions succeeded something else in the
callers would blow up (e.g., create_memory_block_devices() succeeded but
there are no memory block devices anymore).
All relevant call paths (except when adding memory early during boot via
ACPI, which is now documented) hold the device hotplug lock when adding
memory, and when removing memory. Let's document that instead.
Add a simple safety net to create_memory_block_devices() in case we
would actually remove memory blocks while adding them, so we'll never
dereference a NULL pointer. Simplify memory_dev_init() now that the
lock is gone.
Link: http://lkml.kernel.org/r/20190925082621.4927-1-david@redhat.com
Signed-off-by: David Hildenbrand <david@redhat.com>
Reviewed-by: Andrew Morton <akpm@linux-foundation.org>
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: "Rafael J. Wysocki" <rafael@kernel.org>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Oscar Salvador <osalvador@suse.de>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2019-12-01 08:54:14 +07:00
|
|
|
* Initialize the sysfs support for memory devices. At the time this function
|
|
|
|
* is called, we cannot have concurrent creation/deletion of memory block
|
|
|
|
* devices, the device_hotplug_lock is not needed.
|
2005-10-30 08:16:54 +07:00
|
|
|
*/
|
2019-09-24 05:35:46 +07:00
|
|
|
void __init memory_dev_init(void)
|
2005-10-30 08:16:54 +07:00
|
|
|
{
|
|
|
|
int ret;
|
2019-07-19 05:57:37 +07:00
|
|
|
unsigned long block_sz, nr;
|
2005-10-30 08:16:54 +07:00
|
|
|
|
2019-09-24 05:35:46 +07:00
|
|
|
/* Validate the configured memory block size */
|
|
|
|
block_sz = memory_block_size_bytes();
|
|
|
|
if (!is_power_of_2(block_sz) || block_sz < MIN_MEMORY_BLOCK_SIZE)
|
|
|
|
panic("Memory block size not suitable: 0x%lx\n", block_sz);
|
|
|
|
sections_per_block = block_sz / MIN_MEMORY_BLOCK_SIZE;
|
|
|
|
|
2013-06-05 02:42:28 +07:00
|
|
|
ret = subsys_system_register(&memory_subsys, memory_root_attr_groups);
|
2006-12-07 11:37:29 +07:00
|
|
|
if (ret)
|
drivers/base/memory.c: drop the mem_sysfs_mutex
The mem_sysfs_mutex isn't really helpful. Also, it's not really clear
what the mutex protects at all.
The device lists of the memory subsystem are protected separately. We
don't need that mutex when looking up. creating, or removing
independent devices. find_memory_block_by_id() will perform locking on
its own and grab a reference of the returned device.
At the time memory_dev_init() is called, we cannot have concurrent
hot(un)plug operations yet - we're still fairly early during boot. We
don't need any locking.
The creation/removal of memory block devices should be protected on a
higher level - especially using the device hotplug lock to avoid
documented issues (see Documentation/core-api/memory-hotplug.rst) - or
if that is reworked, using similar locking.
Protecting in the context of these functions only doesn't really make
sense. Especially, if we would have a situation where the same memory
blocks are created/deleted at the same time, there is something horribly
going wrong (imagining adding/removing a DIMM at the same time from two
call paths) - after the functions succeeded something else in the
callers would blow up (e.g., create_memory_block_devices() succeeded but
there are no memory block devices anymore).
All relevant call paths (except when adding memory early during boot via
ACPI, which is now documented) hold the device hotplug lock when adding
memory, and when removing memory. Let's document that instead.
Add a simple safety net to create_memory_block_devices() in case we
would actually remove memory blocks while adding them, so we'll never
dereference a NULL pointer. Simplify memory_dev_init() now that the
lock is gone.
Link: http://lkml.kernel.org/r/20190925082621.4927-1-david@redhat.com
Signed-off-by: David Hildenbrand <david@redhat.com>
Reviewed-by: Andrew Morton <akpm@linux-foundation.org>
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: "Rafael J. Wysocki" <rafael@kernel.org>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Oscar Salvador <osalvador@suse.de>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2019-12-01 08:54:14 +07:00
|
|
|
panic("%s() failed to register subsystem: %d\n", __func__, ret);
|
2005-10-30 08:16:54 +07:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Create entries for memory sections that were found
|
|
|
|
* during boot and have been initialized
|
|
|
|
*/
|
2019-07-19 05:57:37 +07:00
|
|
|
for (nr = 0; nr <= __highest_present_section_nr;
|
|
|
|
nr += sections_per_block) {
|
drivers/base/memory.c: drop the mem_sysfs_mutex
The mem_sysfs_mutex isn't really helpful. Also, it's not really clear
what the mutex protects at all.
The device lists of the memory subsystem are protected separately. We
don't need that mutex when looking up. creating, or removing
independent devices. find_memory_block_by_id() will perform locking on
its own and grab a reference of the returned device.
At the time memory_dev_init() is called, we cannot have concurrent
hot(un)plug operations yet - we're still fairly early during boot. We
don't need any locking.
The creation/removal of memory block devices should be protected on a
higher level - especially using the device hotplug lock to avoid
documented issues (see Documentation/core-api/memory-hotplug.rst) - or
if that is reworked, using similar locking.
Protecting in the context of these functions only doesn't really make
sense. Especially, if we would have a situation where the same memory
blocks are created/deleted at the same time, there is something horribly
going wrong (imagining adding/removing a DIMM at the same time from two
call paths) - after the functions succeeded something else in the
callers would blow up (e.g., create_memory_block_devices() succeeded but
there are no memory block devices anymore).
All relevant call paths (except when adding memory early during boot via
ACPI, which is now documented) hold the device hotplug lock when adding
memory, and when removing memory. Let's document that instead.
Add a simple safety net to create_memory_block_devices() in case we
would actually remove memory blocks while adding them, so we'll never
dereference a NULL pointer. Simplify memory_dev_init() now that the
lock is gone.
Link: http://lkml.kernel.org/r/20190925082621.4927-1-david@redhat.com
Signed-off-by: David Hildenbrand <david@redhat.com>
Reviewed-by: Andrew Morton <akpm@linux-foundation.org>
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: "Rafael J. Wysocki" <rafael@kernel.org>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Oscar Salvador <osalvador@suse.de>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2019-12-01 08:54:14 +07:00
|
|
|
ret = add_memory_block(nr);
|
|
|
|
if (ret)
|
|
|
|
panic("%s() failed to add memory block: %d\n", __func__,
|
|
|
|
ret);
|
2005-10-30 08:16:54 +07:00
|
|
|
}
|
|
|
|
}
|
2019-07-19 05:57:50 +07:00
|
|
|
|
|
|
|
/**
|
|
|
|
* walk_memory_blocks - walk through all present memory blocks overlapped
|
|
|
|
* by the range [start, start + size)
|
|
|
|
*
|
|
|
|
* @start: start address of the memory range
|
|
|
|
* @size: size of the memory range
|
|
|
|
* @arg: argument passed to func
|
|
|
|
* @func: callback for each memory section walked
|
|
|
|
*
|
|
|
|
* This function walks through all present memory blocks overlapped by the
|
|
|
|
* range [start, start + size), calling func on each memory block.
|
|
|
|
*
|
|
|
|
* In case func() returns an error, walking is aborted and the error is
|
|
|
|
* returned.
|
|
|
|
*/
|
|
|
|
int walk_memory_blocks(unsigned long start, unsigned long size,
|
|
|
|
void *arg, walk_memory_blocks_func_t func)
|
|
|
|
{
|
|
|
|
const unsigned long start_block_id = phys_to_block_id(start);
|
|
|
|
const unsigned long end_block_id = phys_to_block_id(start + size - 1);
|
|
|
|
struct memory_block *mem;
|
|
|
|
unsigned long block_id;
|
|
|
|
int ret = 0;
|
|
|
|
|
2019-07-19 05:57:53 +07:00
|
|
|
if (!size)
|
|
|
|
return 0;
|
|
|
|
|
2019-07-19 05:57:50 +07:00
|
|
|
for (block_id = start_block_id; block_id <= end_block_id; block_id++) {
|
2019-07-19 05:57:53 +07:00
|
|
|
mem = find_memory_block_by_id(block_id);
|
2019-07-19 05:57:50 +07:00
|
|
|
if (!mem)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
ret = func(mem, arg);
|
|
|
|
put_device(&mem->dev);
|
|
|
|
if (ret)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
return ret;
|
|
|
|
}
|
mm/memory_hotplug: fix try_offline_node()
try_offline_node() is pretty much broken right now:
- The node span is updated when onlining memory, not when adding it. We
ignore memory that was mever onlined. Bad.
- We touch possible garbage memmaps. The pfn_to_nid(pfn) can easily
trigger a kernel panic. Bad for memory that is offline but also bad
for subsection hotadd with ZONE_DEVICE, whereby the memmap of the
first PFN of a section might contain garbage.
- Sections belonging to mixed nodes are not properly considered.
As memory blocks might belong to multiple nodes, we would have to walk
all pageblocks (or at least subsections) within present sections.
However, we don't have a way to identify whether a memmap that is not
online was initialized (relevant for ZONE_DEVICE). This makes things
more complicated.
Luckily, we can piggy pack on the node span and the nid stored in memory
blocks. Currently, the node span is grown when calling
move_pfn_range_to_zone() - e.g., when onlining memory, and shrunk when
removing memory, before calling try_offline_node(). Sysfs links are
created via link_mem_sections(), e.g., during boot or when adding
memory.
If the node still spans memory or if any memory block belongs to the
nid, we don't set the node offline. As memory blocks that span multiple
nodes cannot get offlined, the nid stored in memory blocks is reliable
enough (for such online memory blocks, the node still spans the memory).
Introduce for_each_memory_block() to efficiently walk all memory blocks.
Note: We will soon stop shrinking the ZONE_DEVICE zone and the node span
when removing ZONE_DEVICE memory to fix similar issues (access of
garbage memmaps) - until we have a reliable way to identify whether
these memmaps were properly initialized. This implies later, that once
a node had ZONE_DEVICE memory, we won't be able to set a node offline -
which should be acceptable.
Since commit f1dd2cd13c4b ("mm, memory_hotplug: do not associate
hotadded memory to zones until online") memory that is added is not
assoziated with a zone/node (memmap not initialized). The introducing
commit 60a5a19e7419 ("memory-hotplug: remove sysfs file of node")
already missed that we could have multiple nodes for a section and that
the zone/node span is updated when onlining pages, not when adding them.
I tested this by hotplugging two DIMMs to a memory-less and cpu-less
NUMA node. The node is properly onlined when adding the DIMMs. When
removing the DIMMs, the node is properly offlined.
Masayoshi Mizuma reported:
: Without this patch, memory hotplug fails as panic:
:
: BUG: kernel NULL pointer dereference, address: 0000000000000000
: ...
: Call Trace:
: remove_memory_block_devices+0x81/0xc0
: try_remove_memory+0xb4/0x130
: __remove_memory+0xa/0x20
: acpi_memory_device_remove+0x84/0x100
: acpi_bus_trim+0x57/0x90
: acpi_bus_trim+0x2e/0x90
: acpi_device_hotplug+0x2b2/0x4d0
: acpi_hotplug_work_fn+0x1a/0x30
: process_one_work+0x171/0x380
: worker_thread+0x49/0x3f0
: kthread+0xf8/0x130
: ret_from_fork+0x35/0x40
[david@redhat.com: v3]
Link: http://lkml.kernel.org/r/20191102120221.7553-1-david@redhat.com
Link: http://lkml.kernel.org/r/20191028105458.28320-1-david@redhat.com
Fixes: 60a5a19e7419 ("memory-hotplug: remove sysfs file of node")
Fixes: f1dd2cd13c4b ("mm, memory_hotplug: do not associate hotadded memory to zones until online") # visiable after d0dc12e86b319
Signed-off-by: David Hildenbrand <david@redhat.com>
Tested-by: Masayoshi Mizuma <m.mizuma@jp.fujitsu.com>
Cc: Tang Chen <tangchen@cn.fujitsu.com>
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: "Rafael J. Wysocki" <rafael@kernel.org>
Cc: Keith Busch <keith.busch@intel.com>
Cc: Jiri Olsa <jolsa@kernel.org>
Cc: "Peter Zijlstra (Intel)" <peterz@infradead.org>
Cc: Jani Nikula <jani.nikula@intel.com>
Cc: Nayna Jain <nayna@linux.ibm.com>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Oscar Salvador <osalvador@suse.de>
Cc: Stephen Rothwell <sfr@canb.auug.org.au>
Cc: Dan Williams <dan.j.williams@intel.com>
Cc: Pavel Tatashin <pasha.tatashin@soleen.com>
Cc: <stable@vger.kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2019-11-16 08:34:57 +07:00
|
|
|
|
|
|
|
struct for_each_memory_block_cb_data {
|
|
|
|
walk_memory_blocks_func_t func;
|
|
|
|
void *arg;
|
|
|
|
};
|
|
|
|
|
|
|
|
static int for_each_memory_block_cb(struct device *dev, void *data)
|
|
|
|
{
|
|
|
|
struct memory_block *mem = to_memory_block(dev);
|
|
|
|
struct for_each_memory_block_cb_data *cb_data = data;
|
|
|
|
|
|
|
|
return cb_data->func(mem, cb_data->arg);
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* for_each_memory_block - walk through all present memory blocks
|
|
|
|
*
|
|
|
|
* @arg: argument passed to func
|
|
|
|
* @func: callback for each memory block walked
|
|
|
|
*
|
|
|
|
* This function walks through all present memory blocks, calling func on
|
|
|
|
* each memory block.
|
|
|
|
*
|
|
|
|
* In case func() returns an error, walking is aborted and the error is
|
|
|
|
* returned.
|
|
|
|
*/
|
|
|
|
int for_each_memory_block(void *arg, walk_memory_blocks_func_t func)
|
|
|
|
{
|
|
|
|
struct for_each_memory_block_cb_data cb_data = {
|
|
|
|
.func = func,
|
|
|
|
.arg = arg,
|
|
|
|
};
|
|
|
|
|
|
|
|
return bus_for_each_dev(&memory_subsys, NULL, &cb_data,
|
|
|
|
for_each_memory_block_cb);
|
|
|
|
}
|