mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-11-24 11:50:52 +07:00
f784a3f196
In free_area_init_core(), zone->managed_pages is set to an approximate value for lowmem, and will be adjusted when the bootmem allocator frees pages into the buddy system. But free_area_init_core() is also called by hotadd_new_pgdat() when hot-adding memory. As a result, zone->managed_pages of the newly added node's pgdat is set to an approximate value in the very beginning. Even if the memory on that node has node been onlined, /sys/device/system/node/nodeXXX/meminfo has wrong value: hot-add node2 (memory not onlined) cat /sys/device/system/node/node2/meminfo Node 2 MemTotal: 33554432 kB Node 2 MemFree: 0 kB Node 2 MemUsed: 33554432 kB Node 2 Active: 0 kB This patch fixes this problem by reset node managed pages to 0 after hot-adding a new node. 1. Move reset_managed_pages_done from reset_node_managed_pages() to reset_all_zones_managed_pages() 2. Make reset_node_managed_pages() non-static 3. Call reset_node_managed_pages() in hotadd_new_pgdat() after pgdat is initialized Signed-off-by: Tang Chen <tangchen@cn.fujitsu.com> Signed-off-by: Yasuaki Ishimatsu <isimatu.yasuaki@jp.fujitsu.com> Cc: <stable@vger.kernel.org> [3.16+] Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
369 lines
11 KiB
C
369 lines
11 KiB
C
/*
|
|
* Discontiguous memory support, Kanoj Sarcar, SGI, Nov 1999
|
|
*/
|
|
#ifndef _LINUX_BOOTMEM_H
|
|
#define _LINUX_BOOTMEM_H
|
|
|
|
#include <linux/mmzone.h>
|
|
#include <linux/mm_types.h>
|
|
#include <asm/dma.h>
|
|
|
|
/*
|
|
* simple boot-time physical memory area allocator.
|
|
*/
|
|
|
|
extern unsigned long max_low_pfn;
|
|
extern unsigned long min_low_pfn;
|
|
|
|
/*
|
|
* highest page
|
|
*/
|
|
extern unsigned long max_pfn;
|
|
|
|
#ifndef CONFIG_NO_BOOTMEM
|
|
/*
|
|
* node_bootmem_map is a map pointer - the bits represent all physical
|
|
* memory pages (including holes) on the node.
|
|
*/
|
|
typedef struct bootmem_data {
|
|
unsigned long node_min_pfn;
|
|
unsigned long node_low_pfn;
|
|
void *node_bootmem_map;
|
|
unsigned long last_end_off;
|
|
unsigned long hint_idx;
|
|
struct list_head list;
|
|
} bootmem_data_t;
|
|
|
|
extern bootmem_data_t bootmem_node_data[];
|
|
#endif
|
|
|
|
extern unsigned long bootmem_bootmap_pages(unsigned long);
|
|
|
|
extern unsigned long init_bootmem_node(pg_data_t *pgdat,
|
|
unsigned long freepfn,
|
|
unsigned long startpfn,
|
|
unsigned long endpfn);
|
|
extern unsigned long init_bootmem(unsigned long addr, unsigned long memend);
|
|
|
|
extern unsigned long free_all_bootmem(void);
|
|
extern void reset_node_managed_pages(pg_data_t *pgdat);
|
|
extern void reset_all_zones_managed_pages(void);
|
|
|
|
extern void free_bootmem_node(pg_data_t *pgdat,
|
|
unsigned long addr,
|
|
unsigned long size);
|
|
extern void free_bootmem(unsigned long physaddr, unsigned long size);
|
|
extern void free_bootmem_late(unsigned long physaddr, unsigned long size);
|
|
|
|
/*
|
|
* Flags for reserve_bootmem (also if CONFIG_HAVE_ARCH_BOOTMEM_NODE,
|
|
* the architecture-specific code should honor this).
|
|
*
|
|
* If flags is BOOTMEM_DEFAULT, then the return value is always 0 (success).
|
|
* If flags contains BOOTMEM_EXCLUSIVE, then -EBUSY is returned if the memory
|
|
* already was reserved.
|
|
*/
|
|
#define BOOTMEM_DEFAULT 0
|
|
#define BOOTMEM_EXCLUSIVE (1<<0)
|
|
|
|
extern int reserve_bootmem(unsigned long addr,
|
|
unsigned long size,
|
|
int flags);
|
|
extern int reserve_bootmem_node(pg_data_t *pgdat,
|
|
unsigned long physaddr,
|
|
unsigned long size,
|
|
int flags);
|
|
|
|
extern void *__alloc_bootmem(unsigned long size,
|
|
unsigned long align,
|
|
unsigned long goal);
|
|
extern void *__alloc_bootmem_nopanic(unsigned long size,
|
|
unsigned long align,
|
|
unsigned long goal);
|
|
extern void *__alloc_bootmem_node(pg_data_t *pgdat,
|
|
unsigned long size,
|
|
unsigned long align,
|
|
unsigned long goal);
|
|
void *__alloc_bootmem_node_high(pg_data_t *pgdat,
|
|
unsigned long size,
|
|
unsigned long align,
|
|
unsigned long goal);
|
|
extern void *__alloc_bootmem_node_nopanic(pg_data_t *pgdat,
|
|
unsigned long size,
|
|
unsigned long align,
|
|
unsigned long goal);
|
|
void *___alloc_bootmem_node_nopanic(pg_data_t *pgdat,
|
|
unsigned long size,
|
|
unsigned long align,
|
|
unsigned long goal,
|
|
unsigned long limit);
|
|
extern void *__alloc_bootmem_low(unsigned long size,
|
|
unsigned long align,
|
|
unsigned long goal);
|
|
void *__alloc_bootmem_low_nopanic(unsigned long size,
|
|
unsigned long align,
|
|
unsigned long goal);
|
|
extern void *__alloc_bootmem_low_node(pg_data_t *pgdat,
|
|
unsigned long size,
|
|
unsigned long align,
|
|
unsigned long goal);
|
|
|
|
#ifdef CONFIG_NO_BOOTMEM
|
|
/* We are using top down, so it is safe to use 0 here */
|
|
#define BOOTMEM_LOW_LIMIT 0
|
|
#else
|
|
#define BOOTMEM_LOW_LIMIT __pa(MAX_DMA_ADDRESS)
|
|
#endif
|
|
|
|
#define alloc_bootmem(x) \
|
|
__alloc_bootmem(x, SMP_CACHE_BYTES, BOOTMEM_LOW_LIMIT)
|
|
#define alloc_bootmem_align(x, align) \
|
|
__alloc_bootmem(x, align, BOOTMEM_LOW_LIMIT)
|
|
#define alloc_bootmem_nopanic(x) \
|
|
__alloc_bootmem_nopanic(x, SMP_CACHE_BYTES, BOOTMEM_LOW_LIMIT)
|
|
#define alloc_bootmem_pages(x) \
|
|
__alloc_bootmem(x, PAGE_SIZE, BOOTMEM_LOW_LIMIT)
|
|
#define alloc_bootmem_pages_nopanic(x) \
|
|
__alloc_bootmem_nopanic(x, PAGE_SIZE, BOOTMEM_LOW_LIMIT)
|
|
#define alloc_bootmem_node(pgdat, x) \
|
|
__alloc_bootmem_node(pgdat, x, SMP_CACHE_BYTES, BOOTMEM_LOW_LIMIT)
|
|
#define alloc_bootmem_node_nopanic(pgdat, x) \
|
|
__alloc_bootmem_node_nopanic(pgdat, x, SMP_CACHE_BYTES, BOOTMEM_LOW_LIMIT)
|
|
#define alloc_bootmem_pages_node(pgdat, x) \
|
|
__alloc_bootmem_node(pgdat, x, PAGE_SIZE, BOOTMEM_LOW_LIMIT)
|
|
#define alloc_bootmem_pages_node_nopanic(pgdat, x) \
|
|
__alloc_bootmem_node_nopanic(pgdat, x, PAGE_SIZE, BOOTMEM_LOW_LIMIT)
|
|
|
|
#define alloc_bootmem_low(x) \
|
|
__alloc_bootmem_low(x, SMP_CACHE_BYTES, 0)
|
|
#define alloc_bootmem_low_pages_nopanic(x) \
|
|
__alloc_bootmem_low_nopanic(x, PAGE_SIZE, 0)
|
|
#define alloc_bootmem_low_pages(x) \
|
|
__alloc_bootmem_low(x, PAGE_SIZE, 0)
|
|
#define alloc_bootmem_low_pages_node(pgdat, x) \
|
|
__alloc_bootmem_low_node(pgdat, x, PAGE_SIZE, 0)
|
|
|
|
|
|
#if defined(CONFIG_HAVE_MEMBLOCK) && defined(CONFIG_NO_BOOTMEM)
|
|
|
|
/* FIXME: use MEMBLOCK_ALLOC_* variants here */
|
|
#define BOOTMEM_ALLOC_ACCESSIBLE 0
|
|
#define BOOTMEM_ALLOC_ANYWHERE (~(phys_addr_t)0)
|
|
|
|
/* FIXME: Move to memblock.h at a point where we remove nobootmem.c */
|
|
void *memblock_virt_alloc_try_nid_nopanic(phys_addr_t size,
|
|
phys_addr_t align, phys_addr_t min_addr,
|
|
phys_addr_t max_addr, int nid);
|
|
void *memblock_virt_alloc_try_nid(phys_addr_t size, phys_addr_t align,
|
|
phys_addr_t min_addr, phys_addr_t max_addr, int nid);
|
|
void __memblock_free_early(phys_addr_t base, phys_addr_t size);
|
|
void __memblock_free_late(phys_addr_t base, phys_addr_t size);
|
|
|
|
static inline void * __init memblock_virt_alloc(
|
|
phys_addr_t size, phys_addr_t align)
|
|
{
|
|
return memblock_virt_alloc_try_nid(size, align, BOOTMEM_LOW_LIMIT,
|
|
BOOTMEM_ALLOC_ACCESSIBLE,
|
|
NUMA_NO_NODE);
|
|
}
|
|
|
|
static inline void * __init memblock_virt_alloc_nopanic(
|
|
phys_addr_t size, phys_addr_t align)
|
|
{
|
|
return memblock_virt_alloc_try_nid_nopanic(size, align,
|
|
BOOTMEM_LOW_LIMIT,
|
|
BOOTMEM_ALLOC_ACCESSIBLE,
|
|
NUMA_NO_NODE);
|
|
}
|
|
|
|
#ifndef ARCH_LOW_ADDRESS_LIMIT
|
|
#define ARCH_LOW_ADDRESS_LIMIT 0xffffffffUL
|
|
#endif
|
|
|
|
static inline void * __init memblock_virt_alloc_low(
|
|
phys_addr_t size, phys_addr_t align)
|
|
{
|
|
return memblock_virt_alloc_try_nid(size, align,
|
|
BOOTMEM_LOW_LIMIT,
|
|
ARCH_LOW_ADDRESS_LIMIT,
|
|
NUMA_NO_NODE);
|
|
}
|
|
static inline void * __init memblock_virt_alloc_low_nopanic(
|
|
phys_addr_t size, phys_addr_t align)
|
|
{
|
|
return memblock_virt_alloc_try_nid_nopanic(size, align,
|
|
BOOTMEM_LOW_LIMIT,
|
|
ARCH_LOW_ADDRESS_LIMIT,
|
|
NUMA_NO_NODE);
|
|
}
|
|
|
|
static inline void * __init memblock_virt_alloc_from_nopanic(
|
|
phys_addr_t size, phys_addr_t align, phys_addr_t min_addr)
|
|
{
|
|
return memblock_virt_alloc_try_nid_nopanic(size, align, min_addr,
|
|
BOOTMEM_ALLOC_ACCESSIBLE,
|
|
NUMA_NO_NODE);
|
|
}
|
|
|
|
static inline void * __init memblock_virt_alloc_node(
|
|
phys_addr_t size, int nid)
|
|
{
|
|
return memblock_virt_alloc_try_nid(size, 0, BOOTMEM_LOW_LIMIT,
|
|
BOOTMEM_ALLOC_ACCESSIBLE, nid);
|
|
}
|
|
|
|
static inline void * __init memblock_virt_alloc_node_nopanic(
|
|
phys_addr_t size, int nid)
|
|
{
|
|
return memblock_virt_alloc_try_nid_nopanic(size, 0, BOOTMEM_LOW_LIMIT,
|
|
BOOTMEM_ALLOC_ACCESSIBLE,
|
|
nid);
|
|
}
|
|
|
|
static inline void __init memblock_free_early(
|
|
phys_addr_t base, phys_addr_t size)
|
|
{
|
|
__memblock_free_early(base, size);
|
|
}
|
|
|
|
static inline void __init memblock_free_early_nid(
|
|
phys_addr_t base, phys_addr_t size, int nid)
|
|
{
|
|
__memblock_free_early(base, size);
|
|
}
|
|
|
|
static inline void __init memblock_free_late(
|
|
phys_addr_t base, phys_addr_t size)
|
|
{
|
|
__memblock_free_late(base, size);
|
|
}
|
|
|
|
#else
|
|
|
|
#define BOOTMEM_ALLOC_ACCESSIBLE 0
|
|
|
|
|
|
/* Fall back to all the existing bootmem APIs */
|
|
static inline void * __init memblock_virt_alloc(
|
|
phys_addr_t size, phys_addr_t align)
|
|
{
|
|
if (!align)
|
|
align = SMP_CACHE_BYTES;
|
|
return __alloc_bootmem(size, align, BOOTMEM_LOW_LIMIT);
|
|
}
|
|
|
|
static inline void * __init memblock_virt_alloc_nopanic(
|
|
phys_addr_t size, phys_addr_t align)
|
|
{
|
|
if (!align)
|
|
align = SMP_CACHE_BYTES;
|
|
return __alloc_bootmem_nopanic(size, align, BOOTMEM_LOW_LIMIT);
|
|
}
|
|
|
|
static inline void * __init memblock_virt_alloc_low(
|
|
phys_addr_t size, phys_addr_t align)
|
|
{
|
|
if (!align)
|
|
align = SMP_CACHE_BYTES;
|
|
return __alloc_bootmem_low(size, align, 0);
|
|
}
|
|
|
|
static inline void * __init memblock_virt_alloc_low_nopanic(
|
|
phys_addr_t size, phys_addr_t align)
|
|
{
|
|
if (!align)
|
|
align = SMP_CACHE_BYTES;
|
|
return __alloc_bootmem_low_nopanic(size, align, 0);
|
|
}
|
|
|
|
static inline void * __init memblock_virt_alloc_from_nopanic(
|
|
phys_addr_t size, phys_addr_t align, phys_addr_t min_addr)
|
|
{
|
|
return __alloc_bootmem_nopanic(size, align, min_addr);
|
|
}
|
|
|
|
static inline void * __init memblock_virt_alloc_node(
|
|
phys_addr_t size, int nid)
|
|
{
|
|
return __alloc_bootmem_node(NODE_DATA(nid), size, SMP_CACHE_BYTES,
|
|
BOOTMEM_LOW_LIMIT);
|
|
}
|
|
|
|
static inline void * __init memblock_virt_alloc_node_nopanic(
|
|
phys_addr_t size, int nid)
|
|
{
|
|
return __alloc_bootmem_node_nopanic(NODE_DATA(nid), size,
|
|
SMP_CACHE_BYTES,
|
|
BOOTMEM_LOW_LIMIT);
|
|
}
|
|
|
|
static inline void * __init memblock_virt_alloc_try_nid(phys_addr_t size,
|
|
phys_addr_t align, phys_addr_t min_addr, phys_addr_t max_addr, int nid)
|
|
{
|
|
return __alloc_bootmem_node_high(NODE_DATA(nid), size, align,
|
|
min_addr);
|
|
}
|
|
|
|
static inline void * __init memblock_virt_alloc_try_nid_nopanic(
|
|
phys_addr_t size, phys_addr_t align,
|
|
phys_addr_t min_addr, phys_addr_t max_addr, int nid)
|
|
{
|
|
return ___alloc_bootmem_node_nopanic(NODE_DATA(nid), size, align,
|
|
min_addr, max_addr);
|
|
}
|
|
|
|
static inline void __init memblock_free_early(
|
|
phys_addr_t base, phys_addr_t size)
|
|
{
|
|
free_bootmem(base, size);
|
|
}
|
|
|
|
static inline void __init memblock_free_early_nid(
|
|
phys_addr_t base, phys_addr_t size, int nid)
|
|
{
|
|
free_bootmem_node(NODE_DATA(nid), base, size);
|
|
}
|
|
|
|
static inline void __init memblock_free_late(
|
|
phys_addr_t base, phys_addr_t size)
|
|
{
|
|
free_bootmem_late(base, size);
|
|
}
|
|
#endif /* defined(CONFIG_HAVE_MEMBLOCK) && defined(CONFIG_NO_BOOTMEM) */
|
|
|
|
#ifdef CONFIG_HAVE_ARCH_ALLOC_REMAP
|
|
extern void *alloc_remap(int nid, unsigned long size);
|
|
#else
|
|
static inline void *alloc_remap(int nid, unsigned long size)
|
|
{
|
|
return NULL;
|
|
}
|
|
#endif /* CONFIG_HAVE_ARCH_ALLOC_REMAP */
|
|
|
|
extern void *alloc_large_system_hash(const char *tablename,
|
|
unsigned long bucketsize,
|
|
unsigned long numentries,
|
|
int scale,
|
|
int flags,
|
|
unsigned int *_hash_shift,
|
|
unsigned int *_hash_mask,
|
|
unsigned long low_limit,
|
|
unsigned long high_limit);
|
|
|
|
#define HASH_EARLY 0x00000001 /* Allocating during early boot? */
|
|
#define HASH_SMALL 0x00000002 /* sub-page allocation allowed, min
|
|
* shift passed via *_hash_shift */
|
|
|
|
/* Only NUMA needs hash distribution. 64bit NUMA architectures have
|
|
* sufficient vmalloc space.
|
|
*/
|
|
#if defined(CONFIG_NUMA) && defined(CONFIG_64BIT)
|
|
#define HASHDIST_DEFAULT 1
|
|
#else
|
|
#define HASHDIST_DEFAULT 0
|
|
#endif
|
|
extern int hashdist; /* Distribute hashes across NUMA nodes? */
|
|
|
|
|
|
#endif /* _LINUX_BOOTMEM_H */
|