mirror of
https://github.com/adulau/aha.git
synced 2024-12-28 11:46:19 +00:00
mm, x86: remove MEMORY_HOTPLUG_RESERVE related code
after:
| commit b263295dbf
| Author: Christoph Lameter <clameter@sgi.com>
| Date: Wed Jan 30 13:30:47 2008 +0100
|
| x86: 64-bit, make sparsemem vmemmap the only memory model
we don't have MEMORY_HOTPLUG_RESERVE anymore.
Historically, x86-64 had an architecture-specific method for memory hotplug
whereby it scanned the SRAT for physical memory ranges that could be
potentially used for memory hot-add later. By reserving those ranges
without physical memory, the memmap would be allocated and left dormant
until needed. This depended on the DISCONTIG memory model which has been
removed so the code implementing HOTPLUG_RESERVE is now dead.
This patch removes the dead code used by MEMORY_HOTPLUG_RESERVE.
(Changelog authored by Mel.)
v2: updated changelog, and remove hotadd= in doc
[ Impact: remove dead code ]
Signed-off-by: Yinghai Lu <yinghai@kernel.org>
Reviewed-by: Christoph Lameter <cl@linux-foundation.org>
Reviewed-by: Mel Gorman <mel@csn.ul.ie>
Workflow-found-OK-by: Andrew Morton <akpm@linux-foundation.org>
LKML-Reference: <4A0C4910.7090508@kernel.org>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
This commit is contained in:
parent
b286e21868
commit
888a589f6b
6 changed files with 12 additions and 135 deletions
|
@ -150,11 +150,6 @@ NUMA
|
||||||
Otherwise, the remaining system RAM is allocated to an
|
Otherwise, the remaining system RAM is allocated to an
|
||||||
additional node.
|
additional node.
|
||||||
|
|
||||||
numa=hotadd=percent
|
|
||||||
Only allow hotadd memory to preallocate page structures upto
|
|
||||||
percent of already available memory.
|
|
||||||
numa=hotadd=0 will disable hotadd memory.
|
|
||||||
|
|
||||||
ACPI
|
ACPI
|
||||||
|
|
||||||
acpi=off Don't enable ACPI
|
acpi=off Don't enable ACPI
|
||||||
|
|
|
@ -17,9 +17,6 @@ extern int compute_hash_shift(struct bootnode *nodes, int numblks,
|
||||||
extern void numa_init_array(void);
|
extern void numa_init_array(void);
|
||||||
extern int numa_off;
|
extern int numa_off;
|
||||||
|
|
||||||
extern void srat_reserve_add_area(int nodeid);
|
|
||||||
extern int hotadd_percent;
|
|
||||||
|
|
||||||
extern s16 apicid_to_node[MAX_LOCAL_APIC];
|
extern s16 apicid_to_node[MAX_LOCAL_APIC];
|
||||||
|
|
||||||
extern unsigned long numa_free_all_bootmem(void);
|
extern unsigned long numa_free_all_bootmem(void);
|
||||||
|
|
|
@ -272,9 +272,6 @@ void __init setup_node_bootmem(int nodeid, unsigned long start,
|
||||||
reserve_bootmem_node(NODE_DATA(nodeid), bootmap_start,
|
reserve_bootmem_node(NODE_DATA(nodeid), bootmap_start,
|
||||||
bootmap_pages<<PAGE_SHIFT, BOOTMEM_DEFAULT);
|
bootmap_pages<<PAGE_SHIFT, BOOTMEM_DEFAULT);
|
||||||
|
|
||||||
#ifdef CONFIG_ACPI_NUMA
|
|
||||||
srat_reserve_add_area(nodeid);
|
|
||||||
#endif
|
|
||||||
node_set_online(nodeid);
|
node_set_online(nodeid);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -591,8 +588,6 @@ static __init int numa_setup(char *opt)
|
||||||
#ifdef CONFIG_ACPI_NUMA
|
#ifdef CONFIG_ACPI_NUMA
|
||||||
if (!strncmp(opt, "noacpi", 6))
|
if (!strncmp(opt, "noacpi", 6))
|
||||||
acpi_numa = -1;
|
acpi_numa = -1;
|
||||||
if (!strncmp(opt, "hotadd=", 7))
|
|
||||||
hotadd_percent = simple_strtoul(opt+7, NULL, 10);
|
|
||||||
#endif
|
#endif
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
|
@ -31,8 +31,6 @@ static nodemask_t nodes_parsed __initdata;
|
||||||
static nodemask_t cpu_nodes_parsed __initdata;
|
static nodemask_t cpu_nodes_parsed __initdata;
|
||||||
static struct bootnode nodes[MAX_NUMNODES] __initdata;
|
static struct bootnode nodes[MAX_NUMNODES] __initdata;
|
||||||
static struct bootnode nodes_add[MAX_NUMNODES];
|
static struct bootnode nodes_add[MAX_NUMNODES];
|
||||||
static int found_add_area __initdata;
|
|
||||||
int hotadd_percent __initdata = 0;
|
|
||||||
|
|
||||||
static int num_node_memblks __initdata;
|
static int num_node_memblks __initdata;
|
||||||
static struct bootnode node_memblk_range[NR_NODE_MEMBLKS] __initdata;
|
static struct bootnode node_memblk_range[NR_NODE_MEMBLKS] __initdata;
|
||||||
|
@ -66,9 +64,6 @@ static __init void cutoff_node(int i, unsigned long start, unsigned long end)
|
||||||
{
|
{
|
||||||
struct bootnode *nd = &nodes[i];
|
struct bootnode *nd = &nodes[i];
|
||||||
|
|
||||||
if (found_add_area)
|
|
||||||
return;
|
|
||||||
|
|
||||||
if (nd->start < start) {
|
if (nd->start < start) {
|
||||||
nd->start = start;
|
nd->start = start;
|
||||||
if (nd->end < nd->start)
|
if (nd->end < nd->start)
|
||||||
|
@ -86,7 +81,6 @@ static __init void bad_srat(void)
|
||||||
int i;
|
int i;
|
||||||
printk(KERN_ERR "SRAT: SRAT not used.\n");
|
printk(KERN_ERR "SRAT: SRAT not used.\n");
|
||||||
acpi_numa = -1;
|
acpi_numa = -1;
|
||||||
found_add_area = 0;
|
|
||||||
for (i = 0; i < MAX_LOCAL_APIC; i++)
|
for (i = 0; i < MAX_LOCAL_APIC; i++)
|
||||||
apicid_to_node[i] = NUMA_NO_NODE;
|
apicid_to_node[i] = NUMA_NO_NODE;
|
||||||
for (i = 0; i < MAX_NUMNODES; i++)
|
for (i = 0; i < MAX_NUMNODES; i++)
|
||||||
|
@ -182,24 +176,21 @@ acpi_numa_processor_affinity_init(struct acpi_srat_cpu_affinity *pa)
|
||||||
pxm, apic_id, node);
|
pxm, apic_id, node);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int update_end_of_memory(unsigned long end) {return -1;}
|
|
||||||
static int hotadd_enough_memory(struct bootnode *nd) {return 1;}
|
|
||||||
#ifdef CONFIG_MEMORY_HOTPLUG_SPARSE
|
#ifdef CONFIG_MEMORY_HOTPLUG_SPARSE
|
||||||
static inline int save_add_info(void) {return 1;}
|
static inline int save_add_info(void) {return 1;}
|
||||||
#else
|
#else
|
||||||
static inline int save_add_info(void) {return 0;}
|
static inline int save_add_info(void) {return 0;}
|
||||||
#endif
|
#endif
|
||||||
/*
|
/*
|
||||||
* Update nodes_add and decide if to include add are in the zone.
|
* Update nodes_add[]
|
||||||
* Both SPARSE and RESERVE need nodes_add information.
|
* This code supports one contiguous hot add area per node
|
||||||
* This code supports one contiguous hot add area per node.
|
|
||||||
*/
|
*/
|
||||||
static int __init
|
static void __init
|
||||||
reserve_hotadd(int node, unsigned long start, unsigned long end)
|
update_nodes_add(int node, unsigned long start, unsigned long end)
|
||||||
{
|
{
|
||||||
unsigned long s_pfn = start >> PAGE_SHIFT;
|
unsigned long s_pfn = start >> PAGE_SHIFT;
|
||||||
unsigned long e_pfn = end >> PAGE_SHIFT;
|
unsigned long e_pfn = end >> PAGE_SHIFT;
|
||||||
int ret = 0, changed = 0;
|
int changed = 0;
|
||||||
struct bootnode *nd = &nodes_add[node];
|
struct bootnode *nd = &nodes_add[node];
|
||||||
|
|
||||||
/* I had some trouble with strange memory hotadd regions breaking
|
/* I had some trouble with strange memory hotadd regions breaking
|
||||||
|
@ -210,7 +201,7 @@ reserve_hotadd(int node, unsigned long start, unsigned long end)
|
||||||
mistakes */
|
mistakes */
|
||||||
if ((signed long)(end - start) < NODE_MIN_SIZE) {
|
if ((signed long)(end - start) < NODE_MIN_SIZE) {
|
||||||
printk(KERN_ERR "SRAT: Hotplug area too small\n");
|
printk(KERN_ERR "SRAT: Hotplug area too small\n");
|
||||||
return -1;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* This check might be a bit too strict, but I'm keeping it for now. */
|
/* This check might be a bit too strict, but I'm keeping it for now. */
|
||||||
|
@ -218,12 +209,7 @@ reserve_hotadd(int node, unsigned long start, unsigned long end)
|
||||||
printk(KERN_ERR
|
printk(KERN_ERR
|
||||||
"SRAT: Hotplug area %lu -> %lu has existing memory\n",
|
"SRAT: Hotplug area %lu -> %lu has existing memory\n",
|
||||||
s_pfn, e_pfn);
|
s_pfn, e_pfn);
|
||||||
return -1;
|
return;
|
||||||
}
|
|
||||||
|
|
||||||
if (!hotadd_enough_memory(&nodes_add[node])) {
|
|
||||||
printk(KERN_ERR "SRAT: Hotplug area too large\n");
|
|
||||||
return -1;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Looks good */
|
/* Looks good */
|
||||||
|
@ -245,11 +231,9 @@ reserve_hotadd(int node, unsigned long start, unsigned long end)
|
||||||
printk(KERN_ERR "SRAT: Hotplug zone not continuous. Partly ignored\n");
|
printk(KERN_ERR "SRAT: Hotplug zone not continuous. Partly ignored\n");
|
||||||
}
|
}
|
||||||
|
|
||||||
ret = update_end_of_memory(nd->end);
|
|
||||||
|
|
||||||
if (changed)
|
if (changed)
|
||||||
printk(KERN_INFO "SRAT: hot plug zone found %Lx - %Lx\n", nd->start, nd->end);
|
printk(KERN_INFO "SRAT: hot plug zone found %Lx - %Lx\n",
|
||||||
return ret;
|
nd->start, nd->end);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Callback for parsing of the Proximity Domain <-> Memory Area mappings */
|
/* Callback for parsing of the Proximity Domain <-> Memory Area mappings */
|
||||||
|
@ -310,13 +294,10 @@ acpi_numa_memory_affinity_init(struct acpi_srat_mem_affinity *ma)
|
||||||
start, end);
|
start, end);
|
||||||
e820_register_active_regions(node, start >> PAGE_SHIFT,
|
e820_register_active_regions(node, start >> PAGE_SHIFT,
|
||||||
end >> PAGE_SHIFT);
|
end >> PAGE_SHIFT);
|
||||||
push_node_boundaries(node, nd->start >> PAGE_SHIFT,
|
|
||||||
nd->end >> PAGE_SHIFT);
|
|
||||||
|
|
||||||
if ((ma->flags & ACPI_SRAT_MEM_HOT_PLUGGABLE) &&
|
if (ma->flags & ACPI_SRAT_MEM_HOT_PLUGGABLE) {
|
||||||
(reserve_hotadd(node, start, end) < 0)) {
|
update_nodes_add(node, start, end);
|
||||||
/* Ignore hotadd region. Undo damage */
|
/* restore nodes[node] */
|
||||||
printk(KERN_NOTICE "SRAT: Hotplug region ignored\n");
|
|
||||||
*nd = oldnode;
|
*nd = oldnode;
|
||||||
if ((nd->start | nd->end) == 0)
|
if ((nd->start | nd->end) == 0)
|
||||||
node_clear(node, nodes_parsed);
|
node_clear(node, nodes_parsed);
|
||||||
|
@ -510,26 +491,6 @@ static int null_slit_node_compare(int a, int b)
|
||||||
}
|
}
|
||||||
#endif /* CONFIG_NUMA_EMU */
|
#endif /* CONFIG_NUMA_EMU */
|
||||||
|
|
||||||
void __init srat_reserve_add_area(int nodeid)
|
|
||||||
{
|
|
||||||
if (found_add_area && nodes_add[nodeid].end) {
|
|
||||||
u64 total_mb;
|
|
||||||
|
|
||||||
printk(KERN_INFO "SRAT: Reserving hot-add memory space "
|
|
||||||
"for node %d at %Lx-%Lx\n",
|
|
||||||
nodeid, nodes_add[nodeid].start, nodes_add[nodeid].end);
|
|
||||||
total_mb = (nodes_add[nodeid].end - nodes_add[nodeid].start)
|
|
||||||
>> PAGE_SHIFT;
|
|
||||||
total_mb *= sizeof(struct page);
|
|
||||||
total_mb >>= 20;
|
|
||||||
printk(KERN_INFO "SRAT: This will cost you %Lu MB of "
|
|
||||||
"pre-allocated memory.\n", (unsigned long long)total_mb);
|
|
||||||
reserve_bootmem_node(NODE_DATA(nodeid), nodes_add[nodeid].start,
|
|
||||||
nodes_add[nodeid].end - nodes_add[nodeid].start,
|
|
||||||
BOOTMEM_DEFAULT);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
int __node_distance(int a, int b)
|
int __node_distance(int a, int b)
|
||||||
{
|
{
|
||||||
int index;
|
int index;
|
||||||
|
|
|
@ -1031,8 +1031,6 @@ extern void add_active_range(unsigned int nid, unsigned long start_pfn,
|
||||||
unsigned long end_pfn);
|
unsigned long end_pfn);
|
||||||
extern void remove_active_range(unsigned int nid, unsigned long start_pfn,
|
extern void remove_active_range(unsigned int nid, unsigned long start_pfn,
|
||||||
unsigned long end_pfn);
|
unsigned long end_pfn);
|
||||||
extern void push_node_boundaries(unsigned int nid, unsigned long start_pfn,
|
|
||||||
unsigned long end_pfn);
|
|
||||||
extern void remove_all_active_ranges(void);
|
extern void remove_all_active_ranges(void);
|
||||||
extern unsigned long absent_pages_in_range(unsigned long start_pfn,
|
extern unsigned long absent_pages_in_range(unsigned long start_pfn,
|
||||||
unsigned long end_pfn);
|
unsigned long end_pfn);
|
||||||
|
|
|
@ -149,10 +149,6 @@ static unsigned long __meminitdata dma_reserve;
|
||||||
static int __meminitdata nr_nodemap_entries;
|
static int __meminitdata nr_nodemap_entries;
|
||||||
static unsigned long __meminitdata arch_zone_lowest_possible_pfn[MAX_NR_ZONES];
|
static unsigned long __meminitdata arch_zone_lowest_possible_pfn[MAX_NR_ZONES];
|
||||||
static unsigned long __meminitdata arch_zone_highest_possible_pfn[MAX_NR_ZONES];
|
static unsigned long __meminitdata arch_zone_highest_possible_pfn[MAX_NR_ZONES];
|
||||||
#ifdef CONFIG_MEMORY_HOTPLUG_RESERVE
|
|
||||||
static unsigned long __meminitdata node_boundary_start_pfn[MAX_NUMNODES];
|
|
||||||
static unsigned long __meminitdata node_boundary_end_pfn[MAX_NUMNODES];
|
|
||||||
#endif /* CONFIG_MEMORY_HOTPLUG_RESERVE */
|
|
||||||
static unsigned long __initdata required_kernelcore;
|
static unsigned long __initdata required_kernelcore;
|
||||||
static unsigned long __initdata required_movablecore;
|
static unsigned long __initdata required_movablecore;
|
||||||
static unsigned long __meminitdata zone_movable_pfn[MAX_NUMNODES];
|
static unsigned long __meminitdata zone_movable_pfn[MAX_NUMNODES];
|
||||||
|
@ -3102,64 +3098,6 @@ void __init sparse_memory_present_with_active_regions(int nid)
|
||||||
early_node_map[i].end_pfn);
|
early_node_map[i].end_pfn);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* push_node_boundaries - Push node boundaries to at least the requested boundary
|
|
||||||
* @nid: The nid of the node to push the boundary for
|
|
||||||
* @start_pfn: The start pfn of the node
|
|
||||||
* @end_pfn: The end pfn of the node
|
|
||||||
*
|
|
||||||
* In reserve-based hot-add, mem_map is allocated that is unused until hotadd
|
|
||||||
* time. Specifically, on x86_64, SRAT will report ranges that can potentially
|
|
||||||
* be hotplugged even though no physical memory exists. This function allows
|
|
||||||
* an arch to push out the node boundaries so mem_map is allocated that can
|
|
||||||
* be used later.
|
|
||||||
*/
|
|
||||||
#ifdef CONFIG_MEMORY_HOTPLUG_RESERVE
|
|
||||||
void __init push_node_boundaries(unsigned int nid,
|
|
||||||
unsigned long start_pfn, unsigned long end_pfn)
|
|
||||||
{
|
|
||||||
mminit_dprintk(MMINIT_TRACE, "zoneboundary",
|
|
||||||
"Entering push_node_boundaries(%u, %lu, %lu)\n",
|
|
||||||
nid, start_pfn, end_pfn);
|
|
||||||
|
|
||||||
/* Initialise the boundary for this node if necessary */
|
|
||||||
if (node_boundary_end_pfn[nid] == 0)
|
|
||||||
node_boundary_start_pfn[nid] = -1UL;
|
|
||||||
|
|
||||||
/* Update the boundaries */
|
|
||||||
if (node_boundary_start_pfn[nid] > start_pfn)
|
|
||||||
node_boundary_start_pfn[nid] = start_pfn;
|
|
||||||
if (node_boundary_end_pfn[nid] < end_pfn)
|
|
||||||
node_boundary_end_pfn[nid] = end_pfn;
|
|
||||||
}
|
|
||||||
|
|
||||||
/* If necessary, push the node boundary out for reserve hotadd */
|
|
||||||
static void __meminit account_node_boundary(unsigned int nid,
|
|
||||||
unsigned long *start_pfn, unsigned long *end_pfn)
|
|
||||||
{
|
|
||||||
mminit_dprintk(MMINIT_TRACE, "zoneboundary",
|
|
||||||
"Entering account_node_boundary(%u, %lu, %lu)\n",
|
|
||||||
nid, *start_pfn, *end_pfn);
|
|
||||||
|
|
||||||
/* Return if boundary information has not been provided */
|
|
||||||
if (node_boundary_end_pfn[nid] == 0)
|
|
||||||
return;
|
|
||||||
|
|
||||||
/* Check the boundaries and update if necessary */
|
|
||||||
if (node_boundary_start_pfn[nid] < *start_pfn)
|
|
||||||
*start_pfn = node_boundary_start_pfn[nid];
|
|
||||||
if (node_boundary_end_pfn[nid] > *end_pfn)
|
|
||||||
*end_pfn = node_boundary_end_pfn[nid];
|
|
||||||
}
|
|
||||||
#else
|
|
||||||
void __init push_node_boundaries(unsigned int nid,
|
|
||||||
unsigned long start_pfn, unsigned long end_pfn) {}
|
|
||||||
|
|
||||||
static void __meminit account_node_boundary(unsigned int nid,
|
|
||||||
unsigned long *start_pfn, unsigned long *end_pfn) {}
|
|
||||||
#endif
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* get_pfn_range_for_nid - Return the start and end page frames for a node
|
* get_pfn_range_for_nid - Return the start and end page frames for a node
|
||||||
* @nid: The nid to return the range for. If MAX_NUMNODES, the min and max PFN are returned.
|
* @nid: The nid to return the range for. If MAX_NUMNODES, the min and max PFN are returned.
|
||||||
|
@ -3185,9 +3123,6 @@ void __meminit get_pfn_range_for_nid(unsigned int nid,
|
||||||
|
|
||||||
if (*start_pfn == -1UL)
|
if (*start_pfn == -1UL)
|
||||||
*start_pfn = 0;
|
*start_pfn = 0;
|
||||||
|
|
||||||
/* Push the node boundaries out if requested */
|
|
||||||
account_node_boundary(nid, start_pfn, end_pfn);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -3793,10 +3728,6 @@ void __init remove_all_active_ranges(void)
|
||||||
{
|
{
|
||||||
memset(early_node_map, 0, sizeof(early_node_map));
|
memset(early_node_map, 0, sizeof(early_node_map));
|
||||||
nr_nodemap_entries = 0;
|
nr_nodemap_entries = 0;
|
||||||
#ifdef CONFIG_MEMORY_HOTPLUG_RESERVE
|
|
||||||
memset(node_boundary_start_pfn, 0, sizeof(node_boundary_start_pfn));
|
|
||||||
memset(node_boundary_end_pfn, 0, sizeof(node_boundary_end_pfn));
|
|
||||||
#endif /* CONFIG_MEMORY_HOTPLUG_RESERVE */
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Compare two active node_active_regions */
|
/* Compare two active node_active_regions */
|
||||||
|
|
Loading…
Reference in a new issue