git://ftp.safe.ca
/
safe
/
jmp
/
linux-2.6
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
vmscan: kswapd: don't retry balance_pgdat() if all zones are unreclaimable
[safe/jmp/linux-2.6]
/
mm
/
memory_hotplug.c
diff --git
a/mm/memory_hotplug.c
b/mm/memory_hotplug.c
index
6837a10
..
030ce8a
100644
(file)
--- a/
mm/memory_hotplug.c
+++ b/
mm/memory_hotplug.c
@@
-22,11
+22,12
@@
#include <linux/highmem.h>
#include <linux/vmalloc.h>
#include <linux/ioport.h>
#include <linux/highmem.h>
#include <linux/vmalloc.h>
#include <linux/ioport.h>
-#include <linux/cpuset.h>
#include <linux/delay.h>
#include <linux/migrate.h>
#include <linux/page-isolation.h>
#include <linux/pfn.h>
#include <linux/delay.h>
#include <linux/migrate.h>
#include <linux/page-isolation.h>
#include <linux/pfn.h>
+#include <linux/suspend.h>
+#include <linux/mm_inline.h>
#include <asm/tlbflush.h>
#include <asm/tlbflush.h>
@@
-71,7
+72,9
@@
static void get_page_bootmem(unsigned long info, struct page *page, int type)
atomic_inc(&page->_count);
}
atomic_inc(&page->_count);
}
-void put_page_bootmem(struct page *page)
+/* reference to __meminit __free_pages_bootmem is valid
+ * so use __ref to tell modpost not to generate a warning */
+void __ref put_page_bootmem(struct page *page)
{
int type;
{
int type;
@@
-190,7
+193,7
@@
static void grow_pgdat_span(struct pglist_data *pgdat, unsigned long start_pfn,
pgdat->node_start_pfn;
}
pgdat->node_start_pfn;
}
-static int __add_zone(struct zone *zone, unsigned long phys_start_pfn)
+static int __
meminit __
add_zone(struct zone *zone, unsigned long phys_start_pfn)
{
struct pglist_data *pgdat = zone->zone_pgdat;
int nr_pages = PAGES_PER_SECTION;
{
struct pglist_data *pgdat = zone->zone_pgdat;
int nr_pages = PAGES_PER_SECTION;
@@
-217,7
+220,8
@@
static int __add_zone(struct zone *zone, unsigned long phys_start_pfn)
return 0;
}
return 0;
}
-static int __add_section(struct zone *zone, unsigned long phys_start_pfn)
+static int __meminit __add_section(int nid, struct zone *zone,
+ unsigned long phys_start_pfn)
{
int nr_pages = PAGES_PER_SECTION;
int ret;
{
int nr_pages = PAGES_PER_SECTION;
int ret;
@@
-235,7
+239,7
@@
static int __add_section(struct zone *zone, unsigned long phys_start_pfn)
if (ret < 0)
return ret;
if (ret < 0)
return ret;
- return register_new_memory(__pfn_to_section(phys_start_pfn));
+ return register_new_memory(
nid,
__pfn_to_section(phys_start_pfn));
}
#ifdef CONFIG_SPARSEMEM_VMEMMAP
}
#ifdef CONFIG_SPARSEMEM_VMEMMAP
@@
-274,8
+278,8
@@
static int __remove_section(struct zone *zone, struct mem_section *ms)
* call this function after deciding the zone to which to
* add the new pages.
*/
* call this function after deciding the zone to which to
* add the new pages.
*/
-int __
add_pages(
struct zone *zone, unsigned long phys_start_pfn,
- unsigned long nr_pages)
+int __
ref __add_pages(int nid,
struct zone *zone, unsigned long phys_start_pfn,
+
unsigned long nr_pages)
{
unsigned long i;
int err = 0;
{
unsigned long i;
int err = 0;
@@
-285,7
+289,7
@@
int __add_pages(struct zone *zone, unsigned long phys_start_pfn,
end_sec = pfn_to_section_nr(phys_start_pfn + nr_pages - 1);
for (i = start_sec; i <= end_sec; i++) {
end_sec = pfn_to_section_nr(phys_start_pfn + nr_pages - 1);
for (i = start_sec; i <= end_sec; i++) {
- err = __add_section(zone, i << PFN_SECTION_SHIFT);
+ err = __add_section(
nid,
zone, i << PFN_SECTION_SHIFT);
/*
* EEXIST is finally dealt with by ioresource collision
/*
* EEXIST is finally dealt with by ioresource collision
@@
-339,8
+343,11
@@
EXPORT_SYMBOL_GPL(__remove_pages);
void online_page(struct page *page)
{
void online_page(struct page *page)
{
+ unsigned long pfn = page_to_pfn(page);
+
totalram_pages++;
totalram_pages++;
- num_physpages++;
+ if (pfn >= num_physpages)
+ num_physpages = pfn + 1;
#ifdef CONFIG_HIGHMEM
if (PageHighMem(page))
#ifdef CONFIG_HIGHMEM
if (PageHighMem(page))
@@
-410,7
+417,7
@@
int online_pages(unsigned long pfn, unsigned long nr_pages)
if (!populated_zone(zone))
need_zonelists_rebuild = 1;
if (!populated_zone(zone))
need_zonelists_rebuild = 1;
- ret = walk_
memory_resourc
e(pfn, nr_pages, &onlined_pages,
+ ret = walk_
system_ram_rang
e(pfn, nr_pages, &onlined_pages,
online_pages_range);
if (ret) {
printk(KERN_DEBUG "online_pages %lx at %lx failed\n",
online_pages_range);
if (ret) {
printk(KERN_DEBUG "online_pages %lx at %lx failed\n",
@@
-422,7
+429,9
@@
int online_pages(unsigned long pfn, unsigned long nr_pages)
zone->present_pages += onlined_pages;
zone->zone_pgdat->node_present_pages += onlined_pages;
zone->present_pages += onlined_pages;
zone->zone_pgdat->node_present_pages += onlined_pages;
- setup_per_zone_pages_min();
+ zone_pcp_update(zone);
+ setup_per_zone_wmarks();
+ calculate_zone_inactive_ratio(zone);
if (onlined_pages) {
kswapd_run(zone_to_nid(zone));
node_set_state(zone_to_nid(zone), N_HIGH_MEMORY);
if (onlined_pages) {
kswapd_run(zone_to_nid(zone));
node_set_state(zone_to_nid(zone), N_HIGH_MEMORY);
@@
-442,7
+451,8
@@
int online_pages(unsigned long pfn, unsigned long nr_pages)
}
#endif /* CONFIG_MEMORY_HOTPLUG_SPARSE */
}
#endif /* CONFIG_MEMORY_HOTPLUG_SPARSE */
-static pg_data_t *hotadd_new_pgdat(int nid, u64 start)
+/* we are OK calling __meminit stuff here - we have CONFIG_MEMORY_HOTPLUG */
+static pg_data_t __ref *hotadd_new_pgdat(int nid, u64 start)
{
struct pglist_data *pgdat;
unsigned long zones_size[MAX_NR_ZONES] = {0};
{
struct pglist_data *pgdat;
unsigned long zones_size[MAX_NR_ZONES] = {0};
@@
-471,21
+481,26
@@
static void rollback_node_hotadd(int nid, pg_data_t *pgdat)
}
}
-int add_memory(int nid, u64 start, u64 size)
+/* we are OK calling __meminit stuff here - we have CONFIG_MEMORY_HOTPLUG */
+int __ref add_memory(int nid, u64 start, u64 size)
{
pg_data_t *pgdat = NULL;
int new_pgdat = 0;
struct resource *res;
int ret;
{
pg_data_t *pgdat = NULL;
int new_pgdat = 0;
struct resource *res;
int ret;
+ lock_system_sleep();
+
res = register_memory_resource(start, size);
res = register_memory_resource(start, size);
+ ret = -EEXIST;
if (!res)
if (!res)
-
return -EEXIST
;
+
goto out
;
if (!node_online(nid)) {
pgdat = hotadd_new_pgdat(nid, start);
if (!node_online(nid)) {
pgdat = hotadd_new_pgdat(nid, start);
+ ret = -ENOMEM;
if (!pgdat)
if (!pgdat)
-
return -ENOMEM
;
+
goto out
;
new_pgdat = 1;
}
new_pgdat = 1;
}
@@
-498,8
+513,6
@@
int add_memory(int nid, u64 start, u64 size)
/* we online node here. we can't roll back from here. */
node_set_online(nid);
/* we online node here. we can't roll back from here. */
node_set_online(nid);
- cpuset_track_online_nodes();
-
if (new_pgdat) {
ret = register_one_node(nid);
/*
if (new_pgdat) {
ret = register_one_node(nid);
/*
@@
-510,7
+523,8
@@
int add_memory(int nid, u64 start, u64 size)
BUG_ON(ret);
}
BUG_ON(ret);
}
- return ret;
+ goto out;
+
error:
/* rollback pgdat allocation and others */
if (new_pgdat)
error:
/* rollback pgdat allocation and others */
if (new_pgdat)
@@
-518,6
+532,8
@@
error:
if (res)
release_memory_resource(res);
if (res)
release_memory_resource(res);
+out:
+ unlock_system_sleep();
return ret;
}
EXPORT_SYMBOL_GPL(add_memory);
return ret;
}
EXPORT_SYMBOL_GPL(add_memory);
@@
-628,15
+644,12
@@
int scan_lru_pages(unsigned long start, unsigned long end)
}
static struct page *
}
static struct page *
-hotremove_migrate_alloc(struct page *page,
- unsigned long private,
- int **x)
+hotremove_migrate_alloc(struct page *page, unsigned long private, int **x)
{
{
- /* This should be improoooo
o
ved!! */
- return alloc_page(GFP_HIGHUSER_
PAGECACH
E);
+ /* This should be improooooved!! */
+ return alloc_page(GFP_HIGHUSER_
MOVABL
E);
}
}
-
#define NR_OFFLINE_AT_ONCE_PAGES (256)
static int
do_migrate_range(unsigned long start_pfn, unsigned long end_pfn)
#define NR_OFFLINE_AT_ONCE_PAGES (256)
static int
do_migrate_range(unsigned long start_pfn, unsigned long end_pfn)
@@
-662,6
+675,9
@@
do_migrate_range(unsigned long start_pfn, unsigned long end_pfn)
if (!ret) { /* Success */
list_add_tail(&page->lru, &source);
move_pages--;
if (!ret) { /* Success */
list_add_tail(&page->lru, &source);
move_pages--;
+ inc_zone_page_state(page, NR_ISOLATED_ANON +
+ page_is_file_cache(page));
+
} else {
/* Becasue we don't have big zone->lock. we should
check this again here. */
} else {
/* Becasue we don't have big zone->lock. we should
check this again here. */
@@
-684,7
+700,7
@@
do_migrate_range(unsigned long start_pfn, unsigned long end_pfn)
if (list_empty(&source))
goto out;
/* this function returns # of failed pages */
if (list_empty(&source))
goto out;
/* this function returns # of failed pages */
- ret = migrate_pages(&source, hotremove_migrate_alloc, 0);
+ ret = migrate_pages(&source, hotremove_migrate_alloc, 0
, 1
);
out:
return ret;
out:
return ret;
@@
-704,7
+720,7
@@
offline_isolated_pages_cb(unsigned long start, unsigned long nr_pages,
static void
offline_isolated_pages(unsigned long start_pfn, unsigned long end_pfn)
{
static void
offline_isolated_pages(unsigned long start_pfn, unsigned long end_pfn)
{
- walk_
memory_resourc
e(start_pfn, end_pfn - start_pfn, NULL,
+ walk_
system_ram_rang
e(start_pfn, end_pfn - start_pfn, NULL,
offline_isolated_pages_cb);
}
offline_isolated_pages_cb);
}
@@
-730,14
+746,14
@@
check_pages_isolated(unsigned long start_pfn, unsigned long end_pfn)
long offlined = 0;
int ret;
long offlined = 0;
int ret;
- ret = walk_
memory_resourc
e(start_pfn, end_pfn - start_pfn, &offlined,
+ ret = walk_
system_ram_rang
e(start_pfn, end_pfn - start_pfn, &offlined,
check_pages_isolated_cb);
if (ret < 0)
offlined = (long)ret;
return offlined;
}
check_pages_isolated_cb);
if (ret < 0)
offlined = (long)ret;
return offlined;
}
-int offline_pages(unsigned long start_pfn,
+
static
int offline_pages(unsigned long start_pfn,
unsigned long end_pfn, unsigned long timeout)
{
unsigned long pfn, nr_pages, expire;
unsigned long end_pfn, unsigned long timeout)
{
unsigned long pfn, nr_pages, expire;
@@
-757,6
+773,8
@@
int offline_pages(unsigned long start_pfn,
if (!test_pages_in_a_zone(start_pfn, end_pfn))
return -EINVAL;
if (!test_pages_in_a_zone(start_pfn, end_pfn))
return -EINVAL;
+ lock_system_sleep();
+
zone = page_zone(pfn_to_page(start_pfn));
node = zone_to_nid(zone);
nr_pages = end_pfn - start_pfn;
zone = page_zone(pfn_to_page(start_pfn));
node = zone_to_nid(zone);
nr_pages = end_pfn - start_pfn;
@@
-764,7
+782,7
@@
int offline_pages(unsigned long start_pfn,
/* set above range as isolated */
ret = start_isolate_page_range(start_pfn, end_pfn);
if (ret)
/* set above range as isolated */
ret = start_isolate_page_range(start_pfn, end_pfn);
if (ret)
-
return re
t;
+
goto ou
t;
arg.start_pfn = start_pfn;
arg.nr_pages = nr_pages;
arg.start_pfn = start_pfn;
arg.nr_pages = nr_pages;
@@
-834,12
+852,19
@@
repeat:
zone->present_pages -= offlined_pages;
zone->zone_pgdat->node_present_pages -= offlined_pages;
totalram_pages -= offlined_pages;
zone->present_pages -= offlined_pages;
zone->zone_pgdat->node_present_pages -= offlined_pages;
totalram_pages -= offlined_pages;
- num_physpages -= offlined_pages;
+
+ setup_per_zone_wmarks();
+ calculate_zone_inactive_ratio(zone);
+ if (!node_present_pages(node)) {
+ node_clear_state(node, N_HIGH_MEMORY);
+ kswapd_stop(node);
+ }
vm_total_pages = nr_free_pagecache_pages();
writeback_set_ratelimit();
memory_notify(MEM_OFFLINE, &arg);
vm_total_pages = nr_free_pagecache_pages();
writeback_set_ratelimit();
memory_notify(MEM_OFFLINE, &arg);
+ unlock_system_sleep();
return 0;
failed_removal:
return 0;
failed_removal:
@@
-849,6
+874,8
@@
failed_removal:
/* pushback to free area */
undo_isolate_page_range(start_pfn, end_pfn);
/* pushback to free area */
undo_isolate_page_range(start_pfn, end_pfn);
+out:
+ unlock_system_sleep();
return ret;
}
return ret;
}