HWPOISON: return early on non-LRU pages
[safe/jmp/linux-2.6] / mm / memory_hotplug.c
index b173711..821dee5 100644 (file)
@@ -216,7 +216,8 @@ static int __meminit __add_zone(struct zone *zone, unsigned long phys_start_pfn)
        return 0;
 }
 
-static int __meminit __add_section(struct zone *zone, unsigned long phys_start_pfn)
+static int __meminit __add_section(int nid, struct zone *zone,
+                                       unsigned long phys_start_pfn)
 {
        int nr_pages = PAGES_PER_SECTION;
        int ret;
@@ -234,7 +235,7 @@ static int __meminit __add_section(struct zone *zone, unsigned long phys_start_p
        if (ret < 0)
                return ret;
 
-       return register_new_memory(__pfn_to_section(phys_start_pfn));
+       return register_new_memory(nid, __pfn_to_section(phys_start_pfn));
 }
 
 #ifdef CONFIG_SPARSEMEM_VMEMMAP
@@ -273,8 +274,8 @@ static int __remove_section(struct zone *zone, struct mem_section *ms)
  * call this function after deciding the zone to which to
  * add the new pages.
  */
-int __ref __add_pages(struct zone *zone, unsigned long phys_start_pfn,
-                unsigned long nr_pages)
+int __ref __add_pages(int nid, struct zone *zone, unsigned long phys_start_pfn,
+                       unsigned long nr_pages)
 {
        unsigned long i;
        int err = 0;
@@ -284,7 +285,7 @@ int __ref __add_pages(struct zone *zone, unsigned long phys_start_pfn,
        end_sec = pfn_to_section_nr(phys_start_pfn + nr_pages - 1);
 
        for (i = start_sec; i <= end_sec; i++) {
-               err = __add_section(zone, i << PFN_SECTION_SHIFT);
+               err = __add_section(nid, zone, i << PFN_SECTION_SHIFT);
 
                /*
                 * EEXIST is finally dealt with by ioresource collision
@@ -338,8 +339,11 @@ EXPORT_SYMBOL_GPL(__remove_pages);
 
 void online_page(struct page *page)
 {
+       unsigned long pfn = page_to_pfn(page);
+
        totalram_pages++;
-       num_physpages++;
+       if (pfn >= num_physpages)
+               num_physpages = pfn + 1;
 
 #ifdef CONFIG_HIGHMEM
        if (PageHighMem(page))
@@ -409,7 +413,7 @@ int online_pages(unsigned long pfn, unsigned long nr_pages)
        if (!populated_zone(zone))
                need_zonelists_rebuild = 1;
 
-       ret = walk_memory_resource(pfn, nr_pages, &onlined_pages,
+       ret = walk_system_ram_range(pfn, nr_pages, &onlined_pages,
                online_pages_range);
        if (ret) {
                printk(KERN_DEBUG "online_pages %lx at %lx failed\n",
@@ -421,7 +425,9 @@ int online_pages(unsigned long pfn, unsigned long nr_pages)
        zone->present_pages += onlined_pages;
        zone->zone_pgdat->node_present_pages += onlined_pages;
 
-       setup_per_zone_pages_min();
+       zone_pcp_update(zone);
+       setup_per_zone_wmarks();
+       calculate_zone_inactive_ratio(zone);
        if (onlined_pages) {
                kswapd_run(zone_to_nid(zone));
                node_set_state(zone_to_nid(zone), N_HIGH_MEMORY);
@@ -626,15 +632,12 @@ int scan_lru_pages(unsigned long start, unsigned long end)
 }
 
 static struct page *
-hotremove_migrate_alloc(struct page *page,
-                       unsigned long private,
-                       int **x)
+hotremove_migrate_alloc(struct page *page, unsigned long private, int **x)
 {
-       /* This should be improoooooved!! */
-       return alloc_page(GFP_HIGHUSER_PAGECACHE);
+       /* This should be improooooved!! */
+       return alloc_page(GFP_HIGHUSER_MOVABLE);
 }
 
-
 #define NR_OFFLINE_AT_ONCE_PAGES       (256)
 static int
 do_migrate_range(unsigned long start_pfn, unsigned long end_pfn)
@@ -702,7 +705,7 @@ offline_isolated_pages_cb(unsigned long start, unsigned long nr_pages,
 static void
 offline_isolated_pages(unsigned long start_pfn, unsigned long end_pfn)
 {
-       walk_memory_resource(start_pfn, end_pfn - start_pfn, NULL,
+       walk_system_ram_range(start_pfn, end_pfn - start_pfn, NULL,
                                offline_isolated_pages_cb);
 }
 
@@ -728,7 +731,7 @@ check_pages_isolated(unsigned long start_pfn, unsigned long end_pfn)
        long offlined = 0;
        int ret;
 
-       ret = walk_memory_resource(start_pfn, end_pfn - start_pfn, &offlined,
+       ret = walk_system_ram_range(start_pfn, end_pfn - start_pfn, &offlined,
                        check_pages_isolated_cb);
        if (ret < 0)
                offlined = (long)ret;
@@ -832,7 +835,9 @@ repeat:
        zone->present_pages -= offlined_pages;
        zone->zone_pgdat->node_present_pages -= offlined_pages;
        totalram_pages -= offlined_pages;
-       num_physpages -= offlined_pages;
+
+       setup_per_zone_wmarks();
+       calculate_zone_inactive_ratio(zone);
 
        vm_total_pages = nr_free_pagecache_pages();
        writeback_set_ratelimit();