]> asedeno.scripts.mit.edu Git - linux.git/commitdiff
mm: remove cold parameter from free_hot_cold_page*
authorMel Gorman <mgorman@techsingularity.net>
Thu, 16 Nov 2017 01:37:59 +0000 (17:37 -0800)
committerLinus Torvalds <torvalds@linux-foundation.org>
Thu, 16 Nov 2017 02:21:06 +0000 (18:21 -0800)
Most callers users of free_hot_cold_page claim the pages being released
are cache hot.  The exception is the page reclaim paths where it is
likely that enough pages will be freed in the near future that the
per-cpu lists are going to be recycled and the cache hotness information
is lost.  As no one really cares about the hotness of pages being
released to the allocator, just ditch the parameter.

The APIs are renamed to indicate that it's no longer about hot/cold
pages.  It should also be less confusing as there are subtle differences
between them.  __free_pages drops a reference and frees a page when the
refcount reaches zero.  free_hot_cold_page handled pages whose refcount
was already zero which is non-obvious from the name.  free_unref_page
should be more obvious.

No performance impact is expected as the overhead is marginal.  The
parameter is removed simply because it is a bit stupid to have a useless
parameter copied everywhere.

[mgorman@techsingularity.net: add pages to head, not tail]
Link: http://lkml.kernel.org/r/20171019154321.qtpzaeftoyyw4iey@techsingularity.net
Link: http://lkml.kernel.org/r/20171018075952.10627-8-mgorman@techsingularity.net
Signed-off-by: Mel Gorman <mgorman@techsingularity.net>
Acked-by: Vlastimil Babka <vbabka@suse.cz>
Cc: Andi Kleen <ak@linux.intel.com>
Cc: Dave Chinner <david@fromorbit.com>
Cc: Dave Hansen <dave.hansen@intel.com>
Cc: Jan Kara <jack@suse.cz>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
arch/powerpc/mm/mmu_context_book3s64.c
arch/powerpc/mm/pgtable_64.c
arch/sparc/mm/init_64.c
arch/tile/mm/homecache.c
include/linux/gfp.h
include/trace/events/kmem.h
mm/page_alloc.c
mm/rmap.c
mm/swap.c
mm/vmscan.c

index 05e15386d4cb356da7e5b320be4b92907c99500b..a7e998158f372368e517cb022321004c6fe43ea0 100644 (file)
@@ -200,7 +200,7 @@ static void destroy_pagetable_page(struct mm_struct *mm)
        /* We allow PTE_FRAG_NR fragments from a PTE page */
        if (page_ref_sub_and_test(page, PTE_FRAG_NR - count)) {
                pgtable_page_dtor(page);
-               free_hot_cold_page(page, 0);
+               free_unref_page(page);
        }
 }
 
index ac0717a90ca6bd7be1a9071666e61c4e78782ceb..1ec3aee43624ff36f9937ffe9ed4d75af929e0f3 100644 (file)
@@ -404,7 +404,7 @@ void pte_fragment_free(unsigned long *table, int kernel)
        if (put_page_testzero(page)) {
                if (!kernel)
                        pgtable_page_dtor(page);
-               free_hot_cold_page(page, 0);
+               free_unref_page(page);
        }
 }
 
index 051f7340179389a85bf3fe89e019e7ab5c60c816..55ba62957e644116b2e30359f803081ddcd41313 100644 (file)
@@ -2939,7 +2939,7 @@ pgtable_t pte_alloc_one(struct mm_struct *mm,
        if (!page)
                return NULL;
        if (!pgtable_page_ctor(page)) {
-               free_hot_cold_page(page, 0);
+               free_unref_page(page);
                return NULL;
        }
        return (pte_t *) page_address(page);
index b51cc28acd0a11e27478697c0c87dec0a62930a6..4432f31e84795cca9e452042d09435b0768d5d54 100644 (file)
@@ -409,7 +409,7 @@ void __homecache_free_pages(struct page *page, unsigned int order)
        if (put_page_testzero(page)) {
                homecache_change_page_home(page, order, PAGE_HOME_HASH);
                if (order == 0) {
-                       free_hot_cold_page(page, false);
+                       free_unref_page(page);
                } else {
                        init_page_count(page);
                        __free_pages(page, order);
index b041f94678de5ac8889f22eedc3f5b01ed656f01..f7e62d9096fe8ecbbe5098c97548ed25bc3b7f45 100644 (file)
@@ -530,8 +530,8 @@ void * __meminit alloc_pages_exact_nid(int nid, size_t size, gfp_t gfp_mask);
 
 extern void __free_pages(struct page *page, unsigned int order);
 extern void free_pages(unsigned long addr, unsigned int order);
-extern void free_hot_cold_page(struct page *page, bool cold);
-extern void free_hot_cold_page_list(struct list_head *list, bool cold);
+extern void free_unref_page(struct page *page);
+extern void free_unref_page_list(struct list_head *list);
 
 struct page_frag_cache;
 extern void __page_frag_cache_drain(struct page *page, unsigned int count);
index 285feeadac39fb7729d33c6118421d1a340e71ab..eb57e3037deb0d912975ea6eba5223f47917ca07 100644 (file)
@@ -172,24 +172,21 @@ TRACE_EVENT(mm_page_free,
 
 TRACE_EVENT(mm_page_free_batched,
 
-       TP_PROTO(struct page *page, int cold),
+       TP_PROTO(struct page *page),
 
-       TP_ARGS(page, cold),
+       TP_ARGS(page),
 
        TP_STRUCT__entry(
                __field(        unsigned long,  pfn             )
-               __field(        int,            cold            )
        ),
 
        TP_fast_assign(
                __entry->pfn            = page_to_pfn(page);
-               __entry->cold           = cold;
        ),
 
-       TP_printk("page=%p pfn=%lu order=0 cold=%d",
+       TP_printk("page=%p pfn=%lu order=0",
                        pfn_to_page(__entry->pfn),
-                       __entry->pfn,
-                       __entry->cold)
+                       __entry->pfn)
 );
 
 TRACE_EVENT(mm_page_alloc,
index 6a3c4a1d513f3d659b059d7e788dbb1f07b91a36..f265d37b31522c65b59bfc397c6f1457224cf752 100644 (file)
@@ -2611,7 +2611,7 @@ void mark_free_pages(struct zone *zone)
 }
 #endif /* CONFIG_PM */
 
-static bool free_hot_cold_page_prepare(struct page *page, unsigned long pfn)
+static bool free_unref_page_prepare(struct page *page, unsigned long pfn)
 {
        int migratetype;
 
@@ -2623,8 +2623,7 @@ static bool free_hot_cold_page_prepare(struct page *page, unsigned long pfn)
        return true;
 }
 
-static void free_hot_cold_page_commit(struct page *page, unsigned long pfn,
-                               bool cold)
+static void free_unref_page_commit(struct page *page, unsigned long pfn)
 {
        struct zone *zone = page_zone(page);
        struct per_cpu_pages *pcp;
@@ -2649,10 +2648,7 @@ static void free_hot_cold_page_commit(struct page *page, unsigned long pfn,
        }
 
        pcp = &this_cpu_ptr(zone->pageset)->pcp;
-       if (!cold)
-               list_add(&page->lru, &pcp->lists[migratetype]);
-       else
-               list_add_tail(&page->lru, &pcp->lists[migratetype]);
+       list_add(&page->lru, &pcp->lists[migratetype]);
        pcp->count++;
        if (pcp->count >= pcp->high) {
                unsigned long batch = READ_ONCE(pcp->batch);
@@ -2663,25 +2659,24 @@ static void free_hot_cold_page_commit(struct page *page, unsigned long pfn,
 
 /*
  * Free a 0-order page
- * cold == true ? free a cold page : free a hot page
  */
-void free_hot_cold_page(struct page *page, bool cold)
+void free_unref_page(struct page *page)
 {
        unsigned long flags;
        unsigned long pfn = page_to_pfn(page);
 
-       if (!free_hot_cold_page_prepare(page, pfn))
+       if (!free_unref_page_prepare(page, pfn))
                return;
 
        local_irq_save(flags);
-       free_hot_cold_page_commit(page, pfn, cold);
+       free_unref_page_commit(page, pfn);
        local_irq_restore(flags);
 }
 
 /*
  * Free a list of 0-order pages
  */
-void free_hot_cold_page_list(struct list_head *list, bool cold)
+void free_unref_page_list(struct list_head *list)
 {
        struct page *page, *next;
        unsigned long flags, pfn;
@@ -2689,7 +2684,7 @@ void free_hot_cold_page_list(struct list_head *list, bool cold)
        /* Prepare pages for freeing */
        list_for_each_entry_safe(page, next, list, lru) {
                pfn = page_to_pfn(page);
-               if (!free_hot_cold_page_prepare(page, pfn))
+               if (!free_unref_page_prepare(page, pfn))
                        list_del(&page->lru);
                set_page_private(page, pfn);
        }
@@ -2699,8 +2694,8 @@ void free_hot_cold_page_list(struct list_head *list, bool cold)
                unsigned long pfn = page_private(page);
 
                set_page_private(page, 0);
-               trace_mm_page_free_batched(page, cold);
-               free_hot_cold_page_commit(page, pfn, cold);
+               trace_mm_page_free_batched(page);
+               free_unref_page_commit(page, pfn);
        }
        local_irq_restore(flags);
 }
@@ -4301,7 +4296,7 @@ void __free_pages(struct page *page, unsigned int order)
 {
        if (put_page_testzero(page)) {
                if (order == 0)
-                       free_hot_cold_page(page, false);
+                       free_unref_page(page);
                else
                        __free_pages_ok(page, order);
        }
@@ -4359,7 +4354,7 @@ void __page_frag_cache_drain(struct page *page, unsigned int count)
                unsigned int order = compound_order(page);
 
                if (order == 0)
-                       free_hot_cold_page(page, false);
+                       free_unref_page(page);
                else
                        __free_pages_ok(page, order);
        }
index 6b5a0f219ac0821f9298e6ae6fcd266762ddb3b5..47db27f8049e105b88f1ed60054b1576b8056dac 100644 (file)
--- a/mm/rmap.c
+++ b/mm/rmap.c
@@ -1321,7 +1321,7 @@ void page_remove_rmap(struct page *page, bool compound)
         * It would be tidy to reset the PageAnon mapping here,
         * but that might overwrite a racing page_add_anon_rmap
         * which increments mapcount after us but sets mapping
-        * before us: so leave the reset to free_hot_cold_page,
+        * before us: so leave the reset to free_unref_page,
         * and remember that it's only reliable while mapped.
         * Leaving it set also helps swapoff to reinstate ptes
         * faster for those pages still in swapcache.
index 29cf75f1a860f8a913de870dd255f8c0f35cd82d..b480279c760ca31daa20da2aa7949a2c209937ad 100644 (file)
--- a/mm/swap.c
+++ b/mm/swap.c
@@ -76,7 +76,7 @@ static void __page_cache_release(struct page *page)
 static void __put_single_page(struct page *page)
 {
        __page_cache_release(page);
-       free_hot_cold_page(page, false);
+       free_unref_page(page);
 }
 
 static void __put_compound_page(struct page *page)
@@ -817,7 +817,7 @@ void release_pages(struct page **pages, int nr)
                spin_unlock_irqrestore(&locked_pgdat->lru_lock, flags);
 
        mem_cgroup_uncharge_list(&pages_to_free);
-       free_hot_cold_page_list(&pages_to_free, 0);
+       free_unref_page_list(&pages_to_free);
 }
 EXPORT_SYMBOL(release_pages);
 
index 2852b8c5a917b327a56fb88c29e15db5914dc53d..c02c850ea3490af95fde44f94bd199fbdc500684 100644 (file)
@@ -1349,7 +1349,7 @@ static unsigned long shrink_page_list(struct list_head *page_list,
 
        mem_cgroup_uncharge_list(&free_pages);
        try_to_unmap_flush();
-       free_hot_cold_page_list(&free_pages, true);
+       free_unref_page_list(&free_pages);
 
        list_splice(&ret_pages, page_list);
        count_vm_events(PGACTIVATE, pgactivate);
@@ -1824,7 +1824,7 @@ shrink_inactive_list(unsigned long nr_to_scan, struct lruvec *lruvec,
        spin_unlock_irq(&pgdat->lru_lock);
 
        mem_cgroup_uncharge_list(&page_list);
-       free_hot_cold_page_list(&page_list, true);
+       free_unref_page_list(&page_list);
 
        /*
         * If reclaim is isolating dirty pages under writeback, it implies
@@ -2063,7 +2063,7 @@ static void shrink_active_list(unsigned long nr_to_scan,
        spin_unlock_irq(&pgdat->lru_lock);
 
        mem_cgroup_uncharge_list(&l_hold);
-       free_hot_cold_page_list(&l_hold, true);
+       free_unref_page_list(&l_hold);
        trace_mm_vmscan_lru_shrink_active(pgdat->node_id, nr_taken, nr_activate,
                        nr_deactivate, nr_rotated, sc->priority, file);
 }