]> asedeno.scripts.mit.edu Git - linux.git/blobdiff - mm/list_lru.c
IB/core: Ensure we map P2P memory correctly in rdma_rw_ctx_[init|destroy]()
[linux.git] / mm / list_lru.c
index c9bdde9c03d13e6902dc8ec224d227b625b9af6f..5b30625fd3651a51deab140c3412897a51624f4f 100644 (file)
@@ -209,17 +209,15 @@ unsigned long list_lru_count_node(struct list_lru *lru, int nid)
 EXPORT_SYMBOL_GPL(list_lru_count_node);
 
 static unsigned long
-__list_lru_walk_one(struct list_lru *lru, int nid, int memcg_idx,
+__list_lru_walk_one(struct list_lru_node *nlru, int memcg_idx,
                    list_lru_walk_cb isolate, void *cb_arg,
                    unsigned long *nr_to_walk)
 {
 
-       struct list_lru_node *nlru = &lru->node[nid];
        struct list_lru_one *l;
        struct list_head *item, *n;
        unsigned long isolated = 0;
 
-       spin_lock(&nlru->lock);
        l = list_lru_from_memcg_idx(nlru, memcg_idx);
 restart:
        list_for_each_safe(item, n, &l->list) {
@@ -265,8 +263,6 @@ __list_lru_walk_one(struct list_lru *lru, int nid, int memcg_idx,
                        BUG();
                }
        }
-
-       spin_unlock(&nlru->lock);
        return isolated;
 }
 
@@ -275,11 +271,32 @@ list_lru_walk_one(struct list_lru *lru, int nid, struct mem_cgroup *memcg,
                  list_lru_walk_cb isolate, void *cb_arg,
                  unsigned long *nr_to_walk)
 {
-       return __list_lru_walk_one(lru, nid, memcg_cache_id(memcg),
-                                  isolate, cb_arg, nr_to_walk);
+       struct list_lru_node *nlru = &lru->node[nid];
+       unsigned long ret;
+
+       spin_lock(&nlru->lock);
+       ret = __list_lru_walk_one(nlru, memcg_cache_id(memcg), isolate, cb_arg,
+                                 nr_to_walk);
+       spin_unlock(&nlru->lock);
+       return ret;
 }
 EXPORT_SYMBOL_GPL(list_lru_walk_one);
 
+unsigned long
+list_lru_walk_one_irq(struct list_lru *lru, int nid, struct mem_cgroup *memcg,
+                     list_lru_walk_cb isolate, void *cb_arg,
+                     unsigned long *nr_to_walk)
+{
+       struct list_lru_node *nlru = &lru->node[nid];
+       unsigned long ret;
+
+       spin_lock_irq(&nlru->lock);
+       ret = __list_lru_walk_one(nlru, memcg_cache_id(memcg), isolate, cb_arg,
+                                 nr_to_walk);
+       spin_unlock_irq(&nlru->lock);
+       return ret;
+}
+
 unsigned long list_lru_walk_node(struct list_lru *lru, int nid,
                                 list_lru_walk_cb isolate, void *cb_arg,
                                 unsigned long *nr_to_walk)
@@ -287,12 +304,18 @@ unsigned long list_lru_walk_node(struct list_lru *lru, int nid,
        long isolated = 0;
        int memcg_idx;
 
-       isolated += __list_lru_walk_one(lru, nid, -1, isolate, cb_arg,
-                                       nr_to_walk);
+       isolated += list_lru_walk_one(lru, nid, NULL, isolate, cb_arg,
+                                     nr_to_walk);
        if (*nr_to_walk > 0 && list_lru_memcg_aware(lru)) {
                for_each_memcg_cache_index(memcg_idx) {
-                       isolated += __list_lru_walk_one(lru, nid, memcg_idx,
-                                               isolate, cb_arg, nr_to_walk);
+                       struct list_lru_node *nlru = &lru->node[nid];
+
+                       spin_lock(&nlru->lock);
+                       isolated += __list_lru_walk_one(nlru, memcg_idx,
+                                                       isolate, cb_arg,
+                                                       nr_to_walk);
+                       spin_unlock(&nlru->lock);
+
                        if (*nr_to_walk <= 0)
                                break;
                }