]> git.kernelconcepts.de Git - karo-tx-linux.git/commitdiff
dcache: convert to use new lru list infrastructure
authorDave Chinner <dchinner@redhat.com>
Wed, 3 Jul 2013 00:19:52 +0000 (10:19 +1000)
committerStephen Rothwell <sfr@canb.auug.org.au>
Wed, 17 Jul 2013 02:34:53 +0000 (12:34 +1000)
[glommer@openvz.org: don't reintroduce double decrement of nr_unused_dentries, adapted for new LRU return codes]
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Signed-off-by: Glauber Costa <glommer@openvz.org>
Cc: "Theodore Ts'o" <tytso@mit.edu>
Cc: Adrian Hunter <adrian.hunter@intel.com>
Cc: Al Viro <viro@zeniv.linux.org.uk>
Cc: Artem Bityutskiy <artem.bityutskiy@linux.intel.com>
Cc: Arve Hjønnevåg <arve@android.com>
Cc: Carlos Maiolino <cmaiolino@redhat.com>
Cc: Christoph Hellwig <hch@lst.de>
Cc: Chuck Lever <chuck.lever@oracle.com>
Cc: Daniel Vetter <daniel.vetter@ffwll.ch>
Cc: David Rientjes <rientjes@google.com>
Cc: Gleb Natapov <gleb@redhat.com>
Cc: Greg Thelen <gthelen@google.com>
Cc: J. Bruce Fields <bfields@redhat.com>
Cc: Jan Kara <jack@suse.cz>
Cc: Jerome Glisse <jglisse@redhat.com>
Cc: John Stultz <john.stultz@linaro.org>
Cc: KAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com>
Cc: Kent Overstreet <koverstreet@google.com>
Cc: Kirill A. Shutemov <kirill.shutemov@linux.intel.com>
Cc: Marcelo Tosatti <mtosatti@redhat.com>
Cc: Mel Gorman <mgorman@suse.de>
Cc: Steven Whitehouse <swhiteho@redhat.com>
Cc: Thomas Hellstrom <thellstrom@vmware.com>
Cc: Trond Myklebust <Trond.Myklebust@netapp.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
fs/dcache.c
fs/super.c
include/linux/fs.h

index 66f8a83fc899ef4ae850fdf6aa4f31a325001b9f..2275bd416736a32cf4ba64e4d3865b99186f62aa 100644 (file)
@@ -37,6 +37,7 @@
 #include <linux/rculist_bl.h>
 #include <linux/prefetch.h>
 #include <linux/ratelimit.h>
+#include <linux/list_lru.h>
 #include "internal.h"
 #include "mount.h"
 
@@ -331,20 +332,8 @@ static void dentry_unlink_inode(struct dentry * dentry)
  */
 static void dentry_lru_add(struct dentry *dentry)
 {
-       if (list_empty(&dentry->d_lru)) {
-               spin_lock(&dentry->d_sb->s_dentry_lru_lock);
-               list_add(&dentry->d_lru, &dentry->d_sb->s_dentry_lru);
-               dentry->d_sb->s_nr_dentry_unused++;
+       if (list_lru_add(&dentry->d_sb->s_dentry_lru, &dentry->d_lru))
                this_cpu_inc(nr_dentry_unused);
-               spin_unlock(&dentry->d_sb->s_dentry_lru_lock);
-       }
-}
-
-static void __dentry_lru_del(struct dentry *dentry)
-{
-       list_del_init(&dentry->d_lru);
-       dentry->d_sb->s_nr_dentry_unused--;
-       this_cpu_dec(nr_dentry_unused);
 }
 
 /*
@@ -362,26 +351,8 @@ static void dentry_lru_del(struct dentry *dentry)
                return;
        }
 
-       if (!list_empty(&dentry->d_lru)) {
-               spin_lock(&dentry->d_sb->s_dentry_lru_lock);
-               __dentry_lru_del(dentry);
-               spin_unlock(&dentry->d_sb->s_dentry_lru_lock);
-       }
-}
-
-static void dentry_lru_move_list(struct dentry *dentry, struct list_head *list)
-{
-       BUG_ON(dentry->d_flags & DCACHE_SHRINK_LIST);
-
-       spin_lock(&dentry->d_sb->s_dentry_lru_lock);
-       if (list_empty(&dentry->d_lru)) {
-               list_add_tail(&dentry->d_lru, list);
-       } else {
-               list_move_tail(&dentry->d_lru, list);
-               dentry->d_sb->s_nr_dentry_unused--;
+       if (list_lru_del(&dentry->d_sb->s_dentry_lru, &dentry->d_lru))
                this_cpu_dec(nr_dentry_unused);
-       }
-       spin_unlock(&dentry->d_sb->s_dentry_lru_lock);
 }
 
 /**
@@ -856,12 +827,72 @@ static void shrink_dentry_list(struct list_head *list)
        rcu_read_unlock();
 }
 
+static enum lru_status
+dentry_lru_isolate(struct list_head *item, spinlock_t *lru_lock, void *arg)
+{
+       struct list_head *freeable = arg;
+       struct dentry   *dentry = container_of(item, struct dentry, d_lru);
+
+
+       /*
+        * we are inverting the lru lock/dentry->d_lock here,
+        * so use a trylock. If we fail to get the lock, just skip
+        * it
+        */
+       if (!spin_trylock(&dentry->d_lock))
+               return LRU_SKIP;
+
+       /*
+        * Referenced dentries are still in use. If they have active
+        * counts, just remove them from the LRU. Otherwise give them
+        * another pass through the LRU.
+        */
+       if (dentry->d_count) {
+               list_del_init(&dentry->d_lru);
+               spin_unlock(&dentry->d_lock);
+               return LRU_REMOVED;
+       }
+
+       if (dentry->d_flags & DCACHE_REFERENCED) {
+               dentry->d_flags &= ~DCACHE_REFERENCED;
+               spin_unlock(&dentry->d_lock);
+
+               /*
+                * The list move itself will be made by the common LRU code. At
+                * this point, we've dropped the dentry->d_lock but keep the
+                * lru lock. This is safe to do, since every list movement is
+                * protected by the lru lock even if both locks are held.
+                *
+                * This is guaranteed by the fact that all LRU management
+                * functions are intermediated by the LRU API calls like
+                * list_lru_add and list_lru_del. List movement in this file
+                * only ever occur through this functions or through callbacks
+                * like this one, that are called from the LRU API.
+                *
+                * The only exceptions to this are functions like
+                * shrink_dentry_list, and code that first checks for the
+                * DCACHE_SHRINK_LIST flag.  Those are guaranteed to be
+                * operating only with stack provided lists after they are
+                * properly isolated from the main list.  It is thus, always a
+                * local access.
+                */
+               return LRU_ROTATE;
+       }
+
+       dentry->d_flags |= DCACHE_SHRINK_LIST;
+       list_move_tail(&dentry->d_lru, freeable);
+       this_cpu_dec(nr_dentry_unused);
+       spin_unlock(&dentry->d_lock);
+
+       return LRU_REMOVED;
+}
+
 /**
  * prune_dcache_sb - shrink the dcache
  * @sb: superblock
- * @count: number of entries to try to free
+ * @nr_to_scan : number of entries to try to free
  *
- * Attempt to shrink the superblock dcache LRU by @count entries. This is
+ * Attempt to shrink the superblock dcache LRU by @nr_to_scan entries. This is
  * done when we need more memory an called from the superblock shrinker
  * function.
  *
@@ -870,45 +901,12 @@ static void shrink_dentry_list(struct list_head *list)
  */
 long prune_dcache_sb(struct super_block *sb, unsigned long nr_to_scan)
 {
-       struct dentry *dentry;
-       LIST_HEAD(referenced);
-       LIST_HEAD(tmp);
-       long freed = 0;
-
-relock:
-       spin_lock(&sb->s_dentry_lru_lock);
-       while (!list_empty(&sb->s_dentry_lru)) {
-               dentry = list_entry(sb->s_dentry_lru.prev,
-                               struct dentry, d_lru);
-               BUG_ON(dentry->d_sb != sb);
-
-               if (!spin_trylock(&dentry->d_lock)) {
-                       spin_unlock(&sb->s_dentry_lru_lock);
-                       cpu_relax();
-                       goto relock;
-               }
-
-               if (dentry->d_flags & DCACHE_REFERENCED) {
-                       dentry->d_flags &= ~DCACHE_REFERENCED;
-                       list_move(&dentry->d_lru, &referenced);
-                       spin_unlock(&dentry->d_lock);
-               } else {
-                       list_move(&dentry->d_lru, &tmp);
-                       dentry->d_flags |= DCACHE_SHRINK_LIST;
-                       this_cpu_dec(nr_dentry_unused);
-                       sb->s_nr_dentry_unused--;
-                       spin_unlock(&dentry->d_lock);
-                       freed++;
-                       if (!--nr_to_scan)
-                               break;
-               }
-               cond_resched_lock(&sb->s_dentry_lru_lock);
-       }
-       if (!list_empty(&referenced))
-               list_splice(&referenced, &sb->s_dentry_lru);
-       spin_unlock(&sb->s_dentry_lru_lock);
+       LIST_HEAD(dispose);
+       long freed;
 
-       shrink_dentry_list(&tmp);
+       freed = list_lru_walk(&sb->s_dentry_lru, dentry_lru_isolate,
+                             &dispose, nr_to_scan);
+       shrink_dentry_list(&dispose);
        return freed;
 }
 
@@ -942,24 +940,10 @@ shrink_dcache_list(
  */
 void shrink_dcache_sb(struct super_block *sb)
 {
-       LIST_HEAD(tmp);
-
-       spin_lock(&sb->s_dentry_lru_lock);
-       while (!list_empty(&sb->s_dentry_lru)) {
-               /*
-                * account for removal here so we don't need to handle it later
-                * even though the dentry is no longer on the lru list.
-                */
-               list_splice_init(&sb->s_dentry_lru, &tmp);
-               this_cpu_sub(nr_dentry_unused, sb->s_nr_dentry_unused);
-               sb->s_nr_dentry_unused = 0;
-               spin_unlock(&sb->s_dentry_lru_lock);
+       long disposed;
 
-               shrink_dcache_list(&tmp);
-
-               spin_lock(&sb->s_dentry_lru_lock);
-       }
-       spin_unlock(&sb->s_dentry_lru_lock);
+       disposed = list_lru_dispose_all(&sb->s_dentry_lru, shrink_dcache_list);
+       this_cpu_sub(nr_dentry_unused, disposed);
 }
 EXPORT_SYMBOL(shrink_dcache_sb);
 
@@ -1232,7 +1216,8 @@ resume:
                if (dentry->d_count) {
                        dentry_lru_del(dentry);
                } else if (!(dentry->d_flags & DCACHE_SHRINK_LIST)) {
-                       dentry_lru_move_list(dentry, dispose);
+                       dentry_lru_del(dentry);
+                       list_add_tail(&dentry->d_lru, dispose);
                        dentry->d_flags |= DCACHE_SHRINK_LIST;
                        found++;
                }
index 3784de973f4438eac57f889aa6328a46ea52651b..21201f91548cbbf431812b740266f51f665ea8e5 100644 (file)
@@ -79,11 +79,11 @@ static unsigned long super_cache_scan(struct shrinker *shrink,
                fs_objects = sb->s_op->nr_cached_objects(sb);
 
        inodes = list_lru_count(&sb->s_inode_lru);
-       total_objects = sb->s_nr_dentry_unused + inodes + fs_objects + 1;
+       dentries = list_lru_count(&sb->s_dentry_lru);
+       total_objects = dentries + inodes + fs_objects + 1;
 
        /* proportion the scan between the caches */
-       dentries = mult_frac(sc->nr_to_scan, sb->s_nr_dentry_unused,
-                                                               total_objects);
+       dentries = mult_frac(sc->nr_to_scan, dentries, total_objects);
        inodes = mult_frac(sc->nr_to_scan, inodes, total_objects);
 
        /*
@@ -117,7 +117,7 @@ static unsigned long super_cache_count(struct shrinker *shrink,
        if (sb->s_op && sb->s_op->nr_cached_objects)
                total_objects = sb->s_op->nr_cached_objects(sb);
 
-       total_objects += sb->s_nr_dentry_unused;
+       total_objects += list_lru_count(&sb->s_dentry_lru);
        total_objects += list_lru_count(&sb->s_inode_lru);
 
        total_objects = vfs_pressure_ratio(total_objects);
@@ -197,8 +197,7 @@ static struct super_block *alloc_super(struct file_system_type *type, int flags)
                INIT_HLIST_NODE(&s->s_instances);
                INIT_HLIST_BL_HEAD(&s->s_anon);
                INIT_LIST_HEAD(&s->s_inodes);
-               INIT_LIST_HEAD(&s->s_dentry_lru);
-               spin_lock_init(&s->s_dentry_lru_lock);
+               list_lru_init(&s->s_dentry_lru);
                list_lru_init(&s->s_inode_lru);
                INIT_LIST_HEAD(&s->s_mounts);
                init_rwsem(&s->s_umount);
index eab83e232644b5505bac5a213d44b8f8e6429972..27bc26cb6844e3cdb6f876fbfa7bb8a4063833df 100644 (file)
@@ -1270,14 +1270,6 @@ struct super_block {
        struct list_head        s_files;
 #endif
        struct list_head        s_mounts;       /* list of mounts; _not_ for fs use */
-
-       /* s_dentry_lru_lock protects s_dentry_lru and s_nr_dentry_unused */
-       spinlock_t              s_dentry_lru_lock ____cacheline_aligned_in_smp;
-       struct list_head        s_dentry_lru;   /* unused dentry lru */
-       long                    s_nr_dentry_unused;     /* # of dentry on lru */
-
-       struct list_lru         s_inode_lru ____cacheline_aligned_in_smp;
-
        struct block_device     *s_bdev;
        struct backing_dev_info *s_bdi;
        struct mtd_info         *s_mtd;
@@ -1328,6 +1320,13 @@ struct super_block {
 
        /* Being remounted read-only */
        int s_readonly_remount;
+
+       /*
+        * Keep the lru lists last in the structure so they always sit on their
+        * own individual cachelines.
+        */
+       struct list_lru         s_dentry_lru ____cacheline_aligned_in_smp;
+       struct list_lru         s_inode_lru ____cacheline_aligned_in_smp;
 };
 
 extern struct timespec current_fs_time(struct super_block *sb);