]> git.kernelconcepts.de Git - karo-tx-linux.git/commitdiff
staging: zsmalloc: rename zspage_order with zspage_pages
authorMinchan Kim <minchan@kernel.org>
Thu, 3 May 2012 06:40:39 +0000 (15:40 +0900)
committerGreg Kroah-Hartman <gregkh@linuxfoundation.org>
Wed, 9 May 2012 20:20:22 +0000 (13:20 -0700)
zspage_order defines how many pages are needed to make a zspage.
So _order_ is rather awkward naming. It already deceive Jonathan
- http://lwn.net/Articles/477067/
" For each size, the code calculates an optimum number of pages (up to 16)"

Let's change from _order_ to _pages_ and some function names.

Signed-off-by: Minchan Kim <minchan@kernel.org>
Acked-by: Nitin Gupta <ngupta@vflare.org>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
drivers/staging/zsmalloc/zsmalloc-main.c
drivers/staging/zsmalloc/zsmalloc_int.h

index 504b6c2db22e94adf0fa529ff59e5cd7d77f8081..86428000f88a85901667bf20f9d167a7f8b75f72 100644 (file)
@@ -180,7 +180,7 @@ out:
  * link together 3 PAGE_SIZE sized pages to form a zspage
  * since then we can perfectly fit in 8 such objects.
  */
-static int get_zspage_order(int class_size)
+static int get_pages_per_zspage(int class_size)
 {
        int i, max_usedpc = 0;
        /* zspage order which gives maximum used size per KB */
@@ -368,7 +368,7 @@ static struct page *alloc_zspage(struct size_class *class, gfp_t flags)
         * identify the last page.
         */
        error = -ENOMEM;
-       for (i = 0; i < class->zspage_order; i++) {
+       for (i = 0; i < class->pages_per_zspage; i++) {
                struct page *page, *prev_page;
 
                page = alloc_page(flags);
@@ -388,7 +388,7 @@ static struct page *alloc_zspage(struct size_class *class, gfp_t flags)
                        page->first_page = first_page;
                if (i >= 2)
                        list_add(&page->lru, &prev_page->lru);
-               if (i == class->zspage_order - 1)       /* last page */
+               if (i == class->pages_per_zspage - 1)   /* last page */
                        SetPagePrivate2(page);
                prev_page = page;
        }
@@ -397,7 +397,7 @@ static struct page *alloc_zspage(struct size_class *class, gfp_t flags)
 
        first_page->freelist = obj_location_to_handle(first_page, 0);
        /* Maximum number of objects we can store in this zspage */
-       first_page->objects = class->zspage_order * PAGE_SIZE / class->size;
+       first_page->objects = class->pages_per_zspage * PAGE_SIZE / class->size;
 
        error = 0; /* Success */
 
@@ -512,7 +512,7 @@ struct zs_pool *zs_create_pool(const char *name, gfp_t flags)
                class->size = size;
                class->index = i;
                spin_lock_init(&class->lock);
-               class->zspage_order = get_zspage_order(size);
+               class->pages_per_zspage = get_pages_per_zspage(size);
 
        }
 
@@ -603,7 +603,7 @@ void *zs_malloc(struct zs_pool *pool, size_t size)
 
                set_zspage_mapping(first_page, class->index, ZS_EMPTY);
                spin_lock(&class->lock);
-               class->pages_allocated += class->zspage_order;
+               class->pages_allocated += class->pages_per_zspage;
        }
 
        obj = first_page->freelist;
@@ -658,7 +658,7 @@ void zs_free(struct zs_pool *pool, void *obj)
        fullness = fix_fullness_group(pool, first_page);
 
        if (fullness == ZS_EMPTY)
-               class->pages_allocated -= class->zspage_order;
+               class->pages_allocated -= class->pages_per_zspage;
 
        spin_unlock(&class->lock);
 
index 92eefc663afc3336f73a8114f9c5543301bd1819..6fd32a9e03152f484fb07160e0724d4b18d0700c 100644 (file)
@@ -124,7 +124,7 @@ struct size_class {
        unsigned int index;
 
        /* Number of PAGE_SIZE sized pages to combine to form a 'zspage' */
-       int zspage_order;
+       int pages_per_zspage;
 
        spinlock_t lock;