mm: fix cache mode tracking in vm_insert_mixed()
[cascardo/linux.git] / mm / vmalloc.c
index e11475c..80660a0 100644 (file)
@@ -1359,14 +1359,14 @@ static struct vm_struct *__get_vm_area_node(unsigned long size,
        struct vm_struct *area;
 
        BUG_ON(in_interrupt());
-       if (flags & VM_IOREMAP)
-               align = 1ul << clamp_t(int, fls_long(size),
-                                      PAGE_SHIFT, IOREMAP_MAX_ORDER);
-
        size = PAGE_ALIGN(size);
        if (unlikely(!size))
                return NULL;
 
+       if (flags & VM_IOREMAP)
+               align = 1ul << clamp_t(int, get_count_order_long(size),
+                                      PAGE_SHIFT, IOREMAP_MAX_ORDER);
+
        area = kzalloc_node(sizeof(*area), gfp_mask & GFP_RECLAIM_MASK, node);
        if (unlikely(!area))
                return NULL;
@@ -1501,7 +1501,7 @@ static void __vunmap(const void *addr, int deallocate_pages)
                        struct page *page = area->pages[i];
 
                        BUG_ON(!page);
-                       __free_kmem_pages(page, 0);
+                       __free_pages(page, 0);
                }
 
                kvfree(area->pages);
@@ -1629,9 +1629,9 @@ static void *__vmalloc_area_node(struct vm_struct *area, gfp_t gfp_mask,
                struct page *page;
 
                if (node == NUMA_NO_NODE)
-                       page = alloc_kmem_pages(alloc_mask, order);
+                       page = alloc_pages(alloc_mask, order);
                else
-                       page = alloc_kmem_pages_node(node, alloc_mask, order);
+                       page = alloc_pages_node(node, alloc_mask, order);
 
                if (unlikely(!page)) {
                        /* Successfully allocated i pages, free them in __vunmap() */