drm/ttm: add transparent huge page support for cached allocations v2
authorChristian König <christian.koenig@amd.com>
Wed, 20 Sep 2017 13:06:12 +0000 (15:06 +0200)
committerAlex Deucher <alexander.deucher@amd.com>
Thu, 19 Oct 2017 19:27:01 +0000 (15:27 -0400)
Try to allocate huge pages when it makes sense.

v2: avoid compound pages for now

Signed-off-by: Christian König <christian.koenig@amd.com>
Acked-by: Felix Kuehling <Felix.Kuehling@amd.com>
Acked-by: Alex Deucher <alexander.deucher@amd.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
drivers/gpu/drm/ttm/ttm_page_alloc.c

index 6c852e81660b4dc9b49580b06ab67c2bfd7350b4..1bc6053b458176f18e38793b50da2cb576545e72 100644 (file)
@@ -685,12 +685,24 @@ static void ttm_put_pages(struct page **pages, unsigned npages, int flags,
 
        if (pool == NULL) {
                /* No pool for this memory type so free the pages */
-               for (i = 0; i < npages; i++) {
-                       if (pages[i]) {
-                               if (page_count(pages[i]) != 1)
-                                       pr_err("Erroneous page count. Leaking pages.\n");
-                               __free_page(pages[i]);
-                               pages[i] = NULL;
+               i = 0;
+               while (i < npages) {
+                       unsigned order;
+
+                       if (!pages[i]) {
+                               ++i;
+                               continue;
+                       }
+
+                       if (page_count(pages[i]) != 1)
+                               pr_err("Erroneous page count. Leaking pages.\n");
+                       order = compound_order(pages[i]);
+                       __free_pages(pages[i], order);
+
+                       order = 1 << order;
+                       while (order) {
+                               pages[i++] = NULL;
+                               --order;
                        }
                }
                return;
@@ -740,12 +752,33 @@ static int ttm_get_pages(struct page **pages, unsigned npages, int flags,
 
        /* No pool for cached pages */
        if (pool == NULL) {
+               unsigned i, j;
+
                if (flags & TTM_PAGE_FLAG_DMA32)
                        gfp_flags |= GFP_DMA32;
                else
                        gfp_flags |= GFP_HIGHUSER;
 
-               for (r = 0; r < npages; ++r) {
+               i = 0;
+#ifdef CONFIG_TRANSPARENT_HUGEPAGE
+               while (npages >= HPAGE_PMD_NR) {
+                       gfp_t huge_flags = gfp_flags;
+
+                       huge_flags |= GFP_TRANSHUGE;
+                       huge_flags &= ~__GFP_MOVABLE;
+                       huge_flags &= ~__GFP_COMP;
+                       p = alloc_pages(huge_flags, HPAGE_PMD_ORDER);
+                       if (!p)
+                               break;
+
+                       for (j = 0; j < HPAGE_PMD_NR; ++j)
+                               pages[i++] = p++;
+
+                       npages -= HPAGE_PMD_NR;
+               }
+#endif
+
+               while (npages) {
                        p = alloc_page(gfp_flags);
                        if (!p) {
 
@@ -753,7 +786,8 @@ static int ttm_get_pages(struct page **pages, unsigned npages, int flags,
                                return -ENOMEM;
                        }
 
-                       pages[r] = p;
+                       pages[i++] = p;
+                       --npages;
                }
                return 0;
        }