d_page 286 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c static void __ttm_dma_free_page(struct dma_pool *pool, struct dma_page *d_page) d_page 289 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c dma_addr_t dma = d_page->dma; d_page 290 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c d_page->vaddr &= ~VADDR_FLAG_HUGE_POOL; d_page 294 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c dma_free_attrs(pool->dev, pool->size, (void *)d_page->vaddr, dma, attrs); d_page 296 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c kfree(d_page); d_page 297 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c d_page = NULL; d_page 301 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c struct dma_page *d_page; d_page 305 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c d_page = kmalloc(sizeof(struct dma_page), GFP_KERNEL); d_page 306 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c if (!d_page) d_page 312 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c vaddr = dma_alloc_attrs(pool->dev, pool->size, &d_page->dma, d_page 316 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c d_page->p = vmalloc_to_page(vaddr); d_page 318 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c d_page->p = virt_to_page(vaddr); d_page 319 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c d_page->vaddr = (unsigned long)vaddr; d_page 321 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c d_page->vaddr |= VADDR_FLAG_HUGE_POOL; d_page 323 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c kfree(d_page); d_page 324 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c d_page = NULL; d_page 326 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c return d_page; d_page 353 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c static void ttm_dma_page_put(struct dma_pool *pool, struct dma_page *d_page) d_page 355 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c struct page *page = d_page->p; d_page 366 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c list_del(&d_page->page_list); d_page 367 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c __ttm_dma_free_page(pool, d_page); d_page 373 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c struct dma_page *d_page, *tmp; d_page 376 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c list_for_each_entry_safe(d_page, tmp, d_pages, page_list) d_page 377 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c ttm_dma_page_put(pool, d_page); d_page 388 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c list_for_each_entry_safe(d_page, tmp, d_pages, page_list) { d_page 389 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c list_del(&d_page->page_list); d_page 390 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c __ttm_dma_free_page(pool, d_page); d_page 671 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c struct dma_page *d_page, *tmp; d_page 679 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c list_for_each_entry_safe(d_page, tmp, d_pages, page_list) { d_page 680 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c if (d_page->p != p) d_page 683 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c list_del(&d_page->page_list); d_page 684 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c __ttm_dma_free_page(pool, d_page); d_page 812 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c struct dma_page *d_page; d_page 818 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c list_for_each_entry(d_page, &d_pages, page_list) { d_page 838 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c struct dma_page *d_page = NULL; d_page 846 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c d_page = list_first_entry(&pool->free_list, struct dma_page, page_list); d_page 847 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c ttm->pages[index] = d_page->p; d_page 848 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c ttm_dma->dma_address[index] = d_page->dma; d_page 849 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c list_move_tail(&d_page->page_list, &ttm_dma->pages_list); d_page 854 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c return d_page; d_page 893 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c struct dma_page *d_page; d_page 925 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c d_page = ttm_dma_pool_get_pages(pool, ttm_dma, i); d_page 926 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c if (!d_page) d_page 936 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c d_page->vaddr |= VADDR_FLAG_UPDATED_COUNT; d_page 960 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c d_page = ttm_dma_pool_get_pages(pool, ttm_dma, i); d_page 961 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c if (!d_page) { d_page 973 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c d_page->vaddr |= VADDR_FLAG_UPDATED_COUNT; d_page 997 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c struct dma_page *d_page, *next; d_page 1009 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c list_for_each_entry_safe(d_page, next, &ttm_dma->pages_list, d_page 1011 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c if (!(d_page->vaddr & VADDR_FLAG_HUGE_POOL)) d_page 1015 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c if (d_page->vaddr & VADDR_FLAG_UPDATED_COUNT) { d_page 1016 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c ttm_mem_global_free_page(mem_glob, d_page->p, d_page 1018 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c d_page->vaddr &= ~VADDR_FLAG_UPDATED_COUNT; d_page 1020 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c ttm_dma_page_put(pool, d_page); d_page 1039 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c list_for_each_entry_safe(d_page, next, &ttm_dma->pages_list, d_page 1041 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c ttm->pages[count] = d_page->p; d_page 1044 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c if (d_page->vaddr & VADDR_FLAG_UPDATED_COUNT) { d_page 1045 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c ttm_mem_global_free_page(mem_glob, d_page->p, d_page 1047 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c d_page->vaddr &= ~VADDR_FLAG_UPDATED_COUNT; d_page 1051 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c ttm_dma_page_put(pool, d_page); d_page 1374 kernel/power/snapshot.c struct page *s_page, *d_page; d_page 1378 kernel/power/snapshot.c d_page = pfn_to_page(dst_pfn); d_page 1381 kernel/power/snapshot.c dst = kmap_atomic(d_page); d_page 1386 kernel/power/snapshot.c if (PageHighMem(d_page)) { d_page 1392 kernel/power/snapshot.c dst = kmap_atomic(d_page); d_page 1396 kernel/power/snapshot.c safe_copy_page(page_address(d_page), s_page); d_page 1588 mm/zsmalloc.c struct page *s_page, *d_page; d_page 1598 mm/zsmalloc.c obj_to_location(dst, &d_page, &d_objidx); d_page 1610 mm/zsmalloc.c d_addr = kmap_atomic(d_page); d_page 1630 mm/zsmalloc.c d_addr = kmap_atomic(d_page); d_page 1637 mm/zsmalloc.c d_page = get_next_page(d_page); d_page 1638 mm/zsmalloc.c d_addr = kmap_atomic(d_page); d_page 1689 mm/zsmalloc.c struct page *d_page; d_page 1701 mm/zsmalloc.c struct page *d_page = cc->d_page; d_page 1716 mm/zsmalloc.c if (zspage_full(class, get_zspage(d_page))) { d_page 1723 mm/zsmalloc.c free_obj = obj_malloc(class, get_zspage(d_page), handle); d_page 2281 mm/zsmalloc.c cc.d_page = get_first_page(dst_zspage);