d_page            286 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c static void __ttm_dma_free_page(struct dma_pool *pool, struct dma_page *d_page)
d_page            289 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 	dma_addr_t dma = d_page->dma;
d_page            290 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 	d_page->vaddr &= ~VADDR_FLAG_HUGE_POOL;
d_page            294 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 	dma_free_attrs(pool->dev, pool->size, (void *)d_page->vaddr, dma, attrs);
d_page            296 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 	kfree(d_page);
d_page            297 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 	d_page = NULL;
d_page            301 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 	struct dma_page *d_page;
d_page            305 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 	d_page = kmalloc(sizeof(struct dma_page), GFP_KERNEL);
d_page            306 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 	if (!d_page)
d_page            312 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 	vaddr = dma_alloc_attrs(pool->dev, pool->size, &d_page->dma,
d_page            316 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 			d_page->p = vmalloc_to_page(vaddr);
d_page            318 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 			d_page->p = virt_to_page(vaddr);
d_page            319 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 		d_page->vaddr = (unsigned long)vaddr;
d_page            321 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 			d_page->vaddr |= VADDR_FLAG_HUGE_POOL;
d_page            323 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 		kfree(d_page);
d_page            324 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 		d_page = NULL;
d_page            326 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 	return d_page;
d_page            353 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c static void ttm_dma_page_put(struct dma_pool *pool, struct dma_page *d_page)
d_page            355 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 	struct page *page = d_page->p;
d_page            366 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 	list_del(&d_page->page_list);
d_page            367 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 	__ttm_dma_free_page(pool, d_page);
d_page            373 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 	struct dma_page *d_page, *tmp;
d_page            376 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 		list_for_each_entry_safe(d_page, tmp, d_pages, page_list)
d_page            377 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 			ttm_dma_page_put(pool, d_page);
d_page            388 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 	list_for_each_entry_safe(d_page, tmp, d_pages, page_list) {
d_page            389 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 		list_del(&d_page->page_list);
d_page            390 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 		__ttm_dma_free_page(pool, d_page);
d_page            671 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 	struct dma_page *d_page, *tmp;
d_page            679 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 	list_for_each_entry_safe(d_page, tmp, d_pages, page_list) {
d_page            680 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 		if (d_page->p != p)
d_page            683 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 		list_del(&d_page->page_list);
d_page            684 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 		__ttm_dma_free_page(pool, d_page);
d_page            812 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 			struct dma_page *d_page;
d_page            818 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 			list_for_each_entry(d_page, &d_pages, page_list) {
d_page            838 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 	struct dma_page *d_page = NULL;
d_page            846 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 		d_page = list_first_entry(&pool->free_list, struct dma_page, page_list);
d_page            847 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 		ttm->pages[index] = d_page->p;
d_page            848 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 		ttm_dma->dma_address[index] = d_page->dma;
d_page            849 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 		list_move_tail(&d_page->page_list, &ttm_dma->pages_list);
d_page            854 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 	return d_page;
d_page            893 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 	struct dma_page *d_page;
d_page            925 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 		d_page = ttm_dma_pool_get_pages(pool, ttm_dma, i);
d_page            926 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 		if (!d_page)
d_page            936 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 		d_page->vaddr |= VADDR_FLAG_UPDATED_COUNT;
d_page            960 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 		d_page = ttm_dma_pool_get_pages(pool, ttm_dma, i);
d_page            961 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 		if (!d_page) {
d_page            973 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 		d_page->vaddr |= VADDR_FLAG_UPDATED_COUNT;
d_page            997 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 	struct dma_page *d_page, *next;
d_page           1009 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 		list_for_each_entry_safe(d_page, next, &ttm_dma->pages_list,
d_page           1011 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 			if (!(d_page->vaddr & VADDR_FLAG_HUGE_POOL))
d_page           1015 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 			if (d_page->vaddr & VADDR_FLAG_UPDATED_COUNT) {
d_page           1016 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 				ttm_mem_global_free_page(mem_glob, d_page->p,
d_page           1018 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 				d_page->vaddr &= ~VADDR_FLAG_UPDATED_COUNT;
d_page           1020 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 			ttm_dma_page_put(pool, d_page);
d_page           1039 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 	list_for_each_entry_safe(d_page, next, &ttm_dma->pages_list,
d_page           1041 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 		ttm->pages[count] = d_page->p;
d_page           1044 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 		if (d_page->vaddr & VADDR_FLAG_UPDATED_COUNT) {
d_page           1045 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 			ttm_mem_global_free_page(mem_glob, d_page->p,
d_page           1047 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 			d_page->vaddr &= ~VADDR_FLAG_UPDATED_COUNT;
d_page           1051 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c 			ttm_dma_page_put(pool, d_page);
d_page           1374 kernel/power/snapshot.c 	struct page *s_page, *d_page;
d_page           1378 kernel/power/snapshot.c 	d_page = pfn_to_page(dst_pfn);
d_page           1381 kernel/power/snapshot.c 		dst = kmap_atomic(d_page);
d_page           1386 kernel/power/snapshot.c 		if (PageHighMem(d_page)) {
d_page           1392 kernel/power/snapshot.c 			dst = kmap_atomic(d_page);
d_page           1396 kernel/power/snapshot.c 			safe_copy_page(page_address(d_page), s_page);
d_page           1588 mm/zsmalloc.c  	struct page *s_page, *d_page;
d_page           1598 mm/zsmalloc.c  	obj_to_location(dst, &d_page, &d_objidx);
d_page           1610 mm/zsmalloc.c  	d_addr = kmap_atomic(d_page);
d_page           1630 mm/zsmalloc.c  			d_addr = kmap_atomic(d_page);
d_page           1637 mm/zsmalloc.c  			d_page = get_next_page(d_page);
d_page           1638 mm/zsmalloc.c  			d_addr = kmap_atomic(d_page);
d_page           1689 mm/zsmalloc.c  	struct page *d_page;
d_page           1701 mm/zsmalloc.c  	struct page *d_page = cc->d_page;
d_page           1716 mm/zsmalloc.c  		if (zspage_full(class, get_zspage(d_page))) {
d_page           1723 mm/zsmalloc.c  		free_obj = obj_malloc(class, get_zspage(d_page), handle);
d_page           2281 mm/zsmalloc.c  			cc.d_page = get_first_page(dst_zspage);