| /linux-4.4.14/arch/arm/mm/ | 
| D | flush.c | 288 		mapping = page_mapping(page);  in __sync_icache_dcache() 330 	mapping = page_mapping(page);  in flush_dcache_page() 360 		mapping = page_mapping(page);  in flush_kernel_dcache_page()
  | 
| D | copypage-xscale.c | 93 		__flush_dcache_page(page_mapping(from), from);  in xscale_mc_copy_user_highpage()
  | 
| D | copypage-v4mc.c | 73 		__flush_dcache_page(page_mapping(from), from);  in v4_mc_copy_user_highpage()
  | 
| D | copypage-v6.c | 79 		__flush_dcache_page(page_mapping(from), from);  in v6_copy_user_highpage_aliasing()
  | 
| D | fault-armv.c | 198 	mapping = page_mapping(page);  in update_mmu_cache()
  | 
| /linux-4.4.14/arch/score/mm/ | 
| D | cache.c | 57 	struct address_space *mapping = page_mapping(page);  in flush_dcache_page() 89 	if (page_mapping(page) && test_bit(PG_dcache_dirty, &(page)->flags)) {  in __update_cache()
  | 
| /linux-4.4.14/drivers/net/ethernet/broadcom/bnx2x/ | 
| D | bnx2x_init_ops.h | 622 		BNX2X_ILT_FREE(line->page, line->page_mapping, line->size);  in bnx2x_ilt_line_mem_op() 625 	BNX2X_ILT_ZALLOC(line->page, &line->page_mapping, size);  in bnx2x_ilt_line_mem_op() 677 			      dma_addr_t page_mapping)  in bnx2x_ilt_line_wr()  argument 686 	bnx2x_wr_64(bp, reg, ILT_ADDR1(page_mapping), ILT_ADDR2(page_mapping));  in bnx2x_ilt_line_wr() 700 		bnx2x_ilt_line_wr(bp, abs_idx, ilt->lines[idx].page_mapping);  in bnx2x_ilt_line_init_op()
  | 
| D | bnx2x_init.h | 504 	dma_addr_t page_mapping;  member
  | 
| D | bnx2x_sriov.c | 1656 		ilt->lines[line+i].page_mapping = hw_cxt->mapping;  in bnx2x_iov_init_ilt()
  | 
| D | bnx2x_main.c | 7975 		ilt->lines[cdu_ilt_start + i].page_mapping =  in bnx2x_init_hw_func()
  | 
| /linux-4.4.14/arch/unicore32/mm/ | 
| D | flush.c | 86 	mapping = page_mapping(page);  in flush_dcache_page()
  | 
| D | mmu.c | 506 	mapping = page_mapping(page);  in update_mmu_cache()
  | 
| /linux-4.4.14/arch/nios2/mm/ | 
| D | cacheflush.c | 183 	mapping = page_mapping(page);  in flush_dcache_page() 218 	mapping = page_mapping(page);  in update_mmu_cache()
  | 
| /linux-4.4.14/arch/arm64/mm/ | 
| D | flush.c | 74 	if (!page_mapping(page))  in __sync_icache_dcache()
  | 
| /linux-4.4.14/mm/ | 
| D | memory-failure.c | 97 	mapping = page_mapping(p);  in hwpoison_filter_dev() 595 	mapping = page_mapping(p);  in me_pagecache_clean() 640 	struct address_space *mapping = page_mapping(p);  in me_pagecache_dirty() 750 	if (!(page_mapping(hpage) || PageAnon(hpage))) {  in me_huge_page() 987 	mapping = page_mapping(hpage);  in hwpoison_user_mappings() 1451 	if (page_mapping(page)) {  in unpoison_memory()
  | 
| D | page-writeback.c | 2467 		struct address_space *mapping = page_mapping(page);  in __set_page_dirty_nobuffers() 2476 		BUG_ON(page_mapping(page) != mapping);  in __set_page_dirty_nobuffers() 2549 	struct address_space *mapping = page_mapping(page);  in set_page_dirty() 2615 	struct address_space *mapping = page_mapping(page);  in cancel_dirty_page() 2653 	struct address_space *mapping = page_mapping(page);  in clear_page_dirty_for_io() 2717 	struct address_space *mapping = page_mapping(page);  in test_clear_page_writeback() 2755 	struct address_space *mapping = page_mapping(page);  in __test_set_page_writeback()
  | 
| D | vmscan.c | 513 	if (page_mapping(page) == mapping)  in handle_write_error() 616 	BUG_ON(mapping != page_mapping(page));  in __remove_mapping() 872 	mapping = page_mapping(page);  in page_check_dirty_writeback() 955 		mapping = page_mapping(page);  in shrink_page_list() 1055 			mapping = page_mapping(page);  in shrink_page_list() 1129 				mapping = page_mapping(page);  in shrink_page_list() 1314 			mapping = page_mapping(page);  in __isolate_lru_page() 3834 	return !mapping_unevictable(page_mapping(page)) && !PageMlocked(page);  in page_evictable()
  | 
| D | util.c | 356 struct address_space *page_mapping(struct page *page)  in page_mapping()  function
  | 
| D | truncate.c | 184 	struct address_space *mapping = page_mapping(page);  in invalidate_inode_page()
  | 
| D | compaction.c | 776 		if (!page_mapping(page) &&  in isolate_migratepages_block()
  | 
| D | rmap.c | 1033 	mapping = page_mapping(page);  in page_mkclean()
  | 
| D | migrate.c | 756 	mapping = page_mapping(page);  in move_to_new_page()
  | 
| D | shmem.c | 991 	swap_mapping = page_mapping(oldpage);  in shmem_replace_page()
  | 
| D | memcontrol.c | 4595 		struct address_space *mapping = page_mapping(page);  in mem_cgroup_move_account()
  | 
| /linux-4.4.14/arch/sh/mm/ | 
| D | cache-sh7705.c | 139 	struct address_space *mapping = page_mapping(page);  in sh7705_flush_dcache_page()
  | 
| D | cache-sh4.c | 115 	struct address_space *mapping = page_mapping(page);  in sh4_flush_dcache_page()
  | 
| /linux-4.4.14/arch/sparc/kernel/ | 
| D | smp_64.c | 878 			     page_mapping(page) != NULL));  in __local_flush_dcache_page() 880 	if (page_mapping(page) != NULL &&  in __local_flush_dcache_page() 907 			if (page_mapping(page) != NULL)  in smp_flush_dcache_page_impl() 943 		if (page_mapping(page) != NULL)  in flush_dcache_page_all()
  | 
| /linux-4.4.14/arch/parisc/kernel/ | 
| D | cache.c | 90 	if (page_mapping(page) && test_bit(PG_dcache_dirty, &page->flags)) {  in update_mmu_cache() 288 	struct address_space *mapping = page_mapping(page);  in flush_dcache_page()
  | 
| /linux-4.4.14/arch/mips/mm/ | 
| D | cache.c | 84 	struct address_space *mapping = page_mapping(page);  in __flush_dcache_page()
  | 
| /linux-4.4.14/arch/sparc/mm/ | 
| D | tlb.c | 131 		mapping = page_mapping(page);  in tlb_batch_add()
  | 
| D | init_64.c | 208 			     page_mapping(page) != NULL));  in flush_dcache_page_impl() 210 	if (page_mapping(page) != NULL &&  in flush_dcache_page_impl() 377 	mapping = page_mapping(page);  in flush_dcache_page()
  | 
| /linux-4.4.14/arch/xtensa/mm/ | 
| D | cache.c | 135 	struct address_space *mapping = page_mapping(page);  in flush_dcache_page()
  | 
| /linux-4.4.14/arch/arc/mm/ | 
| D | cache.c | 610 	mapping = page_mapping(page);  in flush_dcache_page()
  | 
| /linux-4.4.14/Documentation/vm/ | 
| D | unevictable-lru.txt | 331 populate_vma_page_range() checks page_mapping() after acquiring the page lock. 443 zeros out the page_mapping of the old page before unlocking it, so m[un]lock
  | 
| /linux-4.4.14/fs/ | 
| D | buffer.c | 677 	struct address_space *mapping = page_mapping(page);  in __set_page_dirty_buffers() 1183 			mapping = page_mapping(page);  in mark_buffer_dirty()
  | 
| D | splice.c | 51 	mapping = page_mapping(page);  in page_cache_pipe_buf_steal()
  | 
| /linux-4.4.14/include/linux/ | 
| D | mm.h | 941 extern struct address_space *page_mapping(struct page *page);
  |