/linux-4.1.27/arch/arm/mm/ |
D | flush.c | 273 mapping = page_mapping(page); in __sync_icache_dcache() 315 mapping = page_mapping(page); in flush_dcache_page() 345 mapping = page_mapping(page); in flush_kernel_dcache_page()
|
D | copypage-xscale.c | 93 __flush_dcache_page(page_mapping(from), from); in xscale_mc_copy_user_highpage()
|
D | copypage-v4mc.c | 73 __flush_dcache_page(page_mapping(from), from); in v4_mc_copy_user_highpage()
|
D | copypage-v6.c | 79 __flush_dcache_page(page_mapping(from), from); in v6_copy_user_highpage_aliasing()
|
D | fault-armv.c | 198 mapping = page_mapping(page); in update_mmu_cache()
|
/linux-4.1.27/arch/score/mm/ |
D | cache.c | 57 struct address_space *mapping = page_mapping(page); in flush_dcache_page() 89 if (page_mapping(page) && test_bit(PG_dcache_dirty, &(page)->flags)) { in __update_cache()
|
/linux-4.1.27/drivers/net/ethernet/broadcom/bnx2x/ |
D | bnx2x_init_ops.h | 620 BNX2X_ILT_FREE(line->page, line->page_mapping, line->size); in bnx2x_ilt_line_mem_op() 623 BNX2X_ILT_ZALLOC(line->page, &line->page_mapping, size); in bnx2x_ilt_line_mem_op() 675 dma_addr_t page_mapping) in bnx2x_ilt_line_wr() argument 684 bnx2x_wr_64(bp, reg, ILT_ADDR1(page_mapping), ILT_ADDR2(page_mapping)); in bnx2x_ilt_line_wr() 698 bnx2x_ilt_line_wr(bp, abs_idx, ilt->lines[idx].page_mapping); in bnx2x_ilt_line_init_op()
|
D | bnx2x_init.h | 502 dma_addr_t page_mapping; member
|
D | bnx2x_sriov.c | 1698 ilt->lines[line+i].page_mapping = hw_cxt->mapping; in bnx2x_iov_init_ilt()
|
D | bnx2x_main.c | 7929 ilt->lines[cdu_ilt_start + i].page_mapping = in bnx2x_init_hw_func()
|
/linux-4.1.27/arch/unicore32/mm/ |
D | flush.c | 86 mapping = page_mapping(page); in flush_dcache_page()
|
D | mmu.c | 506 mapping = page_mapping(page); in update_mmu_cache()
|
/linux-4.1.27/arch/arm64/mm/ |
D | flush.c | 78 if (!page_mapping(page)) in __sync_icache_dcache()
|
/linux-4.1.27/arch/nios2/mm/ |
D | cacheflush.c | 199 mapping = page_mapping(page); in flush_dcache_page() 234 mapping = page_mapping(page); in update_mmu_cache()
|
/linux-4.1.27/mm/ |
D | memory-failure.c | 94 mapping = page_mapping(p); in hwpoison_filter_dev() 638 mapping = page_mapping(p); in me_pagecache_clean() 683 struct address_space *mapping = page_mapping(p); in me_pagecache_dirty() 789 if (!(page_mapping(hpage) || PageAnon(hpage))) { in me_huge_page() 965 mapping = page_mapping(hpage); in hwpoison_user_mappings()
|
D | page-writeback.c | 2147 struct address_space *mapping = page_mapping(page); in __set_page_dirty_nobuffers() 2154 BUG_ON(page_mapping(page) != mapping); in __set_page_dirty_nobuffers() 2217 struct address_space *mapping = page_mapping(page); in set_page_dirty() 2284 struct address_space *mapping = page_mapping(page); in clear_page_dirty_for_io() 2338 struct address_space *mapping = page_mapping(page); in test_clear_page_writeback() 2373 struct address_space *mapping = page_mapping(page); in __test_set_page_writeback()
|
D | vmscan.c | 483 if (page_mapping(page) == mapping) in handle_write_error() 583 BUG_ON(mapping != page_mapping(page)); in __remove_mapping() 835 mapping = page_mapping(page); in page_check_dirty_writeback() 918 mapping = page_mapping(page); in shrink_page_list() 1015 mapping = page_mapping(page); in shrink_page_list() 1083 mapping = page_mapping(page); in shrink_page_list() 1267 mapping = page_mapping(page); in __isolate_lru_page() 3769 return !mapping_unevictable(page_mapping(page)) && !PageMlocked(page); in page_evictable()
|
D | util.c | 356 struct address_space *page_mapping(struct page *page) in page_mapping() function
|
D | truncate.c | 186 struct address_space *mapping = page_mapping(page); in invalidate_inode_page()
|
D | compaction.c | 755 if (!page_mapping(page) && in isolate_migratepages_block()
|
D | rmap.c | 922 mapping = page_mapping(page); in page_mkclean()
|
D | migrate.c | 737 mapping = page_mapping(page); in move_to_new_page()
|
D | shmem.c | 969 swap_mapping = page_mapping(oldpage); in shmem_replace_page()
|
/linux-4.1.27/arch/sh/mm/ |
D | cache-sh7705.c | 139 struct address_space *mapping = page_mapping(page); in sh7705_flush_dcache_page()
|
D | cache-sh4.c | 115 struct address_space *mapping = page_mapping(page); in sh4_flush_dcache_page()
|
/linux-4.1.27/arch/sparc/kernel/ |
D | smp_64.c | 878 page_mapping(page) != NULL)); in __local_flush_dcache_page() 880 if (page_mapping(page) != NULL && in __local_flush_dcache_page() 907 if (page_mapping(page) != NULL) in smp_flush_dcache_page_impl() 943 if (page_mapping(page) != NULL) in flush_dcache_page_all()
|
/linux-4.1.27/arch/parisc/kernel/ |
D | cache.c | 90 if (page_mapping(page) && test_bit(PG_dcache_dirty, &page->flags)) { in update_mmu_cache() 288 struct address_space *mapping = page_mapping(page); in flush_dcache_page()
|
/linux-4.1.27/arch/mips/mm/ |
D | cache.c | 84 struct address_space *mapping = page_mapping(page); in __flush_dcache_page()
|
/linux-4.1.27/arch/xtensa/mm/ |
D | cache.c | 135 struct address_space *mapping = page_mapping(page); in flush_dcache_page()
|
/linux-4.1.27/arch/sparc/mm/ |
D | tlb.c | 121 mapping = page_mapping(page); in tlb_batch_add()
|
D | init_64.c | 206 page_mapping(page) != NULL)); in flush_dcache_page_impl() 208 if (page_mapping(page) != NULL && in flush_dcache_page_impl() 387 mapping = page_mapping(page); in flush_dcache_page()
|
/linux-4.1.27/arch/arc/mm/ |
D | cache_arc700.c | 483 mapping = page_mapping(page); in flush_dcache_page()
|
/linux-4.1.27/fs/ |
D | buffer.c | 671 struct address_space *mapping = page_mapping(page); in __set_page_dirty_buffers() 1162 struct address_space *mapping = page_mapping(page); in mark_buffer_dirty()
|
D | splice.c | 51 mapping = page_mapping(page); in page_cache_pipe_buf_steal()
|
/linux-4.1.27/Documentation/vm/ |
D | unevictable-lru.txt | 331 populate_vma_page_range() checks page_mapping() after acquiring the page lock. 443 zeros out the page_mapping of the old page before unlocking it, so m[un]lock
|
/linux-4.1.27/include/linux/ |
D | mm.h | 958 extern struct address_space *page_mapping(struct page *page);
|