/linux-4.4.14/mm/ |
D | swap_state.c | 295 struct page *found_page, *new_page = NULL; in __read_swap_cache_async() local 313 if (!new_page) { in __read_swap_cache_async() 314 new_page = alloc_page_vma(gfp_mask, vma, addr); in __read_swap_cache_async() 315 if (!new_page) in __read_swap_cache_async() 356 __set_page_locked(new_page); in __read_swap_cache_async() 357 SetPageSwapBacked(new_page); in __read_swap_cache_async() 358 err = __add_to_swap_cache(new_page, entry); in __read_swap_cache_async() 364 lru_cache_add_anon(new_page); in __read_swap_cache_async() 366 return new_page; in __read_swap_cache_async() 369 ClearPageSwapBacked(new_page); in __read_swap_cache_async() [all …]
|
D | migrate.c | 1746 struct page *new_page = NULL; in migrate_misplaced_transhuge_page() local 1760 new_page = alloc_pages_node(node, in migrate_misplaced_transhuge_page() 1763 if (!new_page) in migrate_misplaced_transhuge_page() 1768 put_page(new_page); in migrate_misplaced_transhuge_page() 1776 __set_page_locked(new_page); in migrate_misplaced_transhuge_page() 1777 SetPageSwapBacked(new_page); in migrate_misplaced_transhuge_page() 1780 new_page->mapping = page->mapping; in migrate_misplaced_transhuge_page() 1781 new_page->index = page->index; in migrate_misplaced_transhuge_page() 1782 migrate_page_copy(new_page, page); in migrate_misplaced_transhuge_page() 1783 WARN_ON(PageLRU(new_page)); in migrate_misplaced_transhuge_page() [all …]
|
D | huge_memory.c | 1156 struct page *page = NULL, *new_page; in do_huge_pmd_wp_page() local 1189 new_page = alloc_hugepage_vma(huge_gfp, vma, haddr, HPAGE_PMD_ORDER); in do_huge_pmd_wp_page() 1191 new_page = NULL; in do_huge_pmd_wp_page() 1193 if (unlikely(!new_page)) { in do_huge_pmd_wp_page() 1210 if (unlikely(mem_cgroup_try_charge(new_page, mm, huge_gfp, &memcg))) { in do_huge_pmd_wp_page() 1211 put_page(new_page); in do_huge_pmd_wp_page() 1225 clear_huge_page(new_page, haddr, HPAGE_PMD_NR); in do_huge_pmd_wp_page() 1227 copy_user_huge_page(new_page, page, haddr, vma, HPAGE_PMD_NR); in do_huge_pmd_wp_page() 1228 __SetPageUptodate(new_page); in do_huge_pmd_wp_page() 1239 mem_cgroup_cancel_charge(new_page, memcg); in do_huge_pmd_wp_page() [all …]
|
D | memory.c | 2104 struct page *new_page = NULL; in wp_page_copy() local 2116 new_page = alloc_zeroed_user_highpage_movable(vma, address); in wp_page_copy() 2117 if (!new_page) in wp_page_copy() 2120 new_page = alloc_page_vma(GFP_HIGHUSER_MOVABLE, vma, address); in wp_page_copy() 2121 if (!new_page) in wp_page_copy() 2123 cow_user_page(new_page, old_page, address, vma); in wp_page_copy() 2126 if (mem_cgroup_try_charge(new_page, mm, GFP_KERNEL, &memcg)) in wp_page_copy() 2129 __SetPageUptodate(new_page); in wp_page_copy() 2147 entry = mk_pte(new_page, vma->vm_page_prot); in wp_page_copy() 2156 page_add_new_anon_rmap(new_page, vma, address); in wp_page_copy() [all …]
|
D | ksm.c | 1885 struct page *new_page; in ksm_might_need_to_copy() local 1900 new_page = alloc_page_vma(GFP_HIGHUSER_MOVABLE, vma, address); in ksm_might_need_to_copy() 1901 if (new_page) { in ksm_might_need_to_copy() 1902 copy_user_highpage(new_page, page, address, vma); in ksm_might_need_to_copy() 1904 SetPageDirty(new_page); in ksm_might_need_to_copy() 1905 __SetPageUptodate(new_page); in ksm_might_need_to_copy() 1906 __set_page_locked(new_page); in ksm_might_need_to_copy() 1909 return new_page; in ksm_might_need_to_copy()
|
D | hugetlb.c | 3355 struct page *old_page, *new_page; in hugetlb_cow() local 3391 new_page = alloc_huge_page(vma, address, outside_reserve); in hugetlb_cow() 3393 if (IS_ERR(new_page)) { in hugetlb_cow() 3418 ret = (PTR_ERR(new_page) == -ENOMEM) ? in hugetlb_cow() 3432 copy_user_huge_page(new_page, old_page, address, vma, in hugetlb_cow() 3434 __SetPageUptodate(new_page); in hugetlb_cow() 3435 set_page_huge_active(new_page); in hugetlb_cow() 3448 ClearPagePrivate(new_page); in hugetlb_cow() 3454 make_huge_pte(vma, new_page, 1)); in hugetlb_cow() 3456 hugepage_add_new_anon_rmap(new_page, vma, address); in hugetlb_cow() [all …]
|
D | memory-failure.c | 1514 static struct page *new_page(struct page *p, unsigned long private, int **x) in new_page() function 1617 ret = migrate_pages(&pagelist, new_page, NULL, MPOL_MF_MOVE_ALL, in soft_offline_huge_page() 1697 ret = migrate_pages(&pagelist, new_page, NULL, MPOL_MF_MOVE_ALL, in __soft_offline_page()
|
D | mempolicy.c | 1094 static struct page *new_page(struct page *page, unsigned long start, int **x) in new_page() function 1129 static struct page *new_page(struct page *page, unsigned long start, int **x) in new_page() function 1214 nr_failed = migrate_pages(&pagelist, new_page, NULL, in do_mbind()
|
/linux-4.4.14/arch/s390/mm/ |
D | vmem.c | 228 void *new_page; in vmemmap_populate() local 230 new_page = vmemmap_alloc_block(PMD_SIZE, node); in vmemmap_populate() 231 if (!new_page) in vmemmap_populate() 233 pmd_val(*pm_dir) = __pa(new_page) | in vmemmap_populate() 249 void *new_page; in vmemmap_populate() local 251 new_page = vmemmap_alloc_block(PAGE_SIZE, node); in vmemmap_populate() 252 if (!new_page) in vmemmap_populate() 255 __pa(new_page) | pgprot_val(PAGE_KERNEL); in vmemmap_populate()
|
/linux-4.4.14/fs/f2fs/ |
D | namei.c | 594 struct page *old_page, *new_page = NULL; in f2fs_rename() local 634 &new_page); in f2fs_rename() 650 f2fs_set_link(new_dir, new_entry, new_page, old_inode); in f2fs_rename() 726 if (new_page) { in f2fs_rename() 727 f2fs_dentry_kunmap(new_dir, new_page); in f2fs_rename() 728 f2fs_put_page(new_page, 0); in f2fs_rename() 752 struct page *old_page, *new_page; in f2fs_cross_rename() local 772 new_entry = f2fs_find_entry(new_dir, &new_dentry->d_name, &new_page); in f2fs_cross_rename() 854 f2fs_set_link(new_dir, new_entry, new_page, old_inode); in f2fs_cross_rename() 898 f2fs_dentry_kunmap(new_dir, new_page); in f2fs_cross_rename() [all …]
|
/linux-4.4.14/fs/minix/ |
D | namei.c | 210 struct page * new_page; in minix_rename() local 218 new_de = minix_find_entry(new_dentry, &new_page); in minix_rename() 221 minix_set_link(new_de, new_page, old_inode); in minix_rename()
|
/linux-4.4.14/fs/sysv/ |
D | namei.c | 231 struct page * new_page; in sysv_rename() local 239 new_de = sysv_find_entry(new_dentry, &new_page); in sysv_rename() 242 sysv_set_link(new_de, new_page, old_inode); in sysv_rename()
|
/linux-4.4.14/fs/exofs/ |
D | namei.c | 251 struct page *new_page; in exofs_rename() local 259 new_de = exofs_find_entry(new_dir, new_dentry, &new_page); in exofs_rename() 262 err = exofs_set_link(new_dir, new_de, new_page, old_inode); in exofs_rename()
|
/linux-4.4.14/fs/ufs/ |
D | namei.c | 269 struct page *new_page; in ufs_rename() local 277 new_de = ufs_find_entry(new_dir, &new_dentry->d_name, &new_page); in ufs_rename() 280 ufs_set_link(new_dir, new_de, new_page, old_inode, 1); in ufs_rename()
|
/linux-4.4.14/fs/ubifs/ |
D | budget.c | 379 znodes = req->new_ino + (req->new_page << UBIFS_BLOCKS_PER_PAGE_SHIFT) + in calc_idx_growth() 396 if (req->new_page) in calc_data_growth() 442 ubifs_assert(req->new_page <= 1); in ubifs_budget_space() 529 ubifs_assert(req->new_page <= 1); in ubifs_release_budget()
|
D | file.c | 201 struct ubifs_budget_req req = { .recalculate = 1, .new_page = 1 }; in release_new_page_budget() 227 struct ubifs_budget_req req = { .new_page = 1 }; in write_begin_slow() 368 req.new_page = 1; in allocate_budget() 1481 struct ubifs_budget_req req = { .new_page = 1 }; in ubifs_vm_page_mkwrite()
|
D | ubifs.h | 886 unsigned int new_page:1; member 896 unsigned int new_page; member
|
D | debug.c | 556 req->new_page, req->dirtied_page); in ubifs_dump_budget_req()
|
/linux-4.4.14/fs/ext2/ |
D | namei.c | 362 struct page *new_page; in ext2_rename() local 370 new_de = ext2_find_entry (new_dir, &new_dentry->d_name, &new_page); in ext2_rename() 373 ext2_set_link(new_dir, new_de, new_page, old_inode, 1); in ext2_rename()
|
/linux-4.4.14/kernel/events/ |
D | uprobes.c | 295 struct page *old_page, *new_page; in uprobe_write_opcode() local 314 new_page = alloc_page_vma(GFP_HIGHUSER_MOVABLE, vma, vaddr); in uprobe_write_opcode() 315 if (!new_page) in uprobe_write_opcode() 318 __SetPageUptodate(new_page); in uprobe_write_opcode() 319 copy_highpage(new_page, old_page); in uprobe_write_opcode() 320 copy_to_page(new_page, vaddr, &opcode, UPROBE_SWBP_INSN_SIZE); in uprobe_write_opcode() 322 ret = __replace_page(vma, vaddr, old_page, new_page); in uprobe_write_opcode() 323 page_cache_release(new_page); in uprobe_write_opcode()
|
/linux-4.4.14/fs/nilfs2/ |
D | namei.c | 384 struct page *new_page; in nilfs_rename() local 392 new_de = nilfs_find_entry(new_dir, &new_dentry->d_name, &new_page); in nilfs_rename() 395 nilfs_set_link(new_dir, new_de, new_page, old_inode); in nilfs_rename()
|
/linux-4.4.14/fs/jbd2/ |
D | journal.c | 358 struct page *new_page; in jbd2_journal_write_metadata_buffer() local 387 new_page = virt_to_page(jh_in->b_frozen_data); in jbd2_journal_write_metadata_buffer() 390 new_page = jh2bh(jh_in)->b_page; in jbd2_journal_write_metadata_buffer() 394 mapped_data = kmap_atomic(new_page); in jbd2_journal_write_metadata_buffer() 434 mapped_data = kmap_atomic(new_page); in jbd2_journal_write_metadata_buffer() 438 new_page = virt_to_page(tmp); in jbd2_journal_write_metadata_buffer() 455 mapped_data = kmap_atomic(new_page); in jbd2_journal_write_metadata_buffer() 460 set_bh_page(new_bh, new_page, new_offset); in jbd2_journal_write_metadata_buffer()
|
/linux-4.4.14/fs/ |
D | dax.c | 423 struct page *new_page = vmf->cow_page; in __dax_fault() local 425 error = copy_user_bh(new_page, &bh, blkbits, vaddr); in __dax_fault() 427 clear_user_highpage(new_page, vaddr); in __dax_fault()
|
/linux-4.4.14/drivers/tty/serial/ |
D | icom.c | 361 unsigned char *new_page = NULL; in load_code() local 434 new_page = pci_alloc_consistent(dev, 4096, &temp_pci); in load_code() 436 if (!new_page) { in load_code() 456 new_page[index] = fw->data[index]; in load_code() 513 if (new_page != NULL) in load_code() 514 pci_free_consistent(dev, 4096, new_page, temp_pci); in load_code()
|
/linux-4.4.14/drivers/staging/rts5208/ |
D | xd.c | 1200 u32 old_page, new_page; in xd_copy_page() local 1218 new_page = (new_blk << xd_card->block_shift) + start_page; in xd_copy_page() 1290 xd_assign_phy_addr(chip, new_page, XD_RW_ADDR); in xd_copy_page() 1314 new_page++; in xd_copy_page()
|
/linux-4.4.14/drivers/net/vmxnet3/ |
D | vmxnet3_drv.c | 1251 struct page *new_page = NULL; in vmxnet3_rq_rx_complete() local 1370 new_page = alloc_page(GFP_ATOMIC); in vmxnet3_rq_rx_complete() 1376 if (unlikely(!new_page)) { in vmxnet3_rq_rx_complete() 1384 new_page, in vmxnet3_rq_rx_complete() 1389 put_page(new_page); in vmxnet3_rq_rx_complete() 1404 rbi->page = new_page; in vmxnet3_rq_rx_complete()
|