walk_page_range 106 arch/openrisc/kernel/dma.c if (walk_page_range(&init_mm, va, va + size, &set_nocache_walk_ops, walk_page_range 122 arch/openrisc/kernel/dma.c WARN_ON(walk_page_range(&init_mm, va, va + size, walk_page_range 2550 arch/s390/mm/gmap.c walk_page_range(mm, 0, TASK_SIZE, &zap_zero_walk_ops, NULL); walk_page_range 2619 arch/s390/mm/gmap.c walk_page_range(mm, 0, TASK_SIZE, &enable_skey_walk_ops, NULL); walk_page_range 2644 arch/s390/mm/gmap.c walk_page_range(mm, 0, TASK_SIZE, &reset_cmma_walk_ops, NULL); walk_page_range 1227 fs/proc/task_mmu.c walk_page_range(mm, 0, mm->highest_vm_end, &clear_refs_walk_ops, walk_page_range 1596 fs/proc/task_mmu.c ret = walk_page_range(mm, start_vaddr, end, &pagemap_ops, &pm); walk_page_range 60 include/linux/pagewalk.h int walk_page_range(struct mm_struct *mm, unsigned long start, walk_page_range 927 mm/hmm.c walk_page_range(vma->vm_mm, start, end, &hmm_walk_ops, walk_page_range 931 mm/hmm.c ret = walk_page_range(vma->vm_mm, start, end, walk_page_range 263 mm/madvise.c walk_page_range(vma->vm_mm, start, end, &swapin_walk_ops, vma); walk_page_range 487 mm/madvise.c walk_page_range(vma->vm_mm, addr, end, &cold_walk_ops, &walk_private); walk_page_range 520 mm/madvise.c walk_page_range(vma->vm_mm, addr, end, &cold_walk_ops, &walk_private); walk_page_range 725 mm/madvise.c walk_page_range(vma->vm_mm, range.start, range.end, walk_page_range 5719 mm/memcontrol.c walk_page_range(mm, 0, mm->highest_vm_end, &precharge_walk_ops, NULL); walk_page_range 6019 mm/memcontrol.c walk_page_range(mc.mm, 0, mc.mm->highest_vm_end, &charge_walk_ops, walk_page_range 691 mm/mempolicy.c return walk_page_range(mm, start, end, &queue_pages_walk_ops, &qp); walk_page_range 2373 mm/migrate.c walk_page_range(migrate->vma->vm_mm, migrate->start, migrate->end, walk_page_range 222 mm/mincore.c err = walk_page_range(vma->vm_mm, addr, end, &mincore_walk_ops, vec); walk_page_range 399 mm/mprotect.c error = walk_page_range(current->mm, start, end,