Lines Matching refs:start

360 		struct vm_area_struct *vma, int zoneid,	unsigned long start,  in __munlock_pagevec_fill()  argument
371 pte = get_locked_pte(vma->vm_mm, start, &ptl); in __munlock_pagevec_fill()
373 end = pgd_addr_end(start, end); in __munlock_pagevec_fill()
374 end = pud_addr_end(start, end); in __munlock_pagevec_fill()
375 end = pmd_addr_end(start, end); in __munlock_pagevec_fill()
378 start += PAGE_SIZE; in __munlock_pagevec_fill()
379 while (start < end) { in __munlock_pagevec_fill()
383 page = vm_normal_page(vma, start, *pte); in __munlock_pagevec_fill()
396 start += PAGE_SIZE; in __munlock_pagevec_fill()
401 return start; in __munlock_pagevec_fill()
423 unsigned long start, unsigned long end) in munlock_vma_pages_range() argument
427 while (start < end) { in munlock_vma_pages_range()
443 page = follow_page_mask(vma, start, FOLL_GET | FOLL_DUMP, in munlock_vma_pages_range()
474 start = __munlock_pagevec_fill(&pvec, vma, in munlock_vma_pages_range()
475 zoneid, start, end); in munlock_vma_pages_range()
481 VM_BUG_ON((start >> PAGE_SHIFT) & page_mask); in munlock_vma_pages_range()
483 start += page_increm * PAGE_SIZE; in munlock_vma_pages_range()
499 unsigned long start, unsigned long end, vm_flags_t newflags) in mlock_fixup() argument
512 pgoff = vma->vm_pgoff + ((start - vma->vm_start) >> PAGE_SHIFT); in mlock_fixup()
513 *prev = vma_merge(mm, *prev, start, end, newflags, vma->anon_vma, in mlock_fixup()
521 if (start != vma->vm_start) { in mlock_fixup()
522 ret = split_vma(mm, vma, start, 1); in mlock_fixup()
537 nr_pages = (end - start) >> PAGE_SHIFT; in mlock_fixup()
551 munlock_vma_pages_range(vma, start, end); in mlock_fixup()
558 static int apply_vma_lock_flags(unsigned long start, size_t len, in apply_vma_lock_flags() argument
565 VM_BUG_ON(offset_in_page(start)); in apply_vma_lock_flags()
567 end = start + len; in apply_vma_lock_flags()
568 if (end < start) in apply_vma_lock_flags()
570 if (end == start) in apply_vma_lock_flags()
572 vma = find_vma(current->mm, start); in apply_vma_lock_flags()
573 if (!vma || vma->vm_start > start) in apply_vma_lock_flags()
577 if (start > vma->vm_start) in apply_vma_lock_flags()
580 for (nstart = start ; ; ) { in apply_vma_lock_flags()
607 static int do_mlock(unsigned long start, size_t len, vm_flags_t flags) in do_mlock() argument
618 len = PAGE_ALIGN(len + (offset_in_page(start))); in do_mlock()
619 start &= PAGE_MASK; in do_mlock()
631 error = apply_vma_lock_flags(start, len, flags); in do_mlock()
637 error = __mm_populate(start, len, 0); in do_mlock()
643 SYSCALL_DEFINE2(mlock, unsigned long, start, size_t, len) in SYSCALL_DEFINE2() argument
645 return do_mlock(start, len, VM_LOCKED); in SYSCALL_DEFINE2()
648 SYSCALL_DEFINE3(mlock2, unsigned long, start, size_t, len, int, flags) in SYSCALL_DEFINE3() argument
658 return do_mlock(start, len, vm_flags); in SYSCALL_DEFINE3()
661 SYSCALL_DEFINE2(munlock, unsigned long, start, size_t, len) in SYSCALL_DEFINE2() argument
665 len = PAGE_ALIGN(len + (offset_in_page(start))); in SYSCALL_DEFINE2()
666 start &= PAGE_MASK; in SYSCALL_DEFINE2()
669 ret = apply_vma_lock_flags(start, len, 0); in SYSCALL_DEFINE2()