hmm_vma_walk 227 mm/hmm.c struct hmm_vma_walk *hmm_vma_walk = walk->private; hmm_vma_walk 228 mm/hmm.c struct hmm_range *range = hmm_vma_walk->range; hmm_vma_walk 235 mm/hmm.c if (hmm_vma_walk->flags & HMM_FAULT_ALLOW_RETRY) hmm_vma_walk 259 mm/hmm.c struct hmm_vma_walk *hmm_vma_walk = walk->private; hmm_vma_walk 260 mm/hmm.c struct hmm_range *range = hmm_vma_walk->range; hmm_vma_walk 287 mm/hmm.c struct hmm_vma_walk *hmm_vma_walk = walk->private; hmm_vma_walk 288 mm/hmm.c struct hmm_range *range = hmm_vma_walk->range; hmm_vma_walk 292 mm/hmm.c hmm_vma_walk->last = addr; hmm_vma_walk 313 mm/hmm.c static inline void hmm_pte_need_fault(const struct hmm_vma_walk *hmm_vma_walk, hmm_vma_walk 317 mm/hmm.c struct hmm_range *range = hmm_vma_walk->range; hmm_vma_walk 319 mm/hmm.c if (hmm_vma_walk->flags & HMM_FAULT_SNAPSHOT) hmm_vma_walk 357 mm/hmm.c static void hmm_range_need_fault(const struct hmm_vma_walk *hmm_vma_walk, hmm_vma_walk 364 mm/hmm.c if (hmm_vma_walk->flags & HMM_FAULT_SNAPSHOT) { hmm_vma_walk 371 mm/hmm.c hmm_pte_need_fault(hmm_vma_walk, pfns[i], cpu_flags, hmm_vma_walk 381 mm/hmm.c struct hmm_vma_walk *hmm_vma_walk = walk->private; hmm_vma_walk 382 mm/hmm.c struct hmm_range *range = hmm_vma_walk->range; hmm_vma_walk 390 mm/hmm.c hmm_range_need_fault(hmm_vma_walk, pfns, npages, hmm_vma_walk 408 mm/hmm.c struct hmm_vma_walk *hmm_vma_walk = walk->private; hmm_vma_walk 409 mm/hmm.c struct hmm_range *range = hmm_vma_walk->range; hmm_vma_walk 416 mm/hmm.c hmm_range_need_fault(hmm_vma_walk, pfns, npages, cpu_flags, hmm_vma_walk 425 mm/hmm.c hmm_vma_walk->pgmap = get_dev_pagemap(pfn, hmm_vma_walk 426 mm/hmm.c hmm_vma_walk->pgmap); hmm_vma_walk 427 mm/hmm.c if (unlikely(!hmm_vma_walk->pgmap)) hmm_vma_walk 432 mm/hmm.c if (hmm_vma_walk->pgmap) { hmm_vma_walk 433 mm/hmm.c put_dev_pagemap(hmm_vma_walk->pgmap); hmm_vma_walk 434 mm/hmm.c hmm_vma_walk->pgmap = NULL; hmm_vma_walk 436 mm/hmm.c hmm_vma_walk->last = end; hmm_vma_walk 458 mm/hmm.c struct hmm_vma_walk *hmm_vma_walk = walk->private; hmm_vma_walk 459 mm/hmm.c struct hmm_range *range = hmm_vma_walk->range; hmm_vma_walk 469 mm/hmm.c hmm_pte_need_fault(hmm_vma_walk, orig_pfn, 0, hmm_vma_walk 481 mm/hmm.c hmm_pte_need_fault(hmm_vma_walk, orig_pfn, cpu_flags, hmm_vma_walk 497 mm/hmm.c hmm_pte_need_fault(hmm_vma_walk, orig_pfn, cpu_flags, hmm_vma_walk 510 mm/hmm.c hmm_vma_walk->last = addr; hmm_vma_walk 522 mm/hmm.c hmm_pte_need_fault(hmm_vma_walk, orig_pfn, cpu_flags, hmm_vma_walk 530 mm/hmm.c hmm_vma_walk->pgmap = get_dev_pagemap(pte_pfn(pte), hmm_vma_walk 531 mm/hmm.c hmm_vma_walk->pgmap); hmm_vma_walk 532 mm/hmm.c if (unlikely(!hmm_vma_walk->pgmap)) hmm_vma_walk 543 mm/hmm.c if (hmm_vma_walk->pgmap) { hmm_vma_walk 544 mm/hmm.c put_dev_pagemap(hmm_vma_walk->pgmap); hmm_vma_walk 545 mm/hmm.c hmm_vma_walk->pgmap = NULL; hmm_vma_walk 557 mm/hmm.c struct hmm_vma_walk *hmm_vma_walk = walk->private; hmm_vma_walk 558 mm/hmm.c struct hmm_range *range = hmm_vma_walk->range; hmm_vma_walk 578 mm/hmm.c hmm_range_need_fault(hmm_vma_walk, pfns, npages, hmm_vma_walk 581 mm/hmm.c hmm_vma_walk->last = addr; hmm_vma_walk 625 mm/hmm.c hmm_vma_walk->last = addr; hmm_vma_walk 629 mm/hmm.c if (hmm_vma_walk->pgmap) { hmm_vma_walk 636 mm/hmm.c put_dev_pagemap(hmm_vma_walk->pgmap); hmm_vma_walk 637 mm/hmm.c hmm_vma_walk->pgmap = NULL; hmm_vma_walk 641 mm/hmm.c hmm_vma_walk->last = addr; hmm_vma_walk 659 mm/hmm.c struct hmm_vma_walk *hmm_vma_walk = walk->private; hmm_vma_walk 660 mm/hmm.c struct hmm_range *range = hmm_vma_walk->range; hmm_vma_walk 684 mm/hmm.c hmm_range_need_fault(hmm_vma_walk, pfns, npages, hmm_vma_walk 692 mm/hmm.c hmm_vma_walk->pgmap = get_dev_pagemap(pfn, hmm_vma_walk 693 mm/hmm.c hmm_vma_walk->pgmap); hmm_vma_walk 694 mm/hmm.c if (unlikely(!hmm_vma_walk->pgmap)) hmm_vma_walk 699 mm/hmm.c if (hmm_vma_walk->pgmap) { hmm_vma_walk 700 mm/hmm.c put_dev_pagemap(hmm_vma_walk->pgmap); hmm_vma_walk 701 mm/hmm.c hmm_vma_walk->pgmap = NULL; hmm_vma_walk 703 mm/hmm.c hmm_vma_walk->last = end; hmm_vma_walk 731 mm/hmm.c struct hmm_vma_walk *hmm_vma_walk = walk->private; hmm_vma_walk 732 mm/hmm.c struct hmm_range *range = hmm_vma_walk->range; hmm_vma_walk 748 mm/hmm.c hmm_pte_need_fault(hmm_vma_walk, orig_pfn, cpu_flags, hmm_vma_walk 759 mm/hmm.c hmm_vma_walk->last = end; hmm_vma_walk 894 mm/hmm.c struct hmm_vma_walk hmm_vma_walk; hmm_vma_walk 921 mm/hmm.c hmm_vma_walk.pgmap = NULL; hmm_vma_walk 922 mm/hmm.c hmm_vma_walk.last = start; hmm_vma_walk 923 mm/hmm.c hmm_vma_walk.flags = flags; hmm_vma_walk 924 mm/hmm.c hmm_vma_walk.range = range; hmm_vma_walk 928 mm/hmm.c &hmm_vma_walk); hmm_vma_walk 932 mm/hmm.c &hmm_walk_ops, &hmm_vma_walk); hmm_vma_walk 933 mm/hmm.c start = hmm_vma_walk.last; hmm_vma_walk 941 mm/hmm.c i = (hmm_vma_walk.last - range->start) >> PAGE_SHIFT; hmm_vma_walk 943 mm/hmm.c hmm_vma_walk.last, range->end); hmm_vma_walk 950 mm/hmm.c return (hmm_vma_walk.last - range->start) >> PAGE_SHIFT;