hmm_vma_walk      227 mm/hmm.c       	struct hmm_vma_walk *hmm_vma_walk = walk->private;
hmm_vma_walk      228 mm/hmm.c       	struct hmm_range *range = hmm_vma_walk->range;
hmm_vma_walk      235 mm/hmm.c       	if (hmm_vma_walk->flags & HMM_FAULT_ALLOW_RETRY)
hmm_vma_walk      259 mm/hmm.c       	struct hmm_vma_walk *hmm_vma_walk = walk->private;
hmm_vma_walk      260 mm/hmm.c       	struct hmm_range *range = hmm_vma_walk->range;
hmm_vma_walk      287 mm/hmm.c       	struct hmm_vma_walk *hmm_vma_walk = walk->private;
hmm_vma_walk      288 mm/hmm.c       	struct hmm_range *range = hmm_vma_walk->range;
hmm_vma_walk      292 mm/hmm.c       	hmm_vma_walk->last = addr;
hmm_vma_walk      313 mm/hmm.c       static inline void hmm_pte_need_fault(const struct hmm_vma_walk *hmm_vma_walk,
hmm_vma_walk      317 mm/hmm.c       	struct hmm_range *range = hmm_vma_walk->range;
hmm_vma_walk      319 mm/hmm.c       	if (hmm_vma_walk->flags & HMM_FAULT_SNAPSHOT)
hmm_vma_walk      357 mm/hmm.c       static void hmm_range_need_fault(const struct hmm_vma_walk *hmm_vma_walk,
hmm_vma_walk      364 mm/hmm.c       	if (hmm_vma_walk->flags & HMM_FAULT_SNAPSHOT) {
hmm_vma_walk      371 mm/hmm.c       		hmm_pte_need_fault(hmm_vma_walk, pfns[i], cpu_flags,
hmm_vma_walk      381 mm/hmm.c       	struct hmm_vma_walk *hmm_vma_walk = walk->private;
hmm_vma_walk      382 mm/hmm.c       	struct hmm_range *range = hmm_vma_walk->range;
hmm_vma_walk      390 mm/hmm.c       	hmm_range_need_fault(hmm_vma_walk, pfns, npages,
hmm_vma_walk      408 mm/hmm.c       	struct hmm_vma_walk *hmm_vma_walk = walk->private;
hmm_vma_walk      409 mm/hmm.c       	struct hmm_range *range = hmm_vma_walk->range;
hmm_vma_walk      416 mm/hmm.c       	hmm_range_need_fault(hmm_vma_walk, pfns, npages, cpu_flags,
hmm_vma_walk      425 mm/hmm.c       			hmm_vma_walk->pgmap = get_dev_pagemap(pfn,
hmm_vma_walk      426 mm/hmm.c       					      hmm_vma_walk->pgmap);
hmm_vma_walk      427 mm/hmm.c       			if (unlikely(!hmm_vma_walk->pgmap))
hmm_vma_walk      432 mm/hmm.c       	if (hmm_vma_walk->pgmap) {
hmm_vma_walk      433 mm/hmm.c       		put_dev_pagemap(hmm_vma_walk->pgmap);
hmm_vma_walk      434 mm/hmm.c       		hmm_vma_walk->pgmap = NULL;
hmm_vma_walk      436 mm/hmm.c       	hmm_vma_walk->last = end;
hmm_vma_walk      458 mm/hmm.c       	struct hmm_vma_walk *hmm_vma_walk = walk->private;
hmm_vma_walk      459 mm/hmm.c       	struct hmm_range *range = hmm_vma_walk->range;
hmm_vma_walk      469 mm/hmm.c       		hmm_pte_need_fault(hmm_vma_walk, orig_pfn, 0,
hmm_vma_walk      481 mm/hmm.c       			hmm_pte_need_fault(hmm_vma_walk, orig_pfn, cpu_flags,
hmm_vma_walk      497 mm/hmm.c       			hmm_pte_need_fault(hmm_vma_walk, orig_pfn, cpu_flags,
hmm_vma_walk      510 mm/hmm.c       				hmm_vma_walk->last = addr;
hmm_vma_walk      522 mm/hmm.c       		hmm_pte_need_fault(hmm_vma_walk, orig_pfn, cpu_flags,
hmm_vma_walk      530 mm/hmm.c       		hmm_vma_walk->pgmap = get_dev_pagemap(pte_pfn(pte),
hmm_vma_walk      531 mm/hmm.c       					      hmm_vma_walk->pgmap);
hmm_vma_walk      532 mm/hmm.c       		if (unlikely(!hmm_vma_walk->pgmap))
hmm_vma_walk      543 mm/hmm.c       	if (hmm_vma_walk->pgmap) {
hmm_vma_walk      544 mm/hmm.c       		put_dev_pagemap(hmm_vma_walk->pgmap);
hmm_vma_walk      545 mm/hmm.c       		hmm_vma_walk->pgmap = NULL;
hmm_vma_walk      557 mm/hmm.c       	struct hmm_vma_walk *hmm_vma_walk = walk->private;
hmm_vma_walk      558 mm/hmm.c       	struct hmm_range *range = hmm_vma_walk->range;
hmm_vma_walk      578 mm/hmm.c       		hmm_range_need_fault(hmm_vma_walk, pfns, npages,
hmm_vma_walk      581 mm/hmm.c       			hmm_vma_walk->last = addr;
hmm_vma_walk      625 mm/hmm.c       			hmm_vma_walk->last = addr;
hmm_vma_walk      629 mm/hmm.c       	if (hmm_vma_walk->pgmap) {
hmm_vma_walk      636 mm/hmm.c       		put_dev_pagemap(hmm_vma_walk->pgmap);
hmm_vma_walk      637 mm/hmm.c       		hmm_vma_walk->pgmap = NULL;
hmm_vma_walk      641 mm/hmm.c       	hmm_vma_walk->last = addr;
hmm_vma_walk      659 mm/hmm.c       	struct hmm_vma_walk *hmm_vma_walk = walk->private;
hmm_vma_walk      660 mm/hmm.c       	struct hmm_range *range = hmm_vma_walk->range;
hmm_vma_walk      684 mm/hmm.c       		hmm_range_need_fault(hmm_vma_walk, pfns, npages,
hmm_vma_walk      692 mm/hmm.c       			hmm_vma_walk->pgmap = get_dev_pagemap(pfn,
hmm_vma_walk      693 mm/hmm.c       					      hmm_vma_walk->pgmap);
hmm_vma_walk      694 mm/hmm.c       			if (unlikely(!hmm_vma_walk->pgmap))
hmm_vma_walk      699 mm/hmm.c       		if (hmm_vma_walk->pgmap) {
hmm_vma_walk      700 mm/hmm.c       			put_dev_pagemap(hmm_vma_walk->pgmap);
hmm_vma_walk      701 mm/hmm.c       			hmm_vma_walk->pgmap = NULL;
hmm_vma_walk      703 mm/hmm.c       		hmm_vma_walk->last = end;
hmm_vma_walk      731 mm/hmm.c       	struct hmm_vma_walk *hmm_vma_walk = walk->private;
hmm_vma_walk      732 mm/hmm.c       	struct hmm_range *range = hmm_vma_walk->range;
hmm_vma_walk      748 mm/hmm.c       	hmm_pte_need_fault(hmm_vma_walk, orig_pfn, cpu_flags,
hmm_vma_walk      759 mm/hmm.c       	hmm_vma_walk->last = end;
hmm_vma_walk      894 mm/hmm.c       	struct hmm_vma_walk hmm_vma_walk;
hmm_vma_walk      921 mm/hmm.c       		hmm_vma_walk.pgmap = NULL;
hmm_vma_walk      922 mm/hmm.c       		hmm_vma_walk.last = start;
hmm_vma_walk      923 mm/hmm.c       		hmm_vma_walk.flags = flags;
hmm_vma_walk      924 mm/hmm.c       		hmm_vma_walk.range = range;
hmm_vma_walk      928 mm/hmm.c       				&hmm_vma_walk);
hmm_vma_walk      932 mm/hmm.c       					&hmm_walk_ops, &hmm_vma_walk);
hmm_vma_walk      933 mm/hmm.c       			start = hmm_vma_walk.last;
hmm_vma_walk      941 mm/hmm.c       			i = (hmm_vma_walk.last - range->start) >> PAGE_SHIFT;
hmm_vma_walk      943 mm/hmm.c       				hmm_vma_walk.last, range->end);
hmm_vma_walk      950 mm/hmm.c       	return (hmm_vma_walk.last - range->start) >> PAGE_SHIFT;