vmas 343 drivers/gpu/drm/msm/msm_gem.c list_add_tail(&vma->list, &msm_obj->vmas); vmas 356 drivers/gpu/drm/msm/msm_gem.c list_for_each_entry(vma, &msm_obj->vmas, list) { vmas 382 drivers/gpu/drm/msm/msm_gem.c list_for_each_entry_safe(vma, tmp, &msm_obj->vmas, list) { vmas 831 drivers/gpu/drm/msm/msm_gem.c if (!list_empty(&msm_obj->vmas)) { vmas 835 drivers/gpu/drm/msm/msm_gem.c list_for_each_entry(vma, &msm_obj->vmas, list) vmas 1007 drivers/gpu/drm/msm/msm_gem.c INIT_LIST_HEAD(&msm_obj->vmas); vmas 74 drivers/gpu/drm/msm/msm_gem.h struct list_head vmas; /* list of msm_gem_vma */ vmas 343 drivers/vfio/vfio_iommu_type1.c struct vm_area_struct *vmas[1]; vmas 353 drivers/vfio/vfio_iommu_type1.c vmas); vmas 356 drivers/vfio/vfio_iommu_type1.c vmas, NULL); vmas 364 drivers/vfio/vfio_iommu_type1.c if (ret > 0 && vma_is_fsdax(vmas[0])) { vmas 210 drivers/video/fbdev/vermilion/vermilion.h atomic_t vmas; vmas 3449 fs/io_uring.c struct vm_area_struct **vmas = NULL; vmas 3501 fs/io_uring.c kvfree(vmas); vmas 3505 fs/io_uring.c vmas = kvmalloc_array(nr_pages, vmas 3508 fs/io_uring.c if (!pages || !vmas) { vmas 3530 fs/io_uring.c pages, vmas); vmas 3534 fs/io_uring.c struct vm_area_struct *vma = vmas[j]; vmas 3579 fs/io_uring.c kvfree(vmas); vmas 3583 fs/io_uring.c kvfree(vmas); vmas 1535 include/linux/mm.h struct vm_area_struct **vmas, int *locked); vmas 1538 include/linux/mm.h struct vm_area_struct **vmas); vmas 36 include/linux/mm_types_task.h struct vm_area_struct *vmas[VMACACHE_SIZE]; vmas 10 include/linux/vmacache.h memset(tsk->vmacache.vmas, 0, sizeof(tsk->vmacache.vmas)); vmas 293 kernel/debug/debug_core.c if (!current->vmacache.vmas[i]) vmas 295 kernel/debug/debug_core.c flush_cache_range(current->vmacache.vmas[i], vmas 791 mm/gup.c struct vm_area_struct **vmas, int *nonblocking) vmas 837 mm/gup.c i = follow_hugetlb_page(mm, vma, pages, vmas, vmas 893 mm/gup.c if (vmas) { vmas 894 mm/gup.c vmas[i] = vma; vmas 1018 mm/gup.c struct vm_area_struct **vmas, vmas 1027 mm/gup.c BUG_ON(vmas); vmas 1039 mm/gup.c vmas, locked); vmas 1167 mm/gup.c struct vm_area_struct **vmas, int *locked) vmas 1178 mm/gup.c return __get_user_pages_locked(tsk, mm, start, nr_pages, pages, vmas, vmas 1334 mm/gup.c struct vm_area_struct **vmas, int *locked, vmas 1364 mm/gup.c if (vmas) vmas 1365 mm/gup.c vmas[i] = vma; vmas 1377 mm/gup.c static bool check_dax_vmas(struct vm_area_struct **vmas, long nr_pages) vmas 1383 mm/gup.c struct vm_area_struct *vma = vmas[i]; vmas 1453 mm/gup.c struct vm_area_struct **vmas, vmas 1523 mm/gup.c pages, vmas, NULL, vmas 1540 mm/gup.c struct vm_area_struct **vmas, vmas 1556 mm/gup.c struct vm_area_struct **vmas, vmas 1559 mm/gup.c struct vm_area_struct **vmas_tmp = vmas; vmas 1597 mm/gup.c if (vmas_tmp != vmas) vmas 1607 mm/gup.c struct vm_area_struct **vmas, vmas 1610 mm/gup.c return __get_user_pages_locked(tsk, mm, start, nr_pages, pages, vmas, vmas 1624 mm/gup.c struct vm_area_struct **vmas) vmas 1627 mm/gup.c pages, vmas, gup_flags | FOLL_TOUCH); vmas 4388 mm/hugetlb.c struct page **pages, struct vm_area_struct **vmas, vmas 4517 mm/hugetlb.c if (vmas) vmas 4518 mm/hugetlb.c vmas[i] = vma; vmas 660 mm/nommu.c if (curr->vmacache.vmas[i] == vma) { vmas 39 mm/vmacache.c current->vmacache.vmas[VMACACHE_HASH(addr)] = newvma; vmas 73 mm/vmacache.c struct vm_area_struct *vma = current->vmacache.vmas[idx]; vmas 106 mm/vmacache.c struct vm_area_struct *vma = current->vmacache.vmas[idx];