Lines Matching refs:mm
23 void task_mem(struct seq_file *m, struct mm_struct *mm) in task_mem() argument
35 hiwater_vm = total_vm = mm->total_vm; in task_mem()
36 if (hiwater_vm < mm->hiwater_vm) in task_mem()
37 hiwater_vm = mm->hiwater_vm; in task_mem()
38 hiwater_rss = total_rss = get_mm_rss(mm); in task_mem()
39 if (hiwater_rss < mm->hiwater_rss) in task_mem()
40 hiwater_rss = mm->hiwater_rss; in task_mem()
42 data = mm->total_vm - mm->shared_vm - mm->stack_vm; in task_mem()
43 text = (PAGE_ALIGN(mm->end_code) - (mm->start_code & PAGE_MASK)) >> 10; in task_mem()
44 lib = (mm->exec_vm << (PAGE_SHIFT-10)) - text; in task_mem()
45 swap = get_mm_counter(mm, MM_SWAPENTS); in task_mem()
46 ptes = PTRS_PER_PTE * sizeof(pte_t) * atomic_long_read(&mm->nr_ptes); in task_mem()
47 pmds = PTRS_PER_PMD * sizeof(pmd_t) * mm_nr_pmds(mm); in task_mem()
64 mm->locked_vm << (PAGE_SHIFT-10), in task_mem()
65 mm->pinned_vm << (PAGE_SHIFT-10), in task_mem()
69 mm->stack_vm << (PAGE_SHIFT-10), text, lib, in task_mem()
73 hugetlb_report_usage(m, mm); in task_mem()
76 unsigned long task_vsize(struct mm_struct *mm) in task_vsize() argument
78 return PAGE_SIZE * mm->total_vm; in task_vsize()
81 unsigned long task_statm(struct mm_struct *mm, in task_statm() argument
85 *shared = get_mm_counter(mm, MM_FILEPAGES); in task_statm()
86 *text = (PAGE_ALIGN(mm->end_code) - (mm->start_code & PAGE_MASK)) in task_statm()
88 *data = mm->total_vm - mm->shared_vm; in task_statm()
89 *resident = *shared + get_mm_counter(mm, MM_ANONPAGES); in task_statm()
90 return mm->total_vm; in task_statm()
121 struct mm_struct *mm = priv->mm; in vma_stop() local
124 up_read(&mm->mmap_sem); in vma_stop()
125 mmput(mm); in vma_stop()
146 struct mm_struct *mm; in m_start() local
158 mm = priv->mm; in m_start()
159 if (!mm || !atomic_inc_not_zero(&mm->mm_users)) in m_start()
162 down_read(&mm->mmap_sem); in m_start()
164 priv->tail_vma = get_gate_vma(mm); in m_start()
167 vma = find_vma(mm, last_addr); in m_start()
173 if (pos < mm->map_count) { in m_start()
174 for (vma = mm->mmap; pos; pos--) { in m_start()
182 if (pos == mm->map_count && priv->tail_vma) in m_start()
222 priv->mm = proc_mem_open(inode, PTRACE_MODE_READ); in proc_maps_open()
223 if (IS_ERR(priv->mm)) { in proc_maps_open()
224 int err = PTR_ERR(priv->mm); in proc_maps_open()
238 if (priv->mm) in proc_map_release()
239 mmdrop(priv->mm); in proc_map_release()
273 struct mm_struct *mm = vma->vm_mm; in show_map_vma() local
329 if (!mm) { in show_map_vma()
334 if (vma->vm_start <= mm->brk && in show_map_vma()
335 vma->vm_end >= mm->start_brk) { in show_map_vma()
346 if (!is_pid || (vma->vm_start <= mm->start_stack && in show_map_vma()
347 vma->vm_end >= mm->start_stack)) { in show_map_vma()
669 .mm = vma->vm_mm, in show_smap()
914 struct mm_struct *mm; in clear_refs_write() local
935 mm = get_task_mm(task); in clear_refs_write()
936 if (mm) { in clear_refs_write()
943 .mm = mm, in clear_refs_write()
952 down_write(&mm->mmap_sem); in clear_refs_write()
953 reset_mm_hiwater_rss(mm); in clear_refs_write()
954 up_write(&mm->mmap_sem); in clear_refs_write()
958 down_read(&mm->mmap_sem); in clear_refs_write()
960 for (vma = mm->mmap; vma; vma = vma->vm_next) { in clear_refs_write()
963 up_read(&mm->mmap_sem); in clear_refs_write()
964 down_write(&mm->mmap_sem); in clear_refs_write()
965 for (vma = mm->mmap; vma; vma = vma->vm_next) { in clear_refs_write()
969 downgrade_write(&mm->mmap_sem); in clear_refs_write()
972 mmu_notifier_invalidate_range_start(mm, 0, -1); in clear_refs_write()
976 mmu_notifier_invalidate_range_end(mm, 0, -1); in clear_refs_write()
977 flush_tlb_mm(mm); in clear_refs_write()
978 up_read(&mm->mmap_sem); in clear_refs_write()
980 mmput(mm); in clear_refs_write()
1038 struct vm_area_struct *vma = find_vma(walk->mm, addr); in pagemap_pte_hole()
1161 orig_pte = pte = pte_offset_map_lock(walk->mm, pmdp, addr, &ptl); in pagemap_pmd_range()
1253 struct mm_struct *mm = file->private_data; in pagemap_read() local
1262 if (!mm || !atomic_inc_not_zero(&mm->mm_users)) in pagemap_read()
1288 pagemap_walk.mm = mm; in pagemap_read()
1294 end_vaddr = mm->task_size; in pagemap_read()
1297 if (svpfn > mm->task_size >> PAGE_SHIFT) in pagemap_read()
1316 down_read(&mm->mmap_sem); in pagemap_read()
1318 up_read(&mm->mmap_sem); in pagemap_read()
1337 mmput(mm); in pagemap_read()
1344 struct mm_struct *mm; in pagemap_open() local
1346 mm = proc_mem_open(inode, PTRACE_MODE_READ); in pagemap_open()
1347 if (IS_ERR(mm)) in pagemap_open()
1348 return PTR_ERR(mm); in pagemap_open()
1349 file->private_data = mm; in pagemap_open()
1355 struct mm_struct *mm = file->private_data; in pagemap_release() local
1357 if (mm) in pagemap_release()
1358 mmdrop(mm); in pagemap_release()
1488 orig_pte = pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); in gather_pte_stats()
1537 struct mm_struct *mm = vma->vm_mm; in show_numa_map() local
1542 .mm = mm, in show_numa_map()
1548 if (!mm) in show_numa_map()
1567 } else if (vma->vm_start <= mm->brk && vma->vm_end >= mm->start_brk) { in show_numa_map()
1576 if (!is_pid || (vma->vm_start <= mm->start_stack && in show_numa_map()
1577 vma->vm_end >= mm->start_stack)) in show_numa_map()